generated from tomkerkhove/template
-
-
Notifications
You must be signed in to change notification settings - Fork 24
/
Copy pathvalues.yaml
245 lines (219 loc) · 5.88 KB
/
values.yaml
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
180
181
182
183
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
205
206
207
208
209
210
211
212
213
214
215
216
217
218
219
220
221
222
223
224
225
226
227
228
229
230
231
232
233
234
235
236
237
238
239
240
241
242
243
244
245
# Default values for promitor-agent-scraper.
# This is a YAML-formatted file.
replicaCount: 1
nameOverride: ""
fullnameOverride: ""
image:
repository: containers.promitor.io/tomkerkhove/promitor-agent-scraper
pullPolicy: Always
pullSecrets: []
tag:
azureAuthentication:
appId: "" # [Deprecated] Prefer identity.id
appKey: "" # [Deprecated] Prefer identity.key
mode: "ServicePrincipal"
identity:
id: ""
key: ""
binding: ""
metricSinks:
atlassianStatuspage:
enabled: false
pageId: ""
apiKey: ""
systemMetricMapping: []
# - id: <atlassian-statuspage-system-metric-id>
# promitorMetricName: <promitor-metric-name>
openTelemetryCollector:
enabled: false
collectorUri: ""
prometheusScrapingEndpoint:
enabled: true
baseUriPath: /metrics
enableMetricTimestamps: true
metricUnavailableValue: NaN
labelTransformation: None
enableServiceDiscovery: true
serviceMonitor:
enabled: false
namespace: ""
labels: {}
interval: 60s
timeout: 10s
metricRelabelings: []
relabelings: []
statsd:
enabled: false
host: ""
port: 8125
metricPrefix: ""
prometheusRule:
enabled: false
namespace: ""
additionalLabels: {}
interval: ""
rules: []
# Sample rule below. Can also use templated strings with some limitations regarding possible line length: https://github.com/go-yaml/yaml/issues/166
# To prevent Helm from messing up the rules on fields > 80 characters you can use yaml literal style as shown below.
# - alert: PromitorRemainingArmCalls
# expr: |
# promitor_ratelimit_arm{service="{{ template "promitor-agent-scraper.name" . }}"} < 11999
# for: 5m
# labels:
# severity: warning
# annotations:
# description: |
# Service {{ template "promitor-agent-scraper.name" . }} currently reports {{ "{{ $value }}" }} remaining calls before Azure Resource Manager throttles us.
# summary: Azure Resource Manager may throttle us soon.
resourceDiscovery:
enabled: false
host: ""
port: 80
telemetry:
defaultLogLevel: "Error"
applicationInsights:
enabled: false
key: ""
logLevel: ""
containerLogs:
enabled: true
logLevel: ""
## Metric Declaration YAML
azureMetadata:
tenantId: ""
subscriptionId: ""
resourceGroupName: promitor
cloud: "Global"
metricDefaults:
aggregation:
interval: 00:05:00
scraping:
schedule: "*/5 * * * *"
metrics: []
# Sample metric configuration below
# - name: demo_queue_size
# description: "Amount of active messages of the 'myqueue' queue (determined with ServiceBusQueue provider)"
# resourceType: ServiceBusQueue
# azureMetricConfiguration:
# metricName: ActiveMessages
# aggregation:
# type: Total
# interval: 00:15:00
# resources:
# - namespace: promitor-messaging
# queueName: orders
azureMonitor:
logging:
enabled: false
logLevel:
integration:
metricsBatching:
enabled: false
history:
startingFromInHours:
deployment:
env:
extra: []
# - name: AZURE_STORAGE_QUEUE_SAS_TOKEN
# valueFrom:
# secretKeyRef:
# name: azure-storage-queue
# key: sas-token
secrets:
# To use your own secret, set createSecret to false and define the name/keys that your secret uses
createSecret: true
secretName: ""
appKeySecret: azure-app-key
atlassianStatuspageApiKey: atlassian-statuspage-apikey
extra: {}
service:
port: 8888
## By default this pod is running as a non-root user.
## If you choose targetPort <1024 it will fail to start.
targetPort: 5000
loadBalancer:
enabled: false
azure:
dnsPrefix:
exposeInternally: false
health:
readiness:
enabled: true
verifyDependencies: false
delay: 5
interval: 5
timeoutSeconds: 1
thresholds:
failure: 3
success: 1
liveness:
enabled: true
verifyDependencies: false
delay: 5
interval: 30
timeoutSeconds: 1
thresholds:
failure: 3
success: 1
affinity: {}
podLabels: {}
annotations: {}
priorityClassName: ""
resources: {}
# limits:
# cpu: 100m
# memory: 128Mi
# requests:
# cpu: 100m
# memory: 128Mi
## Role-based access control
## https://kubernetes.io/docs/reference/access-authn-authz/rbac/
rbac:
## If true, create & use RBAC resources
create: true
## If true, create & use Pod Security Policy resources
## ref: https://kubernetes.io/docs/concepts/policy/pod-security-policy/
podSecurityPolicyEnabled: false
## Service Account for pods
## ref: https://kubernetes.io/docs/tasks/configure-pod-container/configure-service-account/
serviceAccount:
## Specifies whether a service account should be created
create: true
## The name of the service account to use if create is false
## If create is true, a name is generated using the fullname template
name: promitor-scraper
annotations: {}
## Set this to true if you plan on using Pod Security Policy
automountServiceAccountToken: false
nodeSelector: {}
## securityContext and containerSecurityContext are using secure defaults.
## Only override if you have a good reason to.
securityContext:
enabled: true
runAsGroup: 10000
runAsNonRoot: true
runAsUser: 10000
seccompProfile:
type: RuntimeDefault
containerSecurityContext:
allowPrivilegeEscalation: false
capabilities:
drop:
- ALL
enabled: true
privileged: false
readOnlyRootFilesystem: true
tolerations: []
## Pass extra volumeMounts to the promitor container
extraVolumeMounts: []
# - name: secrets-store-inline
# mountPath: "/mnt/promitor-secrets"
# readOnly: true
## Pass extra volumes to the promitor deployment
extraVolumes: []
# - name: secrets-store-inline
# csi:
# driver: secrets-store.csi.k8s.io
# readOnly: true
# volumeAttributes:
# secretProviderClass: "promitor-secrets"