-
Notifications
You must be signed in to change notification settings - Fork 0
/
Copy pathstorage.yaml
109 lines (103 loc) · 3.35 KB
/
storage.yaml
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
## These values can be used to override the storage and persistence options of
## all the stateful charts used in Zenko. While all the options listed are the
## default values, this file can be changed and applied only at chart
## installation time by specifying `-f storage.yaml`. Values cannot be changed
## post installation.
mongodb-replicaset:
## Notes:
## - number of PVs required is `nodeCount`
## - Sizing is generally based off max object counts
## e.g. (Number of objects * 0.01) MB is typically used for sizing
## objects with the default amount of metadata
##
persistentVolume:
enabled: true
## mongodb-replicaset data Persistent Volume Storage Class
## If defined, storageClassName: <storageClass>
## If set to "-", storageClassName: "", which disables dynamic provisioning
## If undefined (the default) or set to null, no storageClassName spec is
## set, choosing the default provisioner. (gp2 on AWS, standard on
## GKE, AWS & OpenStack)
##
storageClass: null
## Must match those of existing PV or dynamic provisioner
## Ref: http://kubernetes.io/docs/user-guide/persistent-volumes/
##
accessModes:
- ReadWriteOnce
size: 50Gi
annotations: {}
prometheus:
server:
## Notes:
## - number of PVs required is 2 by default
## - This is used for storing application level metrics
## - Retention policies will attempt to keep 15 days worth of metrics
## if the volume isn't full
##
persistentVolume:
enabled: true
accessModes:
- ReadWriteOnce
annotations: {}
mountPath: /data
size: 8Gi
storageClass: null
## Subdirectory of Prometheus server data Persistent Volume to mount
## Useful if the volume's root directory is not empty
##
subPath: ""
redis-ha:
## Notes:
## - number of PVs required is `nodeCount`
## - in-memory key-value store vital to the replication services
##
persistentVolume:
enabled: true
storageClass: null
accessModes:
- ReadWriteOnce
size: 10Gi
annotations: {}
s3-data:
## Notes:
## - number of PVs required is 1
## - this storage is used for the Zenko transient source (data that can be
## stored locally prior to replication to an external resource)
## - storing files locally on Zenko is mostly limited by the s3-data volume
##
persistentVolume:
enabled: true
accessModes:
- ReadWriteOnce
annotations: {}
existingClaim: ""
size: 90Gi
storageClass: null
zenko-queue:
## Notes:
## - number of PVs required is `nodeCount`
## - Rention is 7 days by default which will generally use at least 300MB
## per pod for lighter work loads
##
persistence:
enabled: true
## The size of the PersistentVolume to allocate to each Kafka Pod in the
## StatefulSet. For production servers this should be much larger.
##
size: 20Gi
## The location within the Kafka container where the PV will mount its
## storage and Kafka will store its logs.
##
mountPath: "/opt/kafka/data"
storageClass: null
zenko-quorum:
## Notes:
## - number of PVs required is `nodeCount`
## - disk performance is vital here and SSD drives are highly recommended
##
persistence:
enabled: true
storageClass: null
accessMode: ReadWriteOnce
size: 5Gi