diff --git a/1.17.5/_gen/postgresql/v1/cluster.libsonnet b/1.17.5/_gen/postgresql/v1/cluster.libsonnet
index 7847c36..71b8f6e 100644
--- a/1.17.5/_gen/postgresql/v1/cluster.libsonnet
+++ b/1.17.5/_gen/postgresql/v1/cluster.libsonnet
@@ -802,6 +802,13 @@
'#withPrefix':: d.fn(help='"An optional identifier to prepend to each key in the ConfigMap. Must be a C_IDENTIFIER."', args=[d.arg(name='prefix', type=d.T.string)]),
withPrefix(prefix): { prefix: prefix },
},
+ '#ephemeralVolumesSizeLimit':: d.obj(help='"EphemeralVolumesSizeLimit allows the user to set the limits for the ephemeral volumes"'),
+ ephemeralVolumesSizeLimit: {
+ '#withShm':: d.fn(help='"Shm is the size limit of the shared memory volume"', args=[d.arg(name='shm', type=d.T.any)]),
+ withShm(shm): { spec+: { ephemeralVolumesSizeLimit+: { shm: shm } } },
+ '#withTemporaryData':: d.fn(help='"TemporaryData is the size limit of the temporary data volume"', args=[d.arg(name='temporaryData', type=d.T.any)]),
+ withTemporaryData(temporaryData): { spec+: { ephemeralVolumesSizeLimit+: { temporaryData: temporaryData } } },
+ },
'#externalClusters':: d.obj(help='"The list of external clusters which are used in the configuration"'),
externalClusters: {
'#barmanObjectStore':: d.obj(help='"The configuration for the barman-cloud tool suite"'),
@@ -1232,7 +1239,7 @@
replicationSlots: {
'#highAvailability':: d.obj(help='"Replication slots for high availability configuration"'),
highAvailability: {
- '#withEnabled':: d.fn(help='"If enabled, the operator will automatically manage replication slots on the primary instance and use them in streaming replication connections with all the standby instances that are part of the HA cluster. If disabled (default), the operator will not take advantage of replication slots in streaming connections with the replicas. This feature also controls replication slots in replica cluster, from the designated primary to its cascading replicas. This can only be set at creation time."', args=[d.arg(name='enabled', type=d.T.boolean)]),
+ '#withEnabled':: d.fn(help='"If enabled (default), the operator will automatically manage replication slots on the primary instance and use them in streaming replication connections with all the standby instances that are part of the HA cluster. If disabled, the operator will not take advantage of replication slots in streaming connections with the replicas. This feature also controls replication slots in replica cluster, from the designated primary to its cascading replicas."', args=[d.arg(name='enabled', type=d.T.boolean)]),
withEnabled(enabled): { spec+: { replicationSlots+: { highAvailability+: { enabled: enabled } } } },
'#withSlotPrefix':: d.fn(help='"Prefix for replication slots managed by the operator for HA. It may only contain lower case letters, numbers, and the underscore character. This can only be set at creation time. By default set to `_cnpg_`."', args=[d.arg(name='slotPrefix', type=d.T.string)]),
withSlotPrefix(slotPrefix): { spec+: { replicationSlots+: { highAvailability+: { slotPrefix: slotPrefix } } } },
@@ -1497,7 +1504,7 @@
},
'#withDescription':: d.fn(help='"Description of this PostgreSQL cluster"', args=[d.arg(name='description', type=d.T.string)]),
withDescription(description): { spec+: { description: description } },
- '#withEnableSuperuserAccess':: d.fn(help='"When this option is enabled, the operator will use the `SuperuserSecret` to update the `postgres` user password (if the secret is not present, the operator will automatically create one). When this option is disabled, the operator will ignore the `SuperuserSecret` content, delete it when automatically created, and then blank the password of the `postgres` user by setting it to `NULL`. Enabled by default."', args=[d.arg(name='enableSuperuserAccess', type=d.T.boolean)]),
+ '#withEnableSuperuserAccess':: d.fn(help='"When this option is enabled, the operator will use the `SuperuserSecret` to update the `postgres` user password (if the secret is not present, the operator will automatically create one). When this option is disabled, the operator will ignore the `SuperuserSecret` content, delete it when automatically created, and then blank the password of the `postgres` user by setting it to `NULL`. Disabled by default."', args=[d.arg(name='enableSuperuserAccess', type=d.T.boolean)]),
withEnableSuperuserAccess(enableSuperuserAccess): { spec+: { enableSuperuserAccess: enableSuperuserAccess } },
'#withEnv':: d.fn(help='"Env follows the Env format to pass environment variables to the pods created in the cluster"', args=[d.arg(name='env', type=d.T.array)]),
withEnv(env): { spec+: { env: if std.isArray(v=env) then env else [env] } },
@@ -1541,8 +1548,8 @@
withPriorityClassName(priorityClassName): { spec+: { priorityClassName: priorityClassName } },
'#withSchedulerName':: d.fn(help='"If specified, the pod will be dispatched by specified Kubernetes scheduler. If not specified, the pod will be dispatched by the default scheduler. More info: https://kubernetes.io/docs/concepts/scheduling-eviction/kube-scheduler/"', args=[d.arg(name='schedulerName', type=d.T.string)]),
withSchedulerName(schedulerName): { spec+: { schedulerName: schedulerName } },
- '#withSmartStopDelay':: d.fn(help='"The time in seconds that controls the window of time reserved for the smart shutdown of Postgres to complete. this formula to compute the timeout of smart shutdown is `max(stopDelay - smartStopDelay, 30)`"', args=[d.arg(name='smartStopDelay', type=d.T.integer)]),
- withSmartStopDelay(smartStopDelay): { spec+: { smartStopDelay: smartStopDelay } },
+ '#withSmartShutdownTimeout':: d.fn(help='"The time in seconds that controls the window of time reserved for the smart shutdown of Postgres to complete. Make sure you reserve enough time for the operator to request a fast shutdown of Postgres (that is: `stopDelay` - `smartShutdownTimeout`)."', args=[d.arg(name='smartShutdownTimeout', type=d.T.integer)]),
+ withSmartShutdownTimeout(smartShutdownTimeout): { spec+: { smartShutdownTimeout: smartShutdownTimeout } },
'#withStartDelay':: d.fn(help='"The time in seconds that is allowed for a PostgreSQL instance to successfully start up (default 3600). The startup probe failure threshold is derived from this value using the formula: ceiling(startDelay / 10)."', args=[d.arg(name='startDelay', type=d.T.integer)]),
withStartDelay(startDelay): { spec+: { startDelay: startDelay } },
'#withStopDelay':: d.fn(help='"The time in seconds that is allowed for a PostgreSQL instance to gracefully shutdown (default 1800)"', args=[d.arg(name='stopDelay', type=d.T.integer)]),
diff --git a/1.18.5/_gen/postgresql/v1/cluster.libsonnet b/1.18.5/_gen/postgresql/v1/cluster.libsonnet
index 7847c36..71b8f6e 100644
--- a/1.18.5/_gen/postgresql/v1/cluster.libsonnet
+++ b/1.18.5/_gen/postgresql/v1/cluster.libsonnet
@@ -802,6 +802,13 @@
'#withPrefix':: d.fn(help='"An optional identifier to prepend to each key in the ConfigMap. Must be a C_IDENTIFIER."', args=[d.arg(name='prefix', type=d.T.string)]),
withPrefix(prefix): { prefix: prefix },
},
+ '#ephemeralVolumesSizeLimit':: d.obj(help='"EphemeralVolumesSizeLimit allows the user to set the limits for the ephemeral volumes"'),
+ ephemeralVolumesSizeLimit: {
+ '#withShm':: d.fn(help='"Shm is the size limit of the shared memory volume"', args=[d.arg(name='shm', type=d.T.any)]),
+ withShm(shm): { spec+: { ephemeralVolumesSizeLimit+: { shm: shm } } },
+ '#withTemporaryData':: d.fn(help='"TemporaryData is the size limit of the temporary data volume"', args=[d.arg(name='temporaryData', type=d.T.any)]),
+ withTemporaryData(temporaryData): { spec+: { ephemeralVolumesSizeLimit+: { temporaryData: temporaryData } } },
+ },
'#externalClusters':: d.obj(help='"The list of external clusters which are used in the configuration"'),
externalClusters: {
'#barmanObjectStore':: d.obj(help='"The configuration for the barman-cloud tool suite"'),
@@ -1232,7 +1239,7 @@
replicationSlots: {
'#highAvailability':: d.obj(help='"Replication slots for high availability configuration"'),
highAvailability: {
- '#withEnabled':: d.fn(help='"If enabled, the operator will automatically manage replication slots on the primary instance and use them in streaming replication connections with all the standby instances that are part of the HA cluster. If disabled (default), the operator will not take advantage of replication slots in streaming connections with the replicas. This feature also controls replication slots in replica cluster, from the designated primary to its cascading replicas. This can only be set at creation time."', args=[d.arg(name='enabled', type=d.T.boolean)]),
+ '#withEnabled':: d.fn(help='"If enabled (default), the operator will automatically manage replication slots on the primary instance and use them in streaming replication connections with all the standby instances that are part of the HA cluster. If disabled, the operator will not take advantage of replication slots in streaming connections with the replicas. This feature also controls replication slots in replica cluster, from the designated primary to its cascading replicas."', args=[d.arg(name='enabled', type=d.T.boolean)]),
withEnabled(enabled): { spec+: { replicationSlots+: { highAvailability+: { enabled: enabled } } } },
'#withSlotPrefix':: d.fn(help='"Prefix for replication slots managed by the operator for HA. It may only contain lower case letters, numbers, and the underscore character. This can only be set at creation time. By default set to `_cnpg_`."', args=[d.arg(name='slotPrefix', type=d.T.string)]),
withSlotPrefix(slotPrefix): { spec+: { replicationSlots+: { highAvailability+: { slotPrefix: slotPrefix } } } },
@@ -1497,7 +1504,7 @@
},
'#withDescription':: d.fn(help='"Description of this PostgreSQL cluster"', args=[d.arg(name='description', type=d.T.string)]),
withDescription(description): { spec+: { description: description } },
- '#withEnableSuperuserAccess':: d.fn(help='"When this option is enabled, the operator will use the `SuperuserSecret` to update the `postgres` user password (if the secret is not present, the operator will automatically create one). When this option is disabled, the operator will ignore the `SuperuserSecret` content, delete it when automatically created, and then blank the password of the `postgres` user by setting it to `NULL`. Enabled by default."', args=[d.arg(name='enableSuperuserAccess', type=d.T.boolean)]),
+ '#withEnableSuperuserAccess':: d.fn(help='"When this option is enabled, the operator will use the `SuperuserSecret` to update the `postgres` user password (if the secret is not present, the operator will automatically create one). When this option is disabled, the operator will ignore the `SuperuserSecret` content, delete it when automatically created, and then blank the password of the `postgres` user by setting it to `NULL`. Disabled by default."', args=[d.arg(name='enableSuperuserAccess', type=d.T.boolean)]),
withEnableSuperuserAccess(enableSuperuserAccess): { spec+: { enableSuperuserAccess: enableSuperuserAccess } },
'#withEnv':: d.fn(help='"Env follows the Env format to pass environment variables to the pods created in the cluster"', args=[d.arg(name='env', type=d.T.array)]),
withEnv(env): { spec+: { env: if std.isArray(v=env) then env else [env] } },
@@ -1541,8 +1548,8 @@
withPriorityClassName(priorityClassName): { spec+: { priorityClassName: priorityClassName } },
'#withSchedulerName':: d.fn(help='"If specified, the pod will be dispatched by specified Kubernetes scheduler. If not specified, the pod will be dispatched by the default scheduler. More info: https://kubernetes.io/docs/concepts/scheduling-eviction/kube-scheduler/"', args=[d.arg(name='schedulerName', type=d.T.string)]),
withSchedulerName(schedulerName): { spec+: { schedulerName: schedulerName } },
- '#withSmartStopDelay':: d.fn(help='"The time in seconds that controls the window of time reserved for the smart shutdown of Postgres to complete. this formula to compute the timeout of smart shutdown is `max(stopDelay - smartStopDelay, 30)`"', args=[d.arg(name='smartStopDelay', type=d.T.integer)]),
- withSmartStopDelay(smartStopDelay): { spec+: { smartStopDelay: smartStopDelay } },
+ '#withSmartShutdownTimeout':: d.fn(help='"The time in seconds that controls the window of time reserved for the smart shutdown of Postgres to complete. Make sure you reserve enough time for the operator to request a fast shutdown of Postgres (that is: `stopDelay` - `smartShutdownTimeout`)."', args=[d.arg(name='smartShutdownTimeout', type=d.T.integer)]),
+ withSmartShutdownTimeout(smartShutdownTimeout): { spec+: { smartShutdownTimeout: smartShutdownTimeout } },
'#withStartDelay':: d.fn(help='"The time in seconds that is allowed for a PostgreSQL instance to successfully start up (default 3600). The startup probe failure threshold is derived from this value using the formula: ceiling(startDelay / 10)."', args=[d.arg(name='startDelay', type=d.T.integer)]),
withStartDelay(startDelay): { spec+: { startDelay: startDelay } },
'#withStopDelay':: d.fn(help='"The time in seconds that is allowed for a PostgreSQL instance to gracefully shutdown (default 1800)"', args=[d.arg(name='stopDelay', type=d.T.integer)]),
diff --git a/1.19.4/_gen/postgresql/v1/cluster.libsonnet b/1.19.4/_gen/postgresql/v1/cluster.libsonnet
index 7847c36..71b8f6e 100644
--- a/1.19.4/_gen/postgresql/v1/cluster.libsonnet
+++ b/1.19.4/_gen/postgresql/v1/cluster.libsonnet
@@ -802,6 +802,13 @@
'#withPrefix':: d.fn(help='"An optional identifier to prepend to each key in the ConfigMap. Must be a C_IDENTIFIER."', args=[d.arg(name='prefix', type=d.T.string)]),
withPrefix(prefix): { prefix: prefix },
},
+ '#ephemeralVolumesSizeLimit':: d.obj(help='"EphemeralVolumesSizeLimit allows the user to set the limits for the ephemeral volumes"'),
+ ephemeralVolumesSizeLimit: {
+ '#withShm':: d.fn(help='"Shm is the size limit of the shared memory volume"', args=[d.arg(name='shm', type=d.T.any)]),
+ withShm(shm): { spec+: { ephemeralVolumesSizeLimit+: { shm: shm } } },
+ '#withTemporaryData':: d.fn(help='"TemporaryData is the size limit of the temporary data volume"', args=[d.arg(name='temporaryData', type=d.T.any)]),
+ withTemporaryData(temporaryData): { spec+: { ephemeralVolumesSizeLimit+: { temporaryData: temporaryData } } },
+ },
'#externalClusters':: d.obj(help='"The list of external clusters which are used in the configuration"'),
externalClusters: {
'#barmanObjectStore':: d.obj(help='"The configuration for the barman-cloud tool suite"'),
@@ -1232,7 +1239,7 @@
replicationSlots: {
'#highAvailability':: d.obj(help='"Replication slots for high availability configuration"'),
highAvailability: {
- '#withEnabled':: d.fn(help='"If enabled, the operator will automatically manage replication slots on the primary instance and use them in streaming replication connections with all the standby instances that are part of the HA cluster. If disabled (default), the operator will not take advantage of replication slots in streaming connections with the replicas. This feature also controls replication slots in replica cluster, from the designated primary to its cascading replicas. This can only be set at creation time."', args=[d.arg(name='enabled', type=d.T.boolean)]),
+ '#withEnabled':: d.fn(help='"If enabled (default), the operator will automatically manage replication slots on the primary instance and use them in streaming replication connections with all the standby instances that are part of the HA cluster. If disabled, the operator will not take advantage of replication slots in streaming connections with the replicas. This feature also controls replication slots in replica cluster, from the designated primary to its cascading replicas."', args=[d.arg(name='enabled', type=d.T.boolean)]),
withEnabled(enabled): { spec+: { replicationSlots+: { highAvailability+: { enabled: enabled } } } },
'#withSlotPrefix':: d.fn(help='"Prefix for replication slots managed by the operator for HA. It may only contain lower case letters, numbers, and the underscore character. This can only be set at creation time. By default set to `_cnpg_`."', args=[d.arg(name='slotPrefix', type=d.T.string)]),
withSlotPrefix(slotPrefix): { spec+: { replicationSlots+: { highAvailability+: { slotPrefix: slotPrefix } } } },
@@ -1497,7 +1504,7 @@
},
'#withDescription':: d.fn(help='"Description of this PostgreSQL cluster"', args=[d.arg(name='description', type=d.T.string)]),
withDescription(description): { spec+: { description: description } },
- '#withEnableSuperuserAccess':: d.fn(help='"When this option is enabled, the operator will use the `SuperuserSecret` to update the `postgres` user password (if the secret is not present, the operator will automatically create one). When this option is disabled, the operator will ignore the `SuperuserSecret` content, delete it when automatically created, and then blank the password of the `postgres` user by setting it to `NULL`. Enabled by default."', args=[d.arg(name='enableSuperuserAccess', type=d.T.boolean)]),
+ '#withEnableSuperuserAccess':: d.fn(help='"When this option is enabled, the operator will use the `SuperuserSecret` to update the `postgres` user password (if the secret is not present, the operator will automatically create one). When this option is disabled, the operator will ignore the `SuperuserSecret` content, delete it when automatically created, and then blank the password of the `postgres` user by setting it to `NULL`. Disabled by default."', args=[d.arg(name='enableSuperuserAccess', type=d.T.boolean)]),
withEnableSuperuserAccess(enableSuperuserAccess): { spec+: { enableSuperuserAccess: enableSuperuserAccess } },
'#withEnv':: d.fn(help='"Env follows the Env format to pass environment variables to the pods created in the cluster"', args=[d.arg(name='env', type=d.T.array)]),
withEnv(env): { spec+: { env: if std.isArray(v=env) then env else [env] } },
@@ -1541,8 +1548,8 @@
withPriorityClassName(priorityClassName): { spec+: { priorityClassName: priorityClassName } },
'#withSchedulerName':: d.fn(help='"If specified, the pod will be dispatched by specified Kubernetes scheduler. If not specified, the pod will be dispatched by the default scheduler. More info: https://kubernetes.io/docs/concepts/scheduling-eviction/kube-scheduler/"', args=[d.arg(name='schedulerName', type=d.T.string)]),
withSchedulerName(schedulerName): { spec+: { schedulerName: schedulerName } },
- '#withSmartStopDelay':: d.fn(help='"The time in seconds that controls the window of time reserved for the smart shutdown of Postgres to complete. this formula to compute the timeout of smart shutdown is `max(stopDelay - smartStopDelay, 30)`"', args=[d.arg(name='smartStopDelay', type=d.T.integer)]),
- withSmartStopDelay(smartStopDelay): { spec+: { smartStopDelay: smartStopDelay } },
+ '#withSmartShutdownTimeout':: d.fn(help='"The time in seconds that controls the window of time reserved for the smart shutdown of Postgres to complete. Make sure you reserve enough time for the operator to request a fast shutdown of Postgres (that is: `stopDelay` - `smartShutdownTimeout`)."', args=[d.arg(name='smartShutdownTimeout', type=d.T.integer)]),
+ withSmartShutdownTimeout(smartShutdownTimeout): { spec+: { smartShutdownTimeout: smartShutdownTimeout } },
'#withStartDelay':: d.fn(help='"The time in seconds that is allowed for a PostgreSQL instance to successfully start up (default 3600). The startup probe failure threshold is derived from this value using the formula: ceiling(startDelay / 10)."', args=[d.arg(name='startDelay', type=d.T.integer)]),
withStartDelay(startDelay): { spec+: { startDelay: startDelay } },
'#withStopDelay':: d.fn(help='"The time in seconds that is allowed for a PostgreSQL instance to gracefully shutdown (default 1800)"', args=[d.arg(name='stopDelay', type=d.T.integer)]),
diff --git a/1.20.2/_gen/postgresql/v1/cluster.libsonnet b/1.20.2/_gen/postgresql/v1/cluster.libsonnet
index 7847c36..71b8f6e 100644
--- a/1.20.2/_gen/postgresql/v1/cluster.libsonnet
+++ b/1.20.2/_gen/postgresql/v1/cluster.libsonnet
@@ -802,6 +802,13 @@
'#withPrefix':: d.fn(help='"An optional identifier to prepend to each key in the ConfigMap. Must be a C_IDENTIFIER."', args=[d.arg(name='prefix', type=d.T.string)]),
withPrefix(prefix): { prefix: prefix },
},
+ '#ephemeralVolumesSizeLimit':: d.obj(help='"EphemeralVolumesSizeLimit allows the user to set the limits for the ephemeral volumes"'),
+ ephemeralVolumesSizeLimit: {
+ '#withShm':: d.fn(help='"Shm is the size limit of the shared memory volume"', args=[d.arg(name='shm', type=d.T.any)]),
+ withShm(shm): { spec+: { ephemeralVolumesSizeLimit+: { shm: shm } } },
+ '#withTemporaryData':: d.fn(help='"TemporaryData is the size limit of the temporary data volume"', args=[d.arg(name='temporaryData', type=d.T.any)]),
+ withTemporaryData(temporaryData): { spec+: { ephemeralVolumesSizeLimit+: { temporaryData: temporaryData } } },
+ },
'#externalClusters':: d.obj(help='"The list of external clusters which are used in the configuration"'),
externalClusters: {
'#barmanObjectStore':: d.obj(help='"The configuration for the barman-cloud tool suite"'),
@@ -1232,7 +1239,7 @@
replicationSlots: {
'#highAvailability':: d.obj(help='"Replication slots for high availability configuration"'),
highAvailability: {
- '#withEnabled':: d.fn(help='"If enabled, the operator will automatically manage replication slots on the primary instance and use them in streaming replication connections with all the standby instances that are part of the HA cluster. If disabled (default), the operator will not take advantage of replication slots in streaming connections with the replicas. This feature also controls replication slots in replica cluster, from the designated primary to its cascading replicas. This can only be set at creation time."', args=[d.arg(name='enabled', type=d.T.boolean)]),
+ '#withEnabled':: d.fn(help='"If enabled (default), the operator will automatically manage replication slots on the primary instance and use them in streaming replication connections with all the standby instances that are part of the HA cluster. If disabled, the operator will not take advantage of replication slots in streaming connections with the replicas. This feature also controls replication slots in replica cluster, from the designated primary to its cascading replicas."', args=[d.arg(name='enabled', type=d.T.boolean)]),
withEnabled(enabled): { spec+: { replicationSlots+: { highAvailability+: { enabled: enabled } } } },
'#withSlotPrefix':: d.fn(help='"Prefix for replication slots managed by the operator for HA. It may only contain lower case letters, numbers, and the underscore character. This can only be set at creation time. By default set to `_cnpg_`."', args=[d.arg(name='slotPrefix', type=d.T.string)]),
withSlotPrefix(slotPrefix): { spec+: { replicationSlots+: { highAvailability+: { slotPrefix: slotPrefix } } } },
@@ -1497,7 +1504,7 @@
},
'#withDescription':: d.fn(help='"Description of this PostgreSQL cluster"', args=[d.arg(name='description', type=d.T.string)]),
withDescription(description): { spec+: { description: description } },
- '#withEnableSuperuserAccess':: d.fn(help='"When this option is enabled, the operator will use the `SuperuserSecret` to update the `postgres` user password (if the secret is not present, the operator will automatically create one). When this option is disabled, the operator will ignore the `SuperuserSecret` content, delete it when automatically created, and then blank the password of the `postgres` user by setting it to `NULL`. Enabled by default."', args=[d.arg(name='enableSuperuserAccess', type=d.T.boolean)]),
+ '#withEnableSuperuserAccess':: d.fn(help='"When this option is enabled, the operator will use the `SuperuserSecret` to update the `postgres` user password (if the secret is not present, the operator will automatically create one). When this option is disabled, the operator will ignore the `SuperuserSecret` content, delete it when automatically created, and then blank the password of the `postgres` user by setting it to `NULL`. Disabled by default."', args=[d.arg(name='enableSuperuserAccess', type=d.T.boolean)]),
withEnableSuperuserAccess(enableSuperuserAccess): { spec+: { enableSuperuserAccess: enableSuperuserAccess } },
'#withEnv':: d.fn(help='"Env follows the Env format to pass environment variables to the pods created in the cluster"', args=[d.arg(name='env', type=d.T.array)]),
withEnv(env): { spec+: { env: if std.isArray(v=env) then env else [env] } },
@@ -1541,8 +1548,8 @@
withPriorityClassName(priorityClassName): { spec+: { priorityClassName: priorityClassName } },
'#withSchedulerName':: d.fn(help='"If specified, the pod will be dispatched by specified Kubernetes scheduler. If not specified, the pod will be dispatched by the default scheduler. More info: https://kubernetes.io/docs/concepts/scheduling-eviction/kube-scheduler/"', args=[d.arg(name='schedulerName', type=d.T.string)]),
withSchedulerName(schedulerName): { spec+: { schedulerName: schedulerName } },
- '#withSmartStopDelay':: d.fn(help='"The time in seconds that controls the window of time reserved for the smart shutdown of Postgres to complete. this formula to compute the timeout of smart shutdown is `max(stopDelay - smartStopDelay, 30)`"', args=[d.arg(name='smartStopDelay', type=d.T.integer)]),
- withSmartStopDelay(smartStopDelay): { spec+: { smartStopDelay: smartStopDelay } },
+ '#withSmartShutdownTimeout':: d.fn(help='"The time in seconds that controls the window of time reserved for the smart shutdown of Postgres to complete. Make sure you reserve enough time for the operator to request a fast shutdown of Postgres (that is: `stopDelay` - `smartShutdownTimeout`)."', args=[d.arg(name='smartShutdownTimeout', type=d.T.integer)]),
+ withSmartShutdownTimeout(smartShutdownTimeout): { spec+: { smartShutdownTimeout: smartShutdownTimeout } },
'#withStartDelay':: d.fn(help='"The time in seconds that is allowed for a PostgreSQL instance to successfully start up (default 3600). The startup probe failure threshold is derived from this value using the formula: ceiling(startDelay / 10)."', args=[d.arg(name='startDelay', type=d.T.integer)]),
withStartDelay(startDelay): { spec+: { startDelay: startDelay } },
'#withStopDelay':: d.fn(help='"The time in seconds that is allowed for a PostgreSQL instance to gracefully shutdown (default 1800)"', args=[d.arg(name='stopDelay', type=d.T.integer)]),
diff --git a/1.21.0/_gen/postgresql/main.libsonnet b/1.21.0/_gen/postgresql/main.libsonnet
new file mode 100644
index 0000000..4e73d97
--- /dev/null
+++ b/1.21.0/_gen/postgresql/main.libsonnet
@@ -0,0 +1,5 @@
+{
+ local d = (import 'doc-util/main.libsonnet'),
+ '#':: d.pkg(name='postgresql', url='', help=''),
+ v1: (import 'v1/main.libsonnet'),
+}
diff --git a/1.21.0/_gen/postgresql/v1/backup.libsonnet b/1.21.0/_gen/postgresql/v1/backup.libsonnet
new file mode 100644
index 0000000..8400ca7
--- /dev/null
+++ b/1.21.0/_gen/postgresql/v1/backup.libsonnet
@@ -0,0 +1,64 @@
+{
+ local d = (import 'doc-util/main.libsonnet'),
+ '#':: d.pkg(name='backup', url='', help='"Backup is the Schema for the backups API"'),
+ '#metadata':: d.obj(help='"ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create."'),
+ metadata: {
+ '#withAnnotations':: d.fn(help='"Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations"', args=[d.arg(name='annotations', type=d.T.object)]),
+ withAnnotations(annotations): { metadata+: { annotations: annotations } },
+ '#withAnnotationsMixin':: d.fn(help='"Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]),
+ withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } },
+ '#withClusterName':: d.fn(help='"The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request."', args=[d.arg(name='clusterName', type=d.T.string)]),
+ withClusterName(clusterName): { metadata+: { clusterName: clusterName } },
+ '#withCreationTimestamp':: d.fn(help='"Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers."', args=[d.arg(name='creationTimestamp', type=d.T.string)]),
+ withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } },
+ '#withDeletionGracePeriodSeconds':: d.fn(help='"Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only."', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]),
+ withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } },
+ '#withDeletionTimestamp':: d.fn(help='"Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers."', args=[d.arg(name='deletionTimestamp', type=d.T.string)]),
+ withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } },
+ '#withFinalizers':: d.fn(help='"Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list."', args=[d.arg(name='finalizers', type=d.T.array)]),
+ withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } },
+ '#withFinalizersMixin':: d.fn(help='"Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]),
+ withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } },
+ '#withGenerateName':: d.fn(help='"GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\\n\\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\\n\\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency"', args=[d.arg(name='generateName', type=d.T.string)]),
+ withGenerateName(generateName): { metadata+: { generateName: generateName } },
+ '#withGeneration':: d.fn(help='"A sequence number representing a specific generation of the desired state. Populated by the system. Read-only."', args=[d.arg(name='generation', type=d.T.integer)]),
+ withGeneration(generation): { metadata+: { generation: generation } },
+ '#withLabels':: d.fn(help='"Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels"', args=[d.arg(name='labels', type=d.T.object)]),
+ withLabels(labels): { metadata+: { labels: labels } },
+ '#withLabelsMixin':: d.fn(help='"Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]),
+ withLabelsMixin(labels): { metadata+: { labels+: labels } },
+ '#withName':: d.fn(help='"Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names"', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { metadata+: { name: name } },
+ '#withNamespace':: d.fn(help='"Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the \\"default\\" namespace, but \\"default\\" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\\n\\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces"', args=[d.arg(name='namespace', type=d.T.string)]),
+ withNamespace(namespace): { metadata+: { namespace: namespace } },
+ '#withOwnerReferences':: d.fn(help='"List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller."', args=[d.arg(name='ownerReferences', type=d.T.array)]),
+ withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } },
+ '#withOwnerReferencesMixin':: d.fn(help='"List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]),
+ withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } },
+ '#withResourceVersion':: d.fn(help='"An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\\n\\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency"', args=[d.arg(name='resourceVersion', type=d.T.string)]),
+ withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } },
+ '#withSelfLink':: d.fn(help='"SelfLink is a URL representing this object. Populated by the system. Read-only.\\n\\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release."', args=[d.arg(name='selfLink', type=d.T.string)]),
+ withSelfLink(selfLink): { metadata+: { selfLink: selfLink } },
+ '#withUid':: d.fn(help='"UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\\n\\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids"', args=[d.arg(name='uid', type=d.T.string)]),
+ withUid(uid): { metadata+: { uid: uid } },
+ },
+ '#new':: d.fn(help='new returns an instance of Backup', args=[d.arg(name='name', type=d.T.string)]),
+ new(name): {
+ apiVersion: 'postgresql.cnpg.io/v1',
+ kind: 'Backup',
+ } + self.metadata.withName(name=name),
+ '#spec':: d.obj(help='"Specification of the desired behavior of the backup. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#spec-and-status"'),
+ spec: {
+ '#cluster':: d.obj(help='"The cluster to backup"'),
+ cluster: {
+ '#withName':: d.fn(help='"Name of the referent."', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { spec+: { cluster+: { name: name } } },
+ },
+ '#withMethod':: d.fn(help='"The backup method to be used, possible options are `barmanObjectStore` and `volumeSnapshot`. Defaults to: `barmanObjectStore`."', args=[d.arg(name='method', type=d.T.string)]),
+ withMethod(method): { spec+: { method: method } },
+ '#withTarget':: d.fn(help='"The policy to decide which instance should perform this backup. If empty, it defaults to `cluster.spec.backup.target`. Available options are empty string, `primary` and `prefer-standby`. `primary` to have backups run always on primary instances, `prefer-standby` to have backups run preferably on the most updated standby, if available."', args=[d.arg(name='target', type=d.T.string)]),
+ withTarget(target): { spec+: { target: target } },
+ },
+ '#mixin': 'ignore',
+ mixin: self,
+}
diff --git a/1.21.0/_gen/postgresql/v1/cluster.libsonnet b/1.21.0/_gen/postgresql/v1/cluster.libsonnet
new file mode 100644
index 0000000..71b8f6e
--- /dev/null
+++ b/1.21.0/_gen/postgresql/v1/cluster.libsonnet
@@ -0,0 +1,1566 @@
+{
+ local d = (import 'doc-util/main.libsonnet'),
+ '#':: d.pkg(name='cluster', url='', help='"Cluster is the Schema for the PostgreSQL API"'),
+ '#metadata':: d.obj(help='"ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create."'),
+ metadata: {
+ '#withAnnotations':: d.fn(help='"Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations"', args=[d.arg(name='annotations', type=d.T.object)]),
+ withAnnotations(annotations): { metadata+: { annotations: annotations } },
+ '#withAnnotationsMixin':: d.fn(help='"Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]),
+ withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } },
+ '#withClusterName':: d.fn(help='"The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request."', args=[d.arg(name='clusterName', type=d.T.string)]),
+ withClusterName(clusterName): { metadata+: { clusterName: clusterName } },
+ '#withCreationTimestamp':: d.fn(help='"Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers."', args=[d.arg(name='creationTimestamp', type=d.T.string)]),
+ withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } },
+ '#withDeletionGracePeriodSeconds':: d.fn(help='"Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only."', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]),
+ withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } },
+ '#withDeletionTimestamp':: d.fn(help='"Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers."', args=[d.arg(name='deletionTimestamp', type=d.T.string)]),
+ withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } },
+ '#withFinalizers':: d.fn(help='"Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list."', args=[d.arg(name='finalizers', type=d.T.array)]),
+ withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } },
+ '#withFinalizersMixin':: d.fn(help='"Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]),
+ withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } },
+ '#withGenerateName':: d.fn(help='"GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\\n\\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\\n\\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency"', args=[d.arg(name='generateName', type=d.T.string)]),
+ withGenerateName(generateName): { metadata+: { generateName: generateName } },
+ '#withGeneration':: d.fn(help='"A sequence number representing a specific generation of the desired state. Populated by the system. Read-only."', args=[d.arg(name='generation', type=d.T.integer)]),
+ withGeneration(generation): { metadata+: { generation: generation } },
+ '#withLabels':: d.fn(help='"Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels"', args=[d.arg(name='labels', type=d.T.object)]),
+ withLabels(labels): { metadata+: { labels: labels } },
+ '#withLabelsMixin':: d.fn(help='"Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]),
+ withLabelsMixin(labels): { metadata+: { labels+: labels } },
+ '#withName':: d.fn(help='"Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names"', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { metadata+: { name: name } },
+ '#withNamespace':: d.fn(help='"Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the \\"default\\" namespace, but \\"default\\" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\\n\\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces"', args=[d.arg(name='namespace', type=d.T.string)]),
+ withNamespace(namespace): { metadata+: { namespace: namespace } },
+ '#withOwnerReferences':: d.fn(help='"List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller."', args=[d.arg(name='ownerReferences', type=d.T.array)]),
+ withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } },
+ '#withOwnerReferencesMixin':: d.fn(help='"List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]),
+ withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } },
+ '#withResourceVersion':: d.fn(help='"An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\\n\\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency"', args=[d.arg(name='resourceVersion', type=d.T.string)]),
+ withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } },
+ '#withSelfLink':: d.fn(help='"SelfLink is a URL representing this object. Populated by the system. Read-only.\\n\\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release."', args=[d.arg(name='selfLink', type=d.T.string)]),
+ withSelfLink(selfLink): { metadata+: { selfLink: selfLink } },
+ '#withUid':: d.fn(help='"UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\\n\\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids"', args=[d.arg(name='uid', type=d.T.string)]),
+ withUid(uid): { metadata+: { uid: uid } },
+ },
+ '#new':: d.fn(help='new returns an instance of Cluster', args=[d.arg(name='name', type=d.T.string)]),
+ new(name): {
+ apiVersion: 'postgresql.cnpg.io/v1',
+ kind: 'Cluster',
+ } + self.metadata.withName(name=name),
+ '#spec':: d.obj(help='"Specification of the desired behavior of the cluster. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#spec-and-status"'),
+ spec: {
+ '#affinity':: d.obj(help='"Affinity/Anti-affinity rules for Pods"'),
+ affinity: {
+ '#additionalPodAffinity':: d.obj(help="\"AdditionalPodAffinity allows to specify pod affinity terms to be passed to all the cluster's pods.\""),
+ additionalPodAffinity: {
+ '#preferredDuringSchedulingIgnoredDuringExecution':: d.obj(help='"The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding \\"weight\\" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred."'),
+ preferredDuringSchedulingIgnoredDuringExecution: {
+ '#podAffinityTerm':: d.obj(help='"Required. A pod affinity term, associated with the corresponding weight."'),
+ podAffinityTerm: {
+ '#labelSelector':: d.obj(help='"A label query over a set of resources, in this case pods."'),
+ labelSelector: {
+ '#matchExpressions':: d.obj(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."'),
+ matchExpressions: {
+ '#withKey':: d.fn(help='"key is the label key that the selector applies to."', args=[d.arg(name='key', type=d.T.string)]),
+ withKey(key): { key: key },
+ '#withOperator':: d.fn(help="\"operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist.\"", args=[d.arg(name='operator', type=d.T.string)]),
+ withOperator(operator): { operator: operator },
+ '#withValues':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."', args=[d.arg(name='values', type=d.T.array)]),
+ withValues(values): { values: if std.isArray(v=values) then values else [values] },
+ '#withValuesMixin':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='values', type=d.T.array)]),
+ withValuesMixin(values): { values+: if std.isArray(v=values) then values else [values] },
+ },
+ '#withMatchExpressions':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."', args=[d.arg(name='matchExpressions', type=d.T.array)]),
+ withMatchExpressions(matchExpressions): { podAffinityTerm+: { labelSelector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } },
+ '#withMatchExpressionsMixin':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]),
+ withMatchExpressionsMixin(matchExpressions): { podAffinityTerm+: { labelSelector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } },
+ '#withMatchLabels':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."', args=[d.arg(name='matchLabels', type=d.T.object)]),
+ withMatchLabels(matchLabels): { podAffinityTerm+: { labelSelector+: { matchLabels: matchLabels } } },
+ '#withMatchLabelsMixin':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]),
+ withMatchLabelsMixin(matchLabels): { podAffinityTerm+: { labelSelector+: { matchLabels+: matchLabels } } },
+ },
+ '#namespaceSelector':: d.obj(help="\"A label query over the set of namespaces that the term applies to. The term is applied to the union of the namespaces selected by this field and the ones listed in the namespaces field. null selector and null or empty namespaces list means \\\"this pod's namespace\\\". An empty selector ({}) matches all namespaces.\""),
+ namespaceSelector: {
+ '#matchExpressions':: d.obj(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."'),
+ matchExpressions: {
+ '#withKey':: d.fn(help='"key is the label key that the selector applies to."', args=[d.arg(name='key', type=d.T.string)]),
+ withKey(key): { key: key },
+ '#withOperator':: d.fn(help="\"operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist.\"", args=[d.arg(name='operator', type=d.T.string)]),
+ withOperator(operator): { operator: operator },
+ '#withValues':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."', args=[d.arg(name='values', type=d.T.array)]),
+ withValues(values): { values: if std.isArray(v=values) then values else [values] },
+ '#withValuesMixin':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='values', type=d.T.array)]),
+ withValuesMixin(values): { values+: if std.isArray(v=values) then values else [values] },
+ },
+ '#withMatchExpressions':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."', args=[d.arg(name='matchExpressions', type=d.T.array)]),
+ withMatchExpressions(matchExpressions): { podAffinityTerm+: { namespaceSelector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } },
+ '#withMatchExpressionsMixin':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]),
+ withMatchExpressionsMixin(matchExpressions): { podAffinityTerm+: { namespaceSelector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } },
+ '#withMatchLabels':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."', args=[d.arg(name='matchLabels', type=d.T.object)]),
+ withMatchLabels(matchLabels): { podAffinityTerm+: { namespaceSelector+: { matchLabels: matchLabels } } },
+ '#withMatchLabelsMixin':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]),
+ withMatchLabelsMixin(matchLabels): { podAffinityTerm+: { namespaceSelector+: { matchLabels+: matchLabels } } },
+ },
+ '#withNamespaces':: d.fn(help="\"namespaces specifies a static list of namespace names that the term applies to. The term is applied to the union of the namespaces listed in this field and the ones selected by namespaceSelector. null or empty namespaces list and null namespaceSelector means \\\"this pod's namespace\\\".\"", args=[d.arg(name='namespaces', type=d.T.array)]),
+ withNamespaces(namespaces): { podAffinityTerm+: { namespaces: if std.isArray(v=namespaces) then namespaces else [namespaces] } },
+ '#withNamespacesMixin':: d.fn(help="\"namespaces specifies a static list of namespace names that the term applies to. The term is applied to the union of the namespaces listed in this field and the ones selected by namespaceSelector. null or empty namespaces list and null namespaceSelector means \\\"this pod's namespace\\\".\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='namespaces', type=d.T.array)]),
+ withNamespacesMixin(namespaces): { podAffinityTerm+: { namespaces+: if std.isArray(v=namespaces) then namespaces else [namespaces] } },
+ '#withTopologyKey':: d.fn(help='"This pod should be co-located (affinity) or not co-located (anti-affinity) with the pods matching the labelSelector in the specified namespaces, where co-located is defined as running on a node whose value of the label with key topologyKey matches that of any node on which any of the selected pods is running. Empty topologyKey is not allowed."', args=[d.arg(name='topologyKey', type=d.T.string)]),
+ withTopologyKey(topologyKey): { podAffinityTerm+: { topologyKey: topologyKey } },
+ },
+ '#withWeight':: d.fn(help='"weight associated with matching the corresponding podAffinityTerm, in the range 1-100."', args=[d.arg(name='weight', type=d.T.integer)]),
+ withWeight(weight): { weight: weight },
+ },
+ '#requiredDuringSchedulingIgnoredDuringExecution':: d.obj(help='"If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied."'),
+ requiredDuringSchedulingIgnoredDuringExecution: {
+ '#labelSelector':: d.obj(help='"A label query over a set of resources, in this case pods."'),
+ labelSelector: {
+ '#matchExpressions':: d.obj(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."'),
+ matchExpressions: {
+ '#withKey':: d.fn(help='"key is the label key that the selector applies to."', args=[d.arg(name='key', type=d.T.string)]),
+ withKey(key): { key: key },
+ '#withOperator':: d.fn(help="\"operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist.\"", args=[d.arg(name='operator', type=d.T.string)]),
+ withOperator(operator): { operator: operator },
+ '#withValues':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."', args=[d.arg(name='values', type=d.T.array)]),
+ withValues(values): { values: if std.isArray(v=values) then values else [values] },
+ '#withValuesMixin':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='values', type=d.T.array)]),
+ withValuesMixin(values): { values+: if std.isArray(v=values) then values else [values] },
+ },
+ '#withMatchExpressions':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."', args=[d.arg(name='matchExpressions', type=d.T.array)]),
+ withMatchExpressions(matchExpressions): { labelSelector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } },
+ '#withMatchExpressionsMixin':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]),
+ withMatchExpressionsMixin(matchExpressions): { labelSelector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } },
+ '#withMatchLabels':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."', args=[d.arg(name='matchLabels', type=d.T.object)]),
+ withMatchLabels(matchLabels): { labelSelector+: { matchLabels: matchLabels } },
+ '#withMatchLabelsMixin':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]),
+ withMatchLabelsMixin(matchLabels): { labelSelector+: { matchLabels+: matchLabels } },
+ },
+ '#namespaceSelector':: d.obj(help="\"A label query over the set of namespaces that the term applies to. The term is applied to the union of the namespaces selected by this field and the ones listed in the namespaces field. null selector and null or empty namespaces list means \\\"this pod's namespace\\\". An empty selector ({}) matches all namespaces.\""),
+ namespaceSelector: {
+ '#matchExpressions':: d.obj(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."'),
+ matchExpressions: {
+ '#withKey':: d.fn(help='"key is the label key that the selector applies to."', args=[d.arg(name='key', type=d.T.string)]),
+ withKey(key): { key: key },
+ '#withOperator':: d.fn(help="\"operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist.\"", args=[d.arg(name='operator', type=d.T.string)]),
+ withOperator(operator): { operator: operator },
+ '#withValues':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."', args=[d.arg(name='values', type=d.T.array)]),
+ withValues(values): { values: if std.isArray(v=values) then values else [values] },
+ '#withValuesMixin':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='values', type=d.T.array)]),
+ withValuesMixin(values): { values+: if std.isArray(v=values) then values else [values] },
+ },
+ '#withMatchExpressions':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."', args=[d.arg(name='matchExpressions', type=d.T.array)]),
+ withMatchExpressions(matchExpressions): { namespaceSelector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } },
+ '#withMatchExpressionsMixin':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]),
+ withMatchExpressionsMixin(matchExpressions): { namespaceSelector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } },
+ '#withMatchLabels':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."', args=[d.arg(name='matchLabels', type=d.T.object)]),
+ withMatchLabels(matchLabels): { namespaceSelector+: { matchLabels: matchLabels } },
+ '#withMatchLabelsMixin':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]),
+ withMatchLabelsMixin(matchLabels): { namespaceSelector+: { matchLabels+: matchLabels } },
+ },
+ '#withNamespaces':: d.fn(help="\"namespaces specifies a static list of namespace names that the term applies to. The term is applied to the union of the namespaces listed in this field and the ones selected by namespaceSelector. null or empty namespaces list and null namespaceSelector means \\\"this pod's namespace\\\".\"", args=[d.arg(name='namespaces', type=d.T.array)]),
+ withNamespaces(namespaces): { namespaces: if std.isArray(v=namespaces) then namespaces else [namespaces] },
+ '#withNamespacesMixin':: d.fn(help="\"namespaces specifies a static list of namespace names that the term applies to. The term is applied to the union of the namespaces listed in this field and the ones selected by namespaceSelector. null or empty namespaces list and null namespaceSelector means \\\"this pod's namespace\\\".\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='namespaces', type=d.T.array)]),
+ withNamespacesMixin(namespaces): { namespaces+: if std.isArray(v=namespaces) then namespaces else [namespaces] },
+ '#withTopologyKey':: d.fn(help='"This pod should be co-located (affinity) or not co-located (anti-affinity) with the pods matching the labelSelector in the specified namespaces, where co-located is defined as running on a node whose value of the label with key topologyKey matches that of any node on which any of the selected pods is running. Empty topologyKey is not allowed."', args=[d.arg(name='topologyKey', type=d.T.string)]),
+ withTopologyKey(topologyKey): { topologyKey: topologyKey },
+ },
+ '#withPreferredDuringSchedulingIgnoredDuringExecution':: d.fn(help='"The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding \\"weight\\" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred."', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]),
+ withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { affinity+: { additionalPodAffinity+: { preferredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } },
+ '#withPreferredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='"The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding \\"weight\\" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]),
+ withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { affinity+: { additionalPodAffinity+: { preferredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } },
+ '#withRequiredDuringSchedulingIgnoredDuringExecution':: d.fn(help='"If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied."', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]),
+ withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution): { spec+: { affinity+: { additionalPodAffinity+: { requiredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } },
+ '#withRequiredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='"If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]),
+ withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution): { spec+: { affinity+: { additionalPodAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } },
+ },
+ '#additionalPodAntiAffinity':: d.obj(help='"AdditionalPodAntiAffinity allows to specify pod anti-affinity terms to be added to the ones generated by the operator if EnablePodAntiAffinity is set to true (default) or to be used exclusively if set to false."'),
+ additionalPodAntiAffinity: {
+ '#preferredDuringSchedulingIgnoredDuringExecution':: d.obj(help='"The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding \\"weight\\" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred."'),
+ preferredDuringSchedulingIgnoredDuringExecution: {
+ '#podAffinityTerm':: d.obj(help='"Required. A pod affinity term, associated with the corresponding weight."'),
+ podAffinityTerm: {
+ '#labelSelector':: d.obj(help='"A label query over a set of resources, in this case pods."'),
+ labelSelector: {
+ '#matchExpressions':: d.obj(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."'),
+ matchExpressions: {
+ '#withKey':: d.fn(help='"key is the label key that the selector applies to."', args=[d.arg(name='key', type=d.T.string)]),
+ withKey(key): { key: key },
+ '#withOperator':: d.fn(help="\"operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist.\"", args=[d.arg(name='operator', type=d.T.string)]),
+ withOperator(operator): { operator: operator },
+ '#withValues':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."', args=[d.arg(name='values', type=d.T.array)]),
+ withValues(values): { values: if std.isArray(v=values) then values else [values] },
+ '#withValuesMixin':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='values', type=d.T.array)]),
+ withValuesMixin(values): { values+: if std.isArray(v=values) then values else [values] },
+ },
+ '#withMatchExpressions':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."', args=[d.arg(name='matchExpressions', type=d.T.array)]),
+ withMatchExpressions(matchExpressions): { podAffinityTerm+: { labelSelector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } },
+ '#withMatchExpressionsMixin':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]),
+ withMatchExpressionsMixin(matchExpressions): { podAffinityTerm+: { labelSelector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } },
+ '#withMatchLabels':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."', args=[d.arg(name='matchLabels', type=d.T.object)]),
+ withMatchLabels(matchLabels): { podAffinityTerm+: { labelSelector+: { matchLabels: matchLabels } } },
+ '#withMatchLabelsMixin':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]),
+ withMatchLabelsMixin(matchLabels): { podAffinityTerm+: { labelSelector+: { matchLabels+: matchLabels } } },
+ },
+ '#namespaceSelector':: d.obj(help="\"A label query over the set of namespaces that the term applies to. The term is applied to the union of the namespaces selected by this field and the ones listed in the namespaces field. null selector and null or empty namespaces list means \\\"this pod's namespace\\\". An empty selector ({}) matches all namespaces.\""),
+ namespaceSelector: {
+ '#matchExpressions':: d.obj(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."'),
+ matchExpressions: {
+ '#withKey':: d.fn(help='"key is the label key that the selector applies to."', args=[d.arg(name='key', type=d.T.string)]),
+ withKey(key): { key: key },
+ '#withOperator':: d.fn(help="\"operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist.\"", args=[d.arg(name='operator', type=d.T.string)]),
+ withOperator(operator): { operator: operator },
+ '#withValues':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."', args=[d.arg(name='values', type=d.T.array)]),
+ withValues(values): { values: if std.isArray(v=values) then values else [values] },
+ '#withValuesMixin':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='values', type=d.T.array)]),
+ withValuesMixin(values): { values+: if std.isArray(v=values) then values else [values] },
+ },
+ '#withMatchExpressions':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."', args=[d.arg(name='matchExpressions', type=d.T.array)]),
+ withMatchExpressions(matchExpressions): { podAffinityTerm+: { namespaceSelector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } },
+ '#withMatchExpressionsMixin':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]),
+ withMatchExpressionsMixin(matchExpressions): { podAffinityTerm+: { namespaceSelector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } },
+ '#withMatchLabels':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."', args=[d.arg(name='matchLabels', type=d.T.object)]),
+ withMatchLabels(matchLabels): { podAffinityTerm+: { namespaceSelector+: { matchLabels: matchLabels } } },
+ '#withMatchLabelsMixin':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]),
+ withMatchLabelsMixin(matchLabels): { podAffinityTerm+: { namespaceSelector+: { matchLabels+: matchLabels } } },
+ },
+ '#withNamespaces':: d.fn(help="\"namespaces specifies a static list of namespace names that the term applies to. The term is applied to the union of the namespaces listed in this field and the ones selected by namespaceSelector. null or empty namespaces list and null namespaceSelector means \\\"this pod's namespace\\\".\"", args=[d.arg(name='namespaces', type=d.T.array)]),
+ withNamespaces(namespaces): { podAffinityTerm+: { namespaces: if std.isArray(v=namespaces) then namespaces else [namespaces] } },
+ '#withNamespacesMixin':: d.fn(help="\"namespaces specifies a static list of namespace names that the term applies to. The term is applied to the union of the namespaces listed in this field and the ones selected by namespaceSelector. null or empty namespaces list and null namespaceSelector means \\\"this pod's namespace\\\".\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='namespaces', type=d.T.array)]),
+ withNamespacesMixin(namespaces): { podAffinityTerm+: { namespaces+: if std.isArray(v=namespaces) then namespaces else [namespaces] } },
+ '#withTopologyKey':: d.fn(help='"This pod should be co-located (affinity) or not co-located (anti-affinity) with the pods matching the labelSelector in the specified namespaces, where co-located is defined as running on a node whose value of the label with key topologyKey matches that of any node on which any of the selected pods is running. Empty topologyKey is not allowed."', args=[d.arg(name='topologyKey', type=d.T.string)]),
+ withTopologyKey(topologyKey): { podAffinityTerm+: { topologyKey: topologyKey } },
+ },
+ '#withWeight':: d.fn(help='"weight associated with matching the corresponding podAffinityTerm, in the range 1-100."', args=[d.arg(name='weight', type=d.T.integer)]),
+ withWeight(weight): { weight: weight },
+ },
+ '#requiredDuringSchedulingIgnoredDuringExecution':: d.obj(help='"If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied."'),
+ requiredDuringSchedulingIgnoredDuringExecution: {
+ '#labelSelector':: d.obj(help='"A label query over a set of resources, in this case pods."'),
+ labelSelector: {
+ '#matchExpressions':: d.obj(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."'),
+ matchExpressions: {
+ '#withKey':: d.fn(help='"key is the label key that the selector applies to."', args=[d.arg(name='key', type=d.T.string)]),
+ withKey(key): { key: key },
+ '#withOperator':: d.fn(help="\"operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist.\"", args=[d.arg(name='operator', type=d.T.string)]),
+ withOperator(operator): { operator: operator },
+ '#withValues':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."', args=[d.arg(name='values', type=d.T.array)]),
+ withValues(values): { values: if std.isArray(v=values) then values else [values] },
+ '#withValuesMixin':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='values', type=d.T.array)]),
+ withValuesMixin(values): { values+: if std.isArray(v=values) then values else [values] },
+ },
+ '#withMatchExpressions':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."', args=[d.arg(name='matchExpressions', type=d.T.array)]),
+ withMatchExpressions(matchExpressions): { labelSelector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } },
+ '#withMatchExpressionsMixin':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]),
+ withMatchExpressionsMixin(matchExpressions): { labelSelector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } },
+ '#withMatchLabels':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."', args=[d.arg(name='matchLabels', type=d.T.object)]),
+ withMatchLabels(matchLabels): { labelSelector+: { matchLabels: matchLabels } },
+ '#withMatchLabelsMixin':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]),
+ withMatchLabelsMixin(matchLabels): { labelSelector+: { matchLabels+: matchLabels } },
+ },
+ '#namespaceSelector':: d.obj(help="\"A label query over the set of namespaces that the term applies to. The term is applied to the union of the namespaces selected by this field and the ones listed in the namespaces field. null selector and null or empty namespaces list means \\\"this pod's namespace\\\". An empty selector ({}) matches all namespaces.\""),
+ namespaceSelector: {
+ '#matchExpressions':: d.obj(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."'),
+ matchExpressions: {
+ '#withKey':: d.fn(help='"key is the label key that the selector applies to."', args=[d.arg(name='key', type=d.T.string)]),
+ withKey(key): { key: key },
+ '#withOperator':: d.fn(help="\"operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist.\"", args=[d.arg(name='operator', type=d.T.string)]),
+ withOperator(operator): { operator: operator },
+ '#withValues':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."', args=[d.arg(name='values', type=d.T.array)]),
+ withValues(values): { values: if std.isArray(v=values) then values else [values] },
+ '#withValuesMixin':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='values', type=d.T.array)]),
+ withValuesMixin(values): { values+: if std.isArray(v=values) then values else [values] },
+ },
+ '#withMatchExpressions':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."', args=[d.arg(name='matchExpressions', type=d.T.array)]),
+ withMatchExpressions(matchExpressions): { namespaceSelector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } },
+ '#withMatchExpressionsMixin':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]),
+ withMatchExpressionsMixin(matchExpressions): { namespaceSelector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } },
+ '#withMatchLabels':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."', args=[d.arg(name='matchLabels', type=d.T.object)]),
+ withMatchLabels(matchLabels): { namespaceSelector+: { matchLabels: matchLabels } },
+ '#withMatchLabelsMixin':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]),
+ withMatchLabelsMixin(matchLabels): { namespaceSelector+: { matchLabels+: matchLabels } },
+ },
+ '#withNamespaces':: d.fn(help="\"namespaces specifies a static list of namespace names that the term applies to. The term is applied to the union of the namespaces listed in this field and the ones selected by namespaceSelector. null or empty namespaces list and null namespaceSelector means \\\"this pod's namespace\\\".\"", args=[d.arg(name='namespaces', type=d.T.array)]),
+ withNamespaces(namespaces): { namespaces: if std.isArray(v=namespaces) then namespaces else [namespaces] },
+ '#withNamespacesMixin':: d.fn(help="\"namespaces specifies a static list of namespace names that the term applies to. The term is applied to the union of the namespaces listed in this field and the ones selected by namespaceSelector. null or empty namespaces list and null namespaceSelector means \\\"this pod's namespace\\\".\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='namespaces', type=d.T.array)]),
+ withNamespacesMixin(namespaces): { namespaces+: if std.isArray(v=namespaces) then namespaces else [namespaces] },
+ '#withTopologyKey':: d.fn(help='"This pod should be co-located (affinity) or not co-located (anti-affinity) with the pods matching the labelSelector in the specified namespaces, where co-located is defined as running on a node whose value of the label with key topologyKey matches that of any node on which any of the selected pods is running. Empty topologyKey is not allowed."', args=[d.arg(name='topologyKey', type=d.T.string)]),
+ withTopologyKey(topologyKey): { topologyKey: topologyKey },
+ },
+ '#withPreferredDuringSchedulingIgnoredDuringExecution':: d.fn(help='"The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding \\"weight\\" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred."', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]),
+ withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { affinity+: { additionalPodAntiAffinity+: { preferredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } },
+ '#withPreferredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='"The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding \\"weight\\" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]),
+ withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { affinity+: { additionalPodAntiAffinity+: { preferredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } },
+ '#withRequiredDuringSchedulingIgnoredDuringExecution':: d.fn(help='"If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied."', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]),
+ withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution): { spec+: { affinity+: { additionalPodAntiAffinity+: { requiredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } },
+ '#withRequiredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='"If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]),
+ withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution): { spec+: { affinity+: { additionalPodAntiAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } },
+ },
+ '#nodeAffinity':: d.obj(help='"NodeAffinity describes node affinity scheduling rules for the pod. More info: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#node-affinity"'),
+ nodeAffinity: {
+ '#preferredDuringSchedulingIgnoredDuringExecution':: d.obj(help='"The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding \\"weight\\" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred."'),
+ preferredDuringSchedulingIgnoredDuringExecution: {
+ '#preference':: d.obj(help='"A node selector term, associated with the corresponding weight."'),
+ preference: {
+ '#matchExpressions':: d.obj(help="\"A list of node selector requirements by node's labels.\""),
+ matchExpressions: {
+ '#withKey':: d.fn(help='"The label key that the selector applies to."', args=[d.arg(name='key', type=d.T.string)]),
+ withKey(key): { key: key },
+ '#withOperator':: d.fn(help="\"Represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists, DoesNotExist. Gt, and Lt.\"", args=[d.arg(name='operator', type=d.T.string)]),
+ withOperator(operator): { operator: operator },
+ '#withValues':: d.fn(help='"An array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. If the operator is Gt or Lt, the values array must have a single element, which will be interpreted as an integer. This array is replaced during a strategic merge patch."', args=[d.arg(name='values', type=d.T.array)]),
+ withValues(values): { values: if std.isArray(v=values) then values else [values] },
+ '#withValuesMixin':: d.fn(help='"An array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. If the operator is Gt or Lt, the values array must have a single element, which will be interpreted as an integer. This array is replaced during a strategic merge patch."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='values', type=d.T.array)]),
+ withValuesMixin(values): { values+: if std.isArray(v=values) then values else [values] },
+ },
+ '#matchFields':: d.obj(help="\"A list of node selector requirements by node's fields.\""),
+ matchFields: {
+ '#withKey':: d.fn(help='"The label key that the selector applies to."', args=[d.arg(name='key', type=d.T.string)]),
+ withKey(key): { key: key },
+ '#withOperator':: d.fn(help="\"Represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists, DoesNotExist. Gt, and Lt.\"", args=[d.arg(name='operator', type=d.T.string)]),
+ withOperator(operator): { operator: operator },
+ '#withValues':: d.fn(help='"An array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. If the operator is Gt or Lt, the values array must have a single element, which will be interpreted as an integer. This array is replaced during a strategic merge patch."', args=[d.arg(name='values', type=d.T.array)]),
+ withValues(values): { values: if std.isArray(v=values) then values else [values] },
+ '#withValuesMixin':: d.fn(help='"An array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. If the operator is Gt or Lt, the values array must have a single element, which will be interpreted as an integer. This array is replaced during a strategic merge patch."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='values', type=d.T.array)]),
+ withValuesMixin(values): { values+: if std.isArray(v=values) then values else [values] },
+ },
+ '#withMatchExpressions':: d.fn(help="\"A list of node selector requirements by node's labels.\"", args=[d.arg(name='matchExpressions', type=d.T.array)]),
+ withMatchExpressions(matchExpressions): { preference+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } },
+ '#withMatchExpressionsMixin':: d.fn(help="\"A list of node selector requirements by node's labels.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='matchExpressions', type=d.T.array)]),
+ withMatchExpressionsMixin(matchExpressions): { preference+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } },
+ '#withMatchFields':: d.fn(help="\"A list of node selector requirements by node's fields.\"", args=[d.arg(name='matchFields', type=d.T.array)]),
+ withMatchFields(matchFields): { preference+: { matchFields: if std.isArray(v=matchFields) then matchFields else [matchFields] } },
+ '#withMatchFieldsMixin':: d.fn(help="\"A list of node selector requirements by node's fields.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='matchFields', type=d.T.array)]),
+ withMatchFieldsMixin(matchFields): { preference+: { matchFields+: if std.isArray(v=matchFields) then matchFields else [matchFields] } },
+ },
+ '#withWeight':: d.fn(help='"Weight associated with matching the corresponding nodeSelectorTerm, in the range 1-100."', args=[d.arg(name='weight', type=d.T.integer)]),
+ withWeight(weight): { weight: weight },
+ },
+ '#requiredDuringSchedulingIgnoredDuringExecution':: d.obj(help='"If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to an update), the system may or may not try to eventually evict the pod from its node."'),
+ requiredDuringSchedulingIgnoredDuringExecution: {
+ '#nodeSelectorTerms':: d.obj(help='"Required. A list of node selector terms. The terms are ORed."'),
+ nodeSelectorTerms: {
+ '#matchExpressions':: d.obj(help="\"A list of node selector requirements by node's labels.\""),
+ matchExpressions: {
+ '#withKey':: d.fn(help='"The label key that the selector applies to."', args=[d.arg(name='key', type=d.T.string)]),
+ withKey(key): { key: key },
+ '#withOperator':: d.fn(help="\"Represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists, DoesNotExist. Gt, and Lt.\"", args=[d.arg(name='operator', type=d.T.string)]),
+ withOperator(operator): { operator: operator },
+ '#withValues':: d.fn(help='"An array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. If the operator is Gt or Lt, the values array must have a single element, which will be interpreted as an integer. This array is replaced during a strategic merge patch."', args=[d.arg(name='values', type=d.T.array)]),
+ withValues(values): { values: if std.isArray(v=values) then values else [values] },
+ '#withValuesMixin':: d.fn(help='"An array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. If the operator is Gt or Lt, the values array must have a single element, which will be interpreted as an integer. This array is replaced during a strategic merge patch."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='values', type=d.T.array)]),
+ withValuesMixin(values): { values+: if std.isArray(v=values) then values else [values] },
+ },
+ '#matchFields':: d.obj(help="\"A list of node selector requirements by node's fields.\""),
+ matchFields: {
+ '#withKey':: d.fn(help='"The label key that the selector applies to."', args=[d.arg(name='key', type=d.T.string)]),
+ withKey(key): { key: key },
+ '#withOperator':: d.fn(help="\"Represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists, DoesNotExist. Gt, and Lt.\"", args=[d.arg(name='operator', type=d.T.string)]),
+ withOperator(operator): { operator: operator },
+ '#withValues':: d.fn(help='"An array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. If the operator is Gt or Lt, the values array must have a single element, which will be interpreted as an integer. This array is replaced during a strategic merge patch."', args=[d.arg(name='values', type=d.T.array)]),
+ withValues(values): { values: if std.isArray(v=values) then values else [values] },
+ '#withValuesMixin':: d.fn(help='"An array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. If the operator is Gt or Lt, the values array must have a single element, which will be interpreted as an integer. This array is replaced during a strategic merge patch."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='values', type=d.T.array)]),
+ withValuesMixin(values): { values+: if std.isArray(v=values) then values else [values] },
+ },
+ '#withMatchExpressions':: d.fn(help="\"A list of node selector requirements by node's labels.\"", args=[d.arg(name='matchExpressions', type=d.T.array)]),
+ withMatchExpressions(matchExpressions): { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] },
+ '#withMatchExpressionsMixin':: d.fn(help="\"A list of node selector requirements by node's labels.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='matchExpressions', type=d.T.array)]),
+ withMatchExpressionsMixin(matchExpressions): { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] },
+ '#withMatchFields':: d.fn(help="\"A list of node selector requirements by node's fields.\"", args=[d.arg(name='matchFields', type=d.T.array)]),
+ withMatchFields(matchFields): { matchFields: if std.isArray(v=matchFields) then matchFields else [matchFields] },
+ '#withMatchFieldsMixin':: d.fn(help="\"A list of node selector requirements by node's fields.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='matchFields', type=d.T.array)]),
+ withMatchFieldsMixin(matchFields): { matchFields+: if std.isArray(v=matchFields) then matchFields else [matchFields] },
+ },
+ '#withNodeSelectorTerms':: d.fn(help='"Required. A list of node selector terms. The terms are ORed."', args=[d.arg(name='nodeSelectorTerms', type=d.T.array)]),
+ withNodeSelectorTerms(nodeSelectorTerms): { spec+: { affinity+: { nodeAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: { nodeSelectorTerms: if std.isArray(v=nodeSelectorTerms) then nodeSelectorTerms else [nodeSelectorTerms] } } } } },
+ '#withNodeSelectorTermsMixin':: d.fn(help='"Required. A list of node selector terms. The terms are ORed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='nodeSelectorTerms', type=d.T.array)]),
+ withNodeSelectorTermsMixin(nodeSelectorTerms): { spec+: { affinity+: { nodeAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: { nodeSelectorTerms+: if std.isArray(v=nodeSelectorTerms) then nodeSelectorTerms else [nodeSelectorTerms] } } } } },
+ },
+ '#withPreferredDuringSchedulingIgnoredDuringExecution':: d.fn(help='"The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding \\"weight\\" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred."', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]),
+ withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { affinity+: { nodeAffinity+: { preferredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } },
+ '#withPreferredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='"The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding \\"weight\\" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]),
+ withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { affinity+: { nodeAffinity+: { preferredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } },
+ },
+ '#tolerations':: d.obj(help='"Tolerations is a list of Tolerations that should be set for all the pods, in order to allow them to run on tainted nodes. More info: https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/"'),
+ tolerations: {
+ '#withEffect':: d.fn(help='"Effect indicates the taint effect to match. Empty means match all taint effects. When specified, allowed values are NoSchedule, PreferNoSchedule and NoExecute."', args=[d.arg(name='effect', type=d.T.string)]),
+ withEffect(effect): { effect: effect },
+ '#withKey':: d.fn(help='"Key is the taint key that the toleration applies to. Empty means match all taint keys. If the key is empty, operator must be Exists; this combination means to match all values and all keys."', args=[d.arg(name='key', type=d.T.string)]),
+ withKey(key): { key: key },
+ '#withOperator':: d.fn(help="\"Operator represents a key's relationship to the value. Valid operators are Exists and Equal. Defaults to Equal. Exists is equivalent to wildcard for value, so that a pod can tolerate all taints of a particular category.\"", args=[d.arg(name='operator', type=d.T.string)]),
+ withOperator(operator): { operator: operator },
+ '#withTolerationSeconds':: d.fn(help='"TolerationSeconds represents the period of time the toleration (which must be of effect NoExecute, otherwise this field is ignored) tolerates the taint. By default, it is not set, which means tolerate the taint forever (do not evict). Zero and negative values will be treated as 0 (evict immediately) by the system."', args=[d.arg(name='tolerationSeconds', type=d.T.integer)]),
+ withTolerationSeconds(tolerationSeconds): { tolerationSeconds: tolerationSeconds },
+ '#withValue':: d.fn(help='"Value is the taint value the toleration matches to. If the operator is Exists, the value should be empty, otherwise just a regular string."', args=[d.arg(name='value', type=d.T.string)]),
+ withValue(value): { value: value },
+ },
+ '#withEnablePodAntiAffinity':: d.fn(help='"Activates anti-affinity for the pods. The operator will define pods anti-affinity unless this field is explicitly set to false"', args=[d.arg(name='enablePodAntiAffinity', type=d.T.boolean)]),
+ withEnablePodAntiAffinity(enablePodAntiAffinity): { spec+: { affinity+: { enablePodAntiAffinity: enablePodAntiAffinity } } },
+ '#withNodeSelector':: d.fn(help='"NodeSelector is map of key-value pairs used to define the nodes on which the pods can run. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/"', args=[d.arg(name='nodeSelector', type=d.T.object)]),
+ withNodeSelector(nodeSelector): { spec+: { affinity+: { nodeSelector: nodeSelector } } },
+ '#withNodeSelectorMixin':: d.fn(help='"NodeSelector is map of key-value pairs used to define the nodes on which the pods can run. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='nodeSelector', type=d.T.object)]),
+ withNodeSelectorMixin(nodeSelector): { spec+: { affinity+: { nodeSelector+: nodeSelector } } },
+ '#withPodAntiAffinityType':: d.fn(help="\"PodAntiAffinityType allows the user to decide whether pod anti-affinity between cluster instance has to be considered a strong requirement during scheduling or not. Allowed values are: \\\"preferred\\\" (default if empty) or \\\"required\\\". Setting it to \\\"required\\\", could lead to instances remaining pending until new kubernetes nodes are added if all the existing nodes don't match the required pod anti-affinity rule. More info: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#inter-pod-affinity-and-anti-affinity\"", args=[d.arg(name='podAntiAffinityType', type=d.T.string)]),
+ withPodAntiAffinityType(podAntiAffinityType): { spec+: { affinity+: { podAntiAffinityType: podAntiAffinityType } } },
+ '#withTolerations':: d.fn(help='"Tolerations is a list of Tolerations that should be set for all the pods, in order to allow them to run on tainted nodes. More info: https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/"', args=[d.arg(name='tolerations', type=d.T.array)]),
+ withTolerations(tolerations): { spec+: { affinity+: { tolerations: if std.isArray(v=tolerations) then tolerations else [tolerations] } } },
+ '#withTolerationsMixin':: d.fn(help='"Tolerations is a list of Tolerations that should be set for all the pods, in order to allow them to run on tainted nodes. More info: https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='tolerations', type=d.T.array)]),
+ withTolerationsMixin(tolerations): { spec+: { affinity+: { tolerations+: if std.isArray(v=tolerations) then tolerations else [tolerations] } } },
+ '#withTopologyKey':: d.fn(help='"TopologyKey to use for anti-affinity configuration. See k8s documentation for more info on that"', args=[d.arg(name='topologyKey', type=d.T.string)]),
+ withTopologyKey(topologyKey): { spec+: { affinity+: { topologyKey: topologyKey } } },
+ },
+ '#backup':: d.obj(help='"The configuration to be used for backups"'),
+ backup: {
+ '#barmanObjectStore':: d.obj(help='"The configuration for the barman-cloud tool suite"'),
+ barmanObjectStore: {
+ '#azureCredentials':: d.obj(help='"The credentials to use to upload data to Azure Blob Storage"'),
+ azureCredentials: {
+ '#connectionString':: d.obj(help='"The connection string to be used"'),
+ connectionString: {
+ '#withKey':: d.fn(help='"The key to select"', args=[d.arg(name='key', type=d.T.string)]),
+ withKey(key): { spec+: { backup+: { barmanObjectStore+: { azureCredentials+: { connectionString+: { key: key } } } } } },
+ '#withName':: d.fn(help='"Name of the referent."', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { spec+: { backup+: { barmanObjectStore+: { azureCredentials+: { connectionString+: { name: name } } } } } },
+ },
+ '#storageAccount':: d.obj(help='"The storage account where to upload data"'),
+ storageAccount: {
+ '#withKey':: d.fn(help='"The key to select"', args=[d.arg(name='key', type=d.T.string)]),
+ withKey(key): { spec+: { backup+: { barmanObjectStore+: { azureCredentials+: { storageAccount+: { key: key } } } } } },
+ '#withName':: d.fn(help='"Name of the referent."', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { spec+: { backup+: { barmanObjectStore+: { azureCredentials+: { storageAccount+: { name: name } } } } } },
+ },
+ '#storageKey':: d.obj(help='"The storage account key to be used in conjunction with the storage account name"'),
+ storageKey: {
+ '#withKey':: d.fn(help='"The key to select"', args=[d.arg(name='key', type=d.T.string)]),
+ withKey(key): { spec+: { backup+: { barmanObjectStore+: { azureCredentials+: { storageKey+: { key: key } } } } } },
+ '#withName':: d.fn(help='"Name of the referent."', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { spec+: { backup+: { barmanObjectStore+: { azureCredentials+: { storageKey+: { name: name } } } } } },
+ },
+ '#storageSasToken':: d.obj(help='"A shared-access-signature to be used in conjunction with the storage account name"'),
+ storageSasToken: {
+ '#withKey':: d.fn(help='"The key to select"', args=[d.arg(name='key', type=d.T.string)]),
+ withKey(key): { spec+: { backup+: { barmanObjectStore+: { azureCredentials+: { storageSasToken+: { key: key } } } } } },
+ '#withName':: d.fn(help='"Name of the referent."', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { spec+: { backup+: { barmanObjectStore+: { azureCredentials+: { storageSasToken+: { name: name } } } } } },
+ },
+ '#withInheritFromAzureAD':: d.fn(help='"Use the Azure AD based authentication without providing explicitly the keys."', args=[d.arg(name='inheritFromAzureAD', type=d.T.boolean)]),
+ withInheritFromAzureAD(inheritFromAzureAD): { spec+: { backup+: { barmanObjectStore+: { azureCredentials+: { inheritFromAzureAD: inheritFromAzureAD } } } } },
+ },
+ '#data':: d.obj(help='"The configuration to be used to backup the data files When not defined, base backups files will be stored uncompressed and may be unencrypted in the object store, according to the bucket default policy."'),
+ data: {
+ '#withCompression':: d.fn(help='"Compress a backup file (a tar file per tablespace) while streaming it to the object store. Available options are empty string (no compression, default), `gzip`, `bzip2` or `snappy`."', args=[d.arg(name='compression', type=d.T.string)]),
+ withCompression(compression): { spec+: { backup+: { barmanObjectStore+: { data+: { compression: compression } } } } },
+ '#withEncryption':: d.fn(help='"Whenever to force the encryption of files (if the bucket is not already configured for that). Allowed options are empty string (use the bucket policy, default), `AES256` and `aws:kms`"', args=[d.arg(name='encryption', type=d.T.string)]),
+ withEncryption(encryption): { spec+: { backup+: { barmanObjectStore+: { data+: { encryption: encryption } } } } },
+ '#withImmediateCheckpoint':: d.fn(help='"Control whether the I/O workload for the backup initial checkpoint will be limited, according to the `checkpoint_completion_target` setting on the PostgreSQL server. If set to true, an immediate checkpoint will be used, meaning PostgreSQL will complete the checkpoint as soon as possible. `false` by default."', args=[d.arg(name='immediateCheckpoint', type=d.T.boolean)]),
+ withImmediateCheckpoint(immediateCheckpoint): { spec+: { backup+: { barmanObjectStore+: { data+: { immediateCheckpoint: immediateCheckpoint } } } } },
+ '#withJobs':: d.fn(help='"The number of parallel jobs to be used to upload the backup, defaults to 2"', args=[d.arg(name='jobs', type=d.T.integer)]),
+ withJobs(jobs): { spec+: { backup+: { barmanObjectStore+: { data+: { jobs: jobs } } } } },
+ },
+ '#endpointCA':: d.obj(help='"EndpointCA store the CA bundle of the barman endpoint. Useful when using self-signed certificates to avoid errors with certificate issuer and barman-cloud-wal-archive"'),
+ endpointCA: {
+ '#withKey':: d.fn(help='"The key to select"', args=[d.arg(name='key', type=d.T.string)]),
+ withKey(key): { spec+: { backup+: { barmanObjectStore+: { endpointCA+: { key: key } } } } },
+ '#withName':: d.fn(help='"Name of the referent."', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { spec+: { backup+: { barmanObjectStore+: { endpointCA+: { name: name } } } } },
+ },
+ '#googleCredentials':: d.obj(help='"The credentials to use to upload data to Google Cloud Storage"'),
+ googleCredentials: {
+ '#applicationCredentials':: d.obj(help='"The secret containing the Google Cloud Storage JSON file with the credentials"'),
+ applicationCredentials: {
+ '#withKey':: d.fn(help='"The key to select"', args=[d.arg(name='key', type=d.T.string)]),
+ withKey(key): { spec+: { backup+: { barmanObjectStore+: { googleCredentials+: { applicationCredentials+: { key: key } } } } } },
+ '#withName':: d.fn(help='"Name of the referent."', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { spec+: { backup+: { barmanObjectStore+: { googleCredentials+: { applicationCredentials+: { name: name } } } } } },
+ },
+ '#withGkeEnvironment':: d.fn(help="\"If set to true, will presume that it's running inside a GKE environment, default to false.\"", args=[d.arg(name='gkeEnvironment', type=d.T.boolean)]),
+ withGkeEnvironment(gkeEnvironment): { spec+: { backup+: { barmanObjectStore+: { googleCredentials+: { gkeEnvironment: gkeEnvironment } } } } },
+ },
+ '#s3Credentials':: d.obj(help='"The credentials to use to upload data to S3"'),
+ s3Credentials: {
+ '#accessKeyId':: d.obj(help='"The reference to the access key id"'),
+ accessKeyId: {
+ '#withKey':: d.fn(help='"The key to select"', args=[d.arg(name='key', type=d.T.string)]),
+ withKey(key): { spec+: { backup+: { barmanObjectStore+: { s3Credentials+: { accessKeyId+: { key: key } } } } } },
+ '#withName':: d.fn(help='"Name of the referent."', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { spec+: { backup+: { barmanObjectStore+: { s3Credentials+: { accessKeyId+: { name: name } } } } } },
+ },
+ '#region':: d.obj(help='"The reference to the secret containing the region name"'),
+ region: {
+ '#withKey':: d.fn(help='"The key to select"', args=[d.arg(name='key', type=d.T.string)]),
+ withKey(key): { spec+: { backup+: { barmanObjectStore+: { s3Credentials+: { region+: { key: key } } } } } },
+ '#withName':: d.fn(help='"Name of the referent."', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { spec+: { backup+: { barmanObjectStore+: { s3Credentials+: { region+: { name: name } } } } } },
+ },
+ '#secretAccessKey':: d.obj(help='"The reference to the secret access key"'),
+ secretAccessKey: {
+ '#withKey':: d.fn(help='"The key to select"', args=[d.arg(name='key', type=d.T.string)]),
+ withKey(key): { spec+: { backup+: { barmanObjectStore+: { s3Credentials+: { secretAccessKey+: { key: key } } } } } },
+ '#withName':: d.fn(help='"Name of the referent."', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { spec+: { backup+: { barmanObjectStore+: { s3Credentials+: { secretAccessKey+: { name: name } } } } } },
+ },
+ '#sessionToken':: d.obj(help='"The references to the session key"'),
+ sessionToken: {
+ '#withKey':: d.fn(help='"The key to select"', args=[d.arg(name='key', type=d.T.string)]),
+ withKey(key): { spec+: { backup+: { barmanObjectStore+: { s3Credentials+: { sessionToken+: { key: key } } } } } },
+ '#withName':: d.fn(help='"Name of the referent."', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { spec+: { backup+: { barmanObjectStore+: { s3Credentials+: { sessionToken+: { name: name } } } } } },
+ },
+ '#withInheritFromIAMRole':: d.fn(help='"Use the role based authentication without providing explicitly the keys."', args=[d.arg(name='inheritFromIAMRole', type=d.T.boolean)]),
+ withInheritFromIAMRole(inheritFromIAMRole): { spec+: { backup+: { barmanObjectStore+: { s3Credentials+: { inheritFromIAMRole: inheritFromIAMRole } } } } },
+ },
+ '#wal':: d.obj(help='"The configuration for the backup of the WAL stream. When not defined, WAL files will be stored uncompressed and may be unencrypted in the object store, according to the bucket default policy."'),
+ wal: {
+ '#withCompression':: d.fn(help='"Compress a WAL file before sending it to the object store. Available options are empty string (no compression, default), `gzip`, `bzip2` or `snappy`."', args=[d.arg(name='compression', type=d.T.string)]),
+ withCompression(compression): { spec+: { backup+: { barmanObjectStore+: { wal+: { compression: compression } } } } },
+ '#withEncryption':: d.fn(help='"Whenever to force the encryption of files (if the bucket is not already configured for that). Allowed options are empty string (use the bucket policy, default), `AES256` and `aws:kms`"', args=[d.arg(name='encryption', type=d.T.string)]),
+ withEncryption(encryption): { spec+: { backup+: { barmanObjectStore+: { wal+: { encryption: encryption } } } } },
+ '#withMaxParallel':: d.fn(help='"Number of WAL files to be either archived in parallel (when the PostgreSQL instance is archiving to a backup object store) or restored in parallel (when a PostgreSQL standby is fetching WAL files from a recovery object store). If not specified, WAL files will be processed one at a time. It accepts a positive integer as a value - with 1 being the minimum accepted value."', args=[d.arg(name='maxParallel', type=d.T.integer)]),
+ withMaxParallel(maxParallel): { spec+: { backup+: { barmanObjectStore+: { wal+: { maxParallel: maxParallel } } } } },
+ },
+ '#withDestinationPath':: d.fn(help='"The path where to store the backup (i.e. s3://bucket/path/to/folder) this path, with different destination folders, will be used for WALs and for data"', args=[d.arg(name='destinationPath', type=d.T.string)]),
+ withDestinationPath(destinationPath): { spec+: { backup+: { barmanObjectStore+: { destinationPath: destinationPath } } } },
+ '#withEndpointURL':: d.fn(help='"Endpoint to be used to upload data to the cloud, overriding the automatic endpoint discovery"', args=[d.arg(name='endpointURL', type=d.T.string)]),
+ withEndpointURL(endpointURL): { spec+: { backup+: { barmanObjectStore+: { endpointURL: endpointURL } } } },
+ '#withHistoryTags':: d.fn(help='"HistoryTags is a list of key value pairs that will be passed to the Barman --history-tags option."', args=[d.arg(name='historyTags', type=d.T.object)]),
+ withHistoryTags(historyTags): { spec+: { backup+: { barmanObjectStore+: { historyTags: historyTags } } } },
+ '#withHistoryTagsMixin':: d.fn(help='"HistoryTags is a list of key value pairs that will be passed to the Barman --history-tags option."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='historyTags', type=d.T.object)]),
+ withHistoryTagsMixin(historyTags): { spec+: { backup+: { barmanObjectStore+: { historyTags+: historyTags } } } },
+ '#withServerName':: d.fn(help='"The server name on S3, the cluster name is used if this parameter is omitted"', args=[d.arg(name='serverName', type=d.T.string)]),
+ withServerName(serverName): { spec+: { backup+: { barmanObjectStore+: { serverName: serverName } } } },
+ '#withTags':: d.fn(help='"Tags is a list of key value pairs that will be passed to the Barman --tags option."', args=[d.arg(name='tags', type=d.T.object)]),
+ withTags(tags): { spec+: { backup+: { barmanObjectStore+: { tags: tags } } } },
+ '#withTagsMixin':: d.fn(help='"Tags is a list of key value pairs that will be passed to the Barman --tags option."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='tags', type=d.T.object)]),
+ withTagsMixin(tags): { spec+: { backup+: { barmanObjectStore+: { tags+: tags } } } },
+ },
+ '#volumeSnapshot':: d.obj(help='"VolumeSnapshot provides the configuration for the execution of volume snapshot backups."'),
+ volumeSnapshot: {
+ '#withAnnotations':: d.fn(help='"Annotations key-value pairs that will be added to .metadata.annotations snapshot resources."', args=[d.arg(name='annotations', type=d.T.object)]),
+ withAnnotations(annotations): { spec+: { backup+: { volumeSnapshot+: { annotations: annotations } } } },
+ '#withAnnotationsMixin':: d.fn(help='"Annotations key-value pairs that will be added to .metadata.annotations snapshot resources."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]),
+ withAnnotationsMixin(annotations): { spec+: { backup+: { volumeSnapshot+: { annotations+: annotations } } } },
+ '#withClassName':: d.fn(help='"ClassName specifies the Snapshot Class to be used for PG_DATA PersistentVolumeClaim. It is the default class for the other types if no specific class is present"', args=[d.arg(name='className', type=d.T.string)]),
+ withClassName(className): { spec+: { backup+: { volumeSnapshot+: { className: className } } } },
+ '#withLabels':: d.fn(help='"Labels are key-value pairs that will be added to .metadata.labels snapshot resources."', args=[d.arg(name='labels', type=d.T.object)]),
+ withLabels(labels): { spec+: { backup+: { volumeSnapshot+: { labels: labels } } } },
+ '#withLabelsMixin':: d.fn(help='"Labels are key-value pairs that will be added to .metadata.labels snapshot resources."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]),
+ withLabelsMixin(labels): { spec+: { backup+: { volumeSnapshot+: { labels+: labels } } } },
+ '#withSnapshotOwnerReference':: d.fn(help='"SnapshotOwnerReference indicates the type of owner reference the snapshot should have"', args=[d.arg(name='snapshotOwnerReference', type=d.T.string)]),
+ withSnapshotOwnerReference(snapshotOwnerReference): { spec+: { backup+: { volumeSnapshot+: { snapshotOwnerReference: snapshotOwnerReference } } } },
+ '#withWalClassName':: d.fn(help='"WalClassName specifies the Snapshot Class to be used for the PG_WAL PersistentVolumeClaim."', args=[d.arg(name='walClassName', type=d.T.string)]),
+ withWalClassName(walClassName): { spec+: { backup+: { volumeSnapshot+: { walClassName: walClassName } } } },
+ },
+ '#withRetentionPolicy':: d.fn(help="\"RetentionPolicy is the retention policy to be used for backups and WALs (i.e. '60d'). The retention policy is expressed in the form of `XXu` where `XX` is a positive integer and `u` is in `[dwm]` - days, weeks, months. It's currently only applicable when using the BarmanObjectStore method.\"", args=[d.arg(name='retentionPolicy', type=d.T.string)]),
+ withRetentionPolicy(retentionPolicy): { spec+: { backup+: { retentionPolicy: retentionPolicy } } },
+ '#withTarget':: d.fn(help='"The policy to decide which instance should perform backups. Available options are empty string, which will default to `prefer-standby` policy, `primary` to have backups run always on primary instances, `prefer-standby` to have backups run preferably on the most updated standby, if available."', args=[d.arg(name='target', type=d.T.string)]),
+ withTarget(target): { spec+: { backup+: { target: target } } },
+ },
+ '#bootstrap':: d.obj(help='"Instructions to bootstrap this cluster"'),
+ bootstrap: {
+ '#initdb':: d.obj(help='"Bootstrap the cluster via initdb"'),
+ initdb: {
+ '#import':: d.obj(help='"Bootstraps the new cluster by importing data from an existing PostgreSQL instance using logical backup (`pg_dump` and `pg_restore`)"'),
+ 'import': {
+ '#source':: d.obj(help='"The source of the import"'),
+ source: {
+ '#withExternalCluster':: d.fn(help='"The name of the externalCluster used for import"', args=[d.arg(name='externalCluster', type=d.T.string)]),
+ withExternalCluster(externalCluster): { spec+: { bootstrap+: { initdb+: { 'import'+: { source+: { externalCluster: externalCluster } } } } } },
+ },
+ '#withDatabases':: d.fn(help='"The databases to import"', args=[d.arg(name='databases', type=d.T.array)]),
+ withDatabases(databases): { spec+: { bootstrap+: { initdb+: { 'import'+: { databases: if std.isArray(v=databases) then databases else [databases] } } } } },
+ '#withDatabasesMixin':: d.fn(help='"The databases to import"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='databases', type=d.T.array)]),
+ withDatabasesMixin(databases): { spec+: { bootstrap+: { initdb+: { 'import'+: { databases+: if std.isArray(v=databases) then databases else [databases] } } } } },
+ '#withPostImportApplicationSQL':: d.fn(help='"List of SQL queries to be executed as a superuser in the application database right after is imported - to be used with extreme care (by default empty). Only available in microservice type."', args=[d.arg(name='postImportApplicationSQL', type=d.T.array)]),
+ withPostImportApplicationSQL(postImportApplicationSQL): { spec+: { bootstrap+: { initdb+: { 'import'+: { postImportApplicationSQL: if std.isArray(v=postImportApplicationSQL) then postImportApplicationSQL else [postImportApplicationSQL] } } } } },
+ '#withPostImportApplicationSQLMixin':: d.fn(help='"List of SQL queries to be executed as a superuser in the application database right after is imported - to be used with extreme care (by default empty). Only available in microservice type."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='postImportApplicationSQL', type=d.T.array)]),
+ withPostImportApplicationSQLMixin(postImportApplicationSQL): { spec+: { bootstrap+: { initdb+: { 'import'+: { postImportApplicationSQL+: if std.isArray(v=postImportApplicationSQL) then postImportApplicationSQL else [postImportApplicationSQL] } } } } },
+ '#withRoles':: d.fn(help='"The roles to import"', args=[d.arg(name='roles', type=d.T.array)]),
+ withRoles(roles): { spec+: { bootstrap+: { initdb+: { 'import'+: { roles: if std.isArray(v=roles) then roles else [roles] } } } } },
+ '#withRolesMixin':: d.fn(help='"The roles to import"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='roles', type=d.T.array)]),
+ withRolesMixin(roles): { spec+: { bootstrap+: { initdb+: { 'import'+: { roles+: if std.isArray(v=roles) then roles else [roles] } } } } },
+ '#withSchemaOnly':: d.fn(help='"When set to true, only the `pre-data` and `post-data` sections of `pg_restore` are invoked, avoiding data import. Default: `false`."', args=[d.arg(name='schemaOnly', type=d.T.boolean)]),
+ withSchemaOnly(schemaOnly): { spec+: { bootstrap+: { initdb+: { 'import'+: { schemaOnly: schemaOnly } } } } },
+ '#withType':: d.fn(help='"The import type. Can be `microservice` or `monolith`."', args=[d.arg(name='type', type=d.T.string)]),
+ withType(type): { spec+: { bootstrap+: { initdb+: { 'import'+: { type: type } } } } },
+ },
+ '#postInitApplicationSQLRefs':: d.obj(help='"PostInitApplicationSQLRefs points references to ConfigMaps or Secrets which contain SQL files, the general implementation order to these references is from all Secrets to all ConfigMaps, and inside Secrets or ConfigMaps, the implementation order is same as the order of each array (by default empty)"'),
+ postInitApplicationSQLRefs: {
+ '#configMapRefs':: d.obj(help='"ConfigMapRefs holds a list of references to ConfigMaps"'),
+ configMapRefs: {
+ '#withKey':: d.fn(help='"The key to select"', args=[d.arg(name='key', type=d.T.string)]),
+ withKey(key): { key: key },
+ '#withName':: d.fn(help='"Name of the referent."', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { name: name },
+ },
+ '#secretRefs':: d.obj(help='"SecretRefs holds a list of references to Secrets"'),
+ secretRefs: {
+ '#withKey':: d.fn(help='"The key to select"', args=[d.arg(name='key', type=d.T.string)]),
+ withKey(key): { key: key },
+ '#withName':: d.fn(help='"Name of the referent."', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { name: name },
+ },
+ '#withConfigMapRefs':: d.fn(help='"ConfigMapRefs holds a list of references to ConfigMaps"', args=[d.arg(name='configMapRefs', type=d.T.array)]),
+ withConfigMapRefs(configMapRefs): { spec+: { bootstrap+: { initdb+: { postInitApplicationSQLRefs+: { configMapRefs: if std.isArray(v=configMapRefs) then configMapRefs else [configMapRefs] } } } } },
+ '#withConfigMapRefsMixin':: d.fn(help='"ConfigMapRefs holds a list of references to ConfigMaps"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='configMapRefs', type=d.T.array)]),
+ withConfigMapRefsMixin(configMapRefs): { spec+: { bootstrap+: { initdb+: { postInitApplicationSQLRefs+: { configMapRefs+: if std.isArray(v=configMapRefs) then configMapRefs else [configMapRefs] } } } } },
+ '#withSecretRefs':: d.fn(help='"SecretRefs holds a list of references to Secrets"', args=[d.arg(name='secretRefs', type=d.T.array)]),
+ withSecretRefs(secretRefs): { spec+: { bootstrap+: { initdb+: { postInitApplicationSQLRefs+: { secretRefs: if std.isArray(v=secretRefs) then secretRefs else [secretRefs] } } } } },
+ '#withSecretRefsMixin':: d.fn(help='"SecretRefs holds a list of references to Secrets"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='secretRefs', type=d.T.array)]),
+ withSecretRefsMixin(secretRefs): { spec+: { bootstrap+: { initdb+: { postInitApplicationSQLRefs+: { secretRefs+: if std.isArray(v=secretRefs) then secretRefs else [secretRefs] } } } } },
+ },
+ '#secret':: d.obj(help='"Name of the secret containing the initial credentials for the owner of the user database. If empty a new secret will be created from scratch"'),
+ secret: {
+ '#withName':: d.fn(help='"Name of the referent."', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { spec+: { bootstrap+: { initdb+: { secret+: { name: name } } } } },
+ },
+ '#withDataChecksums':: d.fn(help='"Whether the `-k` option should be passed to initdb, enabling checksums on data pages (default: `false`)"', args=[d.arg(name='dataChecksums', type=d.T.boolean)]),
+ withDataChecksums(dataChecksums): { spec+: { bootstrap+: { initdb+: { dataChecksums: dataChecksums } } } },
+ '#withDatabase':: d.fn(help='"Name of the database used by the application. Default: `app`."', args=[d.arg(name='database', type=d.T.string)]),
+ withDatabase(database): { spec+: { bootstrap+: { initdb+: { database: database } } } },
+ '#withEncoding':: d.fn(help='"The value to be passed as option `--encoding` for initdb (default:`UTF8`)"', args=[d.arg(name='encoding', type=d.T.string)]),
+ withEncoding(encoding): { spec+: { bootstrap+: { initdb+: { encoding: encoding } } } },
+ '#withLocaleCType':: d.fn(help='"The value to be passed as option `--lc-ctype` for initdb (default:`C`)"', args=[d.arg(name='localeCType', type=d.T.string)]),
+ withLocaleCType(localeCType): { spec+: { bootstrap+: { initdb+: { localeCType: localeCType } } } },
+ '#withLocaleCollate':: d.fn(help='"The value to be passed as option `--lc-collate` for initdb (default:`C`)"', args=[d.arg(name='localeCollate', type=d.T.string)]),
+ withLocaleCollate(localeCollate): { spec+: { bootstrap+: { initdb+: { localeCollate: localeCollate } } } },
+ '#withOptions':: d.fn(help='"The list of options that must be passed to initdb when creating the cluster. Deprecated: This could lead to inconsistent configurations, please use the explicit provided parameters instead. If defined, explicit values will be ignored."', args=[d.arg(name='options', type=d.T.array)]),
+ withOptions(options): { spec+: { bootstrap+: { initdb+: { options: if std.isArray(v=options) then options else [options] } } } },
+ '#withOptionsMixin':: d.fn(help='"The list of options that must be passed to initdb when creating the cluster. Deprecated: This could lead to inconsistent configurations, please use the explicit provided parameters instead. If defined, explicit values will be ignored."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='options', type=d.T.array)]),
+ withOptionsMixin(options): { spec+: { bootstrap+: { initdb+: { options+: if std.isArray(v=options) then options else [options] } } } },
+ '#withOwner':: d.fn(help='"Name of the owner of the database in the instance to be used by applications. Defaults to the value of the `database` key."', args=[d.arg(name='owner', type=d.T.string)]),
+ withOwner(owner): { spec+: { bootstrap+: { initdb+: { owner: owner } } } },
+ '#withPostInitApplicationSQL':: d.fn(help='"List of SQL queries to be executed as a superuser in the application database right after is created - to be used with extreme care (by default empty)"', args=[d.arg(name='postInitApplicationSQL', type=d.T.array)]),
+ withPostInitApplicationSQL(postInitApplicationSQL): { spec+: { bootstrap+: { initdb+: { postInitApplicationSQL: if std.isArray(v=postInitApplicationSQL) then postInitApplicationSQL else [postInitApplicationSQL] } } } },
+ '#withPostInitApplicationSQLMixin':: d.fn(help='"List of SQL queries to be executed as a superuser in the application database right after is created - to be used with extreme care (by default empty)"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='postInitApplicationSQL', type=d.T.array)]),
+ withPostInitApplicationSQLMixin(postInitApplicationSQL): { spec+: { bootstrap+: { initdb+: { postInitApplicationSQL+: if std.isArray(v=postInitApplicationSQL) then postInitApplicationSQL else [postInitApplicationSQL] } } } },
+ '#withPostInitSQL':: d.fn(help='"List of SQL queries to be executed as a superuser immediately after the cluster has been created - to be used with extreme care (by default empty)"', args=[d.arg(name='postInitSQL', type=d.T.array)]),
+ withPostInitSQL(postInitSQL): { spec+: { bootstrap+: { initdb+: { postInitSQL: if std.isArray(v=postInitSQL) then postInitSQL else [postInitSQL] } } } },
+ '#withPostInitSQLMixin':: d.fn(help='"List of SQL queries to be executed as a superuser immediately after the cluster has been created - to be used with extreme care (by default empty)"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='postInitSQL', type=d.T.array)]),
+ withPostInitSQLMixin(postInitSQL): { spec+: { bootstrap+: { initdb+: { postInitSQL+: if std.isArray(v=postInitSQL) then postInitSQL else [postInitSQL] } } } },
+ '#withPostInitTemplateSQL':: d.fn(help='"List of SQL queries to be executed as a superuser in the `template1` after the cluster has been created - to be used with extreme care (by default empty)"', args=[d.arg(name='postInitTemplateSQL', type=d.T.array)]),
+ withPostInitTemplateSQL(postInitTemplateSQL): { spec+: { bootstrap+: { initdb+: { postInitTemplateSQL: if std.isArray(v=postInitTemplateSQL) then postInitTemplateSQL else [postInitTemplateSQL] } } } },
+ '#withPostInitTemplateSQLMixin':: d.fn(help='"List of SQL queries to be executed as a superuser in the `template1` after the cluster has been created - to be used with extreme care (by default empty)"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='postInitTemplateSQL', type=d.T.array)]),
+ withPostInitTemplateSQLMixin(postInitTemplateSQL): { spec+: { bootstrap+: { initdb+: { postInitTemplateSQL+: if std.isArray(v=postInitTemplateSQL) then postInitTemplateSQL else [postInitTemplateSQL] } } } },
+ '#withWalSegmentSize':: d.fn(help='"The value in megabytes (1 to 1024) to be passed to the `--wal-segsize` option for initdb (default: empty, resulting in PostgreSQL default: 16MB)"', args=[d.arg(name='walSegmentSize', type=d.T.integer)]),
+ withWalSegmentSize(walSegmentSize): { spec+: { bootstrap+: { initdb+: { walSegmentSize: walSegmentSize } } } },
+ },
+ '#pg_basebackup':: d.obj(help='"Bootstrap the cluster taking a physical backup of another compatible PostgreSQL instance"'),
+ pg_basebackup: {
+ '#secret':: d.obj(help='"Name of the secret containing the initial credentials for the owner of the user database. If empty a new secret will be created from scratch"'),
+ secret: {
+ '#withName':: d.fn(help='"Name of the referent."', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { spec+: { bootstrap+: { pg_basebackup+: { secret+: { name: name } } } } },
+ },
+ '#withDatabase':: d.fn(help='"Name of the database used by the application. Default: `app`."', args=[d.arg(name='database', type=d.T.string)]),
+ withDatabase(database): { spec+: { bootstrap+: { pg_basebackup+: { database: database } } } },
+ '#withOwner':: d.fn(help='"Name of the owner of the database in the instance to be used by applications. Defaults to the value of the `database` key."', args=[d.arg(name='owner', type=d.T.string)]),
+ withOwner(owner): { spec+: { bootstrap+: { pg_basebackup+: { owner: owner } } } },
+ '#withSource':: d.fn(help='"The name of the server of which we need to take a physical backup"', args=[d.arg(name='source', type=d.T.string)]),
+ withSource(source): { spec+: { bootstrap+: { pg_basebackup+: { source: source } } } },
+ },
+ '#recovery':: d.obj(help='"Bootstrap the cluster from a backup"'),
+ recovery: {
+ '#backup':: d.obj(help='"The backup object containing the physical base backup from which to initiate the recovery procedure. Mutually exclusive with `source` and `volumeSnapshots`."'),
+ backup: {
+ '#endpointCA':: d.obj(help='"EndpointCA store the CA bundle of the barman endpoint. Useful when using self-signed certificates to avoid errors with certificate issuer and barman-cloud-wal-archive."'),
+ endpointCA: {
+ '#withKey':: d.fn(help='"The key to select"', args=[d.arg(name='key', type=d.T.string)]),
+ withKey(key): { spec+: { bootstrap+: { recovery+: { backup+: { endpointCA+: { key: key } } } } } },
+ '#withName':: d.fn(help='"Name of the referent."', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { spec+: { bootstrap+: { recovery+: { backup+: { endpointCA+: { name: name } } } } } },
+ },
+ '#withName':: d.fn(help='"Name of the referent."', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { spec+: { bootstrap+: { recovery+: { backup+: { name: name } } } } },
+ },
+ '#recoveryTarget':: d.obj(help='"By default, the recovery process applies all the available WAL files in the archive (full recovery). However, you can also end the recovery as soon as a consistent state is reached or recover to a point-in-time (PITR) by specifying a `RecoveryTarget` object, as expected by PostgreSQL (i.e., timestamp, transaction Id, LSN, ...). More info: https://www.postgresql.org/docs/current/runtime-config-wal.html#RUNTIME-CONFIG-WAL-RECOVERY-TARGET"'),
+ recoveryTarget: {
+ '#withBackupID':: d.fn(help='"The ID of the backup from which to start the recovery process. If empty (default) the operator will automatically detect the backup based on targetTime or targetLSN if specified. Otherwise use the latest available backup in chronological order."', args=[d.arg(name='backupID', type=d.T.string)]),
+ withBackupID(backupID): { spec+: { bootstrap+: { recovery+: { recoveryTarget+: { backupID: backupID } } } } },
+ '#withExclusive':: d.fn(help='"Set the target to be exclusive. If omitted, defaults to false, so that in Postgres, `recovery_target_inclusive` will be true"', args=[d.arg(name='exclusive', type=d.T.boolean)]),
+ withExclusive(exclusive): { spec+: { bootstrap+: { recovery+: { recoveryTarget+: { exclusive: exclusive } } } } },
+ '#withTargetImmediate':: d.fn(help='"End recovery as soon as a consistent state is reached"', args=[d.arg(name='targetImmediate', type=d.T.boolean)]),
+ withTargetImmediate(targetImmediate): { spec+: { bootstrap+: { recovery+: { recoveryTarget+: { targetImmediate: targetImmediate } } } } },
+ '#withTargetLSN':: d.fn(help='"The target LSN (Log Sequence Number)"', args=[d.arg(name='targetLSN', type=d.T.string)]),
+ withTargetLSN(targetLSN): { spec+: { bootstrap+: { recovery+: { recoveryTarget+: { targetLSN: targetLSN } } } } },
+ '#withTargetName':: d.fn(help='"The target name (to be previously created with `pg_create_restore_point`)"', args=[d.arg(name='targetName', type=d.T.string)]),
+ withTargetName(targetName): { spec+: { bootstrap+: { recovery+: { recoveryTarget+: { targetName: targetName } } } } },
+ '#withTargetTLI':: d.fn(help='"The target timeline (\\"latest\\" or a positive integer)"', args=[d.arg(name='targetTLI', type=d.T.string)]),
+ withTargetTLI(targetTLI): { spec+: { bootstrap+: { recovery+: { recoveryTarget+: { targetTLI: targetTLI } } } } },
+ '#withTargetTime':: d.fn(help='"The target time as a timestamp in the RFC3339 standard"', args=[d.arg(name='targetTime', type=d.T.string)]),
+ withTargetTime(targetTime): { spec+: { bootstrap+: { recovery+: { recoveryTarget+: { targetTime: targetTime } } } } },
+ '#withTargetXID':: d.fn(help='"The target transaction ID"', args=[d.arg(name='targetXID', type=d.T.string)]),
+ withTargetXID(targetXID): { spec+: { bootstrap+: { recovery+: { recoveryTarget+: { targetXID: targetXID } } } } },
+ },
+ '#secret':: d.obj(help='"Name of the secret containing the initial credentials for the owner of the user database. If empty a new secret will be created from scratch"'),
+ secret: {
+ '#withName':: d.fn(help='"Name of the referent."', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { spec+: { bootstrap+: { recovery+: { secret+: { name: name } } } } },
+ },
+ '#volumeSnapshots':: d.obj(help='"The static PVC data source(s) from which to initiate the recovery procedure. Currently supporting `VolumeSnapshot` and `PersistentVolumeClaim` resources that map an existing PVC group, compatible with CloudNativePG, and taken with a cold backup copy on a fenced Postgres instance (limitation which will be removed in the future when online backup will be implemented). Mutually exclusive with `backup`."'),
+ volumeSnapshots: {
+ '#storage':: d.obj(help='"Configuration of the storage of the instances"'),
+ storage: {
+ '#withApiGroup':: d.fn(help='"APIGroup is the group for the resource being referenced. If APIGroup is not specified, the specified Kind must be in the core API group. For any other third-party types, APIGroup is required."', args=[d.arg(name='apiGroup', type=d.T.string)]),
+ withApiGroup(apiGroup): { spec+: { bootstrap+: { recovery+: { volumeSnapshots+: { storage+: { apiGroup: apiGroup } } } } } },
+ '#withKind':: d.fn(help='"Kind is the type of resource being referenced"', args=[d.arg(name='kind', type=d.T.string)]),
+ withKind(kind): { spec+: { bootstrap+: { recovery+: { volumeSnapshots+: { storage+: { kind: kind } } } } } },
+ '#withName':: d.fn(help='"Name is the name of resource being referenced"', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { spec+: { bootstrap+: { recovery+: { volumeSnapshots+: { storage+: { name: name } } } } } },
+ },
+ '#walStorage':: d.obj(help='"Configuration of the storage for PostgreSQL WAL (Write-Ahead Log)"'),
+ walStorage: {
+ '#withApiGroup':: d.fn(help='"APIGroup is the group for the resource being referenced. If APIGroup is not specified, the specified Kind must be in the core API group. For any other third-party types, APIGroup is required."', args=[d.arg(name='apiGroup', type=d.T.string)]),
+ withApiGroup(apiGroup): { spec+: { bootstrap+: { recovery+: { volumeSnapshots+: { walStorage+: { apiGroup: apiGroup } } } } } },
+ '#withKind':: d.fn(help='"Kind is the type of resource being referenced"', args=[d.arg(name='kind', type=d.T.string)]),
+ withKind(kind): { spec+: { bootstrap+: { recovery+: { volumeSnapshots+: { walStorage+: { kind: kind } } } } } },
+ '#withName':: d.fn(help='"Name is the name of resource being referenced"', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { spec+: { bootstrap+: { recovery+: { volumeSnapshots+: { walStorage+: { name: name } } } } } },
+ },
+ },
+ '#withDatabase':: d.fn(help='"Name of the database used by the application. Default: `app`."', args=[d.arg(name='database', type=d.T.string)]),
+ withDatabase(database): { spec+: { bootstrap+: { recovery+: { database: database } } } },
+ '#withOwner':: d.fn(help='"Name of the owner of the database in the instance to be used by applications. Defaults to the value of the `database` key."', args=[d.arg(name='owner', type=d.T.string)]),
+ withOwner(owner): { spec+: { bootstrap+: { recovery+: { owner: owner } } } },
+ '#withSource':: d.fn(help='"The external cluster whose backup we will restore. This is also used as the name of the folder under which the backup is stored, so it must be set to the name of the source cluster Mutually exclusive with `backup`."', args=[d.arg(name='source', type=d.T.string)]),
+ withSource(source): { spec+: { bootstrap+: { recovery+: { source: source } } } },
+ },
+ },
+ '#certificates':: d.obj(help='"The configuration for the CA and related certificates"'),
+ certificates: {
+ '#withClientCASecret':: d.fn(help='"The secret containing the Client CA certificate. If not defined, a new secret will be created with a self-signed CA and will be used to generate all the client certificates.
Contains:
- `ca.crt`: CA that should be used to validate the client certificates, used as `ssl_ca_file` of all the instances.
- `ca.key`: key used to generate client certificates, if ReplicationTLSSecret is provided, this can be omitted.
"', args=[d.arg(name='clientCASecret', type=d.T.string)]),
+ withClientCASecret(clientCASecret): { spec+: { certificates+: { clientCASecret: clientCASecret } } },
+ '#withReplicationTLSSecret':: d.fn(help='"The secret of type kubernetes.io/tls containing the client certificate to authenticate as the `streaming_replica` user. If not defined, ClientCASecret must provide also `ca.key`, and a new secret will be created using the provided CA."', args=[d.arg(name='replicationTLSSecret', type=d.T.string)]),
+ withReplicationTLSSecret(replicationTLSSecret): { spec+: { certificates+: { replicationTLSSecret: replicationTLSSecret } } },
+ '#withServerAltDNSNames':: d.fn(help='"The list of the server alternative DNS names to be added to the generated server TLS certificates, when required."', args=[d.arg(name='serverAltDNSNames', type=d.T.array)]),
+ withServerAltDNSNames(serverAltDNSNames): { spec+: { certificates+: { serverAltDNSNames: if std.isArray(v=serverAltDNSNames) then serverAltDNSNames else [serverAltDNSNames] } } },
+ '#withServerAltDNSNamesMixin':: d.fn(help='"The list of the server alternative DNS names to be added to the generated server TLS certificates, when required."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='serverAltDNSNames', type=d.T.array)]),
+ withServerAltDNSNamesMixin(serverAltDNSNames): { spec+: { certificates+: { serverAltDNSNames+: if std.isArray(v=serverAltDNSNames) then serverAltDNSNames else [serverAltDNSNames] } } },
+ '#withServerCASecret':: d.fn(help='"The secret containing the Server CA certificate. If not defined, a new secret will be created with a self-signed CA and will be used to generate the TLS certificate ServerTLSSecret.
Contains:
- `ca.crt`: CA that should be used to validate the server certificate, used as `sslrootcert` in client connection strings.
- `ca.key`: key used to generate Server SSL certs, if ServerTLSSecret is provided, this can be omitted.
"', args=[d.arg(name='serverCASecret', type=d.T.string)]),
+ withServerCASecret(serverCASecret): { spec+: { certificates+: { serverCASecret: serverCASecret } } },
+ '#withServerTLSSecret':: d.fn(help='"The secret of type kubernetes.io/tls containing the server TLS certificate and key that will be set as `ssl_cert_file` and `ssl_key_file` so that clients can connect to postgres securely. If not defined, ServerCASecret must provide also `ca.key` and a new secret will be created using the provided CA."', args=[d.arg(name='serverTLSSecret', type=d.T.string)]),
+ withServerTLSSecret(serverTLSSecret): { spec+: { certificates+: { serverTLSSecret: serverTLSSecret } } },
+ },
+ '#env':: d.obj(help='"Env follows the Env format to pass environment variables to the pods created in the cluster"'),
+ env: {
+ '#valueFrom':: d.obj(help="\"Source for the environment variable's value. Cannot be used if value is not empty.\""),
+ valueFrom: {
+ '#configMapKeyRef':: d.obj(help='"Selects a key of a ConfigMap."'),
+ configMapKeyRef: {
+ '#withKey':: d.fn(help='"The key to select."', args=[d.arg(name='key', type=d.T.string)]),
+ withKey(key): { valueFrom+: { configMapKeyRef+: { key: key } } },
+ '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { valueFrom+: { configMapKeyRef+: { name: name } } },
+ '#withOptional':: d.fn(help='"Specify whether the ConfigMap or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]),
+ withOptional(optional): { valueFrom+: { configMapKeyRef+: { optional: optional } } },
+ },
+ '#fieldRef':: d.obj(help="\"Selects a field of the pod: supports metadata.name, metadata.namespace, `metadata.labels['\u003cKEY\u003e']`, `metadata.annotations['\u003cKEY\u003e']`, spec.nodeName, spec.serviceAccountName, status.hostIP, status.podIP, status.podIPs.\""),
+ fieldRef: {
+ '#withApiVersion':: d.fn(help='"Version of the schema the FieldPath is written in terms of, defaults to \\"v1\\"."', args=[d.arg(name='apiVersion', type=d.T.string)]),
+ withApiVersion(apiVersion): { valueFrom+: { fieldRef+: { apiVersion: apiVersion } } },
+ '#withFieldPath':: d.fn(help='"Path of the field to select in the specified API version."', args=[d.arg(name='fieldPath', type=d.T.string)]),
+ withFieldPath(fieldPath): { valueFrom+: { fieldRef+: { fieldPath: fieldPath } } },
+ },
+ '#resourceFieldRef':: d.obj(help='"Selects a resource of the container: only resources limits and requests (limits.cpu, limits.memory, limits.ephemeral-storage, requests.cpu, requests.memory and requests.ephemeral-storage) are currently supported."'),
+ resourceFieldRef: {
+ '#withContainerName':: d.fn(help='"Container name: required for volumes, optional for env vars"', args=[d.arg(name='containerName', type=d.T.string)]),
+ withContainerName(containerName): { valueFrom+: { resourceFieldRef+: { containerName: containerName } } },
+ '#withDivisor':: d.fn(help='"Specifies the output format of the exposed resources, defaults to \\"1\\', args=[d.arg(name='divisor', type=d.T.any)]),
+ withDivisor(divisor): { valueFrom+: { resourceFieldRef+: { divisor: divisor } } },
+ '#withResource':: d.fn(help='"Required: resource to select"', args=[d.arg(name='resource', type=d.T.string)]),
+ withResource(resource): { valueFrom+: { resourceFieldRef+: { resource: resource } } },
+ },
+ '#secretKeyRef':: d.obj(help="\"Selects a key of a secret in the pod's namespace\""),
+ secretKeyRef: {
+ '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]),
+ withKey(key): { valueFrom+: { secretKeyRef+: { key: key } } },
+ '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { valueFrom+: { secretKeyRef+: { name: name } } },
+ '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]),
+ withOptional(optional): { valueFrom+: { secretKeyRef+: { optional: optional } } },
+ },
+ },
+ '#withName':: d.fn(help='"Name of the environment variable. Must be a C_IDENTIFIER."', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { name: name },
+ '#withValue':: d.fn(help='"Variable references $(VAR_NAME) are expanded using the previously defined environment variables in the container and any service environment variables. If a variable cannot be resolved, the reference in the input string will be unchanged. Double $$ are reduced to a single $, which allows for escaping the $(VAR_NAME) syntax: i.e. \\"$$(VAR_NAME)\\" will produce the string literal \\"$(VAR_NAME)\\". Escaped references will never be expanded, regardless of whether the variable exists or not. Defaults to \\"\\"."', args=[d.arg(name='value', type=d.T.string)]),
+ withValue(value): { value: value },
+ },
+ '#envFrom':: d.obj(help='"EnvFrom follows the EnvFrom format to pass environment variables sources to the pods to be used by Env"'),
+ envFrom: {
+ '#configMapRef':: d.obj(help='"The ConfigMap to select from"'),
+ configMapRef: {
+ '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { configMapRef+: { name: name } },
+ '#withOptional':: d.fn(help='"Specify whether the ConfigMap must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]),
+ withOptional(optional): { configMapRef+: { optional: optional } },
+ },
+ '#secretRef':: d.obj(help='"The Secret to select from"'),
+ secretRef: {
+ '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { secretRef+: { name: name } },
+ '#withOptional':: d.fn(help='"Specify whether the Secret must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]),
+ withOptional(optional): { secretRef+: { optional: optional } },
+ },
+ '#withPrefix':: d.fn(help='"An optional identifier to prepend to each key in the ConfigMap. Must be a C_IDENTIFIER."', args=[d.arg(name='prefix', type=d.T.string)]),
+ withPrefix(prefix): { prefix: prefix },
+ },
+ '#ephemeralVolumesSizeLimit':: d.obj(help='"EphemeralVolumesSizeLimit allows the user to set the limits for the ephemeral volumes"'),
+ ephemeralVolumesSizeLimit: {
+ '#withShm':: d.fn(help='"Shm is the size limit of the shared memory volume"', args=[d.arg(name='shm', type=d.T.any)]),
+ withShm(shm): { spec+: { ephemeralVolumesSizeLimit+: { shm: shm } } },
+ '#withTemporaryData':: d.fn(help='"TemporaryData is the size limit of the temporary data volume"', args=[d.arg(name='temporaryData', type=d.T.any)]),
+ withTemporaryData(temporaryData): { spec+: { ephemeralVolumesSizeLimit+: { temporaryData: temporaryData } } },
+ },
+ '#externalClusters':: d.obj(help='"The list of external clusters which are used in the configuration"'),
+ externalClusters: {
+ '#barmanObjectStore':: d.obj(help='"The configuration for the barman-cloud tool suite"'),
+ barmanObjectStore: {
+ '#azureCredentials':: d.obj(help='"The credentials to use to upload data to Azure Blob Storage"'),
+ azureCredentials: {
+ '#connectionString':: d.obj(help='"The connection string to be used"'),
+ connectionString: {
+ '#withKey':: d.fn(help='"The key to select"', args=[d.arg(name='key', type=d.T.string)]),
+ withKey(key): { barmanObjectStore+: { azureCredentials+: { connectionString+: { key: key } } } },
+ '#withName':: d.fn(help='"Name of the referent."', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { barmanObjectStore+: { azureCredentials+: { connectionString+: { name: name } } } },
+ },
+ '#storageAccount':: d.obj(help='"The storage account where to upload data"'),
+ storageAccount: {
+ '#withKey':: d.fn(help='"The key to select"', args=[d.arg(name='key', type=d.T.string)]),
+ withKey(key): { barmanObjectStore+: { azureCredentials+: { storageAccount+: { key: key } } } },
+ '#withName':: d.fn(help='"Name of the referent."', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { barmanObjectStore+: { azureCredentials+: { storageAccount+: { name: name } } } },
+ },
+ '#storageKey':: d.obj(help='"The storage account key to be used in conjunction with the storage account name"'),
+ storageKey: {
+ '#withKey':: d.fn(help='"The key to select"', args=[d.arg(name='key', type=d.T.string)]),
+ withKey(key): { barmanObjectStore+: { azureCredentials+: { storageKey+: { key: key } } } },
+ '#withName':: d.fn(help='"Name of the referent."', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { barmanObjectStore+: { azureCredentials+: { storageKey+: { name: name } } } },
+ },
+ '#storageSasToken':: d.obj(help='"A shared-access-signature to be used in conjunction with the storage account name"'),
+ storageSasToken: {
+ '#withKey':: d.fn(help='"The key to select"', args=[d.arg(name='key', type=d.T.string)]),
+ withKey(key): { barmanObjectStore+: { azureCredentials+: { storageSasToken+: { key: key } } } },
+ '#withName':: d.fn(help='"Name of the referent."', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { barmanObjectStore+: { azureCredentials+: { storageSasToken+: { name: name } } } },
+ },
+ '#withInheritFromAzureAD':: d.fn(help='"Use the Azure AD based authentication without providing explicitly the keys."', args=[d.arg(name='inheritFromAzureAD', type=d.T.boolean)]),
+ withInheritFromAzureAD(inheritFromAzureAD): { barmanObjectStore+: { azureCredentials+: { inheritFromAzureAD: inheritFromAzureAD } } },
+ },
+ '#data':: d.obj(help='"The configuration to be used to backup the data files When not defined, base backups files will be stored uncompressed and may be unencrypted in the object store, according to the bucket default policy."'),
+ data: {
+ '#withCompression':: d.fn(help='"Compress a backup file (a tar file per tablespace) while streaming it to the object store. Available options are empty string (no compression, default), `gzip`, `bzip2` or `snappy`."', args=[d.arg(name='compression', type=d.T.string)]),
+ withCompression(compression): { barmanObjectStore+: { data+: { compression: compression } } },
+ '#withEncryption':: d.fn(help='"Whenever to force the encryption of files (if the bucket is not already configured for that). Allowed options are empty string (use the bucket policy, default), `AES256` and `aws:kms`"', args=[d.arg(name='encryption', type=d.T.string)]),
+ withEncryption(encryption): { barmanObjectStore+: { data+: { encryption: encryption } } },
+ '#withImmediateCheckpoint':: d.fn(help='"Control whether the I/O workload for the backup initial checkpoint will be limited, according to the `checkpoint_completion_target` setting on the PostgreSQL server. If set to true, an immediate checkpoint will be used, meaning PostgreSQL will complete the checkpoint as soon as possible. `false` by default."', args=[d.arg(name='immediateCheckpoint', type=d.T.boolean)]),
+ withImmediateCheckpoint(immediateCheckpoint): { barmanObjectStore+: { data+: { immediateCheckpoint: immediateCheckpoint } } },
+ '#withJobs':: d.fn(help='"The number of parallel jobs to be used to upload the backup, defaults to 2"', args=[d.arg(name='jobs', type=d.T.integer)]),
+ withJobs(jobs): { barmanObjectStore+: { data+: { jobs: jobs } } },
+ },
+ '#endpointCA':: d.obj(help='"EndpointCA store the CA bundle of the barman endpoint. Useful when using self-signed certificates to avoid errors with certificate issuer and barman-cloud-wal-archive"'),
+ endpointCA: {
+ '#withKey':: d.fn(help='"The key to select"', args=[d.arg(name='key', type=d.T.string)]),
+ withKey(key): { barmanObjectStore+: { endpointCA+: { key: key } } },
+ '#withName':: d.fn(help='"Name of the referent."', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { barmanObjectStore+: { endpointCA+: { name: name } } },
+ },
+ '#googleCredentials':: d.obj(help='"The credentials to use to upload data to Google Cloud Storage"'),
+ googleCredentials: {
+ '#applicationCredentials':: d.obj(help='"The secret containing the Google Cloud Storage JSON file with the credentials"'),
+ applicationCredentials: {
+ '#withKey':: d.fn(help='"The key to select"', args=[d.arg(name='key', type=d.T.string)]),
+ withKey(key): { barmanObjectStore+: { googleCredentials+: { applicationCredentials+: { key: key } } } },
+ '#withName':: d.fn(help='"Name of the referent."', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { barmanObjectStore+: { googleCredentials+: { applicationCredentials+: { name: name } } } },
+ },
+ '#withGkeEnvironment':: d.fn(help="\"If set to true, will presume that it's running inside a GKE environment, default to false.\"", args=[d.arg(name='gkeEnvironment', type=d.T.boolean)]),
+ withGkeEnvironment(gkeEnvironment): { barmanObjectStore+: { googleCredentials+: { gkeEnvironment: gkeEnvironment } } },
+ },
+ '#s3Credentials':: d.obj(help='"The credentials to use to upload data to S3"'),
+ s3Credentials: {
+ '#accessKeyId':: d.obj(help='"The reference to the access key id"'),
+ accessKeyId: {
+ '#withKey':: d.fn(help='"The key to select"', args=[d.arg(name='key', type=d.T.string)]),
+ withKey(key): { barmanObjectStore+: { s3Credentials+: { accessKeyId+: { key: key } } } },
+ '#withName':: d.fn(help='"Name of the referent."', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { barmanObjectStore+: { s3Credentials+: { accessKeyId+: { name: name } } } },
+ },
+ '#region':: d.obj(help='"The reference to the secret containing the region name"'),
+ region: {
+ '#withKey':: d.fn(help='"The key to select"', args=[d.arg(name='key', type=d.T.string)]),
+ withKey(key): { barmanObjectStore+: { s3Credentials+: { region+: { key: key } } } },
+ '#withName':: d.fn(help='"Name of the referent."', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { barmanObjectStore+: { s3Credentials+: { region+: { name: name } } } },
+ },
+ '#secretAccessKey':: d.obj(help='"The reference to the secret access key"'),
+ secretAccessKey: {
+ '#withKey':: d.fn(help='"The key to select"', args=[d.arg(name='key', type=d.T.string)]),
+ withKey(key): { barmanObjectStore+: { s3Credentials+: { secretAccessKey+: { key: key } } } },
+ '#withName':: d.fn(help='"Name of the referent."', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { barmanObjectStore+: { s3Credentials+: { secretAccessKey+: { name: name } } } },
+ },
+ '#sessionToken':: d.obj(help='"The references to the session key"'),
+ sessionToken: {
+ '#withKey':: d.fn(help='"The key to select"', args=[d.arg(name='key', type=d.T.string)]),
+ withKey(key): { barmanObjectStore+: { s3Credentials+: { sessionToken+: { key: key } } } },
+ '#withName':: d.fn(help='"Name of the referent."', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { barmanObjectStore+: { s3Credentials+: { sessionToken+: { name: name } } } },
+ },
+ '#withInheritFromIAMRole':: d.fn(help='"Use the role based authentication without providing explicitly the keys."', args=[d.arg(name='inheritFromIAMRole', type=d.T.boolean)]),
+ withInheritFromIAMRole(inheritFromIAMRole): { barmanObjectStore+: { s3Credentials+: { inheritFromIAMRole: inheritFromIAMRole } } },
+ },
+ '#wal':: d.obj(help='"The configuration for the backup of the WAL stream. When not defined, WAL files will be stored uncompressed and may be unencrypted in the object store, according to the bucket default policy."'),
+ wal: {
+ '#withCompression':: d.fn(help='"Compress a WAL file before sending it to the object store. Available options are empty string (no compression, default), `gzip`, `bzip2` or `snappy`."', args=[d.arg(name='compression', type=d.T.string)]),
+ withCompression(compression): { barmanObjectStore+: { wal+: { compression: compression } } },
+ '#withEncryption':: d.fn(help='"Whenever to force the encryption of files (if the bucket is not already configured for that). Allowed options are empty string (use the bucket policy, default), `AES256` and `aws:kms`"', args=[d.arg(name='encryption', type=d.T.string)]),
+ withEncryption(encryption): { barmanObjectStore+: { wal+: { encryption: encryption } } },
+ '#withMaxParallel':: d.fn(help='"Number of WAL files to be either archived in parallel (when the PostgreSQL instance is archiving to a backup object store) or restored in parallel (when a PostgreSQL standby is fetching WAL files from a recovery object store). If not specified, WAL files will be processed one at a time. It accepts a positive integer as a value - with 1 being the minimum accepted value."', args=[d.arg(name='maxParallel', type=d.T.integer)]),
+ withMaxParallel(maxParallel): { barmanObjectStore+: { wal+: { maxParallel: maxParallel } } },
+ },
+ '#withDestinationPath':: d.fn(help='"The path where to store the backup (i.e. s3://bucket/path/to/folder) this path, with different destination folders, will be used for WALs and for data"', args=[d.arg(name='destinationPath', type=d.T.string)]),
+ withDestinationPath(destinationPath): { barmanObjectStore+: { destinationPath: destinationPath } },
+ '#withEndpointURL':: d.fn(help='"Endpoint to be used to upload data to the cloud, overriding the automatic endpoint discovery"', args=[d.arg(name='endpointURL', type=d.T.string)]),
+ withEndpointURL(endpointURL): { barmanObjectStore+: { endpointURL: endpointURL } },
+ '#withHistoryTags':: d.fn(help='"HistoryTags is a list of key value pairs that will be passed to the Barman --history-tags option."', args=[d.arg(name='historyTags', type=d.T.object)]),
+ withHistoryTags(historyTags): { barmanObjectStore+: { historyTags: historyTags } },
+ '#withHistoryTagsMixin':: d.fn(help='"HistoryTags is a list of key value pairs that will be passed to the Barman --history-tags option."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='historyTags', type=d.T.object)]),
+ withHistoryTagsMixin(historyTags): { barmanObjectStore+: { historyTags+: historyTags } },
+ '#withServerName':: d.fn(help='"The server name on S3, the cluster name is used if this parameter is omitted"', args=[d.arg(name='serverName', type=d.T.string)]),
+ withServerName(serverName): { barmanObjectStore+: { serverName: serverName } },
+ '#withTags':: d.fn(help='"Tags is a list of key value pairs that will be passed to the Barman --tags option."', args=[d.arg(name='tags', type=d.T.object)]),
+ withTags(tags): { barmanObjectStore+: { tags: tags } },
+ '#withTagsMixin':: d.fn(help='"Tags is a list of key value pairs that will be passed to the Barman --tags option."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='tags', type=d.T.object)]),
+ withTagsMixin(tags): { barmanObjectStore+: { tags+: tags } },
+ },
+ '#password':: d.obj(help='"The reference to the password to be used to connect to the server"'),
+ password: {
+ '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]),
+ withKey(key): { password+: { key: key } },
+ '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { password+: { name: name } },
+ '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]),
+ withOptional(optional): { password+: { optional: optional } },
+ },
+ '#sslCert':: d.obj(help='"The reference to an SSL certificate to be used to connect to this instance"'),
+ sslCert: {
+ '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]),
+ withKey(key): { sslCert+: { key: key } },
+ '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { sslCert+: { name: name } },
+ '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]),
+ withOptional(optional): { sslCert+: { optional: optional } },
+ },
+ '#sslKey':: d.obj(help='"The reference to an SSL private key to be used to connect to this instance"'),
+ sslKey: {
+ '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]),
+ withKey(key): { sslKey+: { key: key } },
+ '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { sslKey+: { name: name } },
+ '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]),
+ withOptional(optional): { sslKey+: { optional: optional } },
+ },
+ '#sslRootCert':: d.obj(help='"The reference to an SSL CA public key to be used to connect to this instance"'),
+ sslRootCert: {
+ '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]),
+ withKey(key): { sslRootCert+: { key: key } },
+ '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { sslRootCert+: { name: name } },
+ '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]),
+ withOptional(optional): { sslRootCert+: { optional: optional } },
+ },
+ '#withConnectionParameters':: d.fn(help='"The list of connection parameters, such as dbname, host, username, etc"', args=[d.arg(name='connectionParameters', type=d.T.object)]),
+ withConnectionParameters(connectionParameters): { connectionParameters: connectionParameters },
+ '#withConnectionParametersMixin':: d.fn(help='"The list of connection parameters, such as dbname, host, username, etc"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='connectionParameters', type=d.T.object)]),
+ withConnectionParametersMixin(connectionParameters): { connectionParameters+: connectionParameters },
+ '#withName':: d.fn(help='"The server name, required"', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { name: name },
+ },
+ '#imagePullSecrets':: d.obj(help='"The list of pull secrets to be used to pull the images"'),
+ imagePullSecrets: {
+ '#withName':: d.fn(help='"Name of the referent."', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { name: name },
+ },
+ '#inheritedMetadata':: d.obj(help='"Metadata that will be inherited by all objects related to the Cluster"'),
+ inheritedMetadata: {
+ '#withAnnotations':: d.fn(help='', args=[d.arg(name='annotations', type=d.T.object)]),
+ withAnnotations(annotations): { spec+: { inheritedMetadata+: { annotations: annotations } } },
+ '#withAnnotationsMixin':: d.fn(help='\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]),
+ withAnnotationsMixin(annotations): { spec+: { inheritedMetadata+: { annotations+: annotations } } },
+ '#withLabels':: d.fn(help='', args=[d.arg(name='labels', type=d.T.object)]),
+ withLabels(labels): { spec+: { inheritedMetadata+: { labels: labels } } },
+ '#withLabelsMixin':: d.fn(help='\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]),
+ withLabelsMixin(labels): { spec+: { inheritedMetadata+: { labels+: labels } } },
+ },
+ '#managed':: d.obj(help='"The configuration that is used by the portions of PostgreSQL that are managed by the instance manager"'),
+ managed: {
+ '#roles':: d.obj(help='"Database roles managed by the `Cluster`"'),
+ roles: {
+ '#passwordSecret':: d.obj(help='"Secret containing the password of the role (if present) If null, the password will be ignored unless DisablePassword is set"'),
+ passwordSecret: {
+ '#withName':: d.fn(help='"Name of the referent."', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { passwordSecret+: { name: name } },
+ },
+ '#withBypassrls':: d.fn(help='"Whether a role bypasses every row-level security (RLS) policy. Default is `false`."', args=[d.arg(name='bypassrls', type=d.T.boolean)]),
+ withBypassrls(bypassrls): { bypassrls: bypassrls },
+ '#withComment':: d.fn(help='"Description of the role"', args=[d.arg(name='comment', type=d.T.string)]),
+ withComment(comment): { comment: comment },
+ '#withConnectionLimit':: d.fn(help='"If the role can log in, this specifies how many concurrent connections the role can make. `-1` (the default) means no limit."', args=[d.arg(name='connectionLimit', type=d.T.integer)]),
+ withConnectionLimit(connectionLimit): { connectionLimit: connectionLimit },
+ '#withCreatedb':: d.fn(help='"When set to `true`, the role being defined will be allowed to create new databases. Specifying `false` (default) will deny a role the ability to create databases."', args=[d.arg(name='createdb', type=d.T.boolean)]),
+ withCreatedb(createdb): { createdb: createdb },
+ '#withCreaterole':: d.fn(help='"Whether the role will be permitted to create, alter, drop, comment on, change the security label for, and grant or revoke membership in other roles. Default is `false`."', args=[d.arg(name='createrole', type=d.T.boolean)]),
+ withCreaterole(createrole): { createrole: createrole },
+ '#withDisablePassword':: d.fn(help="\"DisablePassword indicates that a role's password should be set to NULL in Postgres\"", args=[d.arg(name='disablePassword', type=d.T.boolean)]),
+ withDisablePassword(disablePassword): { disablePassword: disablePassword },
+ '#withEnsure':: d.fn(help='"Ensure the role is `present` or `absent` - defaults to \\"present\\', args=[d.arg(name='ensure', type=d.T.string)]),
+ withEnsure(ensure): { ensure: ensure },
+ '#withInRoles':: d.fn(help='"List of one or more existing roles to which this role will be immediately added as a new member. Default empty."', args=[d.arg(name='inRoles', type=d.T.array)]),
+ withInRoles(inRoles): { inRoles: if std.isArray(v=inRoles) then inRoles else [inRoles] },
+ '#withInRolesMixin':: d.fn(help='"List of one or more existing roles to which this role will be immediately added as a new member. Default empty."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='inRoles', type=d.T.array)]),
+ withInRolesMixin(inRoles): { inRoles+: if std.isArray(v=inRoles) then inRoles else [inRoles] },
+ '#withInherit':: d.fn(help='"Whether a role \\"inherits\\" the privileges of roles it is a member of. Defaults is `true`."', args=[d.arg(name='inherit', type=d.T.boolean)]),
+ withInherit(inherit): { inherit: inherit },
+ '#withLogin':: d.fn(help='"Whether the role is allowed to log in. A role having the `login` attribute can be thought of as a user. Roles without this attribute are useful for managing database privileges, but are not users in the usual sense of the word. Default is `false`."', args=[d.arg(name='login', type=d.T.boolean)]),
+ withLogin(login): { login: login },
+ '#withName':: d.fn(help='"Name of the role"', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { name: name },
+ '#withReplication':: d.fn(help='"Whether a role is a replication role. A role must have this attribute (or be a superuser) in order to be able to connect to the server in replication mode (physical or logical replication) and in order to be able to create or drop replication slots. A role having the `replication` attribute is a very highly privileged role, and should only be used on roles actually used for replication. Default is `false`."', args=[d.arg(name='replication', type=d.T.boolean)]),
+ withReplication(replication): { replication: replication },
+ '#withSuperuser':: d.fn(help='"Whether the role is a `superuser` who can override all access restrictions within the database - superuser status is dangerous and should be used only when really needed. You must yourself be a superuser to create a new superuser. Defaults is `false`."', args=[d.arg(name='superuser', type=d.T.boolean)]),
+ withSuperuser(superuser): { superuser: superuser },
+ '#withValidUntil':: d.fn(help="\"Date and time after which the role's password is no longer valid. When omitted, the password will never expire (default).\"", args=[d.arg(name='validUntil', type=d.T.string)]),
+ withValidUntil(validUntil): { validUntil: validUntil },
+ },
+ '#withRoles':: d.fn(help='"Database roles managed by the `Cluster`"', args=[d.arg(name='roles', type=d.T.array)]),
+ withRoles(roles): { spec+: { managed+: { roles: if std.isArray(v=roles) then roles else [roles] } } },
+ '#withRolesMixin':: d.fn(help='"Database roles managed by the `Cluster`"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='roles', type=d.T.array)]),
+ withRolesMixin(roles): { spec+: { managed+: { roles+: if std.isArray(v=roles) then roles else [roles] } } },
+ },
+ '#monitoring':: d.obj(help='"The configuration of the monitoring infrastructure of this cluster"'),
+ monitoring: {
+ '#customQueriesConfigMap':: d.obj(help='"The list of config maps containing the custom queries"'),
+ customQueriesConfigMap: {
+ '#withKey':: d.fn(help='"The key to select"', args=[d.arg(name='key', type=d.T.string)]),
+ withKey(key): { key: key },
+ '#withName':: d.fn(help='"Name of the referent."', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { name: name },
+ },
+ '#customQueriesSecret':: d.obj(help='"The list of secrets containing the custom queries"'),
+ customQueriesSecret: {
+ '#withKey':: d.fn(help='"The key to select"', args=[d.arg(name='key', type=d.T.string)]),
+ withKey(key): { key: key },
+ '#withName':: d.fn(help='"Name of the referent."', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { name: name },
+ },
+ '#withCustomQueriesConfigMap':: d.fn(help='"The list of config maps containing the custom queries"', args=[d.arg(name='customQueriesConfigMap', type=d.T.array)]),
+ withCustomQueriesConfigMap(customQueriesConfigMap): { spec+: { monitoring+: { customQueriesConfigMap: if std.isArray(v=customQueriesConfigMap) then customQueriesConfigMap else [customQueriesConfigMap] } } },
+ '#withCustomQueriesConfigMapMixin':: d.fn(help='"The list of config maps containing the custom queries"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='customQueriesConfigMap', type=d.T.array)]),
+ withCustomQueriesConfigMapMixin(customQueriesConfigMap): { spec+: { monitoring+: { customQueriesConfigMap+: if std.isArray(v=customQueriesConfigMap) then customQueriesConfigMap else [customQueriesConfigMap] } } },
+ '#withCustomQueriesSecret':: d.fn(help='"The list of secrets containing the custom queries"', args=[d.arg(name='customQueriesSecret', type=d.T.array)]),
+ withCustomQueriesSecret(customQueriesSecret): { spec+: { monitoring+: { customQueriesSecret: if std.isArray(v=customQueriesSecret) then customQueriesSecret else [customQueriesSecret] } } },
+ '#withCustomQueriesSecretMixin':: d.fn(help='"The list of secrets containing the custom queries"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='customQueriesSecret', type=d.T.array)]),
+ withCustomQueriesSecretMixin(customQueriesSecret): { spec+: { monitoring+: { customQueriesSecret+: if std.isArray(v=customQueriesSecret) then customQueriesSecret else [customQueriesSecret] } } },
+ '#withDisableDefaultQueries':: d.fn(help="\"Whether the default queries should be injected. Set it to `true` if you don't want to inject default queries into the cluster. Default: false.\"", args=[d.arg(name='disableDefaultQueries', type=d.T.boolean)]),
+ withDisableDefaultQueries(disableDefaultQueries): { spec+: { monitoring+: { disableDefaultQueries: disableDefaultQueries } } },
+ '#withEnablePodMonitor':: d.fn(help='"Enable or disable the `PodMonitor`"', args=[d.arg(name='enablePodMonitor', type=d.T.boolean)]),
+ withEnablePodMonitor(enablePodMonitor): { spec+: { monitoring+: { enablePodMonitor: enablePodMonitor } } },
+ },
+ '#nodeMaintenanceWindow':: d.obj(help='"Define a maintenance window for the Kubernetes nodes"'),
+ nodeMaintenanceWindow: {
+ '#withInProgress':: d.fn(help='"Is there a node maintenance activity in progress?"', args=[d.arg(name='inProgress', type=d.T.boolean)]),
+ withInProgress(inProgress): { spec+: { nodeMaintenanceWindow+: { inProgress: inProgress } } },
+ '#withReusePVC':: d.fn(help='"Reuse the existing PVC (wait for the node to come up again) or not (recreate it elsewhere - when `instances` >1)"', args=[d.arg(name='reusePVC', type=d.T.boolean)]),
+ withReusePVC(reusePVC): { spec+: { nodeMaintenanceWindow+: { reusePVC: reusePVC } } },
+ },
+ '#postgresql':: d.obj(help='"Configuration of the PostgreSQL server"'),
+ postgresql: {
+ '#ldap':: d.obj(help='"Options to specify LDAP configuration"'),
+ ldap: {
+ '#bindAsAuth':: d.obj(help='"Bind as authentication configuration"'),
+ bindAsAuth: {
+ '#withPrefix':: d.fn(help='"Prefix for the bind authentication option"', args=[d.arg(name='prefix', type=d.T.string)]),
+ withPrefix(prefix): { spec+: { postgresql+: { ldap+: { bindAsAuth+: { prefix: prefix } } } } },
+ '#withSuffix':: d.fn(help='"Suffix for the bind authentication option"', args=[d.arg(name='suffix', type=d.T.string)]),
+ withSuffix(suffix): { spec+: { postgresql+: { ldap+: { bindAsAuth+: { suffix: suffix } } } } },
+ },
+ '#bindSearchAuth':: d.obj(help='"Bind+Search authentication configuration"'),
+ bindSearchAuth: {
+ '#bindPassword':: d.obj(help='"Secret with the password for the user to bind to the directory"'),
+ bindPassword: {
+ '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]),
+ withKey(key): { spec+: { postgresql+: { ldap+: { bindSearchAuth+: { bindPassword+: { key: key } } } } } },
+ '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { spec+: { postgresql+: { ldap+: { bindSearchAuth+: { bindPassword+: { name: name } } } } } },
+ '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]),
+ withOptional(optional): { spec+: { postgresql+: { ldap+: { bindSearchAuth+: { bindPassword+: { optional: optional } } } } } },
+ },
+ '#withBaseDN':: d.fn(help='"Root DN to begin the user search"', args=[d.arg(name='baseDN', type=d.T.string)]),
+ withBaseDN(baseDN): { spec+: { postgresql+: { ldap+: { bindSearchAuth+: { baseDN: baseDN } } } } },
+ '#withBindDN':: d.fn(help='"DN of the user to bind to the directory"', args=[d.arg(name='bindDN', type=d.T.string)]),
+ withBindDN(bindDN): { spec+: { postgresql+: { ldap+: { bindSearchAuth+: { bindDN: bindDN } } } } },
+ '#withSearchAttribute':: d.fn(help='"Attribute to match against the username"', args=[d.arg(name='searchAttribute', type=d.T.string)]),
+ withSearchAttribute(searchAttribute): { spec+: { postgresql+: { ldap+: { bindSearchAuth+: { searchAttribute: searchAttribute } } } } },
+ '#withSearchFilter':: d.fn(help='"Search filter to use when doing the search+bind authentication"', args=[d.arg(name='searchFilter', type=d.T.string)]),
+ withSearchFilter(searchFilter): { spec+: { postgresql+: { ldap+: { bindSearchAuth+: { searchFilter: searchFilter } } } } },
+ },
+ '#withPort':: d.fn(help='"LDAP server port"', args=[d.arg(name='port', type=d.T.integer)]),
+ withPort(port): { spec+: { postgresql+: { ldap+: { port: port } } } },
+ '#withScheme':: d.fn(help='"LDAP schema to be used, possible options are `ldap` and `ldaps`"', args=[d.arg(name='scheme', type=d.T.string)]),
+ withScheme(scheme): { spec+: { postgresql+: { ldap+: { scheme: scheme } } } },
+ '#withServer':: d.fn(help='"LDAP hostname or IP address"', args=[d.arg(name='server', type=d.T.string)]),
+ withServer(server): { spec+: { postgresql+: { ldap+: { server: server } } } },
+ '#withTls':: d.fn(help="\"Set to 'true' to enable LDAP over TLS. 'false' is default\"", args=[d.arg(name='tls', type=d.T.boolean)]),
+ withTls(tls): { spec+: { postgresql+: { ldap+: { tls: tls } } } },
+ },
+ '#syncReplicaElectionConstraint':: d.obj(help='"Requirements to be met by sync replicas. This will affect how the \\"synchronous_standby_names\\" parameter will be set up."'),
+ syncReplicaElectionConstraint: {
+ '#withEnabled':: d.fn(help='"This flag enables the constraints for sync replicas"', args=[d.arg(name='enabled', type=d.T.boolean)]),
+ withEnabled(enabled): { spec+: { postgresql+: { syncReplicaElectionConstraint+: { enabled: enabled } } } },
+ '#withNodeLabelsAntiAffinity':: d.fn(help='"A list of node labels values to extract and compare to evaluate if the pods reside in the same topology or not"', args=[d.arg(name='nodeLabelsAntiAffinity', type=d.T.array)]),
+ withNodeLabelsAntiAffinity(nodeLabelsAntiAffinity): { spec+: { postgresql+: { syncReplicaElectionConstraint+: { nodeLabelsAntiAffinity: if std.isArray(v=nodeLabelsAntiAffinity) then nodeLabelsAntiAffinity else [nodeLabelsAntiAffinity] } } } },
+ '#withNodeLabelsAntiAffinityMixin':: d.fn(help='"A list of node labels values to extract and compare to evaluate if the pods reside in the same topology or not"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='nodeLabelsAntiAffinity', type=d.T.array)]),
+ withNodeLabelsAntiAffinityMixin(nodeLabelsAntiAffinity): { spec+: { postgresql+: { syncReplicaElectionConstraint+: { nodeLabelsAntiAffinity+: if std.isArray(v=nodeLabelsAntiAffinity) then nodeLabelsAntiAffinity else [nodeLabelsAntiAffinity] } } } },
+ },
+ '#withParameters':: d.fn(help='"PostgreSQL configuration options (postgresql.conf)"', args=[d.arg(name='parameters', type=d.T.object)]),
+ withParameters(parameters): { spec+: { postgresql+: { parameters: parameters } } },
+ '#withParametersMixin':: d.fn(help='"PostgreSQL configuration options (postgresql.conf)"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='parameters', type=d.T.object)]),
+ withParametersMixin(parameters): { spec+: { postgresql+: { parameters+: parameters } } },
+ '#withPg_hba':: d.fn(help='"PostgreSQL Host Based Authentication rules (lines to be appended to the pg_hba.conf file)"', args=[d.arg(name='pg_hba', type=d.T.array)]),
+ withPg_hba(pg_hba): { spec+: { postgresql+: { pg_hba: if std.isArray(v=pg_hba) then pg_hba else [pg_hba] } } },
+ '#withPg_hbaMixin':: d.fn(help='"PostgreSQL Host Based Authentication rules (lines to be appended to the pg_hba.conf file)"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='pg_hba', type=d.T.array)]),
+ withPg_hbaMixin(pg_hba): { spec+: { postgresql+: { pg_hba+: if std.isArray(v=pg_hba) then pg_hba else [pg_hba] } } },
+ '#withPromotionTimeout':: d.fn(help='"Specifies the maximum number of seconds to wait when promoting an instance to primary. Default value is 40000000, greater than one year in seconds, big enough to simulate an infinite timeout"', args=[d.arg(name='promotionTimeout', type=d.T.integer)]),
+ withPromotionTimeout(promotionTimeout): { spec+: { postgresql+: { promotionTimeout: promotionTimeout } } },
+ '#withShared_preload_libraries':: d.fn(help='"Lists of shared preload libraries to add to the default ones"', args=[d.arg(name='shared_preload_libraries', type=d.T.array)]),
+ withShared_preload_libraries(shared_preload_libraries): { spec+: { postgresql+: { shared_preload_libraries: if std.isArray(v=shared_preload_libraries) then shared_preload_libraries else [shared_preload_libraries] } } },
+ '#withShared_preload_librariesMixin':: d.fn(help='"Lists of shared preload libraries to add to the default ones"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='shared_preload_libraries', type=d.T.array)]),
+ withShared_preload_librariesMixin(shared_preload_libraries): { spec+: { postgresql+: { shared_preload_libraries+: if std.isArray(v=shared_preload_libraries) then shared_preload_libraries else [shared_preload_libraries] } } },
+ },
+ '#projectedVolumeTemplate':: d.obj(help='"Template to be used to define projected volumes, projected volumes will be mounted under `/projected` base folder"'),
+ projectedVolumeTemplate: {
+ '#sources':: d.obj(help='"sources is the list of volume projections"'),
+ sources: {
+ '#configMap':: d.obj(help='"configMap information about the configMap data to project"'),
+ configMap: {
+ '#items':: d.obj(help="\"items if unspecified, each key-value pair in the Data field of the referenced ConfigMap will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the ConfigMap, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'.\""),
+ items: {
+ '#withKey':: d.fn(help='"key is the key to project."', args=[d.arg(name='key', type=d.T.string)]),
+ withKey(key): { key: key },
+ '#withMode':: d.fn(help='"mode is Optional: mode bits used to set permissions on this file. Must be an octal value between 0000 and 0777 or a decimal value between 0 and 511. YAML accepts both octal and decimal values, JSON requires decimal values for mode bits. If not specified, the volume defaultMode will be used. This might be in conflict with other options that affect the file mode, like fsGroup, and the result can be other mode bits set."', args=[d.arg(name='mode', type=d.T.integer)]),
+ withMode(mode): { mode: mode },
+ '#withPath':: d.fn(help="\"path is the relative path of the file to map the key to. May not be an absolute path. May not contain the path element '..'. May not start with the string '..'.\"", args=[d.arg(name='path', type=d.T.string)]),
+ withPath(path): { path: path },
+ },
+ '#withItems':: d.fn(help="\"items if unspecified, each key-value pair in the Data field of the referenced ConfigMap will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the ConfigMap, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'.\"", args=[d.arg(name='items', type=d.T.array)]),
+ withItems(items): { configMap+: { items: if std.isArray(v=items) then items else [items] } },
+ '#withItemsMixin':: d.fn(help="\"items if unspecified, each key-value pair in the Data field of the referenced ConfigMap will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the ConfigMap, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='items', type=d.T.array)]),
+ withItemsMixin(items): { configMap+: { items+: if std.isArray(v=items) then items else [items] } },
+ '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { configMap+: { name: name } },
+ '#withOptional':: d.fn(help='"optional specify whether the ConfigMap or its keys must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]),
+ withOptional(optional): { configMap+: { optional: optional } },
+ },
+ '#downwardAPI':: d.obj(help='"downwardAPI information about the downwardAPI data to project"'),
+ downwardAPI: {
+ '#items':: d.obj(help='"Items is a list of DownwardAPIVolume file"'),
+ items: {
+ '#fieldRef':: d.obj(help='"Required: Selects a field of the pod: only annotations, labels, name and namespace are supported."'),
+ fieldRef: {
+ '#withApiVersion':: d.fn(help='"Version of the schema the FieldPath is written in terms of, defaults to \\"v1\\"."', args=[d.arg(name='apiVersion', type=d.T.string)]),
+ withApiVersion(apiVersion): { fieldRef+: { apiVersion: apiVersion } },
+ '#withFieldPath':: d.fn(help='"Path of the field to select in the specified API version."', args=[d.arg(name='fieldPath', type=d.T.string)]),
+ withFieldPath(fieldPath): { fieldRef+: { fieldPath: fieldPath } },
+ },
+ '#resourceFieldRef':: d.obj(help='"Selects a resource of the container: only resources limits and requests (limits.cpu, limits.memory, requests.cpu and requests.memory) are currently supported."'),
+ resourceFieldRef: {
+ '#withContainerName':: d.fn(help='"Container name: required for volumes, optional for env vars"', args=[d.arg(name='containerName', type=d.T.string)]),
+ withContainerName(containerName): { resourceFieldRef+: { containerName: containerName } },
+ '#withDivisor':: d.fn(help='"Specifies the output format of the exposed resources, defaults to \\"1\\', args=[d.arg(name='divisor', type=d.T.any)]),
+ withDivisor(divisor): { resourceFieldRef+: { divisor: divisor } },
+ '#withResource':: d.fn(help='"Required: resource to select"', args=[d.arg(name='resource', type=d.T.string)]),
+ withResource(resource): { resourceFieldRef+: { resource: resource } },
+ },
+ '#withMode':: d.fn(help='"Optional: mode bits used to set permissions on this file, must be an octal value between 0000 and 0777 or a decimal value between 0 and 511. YAML accepts both octal and decimal values, JSON requires decimal values for mode bits. If not specified, the volume defaultMode will be used. This might be in conflict with other options that affect the file mode, like fsGroup, and the result can be other mode bits set."', args=[d.arg(name='mode', type=d.T.integer)]),
+ withMode(mode): { mode: mode },
+ '#withPath':: d.fn(help="\"Required: Path is the relative path name of the file to be created. Must not be absolute or contain the '..' path. Must be utf-8 encoded. The first item of the relative path must not start with '..'\"", args=[d.arg(name='path', type=d.T.string)]),
+ withPath(path): { path: path },
+ },
+ '#withItems':: d.fn(help='"Items is a list of DownwardAPIVolume file"', args=[d.arg(name='items', type=d.T.array)]),
+ withItems(items): { downwardAPI+: { items: if std.isArray(v=items) then items else [items] } },
+ '#withItemsMixin':: d.fn(help='"Items is a list of DownwardAPIVolume file"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='items', type=d.T.array)]),
+ withItemsMixin(items): { downwardAPI+: { items+: if std.isArray(v=items) then items else [items] } },
+ },
+ '#secret':: d.obj(help='"secret information about the secret data to project"'),
+ secret: {
+ '#items':: d.obj(help="\"items if unspecified, each key-value pair in the Data field of the referenced Secret will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the Secret, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'.\""),
+ items: {
+ '#withKey':: d.fn(help='"key is the key to project."', args=[d.arg(name='key', type=d.T.string)]),
+ withKey(key): { key: key },
+ '#withMode':: d.fn(help='"mode is Optional: mode bits used to set permissions on this file. Must be an octal value between 0000 and 0777 or a decimal value between 0 and 511. YAML accepts both octal and decimal values, JSON requires decimal values for mode bits. If not specified, the volume defaultMode will be used. This might be in conflict with other options that affect the file mode, like fsGroup, and the result can be other mode bits set."', args=[d.arg(name='mode', type=d.T.integer)]),
+ withMode(mode): { mode: mode },
+ '#withPath':: d.fn(help="\"path is the relative path of the file to map the key to. May not be an absolute path. May not contain the path element '..'. May not start with the string '..'.\"", args=[d.arg(name='path', type=d.T.string)]),
+ withPath(path): { path: path },
+ },
+ '#withItems':: d.fn(help="\"items if unspecified, each key-value pair in the Data field of the referenced Secret will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the Secret, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'.\"", args=[d.arg(name='items', type=d.T.array)]),
+ withItems(items): { secret+: { items: if std.isArray(v=items) then items else [items] } },
+ '#withItemsMixin':: d.fn(help="\"items if unspecified, each key-value pair in the Data field of the referenced Secret will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the Secret, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='items', type=d.T.array)]),
+ withItemsMixin(items): { secret+: { items+: if std.isArray(v=items) then items else [items] } },
+ '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { secret+: { name: name } },
+ '#withOptional':: d.fn(help='"optional field specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]),
+ withOptional(optional): { secret+: { optional: optional } },
+ },
+ '#serviceAccountToken':: d.obj(help='"serviceAccountToken is information about the serviceAccountToken data to project"'),
+ serviceAccountToken: {
+ '#withAudience':: d.fn(help='"audience is the intended audience of the token. A recipient of a token must identify itself with an identifier specified in the audience of the token, and otherwise should reject the token. The audience defaults to the identifier of the apiserver."', args=[d.arg(name='audience', type=d.T.string)]),
+ withAudience(audience): { serviceAccountToken+: { audience: audience } },
+ '#withExpirationSeconds':: d.fn(help='"expirationSeconds is the requested duration of validity of the service account token. As the token approaches expiration, the kubelet volume plugin will proactively rotate the service account token. The kubelet will start trying to rotate the token if the token is older than 80 percent of its time to live or if the token is older than 24 hours.Defaults to 1 hour and must be at least 10 minutes."', args=[d.arg(name='expirationSeconds', type=d.T.integer)]),
+ withExpirationSeconds(expirationSeconds): { serviceAccountToken+: { expirationSeconds: expirationSeconds } },
+ '#withPath':: d.fn(help='"path is the path relative to the mount point of the file to project the token into."', args=[d.arg(name='path', type=d.T.string)]),
+ withPath(path): { serviceAccountToken+: { path: path } },
+ },
+ },
+ '#withDefaultMode':: d.fn(help='"defaultMode are the mode bits used to set permissions on created files by default. Must be an octal value between 0000 and 0777 or a decimal value between 0 and 511. YAML accepts both octal and decimal values, JSON requires decimal values for mode bits. Directories within the path are not affected by this setting. This might be in conflict with other options that affect the file mode, like fsGroup, and the result can be other mode bits set."', args=[d.arg(name='defaultMode', type=d.T.integer)]),
+ withDefaultMode(defaultMode): { spec+: { projectedVolumeTemplate+: { defaultMode: defaultMode } } },
+ '#withSources':: d.fn(help='"sources is the list of volume projections"', args=[d.arg(name='sources', type=d.T.array)]),
+ withSources(sources): { spec+: { projectedVolumeTemplate+: { sources: if std.isArray(v=sources) then sources else [sources] } } },
+ '#withSourcesMixin':: d.fn(help='"sources is the list of volume projections"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='sources', type=d.T.array)]),
+ withSourcesMixin(sources): { spec+: { projectedVolumeTemplate+: { sources+: if std.isArray(v=sources) then sources else [sources] } } },
+ },
+ '#replica':: d.obj(help='"Replica cluster configuration"'),
+ replica: {
+ '#withEnabled':: d.fn(help='"If replica mode is enabled, this cluster will be a replica of an existing cluster. Replica cluster can be created from a recovery object store or via streaming through pg_basebackup. Refer to the Replica clusters page of the documentation for more information."', args=[d.arg(name='enabled', type=d.T.boolean)]),
+ withEnabled(enabled): { spec+: { replica+: { enabled: enabled } } },
+ '#withSource':: d.fn(help='"The name of the external cluster which is the replication origin"', args=[d.arg(name='source', type=d.T.string)]),
+ withSource(source): { spec+: { replica+: { source: source } } },
+ },
+ '#replicationSlots':: d.obj(help='"Replication slots management configuration"'),
+ replicationSlots: {
+ '#highAvailability':: d.obj(help='"Replication slots for high availability configuration"'),
+ highAvailability: {
+ '#withEnabled':: d.fn(help='"If enabled (default), the operator will automatically manage replication slots on the primary instance and use them in streaming replication connections with all the standby instances that are part of the HA cluster. If disabled, the operator will not take advantage of replication slots in streaming connections with the replicas. This feature also controls replication slots in replica cluster, from the designated primary to its cascading replicas."', args=[d.arg(name='enabled', type=d.T.boolean)]),
+ withEnabled(enabled): { spec+: { replicationSlots+: { highAvailability+: { enabled: enabled } } } },
+ '#withSlotPrefix':: d.fn(help='"Prefix for replication slots managed by the operator for HA. It may only contain lower case letters, numbers, and the underscore character. This can only be set at creation time. By default set to `_cnpg_`."', args=[d.arg(name='slotPrefix', type=d.T.string)]),
+ withSlotPrefix(slotPrefix): { spec+: { replicationSlots+: { highAvailability+: { slotPrefix: slotPrefix } } } },
+ },
+ '#withUpdateInterval':: d.fn(help='"Standby will update the status of the local replication slots every `updateInterval` seconds (default 30)."', args=[d.arg(name='updateInterval', type=d.T.integer)]),
+ withUpdateInterval(updateInterval): { spec+: { replicationSlots+: { updateInterval: updateInterval } } },
+ },
+ '#resources':: d.obj(help='"Resources requirements of every generated Pod. Please refer to https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/ for more information."'),
+ resources: {
+ '#claims':: d.obj(help='"Claims lists the names of resources, defined in spec.resourceClaims, that are used by this container. \\n This is an alpha field and requires enabling the DynamicResourceAllocation feature gate. \\n This field is immutable. It can only be set for containers."'),
+ claims: {
+ '#withName':: d.fn(help='"Name must match the name of one entry in pod.spec.resourceClaims of the Pod where this field is used. It makes that resource available inside a container."', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { name: name },
+ },
+ '#withClaims':: d.fn(help='"Claims lists the names of resources, defined in spec.resourceClaims, that are used by this container. \\n This is an alpha field and requires enabling the DynamicResourceAllocation feature gate. \\n This field is immutable. It can only be set for containers."', args=[d.arg(name='claims', type=d.T.array)]),
+ withClaims(claims): { spec+: { resources+: { claims: if std.isArray(v=claims) then claims else [claims] } } },
+ '#withClaimsMixin':: d.fn(help='"Claims lists the names of resources, defined in spec.resourceClaims, that are used by this container. \\n This is an alpha field and requires enabling the DynamicResourceAllocation feature gate. \\n This field is immutable. It can only be set for containers."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='claims', type=d.T.array)]),
+ withClaimsMixin(claims): { spec+: { resources+: { claims+: if std.isArray(v=claims) then claims else [claims] } } },
+ '#withLimits':: d.fn(help='"Limits describes the maximum amount of compute resources allowed. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"', args=[d.arg(name='limits', type=d.T.object)]),
+ withLimits(limits): { spec+: { resources+: { limits: limits } } },
+ '#withLimitsMixin':: d.fn(help='"Limits describes the maximum amount of compute resources allowed. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='limits', type=d.T.object)]),
+ withLimitsMixin(limits): { spec+: { resources+: { limits+: limits } } },
+ '#withRequests':: d.fn(help='"Requests describes the minimum amount of compute resources required. If Requests is omitted for a container, it defaults to Limits if that is explicitly specified, otherwise to an implementation-defined value. Requests cannot exceed Limits. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"', args=[d.arg(name='requests', type=d.T.object)]),
+ withRequests(requests): { spec+: { resources+: { requests: requests } } },
+ '#withRequestsMixin':: d.fn(help='"Requests describes the minimum amount of compute resources required. If Requests is omitted for a container, it defaults to Limits if that is explicitly specified, otherwise to an implementation-defined value. Requests cannot exceed Limits. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='requests', type=d.T.object)]),
+ withRequestsMixin(requests): { spec+: { resources+: { requests+: requests } } },
+ },
+ '#seccompProfile':: d.obj(help='"The SeccompProfile applied to every Pod and Container. Defaults to: `RuntimeDefault`"'),
+ seccompProfile: {
+ '#withLocalhostProfile':: d.fn(help="\"localhostProfile indicates a profile defined in a file on the node should be used. The profile must be preconfigured on the node to work. Must be a descending path, relative to the kubelet's configured seccomp profile location. Must be set if type is \\\"Localhost\\\". Must NOT be set for any other type.\"", args=[d.arg(name='localhostProfile', type=d.T.string)]),
+ withLocalhostProfile(localhostProfile): { spec+: { seccompProfile+: { localhostProfile: localhostProfile } } },
+ '#withType':: d.fn(help='"type indicates which kind of seccomp profile will be applied. Valid options are: \\n Localhost - a profile defined in a file on the node should be used. RuntimeDefault - the container runtime default profile should be used. Unconfined - no profile should be applied."', args=[d.arg(name='type', type=d.T.string)]),
+ withType(type): { spec+: { seccompProfile+: { type: type } } },
+ },
+ '#serviceAccountTemplate':: d.obj(help='"Configure the generation of the service account"'),
+ serviceAccountTemplate: {
+ '#metadata':: d.obj(help='"Metadata are the metadata to be used for the generated service account"'),
+ metadata: {
+ '#withAnnotations':: d.fn(help='"Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations"', args=[d.arg(name='annotations', type=d.T.object)]),
+ withAnnotations(annotations): { spec+: { serviceAccountTemplate+: { metadata+: { annotations: annotations } } } },
+ '#withAnnotationsMixin':: d.fn(help='"Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]),
+ withAnnotationsMixin(annotations): { spec+: { serviceAccountTemplate+: { metadata+: { annotations+: annotations } } } },
+ '#withLabels':: d.fn(help='"Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels"', args=[d.arg(name='labels', type=d.T.object)]),
+ withLabels(labels): { spec+: { serviceAccountTemplate+: { metadata+: { labels: labels } } } },
+ '#withLabelsMixin':: d.fn(help='"Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]),
+ withLabelsMixin(labels): { spec+: { serviceAccountTemplate+: { metadata+: { labels+: labels } } } },
+ },
+ },
+ '#storage':: d.obj(help='"Configuration of the storage of the instances"'),
+ storage: {
+ '#pvcTemplate':: d.obj(help='"Template to be used to generate the Persistent Volume Claim"'),
+ pvcTemplate: {
+ '#dataSource':: d.obj(help='"dataSource field can be used to specify either: * An existing VolumeSnapshot object (snapshot.storage.k8s.io/VolumeSnapshot) * An existing PVC (PersistentVolumeClaim) If the provisioner or an external controller can support the specified data source, it will create a new volume based on the contents of the specified data source. When the AnyVolumeDataSource feature gate is enabled, dataSource contents will be copied to dataSourceRef, and dataSourceRef contents will be copied to dataSource when dataSourceRef.namespace is not specified. If the namespace is specified, then dataSourceRef will not be copied to dataSource."'),
+ dataSource: {
+ '#withApiGroup':: d.fn(help='"APIGroup is the group for the resource being referenced. If APIGroup is not specified, the specified Kind must be in the core API group. For any other third-party types, APIGroup is required."', args=[d.arg(name='apiGroup', type=d.T.string)]),
+ withApiGroup(apiGroup): { spec+: { storage+: { pvcTemplate+: { dataSource+: { apiGroup: apiGroup } } } } },
+ '#withKind':: d.fn(help='"Kind is the type of resource being referenced"', args=[d.arg(name='kind', type=d.T.string)]),
+ withKind(kind): { spec+: { storage+: { pvcTemplate+: { dataSource+: { kind: kind } } } } },
+ '#withName':: d.fn(help='"Name is the name of resource being referenced"', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { spec+: { storage+: { pvcTemplate+: { dataSource+: { name: name } } } } },
+ },
+ '#dataSourceRef':: d.obj(help="\"dataSourceRef specifies the object from which to populate the volume with data, if a non-empty volume is desired. This may be any object from a non-empty API group (non core object) or a PersistentVolumeClaim object. When this field is specified, volume binding will only succeed if the type of the specified object matches some installed volume populator or dynamic provisioner. This field will replace the functionality of the dataSource field and as such if both fields are non-empty, they must have the same value. For backwards compatibility, when namespace isn't specified in dataSourceRef, both fields (dataSource and dataSourceRef) will be set to the same value automatically if one of them is empty and the other is non-empty. When namespace is specified in dataSourceRef, dataSource isn't set to the same value and must be empty. There are three important differences between dataSource and dataSourceRef: * While dataSource only allows two specific types of objects, dataSourceRef allows any non-core object, as well as PersistentVolumeClaim objects. * While dataSource ignores disallowed values (dropping them), dataSourceRef preserves all values, and generates an error if a disallowed value is specified. * While dataSource only allows local objects, dataSourceRef allows objects in any namespaces. (Beta) Using this field requires the AnyVolumeDataSource feature gate to be enabled. (Alpha) Using the namespace field of dataSourceRef requires the CrossNamespaceVolumeDataSource feature gate to be enabled.\""),
+ dataSourceRef: {
+ '#withApiGroup':: d.fn(help='"APIGroup is the group for the resource being referenced. If APIGroup is not specified, the specified Kind must be in the core API group. For any other third-party types, APIGroup is required."', args=[d.arg(name='apiGroup', type=d.T.string)]),
+ withApiGroup(apiGroup): { spec+: { storage+: { pvcTemplate+: { dataSourceRef+: { apiGroup: apiGroup } } } } },
+ '#withKind':: d.fn(help='"Kind is the type of resource being referenced"', args=[d.arg(name='kind', type=d.T.string)]),
+ withKind(kind): { spec+: { storage+: { pvcTemplate+: { dataSourceRef+: { kind: kind } } } } },
+ '#withName':: d.fn(help='"Name is the name of resource being referenced"', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { spec+: { storage+: { pvcTemplate+: { dataSourceRef+: { name: name } } } } },
+ '#withNamespace':: d.fn(help="\"Namespace is the namespace of resource being referenced Note that when a namespace is specified, a gateway.networking.k8s.io/ReferenceGrant object is required in the referent namespace to allow that namespace's owner to accept the reference. See the ReferenceGrant documentation for details. (Alpha) This field requires the CrossNamespaceVolumeDataSource feature gate to be enabled.\"", args=[d.arg(name='namespace', type=d.T.string)]),
+ withNamespace(namespace): { spec+: { storage+: { pvcTemplate+: { dataSourceRef+: { namespace: namespace } } } } },
+ },
+ '#resources':: d.obj(help='"resources represents the minimum resources the volume should have. If RecoverVolumeExpansionFailure feature is enabled users are allowed to specify resource requirements that are lower than previous value but must still be higher than capacity recorded in the status field of the claim. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#resources"'),
+ resources: {
+ '#claims':: d.obj(help='"Claims lists the names of resources, defined in spec.resourceClaims, that are used by this container. \\n This is an alpha field and requires enabling the DynamicResourceAllocation feature gate. \\n This field is immutable. It can only be set for containers."'),
+ claims: {
+ '#withName':: d.fn(help='"Name must match the name of one entry in pod.spec.resourceClaims of the Pod where this field is used. It makes that resource available inside a container."', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { name: name },
+ },
+ '#withClaims':: d.fn(help='"Claims lists the names of resources, defined in spec.resourceClaims, that are used by this container. \\n This is an alpha field and requires enabling the DynamicResourceAllocation feature gate. \\n This field is immutable. It can only be set for containers."', args=[d.arg(name='claims', type=d.T.array)]),
+ withClaims(claims): { spec+: { storage+: { pvcTemplate+: { resources+: { claims: if std.isArray(v=claims) then claims else [claims] } } } } },
+ '#withClaimsMixin':: d.fn(help='"Claims lists the names of resources, defined in spec.resourceClaims, that are used by this container. \\n This is an alpha field and requires enabling the DynamicResourceAllocation feature gate. \\n This field is immutable. It can only be set for containers."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='claims', type=d.T.array)]),
+ withClaimsMixin(claims): { spec+: { storage+: { pvcTemplate+: { resources+: { claims+: if std.isArray(v=claims) then claims else [claims] } } } } },
+ '#withLimits':: d.fn(help='"Limits describes the maximum amount of compute resources allowed. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"', args=[d.arg(name='limits', type=d.T.object)]),
+ withLimits(limits): { spec+: { storage+: { pvcTemplate+: { resources+: { limits: limits } } } } },
+ '#withLimitsMixin':: d.fn(help='"Limits describes the maximum amount of compute resources allowed. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='limits', type=d.T.object)]),
+ withLimitsMixin(limits): { spec+: { storage+: { pvcTemplate+: { resources+: { limits+: limits } } } } },
+ '#withRequests':: d.fn(help='"Requests describes the minimum amount of compute resources required. If Requests is omitted for a container, it defaults to Limits if that is explicitly specified, otherwise to an implementation-defined value. Requests cannot exceed Limits. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"', args=[d.arg(name='requests', type=d.T.object)]),
+ withRequests(requests): { spec+: { storage+: { pvcTemplate+: { resources+: { requests: requests } } } } },
+ '#withRequestsMixin':: d.fn(help='"Requests describes the minimum amount of compute resources required. If Requests is omitted for a container, it defaults to Limits if that is explicitly specified, otherwise to an implementation-defined value. Requests cannot exceed Limits. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='requests', type=d.T.object)]),
+ withRequestsMixin(requests): { spec+: { storage+: { pvcTemplate+: { resources+: { requests+: requests } } } } },
+ },
+ '#selector':: d.obj(help='"selector is a label query over volumes to consider for binding."'),
+ selector: {
+ '#matchExpressions':: d.obj(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."'),
+ matchExpressions: {
+ '#withKey':: d.fn(help='"key is the label key that the selector applies to."', args=[d.arg(name='key', type=d.T.string)]),
+ withKey(key): { key: key },
+ '#withOperator':: d.fn(help="\"operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist.\"", args=[d.arg(name='operator', type=d.T.string)]),
+ withOperator(operator): { operator: operator },
+ '#withValues':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."', args=[d.arg(name='values', type=d.T.array)]),
+ withValues(values): { values: if std.isArray(v=values) then values else [values] },
+ '#withValuesMixin':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='values', type=d.T.array)]),
+ withValuesMixin(values): { values+: if std.isArray(v=values) then values else [values] },
+ },
+ '#withMatchExpressions':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."', args=[d.arg(name='matchExpressions', type=d.T.array)]),
+ withMatchExpressions(matchExpressions): { spec+: { storage+: { pvcTemplate+: { selector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } } } },
+ '#withMatchExpressionsMixin':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]),
+ withMatchExpressionsMixin(matchExpressions): { spec+: { storage+: { pvcTemplate+: { selector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } } } },
+ '#withMatchLabels':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."', args=[d.arg(name='matchLabels', type=d.T.object)]),
+ withMatchLabels(matchLabels): { spec+: { storage+: { pvcTemplate+: { selector+: { matchLabels: matchLabels } } } } },
+ '#withMatchLabelsMixin':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]),
+ withMatchLabelsMixin(matchLabels): { spec+: { storage+: { pvcTemplate+: { selector+: { matchLabels+: matchLabels } } } } },
+ },
+ '#withAccessModes':: d.fn(help='"accessModes contains the desired access modes the volume should have. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#access-modes-1"', args=[d.arg(name='accessModes', type=d.T.array)]),
+ withAccessModes(accessModes): { spec+: { storage+: { pvcTemplate+: { accessModes: if std.isArray(v=accessModes) then accessModes else [accessModes] } } } },
+ '#withAccessModesMixin':: d.fn(help='"accessModes contains the desired access modes the volume should have. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#access-modes-1"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='accessModes', type=d.T.array)]),
+ withAccessModesMixin(accessModes): { spec+: { storage+: { pvcTemplate+: { accessModes+: if std.isArray(v=accessModes) then accessModes else [accessModes] } } } },
+ '#withStorageClassName':: d.fn(help='"storageClassName is the name of the StorageClass required by the claim. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#class-1"', args=[d.arg(name='storageClassName', type=d.T.string)]),
+ withStorageClassName(storageClassName): { spec+: { storage+: { pvcTemplate+: { storageClassName: storageClassName } } } },
+ '#withVolumeMode':: d.fn(help='"volumeMode defines what type of volume is required by the claim. Value of Filesystem is implied when not included in claim spec."', args=[d.arg(name='volumeMode', type=d.T.string)]),
+ withVolumeMode(volumeMode): { spec+: { storage+: { pvcTemplate+: { volumeMode: volumeMode } } } },
+ '#withVolumeName':: d.fn(help='"volumeName is the binding reference to the PersistentVolume backing this claim."', args=[d.arg(name='volumeName', type=d.T.string)]),
+ withVolumeName(volumeName): { spec+: { storage+: { pvcTemplate+: { volumeName: volumeName } } } },
+ },
+ '#withResizeInUseVolumes':: d.fn(help='"Resize existent PVCs, defaults to true"', args=[d.arg(name='resizeInUseVolumes', type=d.T.boolean)]),
+ withResizeInUseVolumes(resizeInUseVolumes): { spec+: { storage+: { resizeInUseVolumes: resizeInUseVolumes } } },
+ '#withSize':: d.fn(help='"Size of the storage. Required if not already specified in the PVC template. Changes to this field are automatically reapplied to the created PVCs. Size cannot be decreased."', args=[d.arg(name='size', type=d.T.string)]),
+ withSize(size): { spec+: { storage+: { size: size } } },
+ '#withStorageClass':: d.fn(help='"StorageClass to use for database data (`PGDATA`). Applied after evaluating the PVC template, if available. If not specified, generated PVCs will be satisfied by the default storage class"', args=[d.arg(name='storageClass', type=d.T.string)]),
+ withStorageClass(storageClass): { spec+: { storage+: { storageClass: storageClass } } },
+ },
+ '#superuserSecret':: d.obj(help='"The secret containing the superuser password. If not defined a new secret will be created with a randomly generated password"'),
+ superuserSecret: {
+ '#withName':: d.fn(help='"Name of the referent."', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { spec+: { superuserSecret+: { name: name } } },
+ },
+ '#topologySpreadConstraints':: d.obj(help='"TopologySpreadConstraints specifies how to spread matching pods among the given topology. More info: https://kubernetes.io/docs/concepts/scheduling-eviction/topology-spread-constraints/"'),
+ topologySpreadConstraints: {
+ '#labelSelector':: d.obj(help='"LabelSelector is used to find matching pods. Pods that match this label selector are counted to determine the number of pods in their corresponding topology domain."'),
+ labelSelector: {
+ '#matchExpressions':: d.obj(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."'),
+ matchExpressions: {
+ '#withKey':: d.fn(help='"key is the label key that the selector applies to."', args=[d.arg(name='key', type=d.T.string)]),
+ withKey(key): { key: key },
+ '#withOperator':: d.fn(help="\"operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist.\"", args=[d.arg(name='operator', type=d.T.string)]),
+ withOperator(operator): { operator: operator },
+ '#withValues':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."', args=[d.arg(name='values', type=d.T.array)]),
+ withValues(values): { values: if std.isArray(v=values) then values else [values] },
+ '#withValuesMixin':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='values', type=d.T.array)]),
+ withValuesMixin(values): { values+: if std.isArray(v=values) then values else [values] },
+ },
+ '#withMatchExpressions':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."', args=[d.arg(name='matchExpressions', type=d.T.array)]),
+ withMatchExpressions(matchExpressions): { labelSelector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } },
+ '#withMatchExpressionsMixin':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]),
+ withMatchExpressionsMixin(matchExpressions): { labelSelector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } },
+ '#withMatchLabels':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."', args=[d.arg(name='matchLabels', type=d.T.object)]),
+ withMatchLabels(matchLabels): { labelSelector+: { matchLabels: matchLabels } },
+ '#withMatchLabelsMixin':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]),
+ withMatchLabelsMixin(matchLabels): { labelSelector+: { matchLabels+: matchLabels } },
+ },
+ '#withMatchLabelKeys':: d.fn(help="\"MatchLabelKeys is a set of pod label keys to select the pods over which spreading will be calculated. The keys are used to lookup values from the incoming pod labels, those key-value labels are ANDed with labelSelector to select the group of existing pods over which spreading will be calculated for the incoming pod. The same key is forbidden to exist in both MatchLabelKeys and LabelSelector. MatchLabelKeys cannot be set when LabelSelector isn't set. Keys that don't exist in the incoming pod labels will be ignored. A null or empty list means only match against labelSelector. \\n This is a beta field and requires the MatchLabelKeysInPodTopologySpread feature gate to be enabled (enabled by default).\"", args=[d.arg(name='matchLabelKeys', type=d.T.array)]),
+ withMatchLabelKeys(matchLabelKeys): { matchLabelKeys: if std.isArray(v=matchLabelKeys) then matchLabelKeys else [matchLabelKeys] },
+ '#withMatchLabelKeysMixin':: d.fn(help="\"MatchLabelKeys is a set of pod label keys to select the pods over which spreading will be calculated. The keys are used to lookup values from the incoming pod labels, those key-value labels are ANDed with labelSelector to select the group of existing pods over which spreading will be calculated for the incoming pod. The same key is forbidden to exist in both MatchLabelKeys and LabelSelector. MatchLabelKeys cannot be set when LabelSelector isn't set. Keys that don't exist in the incoming pod labels will be ignored. A null or empty list means only match against labelSelector. \\n This is a beta field and requires the MatchLabelKeysInPodTopologySpread feature gate to be enabled (enabled by default).\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='matchLabelKeys', type=d.T.array)]),
+ withMatchLabelKeysMixin(matchLabelKeys): { matchLabelKeys+: if std.isArray(v=matchLabelKeys) then matchLabelKeys else [matchLabelKeys] },
+ '#withMaxSkew':: d.fn(help="\"MaxSkew describes the degree to which pods may be unevenly distributed. When `whenUnsatisfiable=DoNotSchedule`, it is the maximum permitted difference between the number of matching pods in the target topology and the global minimum. The global minimum is the minimum number of matching pods in an eligible domain or zero if the number of eligible domains is less than MinDomains. For example, in a 3-zone cluster, MaxSkew is set to 1, and pods with the same labelSelector spread as 2/2/1: In this case, the global minimum is 1. | zone1 | zone2 | zone3 | | P P | P P | P | - if MaxSkew is 1, incoming pod can only be scheduled to zone3 to become 2/2/2; scheduling it onto zone1(zone2) would make the ActualSkew(3-1) on zone1(zone2) violate MaxSkew(1). - if MaxSkew is 2, incoming pod can be scheduled onto any zone. When `whenUnsatisfiable=ScheduleAnyway`, it is used to give higher precedence to topologies that satisfy it. It's a required field. Default value is 1 and 0 is not allowed.\"", args=[d.arg(name='maxSkew', type=d.T.integer)]),
+ withMaxSkew(maxSkew): { maxSkew: maxSkew },
+ '#withMinDomains':: d.fn(help="\"MinDomains indicates a minimum number of eligible domains. When the number of eligible domains with matching topology keys is less than minDomains, Pod Topology Spread treats \\\"global minimum\\\" as 0, and then the calculation of Skew is performed. And when the number of eligible domains with matching topology keys equals or greater than minDomains, this value has no effect on scheduling. As a result, when the number of eligible domains is less than minDomains, scheduler won't schedule more than maxSkew Pods to those domains. If value is nil, the constraint behaves as if MinDomains is equal to 1. Valid values are integers greater than 0. When value is not nil, WhenUnsatisfiable must be DoNotSchedule. \\n For example, in a 3-zone cluster, MaxSkew is set to 2, MinDomains is set to 5 and pods with the same labelSelector spread as 2/2/2: | zone1 | zone2 | zone3 | | P P | P P | P P | The number of domains is less than 5(MinDomains), so \\\"global minimum\\\" is treated as 0. In this situation, new pod with the same labelSelector cannot be scheduled, because computed skew will be 3(3 - 0) if new Pod is scheduled to any of the three zones, it will violate MaxSkew. \\n This is a beta field and requires the MinDomainsInPodTopologySpread feature gate to be enabled (enabled by default).\"", args=[d.arg(name='minDomains', type=d.T.integer)]),
+ withMinDomains(minDomains): { minDomains: minDomains },
+ '#withNodeAffinityPolicy':: d.fn(help="\"NodeAffinityPolicy indicates how we will treat Pod's nodeAffinity/nodeSelector when calculating pod topology spread skew. Options are: - Honor: only nodes matching nodeAffinity/nodeSelector are included in the calculations. - Ignore: nodeAffinity/nodeSelector are ignored. All nodes are included in the calculations. \\n If this value is nil, the behavior is equivalent to the Honor policy. This is a beta-level feature default enabled by the NodeInclusionPolicyInPodTopologySpread feature flag.\"", args=[d.arg(name='nodeAffinityPolicy', type=d.T.string)]),
+ withNodeAffinityPolicy(nodeAffinityPolicy): { nodeAffinityPolicy: nodeAffinityPolicy },
+ '#withNodeTaintsPolicy':: d.fn(help='"NodeTaintsPolicy indicates how we will treat node taints when calculating pod topology spread skew. Options are: - Honor: nodes without taints, along with tainted nodes for which the incoming pod has a toleration, are included. - Ignore: node taints are ignored. All nodes are included. \\n If this value is nil, the behavior is equivalent to the Ignore policy. This is a beta-level feature default enabled by the NodeInclusionPolicyInPodTopologySpread feature flag."', args=[d.arg(name='nodeTaintsPolicy', type=d.T.string)]),
+ withNodeTaintsPolicy(nodeTaintsPolicy): { nodeTaintsPolicy: nodeTaintsPolicy },
+ '#withTopologyKey':: d.fn(help="\"TopologyKey is the key of node labels. Nodes that have a label with this key and identical values are considered to be in the same topology. We consider each \u003ckey, value\u003e as a \\\"bucket\\\", and try to put balanced number of pods into each bucket. We define a domain as a particular instance of a topology. Also, we define an eligible domain as a domain whose nodes meet the requirements of nodeAffinityPolicy and nodeTaintsPolicy. e.g. If TopologyKey is \\\"kubernetes.io/hostname\\\", each Node is a domain of that topology. And, if TopologyKey is \\\"topology.kubernetes.io/zone\\\", each zone is a domain of that topology. It's a required field.\"", args=[d.arg(name='topologyKey', type=d.T.string)]),
+ withTopologyKey(topologyKey): { topologyKey: topologyKey },
+ '#withWhenUnsatisfiable':: d.fn(help="\"WhenUnsatisfiable indicates how to deal with a pod if it doesn't satisfy the spread constraint. - DoNotSchedule (default) tells the scheduler not to schedule it. - ScheduleAnyway tells the scheduler to schedule the pod in any location, but giving higher precedence to topologies that would help reduce the skew. A constraint is considered \\\"Unsatisfiable\\\" for an incoming pod if and only if every possible node assignment for that pod would violate \\\"MaxSkew\\\" on some topology. For example, in a 3-zone cluster, MaxSkew is set to 1, and pods with the same labelSelector spread as 3/1/1: | zone1 | zone2 | zone3 | | P P P | P | P | If WhenUnsatisfiable is set to DoNotSchedule, incoming pod can only be scheduled to zone2(zone3) to become 3/2/1(3/1/2) as ActualSkew(2-1) on zone2(zone3) satisfies MaxSkew(1). In other words, the cluster can still be imbalanced, but scheduler won't make it *more* imbalanced. It's a required field.\"", args=[d.arg(name='whenUnsatisfiable', type=d.T.string)]),
+ withWhenUnsatisfiable(whenUnsatisfiable): { whenUnsatisfiable: whenUnsatisfiable },
+ },
+ '#walStorage':: d.obj(help='"Configuration of the storage for PostgreSQL WAL (Write-Ahead Log)"'),
+ walStorage: {
+ '#pvcTemplate':: d.obj(help='"Template to be used to generate the Persistent Volume Claim"'),
+ pvcTemplate: {
+ '#dataSource':: d.obj(help='"dataSource field can be used to specify either: * An existing VolumeSnapshot object (snapshot.storage.k8s.io/VolumeSnapshot) * An existing PVC (PersistentVolumeClaim) If the provisioner or an external controller can support the specified data source, it will create a new volume based on the contents of the specified data source. When the AnyVolumeDataSource feature gate is enabled, dataSource contents will be copied to dataSourceRef, and dataSourceRef contents will be copied to dataSource when dataSourceRef.namespace is not specified. If the namespace is specified, then dataSourceRef will not be copied to dataSource."'),
+ dataSource: {
+ '#withApiGroup':: d.fn(help='"APIGroup is the group for the resource being referenced. If APIGroup is not specified, the specified Kind must be in the core API group. For any other third-party types, APIGroup is required."', args=[d.arg(name='apiGroup', type=d.T.string)]),
+ withApiGroup(apiGroup): { spec+: { walStorage+: { pvcTemplate+: { dataSource+: { apiGroup: apiGroup } } } } },
+ '#withKind':: d.fn(help='"Kind is the type of resource being referenced"', args=[d.arg(name='kind', type=d.T.string)]),
+ withKind(kind): { spec+: { walStorage+: { pvcTemplate+: { dataSource+: { kind: kind } } } } },
+ '#withName':: d.fn(help='"Name is the name of resource being referenced"', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { spec+: { walStorage+: { pvcTemplate+: { dataSource+: { name: name } } } } },
+ },
+ '#dataSourceRef':: d.obj(help="\"dataSourceRef specifies the object from which to populate the volume with data, if a non-empty volume is desired. This may be any object from a non-empty API group (non core object) or a PersistentVolumeClaim object. When this field is specified, volume binding will only succeed if the type of the specified object matches some installed volume populator or dynamic provisioner. This field will replace the functionality of the dataSource field and as such if both fields are non-empty, they must have the same value. For backwards compatibility, when namespace isn't specified in dataSourceRef, both fields (dataSource and dataSourceRef) will be set to the same value automatically if one of them is empty and the other is non-empty. When namespace is specified in dataSourceRef, dataSource isn't set to the same value and must be empty. There are three important differences between dataSource and dataSourceRef: * While dataSource only allows two specific types of objects, dataSourceRef allows any non-core object, as well as PersistentVolumeClaim objects. * While dataSource ignores disallowed values (dropping them), dataSourceRef preserves all values, and generates an error if a disallowed value is specified. * While dataSource only allows local objects, dataSourceRef allows objects in any namespaces. (Beta) Using this field requires the AnyVolumeDataSource feature gate to be enabled. (Alpha) Using the namespace field of dataSourceRef requires the CrossNamespaceVolumeDataSource feature gate to be enabled.\""),
+ dataSourceRef: {
+ '#withApiGroup':: d.fn(help='"APIGroup is the group for the resource being referenced. If APIGroup is not specified, the specified Kind must be in the core API group. For any other third-party types, APIGroup is required."', args=[d.arg(name='apiGroup', type=d.T.string)]),
+ withApiGroup(apiGroup): { spec+: { walStorage+: { pvcTemplate+: { dataSourceRef+: { apiGroup: apiGroup } } } } },
+ '#withKind':: d.fn(help='"Kind is the type of resource being referenced"', args=[d.arg(name='kind', type=d.T.string)]),
+ withKind(kind): { spec+: { walStorage+: { pvcTemplate+: { dataSourceRef+: { kind: kind } } } } },
+ '#withName':: d.fn(help='"Name is the name of resource being referenced"', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { spec+: { walStorage+: { pvcTemplate+: { dataSourceRef+: { name: name } } } } },
+ '#withNamespace':: d.fn(help="\"Namespace is the namespace of resource being referenced Note that when a namespace is specified, a gateway.networking.k8s.io/ReferenceGrant object is required in the referent namespace to allow that namespace's owner to accept the reference. See the ReferenceGrant documentation for details. (Alpha) This field requires the CrossNamespaceVolumeDataSource feature gate to be enabled.\"", args=[d.arg(name='namespace', type=d.T.string)]),
+ withNamespace(namespace): { spec+: { walStorage+: { pvcTemplate+: { dataSourceRef+: { namespace: namespace } } } } },
+ },
+ '#resources':: d.obj(help='"resources represents the minimum resources the volume should have. If RecoverVolumeExpansionFailure feature is enabled users are allowed to specify resource requirements that are lower than previous value but must still be higher than capacity recorded in the status field of the claim. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#resources"'),
+ resources: {
+ '#claims':: d.obj(help='"Claims lists the names of resources, defined in spec.resourceClaims, that are used by this container. \\n This is an alpha field and requires enabling the DynamicResourceAllocation feature gate. \\n This field is immutable. It can only be set for containers."'),
+ claims: {
+ '#withName':: d.fn(help='"Name must match the name of one entry in pod.spec.resourceClaims of the Pod where this field is used. It makes that resource available inside a container."', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { name: name },
+ },
+ '#withClaims':: d.fn(help='"Claims lists the names of resources, defined in spec.resourceClaims, that are used by this container. \\n This is an alpha field and requires enabling the DynamicResourceAllocation feature gate. \\n This field is immutable. It can only be set for containers."', args=[d.arg(name='claims', type=d.T.array)]),
+ withClaims(claims): { spec+: { walStorage+: { pvcTemplate+: { resources+: { claims: if std.isArray(v=claims) then claims else [claims] } } } } },
+ '#withClaimsMixin':: d.fn(help='"Claims lists the names of resources, defined in spec.resourceClaims, that are used by this container. \\n This is an alpha field and requires enabling the DynamicResourceAllocation feature gate. \\n This field is immutable. It can only be set for containers."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='claims', type=d.T.array)]),
+ withClaimsMixin(claims): { spec+: { walStorage+: { pvcTemplate+: { resources+: { claims+: if std.isArray(v=claims) then claims else [claims] } } } } },
+ '#withLimits':: d.fn(help='"Limits describes the maximum amount of compute resources allowed. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"', args=[d.arg(name='limits', type=d.T.object)]),
+ withLimits(limits): { spec+: { walStorage+: { pvcTemplate+: { resources+: { limits: limits } } } } },
+ '#withLimitsMixin':: d.fn(help='"Limits describes the maximum amount of compute resources allowed. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='limits', type=d.T.object)]),
+ withLimitsMixin(limits): { spec+: { walStorage+: { pvcTemplate+: { resources+: { limits+: limits } } } } },
+ '#withRequests':: d.fn(help='"Requests describes the minimum amount of compute resources required. If Requests is omitted for a container, it defaults to Limits if that is explicitly specified, otherwise to an implementation-defined value. Requests cannot exceed Limits. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"', args=[d.arg(name='requests', type=d.T.object)]),
+ withRequests(requests): { spec+: { walStorage+: { pvcTemplate+: { resources+: { requests: requests } } } } },
+ '#withRequestsMixin':: d.fn(help='"Requests describes the minimum amount of compute resources required. If Requests is omitted for a container, it defaults to Limits if that is explicitly specified, otherwise to an implementation-defined value. Requests cannot exceed Limits. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='requests', type=d.T.object)]),
+ withRequestsMixin(requests): { spec+: { walStorage+: { pvcTemplate+: { resources+: { requests+: requests } } } } },
+ },
+ '#selector':: d.obj(help='"selector is a label query over volumes to consider for binding."'),
+ selector: {
+ '#matchExpressions':: d.obj(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."'),
+ matchExpressions: {
+ '#withKey':: d.fn(help='"key is the label key that the selector applies to."', args=[d.arg(name='key', type=d.T.string)]),
+ withKey(key): { key: key },
+ '#withOperator':: d.fn(help="\"operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist.\"", args=[d.arg(name='operator', type=d.T.string)]),
+ withOperator(operator): { operator: operator },
+ '#withValues':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."', args=[d.arg(name='values', type=d.T.array)]),
+ withValues(values): { values: if std.isArray(v=values) then values else [values] },
+ '#withValuesMixin':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='values', type=d.T.array)]),
+ withValuesMixin(values): { values+: if std.isArray(v=values) then values else [values] },
+ },
+ '#withMatchExpressions':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."', args=[d.arg(name='matchExpressions', type=d.T.array)]),
+ withMatchExpressions(matchExpressions): { spec+: { walStorage+: { pvcTemplate+: { selector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } } } },
+ '#withMatchExpressionsMixin':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]),
+ withMatchExpressionsMixin(matchExpressions): { spec+: { walStorage+: { pvcTemplate+: { selector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } } } },
+ '#withMatchLabels':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."', args=[d.arg(name='matchLabels', type=d.T.object)]),
+ withMatchLabels(matchLabels): { spec+: { walStorage+: { pvcTemplate+: { selector+: { matchLabels: matchLabels } } } } },
+ '#withMatchLabelsMixin':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]),
+ withMatchLabelsMixin(matchLabels): { spec+: { walStorage+: { pvcTemplate+: { selector+: { matchLabels+: matchLabels } } } } },
+ },
+ '#withAccessModes':: d.fn(help='"accessModes contains the desired access modes the volume should have. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#access-modes-1"', args=[d.arg(name='accessModes', type=d.T.array)]),
+ withAccessModes(accessModes): { spec+: { walStorage+: { pvcTemplate+: { accessModes: if std.isArray(v=accessModes) then accessModes else [accessModes] } } } },
+ '#withAccessModesMixin':: d.fn(help='"accessModes contains the desired access modes the volume should have. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#access-modes-1"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='accessModes', type=d.T.array)]),
+ withAccessModesMixin(accessModes): { spec+: { walStorage+: { pvcTemplate+: { accessModes+: if std.isArray(v=accessModes) then accessModes else [accessModes] } } } },
+ '#withStorageClassName':: d.fn(help='"storageClassName is the name of the StorageClass required by the claim. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#class-1"', args=[d.arg(name='storageClassName', type=d.T.string)]),
+ withStorageClassName(storageClassName): { spec+: { walStorage+: { pvcTemplate+: { storageClassName: storageClassName } } } },
+ '#withVolumeMode':: d.fn(help='"volumeMode defines what type of volume is required by the claim. Value of Filesystem is implied when not included in claim spec."', args=[d.arg(name='volumeMode', type=d.T.string)]),
+ withVolumeMode(volumeMode): { spec+: { walStorage+: { pvcTemplate+: { volumeMode: volumeMode } } } },
+ '#withVolumeName':: d.fn(help='"volumeName is the binding reference to the PersistentVolume backing this claim."', args=[d.arg(name='volumeName', type=d.T.string)]),
+ withVolumeName(volumeName): { spec+: { walStorage+: { pvcTemplate+: { volumeName: volumeName } } } },
+ },
+ '#withResizeInUseVolumes':: d.fn(help='"Resize existent PVCs, defaults to true"', args=[d.arg(name='resizeInUseVolumes', type=d.T.boolean)]),
+ withResizeInUseVolumes(resizeInUseVolumes): { spec+: { walStorage+: { resizeInUseVolumes: resizeInUseVolumes } } },
+ '#withSize':: d.fn(help='"Size of the storage. Required if not already specified in the PVC template. Changes to this field are automatically reapplied to the created PVCs. Size cannot be decreased."', args=[d.arg(name='size', type=d.T.string)]),
+ withSize(size): { spec+: { walStorage+: { size: size } } },
+ '#withStorageClass':: d.fn(help='"StorageClass to use for database data (`PGDATA`). Applied after evaluating the PVC template, if available. If not specified, generated PVCs will be satisfied by the default storage class"', args=[d.arg(name='storageClass', type=d.T.string)]),
+ withStorageClass(storageClass): { spec+: { walStorage+: { storageClass: storageClass } } },
+ },
+ '#withDescription':: d.fn(help='"Description of this PostgreSQL cluster"', args=[d.arg(name='description', type=d.T.string)]),
+ withDescription(description): { spec+: { description: description } },
+ '#withEnableSuperuserAccess':: d.fn(help='"When this option is enabled, the operator will use the `SuperuserSecret` to update the `postgres` user password (if the secret is not present, the operator will automatically create one). When this option is disabled, the operator will ignore the `SuperuserSecret` content, delete it when automatically created, and then blank the password of the `postgres` user by setting it to `NULL`. Disabled by default."', args=[d.arg(name='enableSuperuserAccess', type=d.T.boolean)]),
+ withEnableSuperuserAccess(enableSuperuserAccess): { spec+: { enableSuperuserAccess: enableSuperuserAccess } },
+ '#withEnv':: d.fn(help='"Env follows the Env format to pass environment variables to the pods created in the cluster"', args=[d.arg(name='env', type=d.T.array)]),
+ withEnv(env): { spec+: { env: if std.isArray(v=env) then env else [env] } },
+ '#withEnvFrom':: d.fn(help='"EnvFrom follows the EnvFrom format to pass environment variables sources to the pods to be used by Env"', args=[d.arg(name='envFrom', type=d.T.array)]),
+ withEnvFrom(envFrom): { spec+: { envFrom: if std.isArray(v=envFrom) then envFrom else [envFrom] } },
+ '#withEnvFromMixin':: d.fn(help='"EnvFrom follows the EnvFrom format to pass environment variables sources to the pods to be used by Env"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='envFrom', type=d.T.array)]),
+ withEnvFromMixin(envFrom): { spec+: { envFrom+: if std.isArray(v=envFrom) then envFrom else [envFrom] } },
+ '#withEnvMixin':: d.fn(help='"Env follows the Env format to pass environment variables to the pods created in the cluster"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='env', type=d.T.array)]),
+ withEnvMixin(env): { spec+: { env+: if std.isArray(v=env) then env else [env] } },
+ '#withExternalClusters':: d.fn(help='"The list of external clusters which are used in the configuration"', args=[d.arg(name='externalClusters', type=d.T.array)]),
+ withExternalClusters(externalClusters): { spec+: { externalClusters: if std.isArray(v=externalClusters) then externalClusters else [externalClusters] } },
+ '#withExternalClustersMixin':: d.fn(help='"The list of external clusters which are used in the configuration"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='externalClusters', type=d.T.array)]),
+ withExternalClustersMixin(externalClusters): { spec+: { externalClusters+: if std.isArray(v=externalClusters) then externalClusters else [externalClusters] } },
+ '#withFailoverDelay':: d.fn(help='"The amount of time (in seconds) to wait before triggering a failover after the primary PostgreSQL instance in the cluster was detected to be unhealthy"', args=[d.arg(name='failoverDelay', type=d.T.integer)]),
+ withFailoverDelay(failoverDelay): { spec+: { failoverDelay: failoverDelay } },
+ '#withImageName':: d.fn(help='"Name of the container image, supporting both tags (`:`) and digests for deterministic and repeatable deployments (`:@sha256:`)"', args=[d.arg(name='imageName', type=d.T.string)]),
+ withImageName(imageName): { spec+: { imageName: imageName } },
+ '#withImagePullPolicy':: d.fn(help='"Image pull policy. One of `Always`, `Never` or `IfNotPresent`. If not defined, it defaults to `IfNotPresent`. Cannot be updated. More info: https://kubernetes.io/docs/concepts/containers/images#updating-images"', args=[d.arg(name='imagePullPolicy', type=d.T.string)]),
+ withImagePullPolicy(imagePullPolicy): { spec+: { imagePullPolicy: imagePullPolicy } },
+ '#withImagePullSecrets':: d.fn(help='"The list of pull secrets to be used to pull the images"', args=[d.arg(name='imagePullSecrets', type=d.T.array)]),
+ withImagePullSecrets(imagePullSecrets): { spec+: { imagePullSecrets: if std.isArray(v=imagePullSecrets) then imagePullSecrets else [imagePullSecrets] } },
+ '#withImagePullSecretsMixin':: d.fn(help='"The list of pull secrets to be used to pull the images"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='imagePullSecrets', type=d.T.array)]),
+ withImagePullSecretsMixin(imagePullSecrets): { spec+: { imagePullSecrets+: if std.isArray(v=imagePullSecrets) then imagePullSecrets else [imagePullSecrets] } },
+ '#withInstances':: d.fn(help='"Number of instances required in the cluster"', args=[d.arg(name='instances', type=d.T.integer)]),
+ withInstances(instances): { spec+: { instances: instances } },
+ '#withLogLevel':: d.fn(help="\"The instances' log level, one of the following values: error, warning, info (default), debug, trace\"", args=[d.arg(name='logLevel', type=d.T.string)]),
+ withLogLevel(logLevel): { spec+: { logLevel: logLevel } },
+ '#withMaxSyncReplicas':: d.fn(help='"The target value for the synchronous replication quorum, that can be decreased if the number of ready standbys is lower than this. Undefined or 0 disable synchronous replication."', args=[d.arg(name='maxSyncReplicas', type=d.T.integer)]),
+ withMaxSyncReplicas(maxSyncReplicas): { spec+: { maxSyncReplicas: maxSyncReplicas } },
+ '#withMinSyncReplicas':: d.fn(help='"Minimum number of instances required in synchronous replication with the primary. Undefined or 0 allow writes to complete when no standby is available."', args=[d.arg(name='minSyncReplicas', type=d.T.integer)]),
+ withMinSyncReplicas(minSyncReplicas): { spec+: { minSyncReplicas: minSyncReplicas } },
+ '#withPostgresGID':: d.fn(help='"The GID of the `postgres` user inside the image, defaults to `26`"', args=[d.arg(name='postgresGID', type=d.T.integer)]),
+ withPostgresGID(postgresGID): { spec+: { postgresGID: postgresGID } },
+ '#withPostgresUID':: d.fn(help='"The UID of the `postgres` user inside the image, defaults to `26`"', args=[d.arg(name='postgresUID', type=d.T.integer)]),
+ withPostgresUID(postgresUID): { spec+: { postgresUID: postgresUID } },
+ '#withPrimaryUpdateMethod':: d.fn(help='"Method to follow to upgrade the primary server during a rolling update procedure, after all replicas have been successfully updated: it can be with a switchover (`switchover`) or in-place (`restart` - default)"', args=[d.arg(name='primaryUpdateMethod', type=d.T.string)]),
+ withPrimaryUpdateMethod(primaryUpdateMethod): { spec+: { primaryUpdateMethod: primaryUpdateMethod } },
+ '#withPrimaryUpdateStrategy':: d.fn(help='"Deployment strategy to follow to upgrade the primary server during a rolling update procedure, after all replicas have been successfully updated: it can be automated (`unsupervised` - default) or manual (`supervised`)"', args=[d.arg(name='primaryUpdateStrategy', type=d.T.string)]),
+ withPrimaryUpdateStrategy(primaryUpdateStrategy): { spec+: { primaryUpdateStrategy: primaryUpdateStrategy } },
+ '#withPriorityClassName':: d.fn(help='"Name of the priority class which will be used in every generated Pod, if the PriorityClass specified does not exist, the pod will not be able to schedule. Please refer to https://kubernetes.io/docs/concepts/scheduling-eviction/pod-priority-preemption/#priorityclass for more information"', args=[d.arg(name='priorityClassName', type=d.T.string)]),
+ withPriorityClassName(priorityClassName): { spec+: { priorityClassName: priorityClassName } },
+ '#withSchedulerName':: d.fn(help='"If specified, the pod will be dispatched by specified Kubernetes scheduler. If not specified, the pod will be dispatched by the default scheduler. More info: https://kubernetes.io/docs/concepts/scheduling-eviction/kube-scheduler/"', args=[d.arg(name='schedulerName', type=d.T.string)]),
+ withSchedulerName(schedulerName): { spec+: { schedulerName: schedulerName } },
+ '#withSmartShutdownTimeout':: d.fn(help='"The time in seconds that controls the window of time reserved for the smart shutdown of Postgres to complete. Make sure you reserve enough time for the operator to request a fast shutdown of Postgres (that is: `stopDelay` - `smartShutdownTimeout`)."', args=[d.arg(name='smartShutdownTimeout', type=d.T.integer)]),
+ withSmartShutdownTimeout(smartShutdownTimeout): { spec+: { smartShutdownTimeout: smartShutdownTimeout } },
+ '#withStartDelay':: d.fn(help='"The time in seconds that is allowed for a PostgreSQL instance to successfully start up (default 3600). The startup probe failure threshold is derived from this value using the formula: ceiling(startDelay / 10)."', args=[d.arg(name='startDelay', type=d.T.integer)]),
+ withStartDelay(startDelay): { spec+: { startDelay: startDelay } },
+ '#withStopDelay':: d.fn(help='"The time in seconds that is allowed for a PostgreSQL instance to gracefully shutdown (default 1800)"', args=[d.arg(name='stopDelay', type=d.T.integer)]),
+ withStopDelay(stopDelay): { spec+: { stopDelay: stopDelay } },
+ '#withSwitchoverDelay':: d.fn(help='"The time in seconds that is allowed for a primary PostgreSQL instance to gracefully shutdown during a switchover. Default value is 3600 seconds (1 hour)."', args=[d.arg(name='switchoverDelay', type=d.T.integer)]),
+ withSwitchoverDelay(switchoverDelay): { spec+: { switchoverDelay: switchoverDelay } },
+ '#withTopologySpreadConstraints':: d.fn(help='"TopologySpreadConstraints specifies how to spread matching pods among the given topology. More info: https://kubernetes.io/docs/concepts/scheduling-eviction/topology-spread-constraints/"', args=[d.arg(name='topologySpreadConstraints', type=d.T.array)]),
+ withTopologySpreadConstraints(topologySpreadConstraints): { spec+: { topologySpreadConstraints: if std.isArray(v=topologySpreadConstraints) then topologySpreadConstraints else [topologySpreadConstraints] } },
+ '#withTopologySpreadConstraintsMixin':: d.fn(help='"TopologySpreadConstraints specifies how to spread matching pods among the given topology. More info: https://kubernetes.io/docs/concepts/scheduling-eviction/topology-spread-constraints/"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='topologySpreadConstraints', type=d.T.array)]),
+ withTopologySpreadConstraintsMixin(topologySpreadConstraints): { spec+: { topologySpreadConstraints+: if std.isArray(v=topologySpreadConstraints) then topologySpreadConstraints else [topologySpreadConstraints] } },
+ },
+ '#mixin': 'ignore',
+ mixin: self,
+}
diff --git a/1.21.0/_gen/postgresql/v1/main.libsonnet b/1.21.0/_gen/postgresql/v1/main.libsonnet
new file mode 100644
index 0000000..2d20c16
--- /dev/null
+++ b/1.21.0/_gen/postgresql/v1/main.libsonnet
@@ -0,0 +1,8 @@
+{
+ local d = (import 'doc-util/main.libsonnet'),
+ '#':: d.pkg(name='v1', url='', help=''),
+ backup: (import 'backup.libsonnet'),
+ cluster: (import 'cluster.libsonnet'),
+ pooler: (import 'pooler.libsonnet'),
+ scheduledBackup: (import 'scheduledBackup.libsonnet'),
+}
diff --git a/1.21.0/_gen/postgresql/v1/pooler.libsonnet b/1.21.0/_gen/postgresql/v1/pooler.libsonnet
new file mode 100644
index 0000000..95018cc
--- /dev/null
+++ b/1.21.0/_gen/postgresql/v1/pooler.libsonnet
@@ -0,0 +1,2783 @@
+{
+ local d = (import 'doc-util/main.libsonnet'),
+ '#':: d.pkg(name='pooler', url='', help='"Pooler is the Schema for the poolers API"'),
+ '#metadata':: d.obj(help='"ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create."'),
+ metadata: {
+ '#withAnnotations':: d.fn(help='"Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations"', args=[d.arg(name='annotations', type=d.T.object)]),
+ withAnnotations(annotations): { metadata+: { annotations: annotations } },
+ '#withAnnotationsMixin':: d.fn(help='"Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]),
+ withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } },
+ '#withClusterName':: d.fn(help='"The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request."', args=[d.arg(name='clusterName', type=d.T.string)]),
+ withClusterName(clusterName): { metadata+: { clusterName: clusterName } },
+ '#withCreationTimestamp':: d.fn(help='"Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers."', args=[d.arg(name='creationTimestamp', type=d.T.string)]),
+ withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } },
+ '#withDeletionGracePeriodSeconds':: d.fn(help='"Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only."', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]),
+ withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } },
+ '#withDeletionTimestamp':: d.fn(help='"Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers."', args=[d.arg(name='deletionTimestamp', type=d.T.string)]),
+ withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } },
+ '#withFinalizers':: d.fn(help='"Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list."', args=[d.arg(name='finalizers', type=d.T.array)]),
+ withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } },
+ '#withFinalizersMixin':: d.fn(help='"Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]),
+ withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } },
+ '#withGenerateName':: d.fn(help='"GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\\n\\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\\n\\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency"', args=[d.arg(name='generateName', type=d.T.string)]),
+ withGenerateName(generateName): { metadata+: { generateName: generateName } },
+ '#withGeneration':: d.fn(help='"A sequence number representing a specific generation of the desired state. Populated by the system. Read-only."', args=[d.arg(name='generation', type=d.T.integer)]),
+ withGeneration(generation): { metadata+: { generation: generation } },
+ '#withLabels':: d.fn(help='"Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels"', args=[d.arg(name='labels', type=d.T.object)]),
+ withLabels(labels): { metadata+: { labels: labels } },
+ '#withLabelsMixin':: d.fn(help='"Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]),
+ withLabelsMixin(labels): { metadata+: { labels+: labels } },
+ '#withName':: d.fn(help='"Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names"', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { metadata+: { name: name } },
+ '#withNamespace':: d.fn(help='"Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the \\"default\\" namespace, but \\"default\\" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\\n\\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces"', args=[d.arg(name='namespace', type=d.T.string)]),
+ withNamespace(namespace): { metadata+: { namespace: namespace } },
+ '#withOwnerReferences':: d.fn(help='"List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller."', args=[d.arg(name='ownerReferences', type=d.T.array)]),
+ withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } },
+ '#withOwnerReferencesMixin':: d.fn(help='"List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]),
+ withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } },
+ '#withResourceVersion':: d.fn(help='"An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\\n\\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency"', args=[d.arg(name='resourceVersion', type=d.T.string)]),
+ withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } },
+ '#withSelfLink':: d.fn(help='"SelfLink is a URL representing this object. Populated by the system. Read-only.\\n\\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release."', args=[d.arg(name='selfLink', type=d.T.string)]),
+ withSelfLink(selfLink): { metadata+: { selfLink: selfLink } },
+ '#withUid':: d.fn(help='"UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\\n\\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids"', args=[d.arg(name='uid', type=d.T.string)]),
+ withUid(uid): { metadata+: { uid: uid } },
+ },
+ '#new':: d.fn(help='new returns an instance of Pooler', args=[d.arg(name='name', type=d.T.string)]),
+ new(name): {
+ apiVersion: 'postgresql.cnpg.io/v1',
+ kind: 'Pooler',
+ } + self.metadata.withName(name=name),
+ '#spec':: d.obj(help='"Specification of the desired behavior of the Pooler. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#spec-and-status"'),
+ spec: {
+ '#cluster':: d.obj(help='"This is the cluster reference on which the Pooler will work. Pooler name should never match with any cluster name within the same namespace."'),
+ cluster: {
+ '#withName':: d.fn(help='"Name of the referent."', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { spec+: { cluster+: { name: name } } },
+ },
+ '#deploymentStrategy':: d.obj(help='"The deployment strategy to use for pgbouncer to replace existing pods with new ones"'),
+ deploymentStrategy: {
+ '#rollingUpdate':: d.obj(help='"Rolling update config params. Present only if DeploymentStrategyType = RollingUpdate. --- TODO: Update this to follow our convention for oneOf, whatever we decide it to be."'),
+ rollingUpdate: {
+ '#withMaxSurge':: d.fn(help='"The maximum number of pods that can be scheduled above the desired number of pods. Value can be an absolute number (ex: 5) or a percentage of desired pods (ex: 10%). This can not be 0 if MaxUnavailable is 0. Absolute number is calculated from percentage by rounding up. Defaults to 25%. Example: when this is set to 30%, the new ReplicaSet can be scaled up immediately when the rolling update starts, such that the total number of old and new pods do not exceed 130% of desired pods. Once old pods have been killed, new ReplicaSet can be scaled up further, ensuring that total number of pods running at any time during the update is at most 130% of desired pods."', args=[d.arg(name='maxSurge', type=d.T.any)]),
+ withMaxSurge(maxSurge): { spec+: { deploymentStrategy+: { rollingUpdate+: { maxSurge: maxSurge } } } },
+ '#withMaxUnavailable':: d.fn(help='"The maximum number of pods that can be unavailable during the update. Value can be an absolute number (ex: 5) or a percentage of desired pods (ex: 10%). Absolute number is calculated from percentage by rounding down. This can not be 0 if MaxSurge is 0. Defaults to 25%. Example: when this is set to 30%, the old ReplicaSet can be scaled down to 70% of desired pods immediately when the rolling update starts. Once new pods are ready, old ReplicaSet can be scaled down further, followed by scaling up the new ReplicaSet, ensuring that the total number of pods available at all times during the update is at least 70% of desired pods."', args=[d.arg(name='maxUnavailable', type=d.T.any)]),
+ withMaxUnavailable(maxUnavailable): { spec+: { deploymentStrategy+: { rollingUpdate+: { maxUnavailable: maxUnavailable } } } },
+ },
+ '#withType':: d.fn(help='"Type of deployment. Can be \\"Recreate\\" or \\"RollingUpdate\\". Default is RollingUpdate."', args=[d.arg(name='type', type=d.T.string)]),
+ withType(type): { spec+: { deploymentStrategy+: { type: type } } },
+ },
+ '#monitoring':: d.obj(help='"The configuration of the monitoring infrastructure of this pooler."'),
+ monitoring: {
+ '#withEnablePodMonitor':: d.fn(help='"Enable or disable the `PodMonitor`"', args=[d.arg(name='enablePodMonitor', type=d.T.boolean)]),
+ withEnablePodMonitor(enablePodMonitor): { spec+: { monitoring+: { enablePodMonitor: enablePodMonitor } } },
+ },
+ '#pgbouncer':: d.obj(help='"The PgBouncer configuration"'),
+ pgbouncer: {
+ '#authQuerySecret':: d.obj(help='"The credentials of the user that need to be used for the authentication query. In case it is specified, also an AuthQuery (e.g. \\"SELECT usename, passwd FROM pg_shadow WHERE usename=$1\\") has to be specified and no automatic CNPG Cluster integration will be triggered."'),
+ authQuerySecret: {
+ '#withName':: d.fn(help='"Name of the referent."', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { spec+: { pgbouncer+: { authQuerySecret+: { name: name } } } },
+ },
+ '#withAuthQuery':: d.fn(help='"The query that will be used to download the hash of the password of a certain user. Default: \\"SELECT usename, passwd FROM user_search($1)\\". In case it is specified, also an AuthQuerySecret has to be specified and no automatic CNPG Cluster integration will be triggered."', args=[d.arg(name='authQuery', type=d.T.string)]),
+ withAuthQuery(authQuery): { spec+: { pgbouncer+: { authQuery: authQuery } } },
+ '#withParameters':: d.fn(help='"Additional parameters to be passed to PgBouncer - please check the CNPG documentation for a list of options you can configure"', args=[d.arg(name='parameters', type=d.T.object)]),
+ withParameters(parameters): { spec+: { pgbouncer+: { parameters: parameters } } },
+ '#withParametersMixin':: d.fn(help='"Additional parameters to be passed to PgBouncer - please check the CNPG documentation for a list of options you can configure"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='parameters', type=d.T.object)]),
+ withParametersMixin(parameters): { spec+: { pgbouncer+: { parameters+: parameters } } },
+ '#withPaused':: d.fn(help="\"When set to `true`, PgBouncer will disconnect from the PostgreSQL server, first waiting for all queries to complete, and pause all new client connections until this value is set to `false` (default). Internally, the operator calls PgBouncer's `PAUSE` and `RESUME` commands.\"", args=[d.arg(name='paused', type=d.T.boolean)]),
+ withPaused(paused): { spec+: { pgbouncer+: { paused: paused } } },
+ '#withPg_hba':: d.fn(help='"PostgreSQL Host Based Authentication rules (lines to be appended to the pg_hba.conf file)"', args=[d.arg(name='pg_hba', type=d.T.array)]),
+ withPg_hba(pg_hba): { spec+: { pgbouncer+: { pg_hba: if std.isArray(v=pg_hba) then pg_hba else [pg_hba] } } },
+ '#withPg_hbaMixin':: d.fn(help='"PostgreSQL Host Based Authentication rules (lines to be appended to the pg_hba.conf file)"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='pg_hba', type=d.T.array)]),
+ withPg_hbaMixin(pg_hba): { spec+: { pgbouncer+: { pg_hba+: if std.isArray(v=pg_hba) then pg_hba else [pg_hba] } } },
+ '#withPoolMode':: d.fn(help='"The pool mode. Default: `session`."', args=[d.arg(name='poolMode', type=d.T.string)]),
+ withPoolMode(poolMode): { spec+: { pgbouncer+: { poolMode: poolMode } } },
+ },
+ '#template':: d.obj(help='"The template of the Pod to be created"'),
+ template: {
+ '#metadata':: d.obj(help="\"Standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata\""),
+ metadata: {
+ '#withAnnotations':: d.fn(help='"Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations"', args=[d.arg(name='annotations', type=d.T.object)]),
+ withAnnotations(annotations): { spec+: { template+: { metadata+: { annotations: annotations } } } },
+ '#withAnnotationsMixin':: d.fn(help='"Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]),
+ withAnnotationsMixin(annotations): { spec+: { template+: { metadata+: { annotations+: annotations } } } },
+ '#withLabels':: d.fn(help='"Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels"', args=[d.arg(name='labels', type=d.T.object)]),
+ withLabels(labels): { spec+: { template+: { metadata+: { labels: labels } } } },
+ '#withLabelsMixin':: d.fn(help='"Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]),
+ withLabelsMixin(labels): { spec+: { template+: { metadata+: { labels+: labels } } } },
+ },
+ '#spec':: d.obj(help='"Specification of the desired behavior of the pod. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#spec-and-status"'),
+ spec: {
+ '#affinity':: d.obj(help="\"If specified, the pod's scheduling constraints\""),
+ affinity: {
+ '#nodeAffinity':: d.obj(help='"Describes node affinity scheduling rules for the pod."'),
+ nodeAffinity: {
+ '#preferredDuringSchedulingIgnoredDuringExecution':: d.obj(help='"The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding \\"weight\\" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred."'),
+ preferredDuringSchedulingIgnoredDuringExecution: {
+ '#preference':: d.obj(help='"A node selector term, associated with the corresponding weight."'),
+ preference: {
+ '#matchExpressions':: d.obj(help="\"A list of node selector requirements by node's labels.\""),
+ matchExpressions: {
+ '#withKey':: d.fn(help='"The label key that the selector applies to."', args=[d.arg(name='key', type=d.T.string)]),
+ withKey(key): { key: key },
+ '#withOperator':: d.fn(help="\"Represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists, DoesNotExist. Gt, and Lt.\"", args=[d.arg(name='operator', type=d.T.string)]),
+ withOperator(operator): { operator: operator },
+ '#withValues':: d.fn(help='"An array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. If the operator is Gt or Lt, the values array must have a single element, which will be interpreted as an integer. This array is replaced during a strategic merge patch."', args=[d.arg(name='values', type=d.T.array)]),
+ withValues(values): { values: if std.isArray(v=values) then values else [values] },
+ '#withValuesMixin':: d.fn(help='"An array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. If the operator is Gt or Lt, the values array must have a single element, which will be interpreted as an integer. This array is replaced during a strategic merge patch."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='values', type=d.T.array)]),
+ withValuesMixin(values): { values+: if std.isArray(v=values) then values else [values] },
+ },
+ '#matchFields':: d.obj(help="\"A list of node selector requirements by node's fields.\""),
+ matchFields: {
+ '#withKey':: d.fn(help='"The label key that the selector applies to."', args=[d.arg(name='key', type=d.T.string)]),
+ withKey(key): { key: key },
+ '#withOperator':: d.fn(help="\"Represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists, DoesNotExist. Gt, and Lt.\"", args=[d.arg(name='operator', type=d.T.string)]),
+ withOperator(operator): { operator: operator },
+ '#withValues':: d.fn(help='"An array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. If the operator is Gt or Lt, the values array must have a single element, which will be interpreted as an integer. This array is replaced during a strategic merge patch."', args=[d.arg(name='values', type=d.T.array)]),
+ withValues(values): { values: if std.isArray(v=values) then values else [values] },
+ '#withValuesMixin':: d.fn(help='"An array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. If the operator is Gt or Lt, the values array must have a single element, which will be interpreted as an integer. This array is replaced during a strategic merge patch."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='values', type=d.T.array)]),
+ withValuesMixin(values): { values+: if std.isArray(v=values) then values else [values] },
+ },
+ '#withMatchExpressions':: d.fn(help="\"A list of node selector requirements by node's labels.\"", args=[d.arg(name='matchExpressions', type=d.T.array)]),
+ withMatchExpressions(matchExpressions): { preference+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } },
+ '#withMatchExpressionsMixin':: d.fn(help="\"A list of node selector requirements by node's labels.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='matchExpressions', type=d.T.array)]),
+ withMatchExpressionsMixin(matchExpressions): { preference+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } },
+ '#withMatchFields':: d.fn(help="\"A list of node selector requirements by node's fields.\"", args=[d.arg(name='matchFields', type=d.T.array)]),
+ withMatchFields(matchFields): { preference+: { matchFields: if std.isArray(v=matchFields) then matchFields else [matchFields] } },
+ '#withMatchFieldsMixin':: d.fn(help="\"A list of node selector requirements by node's fields.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='matchFields', type=d.T.array)]),
+ withMatchFieldsMixin(matchFields): { preference+: { matchFields+: if std.isArray(v=matchFields) then matchFields else [matchFields] } },
+ },
+ '#withWeight':: d.fn(help='"Weight associated with matching the corresponding nodeSelectorTerm, in the range 1-100."', args=[d.arg(name='weight', type=d.T.integer)]),
+ withWeight(weight): { weight: weight },
+ },
+ '#requiredDuringSchedulingIgnoredDuringExecution':: d.obj(help='"If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to an update), the system may or may not try to eventually evict the pod from its node."'),
+ requiredDuringSchedulingIgnoredDuringExecution: {
+ '#nodeSelectorTerms':: d.obj(help='"Required. A list of node selector terms. The terms are ORed."'),
+ nodeSelectorTerms: {
+ '#matchExpressions':: d.obj(help="\"A list of node selector requirements by node's labels.\""),
+ matchExpressions: {
+ '#withKey':: d.fn(help='"The label key that the selector applies to."', args=[d.arg(name='key', type=d.T.string)]),
+ withKey(key): { key: key },
+ '#withOperator':: d.fn(help="\"Represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists, DoesNotExist. Gt, and Lt.\"", args=[d.arg(name='operator', type=d.T.string)]),
+ withOperator(operator): { operator: operator },
+ '#withValues':: d.fn(help='"An array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. If the operator is Gt or Lt, the values array must have a single element, which will be interpreted as an integer. This array is replaced during a strategic merge patch."', args=[d.arg(name='values', type=d.T.array)]),
+ withValues(values): { values: if std.isArray(v=values) then values else [values] },
+ '#withValuesMixin':: d.fn(help='"An array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. If the operator is Gt or Lt, the values array must have a single element, which will be interpreted as an integer. This array is replaced during a strategic merge patch."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='values', type=d.T.array)]),
+ withValuesMixin(values): { values+: if std.isArray(v=values) then values else [values] },
+ },
+ '#matchFields':: d.obj(help="\"A list of node selector requirements by node's fields.\""),
+ matchFields: {
+ '#withKey':: d.fn(help='"The label key that the selector applies to."', args=[d.arg(name='key', type=d.T.string)]),
+ withKey(key): { key: key },
+ '#withOperator':: d.fn(help="\"Represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists, DoesNotExist. Gt, and Lt.\"", args=[d.arg(name='operator', type=d.T.string)]),
+ withOperator(operator): { operator: operator },
+ '#withValues':: d.fn(help='"An array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. If the operator is Gt or Lt, the values array must have a single element, which will be interpreted as an integer. This array is replaced during a strategic merge patch."', args=[d.arg(name='values', type=d.T.array)]),
+ withValues(values): { values: if std.isArray(v=values) then values else [values] },
+ '#withValuesMixin':: d.fn(help='"An array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. If the operator is Gt or Lt, the values array must have a single element, which will be interpreted as an integer. This array is replaced during a strategic merge patch."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='values', type=d.T.array)]),
+ withValuesMixin(values): { values+: if std.isArray(v=values) then values else [values] },
+ },
+ '#withMatchExpressions':: d.fn(help="\"A list of node selector requirements by node's labels.\"", args=[d.arg(name='matchExpressions', type=d.T.array)]),
+ withMatchExpressions(matchExpressions): { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] },
+ '#withMatchExpressionsMixin':: d.fn(help="\"A list of node selector requirements by node's labels.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='matchExpressions', type=d.T.array)]),
+ withMatchExpressionsMixin(matchExpressions): { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] },
+ '#withMatchFields':: d.fn(help="\"A list of node selector requirements by node's fields.\"", args=[d.arg(name='matchFields', type=d.T.array)]),
+ withMatchFields(matchFields): { matchFields: if std.isArray(v=matchFields) then matchFields else [matchFields] },
+ '#withMatchFieldsMixin':: d.fn(help="\"A list of node selector requirements by node's fields.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='matchFields', type=d.T.array)]),
+ withMatchFieldsMixin(matchFields): { matchFields+: if std.isArray(v=matchFields) then matchFields else [matchFields] },
+ },
+ '#withNodeSelectorTerms':: d.fn(help='"Required. A list of node selector terms. The terms are ORed."', args=[d.arg(name='nodeSelectorTerms', type=d.T.array)]),
+ withNodeSelectorTerms(nodeSelectorTerms): { spec+: { template+: { spec+: { affinity+: { nodeAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: { nodeSelectorTerms: if std.isArray(v=nodeSelectorTerms) then nodeSelectorTerms else [nodeSelectorTerms] } } } } } } },
+ '#withNodeSelectorTermsMixin':: d.fn(help='"Required. A list of node selector terms. The terms are ORed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='nodeSelectorTerms', type=d.T.array)]),
+ withNodeSelectorTermsMixin(nodeSelectorTerms): { spec+: { template+: { spec+: { affinity+: { nodeAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: { nodeSelectorTerms+: if std.isArray(v=nodeSelectorTerms) then nodeSelectorTerms else [nodeSelectorTerms] } } } } } } },
+ },
+ '#withPreferredDuringSchedulingIgnoredDuringExecution':: d.fn(help='"The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding \\"weight\\" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred."', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]),
+ withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { nodeAffinity+: { preferredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } } },
+ '#withPreferredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='"The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding \\"weight\\" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]),
+ withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { nodeAffinity+: { preferredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } } },
+ },
+ '#podAffinity':: d.obj(help='"Describes pod affinity scheduling rules (e.g. co-locate this pod in the same node, zone, etc. as some other pod(s))."'),
+ podAffinity: {
+ '#preferredDuringSchedulingIgnoredDuringExecution':: d.obj(help='"The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding \\"weight\\" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred."'),
+ preferredDuringSchedulingIgnoredDuringExecution: {
+ '#podAffinityTerm':: d.obj(help='"Required. A pod affinity term, associated with the corresponding weight."'),
+ podAffinityTerm: {
+ '#labelSelector':: d.obj(help='"A label query over a set of resources, in this case pods."'),
+ labelSelector: {
+ '#matchExpressions':: d.obj(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."'),
+ matchExpressions: {
+ '#withKey':: d.fn(help='"key is the label key that the selector applies to."', args=[d.arg(name='key', type=d.T.string)]),
+ withKey(key): { key: key },
+ '#withOperator':: d.fn(help="\"operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist.\"", args=[d.arg(name='operator', type=d.T.string)]),
+ withOperator(operator): { operator: operator },
+ '#withValues':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."', args=[d.arg(name='values', type=d.T.array)]),
+ withValues(values): { values: if std.isArray(v=values) then values else [values] },
+ '#withValuesMixin':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='values', type=d.T.array)]),
+ withValuesMixin(values): { values+: if std.isArray(v=values) then values else [values] },
+ },
+ '#withMatchExpressions':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."', args=[d.arg(name='matchExpressions', type=d.T.array)]),
+ withMatchExpressions(matchExpressions): { podAffinityTerm+: { labelSelector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } },
+ '#withMatchExpressionsMixin':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]),
+ withMatchExpressionsMixin(matchExpressions): { podAffinityTerm+: { labelSelector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } },
+ '#withMatchLabels':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."', args=[d.arg(name='matchLabels', type=d.T.object)]),
+ withMatchLabels(matchLabels): { podAffinityTerm+: { labelSelector+: { matchLabels: matchLabels } } },
+ '#withMatchLabelsMixin':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]),
+ withMatchLabelsMixin(matchLabels): { podAffinityTerm+: { labelSelector+: { matchLabels+: matchLabels } } },
+ },
+ '#namespaceSelector':: d.obj(help="\"A label query over the set of namespaces that the term applies to. The term is applied to the union of the namespaces selected by this field and the ones listed in the namespaces field. null selector and null or empty namespaces list means \\\"this pod's namespace\\\". An empty selector ({}) matches all namespaces.\""),
+ namespaceSelector: {
+ '#matchExpressions':: d.obj(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."'),
+ matchExpressions: {
+ '#withKey':: d.fn(help='"key is the label key that the selector applies to."', args=[d.arg(name='key', type=d.T.string)]),
+ withKey(key): { key: key },
+ '#withOperator':: d.fn(help="\"operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist.\"", args=[d.arg(name='operator', type=d.T.string)]),
+ withOperator(operator): { operator: operator },
+ '#withValues':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."', args=[d.arg(name='values', type=d.T.array)]),
+ withValues(values): { values: if std.isArray(v=values) then values else [values] },
+ '#withValuesMixin':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='values', type=d.T.array)]),
+ withValuesMixin(values): { values+: if std.isArray(v=values) then values else [values] },
+ },
+ '#withMatchExpressions':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."', args=[d.arg(name='matchExpressions', type=d.T.array)]),
+ withMatchExpressions(matchExpressions): { podAffinityTerm+: { namespaceSelector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } },
+ '#withMatchExpressionsMixin':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]),
+ withMatchExpressionsMixin(matchExpressions): { podAffinityTerm+: { namespaceSelector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } },
+ '#withMatchLabels':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."', args=[d.arg(name='matchLabels', type=d.T.object)]),
+ withMatchLabels(matchLabels): { podAffinityTerm+: { namespaceSelector+: { matchLabels: matchLabels } } },
+ '#withMatchLabelsMixin':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]),
+ withMatchLabelsMixin(matchLabels): { podAffinityTerm+: { namespaceSelector+: { matchLabels+: matchLabels } } },
+ },
+ '#withNamespaces':: d.fn(help="\"namespaces specifies a static list of namespace names that the term applies to. The term is applied to the union of the namespaces listed in this field and the ones selected by namespaceSelector. null or empty namespaces list and null namespaceSelector means \\\"this pod's namespace\\\".\"", args=[d.arg(name='namespaces', type=d.T.array)]),
+ withNamespaces(namespaces): { podAffinityTerm+: { namespaces: if std.isArray(v=namespaces) then namespaces else [namespaces] } },
+ '#withNamespacesMixin':: d.fn(help="\"namespaces specifies a static list of namespace names that the term applies to. The term is applied to the union of the namespaces listed in this field and the ones selected by namespaceSelector. null or empty namespaces list and null namespaceSelector means \\\"this pod's namespace\\\".\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='namespaces', type=d.T.array)]),
+ withNamespacesMixin(namespaces): { podAffinityTerm+: { namespaces+: if std.isArray(v=namespaces) then namespaces else [namespaces] } },
+ '#withTopologyKey':: d.fn(help='"This pod should be co-located (affinity) or not co-located (anti-affinity) with the pods matching the labelSelector in the specified namespaces, where co-located is defined as running on a node whose value of the label with key topologyKey matches that of any node on which any of the selected pods is running. Empty topologyKey is not allowed."', args=[d.arg(name='topologyKey', type=d.T.string)]),
+ withTopologyKey(topologyKey): { podAffinityTerm+: { topologyKey: topologyKey } },
+ },
+ '#withWeight':: d.fn(help='"weight associated with matching the corresponding podAffinityTerm, in the range 1-100."', args=[d.arg(name='weight', type=d.T.integer)]),
+ withWeight(weight): { weight: weight },
+ },
+ '#requiredDuringSchedulingIgnoredDuringExecution':: d.obj(help='"If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied."'),
+ requiredDuringSchedulingIgnoredDuringExecution: {
+ '#labelSelector':: d.obj(help='"A label query over a set of resources, in this case pods."'),
+ labelSelector: {
+ '#matchExpressions':: d.obj(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."'),
+ matchExpressions: {
+ '#withKey':: d.fn(help='"key is the label key that the selector applies to."', args=[d.arg(name='key', type=d.T.string)]),
+ withKey(key): { key: key },
+ '#withOperator':: d.fn(help="\"operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist.\"", args=[d.arg(name='operator', type=d.T.string)]),
+ withOperator(operator): { operator: operator },
+ '#withValues':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."', args=[d.arg(name='values', type=d.T.array)]),
+ withValues(values): { values: if std.isArray(v=values) then values else [values] },
+ '#withValuesMixin':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='values', type=d.T.array)]),
+ withValuesMixin(values): { values+: if std.isArray(v=values) then values else [values] },
+ },
+ '#withMatchExpressions':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."', args=[d.arg(name='matchExpressions', type=d.T.array)]),
+ withMatchExpressions(matchExpressions): { labelSelector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } },
+ '#withMatchExpressionsMixin':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]),
+ withMatchExpressionsMixin(matchExpressions): { labelSelector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } },
+ '#withMatchLabels':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."', args=[d.arg(name='matchLabels', type=d.T.object)]),
+ withMatchLabels(matchLabels): { labelSelector+: { matchLabels: matchLabels } },
+ '#withMatchLabelsMixin':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]),
+ withMatchLabelsMixin(matchLabels): { labelSelector+: { matchLabels+: matchLabels } },
+ },
+ '#namespaceSelector':: d.obj(help="\"A label query over the set of namespaces that the term applies to. The term is applied to the union of the namespaces selected by this field and the ones listed in the namespaces field. null selector and null or empty namespaces list means \\\"this pod's namespace\\\". An empty selector ({}) matches all namespaces.\""),
+ namespaceSelector: {
+ '#matchExpressions':: d.obj(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."'),
+ matchExpressions: {
+ '#withKey':: d.fn(help='"key is the label key that the selector applies to."', args=[d.arg(name='key', type=d.T.string)]),
+ withKey(key): { key: key },
+ '#withOperator':: d.fn(help="\"operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist.\"", args=[d.arg(name='operator', type=d.T.string)]),
+ withOperator(operator): { operator: operator },
+ '#withValues':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."', args=[d.arg(name='values', type=d.T.array)]),
+ withValues(values): { values: if std.isArray(v=values) then values else [values] },
+ '#withValuesMixin':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='values', type=d.T.array)]),
+ withValuesMixin(values): { values+: if std.isArray(v=values) then values else [values] },
+ },
+ '#withMatchExpressions':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."', args=[d.arg(name='matchExpressions', type=d.T.array)]),
+ withMatchExpressions(matchExpressions): { namespaceSelector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } },
+ '#withMatchExpressionsMixin':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]),
+ withMatchExpressionsMixin(matchExpressions): { namespaceSelector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } },
+ '#withMatchLabels':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."', args=[d.arg(name='matchLabels', type=d.T.object)]),
+ withMatchLabels(matchLabels): { namespaceSelector+: { matchLabels: matchLabels } },
+ '#withMatchLabelsMixin':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]),
+ withMatchLabelsMixin(matchLabels): { namespaceSelector+: { matchLabels+: matchLabels } },
+ },
+ '#withNamespaces':: d.fn(help="\"namespaces specifies a static list of namespace names that the term applies to. The term is applied to the union of the namespaces listed in this field and the ones selected by namespaceSelector. null or empty namespaces list and null namespaceSelector means \\\"this pod's namespace\\\".\"", args=[d.arg(name='namespaces', type=d.T.array)]),
+ withNamespaces(namespaces): { namespaces: if std.isArray(v=namespaces) then namespaces else [namespaces] },
+ '#withNamespacesMixin':: d.fn(help="\"namespaces specifies a static list of namespace names that the term applies to. The term is applied to the union of the namespaces listed in this field and the ones selected by namespaceSelector. null or empty namespaces list and null namespaceSelector means \\\"this pod's namespace\\\".\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='namespaces', type=d.T.array)]),
+ withNamespacesMixin(namespaces): { namespaces+: if std.isArray(v=namespaces) then namespaces else [namespaces] },
+ '#withTopologyKey':: d.fn(help='"This pod should be co-located (affinity) or not co-located (anti-affinity) with the pods matching the labelSelector in the specified namespaces, where co-located is defined as running on a node whose value of the label with key topologyKey matches that of any node on which any of the selected pods is running. Empty topologyKey is not allowed."', args=[d.arg(name='topologyKey', type=d.T.string)]),
+ withTopologyKey(topologyKey): { topologyKey: topologyKey },
+ },
+ '#withPreferredDuringSchedulingIgnoredDuringExecution':: d.fn(help='"The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding \\"weight\\" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred."', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]),
+ withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { podAffinity+: { preferredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } } },
+ '#withPreferredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='"The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding \\"weight\\" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]),
+ withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { podAffinity+: { preferredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } } },
+ '#withRequiredDuringSchedulingIgnoredDuringExecution':: d.fn(help='"If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied."', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]),
+ withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { podAffinity+: { requiredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } } } },
+ '#withRequiredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='"If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]),
+ withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { podAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } } } },
+ },
+ '#podAntiAffinity':: d.obj(help='"Describes pod anti-affinity scheduling rules (e.g. avoid putting this pod in the same node, zone, etc. as some other pod(s))."'),
+ podAntiAffinity: {
+ '#preferredDuringSchedulingIgnoredDuringExecution':: d.obj(help='"The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding \\"weight\\" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred."'),
+ preferredDuringSchedulingIgnoredDuringExecution: {
+ '#podAffinityTerm':: d.obj(help='"Required. A pod affinity term, associated with the corresponding weight."'),
+ podAffinityTerm: {
+ '#labelSelector':: d.obj(help='"A label query over a set of resources, in this case pods."'),
+ labelSelector: {
+ '#matchExpressions':: d.obj(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."'),
+ matchExpressions: {
+ '#withKey':: d.fn(help='"key is the label key that the selector applies to."', args=[d.arg(name='key', type=d.T.string)]),
+ withKey(key): { key: key },
+ '#withOperator':: d.fn(help="\"operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist.\"", args=[d.arg(name='operator', type=d.T.string)]),
+ withOperator(operator): { operator: operator },
+ '#withValues':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."', args=[d.arg(name='values', type=d.T.array)]),
+ withValues(values): { values: if std.isArray(v=values) then values else [values] },
+ '#withValuesMixin':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='values', type=d.T.array)]),
+ withValuesMixin(values): { values+: if std.isArray(v=values) then values else [values] },
+ },
+ '#withMatchExpressions':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."', args=[d.arg(name='matchExpressions', type=d.T.array)]),
+ withMatchExpressions(matchExpressions): { podAffinityTerm+: { labelSelector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } },
+ '#withMatchExpressionsMixin':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]),
+ withMatchExpressionsMixin(matchExpressions): { podAffinityTerm+: { labelSelector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } },
+ '#withMatchLabels':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."', args=[d.arg(name='matchLabels', type=d.T.object)]),
+ withMatchLabels(matchLabels): { podAffinityTerm+: { labelSelector+: { matchLabels: matchLabels } } },
+ '#withMatchLabelsMixin':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]),
+ withMatchLabelsMixin(matchLabels): { podAffinityTerm+: { labelSelector+: { matchLabels+: matchLabels } } },
+ },
+ '#namespaceSelector':: d.obj(help="\"A label query over the set of namespaces that the term applies to. The term is applied to the union of the namespaces selected by this field and the ones listed in the namespaces field. null selector and null or empty namespaces list means \\\"this pod's namespace\\\". An empty selector ({}) matches all namespaces.\""),
+ namespaceSelector: {
+ '#matchExpressions':: d.obj(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."'),
+ matchExpressions: {
+ '#withKey':: d.fn(help='"key is the label key that the selector applies to."', args=[d.arg(name='key', type=d.T.string)]),
+ withKey(key): { key: key },
+ '#withOperator':: d.fn(help="\"operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist.\"", args=[d.arg(name='operator', type=d.T.string)]),
+ withOperator(operator): { operator: operator },
+ '#withValues':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."', args=[d.arg(name='values', type=d.T.array)]),
+ withValues(values): { values: if std.isArray(v=values) then values else [values] },
+ '#withValuesMixin':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='values', type=d.T.array)]),
+ withValuesMixin(values): { values+: if std.isArray(v=values) then values else [values] },
+ },
+ '#withMatchExpressions':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."', args=[d.arg(name='matchExpressions', type=d.T.array)]),
+ withMatchExpressions(matchExpressions): { podAffinityTerm+: { namespaceSelector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } },
+ '#withMatchExpressionsMixin':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]),
+ withMatchExpressionsMixin(matchExpressions): { podAffinityTerm+: { namespaceSelector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } },
+ '#withMatchLabels':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."', args=[d.arg(name='matchLabels', type=d.T.object)]),
+ withMatchLabels(matchLabels): { podAffinityTerm+: { namespaceSelector+: { matchLabels: matchLabels } } },
+ '#withMatchLabelsMixin':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]),
+ withMatchLabelsMixin(matchLabels): { podAffinityTerm+: { namespaceSelector+: { matchLabels+: matchLabels } } },
+ },
+ '#withNamespaces':: d.fn(help="\"namespaces specifies a static list of namespace names that the term applies to. The term is applied to the union of the namespaces listed in this field and the ones selected by namespaceSelector. null or empty namespaces list and null namespaceSelector means \\\"this pod's namespace\\\".\"", args=[d.arg(name='namespaces', type=d.T.array)]),
+ withNamespaces(namespaces): { podAffinityTerm+: { namespaces: if std.isArray(v=namespaces) then namespaces else [namespaces] } },
+ '#withNamespacesMixin':: d.fn(help="\"namespaces specifies a static list of namespace names that the term applies to. The term is applied to the union of the namespaces listed in this field and the ones selected by namespaceSelector. null or empty namespaces list and null namespaceSelector means \\\"this pod's namespace\\\".\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='namespaces', type=d.T.array)]),
+ withNamespacesMixin(namespaces): { podAffinityTerm+: { namespaces+: if std.isArray(v=namespaces) then namespaces else [namespaces] } },
+ '#withTopologyKey':: d.fn(help='"This pod should be co-located (affinity) or not co-located (anti-affinity) with the pods matching the labelSelector in the specified namespaces, where co-located is defined as running on a node whose value of the label with key topologyKey matches that of any node on which any of the selected pods is running. Empty topologyKey is not allowed."', args=[d.arg(name='topologyKey', type=d.T.string)]),
+ withTopologyKey(topologyKey): { podAffinityTerm+: { topologyKey: topologyKey } },
+ },
+ '#withWeight':: d.fn(help='"weight associated with matching the corresponding podAffinityTerm, in the range 1-100."', args=[d.arg(name='weight', type=d.T.integer)]),
+ withWeight(weight): { weight: weight },
+ },
+ '#requiredDuringSchedulingIgnoredDuringExecution':: d.obj(help='"If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied."'),
+ requiredDuringSchedulingIgnoredDuringExecution: {
+ '#labelSelector':: d.obj(help='"A label query over a set of resources, in this case pods."'),
+ labelSelector: {
+ '#matchExpressions':: d.obj(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."'),
+ matchExpressions: {
+ '#withKey':: d.fn(help='"key is the label key that the selector applies to."', args=[d.arg(name='key', type=d.T.string)]),
+ withKey(key): { key: key },
+ '#withOperator':: d.fn(help="\"operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist.\"", args=[d.arg(name='operator', type=d.T.string)]),
+ withOperator(operator): { operator: operator },
+ '#withValues':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."', args=[d.arg(name='values', type=d.T.array)]),
+ withValues(values): { values: if std.isArray(v=values) then values else [values] },
+ '#withValuesMixin':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='values', type=d.T.array)]),
+ withValuesMixin(values): { values+: if std.isArray(v=values) then values else [values] },
+ },
+ '#withMatchExpressions':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."', args=[d.arg(name='matchExpressions', type=d.T.array)]),
+ withMatchExpressions(matchExpressions): { labelSelector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } },
+ '#withMatchExpressionsMixin':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]),
+ withMatchExpressionsMixin(matchExpressions): { labelSelector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } },
+ '#withMatchLabels':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."', args=[d.arg(name='matchLabels', type=d.T.object)]),
+ withMatchLabels(matchLabels): { labelSelector+: { matchLabels: matchLabels } },
+ '#withMatchLabelsMixin':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]),
+ withMatchLabelsMixin(matchLabels): { labelSelector+: { matchLabels+: matchLabels } },
+ },
+ '#namespaceSelector':: d.obj(help="\"A label query over the set of namespaces that the term applies to. The term is applied to the union of the namespaces selected by this field and the ones listed in the namespaces field. null selector and null or empty namespaces list means \\\"this pod's namespace\\\". An empty selector ({}) matches all namespaces.\""),
+ namespaceSelector: {
+ '#matchExpressions':: d.obj(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."'),
+ matchExpressions: {
+ '#withKey':: d.fn(help='"key is the label key that the selector applies to."', args=[d.arg(name='key', type=d.T.string)]),
+ withKey(key): { key: key },
+ '#withOperator':: d.fn(help="\"operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist.\"", args=[d.arg(name='operator', type=d.T.string)]),
+ withOperator(operator): { operator: operator },
+ '#withValues':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."', args=[d.arg(name='values', type=d.T.array)]),
+ withValues(values): { values: if std.isArray(v=values) then values else [values] },
+ '#withValuesMixin':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='values', type=d.T.array)]),
+ withValuesMixin(values): { values+: if std.isArray(v=values) then values else [values] },
+ },
+ '#withMatchExpressions':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."', args=[d.arg(name='matchExpressions', type=d.T.array)]),
+ withMatchExpressions(matchExpressions): { namespaceSelector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } },
+ '#withMatchExpressionsMixin':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]),
+ withMatchExpressionsMixin(matchExpressions): { namespaceSelector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } },
+ '#withMatchLabels':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."', args=[d.arg(name='matchLabels', type=d.T.object)]),
+ withMatchLabels(matchLabels): { namespaceSelector+: { matchLabels: matchLabels } },
+ '#withMatchLabelsMixin':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]),
+ withMatchLabelsMixin(matchLabels): { namespaceSelector+: { matchLabels+: matchLabels } },
+ },
+ '#withNamespaces':: d.fn(help="\"namespaces specifies a static list of namespace names that the term applies to. The term is applied to the union of the namespaces listed in this field and the ones selected by namespaceSelector. null or empty namespaces list and null namespaceSelector means \\\"this pod's namespace\\\".\"", args=[d.arg(name='namespaces', type=d.T.array)]),
+ withNamespaces(namespaces): { namespaces: if std.isArray(v=namespaces) then namespaces else [namespaces] },
+ '#withNamespacesMixin':: d.fn(help="\"namespaces specifies a static list of namespace names that the term applies to. The term is applied to the union of the namespaces listed in this field and the ones selected by namespaceSelector. null or empty namespaces list and null namespaceSelector means \\\"this pod's namespace\\\".\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='namespaces', type=d.T.array)]),
+ withNamespacesMixin(namespaces): { namespaces+: if std.isArray(v=namespaces) then namespaces else [namespaces] },
+ '#withTopologyKey':: d.fn(help='"This pod should be co-located (affinity) or not co-located (anti-affinity) with the pods matching the labelSelector in the specified namespaces, where co-located is defined as running on a node whose value of the label with key topologyKey matches that of any node on which any of the selected pods is running. Empty topologyKey is not allowed."', args=[d.arg(name='topologyKey', type=d.T.string)]),
+ withTopologyKey(topologyKey): { topologyKey: topologyKey },
+ },
+ '#withPreferredDuringSchedulingIgnoredDuringExecution':: d.fn(help='"The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding \\"weight\\" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred."', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]),
+ withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { podAntiAffinity+: { preferredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } } },
+ '#withPreferredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='"The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding \\"weight\\" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='preferredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]),
+ withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { podAntiAffinity+: { preferredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=preferredDuringSchedulingIgnoredDuringExecution) then preferredDuringSchedulingIgnoredDuringExecution else [preferredDuringSchedulingIgnoredDuringExecution] } } } } } },
+ '#withRequiredDuringSchedulingIgnoredDuringExecution':: d.fn(help='"If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied."', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]),
+ withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { podAntiAffinity+: { requiredDuringSchedulingIgnoredDuringExecution: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } } } },
+ '#withRequiredDuringSchedulingIgnoredDuringExecutionMixin':: d.fn(help='"If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='requiredDuringSchedulingIgnoredDuringExecution', type=d.T.array)]),
+ withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution): { spec+: { template+: { spec+: { affinity+: { podAntiAffinity+: { requiredDuringSchedulingIgnoredDuringExecution+: if std.isArray(v=requiredDuringSchedulingIgnoredDuringExecution) then requiredDuringSchedulingIgnoredDuringExecution else [requiredDuringSchedulingIgnoredDuringExecution] } } } } } },
+ },
+ },
+ '#containers':: d.obj(help='"List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated."'),
+ containers: {
+ '#env':: d.obj(help='"List of environment variables to set in the container. Cannot be updated."'),
+ env: {
+ '#valueFrom':: d.obj(help="\"Source for the environment variable's value. Cannot be used if value is not empty.\""),
+ valueFrom: {
+ '#configMapKeyRef':: d.obj(help='"Selects a key of a ConfigMap."'),
+ configMapKeyRef: {
+ '#withKey':: d.fn(help='"The key to select."', args=[d.arg(name='key', type=d.T.string)]),
+ withKey(key): { valueFrom+: { configMapKeyRef+: { key: key } } },
+ '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { valueFrom+: { configMapKeyRef+: { name: name } } },
+ '#withOptional':: d.fn(help='"Specify whether the ConfigMap or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]),
+ withOptional(optional): { valueFrom+: { configMapKeyRef+: { optional: optional } } },
+ },
+ '#fieldRef':: d.obj(help="\"Selects a field of the pod: supports metadata.name, metadata.namespace, `metadata.labels['\u003cKEY\u003e']`, `metadata.annotations['\u003cKEY\u003e']`, spec.nodeName, spec.serviceAccountName, status.hostIP, status.podIP, status.podIPs.\""),
+ fieldRef: {
+ '#withApiVersion':: d.fn(help='"Version of the schema the FieldPath is written in terms of, defaults to \\"v1\\"."', args=[d.arg(name='apiVersion', type=d.T.string)]),
+ withApiVersion(apiVersion): { valueFrom+: { fieldRef+: { apiVersion: apiVersion } } },
+ '#withFieldPath':: d.fn(help='"Path of the field to select in the specified API version."', args=[d.arg(name='fieldPath', type=d.T.string)]),
+ withFieldPath(fieldPath): { valueFrom+: { fieldRef+: { fieldPath: fieldPath } } },
+ },
+ '#resourceFieldRef':: d.obj(help='"Selects a resource of the container: only resources limits and requests (limits.cpu, limits.memory, limits.ephemeral-storage, requests.cpu, requests.memory and requests.ephemeral-storage) are currently supported."'),
+ resourceFieldRef: {
+ '#withContainerName':: d.fn(help='"Container name: required for volumes, optional for env vars"', args=[d.arg(name='containerName', type=d.T.string)]),
+ withContainerName(containerName): { valueFrom+: { resourceFieldRef+: { containerName: containerName } } },
+ '#withDivisor':: d.fn(help='"Specifies the output format of the exposed resources, defaults to \\"1\\', args=[d.arg(name='divisor', type=d.T.any)]),
+ withDivisor(divisor): { valueFrom+: { resourceFieldRef+: { divisor: divisor } } },
+ '#withResource':: d.fn(help='"Required: resource to select"', args=[d.arg(name='resource', type=d.T.string)]),
+ withResource(resource): { valueFrom+: { resourceFieldRef+: { resource: resource } } },
+ },
+ '#secretKeyRef':: d.obj(help="\"Selects a key of a secret in the pod's namespace\""),
+ secretKeyRef: {
+ '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]),
+ withKey(key): { valueFrom+: { secretKeyRef+: { key: key } } },
+ '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { valueFrom+: { secretKeyRef+: { name: name } } },
+ '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]),
+ withOptional(optional): { valueFrom+: { secretKeyRef+: { optional: optional } } },
+ },
+ },
+ '#withName':: d.fn(help='"Name of the environment variable. Must be a C_IDENTIFIER."', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { name: name },
+ '#withValue':: d.fn(help='"Variable references $(VAR_NAME) are expanded using the previously defined environment variables in the container and any service environment variables. If a variable cannot be resolved, the reference in the input string will be unchanged. Double $$ are reduced to a single $, which allows for escaping the $(VAR_NAME) syntax: i.e. \\"$$(VAR_NAME)\\" will produce the string literal \\"$(VAR_NAME)\\". Escaped references will never be expanded, regardless of whether the variable exists or not. Defaults to \\"\\"."', args=[d.arg(name='value', type=d.T.string)]),
+ withValue(value): { value: value },
+ },
+ '#envFrom':: d.obj(help='"List of sources to populate environment variables in the container. The keys defined within a source must be a C_IDENTIFIER. All invalid keys will be reported as an event when the container is starting. When a key exists in multiple sources, the value associated with the last source will take precedence. Values defined by an Env with a duplicate key will take precedence. Cannot be updated."'),
+ envFrom: {
+ '#configMapRef':: d.obj(help='"The ConfigMap to select from"'),
+ configMapRef: {
+ '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { configMapRef+: { name: name } },
+ '#withOptional':: d.fn(help='"Specify whether the ConfigMap must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]),
+ withOptional(optional): { configMapRef+: { optional: optional } },
+ },
+ '#secretRef':: d.obj(help='"The Secret to select from"'),
+ secretRef: {
+ '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { secretRef+: { name: name } },
+ '#withOptional':: d.fn(help='"Specify whether the Secret must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]),
+ withOptional(optional): { secretRef+: { optional: optional } },
+ },
+ '#withPrefix':: d.fn(help='"An optional identifier to prepend to each key in the ConfigMap. Must be a C_IDENTIFIER."', args=[d.arg(name='prefix', type=d.T.string)]),
+ withPrefix(prefix): { prefix: prefix },
+ },
+ '#lifecycle':: d.obj(help='"Actions that the management system should take in response to container lifecycle events. Cannot be updated."'),
+ lifecycle: {
+ '#postStart':: d.obj(help='"PostStart is called immediately after a container is created. If the handler fails, the container is terminated and restarted according to its restart policy. Other management of the container blocks until the hook completes. More info: https://kubernetes.io/docs/concepts/containers/container-lifecycle-hooks/#container-hooks"'),
+ postStart: {
+ '#exec':: d.obj(help='"Exec specifies the action to take."'),
+ exec: {
+ '#withCommand':: d.fn(help="\"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.\"", args=[d.arg(name='command', type=d.T.array)]),
+ withCommand(command): { lifecycle+: { postStart+: { exec+: { command: if std.isArray(v=command) then command else [command] } } } },
+ '#withCommandMixin':: d.fn(help="\"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='command', type=d.T.array)]),
+ withCommandMixin(command): { lifecycle+: { postStart+: { exec+: { command+: if std.isArray(v=command) then command else [command] } } } },
+ },
+ '#httpGet':: d.obj(help='"HTTPGet specifies the http request to perform."'),
+ httpGet: {
+ '#httpHeaders':: d.obj(help='"Custom headers to set in the request. HTTP allows repeated headers."'),
+ httpHeaders: {
+ '#withName':: d.fn(help='"The header field name. This will be canonicalized upon output, so case-variant names will be understood as the same header."', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { name: name },
+ '#withValue':: d.fn(help='"The header field value"', args=[d.arg(name='value', type=d.T.string)]),
+ withValue(value): { value: value },
+ },
+ '#withHost':: d.fn(help='"Host name to connect to, defaults to the pod IP. You probably want to set \\"Host\\" in httpHeaders instead."', args=[d.arg(name='host', type=d.T.string)]),
+ withHost(host): { lifecycle+: { postStart+: { httpGet+: { host: host } } } },
+ '#withHttpHeaders':: d.fn(help='"Custom headers to set in the request. HTTP allows repeated headers."', args=[d.arg(name='httpHeaders', type=d.T.array)]),
+ withHttpHeaders(httpHeaders): { lifecycle+: { postStart+: { httpGet+: { httpHeaders: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } } },
+ '#withHttpHeadersMixin':: d.fn(help='"Custom headers to set in the request. HTTP allows repeated headers."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='httpHeaders', type=d.T.array)]),
+ withHttpHeadersMixin(httpHeaders): { lifecycle+: { postStart+: { httpGet+: { httpHeaders+: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } } },
+ '#withPath':: d.fn(help='"Path to access on the HTTP server."', args=[d.arg(name='path', type=d.T.string)]),
+ withPath(path): { lifecycle+: { postStart+: { httpGet+: { path: path } } } },
+ '#withPort':: d.fn(help='"Name or number of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."', args=[d.arg(name='port', type=d.T.any)]),
+ withPort(port): { lifecycle+: { postStart+: { httpGet+: { port: port } } } },
+ '#withScheme':: d.fn(help='"Scheme to use for connecting to the host. Defaults to HTTP."', args=[d.arg(name='scheme', type=d.T.string)]),
+ withScheme(scheme): { lifecycle+: { postStart+: { httpGet+: { scheme: scheme } } } },
+ },
+ '#tcpSocket':: d.obj(help='"Deprecated. TCPSocket is NOT supported as a LifecycleHandler and kept for the backward compatibility. There are no validation of this field and lifecycle hooks will fail in runtime when tcp handler is specified."'),
+ tcpSocket: {
+ '#withHost':: d.fn(help='"Optional: Host name to connect to, defaults to the pod IP."', args=[d.arg(name='host', type=d.T.string)]),
+ withHost(host): { lifecycle+: { postStart+: { tcpSocket+: { host: host } } } },
+ '#withPort':: d.fn(help='"Number or name of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."', args=[d.arg(name='port', type=d.T.any)]),
+ withPort(port): { lifecycle+: { postStart+: { tcpSocket+: { port: port } } } },
+ },
+ },
+ '#preStop':: d.obj(help="\"PreStop is called immediately before a container is terminated due to an API request or management event such as liveness/startup probe failure, preemption, resource contention, etc. The handler is not called if the container crashes or exits. The Pod's termination grace period countdown begins before the PreStop hook is executed. Regardless of the outcome of the handler, the container will eventually terminate within the Pod's termination grace period (unless delayed by finalizers). Other management of the container blocks until the hook completes or until the termination grace period is reached. More info: https://kubernetes.io/docs/concepts/containers/container-lifecycle-hooks/#container-hooks\""),
+ preStop: {
+ '#exec':: d.obj(help='"Exec specifies the action to take."'),
+ exec: {
+ '#withCommand':: d.fn(help="\"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.\"", args=[d.arg(name='command', type=d.T.array)]),
+ withCommand(command): { lifecycle+: { preStop+: { exec+: { command: if std.isArray(v=command) then command else [command] } } } },
+ '#withCommandMixin':: d.fn(help="\"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='command', type=d.T.array)]),
+ withCommandMixin(command): { lifecycle+: { preStop+: { exec+: { command+: if std.isArray(v=command) then command else [command] } } } },
+ },
+ '#httpGet':: d.obj(help='"HTTPGet specifies the http request to perform."'),
+ httpGet: {
+ '#httpHeaders':: d.obj(help='"Custom headers to set in the request. HTTP allows repeated headers."'),
+ httpHeaders: {
+ '#withName':: d.fn(help='"The header field name. This will be canonicalized upon output, so case-variant names will be understood as the same header."', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { name: name },
+ '#withValue':: d.fn(help='"The header field value"', args=[d.arg(name='value', type=d.T.string)]),
+ withValue(value): { value: value },
+ },
+ '#withHost':: d.fn(help='"Host name to connect to, defaults to the pod IP. You probably want to set \\"Host\\" in httpHeaders instead."', args=[d.arg(name='host', type=d.T.string)]),
+ withHost(host): { lifecycle+: { preStop+: { httpGet+: { host: host } } } },
+ '#withHttpHeaders':: d.fn(help='"Custom headers to set in the request. HTTP allows repeated headers."', args=[d.arg(name='httpHeaders', type=d.T.array)]),
+ withHttpHeaders(httpHeaders): { lifecycle+: { preStop+: { httpGet+: { httpHeaders: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } } },
+ '#withHttpHeadersMixin':: d.fn(help='"Custom headers to set in the request. HTTP allows repeated headers."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='httpHeaders', type=d.T.array)]),
+ withHttpHeadersMixin(httpHeaders): { lifecycle+: { preStop+: { httpGet+: { httpHeaders+: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } } },
+ '#withPath':: d.fn(help='"Path to access on the HTTP server."', args=[d.arg(name='path', type=d.T.string)]),
+ withPath(path): { lifecycle+: { preStop+: { httpGet+: { path: path } } } },
+ '#withPort':: d.fn(help='"Name or number of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."', args=[d.arg(name='port', type=d.T.any)]),
+ withPort(port): { lifecycle+: { preStop+: { httpGet+: { port: port } } } },
+ '#withScheme':: d.fn(help='"Scheme to use for connecting to the host. Defaults to HTTP."', args=[d.arg(name='scheme', type=d.T.string)]),
+ withScheme(scheme): { lifecycle+: { preStop+: { httpGet+: { scheme: scheme } } } },
+ },
+ '#tcpSocket':: d.obj(help='"Deprecated. TCPSocket is NOT supported as a LifecycleHandler and kept for the backward compatibility. There are no validation of this field and lifecycle hooks will fail in runtime when tcp handler is specified."'),
+ tcpSocket: {
+ '#withHost':: d.fn(help='"Optional: Host name to connect to, defaults to the pod IP."', args=[d.arg(name='host', type=d.T.string)]),
+ withHost(host): { lifecycle+: { preStop+: { tcpSocket+: { host: host } } } },
+ '#withPort':: d.fn(help='"Number or name of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."', args=[d.arg(name='port', type=d.T.any)]),
+ withPort(port): { lifecycle+: { preStop+: { tcpSocket+: { port: port } } } },
+ },
+ },
+ },
+ '#livenessProbe':: d.obj(help='"Periodic probe of container liveness. Container will be restarted if the probe fails. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes"'),
+ livenessProbe: {
+ '#exec':: d.obj(help='"Exec specifies the action to take."'),
+ exec: {
+ '#withCommand':: d.fn(help="\"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.\"", args=[d.arg(name='command', type=d.T.array)]),
+ withCommand(command): { livenessProbe+: { exec+: { command: if std.isArray(v=command) then command else [command] } } },
+ '#withCommandMixin':: d.fn(help="\"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='command', type=d.T.array)]),
+ withCommandMixin(command): { livenessProbe+: { exec+: { command+: if std.isArray(v=command) then command else [command] } } },
+ },
+ '#grpc':: d.obj(help='"GRPC specifies an action involving a GRPC port."'),
+ grpc: {
+ '#withPort':: d.fn(help='"Port number of the gRPC service. Number must be in the range 1 to 65535."', args=[d.arg(name='port', type=d.T.integer)]),
+ withPort(port): { livenessProbe+: { grpc+: { port: port } } },
+ '#withService':: d.fn(help='"Service is the name of the service to place in the gRPC HealthCheckRequest (see https://github.com/grpc/grpc/blob/master/doc/health-checking.md). \\n If this is not specified, the default behavior is defined by gRPC."', args=[d.arg(name='service', type=d.T.string)]),
+ withService(service): { livenessProbe+: { grpc+: { service: service } } },
+ },
+ '#httpGet':: d.obj(help='"HTTPGet specifies the http request to perform."'),
+ httpGet: {
+ '#httpHeaders':: d.obj(help='"Custom headers to set in the request. HTTP allows repeated headers."'),
+ httpHeaders: {
+ '#withName':: d.fn(help='"The header field name. This will be canonicalized upon output, so case-variant names will be understood as the same header."', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { name: name },
+ '#withValue':: d.fn(help='"The header field value"', args=[d.arg(name='value', type=d.T.string)]),
+ withValue(value): { value: value },
+ },
+ '#withHost':: d.fn(help='"Host name to connect to, defaults to the pod IP. You probably want to set \\"Host\\" in httpHeaders instead."', args=[d.arg(name='host', type=d.T.string)]),
+ withHost(host): { livenessProbe+: { httpGet+: { host: host } } },
+ '#withHttpHeaders':: d.fn(help='"Custom headers to set in the request. HTTP allows repeated headers."', args=[d.arg(name='httpHeaders', type=d.T.array)]),
+ withHttpHeaders(httpHeaders): { livenessProbe+: { httpGet+: { httpHeaders: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } },
+ '#withHttpHeadersMixin':: d.fn(help='"Custom headers to set in the request. HTTP allows repeated headers."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='httpHeaders', type=d.T.array)]),
+ withHttpHeadersMixin(httpHeaders): { livenessProbe+: { httpGet+: { httpHeaders+: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } },
+ '#withPath':: d.fn(help='"Path to access on the HTTP server."', args=[d.arg(name='path', type=d.T.string)]),
+ withPath(path): { livenessProbe+: { httpGet+: { path: path } } },
+ '#withPort':: d.fn(help='"Name or number of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."', args=[d.arg(name='port', type=d.T.any)]),
+ withPort(port): { livenessProbe+: { httpGet+: { port: port } } },
+ '#withScheme':: d.fn(help='"Scheme to use for connecting to the host. Defaults to HTTP."', args=[d.arg(name='scheme', type=d.T.string)]),
+ withScheme(scheme): { livenessProbe+: { httpGet+: { scheme: scheme } } },
+ },
+ '#tcpSocket':: d.obj(help='"TCPSocket specifies an action involving a TCP port."'),
+ tcpSocket: {
+ '#withHost':: d.fn(help='"Optional: Host name to connect to, defaults to the pod IP."', args=[d.arg(name='host', type=d.T.string)]),
+ withHost(host): { livenessProbe+: { tcpSocket+: { host: host } } },
+ '#withPort':: d.fn(help='"Number or name of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."', args=[d.arg(name='port', type=d.T.any)]),
+ withPort(port): { livenessProbe+: { tcpSocket+: { port: port } } },
+ },
+ '#withFailureThreshold':: d.fn(help='"Minimum consecutive failures for the probe to be considered failed after having succeeded. Defaults to 3. Minimum value is 1."', args=[d.arg(name='failureThreshold', type=d.T.integer)]),
+ withFailureThreshold(failureThreshold): { livenessProbe+: { failureThreshold: failureThreshold } },
+ '#withInitialDelaySeconds':: d.fn(help='"Number of seconds after the container has started before liveness probes are initiated. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes"', args=[d.arg(name='initialDelaySeconds', type=d.T.integer)]),
+ withInitialDelaySeconds(initialDelaySeconds): { livenessProbe+: { initialDelaySeconds: initialDelaySeconds } },
+ '#withPeriodSeconds':: d.fn(help='"How often (in seconds) to perform the probe. Default to 10 seconds. Minimum value is 1."', args=[d.arg(name='periodSeconds', type=d.T.integer)]),
+ withPeriodSeconds(periodSeconds): { livenessProbe+: { periodSeconds: periodSeconds } },
+ '#withSuccessThreshold':: d.fn(help='"Minimum consecutive successes for the probe to be considered successful after having failed. Defaults to 1. Must be 1 for liveness and startup. Minimum value is 1."', args=[d.arg(name='successThreshold', type=d.T.integer)]),
+ withSuccessThreshold(successThreshold): { livenessProbe+: { successThreshold: successThreshold } },
+ '#withTerminationGracePeriodSeconds':: d.fn(help="\"Optional duration in seconds the pod needs to terminate gracefully upon probe failure. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. If this value is nil, the pod's terminationGracePeriodSeconds will be used. Otherwise, this value overrides the value provided by the pod spec. Value must be non-negative integer. The value zero indicates stop immediately via the kill signal (no opportunity to shut down). This is a beta field and requires enabling ProbeTerminationGracePeriod feature gate. Minimum value is 1. spec.terminationGracePeriodSeconds is used if unset.\"", args=[d.arg(name='terminationGracePeriodSeconds', type=d.T.integer)]),
+ withTerminationGracePeriodSeconds(terminationGracePeriodSeconds): { livenessProbe+: { terminationGracePeriodSeconds: terminationGracePeriodSeconds } },
+ '#withTimeoutSeconds':: d.fn(help='"Number of seconds after which the probe times out. Defaults to 1 second. Minimum value is 1. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes"', args=[d.arg(name='timeoutSeconds', type=d.T.integer)]),
+ withTimeoutSeconds(timeoutSeconds): { livenessProbe+: { timeoutSeconds: timeoutSeconds } },
+ },
+ '#ports':: d.obj(help='"List of ports to expose from the container. Not specifying a port here DOES NOT prevent that port from being exposed. Any port which is listening on the default \\"0.0.0.0\\" address inside a container will be accessible from the network. Modifying this array with strategic merge patch may corrupt the data. For more information See https://github.com/kubernetes/kubernetes/issues/108255. Cannot be updated."'),
+ ports: {
+ '#withContainerPort':: d.fn(help="\"Number of port to expose on the pod's IP address. This must be a valid port number, 0 \u003c x \u003c 65536.\"", args=[d.arg(name='containerPort', type=d.T.integer)]),
+ withContainerPort(containerPort): { containerPort: containerPort },
+ '#withHostIP':: d.fn(help='"What host IP to bind the external port to."', args=[d.arg(name='hostIP', type=d.T.string)]),
+ withHostIP(hostIP): { hostIP: hostIP },
+ '#withHostPort':: d.fn(help='"Number of port to expose on the host. If specified, this must be a valid port number, 0 < x < 65536. If HostNetwork is specified, this must match ContainerPort. Most containers do not need this."', args=[d.arg(name='hostPort', type=d.T.integer)]),
+ withHostPort(hostPort): { hostPort: hostPort },
+ '#withName':: d.fn(help='"If specified, this must be an IANA_SVC_NAME and unique within the pod. Each named port in a pod must have a unique name. Name for the port that can be referred to by services."', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { name: name },
+ '#withProtocol':: d.fn(help='"Protocol for port. Must be UDP, TCP, or SCTP. Defaults to \\"TCP\\"."', args=[d.arg(name='protocol', type=d.T.string)]),
+ withProtocol(protocol): { protocol: protocol },
+ },
+ '#readinessProbe':: d.obj(help='"Periodic probe of container service readiness. Container will be removed from service endpoints if the probe fails. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes"'),
+ readinessProbe: {
+ '#exec':: d.obj(help='"Exec specifies the action to take."'),
+ exec: {
+ '#withCommand':: d.fn(help="\"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.\"", args=[d.arg(name='command', type=d.T.array)]),
+ withCommand(command): { readinessProbe+: { exec+: { command: if std.isArray(v=command) then command else [command] } } },
+ '#withCommandMixin':: d.fn(help="\"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='command', type=d.T.array)]),
+ withCommandMixin(command): { readinessProbe+: { exec+: { command+: if std.isArray(v=command) then command else [command] } } },
+ },
+ '#grpc':: d.obj(help='"GRPC specifies an action involving a GRPC port."'),
+ grpc: {
+ '#withPort':: d.fn(help='"Port number of the gRPC service. Number must be in the range 1 to 65535."', args=[d.arg(name='port', type=d.T.integer)]),
+ withPort(port): { readinessProbe+: { grpc+: { port: port } } },
+ '#withService':: d.fn(help='"Service is the name of the service to place in the gRPC HealthCheckRequest (see https://github.com/grpc/grpc/blob/master/doc/health-checking.md). \\n If this is not specified, the default behavior is defined by gRPC."', args=[d.arg(name='service', type=d.T.string)]),
+ withService(service): { readinessProbe+: { grpc+: { service: service } } },
+ },
+ '#httpGet':: d.obj(help='"HTTPGet specifies the http request to perform."'),
+ httpGet: {
+ '#httpHeaders':: d.obj(help='"Custom headers to set in the request. HTTP allows repeated headers."'),
+ httpHeaders: {
+ '#withName':: d.fn(help='"The header field name. This will be canonicalized upon output, so case-variant names will be understood as the same header."', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { name: name },
+ '#withValue':: d.fn(help='"The header field value"', args=[d.arg(name='value', type=d.T.string)]),
+ withValue(value): { value: value },
+ },
+ '#withHost':: d.fn(help='"Host name to connect to, defaults to the pod IP. You probably want to set \\"Host\\" in httpHeaders instead."', args=[d.arg(name='host', type=d.T.string)]),
+ withHost(host): { readinessProbe+: { httpGet+: { host: host } } },
+ '#withHttpHeaders':: d.fn(help='"Custom headers to set in the request. HTTP allows repeated headers."', args=[d.arg(name='httpHeaders', type=d.T.array)]),
+ withHttpHeaders(httpHeaders): { readinessProbe+: { httpGet+: { httpHeaders: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } },
+ '#withHttpHeadersMixin':: d.fn(help='"Custom headers to set in the request. HTTP allows repeated headers."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='httpHeaders', type=d.T.array)]),
+ withHttpHeadersMixin(httpHeaders): { readinessProbe+: { httpGet+: { httpHeaders+: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } },
+ '#withPath':: d.fn(help='"Path to access on the HTTP server."', args=[d.arg(name='path', type=d.T.string)]),
+ withPath(path): { readinessProbe+: { httpGet+: { path: path } } },
+ '#withPort':: d.fn(help='"Name or number of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."', args=[d.arg(name='port', type=d.T.any)]),
+ withPort(port): { readinessProbe+: { httpGet+: { port: port } } },
+ '#withScheme':: d.fn(help='"Scheme to use for connecting to the host. Defaults to HTTP."', args=[d.arg(name='scheme', type=d.T.string)]),
+ withScheme(scheme): { readinessProbe+: { httpGet+: { scheme: scheme } } },
+ },
+ '#tcpSocket':: d.obj(help='"TCPSocket specifies an action involving a TCP port."'),
+ tcpSocket: {
+ '#withHost':: d.fn(help='"Optional: Host name to connect to, defaults to the pod IP."', args=[d.arg(name='host', type=d.T.string)]),
+ withHost(host): { readinessProbe+: { tcpSocket+: { host: host } } },
+ '#withPort':: d.fn(help='"Number or name of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."', args=[d.arg(name='port', type=d.T.any)]),
+ withPort(port): { readinessProbe+: { tcpSocket+: { port: port } } },
+ },
+ '#withFailureThreshold':: d.fn(help='"Minimum consecutive failures for the probe to be considered failed after having succeeded. Defaults to 3. Minimum value is 1."', args=[d.arg(name='failureThreshold', type=d.T.integer)]),
+ withFailureThreshold(failureThreshold): { readinessProbe+: { failureThreshold: failureThreshold } },
+ '#withInitialDelaySeconds':: d.fn(help='"Number of seconds after the container has started before liveness probes are initiated. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes"', args=[d.arg(name='initialDelaySeconds', type=d.T.integer)]),
+ withInitialDelaySeconds(initialDelaySeconds): { readinessProbe+: { initialDelaySeconds: initialDelaySeconds } },
+ '#withPeriodSeconds':: d.fn(help='"How often (in seconds) to perform the probe. Default to 10 seconds. Minimum value is 1."', args=[d.arg(name='periodSeconds', type=d.T.integer)]),
+ withPeriodSeconds(periodSeconds): { readinessProbe+: { periodSeconds: periodSeconds } },
+ '#withSuccessThreshold':: d.fn(help='"Minimum consecutive successes for the probe to be considered successful after having failed. Defaults to 1. Must be 1 for liveness and startup. Minimum value is 1."', args=[d.arg(name='successThreshold', type=d.T.integer)]),
+ withSuccessThreshold(successThreshold): { readinessProbe+: { successThreshold: successThreshold } },
+ '#withTerminationGracePeriodSeconds':: d.fn(help="\"Optional duration in seconds the pod needs to terminate gracefully upon probe failure. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. If this value is nil, the pod's terminationGracePeriodSeconds will be used. Otherwise, this value overrides the value provided by the pod spec. Value must be non-negative integer. The value zero indicates stop immediately via the kill signal (no opportunity to shut down). This is a beta field and requires enabling ProbeTerminationGracePeriod feature gate. Minimum value is 1. spec.terminationGracePeriodSeconds is used if unset.\"", args=[d.arg(name='terminationGracePeriodSeconds', type=d.T.integer)]),
+ withTerminationGracePeriodSeconds(terminationGracePeriodSeconds): { readinessProbe+: { terminationGracePeriodSeconds: terminationGracePeriodSeconds } },
+ '#withTimeoutSeconds':: d.fn(help='"Number of seconds after which the probe times out. Defaults to 1 second. Minimum value is 1. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes"', args=[d.arg(name='timeoutSeconds', type=d.T.integer)]),
+ withTimeoutSeconds(timeoutSeconds): { readinessProbe+: { timeoutSeconds: timeoutSeconds } },
+ },
+ '#resizePolicy':: d.obj(help='"Resources resize policy for the container."'),
+ resizePolicy: {
+ '#withResourceName':: d.fn(help='"Name of the resource to which this resource resize policy applies. Supported values: cpu, memory."', args=[d.arg(name='resourceName', type=d.T.string)]),
+ withResourceName(resourceName): { resourceName: resourceName },
+ '#withRestartPolicy':: d.fn(help='"Restart policy to apply when specified resource is resized. If not specified, it defaults to NotRequired."', args=[d.arg(name='restartPolicy', type=d.T.string)]),
+ withRestartPolicy(restartPolicy): { restartPolicy: restartPolicy },
+ },
+ '#resources':: d.obj(help='"Compute Resources required by this container. Cannot be updated. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"'),
+ resources: {
+ '#claims':: d.obj(help='"Claims lists the names of resources, defined in spec.resourceClaims, that are used by this container. \\n This is an alpha field and requires enabling the DynamicResourceAllocation feature gate. \\n This field is immutable. It can only be set for containers."'),
+ claims: {
+ '#withName':: d.fn(help='"Name must match the name of one entry in pod.spec.resourceClaims of the Pod where this field is used. It makes that resource available inside a container."', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { name: name },
+ },
+ '#withClaims':: d.fn(help='"Claims lists the names of resources, defined in spec.resourceClaims, that are used by this container. \\n This is an alpha field and requires enabling the DynamicResourceAllocation feature gate. \\n This field is immutable. It can only be set for containers."', args=[d.arg(name='claims', type=d.T.array)]),
+ withClaims(claims): { resources+: { claims: if std.isArray(v=claims) then claims else [claims] } },
+ '#withClaimsMixin':: d.fn(help='"Claims lists the names of resources, defined in spec.resourceClaims, that are used by this container. \\n This is an alpha field and requires enabling the DynamicResourceAllocation feature gate. \\n This field is immutable. It can only be set for containers."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='claims', type=d.T.array)]),
+ withClaimsMixin(claims): { resources+: { claims+: if std.isArray(v=claims) then claims else [claims] } },
+ '#withLimits':: d.fn(help='"Limits describes the maximum amount of compute resources allowed. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"', args=[d.arg(name='limits', type=d.T.object)]),
+ withLimits(limits): { resources+: { limits: limits } },
+ '#withLimitsMixin':: d.fn(help='"Limits describes the maximum amount of compute resources allowed. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='limits', type=d.T.object)]),
+ withLimitsMixin(limits): { resources+: { limits+: limits } },
+ '#withRequests':: d.fn(help='"Requests describes the minimum amount of compute resources required. If Requests is omitted for a container, it defaults to Limits if that is explicitly specified, otherwise to an implementation-defined value. Requests cannot exceed Limits. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"', args=[d.arg(name='requests', type=d.T.object)]),
+ withRequests(requests): { resources+: { requests: requests } },
+ '#withRequestsMixin':: d.fn(help='"Requests describes the minimum amount of compute resources required. If Requests is omitted for a container, it defaults to Limits if that is explicitly specified, otherwise to an implementation-defined value. Requests cannot exceed Limits. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='requests', type=d.T.object)]),
+ withRequestsMixin(requests): { resources+: { requests+: requests } },
+ },
+ '#securityContext':: d.obj(help='"SecurityContext defines the security options the container should be run with. If set, the fields of SecurityContext override the equivalent fields of PodSecurityContext. More info: https://kubernetes.io/docs/tasks/configure-pod-container/security-context/"'),
+ securityContext: {
+ '#capabilities':: d.obj(help='"The capabilities to add/drop when running containers. Defaults to the default set of capabilities granted by the container runtime. Note that this field cannot be set when spec.os.name is windows."'),
+ capabilities: {
+ '#withAdd':: d.fn(help='"Added capabilities"', args=[d.arg(name='add', type=d.T.array)]),
+ withAdd(add): { securityContext+: { capabilities+: { add: if std.isArray(v=add) then add else [add] } } },
+ '#withAddMixin':: d.fn(help='"Added capabilities"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='add', type=d.T.array)]),
+ withAddMixin(add): { securityContext+: { capabilities+: { add+: if std.isArray(v=add) then add else [add] } } },
+ '#withDrop':: d.fn(help='"Removed capabilities"', args=[d.arg(name='drop', type=d.T.array)]),
+ withDrop(drop): { securityContext+: { capabilities+: { drop: if std.isArray(v=drop) then drop else [drop] } } },
+ '#withDropMixin':: d.fn(help='"Removed capabilities"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='drop', type=d.T.array)]),
+ withDropMixin(drop): { securityContext+: { capabilities+: { drop+: if std.isArray(v=drop) then drop else [drop] } } },
+ },
+ '#seLinuxOptions':: d.obj(help='"The SELinux context to be applied to the container. If unspecified, the container runtime will allocate a random SELinux context for each container. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. Note that this field cannot be set when spec.os.name is windows."'),
+ seLinuxOptions: {
+ '#withLevel':: d.fn(help='"Level is SELinux level label that applies to the container."', args=[d.arg(name='level', type=d.T.string)]),
+ withLevel(level): { securityContext+: { seLinuxOptions+: { level: level } } },
+ '#withRole':: d.fn(help='"Role is a SELinux role label that applies to the container."', args=[d.arg(name='role', type=d.T.string)]),
+ withRole(role): { securityContext+: { seLinuxOptions+: { role: role } } },
+ '#withType':: d.fn(help='"Type is a SELinux type label that applies to the container."', args=[d.arg(name='type', type=d.T.string)]),
+ withType(type): { securityContext+: { seLinuxOptions+: { type: type } } },
+ '#withUser':: d.fn(help='"User is a SELinux user label that applies to the container."', args=[d.arg(name='user', type=d.T.string)]),
+ withUser(user): { securityContext+: { seLinuxOptions+: { user: user } } },
+ },
+ '#seccompProfile':: d.obj(help='"The seccomp options to use by this container. If seccomp options are provided at both the pod & container level, the container options override the pod options. Note that this field cannot be set when spec.os.name is windows."'),
+ seccompProfile: {
+ '#withLocalhostProfile':: d.fn(help="\"localhostProfile indicates a profile defined in a file on the node should be used. The profile must be preconfigured on the node to work. Must be a descending path, relative to the kubelet's configured seccomp profile location. Must be set if type is \\\"Localhost\\\". Must NOT be set for any other type.\"", args=[d.arg(name='localhostProfile', type=d.T.string)]),
+ withLocalhostProfile(localhostProfile): { securityContext+: { seccompProfile+: { localhostProfile: localhostProfile } } },
+ '#withType':: d.fn(help='"type indicates which kind of seccomp profile will be applied. Valid options are: \\n Localhost - a profile defined in a file on the node should be used. RuntimeDefault - the container runtime default profile should be used. Unconfined - no profile should be applied."', args=[d.arg(name='type', type=d.T.string)]),
+ withType(type): { securityContext+: { seccompProfile+: { type: type } } },
+ },
+ '#windowsOptions':: d.obj(help='"The Windows specific settings applied to all containers. If unspecified, the options from the PodSecurityContext will be used. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. Note that this field cannot be set when spec.os.name is linux."'),
+ windowsOptions: {
+ '#withGmsaCredentialSpec':: d.fn(help='"GMSACredentialSpec is where the GMSA admission webhook (https://github.com/kubernetes-sigs/windows-gmsa) inlines the contents of the GMSA credential spec named by the GMSACredentialSpecName field."', args=[d.arg(name='gmsaCredentialSpec', type=d.T.string)]),
+ withGmsaCredentialSpec(gmsaCredentialSpec): { securityContext+: { windowsOptions+: { gmsaCredentialSpec: gmsaCredentialSpec } } },
+ '#withGmsaCredentialSpecName':: d.fn(help='"GMSACredentialSpecName is the name of the GMSA credential spec to use."', args=[d.arg(name='gmsaCredentialSpecName', type=d.T.string)]),
+ withGmsaCredentialSpecName(gmsaCredentialSpecName): { securityContext+: { windowsOptions+: { gmsaCredentialSpecName: gmsaCredentialSpecName } } },
+ '#withHostProcess':: d.fn(help="\"HostProcess determines if a container should be run as a 'Host Process' container. All of a Pod's containers must have the same effective HostProcess value (it is not allowed to have a mix of HostProcess containers and non-HostProcess containers). In addition, if HostProcess is true then HostNetwork must also be set to true.\"", args=[d.arg(name='hostProcess', type=d.T.boolean)]),
+ withHostProcess(hostProcess): { securityContext+: { windowsOptions+: { hostProcess: hostProcess } } },
+ '#withRunAsUserName':: d.fn(help='"The UserName in Windows to run the entrypoint of the container process. Defaults to the user specified in image metadata if unspecified. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence."', args=[d.arg(name='runAsUserName', type=d.T.string)]),
+ withRunAsUserName(runAsUserName): { securityContext+: { windowsOptions+: { runAsUserName: runAsUserName } } },
+ },
+ '#withAllowPrivilegeEscalation':: d.fn(help='"AllowPrivilegeEscalation controls whether a process can gain more privileges than its parent process. This bool directly controls if the no_new_privs flag will be set on the container process. AllowPrivilegeEscalation is true always when the container is: 1) run as Privileged 2) has CAP_SYS_ADMIN Note that this field cannot be set when spec.os.name is windows."', args=[d.arg(name='allowPrivilegeEscalation', type=d.T.boolean)]),
+ withAllowPrivilegeEscalation(allowPrivilegeEscalation): { securityContext+: { allowPrivilegeEscalation: allowPrivilegeEscalation } },
+ '#withPrivileged':: d.fn(help='"Run container in privileged mode. Processes in privileged containers are essentially equivalent to root on the host. Defaults to false. Note that this field cannot be set when spec.os.name is windows."', args=[d.arg(name='privileged', type=d.T.boolean)]),
+ withPrivileged(privileged): { securityContext+: { privileged: privileged } },
+ '#withProcMount':: d.fn(help='"procMount denotes the type of proc mount to use for the containers. The default is DefaultProcMount which uses the container runtime defaults for readonly paths and masked paths. This requires the ProcMountType feature flag to be enabled. Note that this field cannot be set when spec.os.name is windows."', args=[d.arg(name='procMount', type=d.T.string)]),
+ withProcMount(procMount): { securityContext+: { procMount: procMount } },
+ '#withReadOnlyRootFilesystem':: d.fn(help='"Whether this container has a read-only root filesystem. Default is false. Note that this field cannot be set when spec.os.name is windows."', args=[d.arg(name='readOnlyRootFilesystem', type=d.T.boolean)]),
+ withReadOnlyRootFilesystem(readOnlyRootFilesystem): { securityContext+: { readOnlyRootFilesystem: readOnlyRootFilesystem } },
+ '#withRunAsGroup':: d.fn(help='"The GID to run the entrypoint of the container process. Uses runtime default if unset. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. Note that this field cannot be set when spec.os.name is windows."', args=[d.arg(name='runAsGroup', type=d.T.integer)]),
+ withRunAsGroup(runAsGroup): { securityContext+: { runAsGroup: runAsGroup } },
+ '#withRunAsNonRoot':: d.fn(help='"Indicates that the container must run as a non-root user. If true, the Kubelet will validate the image at runtime to ensure that it does not run as UID 0 (root) and fail to start the container if it does. If unset or false, no such validation will be performed. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence."', args=[d.arg(name='runAsNonRoot', type=d.T.boolean)]),
+ withRunAsNonRoot(runAsNonRoot): { securityContext+: { runAsNonRoot: runAsNonRoot } },
+ '#withRunAsUser':: d.fn(help='"The UID to run the entrypoint of the container process. Defaults to user specified in image metadata if unspecified. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. Note that this field cannot be set when spec.os.name is windows."', args=[d.arg(name='runAsUser', type=d.T.integer)]),
+ withRunAsUser(runAsUser): { securityContext+: { runAsUser: runAsUser } },
+ },
+ '#startupProbe':: d.obj(help="\"StartupProbe indicates that the Pod has successfully initialized. If specified, no other probes are executed until this completes successfully. If this probe fails, the Pod will be restarted, just as if the livenessProbe failed. This can be used to provide different probe parameters at the beginning of a Pod's lifecycle, when it might take a long time to load data or warm a cache, than during steady-state operation. This cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes\""),
+ startupProbe: {
+ '#exec':: d.obj(help='"Exec specifies the action to take."'),
+ exec: {
+ '#withCommand':: d.fn(help="\"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.\"", args=[d.arg(name='command', type=d.T.array)]),
+ withCommand(command): { startupProbe+: { exec+: { command: if std.isArray(v=command) then command else [command] } } },
+ '#withCommandMixin':: d.fn(help="\"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='command', type=d.T.array)]),
+ withCommandMixin(command): { startupProbe+: { exec+: { command+: if std.isArray(v=command) then command else [command] } } },
+ },
+ '#grpc':: d.obj(help='"GRPC specifies an action involving a GRPC port."'),
+ grpc: {
+ '#withPort':: d.fn(help='"Port number of the gRPC service. Number must be in the range 1 to 65535."', args=[d.arg(name='port', type=d.T.integer)]),
+ withPort(port): { startupProbe+: { grpc+: { port: port } } },
+ '#withService':: d.fn(help='"Service is the name of the service to place in the gRPC HealthCheckRequest (see https://github.com/grpc/grpc/blob/master/doc/health-checking.md). \\n If this is not specified, the default behavior is defined by gRPC."', args=[d.arg(name='service', type=d.T.string)]),
+ withService(service): { startupProbe+: { grpc+: { service: service } } },
+ },
+ '#httpGet':: d.obj(help='"HTTPGet specifies the http request to perform."'),
+ httpGet: {
+ '#httpHeaders':: d.obj(help='"Custom headers to set in the request. HTTP allows repeated headers."'),
+ httpHeaders: {
+ '#withName':: d.fn(help='"The header field name. This will be canonicalized upon output, so case-variant names will be understood as the same header."', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { name: name },
+ '#withValue':: d.fn(help='"The header field value"', args=[d.arg(name='value', type=d.T.string)]),
+ withValue(value): { value: value },
+ },
+ '#withHost':: d.fn(help='"Host name to connect to, defaults to the pod IP. You probably want to set \\"Host\\" in httpHeaders instead."', args=[d.arg(name='host', type=d.T.string)]),
+ withHost(host): { startupProbe+: { httpGet+: { host: host } } },
+ '#withHttpHeaders':: d.fn(help='"Custom headers to set in the request. HTTP allows repeated headers."', args=[d.arg(name='httpHeaders', type=d.T.array)]),
+ withHttpHeaders(httpHeaders): { startupProbe+: { httpGet+: { httpHeaders: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } },
+ '#withHttpHeadersMixin':: d.fn(help='"Custom headers to set in the request. HTTP allows repeated headers."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='httpHeaders', type=d.T.array)]),
+ withHttpHeadersMixin(httpHeaders): { startupProbe+: { httpGet+: { httpHeaders+: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } },
+ '#withPath':: d.fn(help='"Path to access on the HTTP server."', args=[d.arg(name='path', type=d.T.string)]),
+ withPath(path): { startupProbe+: { httpGet+: { path: path } } },
+ '#withPort':: d.fn(help='"Name or number of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."', args=[d.arg(name='port', type=d.T.any)]),
+ withPort(port): { startupProbe+: { httpGet+: { port: port } } },
+ '#withScheme':: d.fn(help='"Scheme to use for connecting to the host. Defaults to HTTP."', args=[d.arg(name='scheme', type=d.T.string)]),
+ withScheme(scheme): { startupProbe+: { httpGet+: { scheme: scheme } } },
+ },
+ '#tcpSocket':: d.obj(help='"TCPSocket specifies an action involving a TCP port."'),
+ tcpSocket: {
+ '#withHost':: d.fn(help='"Optional: Host name to connect to, defaults to the pod IP."', args=[d.arg(name='host', type=d.T.string)]),
+ withHost(host): { startupProbe+: { tcpSocket+: { host: host } } },
+ '#withPort':: d.fn(help='"Number or name of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."', args=[d.arg(name='port', type=d.T.any)]),
+ withPort(port): { startupProbe+: { tcpSocket+: { port: port } } },
+ },
+ '#withFailureThreshold':: d.fn(help='"Minimum consecutive failures for the probe to be considered failed after having succeeded. Defaults to 3. Minimum value is 1."', args=[d.arg(name='failureThreshold', type=d.T.integer)]),
+ withFailureThreshold(failureThreshold): { startupProbe+: { failureThreshold: failureThreshold } },
+ '#withInitialDelaySeconds':: d.fn(help='"Number of seconds after the container has started before liveness probes are initiated. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes"', args=[d.arg(name='initialDelaySeconds', type=d.T.integer)]),
+ withInitialDelaySeconds(initialDelaySeconds): { startupProbe+: { initialDelaySeconds: initialDelaySeconds } },
+ '#withPeriodSeconds':: d.fn(help='"How often (in seconds) to perform the probe. Default to 10 seconds. Minimum value is 1."', args=[d.arg(name='periodSeconds', type=d.T.integer)]),
+ withPeriodSeconds(periodSeconds): { startupProbe+: { periodSeconds: periodSeconds } },
+ '#withSuccessThreshold':: d.fn(help='"Minimum consecutive successes for the probe to be considered successful after having failed. Defaults to 1. Must be 1 for liveness and startup. Minimum value is 1."', args=[d.arg(name='successThreshold', type=d.T.integer)]),
+ withSuccessThreshold(successThreshold): { startupProbe+: { successThreshold: successThreshold } },
+ '#withTerminationGracePeriodSeconds':: d.fn(help="\"Optional duration in seconds the pod needs to terminate gracefully upon probe failure. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. If this value is nil, the pod's terminationGracePeriodSeconds will be used. Otherwise, this value overrides the value provided by the pod spec. Value must be non-negative integer. The value zero indicates stop immediately via the kill signal (no opportunity to shut down). This is a beta field and requires enabling ProbeTerminationGracePeriod feature gate. Minimum value is 1. spec.terminationGracePeriodSeconds is used if unset.\"", args=[d.arg(name='terminationGracePeriodSeconds', type=d.T.integer)]),
+ withTerminationGracePeriodSeconds(terminationGracePeriodSeconds): { startupProbe+: { terminationGracePeriodSeconds: terminationGracePeriodSeconds } },
+ '#withTimeoutSeconds':: d.fn(help='"Number of seconds after which the probe times out. Defaults to 1 second. Minimum value is 1. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes"', args=[d.arg(name='timeoutSeconds', type=d.T.integer)]),
+ withTimeoutSeconds(timeoutSeconds): { startupProbe+: { timeoutSeconds: timeoutSeconds } },
+ },
+ '#volumeDevices':: d.obj(help='"volumeDevices is the list of block devices to be used by the container."'),
+ volumeDevices: {
+ '#withDevicePath':: d.fn(help='"devicePath is the path inside of the container that the device will be mapped to."', args=[d.arg(name='devicePath', type=d.T.string)]),
+ withDevicePath(devicePath): { devicePath: devicePath },
+ '#withName':: d.fn(help='"name must match the name of a persistentVolumeClaim in the pod"', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { name: name },
+ },
+ '#volumeMounts':: d.obj(help="\"Pod volumes to mount into the container's filesystem. Cannot be updated.\""),
+ volumeMounts: {
+ '#withMountPath':: d.fn(help="\"Path within the container at which the volume should be mounted. Must not contain ':'.\"", args=[d.arg(name='mountPath', type=d.T.string)]),
+ withMountPath(mountPath): { mountPath: mountPath },
+ '#withMountPropagation':: d.fn(help='"mountPropagation determines how mounts are propagated from the host to container and the other way around. When not set, MountPropagationNone is used. This field is beta in 1.10."', args=[d.arg(name='mountPropagation', type=d.T.string)]),
+ withMountPropagation(mountPropagation): { mountPropagation: mountPropagation },
+ '#withName':: d.fn(help='"This must match the Name of a Volume."', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { name: name },
+ '#withReadOnly':: d.fn(help='"Mounted read-only if true, read-write otherwise (false or unspecified). Defaults to false."', args=[d.arg(name='readOnly', type=d.T.boolean)]),
+ withReadOnly(readOnly): { readOnly: readOnly },
+ '#withSubPath':: d.fn(help="\"Path within the volume from which the container's volume should be mounted. Defaults to \\\"\\\" (volume's root).\"", args=[d.arg(name='subPath', type=d.T.string)]),
+ withSubPath(subPath): { subPath: subPath },
+ '#withSubPathExpr':: d.fn(help="\"Expanded path within the volume from which the container's volume should be mounted. Behaves similarly to SubPath but environment variable references $(VAR_NAME) are expanded using the container's environment. Defaults to \\\"\\\" (volume's root). SubPathExpr and SubPath are mutually exclusive.\"", args=[d.arg(name='subPathExpr', type=d.T.string)]),
+ withSubPathExpr(subPathExpr): { subPathExpr: subPathExpr },
+ },
+ '#withArgs':: d.fn(help="\"Arguments to the entrypoint. The container image's CMD is used if this is not provided. Variable references $(VAR_NAME) are expanded using the container's environment. If a variable cannot be resolved, the reference in the input string will be unchanged. Double $$ are reduced to a single $, which allows for escaping the $(VAR_NAME) syntax: i.e. \\\"$$(VAR_NAME)\\\" will produce the string literal \\\"$(VAR_NAME)\\\". Escaped references will never be expanded, regardless of whether the variable exists or not. Cannot be updated. More info: https://kubernetes.io/docs/tasks/inject-data-application/define-command-argument-container/#running-a-command-in-a-shell\"", args=[d.arg(name='args', type=d.T.array)]),
+ withArgs(args): { args: if std.isArray(v=args) then args else [args] },
+ '#withArgsMixin':: d.fn(help="\"Arguments to the entrypoint. The container image's CMD is used if this is not provided. Variable references $(VAR_NAME) are expanded using the container's environment. If a variable cannot be resolved, the reference in the input string will be unchanged. Double $$ are reduced to a single $, which allows for escaping the $(VAR_NAME) syntax: i.e. \\\"$$(VAR_NAME)\\\" will produce the string literal \\\"$(VAR_NAME)\\\". Escaped references will never be expanded, regardless of whether the variable exists or not. Cannot be updated. More info: https://kubernetes.io/docs/tasks/inject-data-application/define-command-argument-container/#running-a-command-in-a-shell\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='args', type=d.T.array)]),
+ withArgsMixin(args): { args+: if std.isArray(v=args) then args else [args] },
+ '#withCommand':: d.fn(help="\"Entrypoint array. Not executed within a shell. The container image's ENTRYPOINT is used if this is not provided. Variable references $(VAR_NAME) are expanded using the container's environment. If a variable cannot be resolved, the reference in the input string will be unchanged. Double $$ are reduced to a single $, which allows for escaping the $(VAR_NAME) syntax: i.e. \\\"$$(VAR_NAME)\\\" will produce the string literal \\\"$(VAR_NAME)\\\". Escaped references will never be expanded, regardless of whether the variable exists or not. Cannot be updated. More info: https://kubernetes.io/docs/tasks/inject-data-application/define-command-argument-container/#running-a-command-in-a-shell\"", args=[d.arg(name='command', type=d.T.array)]),
+ withCommand(command): { command: if std.isArray(v=command) then command else [command] },
+ '#withCommandMixin':: d.fn(help="\"Entrypoint array. Not executed within a shell. The container image's ENTRYPOINT is used if this is not provided. Variable references $(VAR_NAME) are expanded using the container's environment. If a variable cannot be resolved, the reference in the input string will be unchanged. Double $$ are reduced to a single $, which allows for escaping the $(VAR_NAME) syntax: i.e. \\\"$$(VAR_NAME)\\\" will produce the string literal \\\"$(VAR_NAME)\\\". Escaped references will never be expanded, regardless of whether the variable exists or not. Cannot be updated. More info: https://kubernetes.io/docs/tasks/inject-data-application/define-command-argument-container/#running-a-command-in-a-shell\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='command', type=d.T.array)]),
+ withCommandMixin(command): { command+: if std.isArray(v=command) then command else [command] },
+ '#withEnv':: d.fn(help='"List of environment variables to set in the container. Cannot be updated."', args=[d.arg(name='env', type=d.T.array)]),
+ withEnv(env): { env: if std.isArray(v=env) then env else [env] },
+ '#withEnvFrom':: d.fn(help='"List of sources to populate environment variables in the container. The keys defined within a source must be a C_IDENTIFIER. All invalid keys will be reported as an event when the container is starting. When a key exists in multiple sources, the value associated with the last source will take precedence. Values defined by an Env with a duplicate key will take precedence. Cannot be updated."', args=[d.arg(name='envFrom', type=d.T.array)]),
+ withEnvFrom(envFrom): { envFrom: if std.isArray(v=envFrom) then envFrom else [envFrom] },
+ '#withEnvFromMixin':: d.fn(help='"List of sources to populate environment variables in the container. The keys defined within a source must be a C_IDENTIFIER. All invalid keys will be reported as an event when the container is starting. When a key exists in multiple sources, the value associated with the last source will take precedence. Values defined by an Env with a duplicate key will take precedence. Cannot be updated."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='envFrom', type=d.T.array)]),
+ withEnvFromMixin(envFrom): { envFrom+: if std.isArray(v=envFrom) then envFrom else [envFrom] },
+ '#withEnvMixin':: d.fn(help='"List of environment variables to set in the container. Cannot be updated."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='env', type=d.T.array)]),
+ withEnvMixin(env): { env+: if std.isArray(v=env) then env else [env] },
+ '#withImage':: d.fn(help='"Container image name. More info: https://kubernetes.io/docs/concepts/containers/images This field is optional to allow higher level config management to default or override container images in workload controllers like Deployments and StatefulSets."', args=[d.arg(name='image', type=d.T.string)]),
+ withImage(image): { image: image },
+ '#withImagePullPolicy':: d.fn(help='"Image pull policy. One of Always, Never, IfNotPresent. Defaults to Always if :latest tag is specified, or IfNotPresent otherwise. Cannot be updated. More info: https://kubernetes.io/docs/concepts/containers/images#updating-images"', args=[d.arg(name='imagePullPolicy', type=d.T.string)]),
+ withImagePullPolicy(imagePullPolicy): { imagePullPolicy: imagePullPolicy },
+ '#withName':: d.fn(help='"Name of the container specified as a DNS_LABEL. Each container in a pod must have a unique name (DNS_LABEL). Cannot be updated."', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { name: name },
+ '#withPorts':: d.fn(help='"List of ports to expose from the container. Not specifying a port here DOES NOT prevent that port from being exposed. Any port which is listening on the default \\"0.0.0.0\\" address inside a container will be accessible from the network. Modifying this array with strategic merge patch may corrupt the data. For more information See https://github.com/kubernetes/kubernetes/issues/108255. Cannot be updated."', args=[d.arg(name='ports', type=d.T.array)]),
+ withPorts(ports): { ports: if std.isArray(v=ports) then ports else [ports] },
+ '#withPortsMixin':: d.fn(help='"List of ports to expose from the container. Not specifying a port here DOES NOT prevent that port from being exposed. Any port which is listening on the default \\"0.0.0.0\\" address inside a container will be accessible from the network. Modifying this array with strategic merge patch may corrupt the data. For more information See https://github.com/kubernetes/kubernetes/issues/108255. Cannot be updated."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ports', type=d.T.array)]),
+ withPortsMixin(ports): { ports+: if std.isArray(v=ports) then ports else [ports] },
+ '#withResizePolicy':: d.fn(help='"Resources resize policy for the container."', args=[d.arg(name='resizePolicy', type=d.T.array)]),
+ withResizePolicy(resizePolicy): { resizePolicy: if std.isArray(v=resizePolicy) then resizePolicy else [resizePolicy] },
+ '#withResizePolicyMixin':: d.fn(help='"Resources resize policy for the container."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='resizePolicy', type=d.T.array)]),
+ withResizePolicyMixin(resizePolicy): { resizePolicy+: if std.isArray(v=resizePolicy) then resizePolicy else [resizePolicy] },
+ '#withRestartPolicy':: d.fn(help="\"RestartPolicy defines the restart behavior of individual containers in a pod. This field may only be set for init containers, and the only allowed value is \\\"Always\\\". For non-init containers or when this field is not specified, the restart behavior is defined by the Pod's restart policy and the container type. Setting the RestartPolicy as \\\"Always\\\" for the init container will have the following effect: this init container will be continually restarted on exit until all regular containers have terminated. Once all regular containers have completed, all init containers with restartPolicy \\\"Always\\\" will be shut down. This lifecycle differs from normal init containers and is often referred to as a \\\"sidecar\\\" container. Although this init container still starts in the init container sequence, it does not wait for the container to complete before proceeding to the next init container. Instead, the next init container starts immediately after this init container is started, or after any startupProbe has successfully completed.\"", args=[d.arg(name='restartPolicy', type=d.T.string)]),
+ withRestartPolicy(restartPolicy): { restartPolicy: restartPolicy },
+ '#withStdin':: d.fn(help='"Whether this container should allocate a buffer for stdin in the container runtime. If this is not set, reads from stdin in the container will always result in EOF. Default is false."', args=[d.arg(name='stdin', type=d.T.boolean)]),
+ withStdin(stdin): { stdin: stdin },
+ '#withStdinOnce':: d.fn(help='"Whether the container runtime should close the stdin channel after it has been opened by a single attach. When stdin is true the stdin stream will remain open across multiple attach sessions. If stdinOnce is set to true, stdin is opened on container start, is empty until the first client attaches to stdin, and then remains open and accepts data until the client disconnects, at which time stdin is closed and remains closed until the container is restarted. If this flag is false, a container processes that reads from stdin will never receive an EOF. Default is false"', args=[d.arg(name='stdinOnce', type=d.T.boolean)]),
+ withStdinOnce(stdinOnce): { stdinOnce: stdinOnce },
+ '#withTerminationMessagePath':: d.fn(help="\"Optional: Path at which the file to which the container's termination message will be written is mounted into the container's filesystem. Message written is intended to be brief final status, such as an assertion failure message. Will be truncated by the node if greater than 4096 bytes. The total message length across all containers will be limited to 12kb. Defaults to /dev/termination-log. Cannot be updated.\"", args=[d.arg(name='terminationMessagePath', type=d.T.string)]),
+ withTerminationMessagePath(terminationMessagePath): { terminationMessagePath: terminationMessagePath },
+ '#withTerminationMessagePolicy':: d.fn(help='"Indicate how the termination message should be populated. File will use the contents of terminationMessagePath to populate the container status message on both success and failure. FallbackToLogsOnError will use the last chunk of container log output if the termination message file is empty and the container exited with an error. The log output is limited to 2048 bytes or 80 lines, whichever is smaller. Defaults to File. Cannot be updated."', args=[d.arg(name='terminationMessagePolicy', type=d.T.string)]),
+ withTerminationMessagePolicy(terminationMessagePolicy): { terminationMessagePolicy: terminationMessagePolicy },
+ '#withTty':: d.fn(help="\"Whether this container should allocate a TTY for itself, also requires 'stdin' to be true. Default is false.\"", args=[d.arg(name='tty', type=d.T.boolean)]),
+ withTty(tty): { tty: tty },
+ '#withVolumeDevices':: d.fn(help='"volumeDevices is the list of block devices to be used by the container."', args=[d.arg(name='volumeDevices', type=d.T.array)]),
+ withVolumeDevices(volumeDevices): { volumeDevices: if std.isArray(v=volumeDevices) then volumeDevices else [volumeDevices] },
+ '#withVolumeDevicesMixin':: d.fn(help='"volumeDevices is the list of block devices to be used by the container."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='volumeDevices', type=d.T.array)]),
+ withVolumeDevicesMixin(volumeDevices): { volumeDevices+: if std.isArray(v=volumeDevices) then volumeDevices else [volumeDevices] },
+ '#withVolumeMounts':: d.fn(help="\"Pod volumes to mount into the container's filesystem. Cannot be updated.\"", args=[d.arg(name='volumeMounts', type=d.T.array)]),
+ withVolumeMounts(volumeMounts): { volumeMounts: if std.isArray(v=volumeMounts) then volumeMounts else [volumeMounts] },
+ '#withVolumeMountsMixin':: d.fn(help="\"Pod volumes to mount into the container's filesystem. Cannot be updated.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='volumeMounts', type=d.T.array)]),
+ withVolumeMountsMixin(volumeMounts): { volumeMounts+: if std.isArray(v=volumeMounts) then volumeMounts else [volumeMounts] },
+ '#withWorkingDir':: d.fn(help="\"Container's working directory. If not specified, the container runtime's default will be used, which might be configured in the container image. Cannot be updated.\"", args=[d.arg(name='workingDir', type=d.T.string)]),
+ withWorkingDir(workingDir): { workingDir: workingDir },
+ },
+ '#dnsConfig':: d.obj(help='"Specifies the DNS parameters of a pod. Parameters specified here will be merged to the generated DNS configuration based on DNSPolicy."'),
+ dnsConfig: {
+ '#options':: d.obj(help='"A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy."'),
+ options: {
+ '#withName':: d.fn(help='"Required."', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { name: name },
+ '#withValue':: d.fn(help='', args=[d.arg(name='value', type=d.T.string)]),
+ withValue(value): { value: value },
+ },
+ '#withNameservers':: d.fn(help='"A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed."', args=[d.arg(name='nameservers', type=d.T.array)]),
+ withNameservers(nameservers): { spec+: { template+: { spec+: { dnsConfig+: { nameservers: if std.isArray(v=nameservers) then nameservers else [nameservers] } } } } },
+ '#withNameserversMixin':: d.fn(help='"A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='nameservers', type=d.T.array)]),
+ withNameserversMixin(nameservers): { spec+: { template+: { spec+: { dnsConfig+: { nameservers+: if std.isArray(v=nameservers) then nameservers else [nameservers] } } } } },
+ '#withOptions':: d.fn(help='"A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy."', args=[d.arg(name='options', type=d.T.array)]),
+ withOptions(options): { spec+: { template+: { spec+: { dnsConfig+: { options: if std.isArray(v=options) then options else [options] } } } } },
+ '#withOptionsMixin':: d.fn(help='"A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='options', type=d.T.array)]),
+ withOptionsMixin(options): { spec+: { template+: { spec+: { dnsConfig+: { options+: if std.isArray(v=options) then options else [options] } } } } },
+ '#withSearches':: d.fn(help='"A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed."', args=[d.arg(name='searches', type=d.T.array)]),
+ withSearches(searches): { spec+: { template+: { spec+: { dnsConfig+: { searches: if std.isArray(v=searches) then searches else [searches] } } } } },
+ '#withSearchesMixin':: d.fn(help='"A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='searches', type=d.T.array)]),
+ withSearchesMixin(searches): { spec+: { template+: { spec+: { dnsConfig+: { searches+: if std.isArray(v=searches) then searches else [searches] } } } } },
+ },
+ '#ephemeralContainers':: d.obj(help="\"List of ephemeral containers run in this pod. Ephemeral containers may be run in an existing pod to perform user-initiated actions such as debugging. This list cannot be specified when creating a pod, and it cannot be modified by updating the pod spec. In order to add an ephemeral container to an existing pod, use the pod's ephemeralcontainers subresource.\""),
+ ephemeralContainers: {
+ '#env':: d.obj(help='"List of environment variables to set in the container. Cannot be updated."'),
+ env: {
+ '#valueFrom':: d.obj(help="\"Source for the environment variable's value. Cannot be used if value is not empty.\""),
+ valueFrom: {
+ '#configMapKeyRef':: d.obj(help='"Selects a key of a ConfigMap."'),
+ configMapKeyRef: {
+ '#withKey':: d.fn(help='"The key to select."', args=[d.arg(name='key', type=d.T.string)]),
+ withKey(key): { valueFrom+: { configMapKeyRef+: { key: key } } },
+ '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { valueFrom+: { configMapKeyRef+: { name: name } } },
+ '#withOptional':: d.fn(help='"Specify whether the ConfigMap or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]),
+ withOptional(optional): { valueFrom+: { configMapKeyRef+: { optional: optional } } },
+ },
+ '#fieldRef':: d.obj(help="\"Selects a field of the pod: supports metadata.name, metadata.namespace, `metadata.labels['\u003cKEY\u003e']`, `metadata.annotations['\u003cKEY\u003e']`, spec.nodeName, spec.serviceAccountName, status.hostIP, status.podIP, status.podIPs.\""),
+ fieldRef: {
+ '#withApiVersion':: d.fn(help='"Version of the schema the FieldPath is written in terms of, defaults to \\"v1\\"."', args=[d.arg(name='apiVersion', type=d.T.string)]),
+ withApiVersion(apiVersion): { valueFrom+: { fieldRef+: { apiVersion: apiVersion } } },
+ '#withFieldPath':: d.fn(help='"Path of the field to select in the specified API version."', args=[d.arg(name='fieldPath', type=d.T.string)]),
+ withFieldPath(fieldPath): { valueFrom+: { fieldRef+: { fieldPath: fieldPath } } },
+ },
+ '#resourceFieldRef':: d.obj(help='"Selects a resource of the container: only resources limits and requests (limits.cpu, limits.memory, limits.ephemeral-storage, requests.cpu, requests.memory and requests.ephemeral-storage) are currently supported."'),
+ resourceFieldRef: {
+ '#withContainerName':: d.fn(help='"Container name: required for volumes, optional for env vars"', args=[d.arg(name='containerName', type=d.T.string)]),
+ withContainerName(containerName): { valueFrom+: { resourceFieldRef+: { containerName: containerName } } },
+ '#withDivisor':: d.fn(help='"Specifies the output format of the exposed resources, defaults to \\"1\\', args=[d.arg(name='divisor', type=d.T.any)]),
+ withDivisor(divisor): { valueFrom+: { resourceFieldRef+: { divisor: divisor } } },
+ '#withResource':: d.fn(help='"Required: resource to select"', args=[d.arg(name='resource', type=d.T.string)]),
+ withResource(resource): { valueFrom+: { resourceFieldRef+: { resource: resource } } },
+ },
+ '#secretKeyRef':: d.obj(help="\"Selects a key of a secret in the pod's namespace\""),
+ secretKeyRef: {
+ '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]),
+ withKey(key): { valueFrom+: { secretKeyRef+: { key: key } } },
+ '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { valueFrom+: { secretKeyRef+: { name: name } } },
+ '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]),
+ withOptional(optional): { valueFrom+: { secretKeyRef+: { optional: optional } } },
+ },
+ },
+ '#withName':: d.fn(help='"Name of the environment variable. Must be a C_IDENTIFIER."', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { name: name },
+ '#withValue':: d.fn(help='"Variable references $(VAR_NAME) are expanded using the previously defined environment variables in the container and any service environment variables. If a variable cannot be resolved, the reference in the input string will be unchanged. Double $$ are reduced to a single $, which allows for escaping the $(VAR_NAME) syntax: i.e. \\"$$(VAR_NAME)\\" will produce the string literal \\"$(VAR_NAME)\\". Escaped references will never be expanded, regardless of whether the variable exists or not. Defaults to \\"\\"."', args=[d.arg(name='value', type=d.T.string)]),
+ withValue(value): { value: value },
+ },
+ '#envFrom':: d.obj(help='"List of sources to populate environment variables in the container. The keys defined within a source must be a C_IDENTIFIER. All invalid keys will be reported as an event when the container is starting. When a key exists in multiple sources, the value associated with the last source will take precedence. Values defined by an Env with a duplicate key will take precedence. Cannot be updated."'),
+ envFrom: {
+ '#configMapRef':: d.obj(help='"The ConfigMap to select from"'),
+ configMapRef: {
+ '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { configMapRef+: { name: name } },
+ '#withOptional':: d.fn(help='"Specify whether the ConfigMap must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]),
+ withOptional(optional): { configMapRef+: { optional: optional } },
+ },
+ '#secretRef':: d.obj(help='"The Secret to select from"'),
+ secretRef: {
+ '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { secretRef+: { name: name } },
+ '#withOptional':: d.fn(help='"Specify whether the Secret must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]),
+ withOptional(optional): { secretRef+: { optional: optional } },
+ },
+ '#withPrefix':: d.fn(help='"An optional identifier to prepend to each key in the ConfigMap. Must be a C_IDENTIFIER."', args=[d.arg(name='prefix', type=d.T.string)]),
+ withPrefix(prefix): { prefix: prefix },
+ },
+ '#lifecycle':: d.obj(help='"Lifecycle is not allowed for ephemeral containers."'),
+ lifecycle: {
+ '#postStart':: d.obj(help='"PostStart is called immediately after a container is created. If the handler fails, the container is terminated and restarted according to its restart policy. Other management of the container blocks until the hook completes. More info: https://kubernetes.io/docs/concepts/containers/container-lifecycle-hooks/#container-hooks"'),
+ postStart: {
+ '#exec':: d.obj(help='"Exec specifies the action to take."'),
+ exec: {
+ '#withCommand':: d.fn(help="\"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.\"", args=[d.arg(name='command', type=d.T.array)]),
+ withCommand(command): { lifecycle+: { postStart+: { exec+: { command: if std.isArray(v=command) then command else [command] } } } },
+ '#withCommandMixin':: d.fn(help="\"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='command', type=d.T.array)]),
+ withCommandMixin(command): { lifecycle+: { postStart+: { exec+: { command+: if std.isArray(v=command) then command else [command] } } } },
+ },
+ '#httpGet':: d.obj(help='"HTTPGet specifies the http request to perform."'),
+ httpGet: {
+ '#httpHeaders':: d.obj(help='"Custom headers to set in the request. HTTP allows repeated headers."'),
+ httpHeaders: {
+ '#withName':: d.fn(help='"The header field name. This will be canonicalized upon output, so case-variant names will be understood as the same header."', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { name: name },
+ '#withValue':: d.fn(help='"The header field value"', args=[d.arg(name='value', type=d.T.string)]),
+ withValue(value): { value: value },
+ },
+ '#withHost':: d.fn(help='"Host name to connect to, defaults to the pod IP. You probably want to set \\"Host\\" in httpHeaders instead."', args=[d.arg(name='host', type=d.T.string)]),
+ withHost(host): { lifecycle+: { postStart+: { httpGet+: { host: host } } } },
+ '#withHttpHeaders':: d.fn(help='"Custom headers to set in the request. HTTP allows repeated headers."', args=[d.arg(name='httpHeaders', type=d.T.array)]),
+ withHttpHeaders(httpHeaders): { lifecycle+: { postStart+: { httpGet+: { httpHeaders: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } } },
+ '#withHttpHeadersMixin':: d.fn(help='"Custom headers to set in the request. HTTP allows repeated headers."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='httpHeaders', type=d.T.array)]),
+ withHttpHeadersMixin(httpHeaders): { lifecycle+: { postStart+: { httpGet+: { httpHeaders+: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } } },
+ '#withPath':: d.fn(help='"Path to access on the HTTP server."', args=[d.arg(name='path', type=d.T.string)]),
+ withPath(path): { lifecycle+: { postStart+: { httpGet+: { path: path } } } },
+ '#withPort':: d.fn(help='"Name or number of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."', args=[d.arg(name='port', type=d.T.any)]),
+ withPort(port): { lifecycle+: { postStart+: { httpGet+: { port: port } } } },
+ '#withScheme':: d.fn(help='"Scheme to use for connecting to the host. Defaults to HTTP."', args=[d.arg(name='scheme', type=d.T.string)]),
+ withScheme(scheme): { lifecycle+: { postStart+: { httpGet+: { scheme: scheme } } } },
+ },
+ '#tcpSocket':: d.obj(help='"Deprecated. TCPSocket is NOT supported as a LifecycleHandler and kept for the backward compatibility. There are no validation of this field and lifecycle hooks will fail in runtime when tcp handler is specified."'),
+ tcpSocket: {
+ '#withHost':: d.fn(help='"Optional: Host name to connect to, defaults to the pod IP."', args=[d.arg(name='host', type=d.T.string)]),
+ withHost(host): { lifecycle+: { postStart+: { tcpSocket+: { host: host } } } },
+ '#withPort':: d.fn(help='"Number or name of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."', args=[d.arg(name='port', type=d.T.any)]),
+ withPort(port): { lifecycle+: { postStart+: { tcpSocket+: { port: port } } } },
+ },
+ },
+ '#preStop':: d.obj(help="\"PreStop is called immediately before a container is terminated due to an API request or management event such as liveness/startup probe failure, preemption, resource contention, etc. The handler is not called if the container crashes or exits. The Pod's termination grace period countdown begins before the PreStop hook is executed. Regardless of the outcome of the handler, the container will eventually terminate within the Pod's termination grace period (unless delayed by finalizers). Other management of the container blocks until the hook completes or until the termination grace period is reached. More info: https://kubernetes.io/docs/concepts/containers/container-lifecycle-hooks/#container-hooks\""),
+ preStop: {
+ '#exec':: d.obj(help='"Exec specifies the action to take."'),
+ exec: {
+ '#withCommand':: d.fn(help="\"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.\"", args=[d.arg(name='command', type=d.T.array)]),
+ withCommand(command): { lifecycle+: { preStop+: { exec+: { command: if std.isArray(v=command) then command else [command] } } } },
+ '#withCommandMixin':: d.fn(help="\"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='command', type=d.T.array)]),
+ withCommandMixin(command): { lifecycle+: { preStop+: { exec+: { command+: if std.isArray(v=command) then command else [command] } } } },
+ },
+ '#httpGet':: d.obj(help='"HTTPGet specifies the http request to perform."'),
+ httpGet: {
+ '#httpHeaders':: d.obj(help='"Custom headers to set in the request. HTTP allows repeated headers."'),
+ httpHeaders: {
+ '#withName':: d.fn(help='"The header field name. This will be canonicalized upon output, so case-variant names will be understood as the same header."', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { name: name },
+ '#withValue':: d.fn(help='"The header field value"', args=[d.arg(name='value', type=d.T.string)]),
+ withValue(value): { value: value },
+ },
+ '#withHost':: d.fn(help='"Host name to connect to, defaults to the pod IP. You probably want to set \\"Host\\" in httpHeaders instead."', args=[d.arg(name='host', type=d.T.string)]),
+ withHost(host): { lifecycle+: { preStop+: { httpGet+: { host: host } } } },
+ '#withHttpHeaders':: d.fn(help='"Custom headers to set in the request. HTTP allows repeated headers."', args=[d.arg(name='httpHeaders', type=d.T.array)]),
+ withHttpHeaders(httpHeaders): { lifecycle+: { preStop+: { httpGet+: { httpHeaders: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } } },
+ '#withHttpHeadersMixin':: d.fn(help='"Custom headers to set in the request. HTTP allows repeated headers."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='httpHeaders', type=d.T.array)]),
+ withHttpHeadersMixin(httpHeaders): { lifecycle+: { preStop+: { httpGet+: { httpHeaders+: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } } },
+ '#withPath':: d.fn(help='"Path to access on the HTTP server."', args=[d.arg(name='path', type=d.T.string)]),
+ withPath(path): { lifecycle+: { preStop+: { httpGet+: { path: path } } } },
+ '#withPort':: d.fn(help='"Name or number of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."', args=[d.arg(name='port', type=d.T.any)]),
+ withPort(port): { lifecycle+: { preStop+: { httpGet+: { port: port } } } },
+ '#withScheme':: d.fn(help='"Scheme to use for connecting to the host. Defaults to HTTP."', args=[d.arg(name='scheme', type=d.T.string)]),
+ withScheme(scheme): { lifecycle+: { preStop+: { httpGet+: { scheme: scheme } } } },
+ },
+ '#tcpSocket':: d.obj(help='"Deprecated. TCPSocket is NOT supported as a LifecycleHandler and kept for the backward compatibility. There are no validation of this field and lifecycle hooks will fail in runtime when tcp handler is specified."'),
+ tcpSocket: {
+ '#withHost':: d.fn(help='"Optional: Host name to connect to, defaults to the pod IP."', args=[d.arg(name='host', type=d.T.string)]),
+ withHost(host): { lifecycle+: { preStop+: { tcpSocket+: { host: host } } } },
+ '#withPort':: d.fn(help='"Number or name of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."', args=[d.arg(name='port', type=d.T.any)]),
+ withPort(port): { lifecycle+: { preStop+: { tcpSocket+: { port: port } } } },
+ },
+ },
+ },
+ '#livenessProbe':: d.obj(help='"Probes are not allowed for ephemeral containers."'),
+ livenessProbe: {
+ '#exec':: d.obj(help='"Exec specifies the action to take."'),
+ exec: {
+ '#withCommand':: d.fn(help="\"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.\"", args=[d.arg(name='command', type=d.T.array)]),
+ withCommand(command): { livenessProbe+: { exec+: { command: if std.isArray(v=command) then command else [command] } } },
+ '#withCommandMixin':: d.fn(help="\"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='command', type=d.T.array)]),
+ withCommandMixin(command): { livenessProbe+: { exec+: { command+: if std.isArray(v=command) then command else [command] } } },
+ },
+ '#grpc':: d.obj(help='"GRPC specifies an action involving a GRPC port."'),
+ grpc: {
+ '#withPort':: d.fn(help='"Port number of the gRPC service. Number must be in the range 1 to 65535."', args=[d.arg(name='port', type=d.T.integer)]),
+ withPort(port): { livenessProbe+: { grpc+: { port: port } } },
+ '#withService':: d.fn(help='"Service is the name of the service to place in the gRPC HealthCheckRequest (see https://github.com/grpc/grpc/blob/master/doc/health-checking.md). \\n If this is not specified, the default behavior is defined by gRPC."', args=[d.arg(name='service', type=d.T.string)]),
+ withService(service): { livenessProbe+: { grpc+: { service: service } } },
+ },
+ '#httpGet':: d.obj(help='"HTTPGet specifies the http request to perform."'),
+ httpGet: {
+ '#httpHeaders':: d.obj(help='"Custom headers to set in the request. HTTP allows repeated headers."'),
+ httpHeaders: {
+ '#withName':: d.fn(help='"The header field name. This will be canonicalized upon output, so case-variant names will be understood as the same header."', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { name: name },
+ '#withValue':: d.fn(help='"The header field value"', args=[d.arg(name='value', type=d.T.string)]),
+ withValue(value): { value: value },
+ },
+ '#withHost':: d.fn(help='"Host name to connect to, defaults to the pod IP. You probably want to set \\"Host\\" in httpHeaders instead."', args=[d.arg(name='host', type=d.T.string)]),
+ withHost(host): { livenessProbe+: { httpGet+: { host: host } } },
+ '#withHttpHeaders':: d.fn(help='"Custom headers to set in the request. HTTP allows repeated headers."', args=[d.arg(name='httpHeaders', type=d.T.array)]),
+ withHttpHeaders(httpHeaders): { livenessProbe+: { httpGet+: { httpHeaders: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } },
+ '#withHttpHeadersMixin':: d.fn(help='"Custom headers to set in the request. HTTP allows repeated headers."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='httpHeaders', type=d.T.array)]),
+ withHttpHeadersMixin(httpHeaders): { livenessProbe+: { httpGet+: { httpHeaders+: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } },
+ '#withPath':: d.fn(help='"Path to access on the HTTP server."', args=[d.arg(name='path', type=d.T.string)]),
+ withPath(path): { livenessProbe+: { httpGet+: { path: path } } },
+ '#withPort':: d.fn(help='"Name or number of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."', args=[d.arg(name='port', type=d.T.any)]),
+ withPort(port): { livenessProbe+: { httpGet+: { port: port } } },
+ '#withScheme':: d.fn(help='"Scheme to use for connecting to the host. Defaults to HTTP."', args=[d.arg(name='scheme', type=d.T.string)]),
+ withScheme(scheme): { livenessProbe+: { httpGet+: { scheme: scheme } } },
+ },
+ '#tcpSocket':: d.obj(help='"TCPSocket specifies an action involving a TCP port."'),
+ tcpSocket: {
+ '#withHost':: d.fn(help='"Optional: Host name to connect to, defaults to the pod IP."', args=[d.arg(name='host', type=d.T.string)]),
+ withHost(host): { livenessProbe+: { tcpSocket+: { host: host } } },
+ '#withPort':: d.fn(help='"Number or name of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."', args=[d.arg(name='port', type=d.T.any)]),
+ withPort(port): { livenessProbe+: { tcpSocket+: { port: port } } },
+ },
+ '#withFailureThreshold':: d.fn(help='"Minimum consecutive failures for the probe to be considered failed after having succeeded. Defaults to 3. Minimum value is 1."', args=[d.arg(name='failureThreshold', type=d.T.integer)]),
+ withFailureThreshold(failureThreshold): { livenessProbe+: { failureThreshold: failureThreshold } },
+ '#withInitialDelaySeconds':: d.fn(help='"Number of seconds after the container has started before liveness probes are initiated. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes"', args=[d.arg(name='initialDelaySeconds', type=d.T.integer)]),
+ withInitialDelaySeconds(initialDelaySeconds): { livenessProbe+: { initialDelaySeconds: initialDelaySeconds } },
+ '#withPeriodSeconds':: d.fn(help='"How often (in seconds) to perform the probe. Default to 10 seconds. Minimum value is 1."', args=[d.arg(name='periodSeconds', type=d.T.integer)]),
+ withPeriodSeconds(periodSeconds): { livenessProbe+: { periodSeconds: periodSeconds } },
+ '#withSuccessThreshold':: d.fn(help='"Minimum consecutive successes for the probe to be considered successful after having failed. Defaults to 1. Must be 1 for liveness and startup. Minimum value is 1."', args=[d.arg(name='successThreshold', type=d.T.integer)]),
+ withSuccessThreshold(successThreshold): { livenessProbe+: { successThreshold: successThreshold } },
+ '#withTerminationGracePeriodSeconds':: d.fn(help="\"Optional duration in seconds the pod needs to terminate gracefully upon probe failure. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. If this value is nil, the pod's terminationGracePeriodSeconds will be used. Otherwise, this value overrides the value provided by the pod spec. Value must be non-negative integer. The value zero indicates stop immediately via the kill signal (no opportunity to shut down). This is a beta field and requires enabling ProbeTerminationGracePeriod feature gate. Minimum value is 1. spec.terminationGracePeriodSeconds is used if unset.\"", args=[d.arg(name='terminationGracePeriodSeconds', type=d.T.integer)]),
+ withTerminationGracePeriodSeconds(terminationGracePeriodSeconds): { livenessProbe+: { terminationGracePeriodSeconds: terminationGracePeriodSeconds } },
+ '#withTimeoutSeconds':: d.fn(help='"Number of seconds after which the probe times out. Defaults to 1 second. Minimum value is 1. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes"', args=[d.arg(name='timeoutSeconds', type=d.T.integer)]),
+ withTimeoutSeconds(timeoutSeconds): { livenessProbe+: { timeoutSeconds: timeoutSeconds } },
+ },
+ '#ports':: d.obj(help='"Ports are not allowed for ephemeral containers."'),
+ ports: {
+ '#withContainerPort':: d.fn(help="\"Number of port to expose on the pod's IP address. This must be a valid port number, 0 \u003c x \u003c 65536.\"", args=[d.arg(name='containerPort', type=d.T.integer)]),
+ withContainerPort(containerPort): { containerPort: containerPort },
+ '#withHostIP':: d.fn(help='"What host IP to bind the external port to."', args=[d.arg(name='hostIP', type=d.T.string)]),
+ withHostIP(hostIP): { hostIP: hostIP },
+ '#withHostPort':: d.fn(help='"Number of port to expose on the host. If specified, this must be a valid port number, 0 < x < 65536. If HostNetwork is specified, this must match ContainerPort. Most containers do not need this."', args=[d.arg(name='hostPort', type=d.T.integer)]),
+ withHostPort(hostPort): { hostPort: hostPort },
+ '#withName':: d.fn(help='"If specified, this must be an IANA_SVC_NAME and unique within the pod. Each named port in a pod must have a unique name. Name for the port that can be referred to by services."', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { name: name },
+ '#withProtocol':: d.fn(help='"Protocol for port. Must be UDP, TCP, or SCTP. Defaults to \\"TCP\\"."', args=[d.arg(name='protocol', type=d.T.string)]),
+ withProtocol(protocol): { protocol: protocol },
+ },
+ '#readinessProbe':: d.obj(help='"Probes are not allowed for ephemeral containers."'),
+ readinessProbe: {
+ '#exec':: d.obj(help='"Exec specifies the action to take."'),
+ exec: {
+ '#withCommand':: d.fn(help="\"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.\"", args=[d.arg(name='command', type=d.T.array)]),
+ withCommand(command): { readinessProbe+: { exec+: { command: if std.isArray(v=command) then command else [command] } } },
+ '#withCommandMixin':: d.fn(help="\"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='command', type=d.T.array)]),
+ withCommandMixin(command): { readinessProbe+: { exec+: { command+: if std.isArray(v=command) then command else [command] } } },
+ },
+ '#grpc':: d.obj(help='"GRPC specifies an action involving a GRPC port."'),
+ grpc: {
+ '#withPort':: d.fn(help='"Port number of the gRPC service. Number must be in the range 1 to 65535."', args=[d.arg(name='port', type=d.T.integer)]),
+ withPort(port): { readinessProbe+: { grpc+: { port: port } } },
+ '#withService':: d.fn(help='"Service is the name of the service to place in the gRPC HealthCheckRequest (see https://github.com/grpc/grpc/blob/master/doc/health-checking.md). \\n If this is not specified, the default behavior is defined by gRPC."', args=[d.arg(name='service', type=d.T.string)]),
+ withService(service): { readinessProbe+: { grpc+: { service: service } } },
+ },
+ '#httpGet':: d.obj(help='"HTTPGet specifies the http request to perform."'),
+ httpGet: {
+ '#httpHeaders':: d.obj(help='"Custom headers to set in the request. HTTP allows repeated headers."'),
+ httpHeaders: {
+ '#withName':: d.fn(help='"The header field name. This will be canonicalized upon output, so case-variant names will be understood as the same header."', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { name: name },
+ '#withValue':: d.fn(help='"The header field value"', args=[d.arg(name='value', type=d.T.string)]),
+ withValue(value): { value: value },
+ },
+ '#withHost':: d.fn(help='"Host name to connect to, defaults to the pod IP. You probably want to set \\"Host\\" in httpHeaders instead."', args=[d.arg(name='host', type=d.T.string)]),
+ withHost(host): { readinessProbe+: { httpGet+: { host: host } } },
+ '#withHttpHeaders':: d.fn(help='"Custom headers to set in the request. HTTP allows repeated headers."', args=[d.arg(name='httpHeaders', type=d.T.array)]),
+ withHttpHeaders(httpHeaders): { readinessProbe+: { httpGet+: { httpHeaders: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } },
+ '#withHttpHeadersMixin':: d.fn(help='"Custom headers to set in the request. HTTP allows repeated headers."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='httpHeaders', type=d.T.array)]),
+ withHttpHeadersMixin(httpHeaders): { readinessProbe+: { httpGet+: { httpHeaders+: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } },
+ '#withPath':: d.fn(help='"Path to access on the HTTP server."', args=[d.arg(name='path', type=d.T.string)]),
+ withPath(path): { readinessProbe+: { httpGet+: { path: path } } },
+ '#withPort':: d.fn(help='"Name or number of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."', args=[d.arg(name='port', type=d.T.any)]),
+ withPort(port): { readinessProbe+: { httpGet+: { port: port } } },
+ '#withScheme':: d.fn(help='"Scheme to use for connecting to the host. Defaults to HTTP."', args=[d.arg(name='scheme', type=d.T.string)]),
+ withScheme(scheme): { readinessProbe+: { httpGet+: { scheme: scheme } } },
+ },
+ '#tcpSocket':: d.obj(help='"TCPSocket specifies an action involving a TCP port."'),
+ tcpSocket: {
+ '#withHost':: d.fn(help='"Optional: Host name to connect to, defaults to the pod IP."', args=[d.arg(name='host', type=d.T.string)]),
+ withHost(host): { readinessProbe+: { tcpSocket+: { host: host } } },
+ '#withPort':: d.fn(help='"Number or name of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."', args=[d.arg(name='port', type=d.T.any)]),
+ withPort(port): { readinessProbe+: { tcpSocket+: { port: port } } },
+ },
+ '#withFailureThreshold':: d.fn(help='"Minimum consecutive failures for the probe to be considered failed after having succeeded. Defaults to 3. Minimum value is 1."', args=[d.arg(name='failureThreshold', type=d.T.integer)]),
+ withFailureThreshold(failureThreshold): { readinessProbe+: { failureThreshold: failureThreshold } },
+ '#withInitialDelaySeconds':: d.fn(help='"Number of seconds after the container has started before liveness probes are initiated. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes"', args=[d.arg(name='initialDelaySeconds', type=d.T.integer)]),
+ withInitialDelaySeconds(initialDelaySeconds): { readinessProbe+: { initialDelaySeconds: initialDelaySeconds } },
+ '#withPeriodSeconds':: d.fn(help='"How often (in seconds) to perform the probe. Default to 10 seconds. Minimum value is 1."', args=[d.arg(name='periodSeconds', type=d.T.integer)]),
+ withPeriodSeconds(periodSeconds): { readinessProbe+: { periodSeconds: periodSeconds } },
+ '#withSuccessThreshold':: d.fn(help='"Minimum consecutive successes for the probe to be considered successful after having failed. Defaults to 1. Must be 1 for liveness and startup. Minimum value is 1."', args=[d.arg(name='successThreshold', type=d.T.integer)]),
+ withSuccessThreshold(successThreshold): { readinessProbe+: { successThreshold: successThreshold } },
+ '#withTerminationGracePeriodSeconds':: d.fn(help="\"Optional duration in seconds the pod needs to terminate gracefully upon probe failure. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. If this value is nil, the pod's terminationGracePeriodSeconds will be used. Otherwise, this value overrides the value provided by the pod spec. Value must be non-negative integer. The value zero indicates stop immediately via the kill signal (no opportunity to shut down). This is a beta field and requires enabling ProbeTerminationGracePeriod feature gate. Minimum value is 1. spec.terminationGracePeriodSeconds is used if unset.\"", args=[d.arg(name='terminationGracePeriodSeconds', type=d.T.integer)]),
+ withTerminationGracePeriodSeconds(terminationGracePeriodSeconds): { readinessProbe+: { terminationGracePeriodSeconds: terminationGracePeriodSeconds } },
+ '#withTimeoutSeconds':: d.fn(help='"Number of seconds after which the probe times out. Defaults to 1 second. Minimum value is 1. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes"', args=[d.arg(name='timeoutSeconds', type=d.T.integer)]),
+ withTimeoutSeconds(timeoutSeconds): { readinessProbe+: { timeoutSeconds: timeoutSeconds } },
+ },
+ '#resizePolicy':: d.obj(help='"Resources resize policy for the container."'),
+ resizePolicy: {
+ '#withResourceName':: d.fn(help='"Name of the resource to which this resource resize policy applies. Supported values: cpu, memory."', args=[d.arg(name='resourceName', type=d.T.string)]),
+ withResourceName(resourceName): { resourceName: resourceName },
+ '#withRestartPolicy':: d.fn(help='"Restart policy to apply when specified resource is resized. If not specified, it defaults to NotRequired."', args=[d.arg(name='restartPolicy', type=d.T.string)]),
+ withRestartPolicy(restartPolicy): { restartPolicy: restartPolicy },
+ },
+ '#resources':: d.obj(help='"Resources are not allowed for ephemeral containers. Ephemeral containers use spare resources already allocated to the pod."'),
+ resources: {
+ '#claims':: d.obj(help='"Claims lists the names of resources, defined in spec.resourceClaims, that are used by this container. \\n This is an alpha field and requires enabling the DynamicResourceAllocation feature gate. \\n This field is immutable. It can only be set for containers."'),
+ claims: {
+ '#withName':: d.fn(help='"Name must match the name of one entry in pod.spec.resourceClaims of the Pod where this field is used. It makes that resource available inside a container."', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { name: name },
+ },
+ '#withClaims':: d.fn(help='"Claims lists the names of resources, defined in spec.resourceClaims, that are used by this container. \\n This is an alpha field and requires enabling the DynamicResourceAllocation feature gate. \\n This field is immutable. It can only be set for containers."', args=[d.arg(name='claims', type=d.T.array)]),
+ withClaims(claims): { resources+: { claims: if std.isArray(v=claims) then claims else [claims] } },
+ '#withClaimsMixin':: d.fn(help='"Claims lists the names of resources, defined in spec.resourceClaims, that are used by this container. \\n This is an alpha field and requires enabling the DynamicResourceAllocation feature gate. \\n This field is immutable. It can only be set for containers."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='claims', type=d.T.array)]),
+ withClaimsMixin(claims): { resources+: { claims+: if std.isArray(v=claims) then claims else [claims] } },
+ '#withLimits':: d.fn(help='"Limits describes the maximum amount of compute resources allowed. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"', args=[d.arg(name='limits', type=d.T.object)]),
+ withLimits(limits): { resources+: { limits: limits } },
+ '#withLimitsMixin':: d.fn(help='"Limits describes the maximum amount of compute resources allowed. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='limits', type=d.T.object)]),
+ withLimitsMixin(limits): { resources+: { limits+: limits } },
+ '#withRequests':: d.fn(help='"Requests describes the minimum amount of compute resources required. If Requests is omitted for a container, it defaults to Limits if that is explicitly specified, otherwise to an implementation-defined value. Requests cannot exceed Limits. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"', args=[d.arg(name='requests', type=d.T.object)]),
+ withRequests(requests): { resources+: { requests: requests } },
+ '#withRequestsMixin':: d.fn(help='"Requests describes the minimum amount of compute resources required. If Requests is omitted for a container, it defaults to Limits if that is explicitly specified, otherwise to an implementation-defined value. Requests cannot exceed Limits. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='requests', type=d.T.object)]),
+ withRequestsMixin(requests): { resources+: { requests+: requests } },
+ },
+ '#securityContext':: d.obj(help='"Optional: SecurityContext defines the security options the ephemeral container should be run with. If set, the fields of SecurityContext override the equivalent fields of PodSecurityContext."'),
+ securityContext: {
+ '#capabilities':: d.obj(help='"The capabilities to add/drop when running containers. Defaults to the default set of capabilities granted by the container runtime. Note that this field cannot be set when spec.os.name is windows."'),
+ capabilities: {
+ '#withAdd':: d.fn(help='"Added capabilities"', args=[d.arg(name='add', type=d.T.array)]),
+ withAdd(add): { securityContext+: { capabilities+: { add: if std.isArray(v=add) then add else [add] } } },
+ '#withAddMixin':: d.fn(help='"Added capabilities"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='add', type=d.T.array)]),
+ withAddMixin(add): { securityContext+: { capabilities+: { add+: if std.isArray(v=add) then add else [add] } } },
+ '#withDrop':: d.fn(help='"Removed capabilities"', args=[d.arg(name='drop', type=d.T.array)]),
+ withDrop(drop): { securityContext+: { capabilities+: { drop: if std.isArray(v=drop) then drop else [drop] } } },
+ '#withDropMixin':: d.fn(help='"Removed capabilities"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='drop', type=d.T.array)]),
+ withDropMixin(drop): { securityContext+: { capabilities+: { drop+: if std.isArray(v=drop) then drop else [drop] } } },
+ },
+ '#seLinuxOptions':: d.obj(help='"The SELinux context to be applied to the container. If unspecified, the container runtime will allocate a random SELinux context for each container. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. Note that this field cannot be set when spec.os.name is windows."'),
+ seLinuxOptions: {
+ '#withLevel':: d.fn(help='"Level is SELinux level label that applies to the container."', args=[d.arg(name='level', type=d.T.string)]),
+ withLevel(level): { securityContext+: { seLinuxOptions+: { level: level } } },
+ '#withRole':: d.fn(help='"Role is a SELinux role label that applies to the container."', args=[d.arg(name='role', type=d.T.string)]),
+ withRole(role): { securityContext+: { seLinuxOptions+: { role: role } } },
+ '#withType':: d.fn(help='"Type is a SELinux type label that applies to the container."', args=[d.arg(name='type', type=d.T.string)]),
+ withType(type): { securityContext+: { seLinuxOptions+: { type: type } } },
+ '#withUser':: d.fn(help='"User is a SELinux user label that applies to the container."', args=[d.arg(name='user', type=d.T.string)]),
+ withUser(user): { securityContext+: { seLinuxOptions+: { user: user } } },
+ },
+ '#seccompProfile':: d.obj(help='"The seccomp options to use by this container. If seccomp options are provided at both the pod & container level, the container options override the pod options. Note that this field cannot be set when spec.os.name is windows."'),
+ seccompProfile: {
+ '#withLocalhostProfile':: d.fn(help="\"localhostProfile indicates a profile defined in a file on the node should be used. The profile must be preconfigured on the node to work. Must be a descending path, relative to the kubelet's configured seccomp profile location. Must be set if type is \\\"Localhost\\\". Must NOT be set for any other type.\"", args=[d.arg(name='localhostProfile', type=d.T.string)]),
+ withLocalhostProfile(localhostProfile): { securityContext+: { seccompProfile+: { localhostProfile: localhostProfile } } },
+ '#withType':: d.fn(help='"type indicates which kind of seccomp profile will be applied. Valid options are: \\n Localhost - a profile defined in a file on the node should be used. RuntimeDefault - the container runtime default profile should be used. Unconfined - no profile should be applied."', args=[d.arg(name='type', type=d.T.string)]),
+ withType(type): { securityContext+: { seccompProfile+: { type: type } } },
+ },
+ '#windowsOptions':: d.obj(help='"The Windows specific settings applied to all containers. If unspecified, the options from the PodSecurityContext will be used. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. Note that this field cannot be set when spec.os.name is linux."'),
+ windowsOptions: {
+ '#withGmsaCredentialSpec':: d.fn(help='"GMSACredentialSpec is where the GMSA admission webhook (https://github.com/kubernetes-sigs/windows-gmsa) inlines the contents of the GMSA credential spec named by the GMSACredentialSpecName field."', args=[d.arg(name='gmsaCredentialSpec', type=d.T.string)]),
+ withGmsaCredentialSpec(gmsaCredentialSpec): { securityContext+: { windowsOptions+: { gmsaCredentialSpec: gmsaCredentialSpec } } },
+ '#withGmsaCredentialSpecName':: d.fn(help='"GMSACredentialSpecName is the name of the GMSA credential spec to use."', args=[d.arg(name='gmsaCredentialSpecName', type=d.T.string)]),
+ withGmsaCredentialSpecName(gmsaCredentialSpecName): { securityContext+: { windowsOptions+: { gmsaCredentialSpecName: gmsaCredentialSpecName } } },
+ '#withHostProcess':: d.fn(help="\"HostProcess determines if a container should be run as a 'Host Process' container. All of a Pod's containers must have the same effective HostProcess value (it is not allowed to have a mix of HostProcess containers and non-HostProcess containers). In addition, if HostProcess is true then HostNetwork must also be set to true.\"", args=[d.arg(name='hostProcess', type=d.T.boolean)]),
+ withHostProcess(hostProcess): { securityContext+: { windowsOptions+: { hostProcess: hostProcess } } },
+ '#withRunAsUserName':: d.fn(help='"The UserName in Windows to run the entrypoint of the container process. Defaults to the user specified in image metadata if unspecified. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence."', args=[d.arg(name='runAsUserName', type=d.T.string)]),
+ withRunAsUserName(runAsUserName): { securityContext+: { windowsOptions+: { runAsUserName: runAsUserName } } },
+ },
+ '#withAllowPrivilegeEscalation':: d.fn(help='"AllowPrivilegeEscalation controls whether a process can gain more privileges than its parent process. This bool directly controls if the no_new_privs flag will be set on the container process. AllowPrivilegeEscalation is true always when the container is: 1) run as Privileged 2) has CAP_SYS_ADMIN Note that this field cannot be set when spec.os.name is windows."', args=[d.arg(name='allowPrivilegeEscalation', type=d.T.boolean)]),
+ withAllowPrivilegeEscalation(allowPrivilegeEscalation): { securityContext+: { allowPrivilegeEscalation: allowPrivilegeEscalation } },
+ '#withPrivileged':: d.fn(help='"Run container in privileged mode. Processes in privileged containers are essentially equivalent to root on the host. Defaults to false. Note that this field cannot be set when spec.os.name is windows."', args=[d.arg(name='privileged', type=d.T.boolean)]),
+ withPrivileged(privileged): { securityContext+: { privileged: privileged } },
+ '#withProcMount':: d.fn(help='"procMount denotes the type of proc mount to use for the containers. The default is DefaultProcMount which uses the container runtime defaults for readonly paths and masked paths. This requires the ProcMountType feature flag to be enabled. Note that this field cannot be set when spec.os.name is windows."', args=[d.arg(name='procMount', type=d.T.string)]),
+ withProcMount(procMount): { securityContext+: { procMount: procMount } },
+ '#withReadOnlyRootFilesystem':: d.fn(help='"Whether this container has a read-only root filesystem. Default is false. Note that this field cannot be set when spec.os.name is windows."', args=[d.arg(name='readOnlyRootFilesystem', type=d.T.boolean)]),
+ withReadOnlyRootFilesystem(readOnlyRootFilesystem): { securityContext+: { readOnlyRootFilesystem: readOnlyRootFilesystem } },
+ '#withRunAsGroup':: d.fn(help='"The GID to run the entrypoint of the container process. Uses runtime default if unset. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. Note that this field cannot be set when spec.os.name is windows."', args=[d.arg(name='runAsGroup', type=d.T.integer)]),
+ withRunAsGroup(runAsGroup): { securityContext+: { runAsGroup: runAsGroup } },
+ '#withRunAsNonRoot':: d.fn(help='"Indicates that the container must run as a non-root user. If true, the Kubelet will validate the image at runtime to ensure that it does not run as UID 0 (root) and fail to start the container if it does. If unset or false, no such validation will be performed. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence."', args=[d.arg(name='runAsNonRoot', type=d.T.boolean)]),
+ withRunAsNonRoot(runAsNonRoot): { securityContext+: { runAsNonRoot: runAsNonRoot } },
+ '#withRunAsUser':: d.fn(help='"The UID to run the entrypoint of the container process. Defaults to user specified in image metadata if unspecified. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. Note that this field cannot be set when spec.os.name is windows."', args=[d.arg(name='runAsUser', type=d.T.integer)]),
+ withRunAsUser(runAsUser): { securityContext+: { runAsUser: runAsUser } },
+ },
+ '#startupProbe':: d.obj(help='"Probes are not allowed for ephemeral containers."'),
+ startupProbe: {
+ '#exec':: d.obj(help='"Exec specifies the action to take."'),
+ exec: {
+ '#withCommand':: d.fn(help="\"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.\"", args=[d.arg(name='command', type=d.T.array)]),
+ withCommand(command): { startupProbe+: { exec+: { command: if std.isArray(v=command) then command else [command] } } },
+ '#withCommandMixin':: d.fn(help="\"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='command', type=d.T.array)]),
+ withCommandMixin(command): { startupProbe+: { exec+: { command+: if std.isArray(v=command) then command else [command] } } },
+ },
+ '#grpc':: d.obj(help='"GRPC specifies an action involving a GRPC port."'),
+ grpc: {
+ '#withPort':: d.fn(help='"Port number of the gRPC service. Number must be in the range 1 to 65535."', args=[d.arg(name='port', type=d.T.integer)]),
+ withPort(port): { startupProbe+: { grpc+: { port: port } } },
+ '#withService':: d.fn(help='"Service is the name of the service to place in the gRPC HealthCheckRequest (see https://github.com/grpc/grpc/blob/master/doc/health-checking.md). \\n If this is not specified, the default behavior is defined by gRPC."', args=[d.arg(name='service', type=d.T.string)]),
+ withService(service): { startupProbe+: { grpc+: { service: service } } },
+ },
+ '#httpGet':: d.obj(help='"HTTPGet specifies the http request to perform."'),
+ httpGet: {
+ '#httpHeaders':: d.obj(help='"Custom headers to set in the request. HTTP allows repeated headers."'),
+ httpHeaders: {
+ '#withName':: d.fn(help='"The header field name. This will be canonicalized upon output, so case-variant names will be understood as the same header."', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { name: name },
+ '#withValue':: d.fn(help='"The header field value"', args=[d.arg(name='value', type=d.T.string)]),
+ withValue(value): { value: value },
+ },
+ '#withHost':: d.fn(help='"Host name to connect to, defaults to the pod IP. You probably want to set \\"Host\\" in httpHeaders instead."', args=[d.arg(name='host', type=d.T.string)]),
+ withHost(host): { startupProbe+: { httpGet+: { host: host } } },
+ '#withHttpHeaders':: d.fn(help='"Custom headers to set in the request. HTTP allows repeated headers."', args=[d.arg(name='httpHeaders', type=d.T.array)]),
+ withHttpHeaders(httpHeaders): { startupProbe+: { httpGet+: { httpHeaders: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } },
+ '#withHttpHeadersMixin':: d.fn(help='"Custom headers to set in the request. HTTP allows repeated headers."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='httpHeaders', type=d.T.array)]),
+ withHttpHeadersMixin(httpHeaders): { startupProbe+: { httpGet+: { httpHeaders+: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } },
+ '#withPath':: d.fn(help='"Path to access on the HTTP server."', args=[d.arg(name='path', type=d.T.string)]),
+ withPath(path): { startupProbe+: { httpGet+: { path: path } } },
+ '#withPort':: d.fn(help='"Name or number of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."', args=[d.arg(name='port', type=d.T.any)]),
+ withPort(port): { startupProbe+: { httpGet+: { port: port } } },
+ '#withScheme':: d.fn(help='"Scheme to use for connecting to the host. Defaults to HTTP."', args=[d.arg(name='scheme', type=d.T.string)]),
+ withScheme(scheme): { startupProbe+: { httpGet+: { scheme: scheme } } },
+ },
+ '#tcpSocket':: d.obj(help='"TCPSocket specifies an action involving a TCP port."'),
+ tcpSocket: {
+ '#withHost':: d.fn(help='"Optional: Host name to connect to, defaults to the pod IP."', args=[d.arg(name='host', type=d.T.string)]),
+ withHost(host): { startupProbe+: { tcpSocket+: { host: host } } },
+ '#withPort':: d.fn(help='"Number or name of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."', args=[d.arg(name='port', type=d.T.any)]),
+ withPort(port): { startupProbe+: { tcpSocket+: { port: port } } },
+ },
+ '#withFailureThreshold':: d.fn(help='"Minimum consecutive failures for the probe to be considered failed after having succeeded. Defaults to 3. Minimum value is 1."', args=[d.arg(name='failureThreshold', type=d.T.integer)]),
+ withFailureThreshold(failureThreshold): { startupProbe+: { failureThreshold: failureThreshold } },
+ '#withInitialDelaySeconds':: d.fn(help='"Number of seconds after the container has started before liveness probes are initiated. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes"', args=[d.arg(name='initialDelaySeconds', type=d.T.integer)]),
+ withInitialDelaySeconds(initialDelaySeconds): { startupProbe+: { initialDelaySeconds: initialDelaySeconds } },
+ '#withPeriodSeconds':: d.fn(help='"How often (in seconds) to perform the probe. Default to 10 seconds. Minimum value is 1."', args=[d.arg(name='periodSeconds', type=d.T.integer)]),
+ withPeriodSeconds(periodSeconds): { startupProbe+: { periodSeconds: periodSeconds } },
+ '#withSuccessThreshold':: d.fn(help='"Minimum consecutive successes for the probe to be considered successful after having failed. Defaults to 1. Must be 1 for liveness and startup. Minimum value is 1."', args=[d.arg(name='successThreshold', type=d.T.integer)]),
+ withSuccessThreshold(successThreshold): { startupProbe+: { successThreshold: successThreshold } },
+ '#withTerminationGracePeriodSeconds':: d.fn(help="\"Optional duration in seconds the pod needs to terminate gracefully upon probe failure. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. If this value is nil, the pod's terminationGracePeriodSeconds will be used. Otherwise, this value overrides the value provided by the pod spec. Value must be non-negative integer. The value zero indicates stop immediately via the kill signal (no opportunity to shut down). This is a beta field and requires enabling ProbeTerminationGracePeriod feature gate. Minimum value is 1. spec.terminationGracePeriodSeconds is used if unset.\"", args=[d.arg(name='terminationGracePeriodSeconds', type=d.T.integer)]),
+ withTerminationGracePeriodSeconds(terminationGracePeriodSeconds): { startupProbe+: { terminationGracePeriodSeconds: terminationGracePeriodSeconds } },
+ '#withTimeoutSeconds':: d.fn(help='"Number of seconds after which the probe times out. Defaults to 1 second. Minimum value is 1. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes"', args=[d.arg(name='timeoutSeconds', type=d.T.integer)]),
+ withTimeoutSeconds(timeoutSeconds): { startupProbe+: { timeoutSeconds: timeoutSeconds } },
+ },
+ '#volumeDevices':: d.obj(help='"volumeDevices is the list of block devices to be used by the container."'),
+ volumeDevices: {
+ '#withDevicePath':: d.fn(help='"devicePath is the path inside of the container that the device will be mapped to."', args=[d.arg(name='devicePath', type=d.T.string)]),
+ withDevicePath(devicePath): { devicePath: devicePath },
+ '#withName':: d.fn(help='"name must match the name of a persistentVolumeClaim in the pod"', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { name: name },
+ },
+ '#volumeMounts':: d.obj(help="\"Pod volumes to mount into the container's filesystem. Subpath mounts are not allowed for ephemeral containers. Cannot be updated.\""),
+ volumeMounts: {
+ '#withMountPath':: d.fn(help="\"Path within the container at which the volume should be mounted. Must not contain ':'.\"", args=[d.arg(name='mountPath', type=d.T.string)]),
+ withMountPath(mountPath): { mountPath: mountPath },
+ '#withMountPropagation':: d.fn(help='"mountPropagation determines how mounts are propagated from the host to container and the other way around. When not set, MountPropagationNone is used. This field is beta in 1.10."', args=[d.arg(name='mountPropagation', type=d.T.string)]),
+ withMountPropagation(mountPropagation): { mountPropagation: mountPropagation },
+ '#withName':: d.fn(help='"This must match the Name of a Volume."', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { name: name },
+ '#withReadOnly':: d.fn(help='"Mounted read-only if true, read-write otherwise (false or unspecified). Defaults to false."', args=[d.arg(name='readOnly', type=d.T.boolean)]),
+ withReadOnly(readOnly): { readOnly: readOnly },
+ '#withSubPath':: d.fn(help="\"Path within the volume from which the container's volume should be mounted. Defaults to \\\"\\\" (volume's root).\"", args=[d.arg(name='subPath', type=d.T.string)]),
+ withSubPath(subPath): { subPath: subPath },
+ '#withSubPathExpr':: d.fn(help="\"Expanded path within the volume from which the container's volume should be mounted. Behaves similarly to SubPath but environment variable references $(VAR_NAME) are expanded using the container's environment. Defaults to \\\"\\\" (volume's root). SubPathExpr and SubPath are mutually exclusive.\"", args=[d.arg(name='subPathExpr', type=d.T.string)]),
+ withSubPathExpr(subPathExpr): { subPathExpr: subPathExpr },
+ },
+ '#withArgs':: d.fn(help="\"Arguments to the entrypoint. The image's CMD is used if this is not provided. Variable references $(VAR_NAME) are expanded using the container's environment. If a variable cannot be resolved, the reference in the input string will be unchanged. Double $$ are reduced to a single $, which allows for escaping the $(VAR_NAME) syntax: i.e. \\\"$$(VAR_NAME)\\\" will produce the string literal \\\"$(VAR_NAME)\\\". Escaped references will never be expanded, regardless of whether the variable exists or not. Cannot be updated. More info: https://kubernetes.io/docs/tasks/inject-data-application/define-command-argument-container/#running-a-command-in-a-shell\"", args=[d.arg(name='args', type=d.T.array)]),
+ withArgs(args): { args: if std.isArray(v=args) then args else [args] },
+ '#withArgsMixin':: d.fn(help="\"Arguments to the entrypoint. The image's CMD is used if this is not provided. Variable references $(VAR_NAME) are expanded using the container's environment. If a variable cannot be resolved, the reference in the input string will be unchanged. Double $$ are reduced to a single $, which allows for escaping the $(VAR_NAME) syntax: i.e. \\\"$$(VAR_NAME)\\\" will produce the string literal \\\"$(VAR_NAME)\\\". Escaped references will never be expanded, regardless of whether the variable exists or not. Cannot be updated. More info: https://kubernetes.io/docs/tasks/inject-data-application/define-command-argument-container/#running-a-command-in-a-shell\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='args', type=d.T.array)]),
+ withArgsMixin(args): { args+: if std.isArray(v=args) then args else [args] },
+ '#withCommand':: d.fn(help="\"Entrypoint array. Not executed within a shell. The image's ENTRYPOINT is used if this is not provided. Variable references $(VAR_NAME) are expanded using the container's environment. If a variable cannot be resolved, the reference in the input string will be unchanged. Double $$ are reduced to a single $, which allows for escaping the $(VAR_NAME) syntax: i.e. \\\"$$(VAR_NAME)\\\" will produce the string literal \\\"$(VAR_NAME)\\\". Escaped references will never be expanded, regardless of whether the variable exists or not. Cannot be updated. More info: https://kubernetes.io/docs/tasks/inject-data-application/define-command-argument-container/#running-a-command-in-a-shell\"", args=[d.arg(name='command', type=d.T.array)]),
+ withCommand(command): { command: if std.isArray(v=command) then command else [command] },
+ '#withCommandMixin':: d.fn(help="\"Entrypoint array. Not executed within a shell. The image's ENTRYPOINT is used if this is not provided. Variable references $(VAR_NAME) are expanded using the container's environment. If a variable cannot be resolved, the reference in the input string will be unchanged. Double $$ are reduced to a single $, which allows for escaping the $(VAR_NAME) syntax: i.e. \\\"$$(VAR_NAME)\\\" will produce the string literal \\\"$(VAR_NAME)\\\". Escaped references will never be expanded, regardless of whether the variable exists or not. Cannot be updated. More info: https://kubernetes.io/docs/tasks/inject-data-application/define-command-argument-container/#running-a-command-in-a-shell\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='command', type=d.T.array)]),
+ withCommandMixin(command): { command+: if std.isArray(v=command) then command else [command] },
+ '#withEnv':: d.fn(help='"List of environment variables to set in the container. Cannot be updated."', args=[d.arg(name='env', type=d.T.array)]),
+ withEnv(env): { env: if std.isArray(v=env) then env else [env] },
+ '#withEnvFrom':: d.fn(help='"List of sources to populate environment variables in the container. The keys defined within a source must be a C_IDENTIFIER. All invalid keys will be reported as an event when the container is starting. When a key exists in multiple sources, the value associated with the last source will take precedence. Values defined by an Env with a duplicate key will take precedence. Cannot be updated."', args=[d.arg(name='envFrom', type=d.T.array)]),
+ withEnvFrom(envFrom): { envFrom: if std.isArray(v=envFrom) then envFrom else [envFrom] },
+ '#withEnvFromMixin':: d.fn(help='"List of sources to populate environment variables in the container. The keys defined within a source must be a C_IDENTIFIER. All invalid keys will be reported as an event when the container is starting. When a key exists in multiple sources, the value associated with the last source will take precedence. Values defined by an Env with a duplicate key will take precedence. Cannot be updated."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='envFrom', type=d.T.array)]),
+ withEnvFromMixin(envFrom): { envFrom+: if std.isArray(v=envFrom) then envFrom else [envFrom] },
+ '#withEnvMixin':: d.fn(help='"List of environment variables to set in the container. Cannot be updated."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='env', type=d.T.array)]),
+ withEnvMixin(env): { env+: if std.isArray(v=env) then env else [env] },
+ '#withImage':: d.fn(help='"Container image name. More info: https://kubernetes.io/docs/concepts/containers/images"', args=[d.arg(name='image', type=d.T.string)]),
+ withImage(image): { image: image },
+ '#withImagePullPolicy':: d.fn(help='"Image pull policy. One of Always, Never, IfNotPresent. Defaults to Always if :latest tag is specified, or IfNotPresent otherwise. Cannot be updated. More info: https://kubernetes.io/docs/concepts/containers/images#updating-images"', args=[d.arg(name='imagePullPolicy', type=d.T.string)]),
+ withImagePullPolicy(imagePullPolicy): { imagePullPolicy: imagePullPolicy },
+ '#withName':: d.fn(help='"Name of the ephemeral container specified as a DNS_LABEL. This name must be unique among all containers, init containers and ephemeral containers."', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { name: name },
+ '#withPorts':: d.fn(help='"Ports are not allowed for ephemeral containers."', args=[d.arg(name='ports', type=d.T.array)]),
+ withPorts(ports): { ports: if std.isArray(v=ports) then ports else [ports] },
+ '#withPortsMixin':: d.fn(help='"Ports are not allowed for ephemeral containers."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ports', type=d.T.array)]),
+ withPortsMixin(ports): { ports+: if std.isArray(v=ports) then ports else [ports] },
+ '#withResizePolicy':: d.fn(help='"Resources resize policy for the container."', args=[d.arg(name='resizePolicy', type=d.T.array)]),
+ withResizePolicy(resizePolicy): { resizePolicy: if std.isArray(v=resizePolicy) then resizePolicy else [resizePolicy] },
+ '#withResizePolicyMixin':: d.fn(help='"Resources resize policy for the container."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='resizePolicy', type=d.T.array)]),
+ withResizePolicyMixin(resizePolicy): { resizePolicy+: if std.isArray(v=resizePolicy) then resizePolicy else [resizePolicy] },
+ '#withRestartPolicy':: d.fn(help='"Restart policy for the container to manage the restart behavior of each container within a pod. This may only be set for init containers. You cannot set this field on ephemeral containers."', args=[d.arg(name='restartPolicy', type=d.T.string)]),
+ withRestartPolicy(restartPolicy): { restartPolicy: restartPolicy },
+ '#withStdin':: d.fn(help='"Whether this container should allocate a buffer for stdin in the container runtime. If this is not set, reads from stdin in the container will always result in EOF. Default is false."', args=[d.arg(name='stdin', type=d.T.boolean)]),
+ withStdin(stdin): { stdin: stdin },
+ '#withStdinOnce':: d.fn(help='"Whether the container runtime should close the stdin channel after it has been opened by a single attach. When stdin is true the stdin stream will remain open across multiple attach sessions. If stdinOnce is set to true, stdin is opened on container start, is empty until the first client attaches to stdin, and then remains open and accepts data until the client disconnects, at which time stdin is closed and remains closed until the container is restarted. If this flag is false, a container processes that reads from stdin will never receive an EOF. Default is false"', args=[d.arg(name='stdinOnce', type=d.T.boolean)]),
+ withStdinOnce(stdinOnce): { stdinOnce: stdinOnce },
+ '#withTargetContainerName':: d.fn(help='"If set, the name of the container from PodSpec that this ephemeral container targets. The ephemeral container will be run in the namespaces (IPC, PID, etc) of this container. If not set then the ephemeral container uses the namespaces configured in the Pod spec. \\n The container runtime must implement support for this feature. If the runtime does not support namespace targeting then the result of setting this field is undefined."', args=[d.arg(name='targetContainerName', type=d.T.string)]),
+ withTargetContainerName(targetContainerName): { targetContainerName: targetContainerName },
+ '#withTerminationMessagePath':: d.fn(help="\"Optional: Path at which the file to which the container's termination message will be written is mounted into the container's filesystem. Message written is intended to be brief final status, such as an assertion failure message. Will be truncated by the node if greater than 4096 bytes. The total message length across all containers will be limited to 12kb. Defaults to /dev/termination-log. Cannot be updated.\"", args=[d.arg(name='terminationMessagePath', type=d.T.string)]),
+ withTerminationMessagePath(terminationMessagePath): { terminationMessagePath: terminationMessagePath },
+ '#withTerminationMessagePolicy':: d.fn(help='"Indicate how the termination message should be populated. File will use the contents of terminationMessagePath to populate the container status message on both success and failure. FallbackToLogsOnError will use the last chunk of container log output if the termination message file is empty and the container exited with an error. The log output is limited to 2048 bytes or 80 lines, whichever is smaller. Defaults to File. Cannot be updated."', args=[d.arg(name='terminationMessagePolicy', type=d.T.string)]),
+ withTerminationMessagePolicy(terminationMessagePolicy): { terminationMessagePolicy: terminationMessagePolicy },
+ '#withTty':: d.fn(help="\"Whether this container should allocate a TTY for itself, also requires 'stdin' to be true. Default is false.\"", args=[d.arg(name='tty', type=d.T.boolean)]),
+ withTty(tty): { tty: tty },
+ '#withVolumeDevices':: d.fn(help='"volumeDevices is the list of block devices to be used by the container."', args=[d.arg(name='volumeDevices', type=d.T.array)]),
+ withVolumeDevices(volumeDevices): { volumeDevices: if std.isArray(v=volumeDevices) then volumeDevices else [volumeDevices] },
+ '#withVolumeDevicesMixin':: d.fn(help='"volumeDevices is the list of block devices to be used by the container."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='volumeDevices', type=d.T.array)]),
+ withVolumeDevicesMixin(volumeDevices): { volumeDevices+: if std.isArray(v=volumeDevices) then volumeDevices else [volumeDevices] },
+ '#withVolumeMounts':: d.fn(help="\"Pod volumes to mount into the container's filesystem. Subpath mounts are not allowed for ephemeral containers. Cannot be updated.\"", args=[d.arg(name='volumeMounts', type=d.T.array)]),
+ withVolumeMounts(volumeMounts): { volumeMounts: if std.isArray(v=volumeMounts) then volumeMounts else [volumeMounts] },
+ '#withVolumeMountsMixin':: d.fn(help="\"Pod volumes to mount into the container's filesystem. Subpath mounts are not allowed for ephemeral containers. Cannot be updated.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='volumeMounts', type=d.T.array)]),
+ withVolumeMountsMixin(volumeMounts): { volumeMounts+: if std.isArray(v=volumeMounts) then volumeMounts else [volumeMounts] },
+ '#withWorkingDir':: d.fn(help="\"Container's working directory. If not specified, the container runtime's default will be used, which might be configured in the container image. Cannot be updated.\"", args=[d.arg(name='workingDir', type=d.T.string)]),
+ withWorkingDir(workingDir): { workingDir: workingDir },
+ },
+ '#hostAliases':: d.obj(help="\"HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods.\""),
+ hostAliases: {
+ '#withHostnames':: d.fn(help='"Hostnames for the above IP address."', args=[d.arg(name='hostnames', type=d.T.array)]),
+ withHostnames(hostnames): { hostnames: if std.isArray(v=hostnames) then hostnames else [hostnames] },
+ '#withHostnamesMixin':: d.fn(help='"Hostnames for the above IP address."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='hostnames', type=d.T.array)]),
+ withHostnamesMixin(hostnames): { hostnames+: if std.isArray(v=hostnames) then hostnames else [hostnames] },
+ '#withIp':: d.fn(help='"IP address of the host file entry."', args=[d.arg(name='ip', type=d.T.string)]),
+ withIp(ip): { ip: ip },
+ },
+ '#imagePullSecrets':: d.obj(help='"ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod"'),
+ imagePullSecrets: {
+ '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { name: name },
+ },
+ '#initContainers':: d.obj(help='"List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, Liveness probes, or Startup probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/"'),
+ initContainers: {
+ '#env':: d.obj(help='"List of environment variables to set in the container. Cannot be updated."'),
+ env: {
+ '#valueFrom':: d.obj(help="\"Source for the environment variable's value. Cannot be used if value is not empty.\""),
+ valueFrom: {
+ '#configMapKeyRef':: d.obj(help='"Selects a key of a ConfigMap."'),
+ configMapKeyRef: {
+ '#withKey':: d.fn(help='"The key to select."', args=[d.arg(name='key', type=d.T.string)]),
+ withKey(key): { valueFrom+: { configMapKeyRef+: { key: key } } },
+ '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { valueFrom+: { configMapKeyRef+: { name: name } } },
+ '#withOptional':: d.fn(help='"Specify whether the ConfigMap or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]),
+ withOptional(optional): { valueFrom+: { configMapKeyRef+: { optional: optional } } },
+ },
+ '#fieldRef':: d.obj(help="\"Selects a field of the pod: supports metadata.name, metadata.namespace, `metadata.labels['\u003cKEY\u003e']`, `metadata.annotations['\u003cKEY\u003e']`, spec.nodeName, spec.serviceAccountName, status.hostIP, status.podIP, status.podIPs.\""),
+ fieldRef: {
+ '#withApiVersion':: d.fn(help='"Version of the schema the FieldPath is written in terms of, defaults to \\"v1\\"."', args=[d.arg(name='apiVersion', type=d.T.string)]),
+ withApiVersion(apiVersion): { valueFrom+: { fieldRef+: { apiVersion: apiVersion } } },
+ '#withFieldPath':: d.fn(help='"Path of the field to select in the specified API version."', args=[d.arg(name='fieldPath', type=d.T.string)]),
+ withFieldPath(fieldPath): { valueFrom+: { fieldRef+: { fieldPath: fieldPath } } },
+ },
+ '#resourceFieldRef':: d.obj(help='"Selects a resource of the container: only resources limits and requests (limits.cpu, limits.memory, limits.ephemeral-storage, requests.cpu, requests.memory and requests.ephemeral-storage) are currently supported."'),
+ resourceFieldRef: {
+ '#withContainerName':: d.fn(help='"Container name: required for volumes, optional for env vars"', args=[d.arg(name='containerName', type=d.T.string)]),
+ withContainerName(containerName): { valueFrom+: { resourceFieldRef+: { containerName: containerName } } },
+ '#withDivisor':: d.fn(help='"Specifies the output format of the exposed resources, defaults to \\"1\\', args=[d.arg(name='divisor', type=d.T.any)]),
+ withDivisor(divisor): { valueFrom+: { resourceFieldRef+: { divisor: divisor } } },
+ '#withResource':: d.fn(help='"Required: resource to select"', args=[d.arg(name='resource', type=d.T.string)]),
+ withResource(resource): { valueFrom+: { resourceFieldRef+: { resource: resource } } },
+ },
+ '#secretKeyRef':: d.obj(help="\"Selects a key of a secret in the pod's namespace\""),
+ secretKeyRef: {
+ '#withKey':: d.fn(help='"The key of the secret to select from. Must be a valid secret key."', args=[d.arg(name='key', type=d.T.string)]),
+ withKey(key): { valueFrom+: { secretKeyRef+: { key: key } } },
+ '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { valueFrom+: { secretKeyRef+: { name: name } } },
+ '#withOptional':: d.fn(help='"Specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]),
+ withOptional(optional): { valueFrom+: { secretKeyRef+: { optional: optional } } },
+ },
+ },
+ '#withName':: d.fn(help='"Name of the environment variable. Must be a C_IDENTIFIER."', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { name: name },
+ '#withValue':: d.fn(help='"Variable references $(VAR_NAME) are expanded using the previously defined environment variables in the container and any service environment variables. If a variable cannot be resolved, the reference in the input string will be unchanged. Double $$ are reduced to a single $, which allows for escaping the $(VAR_NAME) syntax: i.e. \\"$$(VAR_NAME)\\" will produce the string literal \\"$(VAR_NAME)\\". Escaped references will never be expanded, regardless of whether the variable exists or not. Defaults to \\"\\"."', args=[d.arg(name='value', type=d.T.string)]),
+ withValue(value): { value: value },
+ },
+ '#envFrom':: d.obj(help='"List of sources to populate environment variables in the container. The keys defined within a source must be a C_IDENTIFIER. All invalid keys will be reported as an event when the container is starting. When a key exists in multiple sources, the value associated with the last source will take precedence. Values defined by an Env with a duplicate key will take precedence. Cannot be updated."'),
+ envFrom: {
+ '#configMapRef':: d.obj(help='"The ConfigMap to select from"'),
+ configMapRef: {
+ '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { configMapRef+: { name: name } },
+ '#withOptional':: d.fn(help='"Specify whether the ConfigMap must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]),
+ withOptional(optional): { configMapRef+: { optional: optional } },
+ },
+ '#secretRef':: d.obj(help='"The Secret to select from"'),
+ secretRef: {
+ '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { secretRef+: { name: name } },
+ '#withOptional':: d.fn(help='"Specify whether the Secret must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]),
+ withOptional(optional): { secretRef+: { optional: optional } },
+ },
+ '#withPrefix':: d.fn(help='"An optional identifier to prepend to each key in the ConfigMap. Must be a C_IDENTIFIER."', args=[d.arg(name='prefix', type=d.T.string)]),
+ withPrefix(prefix): { prefix: prefix },
+ },
+ '#lifecycle':: d.obj(help='"Actions that the management system should take in response to container lifecycle events. Cannot be updated."'),
+ lifecycle: {
+ '#postStart':: d.obj(help='"PostStart is called immediately after a container is created. If the handler fails, the container is terminated and restarted according to its restart policy. Other management of the container blocks until the hook completes. More info: https://kubernetes.io/docs/concepts/containers/container-lifecycle-hooks/#container-hooks"'),
+ postStart: {
+ '#exec':: d.obj(help='"Exec specifies the action to take."'),
+ exec: {
+ '#withCommand':: d.fn(help="\"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.\"", args=[d.arg(name='command', type=d.T.array)]),
+ withCommand(command): { lifecycle+: { postStart+: { exec+: { command: if std.isArray(v=command) then command else [command] } } } },
+ '#withCommandMixin':: d.fn(help="\"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='command', type=d.T.array)]),
+ withCommandMixin(command): { lifecycle+: { postStart+: { exec+: { command+: if std.isArray(v=command) then command else [command] } } } },
+ },
+ '#httpGet':: d.obj(help='"HTTPGet specifies the http request to perform."'),
+ httpGet: {
+ '#httpHeaders':: d.obj(help='"Custom headers to set in the request. HTTP allows repeated headers."'),
+ httpHeaders: {
+ '#withName':: d.fn(help='"The header field name. This will be canonicalized upon output, so case-variant names will be understood as the same header."', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { name: name },
+ '#withValue':: d.fn(help='"The header field value"', args=[d.arg(name='value', type=d.T.string)]),
+ withValue(value): { value: value },
+ },
+ '#withHost':: d.fn(help='"Host name to connect to, defaults to the pod IP. You probably want to set \\"Host\\" in httpHeaders instead."', args=[d.arg(name='host', type=d.T.string)]),
+ withHost(host): { lifecycle+: { postStart+: { httpGet+: { host: host } } } },
+ '#withHttpHeaders':: d.fn(help='"Custom headers to set in the request. HTTP allows repeated headers."', args=[d.arg(name='httpHeaders', type=d.T.array)]),
+ withHttpHeaders(httpHeaders): { lifecycle+: { postStart+: { httpGet+: { httpHeaders: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } } },
+ '#withHttpHeadersMixin':: d.fn(help='"Custom headers to set in the request. HTTP allows repeated headers."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='httpHeaders', type=d.T.array)]),
+ withHttpHeadersMixin(httpHeaders): { lifecycle+: { postStart+: { httpGet+: { httpHeaders+: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } } },
+ '#withPath':: d.fn(help='"Path to access on the HTTP server."', args=[d.arg(name='path', type=d.T.string)]),
+ withPath(path): { lifecycle+: { postStart+: { httpGet+: { path: path } } } },
+ '#withPort':: d.fn(help='"Name or number of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."', args=[d.arg(name='port', type=d.T.any)]),
+ withPort(port): { lifecycle+: { postStart+: { httpGet+: { port: port } } } },
+ '#withScheme':: d.fn(help='"Scheme to use for connecting to the host. Defaults to HTTP."', args=[d.arg(name='scheme', type=d.T.string)]),
+ withScheme(scheme): { lifecycle+: { postStart+: { httpGet+: { scheme: scheme } } } },
+ },
+ '#tcpSocket':: d.obj(help='"Deprecated. TCPSocket is NOT supported as a LifecycleHandler and kept for the backward compatibility. There are no validation of this field and lifecycle hooks will fail in runtime when tcp handler is specified."'),
+ tcpSocket: {
+ '#withHost':: d.fn(help='"Optional: Host name to connect to, defaults to the pod IP."', args=[d.arg(name='host', type=d.T.string)]),
+ withHost(host): { lifecycle+: { postStart+: { tcpSocket+: { host: host } } } },
+ '#withPort':: d.fn(help='"Number or name of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."', args=[d.arg(name='port', type=d.T.any)]),
+ withPort(port): { lifecycle+: { postStart+: { tcpSocket+: { port: port } } } },
+ },
+ },
+ '#preStop':: d.obj(help="\"PreStop is called immediately before a container is terminated due to an API request or management event such as liveness/startup probe failure, preemption, resource contention, etc. The handler is not called if the container crashes or exits. The Pod's termination grace period countdown begins before the PreStop hook is executed. Regardless of the outcome of the handler, the container will eventually terminate within the Pod's termination grace period (unless delayed by finalizers). Other management of the container blocks until the hook completes or until the termination grace period is reached. More info: https://kubernetes.io/docs/concepts/containers/container-lifecycle-hooks/#container-hooks\""),
+ preStop: {
+ '#exec':: d.obj(help='"Exec specifies the action to take."'),
+ exec: {
+ '#withCommand':: d.fn(help="\"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.\"", args=[d.arg(name='command', type=d.T.array)]),
+ withCommand(command): { lifecycle+: { preStop+: { exec+: { command: if std.isArray(v=command) then command else [command] } } } },
+ '#withCommandMixin':: d.fn(help="\"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='command', type=d.T.array)]),
+ withCommandMixin(command): { lifecycle+: { preStop+: { exec+: { command+: if std.isArray(v=command) then command else [command] } } } },
+ },
+ '#httpGet':: d.obj(help='"HTTPGet specifies the http request to perform."'),
+ httpGet: {
+ '#httpHeaders':: d.obj(help='"Custom headers to set in the request. HTTP allows repeated headers."'),
+ httpHeaders: {
+ '#withName':: d.fn(help='"The header field name. This will be canonicalized upon output, so case-variant names will be understood as the same header."', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { name: name },
+ '#withValue':: d.fn(help='"The header field value"', args=[d.arg(name='value', type=d.T.string)]),
+ withValue(value): { value: value },
+ },
+ '#withHost':: d.fn(help='"Host name to connect to, defaults to the pod IP. You probably want to set \\"Host\\" in httpHeaders instead."', args=[d.arg(name='host', type=d.T.string)]),
+ withHost(host): { lifecycle+: { preStop+: { httpGet+: { host: host } } } },
+ '#withHttpHeaders':: d.fn(help='"Custom headers to set in the request. HTTP allows repeated headers."', args=[d.arg(name='httpHeaders', type=d.T.array)]),
+ withHttpHeaders(httpHeaders): { lifecycle+: { preStop+: { httpGet+: { httpHeaders: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } } },
+ '#withHttpHeadersMixin':: d.fn(help='"Custom headers to set in the request. HTTP allows repeated headers."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='httpHeaders', type=d.T.array)]),
+ withHttpHeadersMixin(httpHeaders): { lifecycle+: { preStop+: { httpGet+: { httpHeaders+: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } } },
+ '#withPath':: d.fn(help='"Path to access on the HTTP server."', args=[d.arg(name='path', type=d.T.string)]),
+ withPath(path): { lifecycle+: { preStop+: { httpGet+: { path: path } } } },
+ '#withPort':: d.fn(help='"Name or number of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."', args=[d.arg(name='port', type=d.T.any)]),
+ withPort(port): { lifecycle+: { preStop+: { httpGet+: { port: port } } } },
+ '#withScheme':: d.fn(help='"Scheme to use for connecting to the host. Defaults to HTTP."', args=[d.arg(name='scheme', type=d.T.string)]),
+ withScheme(scheme): { lifecycle+: { preStop+: { httpGet+: { scheme: scheme } } } },
+ },
+ '#tcpSocket':: d.obj(help='"Deprecated. TCPSocket is NOT supported as a LifecycleHandler and kept for the backward compatibility. There are no validation of this field and lifecycle hooks will fail in runtime when tcp handler is specified."'),
+ tcpSocket: {
+ '#withHost':: d.fn(help='"Optional: Host name to connect to, defaults to the pod IP."', args=[d.arg(name='host', type=d.T.string)]),
+ withHost(host): { lifecycle+: { preStop+: { tcpSocket+: { host: host } } } },
+ '#withPort':: d.fn(help='"Number or name of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."', args=[d.arg(name='port', type=d.T.any)]),
+ withPort(port): { lifecycle+: { preStop+: { tcpSocket+: { port: port } } } },
+ },
+ },
+ },
+ '#livenessProbe':: d.obj(help='"Periodic probe of container liveness. Container will be restarted if the probe fails. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes"'),
+ livenessProbe: {
+ '#exec':: d.obj(help='"Exec specifies the action to take."'),
+ exec: {
+ '#withCommand':: d.fn(help="\"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.\"", args=[d.arg(name='command', type=d.T.array)]),
+ withCommand(command): { livenessProbe+: { exec+: { command: if std.isArray(v=command) then command else [command] } } },
+ '#withCommandMixin':: d.fn(help="\"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='command', type=d.T.array)]),
+ withCommandMixin(command): { livenessProbe+: { exec+: { command+: if std.isArray(v=command) then command else [command] } } },
+ },
+ '#grpc':: d.obj(help='"GRPC specifies an action involving a GRPC port."'),
+ grpc: {
+ '#withPort':: d.fn(help='"Port number of the gRPC service. Number must be in the range 1 to 65535."', args=[d.arg(name='port', type=d.T.integer)]),
+ withPort(port): { livenessProbe+: { grpc+: { port: port } } },
+ '#withService':: d.fn(help='"Service is the name of the service to place in the gRPC HealthCheckRequest (see https://github.com/grpc/grpc/blob/master/doc/health-checking.md). \\n If this is not specified, the default behavior is defined by gRPC."', args=[d.arg(name='service', type=d.T.string)]),
+ withService(service): { livenessProbe+: { grpc+: { service: service } } },
+ },
+ '#httpGet':: d.obj(help='"HTTPGet specifies the http request to perform."'),
+ httpGet: {
+ '#httpHeaders':: d.obj(help='"Custom headers to set in the request. HTTP allows repeated headers."'),
+ httpHeaders: {
+ '#withName':: d.fn(help='"The header field name. This will be canonicalized upon output, so case-variant names will be understood as the same header."', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { name: name },
+ '#withValue':: d.fn(help='"The header field value"', args=[d.arg(name='value', type=d.T.string)]),
+ withValue(value): { value: value },
+ },
+ '#withHost':: d.fn(help='"Host name to connect to, defaults to the pod IP. You probably want to set \\"Host\\" in httpHeaders instead."', args=[d.arg(name='host', type=d.T.string)]),
+ withHost(host): { livenessProbe+: { httpGet+: { host: host } } },
+ '#withHttpHeaders':: d.fn(help='"Custom headers to set in the request. HTTP allows repeated headers."', args=[d.arg(name='httpHeaders', type=d.T.array)]),
+ withHttpHeaders(httpHeaders): { livenessProbe+: { httpGet+: { httpHeaders: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } },
+ '#withHttpHeadersMixin':: d.fn(help='"Custom headers to set in the request. HTTP allows repeated headers."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='httpHeaders', type=d.T.array)]),
+ withHttpHeadersMixin(httpHeaders): { livenessProbe+: { httpGet+: { httpHeaders+: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } },
+ '#withPath':: d.fn(help='"Path to access on the HTTP server."', args=[d.arg(name='path', type=d.T.string)]),
+ withPath(path): { livenessProbe+: { httpGet+: { path: path } } },
+ '#withPort':: d.fn(help='"Name or number of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."', args=[d.arg(name='port', type=d.T.any)]),
+ withPort(port): { livenessProbe+: { httpGet+: { port: port } } },
+ '#withScheme':: d.fn(help='"Scheme to use for connecting to the host. Defaults to HTTP."', args=[d.arg(name='scheme', type=d.T.string)]),
+ withScheme(scheme): { livenessProbe+: { httpGet+: { scheme: scheme } } },
+ },
+ '#tcpSocket':: d.obj(help='"TCPSocket specifies an action involving a TCP port."'),
+ tcpSocket: {
+ '#withHost':: d.fn(help='"Optional: Host name to connect to, defaults to the pod IP."', args=[d.arg(name='host', type=d.T.string)]),
+ withHost(host): { livenessProbe+: { tcpSocket+: { host: host } } },
+ '#withPort':: d.fn(help='"Number or name of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."', args=[d.arg(name='port', type=d.T.any)]),
+ withPort(port): { livenessProbe+: { tcpSocket+: { port: port } } },
+ },
+ '#withFailureThreshold':: d.fn(help='"Minimum consecutive failures for the probe to be considered failed after having succeeded. Defaults to 3. Minimum value is 1."', args=[d.arg(name='failureThreshold', type=d.T.integer)]),
+ withFailureThreshold(failureThreshold): { livenessProbe+: { failureThreshold: failureThreshold } },
+ '#withInitialDelaySeconds':: d.fn(help='"Number of seconds after the container has started before liveness probes are initiated. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes"', args=[d.arg(name='initialDelaySeconds', type=d.T.integer)]),
+ withInitialDelaySeconds(initialDelaySeconds): { livenessProbe+: { initialDelaySeconds: initialDelaySeconds } },
+ '#withPeriodSeconds':: d.fn(help='"How often (in seconds) to perform the probe. Default to 10 seconds. Minimum value is 1."', args=[d.arg(name='periodSeconds', type=d.T.integer)]),
+ withPeriodSeconds(periodSeconds): { livenessProbe+: { periodSeconds: periodSeconds } },
+ '#withSuccessThreshold':: d.fn(help='"Minimum consecutive successes for the probe to be considered successful after having failed. Defaults to 1. Must be 1 for liveness and startup. Minimum value is 1."', args=[d.arg(name='successThreshold', type=d.T.integer)]),
+ withSuccessThreshold(successThreshold): { livenessProbe+: { successThreshold: successThreshold } },
+ '#withTerminationGracePeriodSeconds':: d.fn(help="\"Optional duration in seconds the pod needs to terminate gracefully upon probe failure. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. If this value is nil, the pod's terminationGracePeriodSeconds will be used. Otherwise, this value overrides the value provided by the pod spec. Value must be non-negative integer. The value zero indicates stop immediately via the kill signal (no opportunity to shut down). This is a beta field and requires enabling ProbeTerminationGracePeriod feature gate. Minimum value is 1. spec.terminationGracePeriodSeconds is used if unset.\"", args=[d.arg(name='terminationGracePeriodSeconds', type=d.T.integer)]),
+ withTerminationGracePeriodSeconds(terminationGracePeriodSeconds): { livenessProbe+: { terminationGracePeriodSeconds: terminationGracePeriodSeconds } },
+ '#withTimeoutSeconds':: d.fn(help='"Number of seconds after which the probe times out. Defaults to 1 second. Minimum value is 1. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes"', args=[d.arg(name='timeoutSeconds', type=d.T.integer)]),
+ withTimeoutSeconds(timeoutSeconds): { livenessProbe+: { timeoutSeconds: timeoutSeconds } },
+ },
+ '#ports':: d.obj(help='"List of ports to expose from the container. Not specifying a port here DOES NOT prevent that port from being exposed. Any port which is listening on the default \\"0.0.0.0\\" address inside a container will be accessible from the network. Modifying this array with strategic merge patch may corrupt the data. For more information See https://github.com/kubernetes/kubernetes/issues/108255. Cannot be updated."'),
+ ports: {
+ '#withContainerPort':: d.fn(help="\"Number of port to expose on the pod's IP address. This must be a valid port number, 0 \u003c x \u003c 65536.\"", args=[d.arg(name='containerPort', type=d.T.integer)]),
+ withContainerPort(containerPort): { containerPort: containerPort },
+ '#withHostIP':: d.fn(help='"What host IP to bind the external port to."', args=[d.arg(name='hostIP', type=d.T.string)]),
+ withHostIP(hostIP): { hostIP: hostIP },
+ '#withHostPort':: d.fn(help='"Number of port to expose on the host. If specified, this must be a valid port number, 0 < x < 65536. If HostNetwork is specified, this must match ContainerPort. Most containers do not need this."', args=[d.arg(name='hostPort', type=d.T.integer)]),
+ withHostPort(hostPort): { hostPort: hostPort },
+ '#withName':: d.fn(help='"If specified, this must be an IANA_SVC_NAME and unique within the pod. Each named port in a pod must have a unique name. Name for the port that can be referred to by services."', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { name: name },
+ '#withProtocol':: d.fn(help='"Protocol for port. Must be UDP, TCP, or SCTP. Defaults to \\"TCP\\"."', args=[d.arg(name='protocol', type=d.T.string)]),
+ withProtocol(protocol): { protocol: protocol },
+ },
+ '#readinessProbe':: d.obj(help='"Periodic probe of container service readiness. Container will be removed from service endpoints if the probe fails. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes"'),
+ readinessProbe: {
+ '#exec':: d.obj(help='"Exec specifies the action to take."'),
+ exec: {
+ '#withCommand':: d.fn(help="\"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.\"", args=[d.arg(name='command', type=d.T.array)]),
+ withCommand(command): { readinessProbe+: { exec+: { command: if std.isArray(v=command) then command else [command] } } },
+ '#withCommandMixin':: d.fn(help="\"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='command', type=d.T.array)]),
+ withCommandMixin(command): { readinessProbe+: { exec+: { command+: if std.isArray(v=command) then command else [command] } } },
+ },
+ '#grpc':: d.obj(help='"GRPC specifies an action involving a GRPC port."'),
+ grpc: {
+ '#withPort':: d.fn(help='"Port number of the gRPC service. Number must be in the range 1 to 65535."', args=[d.arg(name='port', type=d.T.integer)]),
+ withPort(port): { readinessProbe+: { grpc+: { port: port } } },
+ '#withService':: d.fn(help='"Service is the name of the service to place in the gRPC HealthCheckRequest (see https://github.com/grpc/grpc/blob/master/doc/health-checking.md). \\n If this is not specified, the default behavior is defined by gRPC."', args=[d.arg(name='service', type=d.T.string)]),
+ withService(service): { readinessProbe+: { grpc+: { service: service } } },
+ },
+ '#httpGet':: d.obj(help='"HTTPGet specifies the http request to perform."'),
+ httpGet: {
+ '#httpHeaders':: d.obj(help='"Custom headers to set in the request. HTTP allows repeated headers."'),
+ httpHeaders: {
+ '#withName':: d.fn(help='"The header field name. This will be canonicalized upon output, so case-variant names will be understood as the same header."', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { name: name },
+ '#withValue':: d.fn(help='"The header field value"', args=[d.arg(name='value', type=d.T.string)]),
+ withValue(value): { value: value },
+ },
+ '#withHost':: d.fn(help='"Host name to connect to, defaults to the pod IP. You probably want to set \\"Host\\" in httpHeaders instead."', args=[d.arg(name='host', type=d.T.string)]),
+ withHost(host): { readinessProbe+: { httpGet+: { host: host } } },
+ '#withHttpHeaders':: d.fn(help='"Custom headers to set in the request. HTTP allows repeated headers."', args=[d.arg(name='httpHeaders', type=d.T.array)]),
+ withHttpHeaders(httpHeaders): { readinessProbe+: { httpGet+: { httpHeaders: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } },
+ '#withHttpHeadersMixin':: d.fn(help='"Custom headers to set in the request. HTTP allows repeated headers."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='httpHeaders', type=d.T.array)]),
+ withHttpHeadersMixin(httpHeaders): { readinessProbe+: { httpGet+: { httpHeaders+: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } },
+ '#withPath':: d.fn(help='"Path to access on the HTTP server."', args=[d.arg(name='path', type=d.T.string)]),
+ withPath(path): { readinessProbe+: { httpGet+: { path: path } } },
+ '#withPort':: d.fn(help='"Name or number of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."', args=[d.arg(name='port', type=d.T.any)]),
+ withPort(port): { readinessProbe+: { httpGet+: { port: port } } },
+ '#withScheme':: d.fn(help='"Scheme to use for connecting to the host. Defaults to HTTP."', args=[d.arg(name='scheme', type=d.T.string)]),
+ withScheme(scheme): { readinessProbe+: { httpGet+: { scheme: scheme } } },
+ },
+ '#tcpSocket':: d.obj(help='"TCPSocket specifies an action involving a TCP port."'),
+ tcpSocket: {
+ '#withHost':: d.fn(help='"Optional: Host name to connect to, defaults to the pod IP."', args=[d.arg(name='host', type=d.T.string)]),
+ withHost(host): { readinessProbe+: { tcpSocket+: { host: host } } },
+ '#withPort':: d.fn(help='"Number or name of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."', args=[d.arg(name='port', type=d.T.any)]),
+ withPort(port): { readinessProbe+: { tcpSocket+: { port: port } } },
+ },
+ '#withFailureThreshold':: d.fn(help='"Minimum consecutive failures for the probe to be considered failed after having succeeded. Defaults to 3. Minimum value is 1."', args=[d.arg(name='failureThreshold', type=d.T.integer)]),
+ withFailureThreshold(failureThreshold): { readinessProbe+: { failureThreshold: failureThreshold } },
+ '#withInitialDelaySeconds':: d.fn(help='"Number of seconds after the container has started before liveness probes are initiated. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes"', args=[d.arg(name='initialDelaySeconds', type=d.T.integer)]),
+ withInitialDelaySeconds(initialDelaySeconds): { readinessProbe+: { initialDelaySeconds: initialDelaySeconds } },
+ '#withPeriodSeconds':: d.fn(help='"How often (in seconds) to perform the probe. Default to 10 seconds. Minimum value is 1."', args=[d.arg(name='periodSeconds', type=d.T.integer)]),
+ withPeriodSeconds(periodSeconds): { readinessProbe+: { periodSeconds: periodSeconds } },
+ '#withSuccessThreshold':: d.fn(help='"Minimum consecutive successes for the probe to be considered successful after having failed. Defaults to 1. Must be 1 for liveness and startup. Minimum value is 1."', args=[d.arg(name='successThreshold', type=d.T.integer)]),
+ withSuccessThreshold(successThreshold): { readinessProbe+: { successThreshold: successThreshold } },
+ '#withTerminationGracePeriodSeconds':: d.fn(help="\"Optional duration in seconds the pod needs to terminate gracefully upon probe failure. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. If this value is nil, the pod's terminationGracePeriodSeconds will be used. Otherwise, this value overrides the value provided by the pod spec. Value must be non-negative integer. The value zero indicates stop immediately via the kill signal (no opportunity to shut down). This is a beta field and requires enabling ProbeTerminationGracePeriod feature gate. Minimum value is 1. spec.terminationGracePeriodSeconds is used if unset.\"", args=[d.arg(name='terminationGracePeriodSeconds', type=d.T.integer)]),
+ withTerminationGracePeriodSeconds(terminationGracePeriodSeconds): { readinessProbe+: { terminationGracePeriodSeconds: terminationGracePeriodSeconds } },
+ '#withTimeoutSeconds':: d.fn(help='"Number of seconds after which the probe times out. Defaults to 1 second. Minimum value is 1. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes"', args=[d.arg(name='timeoutSeconds', type=d.T.integer)]),
+ withTimeoutSeconds(timeoutSeconds): { readinessProbe+: { timeoutSeconds: timeoutSeconds } },
+ },
+ '#resizePolicy':: d.obj(help='"Resources resize policy for the container."'),
+ resizePolicy: {
+ '#withResourceName':: d.fn(help='"Name of the resource to which this resource resize policy applies. Supported values: cpu, memory."', args=[d.arg(name='resourceName', type=d.T.string)]),
+ withResourceName(resourceName): { resourceName: resourceName },
+ '#withRestartPolicy':: d.fn(help='"Restart policy to apply when specified resource is resized. If not specified, it defaults to NotRequired."', args=[d.arg(name='restartPolicy', type=d.T.string)]),
+ withRestartPolicy(restartPolicy): { restartPolicy: restartPolicy },
+ },
+ '#resources':: d.obj(help='"Compute Resources required by this container. Cannot be updated. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"'),
+ resources: {
+ '#claims':: d.obj(help='"Claims lists the names of resources, defined in spec.resourceClaims, that are used by this container. \\n This is an alpha field and requires enabling the DynamicResourceAllocation feature gate. \\n This field is immutable. It can only be set for containers."'),
+ claims: {
+ '#withName':: d.fn(help='"Name must match the name of one entry in pod.spec.resourceClaims of the Pod where this field is used. It makes that resource available inside a container."', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { name: name },
+ },
+ '#withClaims':: d.fn(help='"Claims lists the names of resources, defined in spec.resourceClaims, that are used by this container. \\n This is an alpha field and requires enabling the DynamicResourceAllocation feature gate. \\n This field is immutable. It can only be set for containers."', args=[d.arg(name='claims', type=d.T.array)]),
+ withClaims(claims): { resources+: { claims: if std.isArray(v=claims) then claims else [claims] } },
+ '#withClaimsMixin':: d.fn(help='"Claims lists the names of resources, defined in spec.resourceClaims, that are used by this container. \\n This is an alpha field and requires enabling the DynamicResourceAllocation feature gate. \\n This field is immutable. It can only be set for containers."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='claims', type=d.T.array)]),
+ withClaimsMixin(claims): { resources+: { claims+: if std.isArray(v=claims) then claims else [claims] } },
+ '#withLimits':: d.fn(help='"Limits describes the maximum amount of compute resources allowed. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"', args=[d.arg(name='limits', type=d.T.object)]),
+ withLimits(limits): { resources+: { limits: limits } },
+ '#withLimitsMixin':: d.fn(help='"Limits describes the maximum amount of compute resources allowed. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='limits', type=d.T.object)]),
+ withLimitsMixin(limits): { resources+: { limits+: limits } },
+ '#withRequests':: d.fn(help='"Requests describes the minimum amount of compute resources required. If Requests is omitted for a container, it defaults to Limits if that is explicitly specified, otherwise to an implementation-defined value. Requests cannot exceed Limits. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"', args=[d.arg(name='requests', type=d.T.object)]),
+ withRequests(requests): { resources+: { requests: requests } },
+ '#withRequestsMixin':: d.fn(help='"Requests describes the minimum amount of compute resources required. If Requests is omitted for a container, it defaults to Limits if that is explicitly specified, otherwise to an implementation-defined value. Requests cannot exceed Limits. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='requests', type=d.T.object)]),
+ withRequestsMixin(requests): { resources+: { requests+: requests } },
+ },
+ '#securityContext':: d.obj(help='"SecurityContext defines the security options the container should be run with. If set, the fields of SecurityContext override the equivalent fields of PodSecurityContext. More info: https://kubernetes.io/docs/tasks/configure-pod-container/security-context/"'),
+ securityContext: {
+ '#capabilities':: d.obj(help='"The capabilities to add/drop when running containers. Defaults to the default set of capabilities granted by the container runtime. Note that this field cannot be set when spec.os.name is windows."'),
+ capabilities: {
+ '#withAdd':: d.fn(help='"Added capabilities"', args=[d.arg(name='add', type=d.T.array)]),
+ withAdd(add): { securityContext+: { capabilities+: { add: if std.isArray(v=add) then add else [add] } } },
+ '#withAddMixin':: d.fn(help='"Added capabilities"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='add', type=d.T.array)]),
+ withAddMixin(add): { securityContext+: { capabilities+: { add+: if std.isArray(v=add) then add else [add] } } },
+ '#withDrop':: d.fn(help='"Removed capabilities"', args=[d.arg(name='drop', type=d.T.array)]),
+ withDrop(drop): { securityContext+: { capabilities+: { drop: if std.isArray(v=drop) then drop else [drop] } } },
+ '#withDropMixin':: d.fn(help='"Removed capabilities"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='drop', type=d.T.array)]),
+ withDropMixin(drop): { securityContext+: { capabilities+: { drop+: if std.isArray(v=drop) then drop else [drop] } } },
+ },
+ '#seLinuxOptions':: d.obj(help='"The SELinux context to be applied to the container. If unspecified, the container runtime will allocate a random SELinux context for each container. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. Note that this field cannot be set when spec.os.name is windows."'),
+ seLinuxOptions: {
+ '#withLevel':: d.fn(help='"Level is SELinux level label that applies to the container."', args=[d.arg(name='level', type=d.T.string)]),
+ withLevel(level): { securityContext+: { seLinuxOptions+: { level: level } } },
+ '#withRole':: d.fn(help='"Role is a SELinux role label that applies to the container."', args=[d.arg(name='role', type=d.T.string)]),
+ withRole(role): { securityContext+: { seLinuxOptions+: { role: role } } },
+ '#withType':: d.fn(help='"Type is a SELinux type label that applies to the container."', args=[d.arg(name='type', type=d.T.string)]),
+ withType(type): { securityContext+: { seLinuxOptions+: { type: type } } },
+ '#withUser':: d.fn(help='"User is a SELinux user label that applies to the container."', args=[d.arg(name='user', type=d.T.string)]),
+ withUser(user): { securityContext+: { seLinuxOptions+: { user: user } } },
+ },
+ '#seccompProfile':: d.obj(help='"The seccomp options to use by this container. If seccomp options are provided at both the pod & container level, the container options override the pod options. Note that this field cannot be set when spec.os.name is windows."'),
+ seccompProfile: {
+ '#withLocalhostProfile':: d.fn(help="\"localhostProfile indicates a profile defined in a file on the node should be used. The profile must be preconfigured on the node to work. Must be a descending path, relative to the kubelet's configured seccomp profile location. Must be set if type is \\\"Localhost\\\". Must NOT be set for any other type.\"", args=[d.arg(name='localhostProfile', type=d.T.string)]),
+ withLocalhostProfile(localhostProfile): { securityContext+: { seccompProfile+: { localhostProfile: localhostProfile } } },
+ '#withType':: d.fn(help='"type indicates which kind of seccomp profile will be applied. Valid options are: \\n Localhost - a profile defined in a file on the node should be used. RuntimeDefault - the container runtime default profile should be used. Unconfined - no profile should be applied."', args=[d.arg(name='type', type=d.T.string)]),
+ withType(type): { securityContext+: { seccompProfile+: { type: type } } },
+ },
+ '#windowsOptions':: d.obj(help='"The Windows specific settings applied to all containers. If unspecified, the options from the PodSecurityContext will be used. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. Note that this field cannot be set when spec.os.name is linux."'),
+ windowsOptions: {
+ '#withGmsaCredentialSpec':: d.fn(help='"GMSACredentialSpec is where the GMSA admission webhook (https://github.com/kubernetes-sigs/windows-gmsa) inlines the contents of the GMSA credential spec named by the GMSACredentialSpecName field."', args=[d.arg(name='gmsaCredentialSpec', type=d.T.string)]),
+ withGmsaCredentialSpec(gmsaCredentialSpec): { securityContext+: { windowsOptions+: { gmsaCredentialSpec: gmsaCredentialSpec } } },
+ '#withGmsaCredentialSpecName':: d.fn(help='"GMSACredentialSpecName is the name of the GMSA credential spec to use."', args=[d.arg(name='gmsaCredentialSpecName', type=d.T.string)]),
+ withGmsaCredentialSpecName(gmsaCredentialSpecName): { securityContext+: { windowsOptions+: { gmsaCredentialSpecName: gmsaCredentialSpecName } } },
+ '#withHostProcess':: d.fn(help="\"HostProcess determines if a container should be run as a 'Host Process' container. All of a Pod's containers must have the same effective HostProcess value (it is not allowed to have a mix of HostProcess containers and non-HostProcess containers). In addition, if HostProcess is true then HostNetwork must also be set to true.\"", args=[d.arg(name='hostProcess', type=d.T.boolean)]),
+ withHostProcess(hostProcess): { securityContext+: { windowsOptions+: { hostProcess: hostProcess } } },
+ '#withRunAsUserName':: d.fn(help='"The UserName in Windows to run the entrypoint of the container process. Defaults to the user specified in image metadata if unspecified. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence."', args=[d.arg(name='runAsUserName', type=d.T.string)]),
+ withRunAsUserName(runAsUserName): { securityContext+: { windowsOptions+: { runAsUserName: runAsUserName } } },
+ },
+ '#withAllowPrivilegeEscalation':: d.fn(help='"AllowPrivilegeEscalation controls whether a process can gain more privileges than its parent process. This bool directly controls if the no_new_privs flag will be set on the container process. AllowPrivilegeEscalation is true always when the container is: 1) run as Privileged 2) has CAP_SYS_ADMIN Note that this field cannot be set when spec.os.name is windows."', args=[d.arg(name='allowPrivilegeEscalation', type=d.T.boolean)]),
+ withAllowPrivilegeEscalation(allowPrivilegeEscalation): { securityContext+: { allowPrivilegeEscalation: allowPrivilegeEscalation } },
+ '#withPrivileged':: d.fn(help='"Run container in privileged mode. Processes in privileged containers are essentially equivalent to root on the host. Defaults to false. Note that this field cannot be set when spec.os.name is windows."', args=[d.arg(name='privileged', type=d.T.boolean)]),
+ withPrivileged(privileged): { securityContext+: { privileged: privileged } },
+ '#withProcMount':: d.fn(help='"procMount denotes the type of proc mount to use for the containers. The default is DefaultProcMount which uses the container runtime defaults for readonly paths and masked paths. This requires the ProcMountType feature flag to be enabled. Note that this field cannot be set when spec.os.name is windows."', args=[d.arg(name='procMount', type=d.T.string)]),
+ withProcMount(procMount): { securityContext+: { procMount: procMount } },
+ '#withReadOnlyRootFilesystem':: d.fn(help='"Whether this container has a read-only root filesystem. Default is false. Note that this field cannot be set when spec.os.name is windows."', args=[d.arg(name='readOnlyRootFilesystem', type=d.T.boolean)]),
+ withReadOnlyRootFilesystem(readOnlyRootFilesystem): { securityContext+: { readOnlyRootFilesystem: readOnlyRootFilesystem } },
+ '#withRunAsGroup':: d.fn(help='"The GID to run the entrypoint of the container process. Uses runtime default if unset. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. Note that this field cannot be set when spec.os.name is windows."', args=[d.arg(name='runAsGroup', type=d.T.integer)]),
+ withRunAsGroup(runAsGroup): { securityContext+: { runAsGroup: runAsGroup } },
+ '#withRunAsNonRoot':: d.fn(help='"Indicates that the container must run as a non-root user. If true, the Kubelet will validate the image at runtime to ensure that it does not run as UID 0 (root) and fail to start the container if it does. If unset or false, no such validation will be performed. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence."', args=[d.arg(name='runAsNonRoot', type=d.T.boolean)]),
+ withRunAsNonRoot(runAsNonRoot): { securityContext+: { runAsNonRoot: runAsNonRoot } },
+ '#withRunAsUser':: d.fn(help='"The UID to run the entrypoint of the container process. Defaults to user specified in image metadata if unspecified. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. Note that this field cannot be set when spec.os.name is windows."', args=[d.arg(name='runAsUser', type=d.T.integer)]),
+ withRunAsUser(runAsUser): { securityContext+: { runAsUser: runAsUser } },
+ },
+ '#startupProbe':: d.obj(help="\"StartupProbe indicates that the Pod has successfully initialized. If specified, no other probes are executed until this completes successfully. If this probe fails, the Pod will be restarted, just as if the livenessProbe failed. This can be used to provide different probe parameters at the beginning of a Pod's lifecycle, when it might take a long time to load data or warm a cache, than during steady-state operation. This cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes\""),
+ startupProbe: {
+ '#exec':: d.obj(help='"Exec specifies the action to take."'),
+ exec: {
+ '#withCommand':: d.fn(help="\"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.\"", args=[d.arg(name='command', type=d.T.array)]),
+ withCommand(command): { startupProbe+: { exec+: { command: if std.isArray(v=command) then command else [command] } } },
+ '#withCommandMixin':: d.fn(help="\"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='command', type=d.T.array)]),
+ withCommandMixin(command): { startupProbe+: { exec+: { command+: if std.isArray(v=command) then command else [command] } } },
+ },
+ '#grpc':: d.obj(help='"GRPC specifies an action involving a GRPC port."'),
+ grpc: {
+ '#withPort':: d.fn(help='"Port number of the gRPC service. Number must be in the range 1 to 65535."', args=[d.arg(name='port', type=d.T.integer)]),
+ withPort(port): { startupProbe+: { grpc+: { port: port } } },
+ '#withService':: d.fn(help='"Service is the name of the service to place in the gRPC HealthCheckRequest (see https://github.com/grpc/grpc/blob/master/doc/health-checking.md). \\n If this is not specified, the default behavior is defined by gRPC."', args=[d.arg(name='service', type=d.T.string)]),
+ withService(service): { startupProbe+: { grpc+: { service: service } } },
+ },
+ '#httpGet':: d.obj(help='"HTTPGet specifies the http request to perform."'),
+ httpGet: {
+ '#httpHeaders':: d.obj(help='"Custom headers to set in the request. HTTP allows repeated headers."'),
+ httpHeaders: {
+ '#withName':: d.fn(help='"The header field name. This will be canonicalized upon output, so case-variant names will be understood as the same header."', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { name: name },
+ '#withValue':: d.fn(help='"The header field value"', args=[d.arg(name='value', type=d.T.string)]),
+ withValue(value): { value: value },
+ },
+ '#withHost':: d.fn(help='"Host name to connect to, defaults to the pod IP. You probably want to set \\"Host\\" in httpHeaders instead."', args=[d.arg(name='host', type=d.T.string)]),
+ withHost(host): { startupProbe+: { httpGet+: { host: host } } },
+ '#withHttpHeaders':: d.fn(help='"Custom headers to set in the request. HTTP allows repeated headers."', args=[d.arg(name='httpHeaders', type=d.T.array)]),
+ withHttpHeaders(httpHeaders): { startupProbe+: { httpGet+: { httpHeaders: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } },
+ '#withHttpHeadersMixin':: d.fn(help='"Custom headers to set in the request. HTTP allows repeated headers."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='httpHeaders', type=d.T.array)]),
+ withHttpHeadersMixin(httpHeaders): { startupProbe+: { httpGet+: { httpHeaders+: if std.isArray(v=httpHeaders) then httpHeaders else [httpHeaders] } } },
+ '#withPath':: d.fn(help='"Path to access on the HTTP server."', args=[d.arg(name='path', type=d.T.string)]),
+ withPath(path): { startupProbe+: { httpGet+: { path: path } } },
+ '#withPort':: d.fn(help='"Name or number of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."', args=[d.arg(name='port', type=d.T.any)]),
+ withPort(port): { startupProbe+: { httpGet+: { port: port } } },
+ '#withScheme':: d.fn(help='"Scheme to use for connecting to the host. Defaults to HTTP."', args=[d.arg(name='scheme', type=d.T.string)]),
+ withScheme(scheme): { startupProbe+: { httpGet+: { scheme: scheme } } },
+ },
+ '#tcpSocket':: d.obj(help='"TCPSocket specifies an action involving a TCP port."'),
+ tcpSocket: {
+ '#withHost':: d.fn(help='"Optional: Host name to connect to, defaults to the pod IP."', args=[d.arg(name='host', type=d.T.string)]),
+ withHost(host): { startupProbe+: { tcpSocket+: { host: host } } },
+ '#withPort':: d.fn(help='"Number or name of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."', args=[d.arg(name='port', type=d.T.any)]),
+ withPort(port): { startupProbe+: { tcpSocket+: { port: port } } },
+ },
+ '#withFailureThreshold':: d.fn(help='"Minimum consecutive failures for the probe to be considered failed after having succeeded. Defaults to 3. Minimum value is 1."', args=[d.arg(name='failureThreshold', type=d.T.integer)]),
+ withFailureThreshold(failureThreshold): { startupProbe+: { failureThreshold: failureThreshold } },
+ '#withInitialDelaySeconds':: d.fn(help='"Number of seconds after the container has started before liveness probes are initiated. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes"', args=[d.arg(name='initialDelaySeconds', type=d.T.integer)]),
+ withInitialDelaySeconds(initialDelaySeconds): { startupProbe+: { initialDelaySeconds: initialDelaySeconds } },
+ '#withPeriodSeconds':: d.fn(help='"How often (in seconds) to perform the probe. Default to 10 seconds. Minimum value is 1."', args=[d.arg(name='periodSeconds', type=d.T.integer)]),
+ withPeriodSeconds(periodSeconds): { startupProbe+: { periodSeconds: periodSeconds } },
+ '#withSuccessThreshold':: d.fn(help='"Minimum consecutive successes for the probe to be considered successful after having failed. Defaults to 1. Must be 1 for liveness and startup. Minimum value is 1."', args=[d.arg(name='successThreshold', type=d.T.integer)]),
+ withSuccessThreshold(successThreshold): { startupProbe+: { successThreshold: successThreshold } },
+ '#withTerminationGracePeriodSeconds':: d.fn(help="\"Optional duration in seconds the pod needs to terminate gracefully upon probe failure. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. If this value is nil, the pod's terminationGracePeriodSeconds will be used. Otherwise, this value overrides the value provided by the pod spec. Value must be non-negative integer. The value zero indicates stop immediately via the kill signal (no opportunity to shut down). This is a beta field and requires enabling ProbeTerminationGracePeriod feature gate. Minimum value is 1. spec.terminationGracePeriodSeconds is used if unset.\"", args=[d.arg(name='terminationGracePeriodSeconds', type=d.T.integer)]),
+ withTerminationGracePeriodSeconds(terminationGracePeriodSeconds): { startupProbe+: { terminationGracePeriodSeconds: terminationGracePeriodSeconds } },
+ '#withTimeoutSeconds':: d.fn(help='"Number of seconds after which the probe times out. Defaults to 1 second. Minimum value is 1. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes"', args=[d.arg(name='timeoutSeconds', type=d.T.integer)]),
+ withTimeoutSeconds(timeoutSeconds): { startupProbe+: { timeoutSeconds: timeoutSeconds } },
+ },
+ '#volumeDevices':: d.obj(help='"volumeDevices is the list of block devices to be used by the container."'),
+ volumeDevices: {
+ '#withDevicePath':: d.fn(help='"devicePath is the path inside of the container that the device will be mapped to."', args=[d.arg(name='devicePath', type=d.T.string)]),
+ withDevicePath(devicePath): { devicePath: devicePath },
+ '#withName':: d.fn(help='"name must match the name of a persistentVolumeClaim in the pod"', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { name: name },
+ },
+ '#volumeMounts':: d.obj(help="\"Pod volumes to mount into the container's filesystem. Cannot be updated.\""),
+ volumeMounts: {
+ '#withMountPath':: d.fn(help="\"Path within the container at which the volume should be mounted. Must not contain ':'.\"", args=[d.arg(name='mountPath', type=d.T.string)]),
+ withMountPath(mountPath): { mountPath: mountPath },
+ '#withMountPropagation':: d.fn(help='"mountPropagation determines how mounts are propagated from the host to container and the other way around. When not set, MountPropagationNone is used. This field is beta in 1.10."', args=[d.arg(name='mountPropagation', type=d.T.string)]),
+ withMountPropagation(mountPropagation): { mountPropagation: mountPropagation },
+ '#withName':: d.fn(help='"This must match the Name of a Volume."', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { name: name },
+ '#withReadOnly':: d.fn(help='"Mounted read-only if true, read-write otherwise (false or unspecified). Defaults to false."', args=[d.arg(name='readOnly', type=d.T.boolean)]),
+ withReadOnly(readOnly): { readOnly: readOnly },
+ '#withSubPath':: d.fn(help="\"Path within the volume from which the container's volume should be mounted. Defaults to \\\"\\\" (volume's root).\"", args=[d.arg(name='subPath', type=d.T.string)]),
+ withSubPath(subPath): { subPath: subPath },
+ '#withSubPathExpr':: d.fn(help="\"Expanded path within the volume from which the container's volume should be mounted. Behaves similarly to SubPath but environment variable references $(VAR_NAME) are expanded using the container's environment. Defaults to \\\"\\\" (volume's root). SubPathExpr and SubPath are mutually exclusive.\"", args=[d.arg(name='subPathExpr', type=d.T.string)]),
+ withSubPathExpr(subPathExpr): { subPathExpr: subPathExpr },
+ },
+ '#withArgs':: d.fn(help="\"Arguments to the entrypoint. The container image's CMD is used if this is not provided. Variable references $(VAR_NAME) are expanded using the container's environment. If a variable cannot be resolved, the reference in the input string will be unchanged. Double $$ are reduced to a single $, which allows for escaping the $(VAR_NAME) syntax: i.e. \\\"$$(VAR_NAME)\\\" will produce the string literal \\\"$(VAR_NAME)\\\". Escaped references will never be expanded, regardless of whether the variable exists or not. Cannot be updated. More info: https://kubernetes.io/docs/tasks/inject-data-application/define-command-argument-container/#running-a-command-in-a-shell\"", args=[d.arg(name='args', type=d.T.array)]),
+ withArgs(args): { args: if std.isArray(v=args) then args else [args] },
+ '#withArgsMixin':: d.fn(help="\"Arguments to the entrypoint. The container image's CMD is used if this is not provided. Variable references $(VAR_NAME) are expanded using the container's environment. If a variable cannot be resolved, the reference in the input string will be unchanged. Double $$ are reduced to a single $, which allows for escaping the $(VAR_NAME) syntax: i.e. \\\"$$(VAR_NAME)\\\" will produce the string literal \\\"$(VAR_NAME)\\\". Escaped references will never be expanded, regardless of whether the variable exists or not. Cannot be updated. More info: https://kubernetes.io/docs/tasks/inject-data-application/define-command-argument-container/#running-a-command-in-a-shell\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='args', type=d.T.array)]),
+ withArgsMixin(args): { args+: if std.isArray(v=args) then args else [args] },
+ '#withCommand':: d.fn(help="\"Entrypoint array. Not executed within a shell. The container image's ENTRYPOINT is used if this is not provided. Variable references $(VAR_NAME) are expanded using the container's environment. If a variable cannot be resolved, the reference in the input string will be unchanged. Double $$ are reduced to a single $, which allows for escaping the $(VAR_NAME) syntax: i.e. \\\"$$(VAR_NAME)\\\" will produce the string literal \\\"$(VAR_NAME)\\\". Escaped references will never be expanded, regardless of whether the variable exists or not. Cannot be updated. More info: https://kubernetes.io/docs/tasks/inject-data-application/define-command-argument-container/#running-a-command-in-a-shell\"", args=[d.arg(name='command', type=d.T.array)]),
+ withCommand(command): { command: if std.isArray(v=command) then command else [command] },
+ '#withCommandMixin':: d.fn(help="\"Entrypoint array. Not executed within a shell. The container image's ENTRYPOINT is used if this is not provided. Variable references $(VAR_NAME) are expanded using the container's environment. If a variable cannot be resolved, the reference in the input string will be unchanged. Double $$ are reduced to a single $, which allows for escaping the $(VAR_NAME) syntax: i.e. \\\"$$(VAR_NAME)\\\" will produce the string literal \\\"$(VAR_NAME)\\\". Escaped references will never be expanded, regardless of whether the variable exists or not. Cannot be updated. More info: https://kubernetes.io/docs/tasks/inject-data-application/define-command-argument-container/#running-a-command-in-a-shell\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='command', type=d.T.array)]),
+ withCommandMixin(command): { command+: if std.isArray(v=command) then command else [command] },
+ '#withEnv':: d.fn(help='"List of environment variables to set in the container. Cannot be updated."', args=[d.arg(name='env', type=d.T.array)]),
+ withEnv(env): { env: if std.isArray(v=env) then env else [env] },
+ '#withEnvFrom':: d.fn(help='"List of sources to populate environment variables in the container. The keys defined within a source must be a C_IDENTIFIER. All invalid keys will be reported as an event when the container is starting. When a key exists in multiple sources, the value associated with the last source will take precedence. Values defined by an Env with a duplicate key will take precedence. Cannot be updated."', args=[d.arg(name='envFrom', type=d.T.array)]),
+ withEnvFrom(envFrom): { envFrom: if std.isArray(v=envFrom) then envFrom else [envFrom] },
+ '#withEnvFromMixin':: d.fn(help='"List of sources to populate environment variables in the container. The keys defined within a source must be a C_IDENTIFIER. All invalid keys will be reported as an event when the container is starting. When a key exists in multiple sources, the value associated with the last source will take precedence. Values defined by an Env with a duplicate key will take precedence. Cannot be updated."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='envFrom', type=d.T.array)]),
+ withEnvFromMixin(envFrom): { envFrom+: if std.isArray(v=envFrom) then envFrom else [envFrom] },
+ '#withEnvMixin':: d.fn(help='"List of environment variables to set in the container. Cannot be updated."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='env', type=d.T.array)]),
+ withEnvMixin(env): { env+: if std.isArray(v=env) then env else [env] },
+ '#withImage':: d.fn(help='"Container image name. More info: https://kubernetes.io/docs/concepts/containers/images This field is optional to allow higher level config management to default or override container images in workload controllers like Deployments and StatefulSets."', args=[d.arg(name='image', type=d.T.string)]),
+ withImage(image): { image: image },
+ '#withImagePullPolicy':: d.fn(help='"Image pull policy. One of Always, Never, IfNotPresent. Defaults to Always if :latest tag is specified, or IfNotPresent otherwise. Cannot be updated. More info: https://kubernetes.io/docs/concepts/containers/images#updating-images"', args=[d.arg(name='imagePullPolicy', type=d.T.string)]),
+ withImagePullPolicy(imagePullPolicy): { imagePullPolicy: imagePullPolicy },
+ '#withName':: d.fn(help='"Name of the container specified as a DNS_LABEL. Each container in a pod must have a unique name (DNS_LABEL). Cannot be updated."', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { name: name },
+ '#withPorts':: d.fn(help='"List of ports to expose from the container. Not specifying a port here DOES NOT prevent that port from being exposed. Any port which is listening on the default \\"0.0.0.0\\" address inside a container will be accessible from the network. Modifying this array with strategic merge patch may corrupt the data. For more information See https://github.com/kubernetes/kubernetes/issues/108255. Cannot be updated."', args=[d.arg(name='ports', type=d.T.array)]),
+ withPorts(ports): { ports: if std.isArray(v=ports) then ports else [ports] },
+ '#withPortsMixin':: d.fn(help='"List of ports to expose from the container. Not specifying a port here DOES NOT prevent that port from being exposed. Any port which is listening on the default \\"0.0.0.0\\" address inside a container will be accessible from the network. Modifying this array with strategic merge patch may corrupt the data. For more information See https://github.com/kubernetes/kubernetes/issues/108255. Cannot be updated."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ports', type=d.T.array)]),
+ withPortsMixin(ports): { ports+: if std.isArray(v=ports) then ports else [ports] },
+ '#withResizePolicy':: d.fn(help='"Resources resize policy for the container."', args=[d.arg(name='resizePolicy', type=d.T.array)]),
+ withResizePolicy(resizePolicy): { resizePolicy: if std.isArray(v=resizePolicy) then resizePolicy else [resizePolicy] },
+ '#withResizePolicyMixin':: d.fn(help='"Resources resize policy for the container."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='resizePolicy', type=d.T.array)]),
+ withResizePolicyMixin(resizePolicy): { resizePolicy+: if std.isArray(v=resizePolicy) then resizePolicy else [resizePolicy] },
+ '#withRestartPolicy':: d.fn(help="\"RestartPolicy defines the restart behavior of individual containers in a pod. This field may only be set for init containers, and the only allowed value is \\\"Always\\\". For non-init containers or when this field is not specified, the restart behavior is defined by the Pod's restart policy and the container type. Setting the RestartPolicy as \\\"Always\\\" for the init container will have the following effect: this init container will be continually restarted on exit until all regular containers have terminated. Once all regular containers have completed, all init containers with restartPolicy \\\"Always\\\" will be shut down. This lifecycle differs from normal init containers and is often referred to as a \\\"sidecar\\\" container. Although this init container still starts in the init container sequence, it does not wait for the container to complete before proceeding to the next init container. Instead, the next init container starts immediately after this init container is started, or after any startupProbe has successfully completed.\"", args=[d.arg(name='restartPolicy', type=d.T.string)]),
+ withRestartPolicy(restartPolicy): { restartPolicy: restartPolicy },
+ '#withStdin':: d.fn(help='"Whether this container should allocate a buffer for stdin in the container runtime. If this is not set, reads from stdin in the container will always result in EOF. Default is false."', args=[d.arg(name='stdin', type=d.T.boolean)]),
+ withStdin(stdin): { stdin: stdin },
+ '#withStdinOnce':: d.fn(help='"Whether the container runtime should close the stdin channel after it has been opened by a single attach. When stdin is true the stdin stream will remain open across multiple attach sessions. If stdinOnce is set to true, stdin is opened on container start, is empty until the first client attaches to stdin, and then remains open and accepts data until the client disconnects, at which time stdin is closed and remains closed until the container is restarted. If this flag is false, a container processes that reads from stdin will never receive an EOF. Default is false"', args=[d.arg(name='stdinOnce', type=d.T.boolean)]),
+ withStdinOnce(stdinOnce): { stdinOnce: stdinOnce },
+ '#withTerminationMessagePath':: d.fn(help="\"Optional: Path at which the file to which the container's termination message will be written is mounted into the container's filesystem. Message written is intended to be brief final status, such as an assertion failure message. Will be truncated by the node if greater than 4096 bytes. The total message length across all containers will be limited to 12kb. Defaults to /dev/termination-log. Cannot be updated.\"", args=[d.arg(name='terminationMessagePath', type=d.T.string)]),
+ withTerminationMessagePath(terminationMessagePath): { terminationMessagePath: terminationMessagePath },
+ '#withTerminationMessagePolicy':: d.fn(help='"Indicate how the termination message should be populated. File will use the contents of terminationMessagePath to populate the container status message on both success and failure. FallbackToLogsOnError will use the last chunk of container log output if the termination message file is empty and the container exited with an error. The log output is limited to 2048 bytes or 80 lines, whichever is smaller. Defaults to File. Cannot be updated."', args=[d.arg(name='terminationMessagePolicy', type=d.T.string)]),
+ withTerminationMessagePolicy(terminationMessagePolicy): { terminationMessagePolicy: terminationMessagePolicy },
+ '#withTty':: d.fn(help="\"Whether this container should allocate a TTY for itself, also requires 'stdin' to be true. Default is false.\"", args=[d.arg(name='tty', type=d.T.boolean)]),
+ withTty(tty): { tty: tty },
+ '#withVolumeDevices':: d.fn(help='"volumeDevices is the list of block devices to be used by the container."', args=[d.arg(name='volumeDevices', type=d.T.array)]),
+ withVolumeDevices(volumeDevices): { volumeDevices: if std.isArray(v=volumeDevices) then volumeDevices else [volumeDevices] },
+ '#withVolumeDevicesMixin':: d.fn(help='"volumeDevices is the list of block devices to be used by the container."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='volumeDevices', type=d.T.array)]),
+ withVolumeDevicesMixin(volumeDevices): { volumeDevices+: if std.isArray(v=volumeDevices) then volumeDevices else [volumeDevices] },
+ '#withVolumeMounts':: d.fn(help="\"Pod volumes to mount into the container's filesystem. Cannot be updated.\"", args=[d.arg(name='volumeMounts', type=d.T.array)]),
+ withVolumeMounts(volumeMounts): { volumeMounts: if std.isArray(v=volumeMounts) then volumeMounts else [volumeMounts] },
+ '#withVolumeMountsMixin':: d.fn(help="\"Pod volumes to mount into the container's filesystem. Cannot be updated.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='volumeMounts', type=d.T.array)]),
+ withVolumeMountsMixin(volumeMounts): { volumeMounts+: if std.isArray(v=volumeMounts) then volumeMounts else [volumeMounts] },
+ '#withWorkingDir':: d.fn(help="\"Container's working directory. If not specified, the container runtime's default will be used, which might be configured in the container image. Cannot be updated.\"", args=[d.arg(name='workingDir', type=d.T.string)]),
+ withWorkingDir(workingDir): { workingDir: workingDir },
+ },
+ '#os':: d.obj(help='"Specifies the OS of the containers in the pod. Some pod and container fields are restricted if this is set. \\n If the OS field is set to linux, the following fields must be unset: -securityContext.windowsOptions \\n If the OS field is set to windows, following fields must be unset: - spec.hostPID - spec.hostIPC - spec.hostUsers - spec.securityContext.seLinuxOptions - spec.securityContext.seccompProfile - spec.securityContext.fsGroup - spec.securityContext.fsGroupChangePolicy - spec.securityContext.sysctls - spec.shareProcessNamespace - spec.securityContext.runAsUser - spec.securityContext.runAsGroup - spec.securityContext.supplementalGroups - spec.containers[*].securityContext.seLinuxOptions - spec.containers[*].securityContext.seccompProfile - spec.containers[*].securityContext.capabilities - spec.containers[*].securityContext.readOnlyRootFilesystem - spec.containers[*].securityContext.privileged - spec.containers[*].securityContext.allowPrivilegeEscalation - spec.containers[*].securityContext.procMount - spec.containers[*].securityContext.runAsUser - spec.containers[*].securityContext.runAsGroup"'),
+ os: {
+ '#withName':: d.fn(help='"Name is the name of the operating system. The currently supported values are linux and windows. Additional value may be defined in future and can be one of: https://github.com/opencontainers/runtime-spec/blob/master/config.md#platform-specific-configuration Clients should expect to handle additional values and treat unrecognized values in this field as os: null"', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { spec+: { template+: { spec+: { os+: { name: name } } } } },
+ },
+ '#readinessGates':: d.obj(help='"If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to \\"True\\" More info: https://git.k8s.io/enhancements/keps/sig-network/580-pod-readiness-gates"'),
+ readinessGates: {
+ '#withConditionType':: d.fn(help="\"ConditionType refers to a condition in the pod's condition list with matching type.\"", args=[d.arg(name='conditionType', type=d.T.string)]),
+ withConditionType(conditionType): { conditionType: conditionType },
+ },
+ '#resourceClaims':: d.obj(help='"ResourceClaims defines which ResourceClaims must be allocated and reserved before the Pod is allowed to start. The resources will be made available to those containers which consume them by name. \\n This is an alpha field and requires enabling the DynamicResourceAllocation feature gate. \\n This field is immutable."'),
+ resourceClaims: {
+ '#source':: d.obj(help='"Source describes where to find the ResourceClaim."'),
+ source: {
+ '#withResourceClaimName':: d.fn(help='"ResourceClaimName is the name of a ResourceClaim object in the same namespace as this pod."', args=[d.arg(name='resourceClaimName', type=d.T.string)]),
+ withResourceClaimName(resourceClaimName): { source+: { resourceClaimName: resourceClaimName } },
+ '#withResourceClaimTemplateName':: d.fn(help='"ResourceClaimTemplateName is the name of a ResourceClaimTemplate object in the same namespace as this pod. \\n The template will be used to create a new ResourceClaim, which will be bound to this pod. When this pod is deleted, the ResourceClaim will also be deleted. The pod name and resource name, along with a generated component, will be used to form a unique name for the ResourceClaim, which will be recorded in pod.status.resourceClaimStatuses. \\n This field is immutable and no changes will be made to the corresponding ResourceClaim by the control plane after creating the ResourceClaim."', args=[d.arg(name='resourceClaimTemplateName', type=d.T.string)]),
+ withResourceClaimTemplateName(resourceClaimTemplateName): { source+: { resourceClaimTemplateName: resourceClaimTemplateName } },
+ },
+ '#withName':: d.fn(help='"Name uniquely identifies this resource claim inside the pod. This must be a DNS_LABEL."', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { name: name },
+ },
+ '#schedulingGates':: d.obj(help='"SchedulingGates is an opaque list of values that if specified will block scheduling the pod. If schedulingGates is not empty, the pod will stay in the SchedulingGated state and the scheduler will not attempt to schedule the pod. \\n SchedulingGates can only be set at pod creation time, and be removed only afterwards. \\n This is a beta feature enabled by the PodSchedulingReadiness feature gate."'),
+ schedulingGates: {
+ '#withName':: d.fn(help='"Name of the scheduling gate. Each scheduling gate must have a unique name field."', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { name: name },
+ },
+ '#securityContext':: d.obj(help='"SecurityContext holds pod-level security attributes and common container settings. Optional: Defaults to empty. See type description for default values of each field."'),
+ securityContext: {
+ '#seLinuxOptions':: d.obj(help='"The SELinux context to be applied to all containers. If unspecified, the container runtime will allocate a random SELinux context for each container. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. Note that this field cannot be set when spec.os.name is windows."'),
+ seLinuxOptions: {
+ '#withLevel':: d.fn(help='"Level is SELinux level label that applies to the container."', args=[d.arg(name='level', type=d.T.string)]),
+ withLevel(level): { spec+: { template+: { spec+: { securityContext+: { seLinuxOptions+: { level: level } } } } } },
+ '#withRole':: d.fn(help='"Role is a SELinux role label that applies to the container."', args=[d.arg(name='role', type=d.T.string)]),
+ withRole(role): { spec+: { template+: { spec+: { securityContext+: { seLinuxOptions+: { role: role } } } } } },
+ '#withType':: d.fn(help='"Type is a SELinux type label that applies to the container."', args=[d.arg(name='type', type=d.T.string)]),
+ withType(type): { spec+: { template+: { spec+: { securityContext+: { seLinuxOptions+: { type: type } } } } } },
+ '#withUser':: d.fn(help='"User is a SELinux user label that applies to the container."', args=[d.arg(name='user', type=d.T.string)]),
+ withUser(user): { spec+: { template+: { spec+: { securityContext+: { seLinuxOptions+: { user: user } } } } } },
+ },
+ '#seccompProfile':: d.obj(help='"The seccomp options to use by the containers in this pod. Note that this field cannot be set when spec.os.name is windows."'),
+ seccompProfile: {
+ '#withLocalhostProfile':: d.fn(help="\"localhostProfile indicates a profile defined in a file on the node should be used. The profile must be preconfigured on the node to work. Must be a descending path, relative to the kubelet's configured seccomp profile location. Must be set if type is \\\"Localhost\\\". Must NOT be set for any other type.\"", args=[d.arg(name='localhostProfile', type=d.T.string)]),
+ withLocalhostProfile(localhostProfile): { spec+: { template+: { spec+: { securityContext+: { seccompProfile+: { localhostProfile: localhostProfile } } } } } },
+ '#withType':: d.fn(help='"type indicates which kind of seccomp profile will be applied. Valid options are: \\n Localhost - a profile defined in a file on the node should be used. RuntimeDefault - the container runtime default profile should be used. Unconfined - no profile should be applied."', args=[d.arg(name='type', type=d.T.string)]),
+ withType(type): { spec+: { template+: { spec+: { securityContext+: { seccompProfile+: { type: type } } } } } },
+ },
+ '#sysctls':: d.obj(help='"Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch. Note that this field cannot be set when spec.os.name is windows."'),
+ sysctls: {
+ '#withName':: d.fn(help='"Name of a property to set"', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { name: name },
+ '#withValue':: d.fn(help='"Value of a property to set"', args=[d.arg(name='value', type=d.T.string)]),
+ withValue(value): { value: value },
+ },
+ '#windowsOptions':: d.obj(help="\"The Windows specific settings applied to all containers. If unspecified, the options within a container's SecurityContext will be used. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. Note that this field cannot be set when spec.os.name is linux.\""),
+ windowsOptions: {
+ '#withGmsaCredentialSpec':: d.fn(help='"GMSACredentialSpec is where the GMSA admission webhook (https://github.com/kubernetes-sigs/windows-gmsa) inlines the contents of the GMSA credential spec named by the GMSACredentialSpecName field."', args=[d.arg(name='gmsaCredentialSpec', type=d.T.string)]),
+ withGmsaCredentialSpec(gmsaCredentialSpec): { spec+: { template+: { spec+: { securityContext+: { windowsOptions+: { gmsaCredentialSpec: gmsaCredentialSpec } } } } } },
+ '#withGmsaCredentialSpecName':: d.fn(help='"GMSACredentialSpecName is the name of the GMSA credential spec to use."', args=[d.arg(name='gmsaCredentialSpecName', type=d.T.string)]),
+ withGmsaCredentialSpecName(gmsaCredentialSpecName): { spec+: { template+: { spec+: { securityContext+: { windowsOptions+: { gmsaCredentialSpecName: gmsaCredentialSpecName } } } } } },
+ '#withHostProcess':: d.fn(help="\"HostProcess determines if a container should be run as a 'Host Process' container. All of a Pod's containers must have the same effective HostProcess value (it is not allowed to have a mix of HostProcess containers and non-HostProcess containers). In addition, if HostProcess is true then HostNetwork must also be set to true.\"", args=[d.arg(name='hostProcess', type=d.T.boolean)]),
+ withHostProcess(hostProcess): { spec+: { template+: { spec+: { securityContext+: { windowsOptions+: { hostProcess: hostProcess } } } } } },
+ '#withRunAsUserName':: d.fn(help='"The UserName in Windows to run the entrypoint of the container process. Defaults to the user specified in image metadata if unspecified. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence."', args=[d.arg(name='runAsUserName', type=d.T.string)]),
+ withRunAsUserName(runAsUserName): { spec+: { template+: { spec+: { securityContext+: { windowsOptions+: { runAsUserName: runAsUserName } } } } } },
+ },
+ '#withFsGroup':: d.fn(help="\"A special supplemental group that applies to all containers in a pod. Some volume types allow the Kubelet to change the ownership of that volume to be owned by the pod: \\n 1. The owning GID will be the FSGroup 2. The setgid bit is set (new files created in the volume will be owned by FSGroup) 3. The permission bits are OR'd with rw-rw---- \\n If unset, the Kubelet will not modify the ownership and permissions of any volume. Note that this field cannot be set when spec.os.name is windows.\"", args=[d.arg(name='fsGroup', type=d.T.integer)]),
+ withFsGroup(fsGroup): { spec+: { template+: { spec+: { securityContext+: { fsGroup: fsGroup } } } } },
+ '#withFsGroupChangePolicy':: d.fn(help='"fsGroupChangePolicy defines behavior of changing ownership and permission of the volume before being exposed inside Pod. This field will only apply to volume types which support fsGroup based ownership(and permissions). It will have no effect on ephemeral volume types such as: secret, configmaps and emptydir. Valid values are \\"OnRootMismatch\\" and \\"Always\\". If not specified, \\"Always\\" is used. Note that this field cannot be set when spec.os.name is windows."', args=[d.arg(name='fsGroupChangePolicy', type=d.T.string)]),
+ withFsGroupChangePolicy(fsGroupChangePolicy): { spec+: { template+: { spec+: { securityContext+: { fsGroupChangePolicy: fsGroupChangePolicy } } } } },
+ '#withRunAsGroup':: d.fn(help='"The GID to run the entrypoint of the container process. Uses runtime default if unset. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. Note that this field cannot be set when spec.os.name is windows."', args=[d.arg(name='runAsGroup', type=d.T.integer)]),
+ withRunAsGroup(runAsGroup): { spec+: { template+: { spec+: { securityContext+: { runAsGroup: runAsGroup } } } } },
+ '#withRunAsNonRoot':: d.fn(help='"Indicates that the container must run as a non-root user. If true, the Kubelet will validate the image at runtime to ensure that it does not run as UID 0 (root) and fail to start the container if it does. If unset or false, no such validation will be performed. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence."', args=[d.arg(name='runAsNonRoot', type=d.T.boolean)]),
+ withRunAsNonRoot(runAsNonRoot): { spec+: { template+: { spec+: { securityContext+: { runAsNonRoot: runAsNonRoot } } } } },
+ '#withRunAsUser':: d.fn(help='"The UID to run the entrypoint of the container process. Defaults to user specified in image metadata if unspecified. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. Note that this field cannot be set when spec.os.name is windows."', args=[d.arg(name='runAsUser', type=d.T.integer)]),
+ withRunAsUser(runAsUser): { spec+: { template+: { spec+: { securityContext+: { runAsUser: runAsUser } } } } },
+ '#withSupplementalGroups':: d.fn(help="\"A list of groups applied to the first process run in each container, in addition to the container's primary GID, the fsGroup (if specified), and group memberships defined in the container image for the uid of the container process. If unspecified, no additional groups are added to any container. Note that group memberships defined in the container image for the uid of the container process are still effective, even if they are not included in this list. Note that this field cannot be set when spec.os.name is windows.\"", args=[d.arg(name='supplementalGroups', type=d.T.array)]),
+ withSupplementalGroups(supplementalGroups): { spec+: { template+: { spec+: { securityContext+: { supplementalGroups: if std.isArray(v=supplementalGroups) then supplementalGroups else [supplementalGroups] } } } } },
+ '#withSupplementalGroupsMixin':: d.fn(help="\"A list of groups applied to the first process run in each container, in addition to the container's primary GID, the fsGroup (if specified), and group memberships defined in the container image for the uid of the container process. If unspecified, no additional groups are added to any container. Note that group memberships defined in the container image for the uid of the container process are still effective, even if they are not included in this list. Note that this field cannot be set when spec.os.name is windows.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='supplementalGroups', type=d.T.array)]),
+ withSupplementalGroupsMixin(supplementalGroups): { spec+: { template+: { spec+: { securityContext+: { supplementalGroups+: if std.isArray(v=supplementalGroups) then supplementalGroups else [supplementalGroups] } } } } },
+ '#withSysctls':: d.fn(help='"Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch. Note that this field cannot be set when spec.os.name is windows."', args=[d.arg(name='sysctls', type=d.T.array)]),
+ withSysctls(sysctls): { spec+: { template+: { spec+: { securityContext+: { sysctls: if std.isArray(v=sysctls) then sysctls else [sysctls] } } } } },
+ '#withSysctlsMixin':: d.fn(help='"Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch. Note that this field cannot be set when spec.os.name is windows."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='sysctls', type=d.T.array)]),
+ withSysctlsMixin(sysctls): { spec+: { template+: { spec+: { securityContext+: { sysctls+: if std.isArray(v=sysctls) then sysctls else [sysctls] } } } } },
+ },
+ '#tolerations':: d.obj(help="\"If specified, the pod's tolerations.\""),
+ tolerations: {
+ '#withEffect':: d.fn(help='"Effect indicates the taint effect to match. Empty means match all taint effects. When specified, allowed values are NoSchedule, PreferNoSchedule and NoExecute."', args=[d.arg(name='effect', type=d.T.string)]),
+ withEffect(effect): { effect: effect },
+ '#withKey':: d.fn(help='"Key is the taint key that the toleration applies to. Empty means match all taint keys. If the key is empty, operator must be Exists; this combination means to match all values and all keys."', args=[d.arg(name='key', type=d.T.string)]),
+ withKey(key): { key: key },
+ '#withOperator':: d.fn(help="\"Operator represents a key's relationship to the value. Valid operators are Exists and Equal. Defaults to Equal. Exists is equivalent to wildcard for value, so that a pod can tolerate all taints of a particular category.\"", args=[d.arg(name='operator', type=d.T.string)]),
+ withOperator(operator): { operator: operator },
+ '#withTolerationSeconds':: d.fn(help='"TolerationSeconds represents the period of time the toleration (which must be of effect NoExecute, otherwise this field is ignored) tolerates the taint. By default, it is not set, which means tolerate the taint forever (do not evict). Zero and negative values will be treated as 0 (evict immediately) by the system."', args=[d.arg(name='tolerationSeconds', type=d.T.integer)]),
+ withTolerationSeconds(tolerationSeconds): { tolerationSeconds: tolerationSeconds },
+ '#withValue':: d.fn(help='"Value is the taint value the toleration matches to. If the operator is Exists, the value should be empty, otherwise just a regular string."', args=[d.arg(name='value', type=d.T.string)]),
+ withValue(value): { value: value },
+ },
+ '#topologySpreadConstraints':: d.obj(help='"TopologySpreadConstraints describes how a group of pods ought to spread across topology domains. Scheduler will schedule pods in a way which abides by the constraints. All topologySpreadConstraints are ANDed."'),
+ topologySpreadConstraints: {
+ '#labelSelector':: d.obj(help='"LabelSelector is used to find matching pods. Pods that match this label selector are counted to determine the number of pods in their corresponding topology domain."'),
+ labelSelector: {
+ '#matchExpressions':: d.obj(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."'),
+ matchExpressions: {
+ '#withKey':: d.fn(help='"key is the label key that the selector applies to."', args=[d.arg(name='key', type=d.T.string)]),
+ withKey(key): { key: key },
+ '#withOperator':: d.fn(help="\"operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist.\"", args=[d.arg(name='operator', type=d.T.string)]),
+ withOperator(operator): { operator: operator },
+ '#withValues':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."', args=[d.arg(name='values', type=d.T.array)]),
+ withValues(values): { values: if std.isArray(v=values) then values else [values] },
+ '#withValuesMixin':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='values', type=d.T.array)]),
+ withValuesMixin(values): { values+: if std.isArray(v=values) then values else [values] },
+ },
+ '#withMatchExpressions':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."', args=[d.arg(name='matchExpressions', type=d.T.array)]),
+ withMatchExpressions(matchExpressions): { labelSelector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } },
+ '#withMatchExpressionsMixin':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]),
+ withMatchExpressionsMixin(matchExpressions): { labelSelector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } },
+ '#withMatchLabels':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."', args=[d.arg(name='matchLabels', type=d.T.object)]),
+ withMatchLabels(matchLabels): { labelSelector+: { matchLabels: matchLabels } },
+ '#withMatchLabelsMixin':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]),
+ withMatchLabelsMixin(matchLabels): { labelSelector+: { matchLabels+: matchLabels } },
+ },
+ '#withMatchLabelKeys':: d.fn(help="\"MatchLabelKeys is a set of pod label keys to select the pods over which spreading will be calculated. The keys are used to lookup values from the incoming pod labels, those key-value labels are ANDed with labelSelector to select the group of existing pods over which spreading will be calculated for the incoming pod. The same key is forbidden to exist in both MatchLabelKeys and LabelSelector. MatchLabelKeys cannot be set when LabelSelector isn't set. Keys that don't exist in the incoming pod labels will be ignored. A null or empty list means only match against labelSelector. \\n This is a beta field and requires the MatchLabelKeysInPodTopologySpread feature gate to be enabled (enabled by default).\"", args=[d.arg(name='matchLabelKeys', type=d.T.array)]),
+ withMatchLabelKeys(matchLabelKeys): { matchLabelKeys: if std.isArray(v=matchLabelKeys) then matchLabelKeys else [matchLabelKeys] },
+ '#withMatchLabelKeysMixin':: d.fn(help="\"MatchLabelKeys is a set of pod label keys to select the pods over which spreading will be calculated. The keys are used to lookup values from the incoming pod labels, those key-value labels are ANDed with labelSelector to select the group of existing pods over which spreading will be calculated for the incoming pod. The same key is forbidden to exist in both MatchLabelKeys and LabelSelector. MatchLabelKeys cannot be set when LabelSelector isn't set. Keys that don't exist in the incoming pod labels will be ignored. A null or empty list means only match against labelSelector. \\n This is a beta field and requires the MatchLabelKeysInPodTopologySpread feature gate to be enabled (enabled by default).\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='matchLabelKeys', type=d.T.array)]),
+ withMatchLabelKeysMixin(matchLabelKeys): { matchLabelKeys+: if std.isArray(v=matchLabelKeys) then matchLabelKeys else [matchLabelKeys] },
+ '#withMaxSkew':: d.fn(help="\"MaxSkew describes the degree to which pods may be unevenly distributed. When `whenUnsatisfiable=DoNotSchedule`, it is the maximum permitted difference between the number of matching pods in the target topology and the global minimum. The global minimum is the minimum number of matching pods in an eligible domain or zero if the number of eligible domains is less than MinDomains. For example, in a 3-zone cluster, MaxSkew is set to 1, and pods with the same labelSelector spread as 2/2/1: In this case, the global minimum is 1. | zone1 | zone2 | zone3 | | P P | P P | P | - if MaxSkew is 1, incoming pod can only be scheduled to zone3 to become 2/2/2; scheduling it onto zone1(zone2) would make the ActualSkew(3-1) on zone1(zone2) violate MaxSkew(1). - if MaxSkew is 2, incoming pod can be scheduled onto any zone. When `whenUnsatisfiable=ScheduleAnyway`, it is used to give higher precedence to topologies that satisfy it. It's a required field. Default value is 1 and 0 is not allowed.\"", args=[d.arg(name='maxSkew', type=d.T.integer)]),
+ withMaxSkew(maxSkew): { maxSkew: maxSkew },
+ '#withMinDomains':: d.fn(help="\"MinDomains indicates a minimum number of eligible domains. When the number of eligible domains with matching topology keys is less than minDomains, Pod Topology Spread treats \\\"global minimum\\\" as 0, and then the calculation of Skew is performed. And when the number of eligible domains with matching topology keys equals or greater than minDomains, this value has no effect on scheduling. As a result, when the number of eligible domains is less than minDomains, scheduler won't schedule more than maxSkew Pods to those domains. If value is nil, the constraint behaves as if MinDomains is equal to 1. Valid values are integers greater than 0. When value is not nil, WhenUnsatisfiable must be DoNotSchedule. \\n For example, in a 3-zone cluster, MaxSkew is set to 2, MinDomains is set to 5 and pods with the same labelSelector spread as 2/2/2: | zone1 | zone2 | zone3 | | P P | P P | P P | The number of domains is less than 5(MinDomains), so \\\"global minimum\\\" is treated as 0. In this situation, new pod with the same labelSelector cannot be scheduled, because computed skew will be 3(3 - 0) if new Pod is scheduled to any of the three zones, it will violate MaxSkew. \\n This is a beta field and requires the MinDomainsInPodTopologySpread feature gate to be enabled (enabled by default).\"", args=[d.arg(name='minDomains', type=d.T.integer)]),
+ withMinDomains(minDomains): { minDomains: minDomains },
+ '#withNodeAffinityPolicy':: d.fn(help="\"NodeAffinityPolicy indicates how we will treat Pod's nodeAffinity/nodeSelector when calculating pod topology spread skew. Options are: - Honor: only nodes matching nodeAffinity/nodeSelector are included in the calculations. - Ignore: nodeAffinity/nodeSelector are ignored. All nodes are included in the calculations. \\n If this value is nil, the behavior is equivalent to the Honor policy. This is a beta-level feature default enabled by the NodeInclusionPolicyInPodTopologySpread feature flag.\"", args=[d.arg(name='nodeAffinityPolicy', type=d.T.string)]),
+ withNodeAffinityPolicy(nodeAffinityPolicy): { nodeAffinityPolicy: nodeAffinityPolicy },
+ '#withNodeTaintsPolicy':: d.fn(help='"NodeTaintsPolicy indicates how we will treat node taints when calculating pod topology spread skew. Options are: - Honor: nodes without taints, along with tainted nodes for which the incoming pod has a toleration, are included. - Ignore: node taints are ignored. All nodes are included. \\n If this value is nil, the behavior is equivalent to the Ignore policy. This is a beta-level feature default enabled by the NodeInclusionPolicyInPodTopologySpread feature flag."', args=[d.arg(name='nodeTaintsPolicy', type=d.T.string)]),
+ withNodeTaintsPolicy(nodeTaintsPolicy): { nodeTaintsPolicy: nodeTaintsPolicy },
+ '#withTopologyKey':: d.fn(help="\"TopologyKey is the key of node labels. Nodes that have a label with this key and identical values are considered to be in the same topology. We consider each \u003ckey, value\u003e as a \\\"bucket\\\", and try to put balanced number of pods into each bucket. We define a domain as a particular instance of a topology. Also, we define an eligible domain as a domain whose nodes meet the requirements of nodeAffinityPolicy and nodeTaintsPolicy. e.g. If TopologyKey is \\\"kubernetes.io/hostname\\\", each Node is a domain of that topology. And, if TopologyKey is \\\"topology.kubernetes.io/zone\\\", each zone is a domain of that topology. It's a required field.\"", args=[d.arg(name='topologyKey', type=d.T.string)]),
+ withTopologyKey(topologyKey): { topologyKey: topologyKey },
+ '#withWhenUnsatisfiable':: d.fn(help="\"WhenUnsatisfiable indicates how to deal with a pod if it doesn't satisfy the spread constraint. - DoNotSchedule (default) tells the scheduler not to schedule it. - ScheduleAnyway tells the scheduler to schedule the pod in any location, but giving higher precedence to topologies that would help reduce the skew. A constraint is considered \\\"Unsatisfiable\\\" for an incoming pod if and only if every possible node assignment for that pod would violate \\\"MaxSkew\\\" on some topology. For example, in a 3-zone cluster, MaxSkew is set to 1, and pods with the same labelSelector spread as 3/1/1: | zone1 | zone2 | zone3 | | P P P | P | P | If WhenUnsatisfiable is set to DoNotSchedule, incoming pod can only be scheduled to zone2(zone3) to become 3/2/1(3/1/2) as ActualSkew(2-1) on zone2(zone3) satisfies MaxSkew(1). In other words, the cluster can still be imbalanced, but scheduler won't make it *more* imbalanced. It's a required field.\"", args=[d.arg(name='whenUnsatisfiable', type=d.T.string)]),
+ withWhenUnsatisfiable(whenUnsatisfiable): { whenUnsatisfiable: whenUnsatisfiable },
+ },
+ '#volumes':: d.obj(help='"List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes"'),
+ volumes: {
+ '#awsElasticBlockStore':: d.obj(help="\"awsElasticBlockStore represents an AWS Disk resource that is attached to a kubelet's host machine and then exposed to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes#awselasticblockstore\""),
+ awsElasticBlockStore: {
+ '#withFsType':: d.fn(help='"fsType is the filesystem type of the volume that you want to mount. Tip: Ensure that the filesystem type is supported by the host operating system. Examples: \\"ext4\\", \\"xfs\\", \\"ntfs\\". Implicitly inferred to be \\"ext4\\" if unspecified. More info: https://kubernetes.io/docs/concepts/storage/volumes#awselasticblockstore TODO: how do we prevent errors in the filesystem from compromising the machine"', args=[d.arg(name='fsType', type=d.T.string)]),
+ withFsType(fsType): { awsElasticBlockStore+: { fsType: fsType } },
+ '#withPartition':: d.fn(help='"partition is the partition in the volume that you want to mount. If omitted, the default is to mount by volume name. Examples: For volume /dev/sda1, you specify the partition as \\"1\\". Similarly, the volume partition for /dev/sda is \\"0\\" (or you can leave the property empty)."', args=[d.arg(name='partition', type=d.T.integer)]),
+ withPartition(partition): { awsElasticBlockStore+: { partition: partition } },
+ '#withReadOnly':: d.fn(help='"readOnly value true will force the readOnly setting in VolumeMounts. More info: https://kubernetes.io/docs/concepts/storage/volumes#awselasticblockstore"', args=[d.arg(name='readOnly', type=d.T.boolean)]),
+ withReadOnly(readOnly): { awsElasticBlockStore+: { readOnly: readOnly } },
+ '#withVolumeID':: d.fn(help='"volumeID is unique ID of the persistent disk resource in AWS (Amazon EBS volume). More info: https://kubernetes.io/docs/concepts/storage/volumes#awselasticblockstore"', args=[d.arg(name='volumeID', type=d.T.string)]),
+ withVolumeID(volumeID): { awsElasticBlockStore+: { volumeID: volumeID } },
+ },
+ '#azureDisk':: d.obj(help='"azureDisk represents an Azure Data Disk mount on the host and bind mount to the pod."'),
+ azureDisk: {
+ '#withCachingMode':: d.fn(help='"cachingMode is the Host Caching mode: None, Read Only, Read Write."', args=[d.arg(name='cachingMode', type=d.T.string)]),
+ withCachingMode(cachingMode): { azureDisk+: { cachingMode: cachingMode } },
+ '#withDiskName':: d.fn(help='"diskName is the Name of the data disk in the blob storage"', args=[d.arg(name='diskName', type=d.T.string)]),
+ withDiskName(diskName): { azureDisk+: { diskName: diskName } },
+ '#withDiskURI':: d.fn(help='"diskURI is the URI of data disk in the blob storage"', args=[d.arg(name='diskURI', type=d.T.string)]),
+ withDiskURI(diskURI): { azureDisk+: { diskURI: diskURI } },
+ '#withFsType':: d.fn(help='"fsType is Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. \\"ext4\\", \\"xfs\\", \\"ntfs\\". Implicitly inferred to be \\"ext4\\" if unspecified."', args=[d.arg(name='fsType', type=d.T.string)]),
+ withFsType(fsType): { azureDisk+: { fsType: fsType } },
+ '#withKind':: d.fn(help='"kind expected values are Shared: multiple blob disks per storage account Dedicated: single blob disk per storage account Managed: azure managed data disk (only in managed availability set). defaults to shared"', args=[d.arg(name='kind', type=d.T.string)]),
+ withKind(kind): { azureDisk+: { kind: kind } },
+ '#withReadOnly':: d.fn(help='"readOnly Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts."', args=[d.arg(name='readOnly', type=d.T.boolean)]),
+ withReadOnly(readOnly): { azureDisk+: { readOnly: readOnly } },
+ },
+ '#azureFile':: d.obj(help='"azureFile represents an Azure File Service mount on the host and bind mount to the pod."'),
+ azureFile: {
+ '#withReadOnly':: d.fn(help='"readOnly defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts."', args=[d.arg(name='readOnly', type=d.T.boolean)]),
+ withReadOnly(readOnly): { azureFile+: { readOnly: readOnly } },
+ '#withSecretName':: d.fn(help='"secretName is the name of secret that contains Azure Storage Account Name and Key"', args=[d.arg(name='secretName', type=d.T.string)]),
+ withSecretName(secretName): { azureFile+: { secretName: secretName } },
+ '#withShareName':: d.fn(help='"shareName is the azure share Name"', args=[d.arg(name='shareName', type=d.T.string)]),
+ withShareName(shareName): { azureFile+: { shareName: shareName } },
+ },
+ '#cephfs':: d.obj(help="\"cephFS represents a Ceph FS mount on the host that shares a pod's lifetime\""),
+ cephfs: {
+ '#secretRef':: d.obj(help='"secretRef is Optional: SecretRef is reference to the authentication secret for User, default is empty. More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it"'),
+ secretRef: {
+ '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { cephfs+: { secretRef+: { name: name } } },
+ },
+ '#withMonitors':: d.fn(help='"monitors is Required: Monitors is a collection of Ceph monitors More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it"', args=[d.arg(name='monitors', type=d.T.array)]),
+ withMonitors(monitors): { cephfs+: { monitors: if std.isArray(v=monitors) then monitors else [monitors] } },
+ '#withMonitorsMixin':: d.fn(help='"monitors is Required: Monitors is a collection of Ceph monitors More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='monitors', type=d.T.array)]),
+ withMonitorsMixin(monitors): { cephfs+: { monitors+: if std.isArray(v=monitors) then monitors else [monitors] } },
+ '#withPath':: d.fn(help='"path is Optional: Used as the mounted root, rather than the full Ceph tree, default is /"', args=[d.arg(name='path', type=d.T.string)]),
+ withPath(path): { cephfs+: { path: path } },
+ '#withReadOnly':: d.fn(help='"readOnly is Optional: Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts. More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it"', args=[d.arg(name='readOnly', type=d.T.boolean)]),
+ withReadOnly(readOnly): { cephfs+: { readOnly: readOnly } },
+ '#withSecretFile':: d.fn(help='"secretFile is Optional: SecretFile is the path to key ring for User, default is /etc/ceph/user.secret More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it"', args=[d.arg(name='secretFile', type=d.T.string)]),
+ withSecretFile(secretFile): { cephfs+: { secretFile: secretFile } },
+ '#withUser':: d.fn(help='"user is optional: User is the rados user name, default is admin More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it"', args=[d.arg(name='user', type=d.T.string)]),
+ withUser(user): { cephfs+: { user: user } },
+ },
+ '#cinder':: d.obj(help='"cinder represents a cinder volume attached and mounted on kubelets host machine. More info: https://examples.k8s.io/mysql-cinder-pd/README.md"'),
+ cinder: {
+ '#secretRef':: d.obj(help='"secretRef is optional: points to a secret object containing parameters used to connect to OpenStack."'),
+ secretRef: {
+ '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { cinder+: { secretRef+: { name: name } } },
+ },
+ '#withFsType':: d.fn(help='"fsType is the filesystem type to mount. Must be a filesystem type supported by the host operating system. Examples: \\"ext4\\", \\"xfs\\", \\"ntfs\\". Implicitly inferred to be \\"ext4\\" if unspecified. More info: https://examples.k8s.io/mysql-cinder-pd/README.md"', args=[d.arg(name='fsType', type=d.T.string)]),
+ withFsType(fsType): { cinder+: { fsType: fsType } },
+ '#withReadOnly':: d.fn(help='"readOnly defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts. More info: https://examples.k8s.io/mysql-cinder-pd/README.md"', args=[d.arg(name='readOnly', type=d.T.boolean)]),
+ withReadOnly(readOnly): { cinder+: { readOnly: readOnly } },
+ '#withVolumeID':: d.fn(help='"volumeID used to identify the volume in cinder. More info: https://examples.k8s.io/mysql-cinder-pd/README.md"', args=[d.arg(name='volumeID', type=d.T.string)]),
+ withVolumeID(volumeID): { cinder+: { volumeID: volumeID } },
+ },
+ '#configMap':: d.obj(help='"configMap represents a configMap that should populate this volume"'),
+ configMap: {
+ '#items':: d.obj(help="\"items if unspecified, each key-value pair in the Data field of the referenced ConfigMap will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the ConfigMap, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'.\""),
+ items: {
+ '#withKey':: d.fn(help='"key is the key to project."', args=[d.arg(name='key', type=d.T.string)]),
+ withKey(key): { key: key },
+ '#withMode':: d.fn(help='"mode is Optional: mode bits used to set permissions on this file. Must be an octal value between 0000 and 0777 or a decimal value between 0 and 511. YAML accepts both octal and decimal values, JSON requires decimal values for mode bits. If not specified, the volume defaultMode will be used. This might be in conflict with other options that affect the file mode, like fsGroup, and the result can be other mode bits set."', args=[d.arg(name='mode', type=d.T.integer)]),
+ withMode(mode): { mode: mode },
+ '#withPath':: d.fn(help="\"path is the relative path of the file to map the key to. May not be an absolute path. May not contain the path element '..'. May not start with the string '..'.\"", args=[d.arg(name='path', type=d.T.string)]),
+ withPath(path): { path: path },
+ },
+ '#withDefaultMode':: d.fn(help='"defaultMode is optional: mode bits used to set permissions on created files by default. Must be an octal value between 0000 and 0777 or a decimal value between 0 and 511. YAML accepts both octal and decimal values, JSON requires decimal values for mode bits. Defaults to 0644. Directories within the path are not affected by this setting. This might be in conflict with other options that affect the file mode, like fsGroup, and the result can be other mode bits set."', args=[d.arg(name='defaultMode', type=d.T.integer)]),
+ withDefaultMode(defaultMode): { configMap+: { defaultMode: defaultMode } },
+ '#withItems':: d.fn(help="\"items if unspecified, each key-value pair in the Data field of the referenced ConfigMap will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the ConfigMap, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'.\"", args=[d.arg(name='items', type=d.T.array)]),
+ withItems(items): { configMap+: { items: if std.isArray(v=items) then items else [items] } },
+ '#withItemsMixin':: d.fn(help="\"items if unspecified, each key-value pair in the Data field of the referenced ConfigMap will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the ConfigMap, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='items', type=d.T.array)]),
+ withItemsMixin(items): { configMap+: { items+: if std.isArray(v=items) then items else [items] } },
+ '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { configMap+: { name: name } },
+ '#withOptional':: d.fn(help='"optional specify whether the ConfigMap or its keys must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]),
+ withOptional(optional): { configMap+: { optional: optional } },
+ },
+ '#csi':: d.obj(help='"csi (Container Storage Interface) represents ephemeral storage that is handled by certain external CSI drivers (Beta feature)."'),
+ csi: {
+ '#nodePublishSecretRef':: d.obj(help='"nodePublishSecretRef is a reference to the secret object containing sensitive information to pass to the CSI driver to complete the CSI NodePublishVolume and NodeUnpublishVolume calls. This field is optional, and may be empty if no secret is required. If the secret object contains more than one secret, all secret references are passed."'),
+ nodePublishSecretRef: {
+ '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { csi+: { nodePublishSecretRef+: { name: name } } },
+ },
+ '#withDriver':: d.fn(help='"driver is the name of the CSI driver that handles this volume. Consult with your admin for the correct name as registered in the cluster."', args=[d.arg(name='driver', type=d.T.string)]),
+ withDriver(driver): { csi+: { driver: driver } },
+ '#withFsType':: d.fn(help='"fsType to mount. Ex. \\"ext4\\", \\"xfs\\", \\"ntfs\\". If not provided, the empty value is passed to the associated CSI driver which will determine the default filesystem to apply."', args=[d.arg(name='fsType', type=d.T.string)]),
+ withFsType(fsType): { csi+: { fsType: fsType } },
+ '#withReadOnly':: d.fn(help='"readOnly specifies a read-only configuration for the volume. Defaults to false (read/write)."', args=[d.arg(name='readOnly', type=d.T.boolean)]),
+ withReadOnly(readOnly): { csi+: { readOnly: readOnly } },
+ '#withVolumeAttributes':: d.fn(help="\"volumeAttributes stores driver-specific properties that are passed to the CSI driver. Consult your driver's documentation for supported values.\"", args=[d.arg(name='volumeAttributes', type=d.T.object)]),
+ withVolumeAttributes(volumeAttributes): { csi+: { volumeAttributes: volumeAttributes } },
+ '#withVolumeAttributesMixin':: d.fn(help="\"volumeAttributes stores driver-specific properties that are passed to the CSI driver. Consult your driver's documentation for supported values.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='volumeAttributes', type=d.T.object)]),
+ withVolumeAttributesMixin(volumeAttributes): { csi+: { volumeAttributes+: volumeAttributes } },
+ },
+ '#downwardAPI':: d.obj(help='"downwardAPI represents downward API about the pod that should populate this volume"'),
+ downwardAPI: {
+ '#items':: d.obj(help='"Items is a list of downward API volume file"'),
+ items: {
+ '#fieldRef':: d.obj(help='"Required: Selects a field of the pod: only annotations, labels, name and namespace are supported."'),
+ fieldRef: {
+ '#withApiVersion':: d.fn(help='"Version of the schema the FieldPath is written in terms of, defaults to \\"v1\\"."', args=[d.arg(name='apiVersion', type=d.T.string)]),
+ withApiVersion(apiVersion): { fieldRef+: { apiVersion: apiVersion } },
+ '#withFieldPath':: d.fn(help='"Path of the field to select in the specified API version."', args=[d.arg(name='fieldPath', type=d.T.string)]),
+ withFieldPath(fieldPath): { fieldRef+: { fieldPath: fieldPath } },
+ },
+ '#resourceFieldRef':: d.obj(help='"Selects a resource of the container: only resources limits and requests (limits.cpu, limits.memory, requests.cpu and requests.memory) are currently supported."'),
+ resourceFieldRef: {
+ '#withContainerName':: d.fn(help='"Container name: required for volumes, optional for env vars"', args=[d.arg(name='containerName', type=d.T.string)]),
+ withContainerName(containerName): { resourceFieldRef+: { containerName: containerName } },
+ '#withDivisor':: d.fn(help='"Specifies the output format of the exposed resources, defaults to \\"1\\', args=[d.arg(name='divisor', type=d.T.any)]),
+ withDivisor(divisor): { resourceFieldRef+: { divisor: divisor } },
+ '#withResource':: d.fn(help='"Required: resource to select"', args=[d.arg(name='resource', type=d.T.string)]),
+ withResource(resource): { resourceFieldRef+: { resource: resource } },
+ },
+ '#withMode':: d.fn(help='"Optional: mode bits used to set permissions on this file, must be an octal value between 0000 and 0777 or a decimal value between 0 and 511. YAML accepts both octal and decimal values, JSON requires decimal values for mode bits. If not specified, the volume defaultMode will be used. This might be in conflict with other options that affect the file mode, like fsGroup, and the result can be other mode bits set."', args=[d.arg(name='mode', type=d.T.integer)]),
+ withMode(mode): { mode: mode },
+ '#withPath':: d.fn(help="\"Required: Path is the relative path name of the file to be created. Must not be absolute or contain the '..' path. Must be utf-8 encoded. The first item of the relative path must not start with '..'\"", args=[d.arg(name='path', type=d.T.string)]),
+ withPath(path): { path: path },
+ },
+ '#withDefaultMode':: d.fn(help='"Optional: mode bits to use on created files by default. Must be a Optional: mode bits used to set permissions on created files by default. Must be an octal value between 0000 and 0777 or a decimal value between 0 and 511. YAML accepts both octal and decimal values, JSON requires decimal values for mode bits. Defaults to 0644. Directories within the path are not affected by this setting. This might be in conflict with other options that affect the file mode, like fsGroup, and the result can be other mode bits set."', args=[d.arg(name='defaultMode', type=d.T.integer)]),
+ withDefaultMode(defaultMode): { downwardAPI+: { defaultMode: defaultMode } },
+ '#withItems':: d.fn(help='"Items is a list of downward API volume file"', args=[d.arg(name='items', type=d.T.array)]),
+ withItems(items): { downwardAPI+: { items: if std.isArray(v=items) then items else [items] } },
+ '#withItemsMixin':: d.fn(help='"Items is a list of downward API volume file"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='items', type=d.T.array)]),
+ withItemsMixin(items): { downwardAPI+: { items+: if std.isArray(v=items) then items else [items] } },
+ },
+ '#emptyDir':: d.obj(help="\"emptyDir represents a temporary directory that shares a pod's lifetime. More info: https://kubernetes.io/docs/concepts/storage/volumes#emptydir\""),
+ emptyDir: {
+ '#withMedium':: d.fn(help="\"medium represents what type of storage medium should back this directory. The default is \\\"\\\" which means to use the node's default medium. Must be an empty string (default) or Memory. More info: https://kubernetes.io/docs/concepts/storage/volumes#emptydir\"", args=[d.arg(name='medium', type=d.T.string)]),
+ withMedium(medium): { emptyDir+: { medium: medium } },
+ '#withSizeLimit':: d.fn(help='"sizeLimit is the total amount of local storage required for this EmptyDir volume. The size limit is also applicable for memory medium. The maximum usage on memory medium EmptyDir would be the minimum value between the SizeLimit specified here and the sum of memory limits of all containers in a pod. The default is nil which means that the limit is undefined. More info: https://kubernetes.io/docs/concepts/storage/volumes#emptydir"', args=[d.arg(name='sizeLimit', type=d.T.any)]),
+ withSizeLimit(sizeLimit): { emptyDir+: { sizeLimit: sizeLimit } },
+ },
+ '#ephemeral':: d.obj(help="\"ephemeral represents a volume that is handled by a cluster storage driver. The volume's lifecycle is tied to the pod that defines it - it will be created before the pod starts, and deleted when the pod is removed. \\n Use this if: a) the volume is only needed while the pod runs, b) features of normal volumes like restoring from snapshot or capacity tracking are needed, c) the storage driver is specified through a storage class, and d) the storage driver supports dynamic volume provisioning through a PersistentVolumeClaim (see EphemeralVolumeSource for more information on the connection between this volume type and PersistentVolumeClaim). \\n Use PersistentVolumeClaim or one of the vendor-specific APIs for volumes that persist for longer than the lifecycle of an individual pod. \\n Use CSI for light-weight local ephemeral volumes if the CSI driver is meant to be used that way - see the documentation of the driver for more information. \\n A pod can use both types of ephemeral volumes and persistent volumes at the same time.\""),
+ ephemeral: {
+ '#volumeClaimTemplate':: d.obj(help='"Will be used to create a stand-alone PVC to provision the volume. The pod in which this EphemeralVolumeSource is embedded will be the owner of the PVC, i.e. the PVC will be deleted together with the pod. The name of the PVC will be `-` where `` is the name from the `PodSpec.Volumes` array entry. Pod validation will reject the pod if the concatenated name is not valid for a PVC (for example, too long). \\n An existing PVC with that name that is not owned by the pod will *not* be used for the pod to avoid using an unrelated volume by mistake. Starting the pod is then blocked until the unrelated PVC is removed. If such a pre-created PVC is meant to be used by the pod, the PVC has to updated with an owner reference to the pod once the pod exists. Normally this should not be necessary, but it may be useful when manually reconstructing a broken cluster. \\n This field is read-only and no changes will be made by Kubernetes to the PVC after it has been created. \\n Required, must not be nil."'),
+ volumeClaimTemplate: {
+ '#spec':: d.obj(help='"The specification for the PersistentVolumeClaim. The entire content is copied unchanged into the PVC that gets created from this template. The same fields as in a PersistentVolumeClaim are also valid here."'),
+ spec: {
+ '#dataSource':: d.obj(help='"dataSource field can be used to specify either: * An existing VolumeSnapshot object (snapshot.storage.k8s.io/VolumeSnapshot) * An existing PVC (PersistentVolumeClaim) If the provisioner or an external controller can support the specified data source, it will create a new volume based on the contents of the specified data source. When the AnyVolumeDataSource feature gate is enabled, dataSource contents will be copied to dataSourceRef, and dataSourceRef contents will be copied to dataSource when dataSourceRef.namespace is not specified. If the namespace is specified, then dataSourceRef will not be copied to dataSource."'),
+ dataSource: {
+ '#withApiGroup':: d.fn(help='"APIGroup is the group for the resource being referenced. If APIGroup is not specified, the specified Kind must be in the core API group. For any other third-party types, APIGroup is required."', args=[d.arg(name='apiGroup', type=d.T.string)]),
+ withApiGroup(apiGroup): { ephemeral+: { volumeClaimTemplate+: { spec+: { dataSource+: { apiGroup: apiGroup } } } } },
+ '#withKind':: d.fn(help='"Kind is the type of resource being referenced"', args=[d.arg(name='kind', type=d.T.string)]),
+ withKind(kind): { ephemeral+: { volumeClaimTemplate+: { spec+: { dataSource+: { kind: kind } } } } },
+ '#withName':: d.fn(help='"Name is the name of resource being referenced"', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { ephemeral+: { volumeClaimTemplate+: { spec+: { dataSource+: { name: name } } } } },
+ },
+ '#dataSourceRef':: d.obj(help="\"dataSourceRef specifies the object from which to populate the volume with data, if a non-empty volume is desired. This may be any object from a non-empty API group (non core object) or a PersistentVolumeClaim object. When this field is specified, volume binding will only succeed if the type of the specified object matches some installed volume populator or dynamic provisioner. This field will replace the functionality of the dataSource field and as such if both fields are non-empty, they must have the same value. For backwards compatibility, when namespace isn't specified in dataSourceRef, both fields (dataSource and dataSourceRef) will be set to the same value automatically if one of them is empty and the other is non-empty. When namespace is specified in dataSourceRef, dataSource isn't set to the same value and must be empty. There are three important differences between dataSource and dataSourceRef: * While dataSource only allows two specific types of objects, dataSourceRef allows any non-core object, as well as PersistentVolumeClaim objects. * While dataSource ignores disallowed values (dropping them), dataSourceRef preserves all values, and generates an error if a disallowed value is specified. * While dataSource only allows local objects, dataSourceRef allows objects in any namespaces. (Beta) Using this field requires the AnyVolumeDataSource feature gate to be enabled. (Alpha) Using the namespace field of dataSourceRef requires the CrossNamespaceVolumeDataSource feature gate to be enabled.\""),
+ dataSourceRef: {
+ '#withApiGroup':: d.fn(help='"APIGroup is the group for the resource being referenced. If APIGroup is not specified, the specified Kind must be in the core API group. For any other third-party types, APIGroup is required."', args=[d.arg(name='apiGroup', type=d.T.string)]),
+ withApiGroup(apiGroup): { ephemeral+: { volumeClaimTemplate+: { spec+: { dataSourceRef+: { apiGroup: apiGroup } } } } },
+ '#withKind':: d.fn(help='"Kind is the type of resource being referenced"', args=[d.arg(name='kind', type=d.T.string)]),
+ withKind(kind): { ephemeral+: { volumeClaimTemplate+: { spec+: { dataSourceRef+: { kind: kind } } } } },
+ '#withName':: d.fn(help='"Name is the name of resource being referenced"', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { ephemeral+: { volumeClaimTemplate+: { spec+: { dataSourceRef+: { name: name } } } } },
+ '#withNamespace':: d.fn(help="\"Namespace is the namespace of resource being referenced Note that when a namespace is specified, a gateway.networking.k8s.io/ReferenceGrant object is required in the referent namespace to allow that namespace's owner to accept the reference. See the ReferenceGrant documentation for details. (Alpha) This field requires the CrossNamespaceVolumeDataSource feature gate to be enabled.\"", args=[d.arg(name='namespace', type=d.T.string)]),
+ withNamespace(namespace): { ephemeral+: { volumeClaimTemplate+: { spec+: { dataSourceRef+: { namespace: namespace } } } } },
+ },
+ '#resources':: d.obj(help='"resources represents the minimum resources the volume should have. If RecoverVolumeExpansionFailure feature is enabled users are allowed to specify resource requirements that are lower than previous value but must still be higher than capacity recorded in the status field of the claim. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#resources"'),
+ resources: {
+ '#claims':: d.obj(help='"Claims lists the names of resources, defined in spec.resourceClaims, that are used by this container. \\n This is an alpha field and requires enabling the DynamicResourceAllocation feature gate. \\n This field is immutable. It can only be set for containers."'),
+ claims: {
+ '#withName':: d.fn(help='"Name must match the name of one entry in pod.spec.resourceClaims of the Pod where this field is used. It makes that resource available inside a container."', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { name: name },
+ },
+ '#withClaims':: d.fn(help='"Claims lists the names of resources, defined in spec.resourceClaims, that are used by this container. \\n This is an alpha field and requires enabling the DynamicResourceAllocation feature gate. \\n This field is immutable. It can only be set for containers."', args=[d.arg(name='claims', type=d.T.array)]),
+ withClaims(claims): { ephemeral+: { volumeClaimTemplate+: { spec+: { resources+: { claims: if std.isArray(v=claims) then claims else [claims] } } } } },
+ '#withClaimsMixin':: d.fn(help='"Claims lists the names of resources, defined in spec.resourceClaims, that are used by this container. \\n This is an alpha field and requires enabling the DynamicResourceAllocation feature gate. \\n This field is immutable. It can only be set for containers."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='claims', type=d.T.array)]),
+ withClaimsMixin(claims): { ephemeral+: { volumeClaimTemplate+: { spec+: { resources+: { claims+: if std.isArray(v=claims) then claims else [claims] } } } } },
+ '#withLimits':: d.fn(help='"Limits describes the maximum amount of compute resources allowed. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"', args=[d.arg(name='limits', type=d.T.object)]),
+ withLimits(limits): { ephemeral+: { volumeClaimTemplate+: { spec+: { resources+: { limits: limits } } } } },
+ '#withLimitsMixin':: d.fn(help='"Limits describes the maximum amount of compute resources allowed. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='limits', type=d.T.object)]),
+ withLimitsMixin(limits): { ephemeral+: { volumeClaimTemplate+: { spec+: { resources+: { limits+: limits } } } } },
+ '#withRequests':: d.fn(help='"Requests describes the minimum amount of compute resources required. If Requests is omitted for a container, it defaults to Limits if that is explicitly specified, otherwise to an implementation-defined value. Requests cannot exceed Limits. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"', args=[d.arg(name='requests', type=d.T.object)]),
+ withRequests(requests): { ephemeral+: { volumeClaimTemplate+: { spec+: { resources+: { requests: requests } } } } },
+ '#withRequestsMixin':: d.fn(help='"Requests describes the minimum amount of compute resources required. If Requests is omitted for a container, it defaults to Limits if that is explicitly specified, otherwise to an implementation-defined value. Requests cannot exceed Limits. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='requests', type=d.T.object)]),
+ withRequestsMixin(requests): { ephemeral+: { volumeClaimTemplate+: { spec+: { resources+: { requests+: requests } } } } },
+ },
+ '#selector':: d.obj(help='"selector is a label query over volumes to consider for binding."'),
+ selector: {
+ '#matchExpressions':: d.obj(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."'),
+ matchExpressions: {
+ '#withKey':: d.fn(help='"key is the label key that the selector applies to."', args=[d.arg(name='key', type=d.T.string)]),
+ withKey(key): { key: key },
+ '#withOperator':: d.fn(help="\"operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist.\"", args=[d.arg(name='operator', type=d.T.string)]),
+ withOperator(operator): { operator: operator },
+ '#withValues':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."', args=[d.arg(name='values', type=d.T.array)]),
+ withValues(values): { values: if std.isArray(v=values) then values else [values] },
+ '#withValuesMixin':: d.fn(help='"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='values', type=d.T.array)]),
+ withValuesMixin(values): { values+: if std.isArray(v=values) then values else [values] },
+ },
+ '#withMatchExpressions':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."', args=[d.arg(name='matchExpressions', type=d.T.array)]),
+ withMatchExpressions(matchExpressions): { ephemeral+: { volumeClaimTemplate+: { spec+: { selector+: { matchExpressions: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } } } },
+ '#withMatchExpressionsMixin':: d.fn(help='"matchExpressions is a list of label selector requirements. The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchExpressions', type=d.T.array)]),
+ withMatchExpressionsMixin(matchExpressions): { ephemeral+: { volumeClaimTemplate+: { spec+: { selector+: { matchExpressions+: if std.isArray(v=matchExpressions) then matchExpressions else [matchExpressions] } } } } },
+ '#withMatchLabels':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."', args=[d.arg(name='matchLabels', type=d.T.object)]),
+ withMatchLabels(matchLabels): { ephemeral+: { volumeClaimTemplate+: { spec+: { selector+: { matchLabels: matchLabels } } } } },
+ '#withMatchLabelsMixin':: d.fn(help='"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \\"key\\", the operator is \\"In\\", and the values array contains only \\"value\\". The requirements are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='matchLabels', type=d.T.object)]),
+ withMatchLabelsMixin(matchLabels): { ephemeral+: { volumeClaimTemplate+: { spec+: { selector+: { matchLabels+: matchLabels } } } } },
+ },
+ '#withAccessModes':: d.fn(help='"accessModes contains the desired access modes the volume should have. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#access-modes-1"', args=[d.arg(name='accessModes', type=d.T.array)]),
+ withAccessModes(accessModes): { ephemeral+: { volumeClaimTemplate+: { spec+: { accessModes: if std.isArray(v=accessModes) then accessModes else [accessModes] } } } },
+ '#withAccessModesMixin':: d.fn(help='"accessModes contains the desired access modes the volume should have. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#access-modes-1"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='accessModes', type=d.T.array)]),
+ withAccessModesMixin(accessModes): { ephemeral+: { volumeClaimTemplate+: { spec+: { accessModes+: if std.isArray(v=accessModes) then accessModes else [accessModes] } } } },
+ '#withStorageClassName':: d.fn(help='"storageClassName is the name of the StorageClass required by the claim. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#class-1"', args=[d.arg(name='storageClassName', type=d.T.string)]),
+ withStorageClassName(storageClassName): { ephemeral+: { volumeClaimTemplate+: { spec+: { storageClassName: storageClassName } } } },
+ '#withVolumeMode':: d.fn(help='"volumeMode defines what type of volume is required by the claim. Value of Filesystem is implied when not included in claim spec."', args=[d.arg(name='volumeMode', type=d.T.string)]),
+ withVolumeMode(volumeMode): { ephemeral+: { volumeClaimTemplate+: { spec+: { volumeMode: volumeMode } } } },
+ '#withVolumeName':: d.fn(help='"volumeName is the binding reference to the PersistentVolume backing this claim."', args=[d.arg(name='volumeName', type=d.T.string)]),
+ withVolumeName(volumeName): { ephemeral+: { volumeClaimTemplate+: { spec+: { volumeName: volumeName } } } },
+ },
+ '#withMetadata':: d.fn(help='"May contain labels and annotations that will be copied into the PVC when creating it. No other fields are allowed and will be rejected during validation."', args=[d.arg(name='metadata', type=d.T.object)]),
+ withMetadata(metadata): { ephemeral+: { volumeClaimTemplate+: { metadata: metadata } } },
+ '#withMetadataMixin':: d.fn(help='"May contain labels and annotations that will be copied into the PVC when creating it. No other fields are allowed and will be rejected during validation."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='metadata', type=d.T.object)]),
+ withMetadataMixin(metadata): { ephemeral+: { volumeClaimTemplate+: { metadata+: metadata } } },
+ },
+ },
+ '#fc':: d.obj(help="\"fc represents a Fibre Channel resource that is attached to a kubelet's host machine and then exposed to the pod.\""),
+ fc: {
+ '#withFsType':: d.fn(help='"fsType is the filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. \\"ext4\\", \\"xfs\\", \\"ntfs\\". Implicitly inferred to be \\"ext4\\" if unspecified. TODO: how do we prevent errors in the filesystem from compromising the machine"', args=[d.arg(name='fsType', type=d.T.string)]),
+ withFsType(fsType): { fc+: { fsType: fsType } },
+ '#withLun':: d.fn(help='"lun is Optional: FC target lun number"', args=[d.arg(name='lun', type=d.T.integer)]),
+ withLun(lun): { fc+: { lun: lun } },
+ '#withReadOnly':: d.fn(help='"readOnly is Optional: Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts."', args=[d.arg(name='readOnly', type=d.T.boolean)]),
+ withReadOnly(readOnly): { fc+: { readOnly: readOnly } },
+ '#withTargetWWNs':: d.fn(help='"targetWWNs is Optional: FC target worldwide names (WWNs)"', args=[d.arg(name='targetWWNs', type=d.T.array)]),
+ withTargetWWNs(targetWWNs): { fc+: { targetWWNs: if std.isArray(v=targetWWNs) then targetWWNs else [targetWWNs] } },
+ '#withTargetWWNsMixin':: d.fn(help='"targetWWNs is Optional: FC target worldwide names (WWNs)"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='targetWWNs', type=d.T.array)]),
+ withTargetWWNsMixin(targetWWNs): { fc+: { targetWWNs+: if std.isArray(v=targetWWNs) then targetWWNs else [targetWWNs] } },
+ '#withWwids':: d.fn(help='"wwids Optional: FC volume world wide identifiers (wwids) Either wwids or combination of targetWWNs and lun must be set, but not both simultaneously."', args=[d.arg(name='wwids', type=d.T.array)]),
+ withWwids(wwids): { fc+: { wwids: if std.isArray(v=wwids) then wwids else [wwids] } },
+ '#withWwidsMixin':: d.fn(help='"wwids Optional: FC volume world wide identifiers (wwids) Either wwids or combination of targetWWNs and lun must be set, but not both simultaneously."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='wwids', type=d.T.array)]),
+ withWwidsMixin(wwids): { fc+: { wwids+: if std.isArray(v=wwids) then wwids else [wwids] } },
+ },
+ '#flexVolume':: d.obj(help='"flexVolume represents a generic volume resource that is provisioned/attached using an exec based plugin."'),
+ flexVolume: {
+ '#secretRef':: d.obj(help='"secretRef is Optional: secretRef is reference to the secret object containing sensitive information to pass to the plugin scripts. This may be empty if no secret object is specified. If the secret object contains more than one secret, all secrets are passed to the plugin scripts."'),
+ secretRef: {
+ '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { flexVolume+: { secretRef+: { name: name } } },
+ },
+ '#withDriver':: d.fn(help='"driver is the name of the driver to use for this volume."', args=[d.arg(name='driver', type=d.T.string)]),
+ withDriver(driver): { flexVolume+: { driver: driver } },
+ '#withFsType':: d.fn(help='"fsType is the filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. \\"ext4\\", \\"xfs\\", \\"ntfs\\". The default filesystem depends on FlexVolume script."', args=[d.arg(name='fsType', type=d.T.string)]),
+ withFsType(fsType): { flexVolume+: { fsType: fsType } },
+ '#withOptions':: d.fn(help='"options is Optional: this field holds extra command options if any."', args=[d.arg(name='options', type=d.T.object)]),
+ withOptions(options): { flexVolume+: { options: options } },
+ '#withOptionsMixin':: d.fn(help='"options is Optional: this field holds extra command options if any."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='options', type=d.T.object)]),
+ withOptionsMixin(options): { flexVolume+: { options+: options } },
+ '#withReadOnly':: d.fn(help='"readOnly is Optional: defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts."', args=[d.arg(name='readOnly', type=d.T.boolean)]),
+ withReadOnly(readOnly): { flexVolume+: { readOnly: readOnly } },
+ },
+ '#flocker':: d.obj(help="\"flocker represents a Flocker volume attached to a kubelet's host machine. This depends on the Flocker control service being running\""),
+ flocker: {
+ '#withDatasetName':: d.fn(help='"datasetName is Name of the dataset stored as metadata -> name on the dataset for Flocker should be considered as deprecated"', args=[d.arg(name='datasetName', type=d.T.string)]),
+ withDatasetName(datasetName): { flocker+: { datasetName: datasetName } },
+ '#withDatasetUUID':: d.fn(help='"datasetUUID is the UUID of the dataset. This is unique identifier of a Flocker dataset"', args=[d.arg(name='datasetUUID', type=d.T.string)]),
+ withDatasetUUID(datasetUUID): { flocker+: { datasetUUID: datasetUUID } },
+ },
+ '#gcePersistentDisk':: d.obj(help="\"gcePersistentDisk represents a GCE Disk resource that is attached to a kubelet's host machine and then exposed to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk\""),
+ gcePersistentDisk: {
+ '#withFsType':: d.fn(help='"fsType is filesystem type of the volume that you want to mount. Tip: Ensure that the filesystem type is supported by the host operating system. Examples: \\"ext4\\", \\"xfs\\", \\"ntfs\\". Implicitly inferred to be \\"ext4\\" if unspecified. More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk TODO: how do we prevent errors in the filesystem from compromising the machine"', args=[d.arg(name='fsType', type=d.T.string)]),
+ withFsType(fsType): { gcePersistentDisk+: { fsType: fsType } },
+ '#withPartition':: d.fn(help='"partition is the partition in the volume that you want to mount. If omitted, the default is to mount by volume name. Examples: For volume /dev/sda1, you specify the partition as \\"1\\". Similarly, the volume partition for /dev/sda is \\"0\\" (or you can leave the property empty). More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk"', args=[d.arg(name='partition', type=d.T.integer)]),
+ withPartition(partition): { gcePersistentDisk+: { partition: partition } },
+ '#withPdName':: d.fn(help='"pdName is unique name of the PD resource in GCE. Used to identify the disk in GCE. More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk"', args=[d.arg(name='pdName', type=d.T.string)]),
+ withPdName(pdName): { gcePersistentDisk+: { pdName: pdName } },
+ '#withReadOnly':: d.fn(help='"readOnly here will force the ReadOnly setting in VolumeMounts. Defaults to false. More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk"', args=[d.arg(name='readOnly', type=d.T.boolean)]),
+ withReadOnly(readOnly): { gcePersistentDisk+: { readOnly: readOnly } },
+ },
+ '#gitRepo':: d.obj(help="\"gitRepo represents a git repository at a particular revision. DEPRECATED: GitRepo is deprecated. To provision a container with a git repo, mount an EmptyDir into an InitContainer that clones the repo using git, then mount the EmptyDir into the Pod's container.\""),
+ gitRepo: {
+ '#withDirectory':: d.fn(help="\"directory is the target directory name. Must not contain or start with '..'. If '.' is supplied, the volume directory will be the git repository. Otherwise, if specified, the volume will contain the git repository in the subdirectory with the given name.\"", args=[d.arg(name='directory', type=d.T.string)]),
+ withDirectory(directory): { gitRepo+: { directory: directory } },
+ '#withRepository':: d.fn(help='"repository is the URL"', args=[d.arg(name='repository', type=d.T.string)]),
+ withRepository(repository): { gitRepo+: { repository: repository } },
+ '#withRevision':: d.fn(help='"revision is the commit hash for the specified revision."', args=[d.arg(name='revision', type=d.T.string)]),
+ withRevision(revision): { gitRepo+: { revision: revision } },
+ },
+ '#glusterfs':: d.obj(help="\"glusterfs represents a Glusterfs mount on the host that shares a pod's lifetime. More info: https://examples.k8s.io/volumes/glusterfs/README.md\""),
+ glusterfs: {
+ '#withEndpoints':: d.fn(help='"endpoints is the endpoint name that details Glusterfs topology. More info: https://examples.k8s.io/volumes/glusterfs/README.md#create-a-pod"', args=[d.arg(name='endpoints', type=d.T.string)]),
+ withEndpoints(endpoints): { glusterfs+: { endpoints: endpoints } },
+ '#withPath':: d.fn(help='"path is the Glusterfs volume path. More info: https://examples.k8s.io/volumes/glusterfs/README.md#create-a-pod"', args=[d.arg(name='path', type=d.T.string)]),
+ withPath(path): { glusterfs+: { path: path } },
+ '#withReadOnly':: d.fn(help='"readOnly here will force the Glusterfs volume to be mounted with read-only permissions. Defaults to false. More info: https://examples.k8s.io/volumes/glusterfs/README.md#create-a-pod"', args=[d.arg(name='readOnly', type=d.T.boolean)]),
+ withReadOnly(readOnly): { glusterfs+: { readOnly: readOnly } },
+ },
+ '#hostPath':: d.obj(help='"hostPath represents a pre-existing file or directory on the host machine that is directly exposed to the container. This is generally used for system agents or other privileged things that are allowed to see the host machine. Most containers will NOT need this. More info: https://kubernetes.io/docs/concepts/storage/volumes#hostpath --- TODO(jonesdl) We need to restrict who can use host directory mounts and who can/can not mount host directories as read/write."'),
+ hostPath: {
+ '#withPath':: d.fn(help='"path of the directory on the host. If the path is a symlink, it will follow the link to the real path. More info: https://kubernetes.io/docs/concepts/storage/volumes#hostpath"', args=[d.arg(name='path', type=d.T.string)]),
+ withPath(path): { hostPath+: { path: path } },
+ '#withType':: d.fn(help='"type for HostPath Volume Defaults to \\"\\" More info: https://kubernetes.io/docs/concepts/storage/volumes#hostpath"', args=[d.arg(name='type', type=d.T.string)]),
+ withType(type): { hostPath+: { type: type } },
+ },
+ '#iscsi':: d.obj(help="\"iscsi represents an ISCSI Disk resource that is attached to a kubelet's host machine and then exposed to the pod. More info: https://examples.k8s.io/volumes/iscsi/README.md\""),
+ iscsi: {
+ '#secretRef':: d.obj(help='"secretRef is the CHAP Secret for iSCSI target and initiator authentication"'),
+ secretRef: {
+ '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { iscsi+: { secretRef+: { name: name } } },
+ },
+ '#withChapAuthDiscovery':: d.fn(help='"chapAuthDiscovery defines whether support iSCSI Discovery CHAP authentication"', args=[d.arg(name='chapAuthDiscovery', type=d.T.boolean)]),
+ withChapAuthDiscovery(chapAuthDiscovery): { iscsi+: { chapAuthDiscovery: chapAuthDiscovery } },
+ '#withChapAuthSession':: d.fn(help='"chapAuthSession defines whether support iSCSI Session CHAP authentication"', args=[d.arg(name='chapAuthSession', type=d.T.boolean)]),
+ withChapAuthSession(chapAuthSession): { iscsi+: { chapAuthSession: chapAuthSession } },
+ '#withFsType':: d.fn(help='"fsType is the filesystem type of the volume that you want to mount. Tip: Ensure that the filesystem type is supported by the host operating system. Examples: \\"ext4\\", \\"xfs\\", \\"ntfs\\". Implicitly inferred to be \\"ext4\\" if unspecified. More info: https://kubernetes.io/docs/concepts/storage/volumes#iscsi TODO: how do we prevent errors in the filesystem from compromising the machine"', args=[d.arg(name='fsType', type=d.T.string)]),
+ withFsType(fsType): { iscsi+: { fsType: fsType } },
+ '#withInitiatorName':: d.fn(help='"initiatorName is the custom iSCSI Initiator Name. If initiatorName is specified with iscsiInterface simultaneously, new iSCSI interface : will be created for the connection."', args=[d.arg(name='initiatorName', type=d.T.string)]),
+ withInitiatorName(initiatorName): { iscsi+: { initiatorName: initiatorName } },
+ '#withIqn':: d.fn(help='"iqn is the target iSCSI Qualified Name."', args=[d.arg(name='iqn', type=d.T.string)]),
+ withIqn(iqn): { iscsi+: { iqn: iqn } },
+ '#withIscsiInterface':: d.fn(help="\"iscsiInterface is the interface Name that uses an iSCSI transport. Defaults to 'default' (tcp).\"", args=[d.arg(name='iscsiInterface', type=d.T.string)]),
+ withIscsiInterface(iscsiInterface): { iscsi+: { iscsiInterface: iscsiInterface } },
+ '#withLun':: d.fn(help='"lun represents iSCSI Target Lun number."', args=[d.arg(name='lun', type=d.T.integer)]),
+ withLun(lun): { iscsi+: { lun: lun } },
+ '#withPortals':: d.fn(help='"portals is the iSCSI Target Portal List. The portal is either an IP or ip_addr:port if the port is other than default (typically TCP ports 860 and 3260)."', args=[d.arg(name='portals', type=d.T.array)]),
+ withPortals(portals): { iscsi+: { portals: if std.isArray(v=portals) then portals else [portals] } },
+ '#withPortalsMixin':: d.fn(help='"portals is the iSCSI Target Portal List. The portal is either an IP or ip_addr:port if the port is other than default (typically TCP ports 860 and 3260)."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='portals', type=d.T.array)]),
+ withPortalsMixin(portals): { iscsi+: { portals+: if std.isArray(v=portals) then portals else [portals] } },
+ '#withReadOnly':: d.fn(help='"readOnly here will force the ReadOnly setting in VolumeMounts. Defaults to false."', args=[d.arg(name='readOnly', type=d.T.boolean)]),
+ withReadOnly(readOnly): { iscsi+: { readOnly: readOnly } },
+ '#withTargetPortal':: d.fn(help='"targetPortal is iSCSI Target Portal. The Portal is either an IP or ip_addr:port if the port is other than default (typically TCP ports 860 and 3260)."', args=[d.arg(name='targetPortal', type=d.T.string)]),
+ withTargetPortal(targetPortal): { iscsi+: { targetPortal: targetPortal } },
+ },
+ '#nfs':: d.obj(help="\"nfs represents an NFS mount on the host that shares a pod's lifetime More info: https://kubernetes.io/docs/concepts/storage/volumes#nfs\""),
+ nfs: {
+ '#withPath':: d.fn(help='"path that is exported by the NFS server. More info: https://kubernetes.io/docs/concepts/storage/volumes#nfs"', args=[d.arg(name='path', type=d.T.string)]),
+ withPath(path): { nfs+: { path: path } },
+ '#withReadOnly':: d.fn(help='"readOnly here will force the NFS export to be mounted with read-only permissions. Defaults to false. More info: https://kubernetes.io/docs/concepts/storage/volumes#nfs"', args=[d.arg(name='readOnly', type=d.T.boolean)]),
+ withReadOnly(readOnly): { nfs+: { readOnly: readOnly } },
+ '#withServer':: d.fn(help='"server is the hostname or IP address of the NFS server. More info: https://kubernetes.io/docs/concepts/storage/volumes#nfs"', args=[d.arg(name='server', type=d.T.string)]),
+ withServer(server): { nfs+: { server: server } },
+ },
+ '#persistentVolumeClaim':: d.obj(help='"persistentVolumeClaimVolumeSource represents a reference to a PersistentVolumeClaim in the same namespace. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#persistentvolumeclaims"'),
+ persistentVolumeClaim: {
+ '#withClaimName':: d.fn(help='"claimName is the name of a PersistentVolumeClaim in the same namespace as the pod using this volume. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#persistentvolumeclaims"', args=[d.arg(name='claimName', type=d.T.string)]),
+ withClaimName(claimName): { persistentVolumeClaim+: { claimName: claimName } },
+ '#withReadOnly':: d.fn(help='"readOnly Will force the ReadOnly setting in VolumeMounts. Default false."', args=[d.arg(name='readOnly', type=d.T.boolean)]),
+ withReadOnly(readOnly): { persistentVolumeClaim+: { readOnly: readOnly } },
+ },
+ '#photonPersistentDisk':: d.obj(help='"photonPersistentDisk represents a PhotonController persistent disk attached and mounted on kubelets host machine"'),
+ photonPersistentDisk: {
+ '#withFsType':: d.fn(help='"fsType is the filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. \\"ext4\\", \\"xfs\\", \\"ntfs\\". Implicitly inferred to be \\"ext4\\" if unspecified."', args=[d.arg(name='fsType', type=d.T.string)]),
+ withFsType(fsType): { photonPersistentDisk+: { fsType: fsType } },
+ '#withPdID':: d.fn(help='"pdID is the ID that identifies Photon Controller persistent disk"', args=[d.arg(name='pdID', type=d.T.string)]),
+ withPdID(pdID): { photonPersistentDisk+: { pdID: pdID } },
+ },
+ '#portworxVolume':: d.obj(help='"portworxVolume represents a portworx volume attached and mounted on kubelets host machine"'),
+ portworxVolume: {
+ '#withFsType':: d.fn(help='"fSType represents the filesystem type to mount Must be a filesystem type supported by the host operating system. Ex. \\"ext4\\", \\"xfs\\". Implicitly inferred to be \\"ext4\\" if unspecified."', args=[d.arg(name='fsType', type=d.T.string)]),
+ withFsType(fsType): { portworxVolume+: { fsType: fsType } },
+ '#withReadOnly':: d.fn(help='"readOnly defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts."', args=[d.arg(name='readOnly', type=d.T.boolean)]),
+ withReadOnly(readOnly): { portworxVolume+: { readOnly: readOnly } },
+ '#withVolumeID':: d.fn(help='"volumeID uniquely identifies a Portworx volume"', args=[d.arg(name='volumeID', type=d.T.string)]),
+ withVolumeID(volumeID): { portworxVolume+: { volumeID: volumeID } },
+ },
+ '#projected':: d.obj(help='"projected items for all in one resources secrets, configmaps, and downward API"'),
+ projected: {
+ '#sources':: d.obj(help='"sources is the list of volume projections"'),
+ sources: {
+ '#configMap':: d.obj(help='"configMap information about the configMap data to project"'),
+ configMap: {
+ '#items':: d.obj(help="\"items if unspecified, each key-value pair in the Data field of the referenced ConfigMap will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the ConfigMap, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'.\""),
+ items: {
+ '#withKey':: d.fn(help='"key is the key to project."', args=[d.arg(name='key', type=d.T.string)]),
+ withKey(key): { key: key },
+ '#withMode':: d.fn(help='"mode is Optional: mode bits used to set permissions on this file. Must be an octal value between 0000 and 0777 or a decimal value between 0 and 511. YAML accepts both octal and decimal values, JSON requires decimal values for mode bits. If not specified, the volume defaultMode will be used. This might be in conflict with other options that affect the file mode, like fsGroup, and the result can be other mode bits set."', args=[d.arg(name='mode', type=d.T.integer)]),
+ withMode(mode): { mode: mode },
+ '#withPath':: d.fn(help="\"path is the relative path of the file to map the key to. May not be an absolute path. May not contain the path element '..'. May not start with the string '..'.\"", args=[d.arg(name='path', type=d.T.string)]),
+ withPath(path): { path: path },
+ },
+ '#withItems':: d.fn(help="\"items if unspecified, each key-value pair in the Data field of the referenced ConfigMap will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the ConfigMap, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'.\"", args=[d.arg(name='items', type=d.T.array)]),
+ withItems(items): { configMap+: { items: if std.isArray(v=items) then items else [items] } },
+ '#withItemsMixin':: d.fn(help="\"items if unspecified, each key-value pair in the Data field of the referenced ConfigMap will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the ConfigMap, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='items', type=d.T.array)]),
+ withItemsMixin(items): { configMap+: { items+: if std.isArray(v=items) then items else [items] } },
+ '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { configMap+: { name: name } },
+ '#withOptional':: d.fn(help='"optional specify whether the ConfigMap or its keys must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]),
+ withOptional(optional): { configMap+: { optional: optional } },
+ },
+ '#downwardAPI':: d.obj(help='"downwardAPI information about the downwardAPI data to project"'),
+ downwardAPI: {
+ '#items':: d.obj(help='"Items is a list of DownwardAPIVolume file"'),
+ items: {
+ '#fieldRef':: d.obj(help='"Required: Selects a field of the pod: only annotations, labels, name and namespace are supported."'),
+ fieldRef: {
+ '#withApiVersion':: d.fn(help='"Version of the schema the FieldPath is written in terms of, defaults to \\"v1\\"."', args=[d.arg(name='apiVersion', type=d.T.string)]),
+ withApiVersion(apiVersion): { fieldRef+: { apiVersion: apiVersion } },
+ '#withFieldPath':: d.fn(help='"Path of the field to select in the specified API version."', args=[d.arg(name='fieldPath', type=d.T.string)]),
+ withFieldPath(fieldPath): { fieldRef+: { fieldPath: fieldPath } },
+ },
+ '#resourceFieldRef':: d.obj(help='"Selects a resource of the container: only resources limits and requests (limits.cpu, limits.memory, requests.cpu and requests.memory) are currently supported."'),
+ resourceFieldRef: {
+ '#withContainerName':: d.fn(help='"Container name: required for volumes, optional for env vars"', args=[d.arg(name='containerName', type=d.T.string)]),
+ withContainerName(containerName): { resourceFieldRef+: { containerName: containerName } },
+ '#withDivisor':: d.fn(help='"Specifies the output format of the exposed resources, defaults to \\"1\\', args=[d.arg(name='divisor', type=d.T.any)]),
+ withDivisor(divisor): { resourceFieldRef+: { divisor: divisor } },
+ '#withResource':: d.fn(help='"Required: resource to select"', args=[d.arg(name='resource', type=d.T.string)]),
+ withResource(resource): { resourceFieldRef+: { resource: resource } },
+ },
+ '#withMode':: d.fn(help='"Optional: mode bits used to set permissions on this file, must be an octal value between 0000 and 0777 or a decimal value between 0 and 511. YAML accepts both octal and decimal values, JSON requires decimal values for mode bits. If not specified, the volume defaultMode will be used. This might be in conflict with other options that affect the file mode, like fsGroup, and the result can be other mode bits set."', args=[d.arg(name='mode', type=d.T.integer)]),
+ withMode(mode): { mode: mode },
+ '#withPath':: d.fn(help="\"Required: Path is the relative path name of the file to be created. Must not be absolute or contain the '..' path. Must be utf-8 encoded. The first item of the relative path must not start with '..'\"", args=[d.arg(name='path', type=d.T.string)]),
+ withPath(path): { path: path },
+ },
+ '#withItems':: d.fn(help='"Items is a list of DownwardAPIVolume file"', args=[d.arg(name='items', type=d.T.array)]),
+ withItems(items): { downwardAPI+: { items: if std.isArray(v=items) then items else [items] } },
+ '#withItemsMixin':: d.fn(help='"Items is a list of DownwardAPIVolume file"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='items', type=d.T.array)]),
+ withItemsMixin(items): { downwardAPI+: { items+: if std.isArray(v=items) then items else [items] } },
+ },
+ '#secret':: d.obj(help='"secret information about the secret data to project"'),
+ secret: {
+ '#items':: d.obj(help="\"items if unspecified, each key-value pair in the Data field of the referenced Secret will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the Secret, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'.\""),
+ items: {
+ '#withKey':: d.fn(help='"key is the key to project."', args=[d.arg(name='key', type=d.T.string)]),
+ withKey(key): { key: key },
+ '#withMode':: d.fn(help='"mode is Optional: mode bits used to set permissions on this file. Must be an octal value between 0000 and 0777 or a decimal value between 0 and 511. YAML accepts both octal and decimal values, JSON requires decimal values for mode bits. If not specified, the volume defaultMode will be used. This might be in conflict with other options that affect the file mode, like fsGroup, and the result can be other mode bits set."', args=[d.arg(name='mode', type=d.T.integer)]),
+ withMode(mode): { mode: mode },
+ '#withPath':: d.fn(help="\"path is the relative path of the file to map the key to. May not be an absolute path. May not contain the path element '..'. May not start with the string '..'.\"", args=[d.arg(name='path', type=d.T.string)]),
+ withPath(path): { path: path },
+ },
+ '#withItems':: d.fn(help="\"items if unspecified, each key-value pair in the Data field of the referenced Secret will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the Secret, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'.\"", args=[d.arg(name='items', type=d.T.array)]),
+ withItems(items): { secret+: { items: if std.isArray(v=items) then items else [items] } },
+ '#withItemsMixin':: d.fn(help="\"items if unspecified, each key-value pair in the Data field of the referenced Secret will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the Secret, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='items', type=d.T.array)]),
+ withItemsMixin(items): { secret+: { items+: if std.isArray(v=items) then items else [items] } },
+ '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { secret+: { name: name } },
+ '#withOptional':: d.fn(help='"optional field specify whether the Secret or its key must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]),
+ withOptional(optional): { secret+: { optional: optional } },
+ },
+ '#serviceAccountToken':: d.obj(help='"serviceAccountToken is information about the serviceAccountToken data to project"'),
+ serviceAccountToken: {
+ '#withAudience':: d.fn(help='"audience is the intended audience of the token. A recipient of a token must identify itself with an identifier specified in the audience of the token, and otherwise should reject the token. The audience defaults to the identifier of the apiserver."', args=[d.arg(name='audience', type=d.T.string)]),
+ withAudience(audience): { serviceAccountToken+: { audience: audience } },
+ '#withExpirationSeconds':: d.fn(help='"expirationSeconds is the requested duration of validity of the service account token. As the token approaches expiration, the kubelet volume plugin will proactively rotate the service account token. The kubelet will start trying to rotate the token if the token is older than 80 percent of its time to live or if the token is older than 24 hours.Defaults to 1 hour and must be at least 10 minutes."', args=[d.arg(name='expirationSeconds', type=d.T.integer)]),
+ withExpirationSeconds(expirationSeconds): { serviceAccountToken+: { expirationSeconds: expirationSeconds } },
+ '#withPath':: d.fn(help='"path is the path relative to the mount point of the file to project the token into."', args=[d.arg(name='path', type=d.T.string)]),
+ withPath(path): { serviceAccountToken+: { path: path } },
+ },
+ },
+ '#withDefaultMode':: d.fn(help='"defaultMode are the mode bits used to set permissions on created files by default. Must be an octal value between 0000 and 0777 or a decimal value between 0 and 511. YAML accepts both octal and decimal values, JSON requires decimal values for mode bits. Directories within the path are not affected by this setting. This might be in conflict with other options that affect the file mode, like fsGroup, and the result can be other mode bits set."', args=[d.arg(name='defaultMode', type=d.T.integer)]),
+ withDefaultMode(defaultMode): { projected+: { defaultMode: defaultMode } },
+ '#withSources':: d.fn(help='"sources is the list of volume projections"', args=[d.arg(name='sources', type=d.T.array)]),
+ withSources(sources): { projected+: { sources: if std.isArray(v=sources) then sources else [sources] } },
+ '#withSourcesMixin':: d.fn(help='"sources is the list of volume projections"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='sources', type=d.T.array)]),
+ withSourcesMixin(sources): { projected+: { sources+: if std.isArray(v=sources) then sources else [sources] } },
+ },
+ '#quobyte':: d.obj(help="\"quobyte represents a Quobyte mount on the host that shares a pod's lifetime\""),
+ quobyte: {
+ '#withGroup':: d.fn(help='"group to map volume access to Default is no group"', args=[d.arg(name='group', type=d.T.string)]),
+ withGroup(group): { quobyte+: { group: group } },
+ '#withReadOnly':: d.fn(help='"readOnly here will force the Quobyte volume to be mounted with read-only permissions. Defaults to false."', args=[d.arg(name='readOnly', type=d.T.boolean)]),
+ withReadOnly(readOnly): { quobyte+: { readOnly: readOnly } },
+ '#withRegistry':: d.fn(help='"registry represents a single or multiple Quobyte Registry services specified as a string as host:port pair (multiple entries are separated with commas) which acts as the central registry for volumes"', args=[d.arg(name='registry', type=d.T.string)]),
+ withRegistry(registry): { quobyte+: { registry: registry } },
+ '#withTenant':: d.fn(help='"tenant owning the given Quobyte volume in the Backend Used with dynamically provisioned Quobyte volumes, value is set by the plugin"', args=[d.arg(name='tenant', type=d.T.string)]),
+ withTenant(tenant): { quobyte+: { tenant: tenant } },
+ '#withUser':: d.fn(help='"user to map volume access to Defaults to serivceaccount user"', args=[d.arg(name='user', type=d.T.string)]),
+ withUser(user): { quobyte+: { user: user } },
+ '#withVolume':: d.fn(help='"volume is a string that references an already created Quobyte volume by name."', args=[d.arg(name='volume', type=d.T.string)]),
+ withVolume(volume): { quobyte+: { volume: volume } },
+ },
+ '#rbd':: d.obj(help="\"rbd represents a Rados Block Device mount on the host that shares a pod's lifetime. More info: https://examples.k8s.io/volumes/rbd/README.md\""),
+ rbd: {
+ '#secretRef':: d.obj(help='"secretRef is name of the authentication secret for RBDUser. If provided overrides keyring. Default is nil. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it"'),
+ secretRef: {
+ '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { rbd+: { secretRef+: { name: name } } },
+ },
+ '#withFsType':: d.fn(help='"fsType is the filesystem type of the volume that you want to mount. Tip: Ensure that the filesystem type is supported by the host operating system. Examples: \\"ext4\\", \\"xfs\\", \\"ntfs\\". Implicitly inferred to be \\"ext4\\" if unspecified. More info: https://kubernetes.io/docs/concepts/storage/volumes#rbd TODO: how do we prevent errors in the filesystem from compromising the machine"', args=[d.arg(name='fsType', type=d.T.string)]),
+ withFsType(fsType): { rbd+: { fsType: fsType } },
+ '#withImage':: d.fn(help='"image is the rados image name. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it"', args=[d.arg(name='image', type=d.T.string)]),
+ withImage(image): { rbd+: { image: image } },
+ '#withKeyring':: d.fn(help='"keyring is the path to key ring for RBDUser. Default is /etc/ceph/keyring. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it"', args=[d.arg(name='keyring', type=d.T.string)]),
+ withKeyring(keyring): { rbd+: { keyring: keyring } },
+ '#withMonitors':: d.fn(help='"monitors is a collection of Ceph monitors. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it"', args=[d.arg(name='monitors', type=d.T.array)]),
+ withMonitors(monitors): { rbd+: { monitors: if std.isArray(v=monitors) then monitors else [monitors] } },
+ '#withMonitorsMixin':: d.fn(help='"monitors is a collection of Ceph monitors. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='monitors', type=d.T.array)]),
+ withMonitorsMixin(monitors): { rbd+: { monitors+: if std.isArray(v=monitors) then monitors else [monitors] } },
+ '#withPool':: d.fn(help='"pool is the rados pool name. Default is rbd. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it"', args=[d.arg(name='pool', type=d.T.string)]),
+ withPool(pool): { rbd+: { pool: pool } },
+ '#withReadOnly':: d.fn(help='"readOnly here will force the ReadOnly setting in VolumeMounts. Defaults to false. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it"', args=[d.arg(name='readOnly', type=d.T.boolean)]),
+ withReadOnly(readOnly): { rbd+: { readOnly: readOnly } },
+ '#withUser':: d.fn(help='"user is the rados user name. Default is admin. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it"', args=[d.arg(name='user', type=d.T.string)]),
+ withUser(user): { rbd+: { user: user } },
+ },
+ '#scaleIO':: d.obj(help='"scaleIO represents a ScaleIO persistent volume attached and mounted on Kubernetes nodes."'),
+ scaleIO: {
+ '#secretRef':: d.obj(help='"secretRef references to the secret for ScaleIO user and other sensitive information. If this is not provided, Login operation will fail."'),
+ secretRef: {
+ '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { scaleIO+: { secretRef+: { name: name } } },
+ },
+ '#withFsType':: d.fn(help='"fsType is the filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. \\"ext4\\", \\"xfs\\", \\"ntfs\\". Default is \\"xfs\\"."', args=[d.arg(name='fsType', type=d.T.string)]),
+ withFsType(fsType): { scaleIO+: { fsType: fsType } },
+ '#withGateway':: d.fn(help='"gateway is the host address of the ScaleIO API Gateway."', args=[d.arg(name='gateway', type=d.T.string)]),
+ withGateway(gateway): { scaleIO+: { gateway: gateway } },
+ '#withProtectionDomain':: d.fn(help='"protectionDomain is the name of the ScaleIO Protection Domain for the configured storage."', args=[d.arg(name='protectionDomain', type=d.T.string)]),
+ withProtectionDomain(protectionDomain): { scaleIO+: { protectionDomain: protectionDomain } },
+ '#withReadOnly':: d.fn(help='"readOnly Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts."', args=[d.arg(name='readOnly', type=d.T.boolean)]),
+ withReadOnly(readOnly): { scaleIO+: { readOnly: readOnly } },
+ '#withSslEnabled':: d.fn(help='"sslEnabled Flag enable/disable SSL communication with Gateway, default false"', args=[d.arg(name='sslEnabled', type=d.T.boolean)]),
+ withSslEnabled(sslEnabled): { scaleIO+: { sslEnabled: sslEnabled } },
+ '#withStorageMode':: d.fn(help='"storageMode indicates whether the storage for a volume should be ThickProvisioned or ThinProvisioned. Default is ThinProvisioned."', args=[d.arg(name='storageMode', type=d.T.string)]),
+ withStorageMode(storageMode): { scaleIO+: { storageMode: storageMode } },
+ '#withStoragePool':: d.fn(help='"storagePool is the ScaleIO Storage Pool associated with the protection domain."', args=[d.arg(name='storagePool', type=d.T.string)]),
+ withStoragePool(storagePool): { scaleIO+: { storagePool: storagePool } },
+ '#withSystem':: d.fn(help='"system is the name of the storage system as configured in ScaleIO."', args=[d.arg(name='system', type=d.T.string)]),
+ withSystem(system): { scaleIO+: { system: system } },
+ '#withVolumeName':: d.fn(help='"volumeName is the name of a volume already created in the ScaleIO system that is associated with this volume source."', args=[d.arg(name='volumeName', type=d.T.string)]),
+ withVolumeName(volumeName): { scaleIO+: { volumeName: volumeName } },
+ },
+ '#secret':: d.obj(help='"secret represents a secret that should populate this volume. More info: https://kubernetes.io/docs/concepts/storage/volumes#secret"'),
+ secret: {
+ '#items':: d.obj(help="\"items If unspecified, each key-value pair in the Data field of the referenced Secret will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the Secret, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'.\""),
+ items: {
+ '#withKey':: d.fn(help='"key is the key to project."', args=[d.arg(name='key', type=d.T.string)]),
+ withKey(key): { key: key },
+ '#withMode':: d.fn(help='"mode is Optional: mode bits used to set permissions on this file. Must be an octal value between 0000 and 0777 or a decimal value between 0 and 511. YAML accepts both octal and decimal values, JSON requires decimal values for mode bits. If not specified, the volume defaultMode will be used. This might be in conflict with other options that affect the file mode, like fsGroup, and the result can be other mode bits set."', args=[d.arg(name='mode', type=d.T.integer)]),
+ withMode(mode): { mode: mode },
+ '#withPath':: d.fn(help="\"path is the relative path of the file to map the key to. May not be an absolute path. May not contain the path element '..'. May not start with the string '..'.\"", args=[d.arg(name='path', type=d.T.string)]),
+ withPath(path): { path: path },
+ },
+ '#withDefaultMode':: d.fn(help='"defaultMode is Optional: mode bits used to set permissions on created files by default. Must be an octal value between 0000 and 0777 or a decimal value between 0 and 511. YAML accepts both octal and decimal values, JSON requires decimal values for mode bits. Defaults to 0644. Directories within the path are not affected by this setting. This might be in conflict with other options that affect the file mode, like fsGroup, and the result can be other mode bits set."', args=[d.arg(name='defaultMode', type=d.T.integer)]),
+ withDefaultMode(defaultMode): { secret+: { defaultMode: defaultMode } },
+ '#withItems':: d.fn(help="\"items If unspecified, each key-value pair in the Data field of the referenced Secret will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the Secret, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'.\"", args=[d.arg(name='items', type=d.T.array)]),
+ withItems(items): { secret+: { items: if std.isArray(v=items) then items else [items] } },
+ '#withItemsMixin':: d.fn(help="\"items If unspecified, each key-value pair in the Data field of the referenced Secret will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the Secret, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='items', type=d.T.array)]),
+ withItemsMixin(items): { secret+: { items+: if std.isArray(v=items) then items else [items] } },
+ '#withOptional':: d.fn(help='"optional field specify whether the Secret or its keys must be defined"', args=[d.arg(name='optional', type=d.T.boolean)]),
+ withOptional(optional): { secret+: { optional: optional } },
+ '#withSecretName':: d.fn(help="\"secretName is the name of the secret in the pod's namespace to use. More info: https://kubernetes.io/docs/concepts/storage/volumes#secret\"", args=[d.arg(name='secretName', type=d.T.string)]),
+ withSecretName(secretName): { secret+: { secretName: secretName } },
+ },
+ '#storageos':: d.obj(help='"storageOS represents a StorageOS volume attached and mounted on Kubernetes nodes."'),
+ storageos: {
+ '#secretRef':: d.obj(help='"secretRef specifies the secret to use for obtaining the StorageOS API credentials. If not specified, default values will be attempted."'),
+ secretRef: {
+ '#withName':: d.fn(help='"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { storageos+: { secretRef+: { name: name } } },
+ },
+ '#withFsType':: d.fn(help='"fsType is the filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. \\"ext4\\", \\"xfs\\", \\"ntfs\\". Implicitly inferred to be \\"ext4\\" if unspecified."', args=[d.arg(name='fsType', type=d.T.string)]),
+ withFsType(fsType): { storageos+: { fsType: fsType } },
+ '#withReadOnly':: d.fn(help='"readOnly defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts."', args=[d.arg(name='readOnly', type=d.T.boolean)]),
+ withReadOnly(readOnly): { storageos+: { readOnly: readOnly } },
+ '#withVolumeName':: d.fn(help='"volumeName is the human-readable name of the StorageOS volume. Volume names are only unique within a namespace."', args=[d.arg(name='volumeName', type=d.T.string)]),
+ withVolumeName(volumeName): { storageos+: { volumeName: volumeName } },
+ '#withVolumeNamespace':: d.fn(help="\"volumeNamespace specifies the scope of the volume within StorageOS. If no namespace is specified then the Pod's namespace will be used. This allows the Kubernetes name scoping to be mirrored within StorageOS for tighter integration. Set VolumeName to any name to override the default behaviour. Set to \\\"default\\\" if you are not using namespaces within StorageOS. Namespaces that do not pre-exist within StorageOS will be created.\"", args=[d.arg(name='volumeNamespace', type=d.T.string)]),
+ withVolumeNamespace(volumeNamespace): { storageos+: { volumeNamespace: volumeNamespace } },
+ },
+ '#vsphereVolume':: d.obj(help='"vsphereVolume represents a vSphere volume attached and mounted on kubelets host machine"'),
+ vsphereVolume: {
+ '#withFsType':: d.fn(help='"fsType is filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. \\"ext4\\", \\"xfs\\", \\"ntfs\\". Implicitly inferred to be \\"ext4\\" if unspecified."', args=[d.arg(name='fsType', type=d.T.string)]),
+ withFsType(fsType): { vsphereVolume+: { fsType: fsType } },
+ '#withStoragePolicyID':: d.fn(help='"storagePolicyID is the storage Policy Based Management (SPBM) profile ID associated with the StoragePolicyName."', args=[d.arg(name='storagePolicyID', type=d.T.string)]),
+ withStoragePolicyID(storagePolicyID): { vsphereVolume+: { storagePolicyID: storagePolicyID } },
+ '#withStoragePolicyName':: d.fn(help='"storagePolicyName is the storage Policy Based Management (SPBM) profile name."', args=[d.arg(name='storagePolicyName', type=d.T.string)]),
+ withStoragePolicyName(storagePolicyName): { vsphereVolume+: { storagePolicyName: storagePolicyName } },
+ '#withVolumePath':: d.fn(help='"volumePath is the path that identifies vSphere volume vmdk"', args=[d.arg(name='volumePath', type=d.T.string)]),
+ withVolumePath(volumePath): { vsphereVolume+: { volumePath: volumePath } },
+ },
+ '#withName':: d.fn(help='"name of the volume. Must be a DNS_LABEL and unique within the pod. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names"', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { name: name },
+ },
+ '#withActiveDeadlineSeconds':: d.fn(help='"Optional duration in seconds the pod may be active on the node relative to StartTime before the system will actively try to mark it failed and kill associated containers. Value must be a positive integer."', args=[d.arg(name='activeDeadlineSeconds', type=d.T.integer)]),
+ withActiveDeadlineSeconds(activeDeadlineSeconds): { spec+: { template+: { spec+: { activeDeadlineSeconds: activeDeadlineSeconds } } } },
+ '#withAutomountServiceAccountToken':: d.fn(help='"AutomountServiceAccountToken indicates whether a service account token should be automatically mounted."', args=[d.arg(name='automountServiceAccountToken', type=d.T.boolean)]),
+ withAutomountServiceAccountToken(automountServiceAccountToken): { spec+: { template+: { spec+: { automountServiceAccountToken: automountServiceAccountToken } } } },
+ '#withContainers':: d.fn(help='"List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated."', args=[d.arg(name='containers', type=d.T.array)]),
+ withContainers(containers): { spec+: { template+: { spec+: { containers: if std.isArray(v=containers) then containers else [containers] } } } },
+ '#withContainersMixin':: d.fn(help='"List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='containers', type=d.T.array)]),
+ withContainersMixin(containers): { spec+: { template+: { spec+: { containers+: if std.isArray(v=containers) then containers else [containers] } } } },
+ '#withDnsPolicy':: d.fn(help="\"Set DNS policy for the pod. Defaults to \\\"ClusterFirst\\\". Valid values are 'ClusterFirstWithHostNet', 'ClusterFirst', 'Default' or 'None'. DNS parameters given in DNSConfig will be merged with the policy selected with DNSPolicy. To have DNS options set along with hostNetwork, you have to specify DNS policy explicitly to 'ClusterFirstWithHostNet'.\"", args=[d.arg(name='dnsPolicy', type=d.T.string)]),
+ withDnsPolicy(dnsPolicy): { spec+: { template+: { spec+: { dnsPolicy: dnsPolicy } } } },
+ '#withEnableServiceLinks':: d.fn(help="\"EnableServiceLinks indicates whether information about services should be injected into pod's environment variables, matching the syntax of Docker links. Optional: Defaults to true.\"", args=[d.arg(name='enableServiceLinks', type=d.T.boolean)]),
+ withEnableServiceLinks(enableServiceLinks): { spec+: { template+: { spec+: { enableServiceLinks: enableServiceLinks } } } },
+ '#withEphemeralContainers':: d.fn(help="\"List of ephemeral containers run in this pod. Ephemeral containers may be run in an existing pod to perform user-initiated actions such as debugging. This list cannot be specified when creating a pod, and it cannot be modified by updating the pod spec. In order to add an ephemeral container to an existing pod, use the pod's ephemeralcontainers subresource.\"", args=[d.arg(name='ephemeralContainers', type=d.T.array)]),
+ withEphemeralContainers(ephemeralContainers): { spec+: { template+: { spec+: { ephemeralContainers: if std.isArray(v=ephemeralContainers) then ephemeralContainers else [ephemeralContainers] } } } },
+ '#withEphemeralContainersMixin':: d.fn(help="\"List of ephemeral containers run in this pod. Ephemeral containers may be run in an existing pod to perform user-initiated actions such as debugging. This list cannot be specified when creating a pod, and it cannot be modified by updating the pod spec. In order to add an ephemeral container to an existing pod, use the pod's ephemeralcontainers subresource.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='ephemeralContainers', type=d.T.array)]),
+ withEphemeralContainersMixin(ephemeralContainers): { spec+: { template+: { spec+: { ephemeralContainers+: if std.isArray(v=ephemeralContainers) then ephemeralContainers else [ephemeralContainers] } } } },
+ '#withHostAliases':: d.fn(help="\"HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods.\"", args=[d.arg(name='hostAliases', type=d.T.array)]),
+ withHostAliases(hostAliases): { spec+: { template+: { spec+: { hostAliases: if std.isArray(v=hostAliases) then hostAliases else [hostAliases] } } } },
+ '#withHostAliasesMixin':: d.fn(help="\"HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='hostAliases', type=d.T.array)]),
+ withHostAliasesMixin(hostAliases): { spec+: { template+: { spec+: { hostAliases+: if std.isArray(v=hostAliases) then hostAliases else [hostAliases] } } } },
+ '#withHostIPC':: d.fn(help="\"Use the host's ipc namespace. Optional: Default to false.\"", args=[d.arg(name='hostIPC', type=d.T.boolean)]),
+ withHostIPC(hostIPC): { spec+: { template+: { spec+: { hostIPC: hostIPC } } } },
+ '#withHostNetwork':: d.fn(help="\"Host networking requested for this pod. Use the host's network namespace. If this option is set, the ports that will be used must be specified. Default to false.\"", args=[d.arg(name='hostNetwork', type=d.T.boolean)]),
+ withHostNetwork(hostNetwork): { spec+: { template+: { spec+: { hostNetwork: hostNetwork } } } },
+ '#withHostPID':: d.fn(help="\"Use the host's pid namespace. Optional: Default to false.\"", args=[d.arg(name='hostPID', type=d.T.boolean)]),
+ withHostPID(hostPID): { spec+: { template+: { spec+: { hostPID: hostPID } } } },
+ '#withHostUsers':: d.fn(help="\"Use the host's user namespace. Optional: Default to true. If set to true or not present, the pod will be run in the host user namespace, useful for when the pod needs a feature only available to the host user namespace, such as loading a kernel module with CAP_SYS_MODULE. When set to false, a new userns is created for the pod. Setting false is useful for mitigating container breakout vulnerabilities even allowing users to run their containers as root without actually having root privileges on the host. This field is alpha-level and is only honored by servers that enable the UserNamespacesSupport feature.\"", args=[d.arg(name='hostUsers', type=d.T.boolean)]),
+ withHostUsers(hostUsers): { spec+: { template+: { spec+: { hostUsers: hostUsers } } } },
+ '#withHostname':: d.fn(help="\"Specifies the hostname of the Pod If not specified, the pod's hostname will be set to a system-defined value.\"", args=[d.arg(name='hostname', type=d.T.string)]),
+ withHostname(hostname): { spec+: { template+: { spec+: { hostname: hostname } } } },
+ '#withImagePullSecrets':: d.fn(help='"ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod"', args=[d.arg(name='imagePullSecrets', type=d.T.array)]),
+ withImagePullSecrets(imagePullSecrets): { spec+: { template+: { spec+: { imagePullSecrets: if std.isArray(v=imagePullSecrets) then imagePullSecrets else [imagePullSecrets] } } } },
+ '#withImagePullSecretsMixin':: d.fn(help='"ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='imagePullSecrets', type=d.T.array)]),
+ withImagePullSecretsMixin(imagePullSecrets): { spec+: { template+: { spec+: { imagePullSecrets+: if std.isArray(v=imagePullSecrets) then imagePullSecrets else [imagePullSecrets] } } } },
+ '#withInitContainers':: d.fn(help='"List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, Liveness probes, or Startup probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/"', args=[d.arg(name='initContainers', type=d.T.array)]),
+ withInitContainers(initContainers): { spec+: { template+: { spec+: { initContainers: if std.isArray(v=initContainers) then initContainers else [initContainers] } } } },
+ '#withInitContainersMixin':: d.fn(help='"List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, Liveness probes, or Startup probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='initContainers', type=d.T.array)]),
+ withInitContainersMixin(initContainers): { spec+: { template+: { spec+: { initContainers+: if std.isArray(v=initContainers) then initContainers else [initContainers] } } } },
+ '#withNodeName':: d.fn(help='"NodeName is a request to schedule this pod onto a specific node. If it is non-empty, the scheduler simply schedules this pod onto that node, assuming that it fits resource requirements."', args=[d.arg(name='nodeName', type=d.T.string)]),
+ withNodeName(nodeName): { spec+: { template+: { spec+: { nodeName: nodeName } } } },
+ '#withNodeSelector':: d.fn(help="\"NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/\"", args=[d.arg(name='nodeSelector', type=d.T.object)]),
+ withNodeSelector(nodeSelector): { spec+: { template+: { spec+: { nodeSelector: nodeSelector } } } },
+ '#withNodeSelectorMixin':: d.fn(help="\"NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='nodeSelector', type=d.T.object)]),
+ withNodeSelectorMixin(nodeSelector): { spec+: { template+: { spec+: { nodeSelector+: nodeSelector } } } },
+ '#withOverhead':: d.fn(help='"Overhead represents the resource overhead associated with running a pod for a given RuntimeClass. This field will be autopopulated at admission time by the RuntimeClass admission controller. If the RuntimeClass admission controller is enabled, overhead must not be set in Pod create requests. The RuntimeClass admission controller will reject Pod create requests which have the overhead already set. If RuntimeClass is configured and selected in the PodSpec, Overhead will be set to the value defined in the corresponding RuntimeClass, otherwise it will remain unset and treated as zero. More info: https://git.k8s.io/enhancements/keps/sig-node/688-pod-overhead/README.md"', args=[d.arg(name='overhead', type=d.T.object)]),
+ withOverhead(overhead): { spec+: { template+: { spec+: { overhead: overhead } } } },
+ '#withOverheadMixin':: d.fn(help='"Overhead represents the resource overhead associated with running a pod for a given RuntimeClass. This field will be autopopulated at admission time by the RuntimeClass admission controller. If the RuntimeClass admission controller is enabled, overhead must not be set in Pod create requests. The RuntimeClass admission controller will reject Pod create requests which have the overhead already set. If RuntimeClass is configured and selected in the PodSpec, Overhead will be set to the value defined in the corresponding RuntimeClass, otherwise it will remain unset and treated as zero. More info: https://git.k8s.io/enhancements/keps/sig-node/688-pod-overhead/README.md"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='overhead', type=d.T.object)]),
+ withOverheadMixin(overhead): { spec+: { template+: { spec+: { overhead+: overhead } } } },
+ '#withPreemptionPolicy':: d.fn(help='"PreemptionPolicy is the Policy for preempting pods with lower priority. One of Never, PreemptLowerPriority. Defaults to PreemptLowerPriority if unset."', args=[d.arg(name='preemptionPolicy', type=d.T.string)]),
+ withPreemptionPolicy(preemptionPolicy): { spec+: { template+: { spec+: { preemptionPolicy: preemptionPolicy } } } },
+ '#withPriority':: d.fn(help='"The priority value. Various system components use this field to find the priority of the pod. When Priority Admission Controller is enabled, it prevents users from setting this field. The admission controller populates this field from PriorityClassName. The higher the value, the higher the priority."', args=[d.arg(name='priority', type=d.T.integer)]),
+ withPriority(priority): { spec+: { template+: { spec+: { priority: priority } } } },
+ '#withPriorityClassName':: d.fn(help="\"If specified, indicates the pod's priority. \\\"system-node-critical\\\" and \\\"system-cluster-critical\\\" are two special keywords which indicate the highest priorities with the former being the highest priority. Any other name must be defined by creating a PriorityClass object with that name. If not specified, the pod priority will be default or zero if there is no default.\"", args=[d.arg(name='priorityClassName', type=d.T.string)]),
+ withPriorityClassName(priorityClassName): { spec+: { template+: { spec+: { priorityClassName: priorityClassName } } } },
+ '#withReadinessGates':: d.fn(help='"If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to \\"True\\" More info: https://git.k8s.io/enhancements/keps/sig-network/580-pod-readiness-gates"', args=[d.arg(name='readinessGates', type=d.T.array)]),
+ withReadinessGates(readinessGates): { spec+: { template+: { spec+: { readinessGates: if std.isArray(v=readinessGates) then readinessGates else [readinessGates] } } } },
+ '#withReadinessGatesMixin':: d.fn(help='"If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to \\"True\\" More info: https://git.k8s.io/enhancements/keps/sig-network/580-pod-readiness-gates"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='readinessGates', type=d.T.array)]),
+ withReadinessGatesMixin(readinessGates): { spec+: { template+: { spec+: { readinessGates+: if std.isArray(v=readinessGates) then readinessGates else [readinessGates] } } } },
+ '#withResourceClaims':: d.fn(help='"ResourceClaims defines which ResourceClaims must be allocated and reserved before the Pod is allowed to start. The resources will be made available to those containers which consume them by name. \\n This is an alpha field and requires enabling the DynamicResourceAllocation feature gate. \\n This field is immutable."', args=[d.arg(name='resourceClaims', type=d.T.array)]),
+ withResourceClaims(resourceClaims): { spec+: { template+: { spec+: { resourceClaims: if std.isArray(v=resourceClaims) then resourceClaims else [resourceClaims] } } } },
+ '#withResourceClaimsMixin':: d.fn(help='"ResourceClaims defines which ResourceClaims must be allocated and reserved before the Pod is allowed to start. The resources will be made available to those containers which consume them by name. \\n This is an alpha field and requires enabling the DynamicResourceAllocation feature gate. \\n This field is immutable."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='resourceClaims', type=d.T.array)]),
+ withResourceClaimsMixin(resourceClaims): { spec+: { template+: { spec+: { resourceClaims+: if std.isArray(v=resourceClaims) then resourceClaims else [resourceClaims] } } } },
+ '#withRestartPolicy':: d.fn(help='"Restart policy for all containers within the pod. One of Always, OnFailure, Never. In some contexts, only a subset of those values may be permitted. Default to Always. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle/#restart-policy"', args=[d.arg(name='restartPolicy', type=d.T.string)]),
+ withRestartPolicy(restartPolicy): { spec+: { template+: { spec+: { restartPolicy: restartPolicy } } } },
+ '#withRuntimeClassName':: d.fn(help='"RuntimeClassName refers to a RuntimeClass object in the node.k8s.io group, which should be used to run this pod. If no RuntimeClass resource matches the named class, the pod will not be run. If unset or empty, the \\"legacy\\" RuntimeClass will be used, which is an implicit class with an empty definition that uses the default runtime handler. More info: https://git.k8s.io/enhancements/keps/sig-node/585-runtime-class"', args=[d.arg(name='runtimeClassName', type=d.T.string)]),
+ withRuntimeClassName(runtimeClassName): { spec+: { template+: { spec+: { runtimeClassName: runtimeClassName } } } },
+ '#withSchedulerName':: d.fn(help='"If specified, the pod will be dispatched by specified scheduler. If not specified, the pod will be dispatched by default scheduler."', args=[d.arg(name='schedulerName', type=d.T.string)]),
+ withSchedulerName(schedulerName): { spec+: { template+: { spec+: { schedulerName: schedulerName } } } },
+ '#withSchedulingGates':: d.fn(help='"SchedulingGates is an opaque list of values that if specified will block scheduling the pod. If schedulingGates is not empty, the pod will stay in the SchedulingGated state and the scheduler will not attempt to schedule the pod. \\n SchedulingGates can only be set at pod creation time, and be removed only afterwards. \\n This is a beta feature enabled by the PodSchedulingReadiness feature gate."', args=[d.arg(name='schedulingGates', type=d.T.array)]),
+ withSchedulingGates(schedulingGates): { spec+: { template+: { spec+: { schedulingGates: if std.isArray(v=schedulingGates) then schedulingGates else [schedulingGates] } } } },
+ '#withSchedulingGatesMixin':: d.fn(help='"SchedulingGates is an opaque list of values that if specified will block scheduling the pod. If schedulingGates is not empty, the pod will stay in the SchedulingGated state and the scheduler will not attempt to schedule the pod. \\n SchedulingGates can only be set at pod creation time, and be removed only afterwards. \\n This is a beta feature enabled by the PodSchedulingReadiness feature gate."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='schedulingGates', type=d.T.array)]),
+ withSchedulingGatesMixin(schedulingGates): { spec+: { template+: { spec+: { schedulingGates+: if std.isArray(v=schedulingGates) then schedulingGates else [schedulingGates] } } } },
+ '#withServiceAccount':: d.fn(help='"DeprecatedServiceAccount is a depreciated alias for ServiceAccountName. Deprecated: Use serviceAccountName instead."', args=[d.arg(name='serviceAccount', type=d.T.string)]),
+ withServiceAccount(serviceAccount): { spec+: { template+: { spec+: { serviceAccount: serviceAccount } } } },
+ '#withServiceAccountName':: d.fn(help='"ServiceAccountName is the name of the ServiceAccount to use to run this pod. More info: https://kubernetes.io/docs/tasks/configure-pod-container/configure-service-account/"', args=[d.arg(name='serviceAccountName', type=d.T.string)]),
+ withServiceAccountName(serviceAccountName): { spec+: { template+: { spec+: { serviceAccountName: serviceAccountName } } } },
+ '#withSetHostnameAsFQDN':: d.fn(help="\"If true the pod's hostname will be configured as the pod's FQDN, rather than the leaf name (the default). In Linux containers, this means setting the FQDN in the hostname field of the kernel (the nodename field of struct utsname). In Windows containers, this means setting the registry value of hostname for the registry key HKEY_LOCAL_MACHINE\\\\\\\\SYSTEM\\\\\\\\CurrentControlSet\\\\\\\\Services\\\\\\\\Tcpip\\\\\\\\Parameters to FQDN. If a pod does not have FQDN, this has no effect. Default to false.\"", args=[d.arg(name='setHostnameAsFQDN', type=d.T.boolean)]),
+ withSetHostnameAsFQDN(setHostnameAsFQDN): { spec+: { template+: { spec+: { setHostnameAsFQDN: setHostnameAsFQDN } } } },
+ '#withShareProcessNamespace':: d.fn(help='"Share a single process namespace between all of the containers in a pod. When this is set containers will be able to view and signal processes from other containers in the same pod, and the first process in each container will not be assigned PID 1. HostPID and ShareProcessNamespace cannot both be set. Optional: Default to false."', args=[d.arg(name='shareProcessNamespace', type=d.T.boolean)]),
+ withShareProcessNamespace(shareProcessNamespace): { spec+: { template+: { spec+: { shareProcessNamespace: shareProcessNamespace } } } },
+ '#withSubdomain':: d.fn(help='"If specified, the fully qualified Pod hostname will be \\"...svc.\\". If not specified, the pod will not have a domainname at all."', args=[d.arg(name='subdomain', type=d.T.string)]),
+ withSubdomain(subdomain): { spec+: { template+: { spec+: { subdomain: subdomain } } } },
+ '#withTerminationGracePeriodSeconds':: d.fn(help='"Optional duration in seconds the pod needs to terminate gracefully. May be decreased in delete request. Value must be non-negative integer. The value zero indicates stop immediately via the kill signal (no opportunity to shut down). If this value is nil, the default grace period will be used instead. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. Defaults to 30 seconds."', args=[d.arg(name='terminationGracePeriodSeconds', type=d.T.integer)]),
+ withTerminationGracePeriodSeconds(terminationGracePeriodSeconds): { spec+: { template+: { spec+: { terminationGracePeriodSeconds: terminationGracePeriodSeconds } } } },
+ '#withTolerations':: d.fn(help="\"If specified, the pod's tolerations.\"", args=[d.arg(name='tolerations', type=d.T.array)]),
+ withTolerations(tolerations): { spec+: { template+: { spec+: { tolerations: if std.isArray(v=tolerations) then tolerations else [tolerations] } } } },
+ '#withTolerationsMixin':: d.fn(help="\"If specified, the pod's tolerations.\"\n\n**Note:** This function appends passed data to existing values", args=[d.arg(name='tolerations', type=d.T.array)]),
+ withTolerationsMixin(tolerations): { spec+: { template+: { spec+: { tolerations+: if std.isArray(v=tolerations) then tolerations else [tolerations] } } } },
+ '#withTopologySpreadConstraints':: d.fn(help='"TopologySpreadConstraints describes how a group of pods ought to spread across topology domains. Scheduler will schedule pods in a way which abides by the constraints. All topologySpreadConstraints are ANDed."', args=[d.arg(name='topologySpreadConstraints', type=d.T.array)]),
+ withTopologySpreadConstraints(topologySpreadConstraints): { spec+: { template+: { spec+: { topologySpreadConstraints: if std.isArray(v=topologySpreadConstraints) then topologySpreadConstraints else [topologySpreadConstraints] } } } },
+ '#withTopologySpreadConstraintsMixin':: d.fn(help='"TopologySpreadConstraints describes how a group of pods ought to spread across topology domains. Scheduler will schedule pods in a way which abides by the constraints. All topologySpreadConstraints are ANDed."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='topologySpreadConstraints', type=d.T.array)]),
+ withTopologySpreadConstraintsMixin(topologySpreadConstraints): { spec+: { template+: { spec+: { topologySpreadConstraints+: if std.isArray(v=topologySpreadConstraints) then topologySpreadConstraints else [topologySpreadConstraints] } } } },
+ '#withVolumes':: d.fn(help='"List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes"', args=[d.arg(name='volumes', type=d.T.array)]),
+ withVolumes(volumes): { spec+: { template+: { spec+: { volumes: if std.isArray(v=volumes) then volumes else [volumes] } } } },
+ '#withVolumesMixin':: d.fn(help='"List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='volumes', type=d.T.array)]),
+ withVolumesMixin(volumes): { spec+: { template+: { spec+: { volumes+: if std.isArray(v=volumes) then volumes else [volumes] } } } },
+ },
+ },
+ '#withInstances':: d.fn(help='"The number of replicas we want. Default: 1."', args=[d.arg(name='instances', type=d.T.integer)]),
+ withInstances(instances): { spec+: { instances: instances } },
+ '#withType':: d.fn(help='"Type of service to forward traffic to. Default: `rw`."', args=[d.arg(name='type', type=d.T.string)]),
+ withType(type): { spec+: { type: type } },
+ },
+ '#mixin': 'ignore',
+ mixin: self,
+}
diff --git a/1.21.0/_gen/postgresql/v1/scheduledBackup.libsonnet b/1.21.0/_gen/postgresql/v1/scheduledBackup.libsonnet
new file mode 100644
index 0000000..31dfebb
--- /dev/null
+++ b/1.21.0/_gen/postgresql/v1/scheduledBackup.libsonnet
@@ -0,0 +1,72 @@
+{
+ local d = (import 'doc-util/main.libsonnet'),
+ '#':: d.pkg(name='scheduledBackup', url='', help='"ScheduledBackup is the Schema for the scheduledbackups API"'),
+ '#metadata':: d.obj(help='"ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create."'),
+ metadata: {
+ '#withAnnotations':: d.fn(help='"Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations"', args=[d.arg(name='annotations', type=d.T.object)]),
+ withAnnotations(annotations): { metadata+: { annotations: annotations } },
+ '#withAnnotationsMixin':: d.fn(help='"Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='annotations', type=d.T.object)]),
+ withAnnotationsMixin(annotations): { metadata+: { annotations+: annotations } },
+ '#withClusterName':: d.fn(help='"The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request."', args=[d.arg(name='clusterName', type=d.T.string)]),
+ withClusterName(clusterName): { metadata+: { clusterName: clusterName } },
+ '#withCreationTimestamp':: d.fn(help='"Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers."', args=[d.arg(name='creationTimestamp', type=d.T.string)]),
+ withCreationTimestamp(creationTimestamp): { metadata+: { creationTimestamp: creationTimestamp } },
+ '#withDeletionGracePeriodSeconds':: d.fn(help='"Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only."', args=[d.arg(name='deletionGracePeriodSeconds', type=d.T.integer)]),
+ withDeletionGracePeriodSeconds(deletionGracePeriodSeconds): { metadata+: { deletionGracePeriodSeconds: deletionGracePeriodSeconds } },
+ '#withDeletionTimestamp':: d.fn(help='"Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers."', args=[d.arg(name='deletionTimestamp', type=d.T.string)]),
+ withDeletionTimestamp(deletionTimestamp): { metadata+: { deletionTimestamp: deletionTimestamp } },
+ '#withFinalizers':: d.fn(help='"Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list."', args=[d.arg(name='finalizers', type=d.T.array)]),
+ withFinalizers(finalizers): { metadata+: { finalizers: if std.isArray(v=finalizers) then finalizers else [finalizers] } },
+ '#withFinalizersMixin':: d.fn(help='"Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='finalizers', type=d.T.array)]),
+ withFinalizersMixin(finalizers): { metadata+: { finalizers+: if std.isArray(v=finalizers) then finalizers else [finalizers] } },
+ '#withGenerateName':: d.fn(help='"GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\\n\\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\\n\\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency"', args=[d.arg(name='generateName', type=d.T.string)]),
+ withGenerateName(generateName): { metadata+: { generateName: generateName } },
+ '#withGeneration':: d.fn(help='"A sequence number representing a specific generation of the desired state. Populated by the system. Read-only."', args=[d.arg(name='generation', type=d.T.integer)]),
+ withGeneration(generation): { metadata+: { generation: generation } },
+ '#withLabels':: d.fn(help='"Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels"', args=[d.arg(name='labels', type=d.T.object)]),
+ withLabels(labels): { metadata+: { labels: labels } },
+ '#withLabelsMixin':: d.fn(help='"Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels"\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='labels', type=d.T.object)]),
+ withLabelsMixin(labels): { metadata+: { labels+: labels } },
+ '#withName':: d.fn(help='"Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names"', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { metadata+: { name: name } },
+ '#withNamespace':: d.fn(help='"Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the \\"default\\" namespace, but \\"default\\" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\\n\\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces"', args=[d.arg(name='namespace', type=d.T.string)]),
+ withNamespace(namespace): { metadata+: { namespace: namespace } },
+ '#withOwnerReferences':: d.fn(help='"List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller."', args=[d.arg(name='ownerReferences', type=d.T.array)]),
+ withOwnerReferences(ownerReferences): { metadata+: { ownerReferences: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } },
+ '#withOwnerReferencesMixin':: d.fn(help='"List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller."\n\n**Note:** This function appends passed data to existing values', args=[d.arg(name='ownerReferences', type=d.T.array)]),
+ withOwnerReferencesMixin(ownerReferences): { metadata+: { ownerReferences+: if std.isArray(v=ownerReferences) then ownerReferences else [ownerReferences] } },
+ '#withResourceVersion':: d.fn(help='"An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\\n\\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency"', args=[d.arg(name='resourceVersion', type=d.T.string)]),
+ withResourceVersion(resourceVersion): { metadata+: { resourceVersion: resourceVersion } },
+ '#withSelfLink':: d.fn(help='"SelfLink is a URL representing this object. Populated by the system. Read-only.\\n\\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release."', args=[d.arg(name='selfLink', type=d.T.string)]),
+ withSelfLink(selfLink): { metadata+: { selfLink: selfLink } },
+ '#withUid':: d.fn(help='"UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\\n\\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids"', args=[d.arg(name='uid', type=d.T.string)]),
+ withUid(uid): { metadata+: { uid: uid } },
+ },
+ '#new':: d.fn(help='new returns an instance of ScheduledBackup', args=[d.arg(name='name', type=d.T.string)]),
+ new(name): {
+ apiVersion: 'postgresql.cnpg.io/v1',
+ kind: 'ScheduledBackup',
+ } + self.metadata.withName(name=name),
+ '#spec':: d.obj(help='"Specification of the desired behavior of the ScheduledBackup. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#spec-and-status"'),
+ spec: {
+ '#cluster':: d.obj(help='"The cluster to backup"'),
+ cluster: {
+ '#withName':: d.fn(help='"Name of the referent."', args=[d.arg(name='name', type=d.T.string)]),
+ withName(name): { spec+: { cluster+: { name: name } } },
+ },
+ '#withBackupOwnerReference':: d.fn(help='"Indicates which ownerReference should be put inside the created backup resources.
- none: no owner reference for created backup objects (same behavior as before the field was introduced)
- self: sets the Scheduled backup object as owner of the backup
- cluster: set the cluster as owner of the backup
"', args=[d.arg(name='backupOwnerReference', type=d.T.string)]),
+ withBackupOwnerReference(backupOwnerReference): { spec+: { backupOwnerReference: backupOwnerReference } },
+ '#withImmediate':: d.fn(help='"If the first backup has to be immediately start after creation or not"', args=[d.arg(name='immediate', type=d.T.boolean)]),
+ withImmediate(immediate): { spec+: { immediate: immediate } },
+ '#withMethod':: d.fn(help='"The backup method to be used, possible options are `barmanObjectStore` and `volumeSnapshot`. Defaults to: `barmanObjectStore`."', args=[d.arg(name='method', type=d.T.string)]),
+ withMethod(method): { spec+: { method: method } },
+ '#withSchedule':: d.fn(help='"The schedule does not follow the same format used in Kubernetes CronJobs as it includes an additional seconds specifier, see https://pkg.go.dev/github.com/robfig/cron#hdr-CRON_Expression_Format"', args=[d.arg(name='schedule', type=d.T.string)]),
+ withSchedule(schedule): { spec+: { schedule: schedule } },
+ '#withSuspend':: d.fn(help='"If this backup is suspended or not"', args=[d.arg(name='suspend', type=d.T.boolean)]),
+ withSuspend(suspend): { spec+: { suspend: suspend } },
+ '#withTarget':: d.fn(help='"The policy to decide which instance should perform this backup. If empty, it defaults to `cluster.spec.backup.target`. Available options are empty string, `primary` and `prefer-standby`. `primary` to have backups run always on primary instances, `prefer-standby` to have backups run preferably on the most updated standby, if available."', args=[d.arg(name='target', type=d.T.string)]),
+ withTarget(target): { spec+: { target: target } },
+ },
+ '#mixin': 'ignore',
+ mixin: self,
+}
diff --git a/1.21.0/gen.libsonnet b/1.21.0/gen.libsonnet
new file mode 100644
index 0000000..1ff9218
--- /dev/null
+++ b/1.21.0/gen.libsonnet
@@ -0,0 +1,5 @@
+{
+ local d = (import 'doc-util/main.libsonnet'),
+ '#':: d.pkg(name='cloudnative-pg', url='github.com/jsonnet-libs/cloudnative-pg-libsonnet/1.21.0/main.libsonnet', help=''),
+ postgresql:: (import '_gen/postgresql/main.libsonnet'),
+}
diff --git a/1.21.0/main.libsonnet b/1.21.0/main.libsonnet
new file mode 100644
index 0000000..f5597a5
--- /dev/null
+++ b/1.21.0/main.libsonnet
@@ -0,0 +1 @@
+(import 'gen.libsonnet')
diff --git a/docs/1.17.5/postgresql/v1/cluster.md b/docs/1.17.5/postgresql/v1/cluster.md
index 6fe445b..a2683da 100644
--- a/docs/1.17.5/postgresql/v1/cluster.md
+++ b/docs/1.17.5/postgresql/v1/cluster.md
@@ -53,7 +53,7 @@ permalink: /1.17.5/postgresql/v1/cluster/
* [`fn withPrimaryUpdateStrategy(primaryUpdateStrategy)`](#fn-specwithprimaryupdatestrategy)
* [`fn withPriorityClassName(priorityClassName)`](#fn-specwithpriorityclassname)
* [`fn withSchedulerName(schedulerName)`](#fn-specwithschedulername)
- * [`fn withSmartStopDelay(smartStopDelay)`](#fn-specwithsmartstopdelay)
+ * [`fn withSmartShutdownTimeout(smartShutdownTimeout)`](#fn-specwithsmartshutdowntimeout)
* [`fn withStartDelay(startDelay)`](#fn-specwithstartdelay)
* [`fn withStopDelay(stopDelay)`](#fn-specwithstopdelay)
* [`fn withSwitchoverDelay(switchoverDelay)`](#fn-specwithswitchoverdelay)
@@ -395,6 +395,9 @@ permalink: /1.17.5/postgresql/v1/cluster/
* [`obj spec.envFrom.secretRef`](#obj-specenvfromsecretref)
* [`fn withName(name)`](#fn-specenvfromsecretrefwithname)
* [`fn withOptional(optional)`](#fn-specenvfromsecretrefwithoptional)
+ * [`obj spec.ephemeralVolumesSizeLimit`](#obj-specephemeralvolumessizelimit)
+ * [`fn withShm(shm)`](#fn-specephemeralvolumessizelimitwithshm)
+ * [`fn withTemporaryData(temporaryData)`](#fn-specephemeralvolumessizelimitwithtemporarydata)
* [`obj spec.externalClusters`](#obj-specexternalclusters)
* [`fn withConnectionParameters(connectionParameters)`](#fn-specexternalclusterswithconnectionparameters)
* [`fn withConnectionParametersMixin(connectionParameters)`](#fn-specexternalclusterswithconnectionparametersmixin)
@@ -897,7 +900,7 @@ withDescription(description)
withEnableSuperuserAccess(enableSuperuserAccess)
```
-"When this option is enabled, the operator will use the `SuperuserSecret` to update the `postgres` user password (if the secret is not present, the operator will automatically create one). When this option is disabled, the operator will ignore the `SuperuserSecret` content, delete it when automatically created, and then blank the password of the `postgres` user by setting it to `NULL`. Enabled by default."
+"When this option is enabled, the operator will use the `SuperuserSecret` to update the `postgres` user password (if the secret is not present, the operator will automatically create one). When this option is disabled, the operator will ignore the `SuperuserSecret` content, delete it when automatically created, and then blank the password of the `postgres` user by setting it to `NULL`. Disabled by default."
### fn spec.withEnv
@@ -1075,13 +1078,13 @@ withSchedulerName(schedulerName)
"If specified, the pod will be dispatched by specified Kubernetes scheduler. If not specified, the pod will be dispatched by the default scheduler. More info: https://kubernetes.io/docs/concepts/scheduling-eviction/kube-scheduler/"
-### fn spec.withSmartStopDelay
+### fn spec.withSmartShutdownTimeout
```ts
-withSmartStopDelay(smartStopDelay)
+withSmartShutdownTimeout(smartShutdownTimeout)
```
-"The time in seconds that controls the window of time reserved for the smart shutdown of Postgres to complete. this formula to compute the timeout of smart shutdown is `max(stopDelay - smartStopDelay, 30)`"
+"The time in seconds that controls the window of time reserved for the smart shutdown of Postgres to complete. Make sure you reserve enough time for the operator to request a fast shutdown of Postgres (that is: `stopDelay` - `smartShutdownTimeout`)."
### fn spec.withStartDelay
@@ -3605,6 +3608,26 @@ withOptional(optional)
"Specify whether the Secret must be defined"
+## obj spec.ephemeralVolumesSizeLimit
+
+"EphemeralVolumesSizeLimit allows the user to set the limits for the ephemeral volumes"
+
+### fn spec.ephemeralVolumesSizeLimit.withShm
+
+```ts
+withShm(shm)
+```
+
+"Shm is the size limit of the shared memory volume"
+
+### fn spec.ephemeralVolumesSizeLimit.withTemporaryData
+
+```ts
+withTemporaryData(temporaryData)
+```
+
+"TemporaryData is the size limit of the temporary data volume"
+
## obj spec.externalClusters
"The list of external clusters which are used in the configuration"
@@ -4981,7 +5004,7 @@ withUpdateInterval(updateInterval)
withEnabled(enabled)
```
-"If enabled, the operator will automatically manage replication slots on the primary instance and use them in streaming replication connections with all the standby instances that are part of the HA cluster. If disabled (default), the operator will not take advantage of replication slots in streaming connections with the replicas. This feature also controls replication slots in replica cluster, from the designated primary to its cascading replicas. This can only be set at creation time."
+"If enabled (default), the operator will automatically manage replication slots on the primary instance and use them in streaming replication connections with all the standby instances that are part of the HA cluster. If disabled, the operator will not take advantage of replication slots in streaming connections with the replicas. This feature also controls replication slots in replica cluster, from the designated primary to its cascading replicas."
### fn spec.replicationSlots.highAvailability.withSlotPrefix
diff --git a/docs/1.18.5/postgresql/v1/cluster.md b/docs/1.18.5/postgresql/v1/cluster.md
index e09e2fe..d12f718 100644
--- a/docs/1.18.5/postgresql/v1/cluster.md
+++ b/docs/1.18.5/postgresql/v1/cluster.md
@@ -53,7 +53,7 @@ permalink: /1.18.5/postgresql/v1/cluster/
* [`fn withPrimaryUpdateStrategy(primaryUpdateStrategy)`](#fn-specwithprimaryupdatestrategy)
* [`fn withPriorityClassName(priorityClassName)`](#fn-specwithpriorityclassname)
* [`fn withSchedulerName(schedulerName)`](#fn-specwithschedulername)
- * [`fn withSmartStopDelay(smartStopDelay)`](#fn-specwithsmartstopdelay)
+ * [`fn withSmartShutdownTimeout(smartShutdownTimeout)`](#fn-specwithsmartshutdowntimeout)
* [`fn withStartDelay(startDelay)`](#fn-specwithstartdelay)
* [`fn withStopDelay(stopDelay)`](#fn-specwithstopdelay)
* [`fn withSwitchoverDelay(switchoverDelay)`](#fn-specwithswitchoverdelay)
@@ -395,6 +395,9 @@ permalink: /1.18.5/postgresql/v1/cluster/
* [`obj spec.envFrom.secretRef`](#obj-specenvfromsecretref)
* [`fn withName(name)`](#fn-specenvfromsecretrefwithname)
* [`fn withOptional(optional)`](#fn-specenvfromsecretrefwithoptional)
+ * [`obj spec.ephemeralVolumesSizeLimit`](#obj-specephemeralvolumessizelimit)
+ * [`fn withShm(shm)`](#fn-specephemeralvolumessizelimitwithshm)
+ * [`fn withTemporaryData(temporaryData)`](#fn-specephemeralvolumessizelimitwithtemporarydata)
* [`obj spec.externalClusters`](#obj-specexternalclusters)
* [`fn withConnectionParameters(connectionParameters)`](#fn-specexternalclusterswithconnectionparameters)
* [`fn withConnectionParametersMixin(connectionParameters)`](#fn-specexternalclusterswithconnectionparametersmixin)
@@ -897,7 +900,7 @@ withDescription(description)
withEnableSuperuserAccess(enableSuperuserAccess)
```
-"When this option is enabled, the operator will use the `SuperuserSecret` to update the `postgres` user password (if the secret is not present, the operator will automatically create one). When this option is disabled, the operator will ignore the `SuperuserSecret` content, delete it when automatically created, and then blank the password of the `postgres` user by setting it to `NULL`. Enabled by default."
+"When this option is enabled, the operator will use the `SuperuserSecret` to update the `postgres` user password (if the secret is not present, the operator will automatically create one). When this option is disabled, the operator will ignore the `SuperuserSecret` content, delete it when automatically created, and then blank the password of the `postgres` user by setting it to `NULL`. Disabled by default."
### fn spec.withEnv
@@ -1075,13 +1078,13 @@ withSchedulerName(schedulerName)
"If specified, the pod will be dispatched by specified Kubernetes scheduler. If not specified, the pod will be dispatched by the default scheduler. More info: https://kubernetes.io/docs/concepts/scheduling-eviction/kube-scheduler/"
-### fn spec.withSmartStopDelay
+### fn spec.withSmartShutdownTimeout
```ts
-withSmartStopDelay(smartStopDelay)
+withSmartShutdownTimeout(smartShutdownTimeout)
```
-"The time in seconds that controls the window of time reserved for the smart shutdown of Postgres to complete. this formula to compute the timeout of smart shutdown is `max(stopDelay - smartStopDelay, 30)`"
+"The time in seconds that controls the window of time reserved for the smart shutdown of Postgres to complete. Make sure you reserve enough time for the operator to request a fast shutdown of Postgres (that is: `stopDelay` - `smartShutdownTimeout`)."
### fn spec.withStartDelay
@@ -3605,6 +3608,26 @@ withOptional(optional)
"Specify whether the Secret must be defined"
+## obj spec.ephemeralVolumesSizeLimit
+
+"EphemeralVolumesSizeLimit allows the user to set the limits for the ephemeral volumes"
+
+### fn spec.ephemeralVolumesSizeLimit.withShm
+
+```ts
+withShm(shm)
+```
+
+"Shm is the size limit of the shared memory volume"
+
+### fn spec.ephemeralVolumesSizeLimit.withTemporaryData
+
+```ts
+withTemporaryData(temporaryData)
+```
+
+"TemporaryData is the size limit of the temporary data volume"
+
## obj spec.externalClusters
"The list of external clusters which are used in the configuration"
@@ -4981,7 +5004,7 @@ withUpdateInterval(updateInterval)
withEnabled(enabled)
```
-"If enabled, the operator will automatically manage replication slots on the primary instance and use them in streaming replication connections with all the standby instances that are part of the HA cluster. If disabled (default), the operator will not take advantage of replication slots in streaming connections with the replicas. This feature also controls replication slots in replica cluster, from the designated primary to its cascading replicas. This can only be set at creation time."
+"If enabled (default), the operator will automatically manage replication slots on the primary instance and use them in streaming replication connections with all the standby instances that are part of the HA cluster. If disabled, the operator will not take advantage of replication slots in streaming connections with the replicas. This feature also controls replication slots in replica cluster, from the designated primary to its cascading replicas."
### fn spec.replicationSlots.highAvailability.withSlotPrefix
diff --git a/docs/1.19.4/postgresql/v1/cluster.md b/docs/1.19.4/postgresql/v1/cluster.md
index 29120b0..e854170 100644
--- a/docs/1.19.4/postgresql/v1/cluster.md
+++ b/docs/1.19.4/postgresql/v1/cluster.md
@@ -53,7 +53,7 @@ permalink: /1.19.4/postgresql/v1/cluster/
* [`fn withPrimaryUpdateStrategy(primaryUpdateStrategy)`](#fn-specwithprimaryupdatestrategy)
* [`fn withPriorityClassName(priorityClassName)`](#fn-specwithpriorityclassname)
* [`fn withSchedulerName(schedulerName)`](#fn-specwithschedulername)
- * [`fn withSmartStopDelay(smartStopDelay)`](#fn-specwithsmartstopdelay)
+ * [`fn withSmartShutdownTimeout(smartShutdownTimeout)`](#fn-specwithsmartshutdowntimeout)
* [`fn withStartDelay(startDelay)`](#fn-specwithstartdelay)
* [`fn withStopDelay(stopDelay)`](#fn-specwithstopdelay)
* [`fn withSwitchoverDelay(switchoverDelay)`](#fn-specwithswitchoverdelay)
@@ -395,6 +395,9 @@ permalink: /1.19.4/postgresql/v1/cluster/
* [`obj spec.envFrom.secretRef`](#obj-specenvfromsecretref)
* [`fn withName(name)`](#fn-specenvfromsecretrefwithname)
* [`fn withOptional(optional)`](#fn-specenvfromsecretrefwithoptional)
+ * [`obj spec.ephemeralVolumesSizeLimit`](#obj-specephemeralvolumessizelimit)
+ * [`fn withShm(shm)`](#fn-specephemeralvolumessizelimitwithshm)
+ * [`fn withTemporaryData(temporaryData)`](#fn-specephemeralvolumessizelimitwithtemporarydata)
* [`obj spec.externalClusters`](#obj-specexternalclusters)
* [`fn withConnectionParameters(connectionParameters)`](#fn-specexternalclusterswithconnectionparameters)
* [`fn withConnectionParametersMixin(connectionParameters)`](#fn-specexternalclusterswithconnectionparametersmixin)
@@ -897,7 +900,7 @@ withDescription(description)
withEnableSuperuserAccess(enableSuperuserAccess)
```
-"When this option is enabled, the operator will use the `SuperuserSecret` to update the `postgres` user password (if the secret is not present, the operator will automatically create one). When this option is disabled, the operator will ignore the `SuperuserSecret` content, delete it when automatically created, and then blank the password of the `postgres` user by setting it to `NULL`. Enabled by default."
+"When this option is enabled, the operator will use the `SuperuserSecret` to update the `postgres` user password (if the secret is not present, the operator will automatically create one). When this option is disabled, the operator will ignore the `SuperuserSecret` content, delete it when automatically created, and then blank the password of the `postgres` user by setting it to `NULL`. Disabled by default."
### fn spec.withEnv
@@ -1075,13 +1078,13 @@ withSchedulerName(schedulerName)
"If specified, the pod will be dispatched by specified Kubernetes scheduler. If not specified, the pod will be dispatched by the default scheduler. More info: https://kubernetes.io/docs/concepts/scheduling-eviction/kube-scheduler/"
-### fn spec.withSmartStopDelay
+### fn spec.withSmartShutdownTimeout
```ts
-withSmartStopDelay(smartStopDelay)
+withSmartShutdownTimeout(smartShutdownTimeout)
```
-"The time in seconds that controls the window of time reserved for the smart shutdown of Postgres to complete. this formula to compute the timeout of smart shutdown is `max(stopDelay - smartStopDelay, 30)`"
+"The time in seconds that controls the window of time reserved for the smart shutdown of Postgres to complete. Make sure you reserve enough time for the operator to request a fast shutdown of Postgres (that is: `stopDelay` - `smartShutdownTimeout`)."
### fn spec.withStartDelay
@@ -3605,6 +3608,26 @@ withOptional(optional)
"Specify whether the Secret must be defined"
+## obj spec.ephemeralVolumesSizeLimit
+
+"EphemeralVolumesSizeLimit allows the user to set the limits for the ephemeral volumes"
+
+### fn spec.ephemeralVolumesSizeLimit.withShm
+
+```ts
+withShm(shm)
+```
+
+"Shm is the size limit of the shared memory volume"
+
+### fn spec.ephemeralVolumesSizeLimit.withTemporaryData
+
+```ts
+withTemporaryData(temporaryData)
+```
+
+"TemporaryData is the size limit of the temporary data volume"
+
## obj spec.externalClusters
"The list of external clusters which are used in the configuration"
@@ -4981,7 +5004,7 @@ withUpdateInterval(updateInterval)
withEnabled(enabled)
```
-"If enabled, the operator will automatically manage replication slots on the primary instance and use them in streaming replication connections with all the standby instances that are part of the HA cluster. If disabled (default), the operator will not take advantage of replication slots in streaming connections with the replicas. This feature also controls replication slots in replica cluster, from the designated primary to its cascading replicas. This can only be set at creation time."
+"If enabled (default), the operator will automatically manage replication slots on the primary instance and use them in streaming replication connections with all the standby instances that are part of the HA cluster. If disabled, the operator will not take advantage of replication slots in streaming connections with the replicas. This feature also controls replication slots in replica cluster, from the designated primary to its cascading replicas."
### fn spec.replicationSlots.highAvailability.withSlotPrefix
diff --git a/docs/1.20.2/postgresql/v1/cluster.md b/docs/1.20.2/postgresql/v1/cluster.md
index dfe89af..200cd4d 100644
--- a/docs/1.20.2/postgresql/v1/cluster.md
+++ b/docs/1.20.2/postgresql/v1/cluster.md
@@ -53,7 +53,7 @@ permalink: /1.20.2/postgresql/v1/cluster/
* [`fn withPrimaryUpdateStrategy(primaryUpdateStrategy)`](#fn-specwithprimaryupdatestrategy)
* [`fn withPriorityClassName(priorityClassName)`](#fn-specwithpriorityclassname)
* [`fn withSchedulerName(schedulerName)`](#fn-specwithschedulername)
- * [`fn withSmartStopDelay(smartStopDelay)`](#fn-specwithsmartstopdelay)
+ * [`fn withSmartShutdownTimeout(smartShutdownTimeout)`](#fn-specwithsmartshutdowntimeout)
* [`fn withStartDelay(startDelay)`](#fn-specwithstartdelay)
* [`fn withStopDelay(stopDelay)`](#fn-specwithstopdelay)
* [`fn withSwitchoverDelay(switchoverDelay)`](#fn-specwithswitchoverdelay)
@@ -395,6 +395,9 @@ permalink: /1.20.2/postgresql/v1/cluster/
* [`obj spec.envFrom.secretRef`](#obj-specenvfromsecretref)
* [`fn withName(name)`](#fn-specenvfromsecretrefwithname)
* [`fn withOptional(optional)`](#fn-specenvfromsecretrefwithoptional)
+ * [`obj spec.ephemeralVolumesSizeLimit`](#obj-specephemeralvolumessizelimit)
+ * [`fn withShm(shm)`](#fn-specephemeralvolumessizelimitwithshm)
+ * [`fn withTemporaryData(temporaryData)`](#fn-specephemeralvolumessizelimitwithtemporarydata)
* [`obj spec.externalClusters`](#obj-specexternalclusters)
* [`fn withConnectionParameters(connectionParameters)`](#fn-specexternalclusterswithconnectionparameters)
* [`fn withConnectionParametersMixin(connectionParameters)`](#fn-specexternalclusterswithconnectionparametersmixin)
@@ -897,7 +900,7 @@ withDescription(description)
withEnableSuperuserAccess(enableSuperuserAccess)
```
-"When this option is enabled, the operator will use the `SuperuserSecret` to update the `postgres` user password (if the secret is not present, the operator will automatically create one). When this option is disabled, the operator will ignore the `SuperuserSecret` content, delete it when automatically created, and then blank the password of the `postgres` user by setting it to `NULL`. Enabled by default."
+"When this option is enabled, the operator will use the `SuperuserSecret` to update the `postgres` user password (if the secret is not present, the operator will automatically create one). When this option is disabled, the operator will ignore the `SuperuserSecret` content, delete it when automatically created, and then blank the password of the `postgres` user by setting it to `NULL`. Disabled by default."
### fn spec.withEnv
@@ -1075,13 +1078,13 @@ withSchedulerName(schedulerName)
"If specified, the pod will be dispatched by specified Kubernetes scheduler. If not specified, the pod will be dispatched by the default scheduler. More info: https://kubernetes.io/docs/concepts/scheduling-eviction/kube-scheduler/"
-### fn spec.withSmartStopDelay
+### fn spec.withSmartShutdownTimeout
```ts
-withSmartStopDelay(smartStopDelay)
+withSmartShutdownTimeout(smartShutdownTimeout)
```
-"The time in seconds that controls the window of time reserved for the smart shutdown of Postgres to complete. this formula to compute the timeout of smart shutdown is `max(stopDelay - smartStopDelay, 30)`"
+"The time in seconds that controls the window of time reserved for the smart shutdown of Postgres to complete. Make sure you reserve enough time for the operator to request a fast shutdown of Postgres (that is: `stopDelay` - `smartShutdownTimeout`)."
### fn spec.withStartDelay
@@ -3605,6 +3608,26 @@ withOptional(optional)
"Specify whether the Secret must be defined"
+## obj spec.ephemeralVolumesSizeLimit
+
+"EphemeralVolumesSizeLimit allows the user to set the limits for the ephemeral volumes"
+
+### fn spec.ephemeralVolumesSizeLimit.withShm
+
+```ts
+withShm(shm)
+```
+
+"Shm is the size limit of the shared memory volume"
+
+### fn spec.ephemeralVolumesSizeLimit.withTemporaryData
+
+```ts
+withTemporaryData(temporaryData)
+```
+
+"TemporaryData is the size limit of the temporary data volume"
+
## obj spec.externalClusters
"The list of external clusters which are used in the configuration"
@@ -4981,7 +5004,7 @@ withUpdateInterval(updateInterval)
withEnabled(enabled)
```
-"If enabled, the operator will automatically manage replication slots on the primary instance and use them in streaming replication connections with all the standby instances that are part of the HA cluster. If disabled (default), the operator will not take advantage of replication slots in streaming connections with the replicas. This feature also controls replication slots in replica cluster, from the designated primary to its cascading replicas. This can only be set at creation time."
+"If enabled (default), the operator will automatically manage replication slots on the primary instance and use them in streaming replication connections with all the standby instances that are part of the HA cluster. If disabled, the operator will not take advantage of replication slots in streaming connections with the replicas. This feature also controls replication slots in replica cluster, from the designated primary to its cascading replicas."
### fn spec.replicationSlots.highAvailability.withSlotPrefix
diff --git a/docs/1.21.0/README.md b/docs/1.21.0/README.md
new file mode 100644
index 0000000..e3b61db
--- /dev/null
+++ b/docs/1.21.0/README.md
@@ -0,0 +1,13 @@
+---
+permalink: /1.21.0/
+---
+
+# cloudnative-pg
+
+```jsonnet
+local cloudnative-pg = import "github.com/jsonnet-libs/cloudnative-pg-libsonnet/1.21.0/main.libsonnet"
+```
+
+
+
+* [postgresql](postgresql/index.md)
\ No newline at end of file
diff --git a/docs/1.21.0/postgresql/index.md b/docs/1.21.0/postgresql/index.md
new file mode 100644
index 0000000..7ec274e
--- /dev/null
+++ b/docs/1.21.0/postgresql/index.md
@@ -0,0 +1,9 @@
+---
+permalink: /1.21.0/postgresql/
+---
+
+# postgresql
+
+
+
+* [v1](v1/index.md)
\ No newline at end of file
diff --git a/docs/1.21.0/postgresql/v1/backup.md b/docs/1.21.0/postgresql/v1/backup.md
new file mode 100644
index 0000000..6c3587e
--- /dev/null
+++ b/docs/1.21.0/postgresql/v1/backup.md
@@ -0,0 +1,242 @@
+---
+permalink: /1.21.0/postgresql/v1/backup/
+---
+
+# postgresql.v1.backup
+
+"Backup is the Schema for the backups API"
+
+## Index
+
+* [`fn new(name)`](#fn-new)
+* [`obj metadata`](#obj-metadata)
+ * [`fn withAnnotations(annotations)`](#fn-metadatawithannotations)
+ * [`fn withAnnotationsMixin(annotations)`](#fn-metadatawithannotationsmixin)
+ * [`fn withClusterName(clusterName)`](#fn-metadatawithclustername)
+ * [`fn withCreationTimestamp(creationTimestamp)`](#fn-metadatawithcreationtimestamp)
+ * [`fn withDeletionGracePeriodSeconds(deletionGracePeriodSeconds)`](#fn-metadatawithdeletiongraceperiodseconds)
+ * [`fn withDeletionTimestamp(deletionTimestamp)`](#fn-metadatawithdeletiontimestamp)
+ * [`fn withFinalizers(finalizers)`](#fn-metadatawithfinalizers)
+ * [`fn withFinalizersMixin(finalizers)`](#fn-metadatawithfinalizersmixin)
+ * [`fn withGenerateName(generateName)`](#fn-metadatawithgeneratename)
+ * [`fn withGeneration(generation)`](#fn-metadatawithgeneration)
+ * [`fn withLabels(labels)`](#fn-metadatawithlabels)
+ * [`fn withLabelsMixin(labels)`](#fn-metadatawithlabelsmixin)
+ * [`fn withName(name)`](#fn-metadatawithname)
+ * [`fn withNamespace(namespace)`](#fn-metadatawithnamespace)
+ * [`fn withOwnerReferences(ownerReferences)`](#fn-metadatawithownerreferences)
+ * [`fn withOwnerReferencesMixin(ownerReferences)`](#fn-metadatawithownerreferencesmixin)
+ * [`fn withResourceVersion(resourceVersion)`](#fn-metadatawithresourceversion)
+ * [`fn withSelfLink(selfLink)`](#fn-metadatawithselflink)
+ * [`fn withUid(uid)`](#fn-metadatawithuid)
+* [`obj spec`](#obj-spec)
+ * [`fn withMethod(method)`](#fn-specwithmethod)
+ * [`fn withTarget(target)`](#fn-specwithtarget)
+ * [`obj spec.cluster`](#obj-speccluster)
+ * [`fn withName(name)`](#fn-specclusterwithname)
+
+## Fields
+
+### fn new
+
+```ts
+new(name)
+```
+
+new returns an instance of Backup
+
+## obj metadata
+
+"ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create."
+
+### fn metadata.withAnnotations
+
+```ts
+withAnnotations(annotations)
+```
+
+"Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations"
+
+### fn metadata.withAnnotationsMixin
+
+```ts
+withAnnotationsMixin(annotations)
+```
+
+"Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations"
+
+**Note:** This function appends passed data to existing values
+
+### fn metadata.withClusterName
+
+```ts
+withClusterName(clusterName)
+```
+
+"The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request."
+
+### fn metadata.withCreationTimestamp
+
+```ts
+withCreationTimestamp(creationTimestamp)
+```
+
+"Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers."
+
+### fn metadata.withDeletionGracePeriodSeconds
+
+```ts
+withDeletionGracePeriodSeconds(deletionGracePeriodSeconds)
+```
+
+"Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only."
+
+### fn metadata.withDeletionTimestamp
+
+```ts
+withDeletionTimestamp(deletionTimestamp)
+```
+
+"Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers."
+
+### fn metadata.withFinalizers
+
+```ts
+withFinalizers(finalizers)
+```
+
+"Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list."
+
+### fn metadata.withFinalizersMixin
+
+```ts
+withFinalizersMixin(finalizers)
+```
+
+"Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list."
+
+**Note:** This function appends passed data to existing values
+
+### fn metadata.withGenerateName
+
+```ts
+withGenerateName(generateName)
+```
+
+"GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency"
+
+### fn metadata.withGeneration
+
+```ts
+withGeneration(generation)
+```
+
+"A sequence number representing a specific generation of the desired state. Populated by the system. Read-only."
+
+### fn metadata.withLabels
+
+```ts
+withLabels(labels)
+```
+
+"Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels"
+
+### fn metadata.withLabelsMixin
+
+```ts
+withLabelsMixin(labels)
+```
+
+"Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels"
+
+**Note:** This function appends passed data to existing values
+
+### fn metadata.withName
+
+```ts
+withName(name)
+```
+
+"Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names"
+
+### fn metadata.withNamespace
+
+```ts
+withNamespace(namespace)
+```
+
+"Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the \"default\" namespace, but \"default\" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces"
+
+### fn metadata.withOwnerReferences
+
+```ts
+withOwnerReferences(ownerReferences)
+```
+
+"List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller."
+
+### fn metadata.withOwnerReferencesMixin
+
+```ts
+withOwnerReferencesMixin(ownerReferences)
+```
+
+"List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller."
+
+**Note:** This function appends passed data to existing values
+
+### fn metadata.withResourceVersion
+
+```ts
+withResourceVersion(resourceVersion)
+```
+
+"An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency"
+
+### fn metadata.withSelfLink
+
+```ts
+withSelfLink(selfLink)
+```
+
+"SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release."
+
+### fn metadata.withUid
+
+```ts
+withUid(uid)
+```
+
+"UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids"
+
+## obj spec
+
+"Specification of the desired behavior of the backup. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#spec-and-status"
+
+### fn spec.withMethod
+
+```ts
+withMethod(method)
+```
+
+"The backup method to be used, possible options are `barmanObjectStore` and `volumeSnapshot`. Defaults to: `barmanObjectStore`."
+
+### fn spec.withTarget
+
+```ts
+withTarget(target)
+```
+
+"The policy to decide which instance should perform this backup. If empty, it defaults to `cluster.spec.backup.target`. Available options are empty string, `primary` and `prefer-standby`. `primary` to have backups run always on primary instances, `prefer-standby` to have backups run preferably on the most updated standby, if available."
+
+## obj spec.cluster
+
+"The cluster to backup"
+
+### fn spec.cluster.withName
+
+```ts
+withName(name)
+```
+
+"Name of the referent."
\ No newline at end of file
diff --git a/docs/1.21.0/postgresql/v1/cluster.md b/docs/1.21.0/postgresql/v1/cluster.md
new file mode 100644
index 0000000..d8514a5
--- /dev/null
+++ b/docs/1.21.0/postgresql/v1/cluster.md
@@ -0,0 +1,5881 @@
+---
+permalink: /1.21.0/postgresql/v1/cluster/
+---
+
+# postgresql.v1.cluster
+
+"Cluster is the Schema for the PostgreSQL API"
+
+## Index
+
+* [`fn new(name)`](#fn-new)
+* [`obj metadata`](#obj-metadata)
+ * [`fn withAnnotations(annotations)`](#fn-metadatawithannotations)
+ * [`fn withAnnotationsMixin(annotations)`](#fn-metadatawithannotationsmixin)
+ * [`fn withClusterName(clusterName)`](#fn-metadatawithclustername)
+ * [`fn withCreationTimestamp(creationTimestamp)`](#fn-metadatawithcreationtimestamp)
+ * [`fn withDeletionGracePeriodSeconds(deletionGracePeriodSeconds)`](#fn-metadatawithdeletiongraceperiodseconds)
+ * [`fn withDeletionTimestamp(deletionTimestamp)`](#fn-metadatawithdeletiontimestamp)
+ * [`fn withFinalizers(finalizers)`](#fn-metadatawithfinalizers)
+ * [`fn withFinalizersMixin(finalizers)`](#fn-metadatawithfinalizersmixin)
+ * [`fn withGenerateName(generateName)`](#fn-metadatawithgeneratename)
+ * [`fn withGeneration(generation)`](#fn-metadatawithgeneration)
+ * [`fn withLabels(labels)`](#fn-metadatawithlabels)
+ * [`fn withLabelsMixin(labels)`](#fn-metadatawithlabelsmixin)
+ * [`fn withName(name)`](#fn-metadatawithname)
+ * [`fn withNamespace(namespace)`](#fn-metadatawithnamespace)
+ * [`fn withOwnerReferences(ownerReferences)`](#fn-metadatawithownerreferences)
+ * [`fn withOwnerReferencesMixin(ownerReferences)`](#fn-metadatawithownerreferencesmixin)
+ * [`fn withResourceVersion(resourceVersion)`](#fn-metadatawithresourceversion)
+ * [`fn withSelfLink(selfLink)`](#fn-metadatawithselflink)
+ * [`fn withUid(uid)`](#fn-metadatawithuid)
+* [`obj spec`](#obj-spec)
+ * [`fn withDescription(description)`](#fn-specwithdescription)
+ * [`fn withEnableSuperuserAccess(enableSuperuserAccess)`](#fn-specwithenablesuperuseraccess)
+ * [`fn withEnv(env)`](#fn-specwithenv)
+ * [`fn withEnvFrom(envFrom)`](#fn-specwithenvfrom)
+ * [`fn withEnvFromMixin(envFrom)`](#fn-specwithenvfrommixin)
+ * [`fn withEnvMixin(env)`](#fn-specwithenvmixin)
+ * [`fn withExternalClusters(externalClusters)`](#fn-specwithexternalclusters)
+ * [`fn withExternalClustersMixin(externalClusters)`](#fn-specwithexternalclustersmixin)
+ * [`fn withFailoverDelay(failoverDelay)`](#fn-specwithfailoverdelay)
+ * [`fn withImageName(imageName)`](#fn-specwithimagename)
+ * [`fn withImagePullPolicy(imagePullPolicy)`](#fn-specwithimagepullpolicy)
+ * [`fn withImagePullSecrets(imagePullSecrets)`](#fn-specwithimagepullsecrets)
+ * [`fn withImagePullSecretsMixin(imagePullSecrets)`](#fn-specwithimagepullsecretsmixin)
+ * [`fn withInstances(instances)`](#fn-specwithinstances)
+ * [`fn withLogLevel(logLevel)`](#fn-specwithloglevel)
+ * [`fn withMaxSyncReplicas(maxSyncReplicas)`](#fn-specwithmaxsyncreplicas)
+ * [`fn withMinSyncReplicas(minSyncReplicas)`](#fn-specwithminsyncreplicas)
+ * [`fn withPostgresGID(postgresGID)`](#fn-specwithpostgresgid)
+ * [`fn withPostgresUID(postgresUID)`](#fn-specwithpostgresuid)
+ * [`fn withPrimaryUpdateMethod(primaryUpdateMethod)`](#fn-specwithprimaryupdatemethod)
+ * [`fn withPrimaryUpdateStrategy(primaryUpdateStrategy)`](#fn-specwithprimaryupdatestrategy)
+ * [`fn withPriorityClassName(priorityClassName)`](#fn-specwithpriorityclassname)
+ * [`fn withSchedulerName(schedulerName)`](#fn-specwithschedulername)
+ * [`fn withSmartShutdownTimeout(smartShutdownTimeout)`](#fn-specwithsmartshutdowntimeout)
+ * [`fn withStartDelay(startDelay)`](#fn-specwithstartdelay)
+ * [`fn withStopDelay(stopDelay)`](#fn-specwithstopdelay)
+ * [`fn withSwitchoverDelay(switchoverDelay)`](#fn-specwithswitchoverdelay)
+ * [`fn withTopologySpreadConstraints(topologySpreadConstraints)`](#fn-specwithtopologyspreadconstraints)
+ * [`fn withTopologySpreadConstraintsMixin(topologySpreadConstraints)`](#fn-specwithtopologyspreadconstraintsmixin)
+ * [`obj spec.affinity`](#obj-specaffinity)
+ * [`fn withEnablePodAntiAffinity(enablePodAntiAffinity)`](#fn-specaffinitywithenablepodantiaffinity)
+ * [`fn withNodeSelector(nodeSelector)`](#fn-specaffinitywithnodeselector)
+ * [`fn withNodeSelectorMixin(nodeSelector)`](#fn-specaffinitywithnodeselectormixin)
+ * [`fn withPodAntiAffinityType(podAntiAffinityType)`](#fn-specaffinitywithpodantiaffinitytype)
+ * [`fn withTolerations(tolerations)`](#fn-specaffinitywithtolerations)
+ * [`fn withTolerationsMixin(tolerations)`](#fn-specaffinitywithtolerationsmixin)
+ * [`fn withTopologyKey(topologyKey)`](#fn-specaffinitywithtopologykey)
+ * [`obj spec.affinity.additionalPodAffinity`](#obj-specaffinityadditionalpodaffinity)
+ * [`fn withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution)`](#fn-specaffinityadditionalpodaffinitywithpreferredduringschedulingignoredduringexecution)
+ * [`fn withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution)`](#fn-specaffinityadditionalpodaffinitywithpreferredduringschedulingignoredduringexecutionmixin)
+ * [`fn withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution)`](#fn-specaffinityadditionalpodaffinitywithrequiredduringschedulingignoredduringexecution)
+ * [`fn withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution)`](#fn-specaffinityadditionalpodaffinitywithrequiredduringschedulingignoredduringexecutionmixin)
+ * [`obj spec.affinity.additionalPodAffinity.preferredDuringSchedulingIgnoredDuringExecution`](#obj-specaffinityadditionalpodaffinitypreferredduringschedulingignoredduringexecution)
+ * [`fn withWeight(weight)`](#fn-specaffinityadditionalpodaffinitypreferredduringschedulingignoredduringexecutionwithweight)
+ * [`obj spec.affinity.additionalPodAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm`](#obj-specaffinityadditionalpodaffinitypreferredduringschedulingignoredduringexecutionpodaffinityterm)
+ * [`fn withNamespaces(namespaces)`](#fn-specaffinityadditionalpodaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermwithnamespaces)
+ * [`fn withNamespacesMixin(namespaces)`](#fn-specaffinityadditionalpodaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermwithnamespacesmixin)
+ * [`fn withTopologyKey(topologyKey)`](#fn-specaffinityadditionalpodaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermwithtopologykey)
+ * [`obj spec.affinity.additionalPodAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.labelSelector`](#obj-specaffinityadditionalpodaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermlabelselector)
+ * [`fn withMatchExpressions(matchExpressions)`](#fn-specaffinityadditionalpodaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermlabelselectorwithmatchexpressions)
+ * [`fn withMatchExpressionsMixin(matchExpressions)`](#fn-specaffinityadditionalpodaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermlabelselectorwithmatchexpressionsmixin)
+ * [`fn withMatchLabels(matchLabels)`](#fn-specaffinityadditionalpodaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermlabelselectorwithmatchlabels)
+ * [`fn withMatchLabelsMixin(matchLabels)`](#fn-specaffinityadditionalpodaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermlabelselectorwithmatchlabelsmixin)
+ * [`obj spec.affinity.additionalPodAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.labelSelector.matchExpressions`](#obj-specaffinityadditionalpodaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermlabelselectormatchexpressions)
+ * [`fn withKey(key)`](#fn-specaffinityadditionalpodaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermlabelselectormatchexpressionswithkey)
+ * [`fn withOperator(operator)`](#fn-specaffinityadditionalpodaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermlabelselectormatchexpressionswithoperator)
+ * [`fn withValues(values)`](#fn-specaffinityadditionalpodaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermlabelselectormatchexpressionswithvalues)
+ * [`fn withValuesMixin(values)`](#fn-specaffinityadditionalpodaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermlabelselectormatchexpressionswithvaluesmixin)
+ * [`obj spec.affinity.additionalPodAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.namespaceSelector`](#obj-specaffinityadditionalpodaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermnamespaceselector)
+ * [`fn withMatchExpressions(matchExpressions)`](#fn-specaffinityadditionalpodaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermnamespaceselectorwithmatchexpressions)
+ * [`fn withMatchExpressionsMixin(matchExpressions)`](#fn-specaffinityadditionalpodaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermnamespaceselectorwithmatchexpressionsmixin)
+ * [`fn withMatchLabels(matchLabels)`](#fn-specaffinityadditionalpodaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermnamespaceselectorwithmatchlabels)
+ * [`fn withMatchLabelsMixin(matchLabels)`](#fn-specaffinityadditionalpodaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermnamespaceselectorwithmatchlabelsmixin)
+ * [`obj spec.affinity.additionalPodAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.namespaceSelector.matchExpressions`](#obj-specaffinityadditionalpodaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermnamespaceselectormatchexpressions)
+ * [`fn withKey(key)`](#fn-specaffinityadditionalpodaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermnamespaceselectormatchexpressionswithkey)
+ * [`fn withOperator(operator)`](#fn-specaffinityadditionalpodaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermnamespaceselectormatchexpressionswithoperator)
+ * [`fn withValues(values)`](#fn-specaffinityadditionalpodaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermnamespaceselectormatchexpressionswithvalues)
+ * [`fn withValuesMixin(values)`](#fn-specaffinityadditionalpodaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermnamespaceselectormatchexpressionswithvaluesmixin)
+ * [`obj spec.affinity.additionalPodAffinity.requiredDuringSchedulingIgnoredDuringExecution`](#obj-specaffinityadditionalpodaffinityrequiredduringschedulingignoredduringexecution)
+ * [`fn withNamespaces(namespaces)`](#fn-specaffinityadditionalpodaffinityrequiredduringschedulingignoredduringexecutionwithnamespaces)
+ * [`fn withNamespacesMixin(namespaces)`](#fn-specaffinityadditionalpodaffinityrequiredduringschedulingignoredduringexecutionwithnamespacesmixin)
+ * [`fn withTopologyKey(topologyKey)`](#fn-specaffinityadditionalpodaffinityrequiredduringschedulingignoredduringexecutionwithtopologykey)
+ * [`obj spec.affinity.additionalPodAffinity.requiredDuringSchedulingIgnoredDuringExecution.labelSelector`](#obj-specaffinityadditionalpodaffinityrequiredduringschedulingignoredduringexecutionlabelselector)
+ * [`fn withMatchExpressions(matchExpressions)`](#fn-specaffinityadditionalpodaffinityrequiredduringschedulingignoredduringexecutionlabelselectorwithmatchexpressions)
+ * [`fn withMatchExpressionsMixin(matchExpressions)`](#fn-specaffinityadditionalpodaffinityrequiredduringschedulingignoredduringexecutionlabelselectorwithmatchexpressionsmixin)
+ * [`fn withMatchLabels(matchLabels)`](#fn-specaffinityadditionalpodaffinityrequiredduringschedulingignoredduringexecutionlabelselectorwithmatchlabels)
+ * [`fn withMatchLabelsMixin(matchLabels)`](#fn-specaffinityadditionalpodaffinityrequiredduringschedulingignoredduringexecutionlabelselectorwithmatchlabelsmixin)
+ * [`obj spec.affinity.additionalPodAffinity.requiredDuringSchedulingIgnoredDuringExecution.labelSelector.matchExpressions`](#obj-specaffinityadditionalpodaffinityrequiredduringschedulingignoredduringexecutionlabelselectormatchexpressions)
+ * [`fn withKey(key)`](#fn-specaffinityadditionalpodaffinityrequiredduringschedulingignoredduringexecutionlabelselectormatchexpressionswithkey)
+ * [`fn withOperator(operator)`](#fn-specaffinityadditionalpodaffinityrequiredduringschedulingignoredduringexecutionlabelselectormatchexpressionswithoperator)
+ * [`fn withValues(values)`](#fn-specaffinityadditionalpodaffinityrequiredduringschedulingignoredduringexecutionlabelselectormatchexpressionswithvalues)
+ * [`fn withValuesMixin(values)`](#fn-specaffinityadditionalpodaffinityrequiredduringschedulingignoredduringexecutionlabelselectormatchexpressionswithvaluesmixin)
+ * [`obj spec.affinity.additionalPodAffinity.requiredDuringSchedulingIgnoredDuringExecution.namespaceSelector`](#obj-specaffinityadditionalpodaffinityrequiredduringschedulingignoredduringexecutionnamespaceselector)
+ * [`fn withMatchExpressions(matchExpressions)`](#fn-specaffinityadditionalpodaffinityrequiredduringschedulingignoredduringexecutionnamespaceselectorwithmatchexpressions)
+ * [`fn withMatchExpressionsMixin(matchExpressions)`](#fn-specaffinityadditionalpodaffinityrequiredduringschedulingignoredduringexecutionnamespaceselectorwithmatchexpressionsmixin)
+ * [`fn withMatchLabels(matchLabels)`](#fn-specaffinityadditionalpodaffinityrequiredduringschedulingignoredduringexecutionnamespaceselectorwithmatchlabels)
+ * [`fn withMatchLabelsMixin(matchLabels)`](#fn-specaffinityadditionalpodaffinityrequiredduringschedulingignoredduringexecutionnamespaceselectorwithmatchlabelsmixin)
+ * [`obj spec.affinity.additionalPodAffinity.requiredDuringSchedulingIgnoredDuringExecution.namespaceSelector.matchExpressions`](#obj-specaffinityadditionalpodaffinityrequiredduringschedulingignoredduringexecutionnamespaceselectormatchexpressions)
+ * [`fn withKey(key)`](#fn-specaffinityadditionalpodaffinityrequiredduringschedulingignoredduringexecutionnamespaceselectormatchexpressionswithkey)
+ * [`fn withOperator(operator)`](#fn-specaffinityadditionalpodaffinityrequiredduringschedulingignoredduringexecutionnamespaceselectormatchexpressionswithoperator)
+ * [`fn withValues(values)`](#fn-specaffinityadditionalpodaffinityrequiredduringschedulingignoredduringexecutionnamespaceselectormatchexpressionswithvalues)
+ * [`fn withValuesMixin(values)`](#fn-specaffinityadditionalpodaffinityrequiredduringschedulingignoredduringexecutionnamespaceselectormatchexpressionswithvaluesmixin)
+ * [`obj spec.affinity.additionalPodAntiAffinity`](#obj-specaffinityadditionalpodantiaffinity)
+ * [`fn withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution)`](#fn-specaffinityadditionalpodantiaffinitywithpreferredduringschedulingignoredduringexecution)
+ * [`fn withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution)`](#fn-specaffinityadditionalpodantiaffinitywithpreferredduringschedulingignoredduringexecutionmixin)
+ * [`fn withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution)`](#fn-specaffinityadditionalpodantiaffinitywithrequiredduringschedulingignoredduringexecution)
+ * [`fn withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution)`](#fn-specaffinityadditionalpodantiaffinitywithrequiredduringschedulingignoredduringexecutionmixin)
+ * [`obj spec.affinity.additionalPodAntiAffinity.preferredDuringSchedulingIgnoredDuringExecution`](#obj-specaffinityadditionalpodantiaffinitypreferredduringschedulingignoredduringexecution)
+ * [`fn withWeight(weight)`](#fn-specaffinityadditionalpodantiaffinitypreferredduringschedulingignoredduringexecutionwithweight)
+ * [`obj spec.affinity.additionalPodAntiAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm`](#obj-specaffinityadditionalpodantiaffinitypreferredduringschedulingignoredduringexecutionpodaffinityterm)
+ * [`fn withNamespaces(namespaces)`](#fn-specaffinityadditionalpodantiaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermwithnamespaces)
+ * [`fn withNamespacesMixin(namespaces)`](#fn-specaffinityadditionalpodantiaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermwithnamespacesmixin)
+ * [`fn withTopologyKey(topologyKey)`](#fn-specaffinityadditionalpodantiaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermwithtopologykey)
+ * [`obj spec.affinity.additionalPodAntiAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.labelSelector`](#obj-specaffinityadditionalpodantiaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermlabelselector)
+ * [`fn withMatchExpressions(matchExpressions)`](#fn-specaffinityadditionalpodantiaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermlabelselectorwithmatchexpressions)
+ * [`fn withMatchExpressionsMixin(matchExpressions)`](#fn-specaffinityadditionalpodantiaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermlabelselectorwithmatchexpressionsmixin)
+ * [`fn withMatchLabels(matchLabels)`](#fn-specaffinityadditionalpodantiaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermlabelselectorwithmatchlabels)
+ * [`fn withMatchLabelsMixin(matchLabels)`](#fn-specaffinityadditionalpodantiaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermlabelselectorwithmatchlabelsmixin)
+ * [`obj spec.affinity.additionalPodAntiAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.labelSelector.matchExpressions`](#obj-specaffinityadditionalpodantiaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermlabelselectormatchexpressions)
+ * [`fn withKey(key)`](#fn-specaffinityadditionalpodantiaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermlabelselectormatchexpressionswithkey)
+ * [`fn withOperator(operator)`](#fn-specaffinityadditionalpodantiaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermlabelselectormatchexpressionswithoperator)
+ * [`fn withValues(values)`](#fn-specaffinityadditionalpodantiaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermlabelselectormatchexpressionswithvalues)
+ * [`fn withValuesMixin(values)`](#fn-specaffinityadditionalpodantiaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermlabelselectormatchexpressionswithvaluesmixin)
+ * [`obj spec.affinity.additionalPodAntiAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.namespaceSelector`](#obj-specaffinityadditionalpodantiaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermnamespaceselector)
+ * [`fn withMatchExpressions(matchExpressions)`](#fn-specaffinityadditionalpodantiaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermnamespaceselectorwithmatchexpressions)
+ * [`fn withMatchExpressionsMixin(matchExpressions)`](#fn-specaffinityadditionalpodantiaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermnamespaceselectorwithmatchexpressionsmixin)
+ * [`fn withMatchLabels(matchLabels)`](#fn-specaffinityadditionalpodantiaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermnamespaceselectorwithmatchlabels)
+ * [`fn withMatchLabelsMixin(matchLabels)`](#fn-specaffinityadditionalpodantiaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermnamespaceselectorwithmatchlabelsmixin)
+ * [`obj spec.affinity.additionalPodAntiAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.namespaceSelector.matchExpressions`](#obj-specaffinityadditionalpodantiaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermnamespaceselectormatchexpressions)
+ * [`fn withKey(key)`](#fn-specaffinityadditionalpodantiaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermnamespaceselectormatchexpressionswithkey)
+ * [`fn withOperator(operator)`](#fn-specaffinityadditionalpodantiaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermnamespaceselectormatchexpressionswithoperator)
+ * [`fn withValues(values)`](#fn-specaffinityadditionalpodantiaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermnamespaceselectormatchexpressionswithvalues)
+ * [`fn withValuesMixin(values)`](#fn-specaffinityadditionalpodantiaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermnamespaceselectormatchexpressionswithvaluesmixin)
+ * [`obj spec.affinity.additionalPodAntiAffinity.requiredDuringSchedulingIgnoredDuringExecution`](#obj-specaffinityadditionalpodantiaffinityrequiredduringschedulingignoredduringexecution)
+ * [`fn withNamespaces(namespaces)`](#fn-specaffinityadditionalpodantiaffinityrequiredduringschedulingignoredduringexecutionwithnamespaces)
+ * [`fn withNamespacesMixin(namespaces)`](#fn-specaffinityadditionalpodantiaffinityrequiredduringschedulingignoredduringexecutionwithnamespacesmixin)
+ * [`fn withTopologyKey(topologyKey)`](#fn-specaffinityadditionalpodantiaffinityrequiredduringschedulingignoredduringexecutionwithtopologykey)
+ * [`obj spec.affinity.additionalPodAntiAffinity.requiredDuringSchedulingIgnoredDuringExecution.labelSelector`](#obj-specaffinityadditionalpodantiaffinityrequiredduringschedulingignoredduringexecutionlabelselector)
+ * [`fn withMatchExpressions(matchExpressions)`](#fn-specaffinityadditionalpodantiaffinityrequiredduringschedulingignoredduringexecutionlabelselectorwithmatchexpressions)
+ * [`fn withMatchExpressionsMixin(matchExpressions)`](#fn-specaffinityadditionalpodantiaffinityrequiredduringschedulingignoredduringexecutionlabelselectorwithmatchexpressionsmixin)
+ * [`fn withMatchLabels(matchLabels)`](#fn-specaffinityadditionalpodantiaffinityrequiredduringschedulingignoredduringexecutionlabelselectorwithmatchlabels)
+ * [`fn withMatchLabelsMixin(matchLabels)`](#fn-specaffinityadditionalpodantiaffinityrequiredduringschedulingignoredduringexecutionlabelselectorwithmatchlabelsmixin)
+ * [`obj spec.affinity.additionalPodAntiAffinity.requiredDuringSchedulingIgnoredDuringExecution.labelSelector.matchExpressions`](#obj-specaffinityadditionalpodantiaffinityrequiredduringschedulingignoredduringexecutionlabelselectormatchexpressions)
+ * [`fn withKey(key)`](#fn-specaffinityadditionalpodantiaffinityrequiredduringschedulingignoredduringexecutionlabelselectormatchexpressionswithkey)
+ * [`fn withOperator(operator)`](#fn-specaffinityadditionalpodantiaffinityrequiredduringschedulingignoredduringexecutionlabelselectormatchexpressionswithoperator)
+ * [`fn withValues(values)`](#fn-specaffinityadditionalpodantiaffinityrequiredduringschedulingignoredduringexecutionlabelselectormatchexpressionswithvalues)
+ * [`fn withValuesMixin(values)`](#fn-specaffinityadditionalpodantiaffinityrequiredduringschedulingignoredduringexecutionlabelselectormatchexpressionswithvaluesmixin)
+ * [`obj spec.affinity.additionalPodAntiAffinity.requiredDuringSchedulingIgnoredDuringExecution.namespaceSelector`](#obj-specaffinityadditionalpodantiaffinityrequiredduringschedulingignoredduringexecutionnamespaceselector)
+ * [`fn withMatchExpressions(matchExpressions)`](#fn-specaffinityadditionalpodantiaffinityrequiredduringschedulingignoredduringexecutionnamespaceselectorwithmatchexpressions)
+ * [`fn withMatchExpressionsMixin(matchExpressions)`](#fn-specaffinityadditionalpodantiaffinityrequiredduringschedulingignoredduringexecutionnamespaceselectorwithmatchexpressionsmixin)
+ * [`fn withMatchLabels(matchLabels)`](#fn-specaffinityadditionalpodantiaffinityrequiredduringschedulingignoredduringexecutionnamespaceselectorwithmatchlabels)
+ * [`fn withMatchLabelsMixin(matchLabels)`](#fn-specaffinityadditionalpodantiaffinityrequiredduringschedulingignoredduringexecutionnamespaceselectorwithmatchlabelsmixin)
+ * [`obj spec.affinity.additionalPodAntiAffinity.requiredDuringSchedulingIgnoredDuringExecution.namespaceSelector.matchExpressions`](#obj-specaffinityadditionalpodantiaffinityrequiredduringschedulingignoredduringexecutionnamespaceselectormatchexpressions)
+ * [`fn withKey(key)`](#fn-specaffinityadditionalpodantiaffinityrequiredduringschedulingignoredduringexecutionnamespaceselectormatchexpressionswithkey)
+ * [`fn withOperator(operator)`](#fn-specaffinityadditionalpodantiaffinityrequiredduringschedulingignoredduringexecutionnamespaceselectormatchexpressionswithoperator)
+ * [`fn withValues(values)`](#fn-specaffinityadditionalpodantiaffinityrequiredduringschedulingignoredduringexecutionnamespaceselectormatchexpressionswithvalues)
+ * [`fn withValuesMixin(values)`](#fn-specaffinityadditionalpodantiaffinityrequiredduringschedulingignoredduringexecutionnamespaceselectormatchexpressionswithvaluesmixin)
+ * [`obj spec.affinity.nodeAffinity`](#obj-specaffinitynodeaffinity)
+ * [`fn withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution)`](#fn-specaffinitynodeaffinitywithpreferredduringschedulingignoredduringexecution)
+ * [`fn withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution)`](#fn-specaffinitynodeaffinitywithpreferredduringschedulingignoredduringexecutionmixin)
+ * [`obj spec.affinity.nodeAffinity.preferredDuringSchedulingIgnoredDuringExecution`](#obj-specaffinitynodeaffinitypreferredduringschedulingignoredduringexecution)
+ * [`fn withWeight(weight)`](#fn-specaffinitynodeaffinitypreferredduringschedulingignoredduringexecutionwithweight)
+ * [`obj spec.affinity.nodeAffinity.preferredDuringSchedulingIgnoredDuringExecution.preference`](#obj-specaffinitynodeaffinitypreferredduringschedulingignoredduringexecutionpreference)
+ * [`fn withMatchExpressions(matchExpressions)`](#fn-specaffinitynodeaffinitypreferredduringschedulingignoredduringexecutionpreferencewithmatchexpressions)
+ * [`fn withMatchExpressionsMixin(matchExpressions)`](#fn-specaffinitynodeaffinitypreferredduringschedulingignoredduringexecutionpreferencewithmatchexpressionsmixin)
+ * [`fn withMatchFields(matchFields)`](#fn-specaffinitynodeaffinitypreferredduringschedulingignoredduringexecutionpreferencewithmatchfields)
+ * [`fn withMatchFieldsMixin(matchFields)`](#fn-specaffinitynodeaffinitypreferredduringschedulingignoredduringexecutionpreferencewithmatchfieldsmixin)
+ * [`obj spec.affinity.nodeAffinity.preferredDuringSchedulingIgnoredDuringExecution.preference.matchExpressions`](#obj-specaffinitynodeaffinitypreferredduringschedulingignoredduringexecutionpreferencematchexpressions)
+ * [`fn withKey(key)`](#fn-specaffinitynodeaffinitypreferredduringschedulingignoredduringexecutionpreferencematchexpressionswithkey)
+ * [`fn withOperator(operator)`](#fn-specaffinitynodeaffinitypreferredduringschedulingignoredduringexecutionpreferencematchexpressionswithoperator)
+ * [`fn withValues(values)`](#fn-specaffinitynodeaffinitypreferredduringschedulingignoredduringexecutionpreferencematchexpressionswithvalues)
+ * [`fn withValuesMixin(values)`](#fn-specaffinitynodeaffinitypreferredduringschedulingignoredduringexecutionpreferencematchexpressionswithvaluesmixin)
+ * [`obj spec.affinity.nodeAffinity.preferredDuringSchedulingIgnoredDuringExecution.preference.matchFields`](#obj-specaffinitynodeaffinitypreferredduringschedulingignoredduringexecutionpreferencematchfields)
+ * [`fn withKey(key)`](#fn-specaffinitynodeaffinitypreferredduringschedulingignoredduringexecutionpreferencematchfieldswithkey)
+ * [`fn withOperator(operator)`](#fn-specaffinitynodeaffinitypreferredduringschedulingignoredduringexecutionpreferencematchfieldswithoperator)
+ * [`fn withValues(values)`](#fn-specaffinitynodeaffinitypreferredduringschedulingignoredduringexecutionpreferencematchfieldswithvalues)
+ * [`fn withValuesMixin(values)`](#fn-specaffinitynodeaffinitypreferredduringschedulingignoredduringexecutionpreferencematchfieldswithvaluesmixin)
+ * [`obj spec.affinity.nodeAffinity.requiredDuringSchedulingIgnoredDuringExecution`](#obj-specaffinitynodeaffinityrequiredduringschedulingignoredduringexecution)
+ * [`fn withNodeSelectorTerms(nodeSelectorTerms)`](#fn-specaffinitynodeaffinityrequiredduringschedulingignoredduringexecutionwithnodeselectorterms)
+ * [`fn withNodeSelectorTermsMixin(nodeSelectorTerms)`](#fn-specaffinitynodeaffinityrequiredduringschedulingignoredduringexecutionwithnodeselectortermsmixin)
+ * [`obj spec.affinity.nodeAffinity.requiredDuringSchedulingIgnoredDuringExecution.nodeSelectorTerms`](#obj-specaffinitynodeaffinityrequiredduringschedulingignoredduringexecutionnodeselectorterms)
+ * [`fn withMatchExpressions(matchExpressions)`](#fn-specaffinitynodeaffinityrequiredduringschedulingignoredduringexecutionnodeselectortermswithmatchexpressions)
+ * [`fn withMatchExpressionsMixin(matchExpressions)`](#fn-specaffinitynodeaffinityrequiredduringschedulingignoredduringexecutionnodeselectortermswithmatchexpressionsmixin)
+ * [`fn withMatchFields(matchFields)`](#fn-specaffinitynodeaffinityrequiredduringschedulingignoredduringexecutionnodeselectortermswithmatchfields)
+ * [`fn withMatchFieldsMixin(matchFields)`](#fn-specaffinitynodeaffinityrequiredduringschedulingignoredduringexecutionnodeselectortermswithmatchfieldsmixin)
+ * [`obj spec.affinity.nodeAffinity.requiredDuringSchedulingIgnoredDuringExecution.nodeSelectorTerms.matchExpressions`](#obj-specaffinitynodeaffinityrequiredduringschedulingignoredduringexecutionnodeselectortermsmatchexpressions)
+ * [`fn withKey(key)`](#fn-specaffinitynodeaffinityrequiredduringschedulingignoredduringexecutionnodeselectortermsmatchexpressionswithkey)
+ * [`fn withOperator(operator)`](#fn-specaffinitynodeaffinityrequiredduringschedulingignoredduringexecutionnodeselectortermsmatchexpressionswithoperator)
+ * [`fn withValues(values)`](#fn-specaffinitynodeaffinityrequiredduringschedulingignoredduringexecutionnodeselectortermsmatchexpressionswithvalues)
+ * [`fn withValuesMixin(values)`](#fn-specaffinitynodeaffinityrequiredduringschedulingignoredduringexecutionnodeselectortermsmatchexpressionswithvaluesmixin)
+ * [`obj spec.affinity.nodeAffinity.requiredDuringSchedulingIgnoredDuringExecution.nodeSelectorTerms.matchFields`](#obj-specaffinitynodeaffinityrequiredduringschedulingignoredduringexecutionnodeselectortermsmatchfields)
+ * [`fn withKey(key)`](#fn-specaffinitynodeaffinityrequiredduringschedulingignoredduringexecutionnodeselectortermsmatchfieldswithkey)
+ * [`fn withOperator(operator)`](#fn-specaffinitynodeaffinityrequiredduringschedulingignoredduringexecutionnodeselectortermsmatchfieldswithoperator)
+ * [`fn withValues(values)`](#fn-specaffinitynodeaffinityrequiredduringschedulingignoredduringexecutionnodeselectortermsmatchfieldswithvalues)
+ * [`fn withValuesMixin(values)`](#fn-specaffinitynodeaffinityrequiredduringschedulingignoredduringexecutionnodeselectortermsmatchfieldswithvaluesmixin)
+ * [`obj spec.affinity.tolerations`](#obj-specaffinitytolerations)
+ * [`fn withEffect(effect)`](#fn-specaffinitytolerationswitheffect)
+ * [`fn withKey(key)`](#fn-specaffinitytolerationswithkey)
+ * [`fn withOperator(operator)`](#fn-specaffinitytolerationswithoperator)
+ * [`fn withTolerationSeconds(tolerationSeconds)`](#fn-specaffinitytolerationswithtolerationseconds)
+ * [`fn withValue(value)`](#fn-specaffinitytolerationswithvalue)
+ * [`obj spec.backup`](#obj-specbackup)
+ * [`fn withRetentionPolicy(retentionPolicy)`](#fn-specbackupwithretentionpolicy)
+ * [`fn withTarget(target)`](#fn-specbackupwithtarget)
+ * [`obj spec.backup.barmanObjectStore`](#obj-specbackupbarmanobjectstore)
+ * [`fn withDestinationPath(destinationPath)`](#fn-specbackupbarmanobjectstorewithdestinationpath)
+ * [`fn withEndpointURL(endpointURL)`](#fn-specbackupbarmanobjectstorewithendpointurl)
+ * [`fn withHistoryTags(historyTags)`](#fn-specbackupbarmanobjectstorewithhistorytags)
+ * [`fn withHistoryTagsMixin(historyTags)`](#fn-specbackupbarmanobjectstorewithhistorytagsmixin)
+ * [`fn withServerName(serverName)`](#fn-specbackupbarmanobjectstorewithservername)
+ * [`fn withTags(tags)`](#fn-specbackupbarmanobjectstorewithtags)
+ * [`fn withTagsMixin(tags)`](#fn-specbackupbarmanobjectstorewithtagsmixin)
+ * [`obj spec.backup.barmanObjectStore.azureCredentials`](#obj-specbackupbarmanobjectstoreazurecredentials)
+ * [`fn withInheritFromAzureAD(inheritFromAzureAD)`](#fn-specbackupbarmanobjectstoreazurecredentialswithinheritfromazuread)
+ * [`obj spec.backup.barmanObjectStore.azureCredentials.connectionString`](#obj-specbackupbarmanobjectstoreazurecredentialsconnectionstring)
+ * [`fn withKey(key)`](#fn-specbackupbarmanobjectstoreazurecredentialsconnectionstringwithkey)
+ * [`fn withName(name)`](#fn-specbackupbarmanobjectstoreazurecredentialsconnectionstringwithname)
+ * [`obj spec.backup.barmanObjectStore.azureCredentials.storageAccount`](#obj-specbackupbarmanobjectstoreazurecredentialsstorageaccount)
+ * [`fn withKey(key)`](#fn-specbackupbarmanobjectstoreazurecredentialsstorageaccountwithkey)
+ * [`fn withName(name)`](#fn-specbackupbarmanobjectstoreazurecredentialsstorageaccountwithname)
+ * [`obj spec.backup.barmanObjectStore.azureCredentials.storageKey`](#obj-specbackupbarmanobjectstoreazurecredentialsstoragekey)
+ * [`fn withKey(key)`](#fn-specbackupbarmanobjectstoreazurecredentialsstoragekeywithkey)
+ * [`fn withName(name)`](#fn-specbackupbarmanobjectstoreazurecredentialsstoragekeywithname)
+ * [`obj spec.backup.barmanObjectStore.azureCredentials.storageSasToken`](#obj-specbackupbarmanobjectstoreazurecredentialsstoragesastoken)
+ * [`fn withKey(key)`](#fn-specbackupbarmanobjectstoreazurecredentialsstoragesastokenwithkey)
+ * [`fn withName(name)`](#fn-specbackupbarmanobjectstoreazurecredentialsstoragesastokenwithname)
+ * [`obj spec.backup.barmanObjectStore.data`](#obj-specbackupbarmanobjectstoredata)
+ * [`fn withCompression(compression)`](#fn-specbackupbarmanobjectstoredatawithcompression)
+ * [`fn withEncryption(encryption)`](#fn-specbackupbarmanobjectstoredatawithencryption)
+ * [`fn withImmediateCheckpoint(immediateCheckpoint)`](#fn-specbackupbarmanobjectstoredatawithimmediatecheckpoint)
+ * [`fn withJobs(jobs)`](#fn-specbackupbarmanobjectstoredatawithjobs)
+ * [`obj spec.backup.barmanObjectStore.endpointCA`](#obj-specbackupbarmanobjectstoreendpointca)
+ * [`fn withKey(key)`](#fn-specbackupbarmanobjectstoreendpointcawithkey)
+ * [`fn withName(name)`](#fn-specbackupbarmanobjectstoreendpointcawithname)
+ * [`obj spec.backup.barmanObjectStore.googleCredentials`](#obj-specbackupbarmanobjectstoregooglecredentials)
+ * [`fn withGkeEnvironment(gkeEnvironment)`](#fn-specbackupbarmanobjectstoregooglecredentialswithgkeenvironment)
+ * [`obj spec.backup.barmanObjectStore.googleCredentials.applicationCredentials`](#obj-specbackupbarmanobjectstoregooglecredentialsapplicationcredentials)
+ * [`fn withKey(key)`](#fn-specbackupbarmanobjectstoregooglecredentialsapplicationcredentialswithkey)
+ * [`fn withName(name)`](#fn-specbackupbarmanobjectstoregooglecredentialsapplicationcredentialswithname)
+ * [`obj spec.backup.barmanObjectStore.s3Credentials`](#obj-specbackupbarmanobjectstores3credentials)
+ * [`fn withInheritFromIAMRole(inheritFromIAMRole)`](#fn-specbackupbarmanobjectstores3credentialswithinheritfromiamrole)
+ * [`obj spec.backup.barmanObjectStore.s3Credentials.accessKeyId`](#obj-specbackupbarmanobjectstores3credentialsaccesskeyid)
+ * [`fn withKey(key)`](#fn-specbackupbarmanobjectstores3credentialsaccesskeyidwithkey)
+ * [`fn withName(name)`](#fn-specbackupbarmanobjectstores3credentialsaccesskeyidwithname)
+ * [`obj spec.backup.barmanObjectStore.s3Credentials.region`](#obj-specbackupbarmanobjectstores3credentialsregion)
+ * [`fn withKey(key)`](#fn-specbackupbarmanobjectstores3credentialsregionwithkey)
+ * [`fn withName(name)`](#fn-specbackupbarmanobjectstores3credentialsregionwithname)
+ * [`obj spec.backup.barmanObjectStore.s3Credentials.secretAccessKey`](#obj-specbackupbarmanobjectstores3credentialssecretaccesskey)
+ * [`fn withKey(key)`](#fn-specbackupbarmanobjectstores3credentialssecretaccesskeywithkey)
+ * [`fn withName(name)`](#fn-specbackupbarmanobjectstores3credentialssecretaccesskeywithname)
+ * [`obj spec.backup.barmanObjectStore.s3Credentials.sessionToken`](#obj-specbackupbarmanobjectstores3credentialssessiontoken)
+ * [`fn withKey(key)`](#fn-specbackupbarmanobjectstores3credentialssessiontokenwithkey)
+ * [`fn withName(name)`](#fn-specbackupbarmanobjectstores3credentialssessiontokenwithname)
+ * [`obj spec.backup.barmanObjectStore.wal`](#obj-specbackupbarmanobjectstorewal)
+ * [`fn withCompression(compression)`](#fn-specbackupbarmanobjectstorewalwithcompression)
+ * [`fn withEncryption(encryption)`](#fn-specbackupbarmanobjectstorewalwithencryption)
+ * [`fn withMaxParallel(maxParallel)`](#fn-specbackupbarmanobjectstorewalwithmaxparallel)
+ * [`obj spec.backup.volumeSnapshot`](#obj-specbackupvolumesnapshot)
+ * [`fn withAnnotations(annotations)`](#fn-specbackupvolumesnapshotwithannotations)
+ * [`fn withAnnotationsMixin(annotations)`](#fn-specbackupvolumesnapshotwithannotationsmixin)
+ * [`fn withClassName(className)`](#fn-specbackupvolumesnapshotwithclassname)
+ * [`fn withLabels(labels)`](#fn-specbackupvolumesnapshotwithlabels)
+ * [`fn withLabelsMixin(labels)`](#fn-specbackupvolumesnapshotwithlabelsmixin)
+ * [`fn withSnapshotOwnerReference(snapshotOwnerReference)`](#fn-specbackupvolumesnapshotwithsnapshotownerreference)
+ * [`fn withWalClassName(walClassName)`](#fn-specbackupvolumesnapshotwithwalclassname)
+ * [`obj spec.bootstrap`](#obj-specbootstrap)
+ * [`obj spec.bootstrap.initdb`](#obj-specbootstrapinitdb)
+ * [`fn withDataChecksums(dataChecksums)`](#fn-specbootstrapinitdbwithdatachecksums)
+ * [`fn withDatabase(database)`](#fn-specbootstrapinitdbwithdatabase)
+ * [`fn withEncoding(encoding)`](#fn-specbootstrapinitdbwithencoding)
+ * [`fn withLocaleCType(localeCType)`](#fn-specbootstrapinitdbwithlocalectype)
+ * [`fn withLocaleCollate(localeCollate)`](#fn-specbootstrapinitdbwithlocalecollate)
+ * [`fn withOptions(options)`](#fn-specbootstrapinitdbwithoptions)
+ * [`fn withOptionsMixin(options)`](#fn-specbootstrapinitdbwithoptionsmixin)
+ * [`fn withOwner(owner)`](#fn-specbootstrapinitdbwithowner)
+ * [`fn withPostInitApplicationSQL(postInitApplicationSQL)`](#fn-specbootstrapinitdbwithpostinitapplicationsql)
+ * [`fn withPostInitApplicationSQLMixin(postInitApplicationSQL)`](#fn-specbootstrapinitdbwithpostinitapplicationsqlmixin)
+ * [`fn withPostInitSQL(postInitSQL)`](#fn-specbootstrapinitdbwithpostinitsql)
+ * [`fn withPostInitSQLMixin(postInitSQL)`](#fn-specbootstrapinitdbwithpostinitsqlmixin)
+ * [`fn withPostInitTemplateSQL(postInitTemplateSQL)`](#fn-specbootstrapinitdbwithpostinittemplatesql)
+ * [`fn withPostInitTemplateSQLMixin(postInitTemplateSQL)`](#fn-specbootstrapinitdbwithpostinittemplatesqlmixin)
+ * [`fn withWalSegmentSize(walSegmentSize)`](#fn-specbootstrapinitdbwithwalsegmentsize)
+ * [`obj spec.bootstrap.initdb.import`](#obj-specbootstrapinitdbimport)
+ * [`fn withDatabases(databases)`](#fn-specbootstrapinitdbimportwithdatabases)
+ * [`fn withDatabasesMixin(databases)`](#fn-specbootstrapinitdbimportwithdatabasesmixin)
+ * [`fn withPostImportApplicationSQL(postImportApplicationSQL)`](#fn-specbootstrapinitdbimportwithpostimportapplicationsql)
+ * [`fn withPostImportApplicationSQLMixin(postImportApplicationSQL)`](#fn-specbootstrapinitdbimportwithpostimportapplicationsqlmixin)
+ * [`fn withRoles(roles)`](#fn-specbootstrapinitdbimportwithroles)
+ * [`fn withRolesMixin(roles)`](#fn-specbootstrapinitdbimportwithrolesmixin)
+ * [`fn withSchemaOnly(schemaOnly)`](#fn-specbootstrapinitdbimportwithschemaonly)
+ * [`fn withType(type)`](#fn-specbootstrapinitdbimportwithtype)
+ * [`obj spec.bootstrap.initdb.import.source`](#obj-specbootstrapinitdbimportsource)
+ * [`fn withExternalCluster(externalCluster)`](#fn-specbootstrapinitdbimportsourcewithexternalcluster)
+ * [`obj spec.bootstrap.initdb.postInitApplicationSQLRefs`](#obj-specbootstrapinitdbpostinitapplicationsqlrefs)
+ * [`fn withConfigMapRefs(configMapRefs)`](#fn-specbootstrapinitdbpostinitapplicationsqlrefswithconfigmaprefs)
+ * [`fn withConfigMapRefsMixin(configMapRefs)`](#fn-specbootstrapinitdbpostinitapplicationsqlrefswithconfigmaprefsmixin)
+ * [`fn withSecretRefs(secretRefs)`](#fn-specbootstrapinitdbpostinitapplicationsqlrefswithsecretrefs)
+ * [`fn withSecretRefsMixin(secretRefs)`](#fn-specbootstrapinitdbpostinitapplicationsqlrefswithsecretrefsmixin)
+ * [`obj spec.bootstrap.initdb.postInitApplicationSQLRefs.configMapRefs`](#obj-specbootstrapinitdbpostinitapplicationsqlrefsconfigmaprefs)
+ * [`fn withKey(key)`](#fn-specbootstrapinitdbpostinitapplicationsqlrefsconfigmaprefswithkey)
+ * [`fn withName(name)`](#fn-specbootstrapinitdbpostinitapplicationsqlrefsconfigmaprefswithname)
+ * [`obj spec.bootstrap.initdb.postInitApplicationSQLRefs.secretRefs`](#obj-specbootstrapinitdbpostinitapplicationsqlrefssecretrefs)
+ * [`fn withKey(key)`](#fn-specbootstrapinitdbpostinitapplicationsqlrefssecretrefswithkey)
+ * [`fn withName(name)`](#fn-specbootstrapinitdbpostinitapplicationsqlrefssecretrefswithname)
+ * [`obj spec.bootstrap.initdb.secret`](#obj-specbootstrapinitdbsecret)
+ * [`fn withName(name)`](#fn-specbootstrapinitdbsecretwithname)
+ * [`obj spec.bootstrap.pg_basebackup`](#obj-specbootstrappg_basebackup)
+ * [`fn withDatabase(database)`](#fn-specbootstrappg_basebackupwithdatabase)
+ * [`fn withOwner(owner)`](#fn-specbootstrappg_basebackupwithowner)
+ * [`fn withSource(source)`](#fn-specbootstrappg_basebackupwithsource)
+ * [`obj spec.bootstrap.pg_basebackup.secret`](#obj-specbootstrappg_basebackupsecret)
+ * [`fn withName(name)`](#fn-specbootstrappg_basebackupsecretwithname)
+ * [`obj spec.bootstrap.recovery`](#obj-specbootstraprecovery)
+ * [`fn withDatabase(database)`](#fn-specbootstraprecoverywithdatabase)
+ * [`fn withOwner(owner)`](#fn-specbootstraprecoverywithowner)
+ * [`fn withSource(source)`](#fn-specbootstraprecoverywithsource)
+ * [`obj spec.bootstrap.recovery.backup`](#obj-specbootstraprecoverybackup)
+ * [`fn withName(name)`](#fn-specbootstraprecoverybackupwithname)
+ * [`obj spec.bootstrap.recovery.backup.endpointCA`](#obj-specbootstraprecoverybackupendpointca)
+ * [`fn withKey(key)`](#fn-specbootstraprecoverybackupendpointcawithkey)
+ * [`fn withName(name)`](#fn-specbootstraprecoverybackupendpointcawithname)
+ * [`obj spec.bootstrap.recovery.recoveryTarget`](#obj-specbootstraprecoveryrecoverytarget)
+ * [`fn withBackupID(backupID)`](#fn-specbootstraprecoveryrecoverytargetwithbackupid)
+ * [`fn withExclusive(exclusive)`](#fn-specbootstraprecoveryrecoverytargetwithexclusive)
+ * [`fn withTargetImmediate(targetImmediate)`](#fn-specbootstraprecoveryrecoverytargetwithtargetimmediate)
+ * [`fn withTargetLSN(targetLSN)`](#fn-specbootstraprecoveryrecoverytargetwithtargetlsn)
+ * [`fn withTargetName(targetName)`](#fn-specbootstraprecoveryrecoverytargetwithtargetname)
+ * [`fn withTargetTLI(targetTLI)`](#fn-specbootstraprecoveryrecoverytargetwithtargettli)
+ * [`fn withTargetTime(targetTime)`](#fn-specbootstraprecoveryrecoverytargetwithtargettime)
+ * [`fn withTargetXID(targetXID)`](#fn-specbootstraprecoveryrecoverytargetwithtargetxid)
+ * [`obj spec.bootstrap.recovery.secret`](#obj-specbootstraprecoverysecret)
+ * [`fn withName(name)`](#fn-specbootstraprecoverysecretwithname)
+ * [`obj spec.bootstrap.recovery.volumeSnapshots`](#obj-specbootstraprecoveryvolumesnapshots)
+ * [`obj spec.bootstrap.recovery.volumeSnapshots.storage`](#obj-specbootstraprecoveryvolumesnapshotsstorage)
+ * [`fn withApiGroup(apiGroup)`](#fn-specbootstraprecoveryvolumesnapshotsstoragewithapigroup)
+ * [`fn withKind(kind)`](#fn-specbootstraprecoveryvolumesnapshotsstoragewithkind)
+ * [`fn withName(name)`](#fn-specbootstraprecoveryvolumesnapshotsstoragewithname)
+ * [`obj spec.bootstrap.recovery.volumeSnapshots.walStorage`](#obj-specbootstraprecoveryvolumesnapshotswalstorage)
+ * [`fn withApiGroup(apiGroup)`](#fn-specbootstraprecoveryvolumesnapshotswalstoragewithapigroup)
+ * [`fn withKind(kind)`](#fn-specbootstraprecoveryvolumesnapshotswalstoragewithkind)
+ * [`fn withName(name)`](#fn-specbootstraprecoveryvolumesnapshotswalstoragewithname)
+ * [`obj spec.certificates`](#obj-speccertificates)
+ * [`fn withClientCASecret(clientCASecret)`](#fn-speccertificateswithclientcasecret)
+ * [`fn withReplicationTLSSecret(replicationTLSSecret)`](#fn-speccertificateswithreplicationtlssecret)
+ * [`fn withServerAltDNSNames(serverAltDNSNames)`](#fn-speccertificateswithserveraltdnsnames)
+ * [`fn withServerAltDNSNamesMixin(serverAltDNSNames)`](#fn-speccertificateswithserveraltdnsnamesmixin)
+ * [`fn withServerCASecret(serverCASecret)`](#fn-speccertificateswithservercasecret)
+ * [`fn withServerTLSSecret(serverTLSSecret)`](#fn-speccertificateswithservertlssecret)
+ * [`obj spec.env`](#obj-specenv)
+ * [`fn withName(name)`](#fn-specenvwithname)
+ * [`fn withValue(value)`](#fn-specenvwithvalue)
+ * [`obj spec.env.valueFrom`](#obj-specenvvaluefrom)
+ * [`obj spec.env.valueFrom.configMapKeyRef`](#obj-specenvvaluefromconfigmapkeyref)
+ * [`fn withKey(key)`](#fn-specenvvaluefromconfigmapkeyrefwithkey)
+ * [`fn withName(name)`](#fn-specenvvaluefromconfigmapkeyrefwithname)
+ * [`fn withOptional(optional)`](#fn-specenvvaluefromconfigmapkeyrefwithoptional)
+ * [`obj spec.env.valueFrom.fieldRef`](#obj-specenvvaluefromfieldref)
+ * [`fn withApiVersion(apiVersion)`](#fn-specenvvaluefromfieldrefwithapiversion)
+ * [`fn withFieldPath(fieldPath)`](#fn-specenvvaluefromfieldrefwithfieldpath)
+ * [`obj spec.env.valueFrom.resourceFieldRef`](#obj-specenvvaluefromresourcefieldref)
+ * [`fn withContainerName(containerName)`](#fn-specenvvaluefromresourcefieldrefwithcontainername)
+ * [`fn withDivisor(divisor)`](#fn-specenvvaluefromresourcefieldrefwithdivisor)
+ * [`fn withResource(resource)`](#fn-specenvvaluefromresourcefieldrefwithresource)
+ * [`obj spec.env.valueFrom.secretKeyRef`](#obj-specenvvaluefromsecretkeyref)
+ * [`fn withKey(key)`](#fn-specenvvaluefromsecretkeyrefwithkey)
+ * [`fn withName(name)`](#fn-specenvvaluefromsecretkeyrefwithname)
+ * [`fn withOptional(optional)`](#fn-specenvvaluefromsecretkeyrefwithoptional)
+ * [`obj spec.envFrom`](#obj-specenvfrom)
+ * [`fn withPrefix(prefix)`](#fn-specenvfromwithprefix)
+ * [`obj spec.envFrom.configMapRef`](#obj-specenvfromconfigmapref)
+ * [`fn withName(name)`](#fn-specenvfromconfigmaprefwithname)
+ * [`fn withOptional(optional)`](#fn-specenvfromconfigmaprefwithoptional)
+ * [`obj spec.envFrom.secretRef`](#obj-specenvfromsecretref)
+ * [`fn withName(name)`](#fn-specenvfromsecretrefwithname)
+ * [`fn withOptional(optional)`](#fn-specenvfromsecretrefwithoptional)
+ * [`obj spec.ephemeralVolumesSizeLimit`](#obj-specephemeralvolumessizelimit)
+ * [`fn withShm(shm)`](#fn-specephemeralvolumessizelimitwithshm)
+ * [`fn withTemporaryData(temporaryData)`](#fn-specephemeralvolumessizelimitwithtemporarydata)
+ * [`obj spec.externalClusters`](#obj-specexternalclusters)
+ * [`fn withConnectionParameters(connectionParameters)`](#fn-specexternalclusterswithconnectionparameters)
+ * [`fn withConnectionParametersMixin(connectionParameters)`](#fn-specexternalclusterswithconnectionparametersmixin)
+ * [`fn withName(name)`](#fn-specexternalclusterswithname)
+ * [`obj spec.externalClusters.barmanObjectStore`](#obj-specexternalclustersbarmanobjectstore)
+ * [`fn withDestinationPath(destinationPath)`](#fn-specexternalclustersbarmanobjectstorewithdestinationpath)
+ * [`fn withEndpointURL(endpointURL)`](#fn-specexternalclustersbarmanobjectstorewithendpointurl)
+ * [`fn withHistoryTags(historyTags)`](#fn-specexternalclustersbarmanobjectstorewithhistorytags)
+ * [`fn withHistoryTagsMixin(historyTags)`](#fn-specexternalclustersbarmanobjectstorewithhistorytagsmixin)
+ * [`fn withServerName(serverName)`](#fn-specexternalclustersbarmanobjectstorewithservername)
+ * [`fn withTags(tags)`](#fn-specexternalclustersbarmanobjectstorewithtags)
+ * [`fn withTagsMixin(tags)`](#fn-specexternalclustersbarmanobjectstorewithtagsmixin)
+ * [`obj spec.externalClusters.barmanObjectStore.azureCredentials`](#obj-specexternalclustersbarmanobjectstoreazurecredentials)
+ * [`fn withInheritFromAzureAD(inheritFromAzureAD)`](#fn-specexternalclustersbarmanobjectstoreazurecredentialswithinheritfromazuread)
+ * [`obj spec.externalClusters.barmanObjectStore.azureCredentials.connectionString`](#obj-specexternalclustersbarmanobjectstoreazurecredentialsconnectionstring)
+ * [`fn withKey(key)`](#fn-specexternalclustersbarmanobjectstoreazurecredentialsconnectionstringwithkey)
+ * [`fn withName(name)`](#fn-specexternalclustersbarmanobjectstoreazurecredentialsconnectionstringwithname)
+ * [`obj spec.externalClusters.barmanObjectStore.azureCredentials.storageAccount`](#obj-specexternalclustersbarmanobjectstoreazurecredentialsstorageaccount)
+ * [`fn withKey(key)`](#fn-specexternalclustersbarmanobjectstoreazurecredentialsstorageaccountwithkey)
+ * [`fn withName(name)`](#fn-specexternalclustersbarmanobjectstoreazurecredentialsstorageaccountwithname)
+ * [`obj spec.externalClusters.barmanObjectStore.azureCredentials.storageKey`](#obj-specexternalclustersbarmanobjectstoreazurecredentialsstoragekey)
+ * [`fn withKey(key)`](#fn-specexternalclustersbarmanobjectstoreazurecredentialsstoragekeywithkey)
+ * [`fn withName(name)`](#fn-specexternalclustersbarmanobjectstoreazurecredentialsstoragekeywithname)
+ * [`obj spec.externalClusters.barmanObjectStore.azureCredentials.storageSasToken`](#obj-specexternalclustersbarmanobjectstoreazurecredentialsstoragesastoken)
+ * [`fn withKey(key)`](#fn-specexternalclustersbarmanobjectstoreazurecredentialsstoragesastokenwithkey)
+ * [`fn withName(name)`](#fn-specexternalclustersbarmanobjectstoreazurecredentialsstoragesastokenwithname)
+ * [`obj spec.externalClusters.barmanObjectStore.data`](#obj-specexternalclustersbarmanobjectstoredata)
+ * [`fn withCompression(compression)`](#fn-specexternalclustersbarmanobjectstoredatawithcompression)
+ * [`fn withEncryption(encryption)`](#fn-specexternalclustersbarmanobjectstoredatawithencryption)
+ * [`fn withImmediateCheckpoint(immediateCheckpoint)`](#fn-specexternalclustersbarmanobjectstoredatawithimmediatecheckpoint)
+ * [`fn withJobs(jobs)`](#fn-specexternalclustersbarmanobjectstoredatawithjobs)
+ * [`obj spec.externalClusters.barmanObjectStore.endpointCA`](#obj-specexternalclustersbarmanobjectstoreendpointca)
+ * [`fn withKey(key)`](#fn-specexternalclustersbarmanobjectstoreendpointcawithkey)
+ * [`fn withName(name)`](#fn-specexternalclustersbarmanobjectstoreendpointcawithname)
+ * [`obj spec.externalClusters.barmanObjectStore.googleCredentials`](#obj-specexternalclustersbarmanobjectstoregooglecredentials)
+ * [`fn withGkeEnvironment(gkeEnvironment)`](#fn-specexternalclustersbarmanobjectstoregooglecredentialswithgkeenvironment)
+ * [`obj spec.externalClusters.barmanObjectStore.googleCredentials.applicationCredentials`](#obj-specexternalclustersbarmanobjectstoregooglecredentialsapplicationcredentials)
+ * [`fn withKey(key)`](#fn-specexternalclustersbarmanobjectstoregooglecredentialsapplicationcredentialswithkey)
+ * [`fn withName(name)`](#fn-specexternalclustersbarmanobjectstoregooglecredentialsapplicationcredentialswithname)
+ * [`obj spec.externalClusters.barmanObjectStore.s3Credentials`](#obj-specexternalclustersbarmanobjectstores3credentials)
+ * [`fn withInheritFromIAMRole(inheritFromIAMRole)`](#fn-specexternalclustersbarmanobjectstores3credentialswithinheritfromiamrole)
+ * [`obj spec.externalClusters.barmanObjectStore.s3Credentials.accessKeyId`](#obj-specexternalclustersbarmanobjectstores3credentialsaccesskeyid)
+ * [`fn withKey(key)`](#fn-specexternalclustersbarmanobjectstores3credentialsaccesskeyidwithkey)
+ * [`fn withName(name)`](#fn-specexternalclustersbarmanobjectstores3credentialsaccesskeyidwithname)
+ * [`obj spec.externalClusters.barmanObjectStore.s3Credentials.region`](#obj-specexternalclustersbarmanobjectstores3credentialsregion)
+ * [`fn withKey(key)`](#fn-specexternalclustersbarmanobjectstores3credentialsregionwithkey)
+ * [`fn withName(name)`](#fn-specexternalclustersbarmanobjectstores3credentialsregionwithname)
+ * [`obj spec.externalClusters.barmanObjectStore.s3Credentials.secretAccessKey`](#obj-specexternalclustersbarmanobjectstores3credentialssecretaccesskey)
+ * [`fn withKey(key)`](#fn-specexternalclustersbarmanobjectstores3credentialssecretaccesskeywithkey)
+ * [`fn withName(name)`](#fn-specexternalclustersbarmanobjectstores3credentialssecretaccesskeywithname)
+ * [`obj spec.externalClusters.barmanObjectStore.s3Credentials.sessionToken`](#obj-specexternalclustersbarmanobjectstores3credentialssessiontoken)
+ * [`fn withKey(key)`](#fn-specexternalclustersbarmanobjectstores3credentialssessiontokenwithkey)
+ * [`fn withName(name)`](#fn-specexternalclustersbarmanobjectstores3credentialssessiontokenwithname)
+ * [`obj spec.externalClusters.barmanObjectStore.wal`](#obj-specexternalclustersbarmanobjectstorewal)
+ * [`fn withCompression(compression)`](#fn-specexternalclustersbarmanobjectstorewalwithcompression)
+ * [`fn withEncryption(encryption)`](#fn-specexternalclustersbarmanobjectstorewalwithencryption)
+ * [`fn withMaxParallel(maxParallel)`](#fn-specexternalclustersbarmanobjectstorewalwithmaxparallel)
+ * [`obj spec.externalClusters.password`](#obj-specexternalclusterspassword)
+ * [`fn withKey(key)`](#fn-specexternalclusterspasswordwithkey)
+ * [`fn withName(name)`](#fn-specexternalclusterspasswordwithname)
+ * [`fn withOptional(optional)`](#fn-specexternalclusterspasswordwithoptional)
+ * [`obj spec.externalClusters.sslCert`](#obj-specexternalclusterssslcert)
+ * [`fn withKey(key)`](#fn-specexternalclusterssslcertwithkey)
+ * [`fn withName(name)`](#fn-specexternalclusterssslcertwithname)
+ * [`fn withOptional(optional)`](#fn-specexternalclusterssslcertwithoptional)
+ * [`obj spec.externalClusters.sslKey`](#obj-specexternalclusterssslkey)
+ * [`fn withKey(key)`](#fn-specexternalclusterssslkeywithkey)
+ * [`fn withName(name)`](#fn-specexternalclusterssslkeywithname)
+ * [`fn withOptional(optional)`](#fn-specexternalclusterssslkeywithoptional)
+ * [`obj spec.externalClusters.sslRootCert`](#obj-specexternalclusterssslrootcert)
+ * [`fn withKey(key)`](#fn-specexternalclusterssslrootcertwithkey)
+ * [`fn withName(name)`](#fn-specexternalclusterssslrootcertwithname)
+ * [`fn withOptional(optional)`](#fn-specexternalclusterssslrootcertwithoptional)
+ * [`obj spec.imagePullSecrets`](#obj-specimagepullsecrets)
+ * [`fn withName(name)`](#fn-specimagepullsecretswithname)
+ * [`obj spec.inheritedMetadata`](#obj-specinheritedmetadata)
+ * [`fn withAnnotations(annotations)`](#fn-specinheritedmetadatawithannotations)
+ * [`fn withAnnotationsMixin(annotations)`](#fn-specinheritedmetadatawithannotationsmixin)
+ * [`fn withLabels(labels)`](#fn-specinheritedmetadatawithlabels)
+ * [`fn withLabelsMixin(labels)`](#fn-specinheritedmetadatawithlabelsmixin)
+ * [`obj spec.managed`](#obj-specmanaged)
+ * [`fn withRoles(roles)`](#fn-specmanagedwithroles)
+ * [`fn withRolesMixin(roles)`](#fn-specmanagedwithrolesmixin)
+ * [`obj spec.managed.roles`](#obj-specmanagedroles)
+ * [`fn withBypassrls(bypassrls)`](#fn-specmanagedroleswithbypassrls)
+ * [`fn withComment(comment)`](#fn-specmanagedroleswithcomment)
+ * [`fn withConnectionLimit(connectionLimit)`](#fn-specmanagedroleswithconnectionlimit)
+ * [`fn withCreatedb(createdb)`](#fn-specmanagedroleswithcreatedb)
+ * [`fn withCreaterole(createrole)`](#fn-specmanagedroleswithcreaterole)
+ * [`fn withDisablePassword(disablePassword)`](#fn-specmanagedroleswithdisablepassword)
+ * [`fn withEnsure(ensure)`](#fn-specmanagedroleswithensure)
+ * [`fn withInRoles(inRoles)`](#fn-specmanagedroleswithinroles)
+ * [`fn withInRolesMixin(inRoles)`](#fn-specmanagedroleswithinrolesmixin)
+ * [`fn withInherit(inherit)`](#fn-specmanagedroleswithinherit)
+ * [`fn withLogin(login)`](#fn-specmanagedroleswithlogin)
+ * [`fn withName(name)`](#fn-specmanagedroleswithname)
+ * [`fn withReplication(replication)`](#fn-specmanagedroleswithreplication)
+ * [`fn withSuperuser(superuser)`](#fn-specmanagedroleswithsuperuser)
+ * [`fn withValidUntil(validUntil)`](#fn-specmanagedroleswithvaliduntil)
+ * [`obj spec.managed.roles.passwordSecret`](#obj-specmanagedrolespasswordsecret)
+ * [`fn withName(name)`](#fn-specmanagedrolespasswordsecretwithname)
+ * [`obj spec.monitoring`](#obj-specmonitoring)
+ * [`fn withCustomQueriesConfigMap(customQueriesConfigMap)`](#fn-specmonitoringwithcustomqueriesconfigmap)
+ * [`fn withCustomQueriesConfigMapMixin(customQueriesConfigMap)`](#fn-specmonitoringwithcustomqueriesconfigmapmixin)
+ * [`fn withCustomQueriesSecret(customQueriesSecret)`](#fn-specmonitoringwithcustomqueriessecret)
+ * [`fn withCustomQueriesSecretMixin(customQueriesSecret)`](#fn-specmonitoringwithcustomqueriessecretmixin)
+ * [`fn withDisableDefaultQueries(disableDefaultQueries)`](#fn-specmonitoringwithdisabledefaultqueries)
+ * [`fn withEnablePodMonitor(enablePodMonitor)`](#fn-specmonitoringwithenablepodmonitor)
+ * [`obj spec.monitoring.customQueriesConfigMap`](#obj-specmonitoringcustomqueriesconfigmap)
+ * [`fn withKey(key)`](#fn-specmonitoringcustomqueriesconfigmapwithkey)
+ * [`fn withName(name)`](#fn-specmonitoringcustomqueriesconfigmapwithname)
+ * [`obj spec.monitoring.customQueriesSecret`](#obj-specmonitoringcustomqueriessecret)
+ * [`fn withKey(key)`](#fn-specmonitoringcustomqueriessecretwithkey)
+ * [`fn withName(name)`](#fn-specmonitoringcustomqueriessecretwithname)
+ * [`obj spec.nodeMaintenanceWindow`](#obj-specnodemaintenancewindow)
+ * [`fn withInProgress(inProgress)`](#fn-specnodemaintenancewindowwithinprogress)
+ * [`fn withReusePVC(reusePVC)`](#fn-specnodemaintenancewindowwithreusepvc)
+ * [`obj spec.postgresql`](#obj-specpostgresql)
+ * [`fn withParameters(parameters)`](#fn-specpostgresqlwithparameters)
+ * [`fn withParametersMixin(parameters)`](#fn-specpostgresqlwithparametersmixin)
+ * [`fn withPg_hba(pg_hba)`](#fn-specpostgresqlwithpg_hba)
+ * [`fn withPg_hbaMixin(pg_hba)`](#fn-specpostgresqlwithpg_hbamixin)
+ * [`fn withPromotionTimeout(promotionTimeout)`](#fn-specpostgresqlwithpromotiontimeout)
+ * [`fn withShared_preload_libraries(shared_preload_libraries)`](#fn-specpostgresqlwithshared_preload_libraries)
+ * [`fn withShared_preload_librariesMixin(shared_preload_libraries)`](#fn-specpostgresqlwithshared_preload_librariesmixin)
+ * [`obj spec.postgresql.ldap`](#obj-specpostgresqlldap)
+ * [`fn withPort(port)`](#fn-specpostgresqlldapwithport)
+ * [`fn withScheme(scheme)`](#fn-specpostgresqlldapwithscheme)
+ * [`fn withServer(server)`](#fn-specpostgresqlldapwithserver)
+ * [`fn withTls(tls)`](#fn-specpostgresqlldapwithtls)
+ * [`obj spec.postgresql.ldap.bindAsAuth`](#obj-specpostgresqlldapbindasauth)
+ * [`fn withPrefix(prefix)`](#fn-specpostgresqlldapbindasauthwithprefix)
+ * [`fn withSuffix(suffix)`](#fn-specpostgresqlldapbindasauthwithsuffix)
+ * [`obj spec.postgresql.ldap.bindSearchAuth`](#obj-specpostgresqlldapbindsearchauth)
+ * [`fn withBaseDN(baseDN)`](#fn-specpostgresqlldapbindsearchauthwithbasedn)
+ * [`fn withBindDN(bindDN)`](#fn-specpostgresqlldapbindsearchauthwithbinddn)
+ * [`fn withSearchAttribute(searchAttribute)`](#fn-specpostgresqlldapbindsearchauthwithsearchattribute)
+ * [`fn withSearchFilter(searchFilter)`](#fn-specpostgresqlldapbindsearchauthwithsearchfilter)
+ * [`obj spec.postgresql.ldap.bindSearchAuth.bindPassword`](#obj-specpostgresqlldapbindsearchauthbindpassword)
+ * [`fn withKey(key)`](#fn-specpostgresqlldapbindsearchauthbindpasswordwithkey)
+ * [`fn withName(name)`](#fn-specpostgresqlldapbindsearchauthbindpasswordwithname)
+ * [`fn withOptional(optional)`](#fn-specpostgresqlldapbindsearchauthbindpasswordwithoptional)
+ * [`obj spec.postgresql.syncReplicaElectionConstraint`](#obj-specpostgresqlsyncreplicaelectionconstraint)
+ * [`fn withEnabled(enabled)`](#fn-specpostgresqlsyncreplicaelectionconstraintwithenabled)
+ * [`fn withNodeLabelsAntiAffinity(nodeLabelsAntiAffinity)`](#fn-specpostgresqlsyncreplicaelectionconstraintwithnodelabelsantiaffinity)
+ * [`fn withNodeLabelsAntiAffinityMixin(nodeLabelsAntiAffinity)`](#fn-specpostgresqlsyncreplicaelectionconstraintwithnodelabelsantiaffinitymixin)
+ * [`obj spec.projectedVolumeTemplate`](#obj-specprojectedvolumetemplate)
+ * [`fn withDefaultMode(defaultMode)`](#fn-specprojectedvolumetemplatewithdefaultmode)
+ * [`fn withSources(sources)`](#fn-specprojectedvolumetemplatewithsources)
+ * [`fn withSourcesMixin(sources)`](#fn-specprojectedvolumetemplatewithsourcesmixin)
+ * [`obj spec.projectedVolumeTemplate.sources`](#obj-specprojectedvolumetemplatesources)
+ * [`obj spec.projectedVolumeTemplate.sources.configMap`](#obj-specprojectedvolumetemplatesourcesconfigmap)
+ * [`fn withItems(items)`](#fn-specprojectedvolumetemplatesourcesconfigmapwithitems)
+ * [`fn withItemsMixin(items)`](#fn-specprojectedvolumetemplatesourcesconfigmapwithitemsmixin)
+ * [`fn withName(name)`](#fn-specprojectedvolumetemplatesourcesconfigmapwithname)
+ * [`fn withOptional(optional)`](#fn-specprojectedvolumetemplatesourcesconfigmapwithoptional)
+ * [`obj spec.projectedVolumeTemplate.sources.configMap.items`](#obj-specprojectedvolumetemplatesourcesconfigmapitems)
+ * [`fn withKey(key)`](#fn-specprojectedvolumetemplatesourcesconfigmapitemswithkey)
+ * [`fn withMode(mode)`](#fn-specprojectedvolumetemplatesourcesconfigmapitemswithmode)
+ * [`fn withPath(path)`](#fn-specprojectedvolumetemplatesourcesconfigmapitemswithpath)
+ * [`obj spec.projectedVolumeTemplate.sources.downwardAPI`](#obj-specprojectedvolumetemplatesourcesdownwardapi)
+ * [`fn withItems(items)`](#fn-specprojectedvolumetemplatesourcesdownwardapiwithitems)
+ * [`fn withItemsMixin(items)`](#fn-specprojectedvolumetemplatesourcesdownwardapiwithitemsmixin)
+ * [`obj spec.projectedVolumeTemplate.sources.downwardAPI.items`](#obj-specprojectedvolumetemplatesourcesdownwardapiitems)
+ * [`fn withMode(mode)`](#fn-specprojectedvolumetemplatesourcesdownwardapiitemswithmode)
+ * [`fn withPath(path)`](#fn-specprojectedvolumetemplatesourcesdownwardapiitemswithpath)
+ * [`obj spec.projectedVolumeTemplate.sources.downwardAPI.items.fieldRef`](#obj-specprojectedvolumetemplatesourcesdownwardapiitemsfieldref)
+ * [`fn withApiVersion(apiVersion)`](#fn-specprojectedvolumetemplatesourcesdownwardapiitemsfieldrefwithapiversion)
+ * [`fn withFieldPath(fieldPath)`](#fn-specprojectedvolumetemplatesourcesdownwardapiitemsfieldrefwithfieldpath)
+ * [`obj spec.projectedVolumeTemplate.sources.downwardAPI.items.resourceFieldRef`](#obj-specprojectedvolumetemplatesourcesdownwardapiitemsresourcefieldref)
+ * [`fn withContainerName(containerName)`](#fn-specprojectedvolumetemplatesourcesdownwardapiitemsresourcefieldrefwithcontainername)
+ * [`fn withDivisor(divisor)`](#fn-specprojectedvolumetemplatesourcesdownwardapiitemsresourcefieldrefwithdivisor)
+ * [`fn withResource(resource)`](#fn-specprojectedvolumetemplatesourcesdownwardapiitemsresourcefieldrefwithresource)
+ * [`obj spec.projectedVolumeTemplate.sources.secret`](#obj-specprojectedvolumetemplatesourcessecret)
+ * [`fn withItems(items)`](#fn-specprojectedvolumetemplatesourcessecretwithitems)
+ * [`fn withItemsMixin(items)`](#fn-specprojectedvolumetemplatesourcessecretwithitemsmixin)
+ * [`fn withName(name)`](#fn-specprojectedvolumetemplatesourcessecretwithname)
+ * [`fn withOptional(optional)`](#fn-specprojectedvolumetemplatesourcessecretwithoptional)
+ * [`obj spec.projectedVolumeTemplate.sources.secret.items`](#obj-specprojectedvolumetemplatesourcessecretitems)
+ * [`fn withKey(key)`](#fn-specprojectedvolumetemplatesourcessecretitemswithkey)
+ * [`fn withMode(mode)`](#fn-specprojectedvolumetemplatesourcessecretitemswithmode)
+ * [`fn withPath(path)`](#fn-specprojectedvolumetemplatesourcessecretitemswithpath)
+ * [`obj spec.projectedVolumeTemplate.sources.serviceAccountToken`](#obj-specprojectedvolumetemplatesourcesserviceaccounttoken)
+ * [`fn withAudience(audience)`](#fn-specprojectedvolumetemplatesourcesserviceaccounttokenwithaudience)
+ * [`fn withExpirationSeconds(expirationSeconds)`](#fn-specprojectedvolumetemplatesourcesserviceaccounttokenwithexpirationseconds)
+ * [`fn withPath(path)`](#fn-specprojectedvolumetemplatesourcesserviceaccounttokenwithpath)
+ * [`obj spec.replica`](#obj-specreplica)
+ * [`fn withEnabled(enabled)`](#fn-specreplicawithenabled)
+ * [`fn withSource(source)`](#fn-specreplicawithsource)
+ * [`obj spec.replicationSlots`](#obj-specreplicationslots)
+ * [`fn withUpdateInterval(updateInterval)`](#fn-specreplicationslotswithupdateinterval)
+ * [`obj spec.replicationSlots.highAvailability`](#obj-specreplicationslotshighavailability)
+ * [`fn withEnabled(enabled)`](#fn-specreplicationslotshighavailabilitywithenabled)
+ * [`fn withSlotPrefix(slotPrefix)`](#fn-specreplicationslotshighavailabilitywithslotprefix)
+ * [`obj spec.resources`](#obj-specresources)
+ * [`fn withClaims(claims)`](#fn-specresourceswithclaims)
+ * [`fn withClaimsMixin(claims)`](#fn-specresourceswithclaimsmixin)
+ * [`fn withLimits(limits)`](#fn-specresourceswithlimits)
+ * [`fn withLimitsMixin(limits)`](#fn-specresourceswithlimitsmixin)
+ * [`fn withRequests(requests)`](#fn-specresourceswithrequests)
+ * [`fn withRequestsMixin(requests)`](#fn-specresourceswithrequestsmixin)
+ * [`obj spec.resources.claims`](#obj-specresourcesclaims)
+ * [`fn withName(name)`](#fn-specresourcesclaimswithname)
+ * [`obj spec.seccompProfile`](#obj-specseccompprofile)
+ * [`fn withLocalhostProfile(localhostProfile)`](#fn-specseccompprofilewithlocalhostprofile)
+ * [`fn withType(type)`](#fn-specseccompprofilewithtype)
+ * [`obj spec.serviceAccountTemplate`](#obj-specserviceaccounttemplate)
+ * [`obj spec.serviceAccountTemplate.metadata`](#obj-specserviceaccounttemplatemetadata)
+ * [`fn withAnnotations(annotations)`](#fn-specserviceaccounttemplatemetadatawithannotations)
+ * [`fn withAnnotationsMixin(annotations)`](#fn-specserviceaccounttemplatemetadatawithannotationsmixin)
+ * [`fn withLabels(labels)`](#fn-specserviceaccounttemplatemetadatawithlabels)
+ * [`fn withLabelsMixin(labels)`](#fn-specserviceaccounttemplatemetadatawithlabelsmixin)
+ * [`obj spec.storage`](#obj-specstorage)
+ * [`fn withResizeInUseVolumes(resizeInUseVolumes)`](#fn-specstoragewithresizeinusevolumes)
+ * [`fn withSize(size)`](#fn-specstoragewithsize)
+ * [`fn withStorageClass(storageClass)`](#fn-specstoragewithstorageclass)
+ * [`obj spec.storage.pvcTemplate`](#obj-specstoragepvctemplate)
+ * [`fn withAccessModes(accessModes)`](#fn-specstoragepvctemplatewithaccessmodes)
+ * [`fn withAccessModesMixin(accessModes)`](#fn-specstoragepvctemplatewithaccessmodesmixin)
+ * [`fn withStorageClassName(storageClassName)`](#fn-specstoragepvctemplatewithstorageclassname)
+ * [`fn withVolumeMode(volumeMode)`](#fn-specstoragepvctemplatewithvolumemode)
+ * [`fn withVolumeName(volumeName)`](#fn-specstoragepvctemplatewithvolumename)
+ * [`obj spec.storage.pvcTemplate.dataSource`](#obj-specstoragepvctemplatedatasource)
+ * [`fn withApiGroup(apiGroup)`](#fn-specstoragepvctemplatedatasourcewithapigroup)
+ * [`fn withKind(kind)`](#fn-specstoragepvctemplatedatasourcewithkind)
+ * [`fn withName(name)`](#fn-specstoragepvctemplatedatasourcewithname)
+ * [`obj spec.storage.pvcTemplate.dataSourceRef`](#obj-specstoragepvctemplatedatasourceref)
+ * [`fn withApiGroup(apiGroup)`](#fn-specstoragepvctemplatedatasourcerefwithapigroup)
+ * [`fn withKind(kind)`](#fn-specstoragepvctemplatedatasourcerefwithkind)
+ * [`fn withName(name)`](#fn-specstoragepvctemplatedatasourcerefwithname)
+ * [`fn withNamespace(namespace)`](#fn-specstoragepvctemplatedatasourcerefwithnamespace)
+ * [`obj spec.storage.pvcTemplate.resources`](#obj-specstoragepvctemplateresources)
+ * [`fn withClaims(claims)`](#fn-specstoragepvctemplateresourceswithclaims)
+ * [`fn withClaimsMixin(claims)`](#fn-specstoragepvctemplateresourceswithclaimsmixin)
+ * [`fn withLimits(limits)`](#fn-specstoragepvctemplateresourceswithlimits)
+ * [`fn withLimitsMixin(limits)`](#fn-specstoragepvctemplateresourceswithlimitsmixin)
+ * [`fn withRequests(requests)`](#fn-specstoragepvctemplateresourceswithrequests)
+ * [`fn withRequestsMixin(requests)`](#fn-specstoragepvctemplateresourceswithrequestsmixin)
+ * [`obj spec.storage.pvcTemplate.resources.claims`](#obj-specstoragepvctemplateresourcesclaims)
+ * [`fn withName(name)`](#fn-specstoragepvctemplateresourcesclaimswithname)
+ * [`obj spec.storage.pvcTemplate.selector`](#obj-specstoragepvctemplateselector)
+ * [`fn withMatchExpressions(matchExpressions)`](#fn-specstoragepvctemplateselectorwithmatchexpressions)
+ * [`fn withMatchExpressionsMixin(matchExpressions)`](#fn-specstoragepvctemplateselectorwithmatchexpressionsmixin)
+ * [`fn withMatchLabels(matchLabels)`](#fn-specstoragepvctemplateselectorwithmatchlabels)
+ * [`fn withMatchLabelsMixin(matchLabels)`](#fn-specstoragepvctemplateselectorwithmatchlabelsmixin)
+ * [`obj spec.storage.pvcTemplate.selector.matchExpressions`](#obj-specstoragepvctemplateselectormatchexpressions)
+ * [`fn withKey(key)`](#fn-specstoragepvctemplateselectormatchexpressionswithkey)
+ * [`fn withOperator(operator)`](#fn-specstoragepvctemplateselectormatchexpressionswithoperator)
+ * [`fn withValues(values)`](#fn-specstoragepvctemplateselectormatchexpressionswithvalues)
+ * [`fn withValuesMixin(values)`](#fn-specstoragepvctemplateselectormatchexpressionswithvaluesmixin)
+ * [`obj spec.superuserSecret`](#obj-specsuperusersecret)
+ * [`fn withName(name)`](#fn-specsuperusersecretwithname)
+ * [`obj spec.topologySpreadConstraints`](#obj-spectopologyspreadconstraints)
+ * [`fn withMatchLabelKeys(matchLabelKeys)`](#fn-spectopologyspreadconstraintswithmatchlabelkeys)
+ * [`fn withMatchLabelKeysMixin(matchLabelKeys)`](#fn-spectopologyspreadconstraintswithmatchlabelkeysmixin)
+ * [`fn withMaxSkew(maxSkew)`](#fn-spectopologyspreadconstraintswithmaxskew)
+ * [`fn withMinDomains(minDomains)`](#fn-spectopologyspreadconstraintswithmindomains)
+ * [`fn withNodeAffinityPolicy(nodeAffinityPolicy)`](#fn-spectopologyspreadconstraintswithnodeaffinitypolicy)
+ * [`fn withNodeTaintsPolicy(nodeTaintsPolicy)`](#fn-spectopologyspreadconstraintswithnodetaintspolicy)
+ * [`fn withTopologyKey(topologyKey)`](#fn-spectopologyspreadconstraintswithtopologykey)
+ * [`fn withWhenUnsatisfiable(whenUnsatisfiable)`](#fn-spectopologyspreadconstraintswithwhenunsatisfiable)
+ * [`obj spec.topologySpreadConstraints.labelSelector`](#obj-spectopologyspreadconstraintslabelselector)
+ * [`fn withMatchExpressions(matchExpressions)`](#fn-spectopologyspreadconstraintslabelselectorwithmatchexpressions)
+ * [`fn withMatchExpressionsMixin(matchExpressions)`](#fn-spectopologyspreadconstraintslabelselectorwithmatchexpressionsmixin)
+ * [`fn withMatchLabels(matchLabels)`](#fn-spectopologyspreadconstraintslabelselectorwithmatchlabels)
+ * [`fn withMatchLabelsMixin(matchLabels)`](#fn-spectopologyspreadconstraintslabelselectorwithmatchlabelsmixin)
+ * [`obj spec.topologySpreadConstraints.labelSelector.matchExpressions`](#obj-spectopologyspreadconstraintslabelselectormatchexpressions)
+ * [`fn withKey(key)`](#fn-spectopologyspreadconstraintslabelselectormatchexpressionswithkey)
+ * [`fn withOperator(operator)`](#fn-spectopologyspreadconstraintslabelselectormatchexpressionswithoperator)
+ * [`fn withValues(values)`](#fn-spectopologyspreadconstraintslabelselectormatchexpressionswithvalues)
+ * [`fn withValuesMixin(values)`](#fn-spectopologyspreadconstraintslabelselectormatchexpressionswithvaluesmixin)
+ * [`obj spec.walStorage`](#obj-specwalstorage)
+ * [`fn withResizeInUseVolumes(resizeInUseVolumes)`](#fn-specwalstoragewithresizeinusevolumes)
+ * [`fn withSize(size)`](#fn-specwalstoragewithsize)
+ * [`fn withStorageClass(storageClass)`](#fn-specwalstoragewithstorageclass)
+ * [`obj spec.walStorage.pvcTemplate`](#obj-specwalstoragepvctemplate)
+ * [`fn withAccessModes(accessModes)`](#fn-specwalstoragepvctemplatewithaccessmodes)
+ * [`fn withAccessModesMixin(accessModes)`](#fn-specwalstoragepvctemplatewithaccessmodesmixin)
+ * [`fn withStorageClassName(storageClassName)`](#fn-specwalstoragepvctemplatewithstorageclassname)
+ * [`fn withVolumeMode(volumeMode)`](#fn-specwalstoragepvctemplatewithvolumemode)
+ * [`fn withVolumeName(volumeName)`](#fn-specwalstoragepvctemplatewithvolumename)
+ * [`obj spec.walStorage.pvcTemplate.dataSource`](#obj-specwalstoragepvctemplatedatasource)
+ * [`fn withApiGroup(apiGroup)`](#fn-specwalstoragepvctemplatedatasourcewithapigroup)
+ * [`fn withKind(kind)`](#fn-specwalstoragepvctemplatedatasourcewithkind)
+ * [`fn withName(name)`](#fn-specwalstoragepvctemplatedatasourcewithname)
+ * [`obj spec.walStorage.pvcTemplate.dataSourceRef`](#obj-specwalstoragepvctemplatedatasourceref)
+ * [`fn withApiGroup(apiGroup)`](#fn-specwalstoragepvctemplatedatasourcerefwithapigroup)
+ * [`fn withKind(kind)`](#fn-specwalstoragepvctemplatedatasourcerefwithkind)
+ * [`fn withName(name)`](#fn-specwalstoragepvctemplatedatasourcerefwithname)
+ * [`fn withNamespace(namespace)`](#fn-specwalstoragepvctemplatedatasourcerefwithnamespace)
+ * [`obj spec.walStorage.pvcTemplate.resources`](#obj-specwalstoragepvctemplateresources)
+ * [`fn withClaims(claims)`](#fn-specwalstoragepvctemplateresourceswithclaims)
+ * [`fn withClaimsMixin(claims)`](#fn-specwalstoragepvctemplateresourceswithclaimsmixin)
+ * [`fn withLimits(limits)`](#fn-specwalstoragepvctemplateresourceswithlimits)
+ * [`fn withLimitsMixin(limits)`](#fn-specwalstoragepvctemplateresourceswithlimitsmixin)
+ * [`fn withRequests(requests)`](#fn-specwalstoragepvctemplateresourceswithrequests)
+ * [`fn withRequestsMixin(requests)`](#fn-specwalstoragepvctemplateresourceswithrequestsmixin)
+ * [`obj spec.walStorage.pvcTemplate.resources.claims`](#obj-specwalstoragepvctemplateresourcesclaims)
+ * [`fn withName(name)`](#fn-specwalstoragepvctemplateresourcesclaimswithname)
+ * [`obj spec.walStorage.pvcTemplate.selector`](#obj-specwalstoragepvctemplateselector)
+ * [`fn withMatchExpressions(matchExpressions)`](#fn-specwalstoragepvctemplateselectorwithmatchexpressions)
+ * [`fn withMatchExpressionsMixin(matchExpressions)`](#fn-specwalstoragepvctemplateselectorwithmatchexpressionsmixin)
+ * [`fn withMatchLabels(matchLabels)`](#fn-specwalstoragepvctemplateselectorwithmatchlabels)
+ * [`fn withMatchLabelsMixin(matchLabels)`](#fn-specwalstoragepvctemplateselectorwithmatchlabelsmixin)
+ * [`obj spec.walStorage.pvcTemplate.selector.matchExpressions`](#obj-specwalstoragepvctemplateselectormatchexpressions)
+ * [`fn withKey(key)`](#fn-specwalstoragepvctemplateselectormatchexpressionswithkey)
+ * [`fn withOperator(operator)`](#fn-specwalstoragepvctemplateselectormatchexpressionswithoperator)
+ * [`fn withValues(values)`](#fn-specwalstoragepvctemplateselectormatchexpressionswithvalues)
+ * [`fn withValuesMixin(values)`](#fn-specwalstoragepvctemplateselectormatchexpressionswithvaluesmixin)
+
+## Fields
+
+### fn new
+
+```ts
+new(name)
+```
+
+new returns an instance of Cluster
+
+## obj metadata
+
+"ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create."
+
+### fn metadata.withAnnotations
+
+```ts
+withAnnotations(annotations)
+```
+
+"Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations"
+
+### fn metadata.withAnnotationsMixin
+
+```ts
+withAnnotationsMixin(annotations)
+```
+
+"Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations"
+
+**Note:** This function appends passed data to existing values
+
+### fn metadata.withClusterName
+
+```ts
+withClusterName(clusterName)
+```
+
+"The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request."
+
+### fn metadata.withCreationTimestamp
+
+```ts
+withCreationTimestamp(creationTimestamp)
+```
+
+"Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers."
+
+### fn metadata.withDeletionGracePeriodSeconds
+
+```ts
+withDeletionGracePeriodSeconds(deletionGracePeriodSeconds)
+```
+
+"Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only."
+
+### fn metadata.withDeletionTimestamp
+
+```ts
+withDeletionTimestamp(deletionTimestamp)
+```
+
+"Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers."
+
+### fn metadata.withFinalizers
+
+```ts
+withFinalizers(finalizers)
+```
+
+"Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list."
+
+### fn metadata.withFinalizersMixin
+
+```ts
+withFinalizersMixin(finalizers)
+```
+
+"Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list."
+
+**Note:** This function appends passed data to existing values
+
+### fn metadata.withGenerateName
+
+```ts
+withGenerateName(generateName)
+```
+
+"GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency"
+
+### fn metadata.withGeneration
+
+```ts
+withGeneration(generation)
+```
+
+"A sequence number representing a specific generation of the desired state. Populated by the system. Read-only."
+
+### fn metadata.withLabels
+
+```ts
+withLabels(labels)
+```
+
+"Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels"
+
+### fn metadata.withLabelsMixin
+
+```ts
+withLabelsMixin(labels)
+```
+
+"Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels"
+
+**Note:** This function appends passed data to existing values
+
+### fn metadata.withName
+
+```ts
+withName(name)
+```
+
+"Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names"
+
+### fn metadata.withNamespace
+
+```ts
+withNamespace(namespace)
+```
+
+"Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the \"default\" namespace, but \"default\" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces"
+
+### fn metadata.withOwnerReferences
+
+```ts
+withOwnerReferences(ownerReferences)
+```
+
+"List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller."
+
+### fn metadata.withOwnerReferencesMixin
+
+```ts
+withOwnerReferencesMixin(ownerReferences)
+```
+
+"List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller."
+
+**Note:** This function appends passed data to existing values
+
+### fn metadata.withResourceVersion
+
+```ts
+withResourceVersion(resourceVersion)
+```
+
+"An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency"
+
+### fn metadata.withSelfLink
+
+```ts
+withSelfLink(selfLink)
+```
+
+"SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release."
+
+### fn metadata.withUid
+
+```ts
+withUid(uid)
+```
+
+"UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids"
+
+## obj spec
+
+"Specification of the desired behavior of the cluster. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#spec-and-status"
+
+### fn spec.withDescription
+
+```ts
+withDescription(description)
+```
+
+"Description of this PostgreSQL cluster"
+
+### fn spec.withEnableSuperuserAccess
+
+```ts
+withEnableSuperuserAccess(enableSuperuserAccess)
+```
+
+"When this option is enabled, the operator will use the `SuperuserSecret` to update the `postgres` user password (if the secret is not present, the operator will automatically create one). When this option is disabled, the operator will ignore the `SuperuserSecret` content, delete it when automatically created, and then blank the password of the `postgres` user by setting it to `NULL`. Disabled by default."
+
+### fn spec.withEnv
+
+```ts
+withEnv(env)
+```
+
+"Env follows the Env format to pass environment variables to the pods created in the cluster"
+
+### fn spec.withEnvFrom
+
+```ts
+withEnvFrom(envFrom)
+```
+
+"EnvFrom follows the EnvFrom format to pass environment variables sources to the pods to be used by Env"
+
+### fn spec.withEnvFromMixin
+
+```ts
+withEnvFromMixin(envFrom)
+```
+
+"EnvFrom follows the EnvFrom format to pass environment variables sources to the pods to be used by Env"
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.withEnvMixin
+
+```ts
+withEnvMixin(env)
+```
+
+"Env follows the Env format to pass environment variables to the pods created in the cluster"
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.withExternalClusters
+
+```ts
+withExternalClusters(externalClusters)
+```
+
+"The list of external clusters which are used in the configuration"
+
+### fn spec.withExternalClustersMixin
+
+```ts
+withExternalClustersMixin(externalClusters)
+```
+
+"The list of external clusters which are used in the configuration"
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.withFailoverDelay
+
+```ts
+withFailoverDelay(failoverDelay)
+```
+
+"The amount of time (in seconds) to wait before triggering a failover after the primary PostgreSQL instance in the cluster was detected to be unhealthy"
+
+### fn spec.withImageName
+
+```ts
+withImageName(imageName)
+```
+
+"Name of the container image, supporting both tags (`:`) and digests for deterministic and repeatable deployments (`:@sha256:`)"
+
+### fn spec.withImagePullPolicy
+
+```ts
+withImagePullPolicy(imagePullPolicy)
+```
+
+"Image pull policy. One of `Always`, `Never` or `IfNotPresent`. If not defined, it defaults to `IfNotPresent`. Cannot be updated. More info: https://kubernetes.io/docs/concepts/containers/images#updating-images"
+
+### fn spec.withImagePullSecrets
+
+```ts
+withImagePullSecrets(imagePullSecrets)
+```
+
+"The list of pull secrets to be used to pull the images"
+
+### fn spec.withImagePullSecretsMixin
+
+```ts
+withImagePullSecretsMixin(imagePullSecrets)
+```
+
+"The list of pull secrets to be used to pull the images"
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.withInstances
+
+```ts
+withInstances(instances)
+```
+
+"Number of instances required in the cluster"
+
+### fn spec.withLogLevel
+
+```ts
+withLogLevel(logLevel)
+```
+
+"The instances' log level, one of the following values: error, warning, info (default), debug, trace"
+
+### fn spec.withMaxSyncReplicas
+
+```ts
+withMaxSyncReplicas(maxSyncReplicas)
+```
+
+"The target value for the synchronous replication quorum, that can be decreased if the number of ready standbys is lower than this. Undefined or 0 disable synchronous replication."
+
+### fn spec.withMinSyncReplicas
+
+```ts
+withMinSyncReplicas(minSyncReplicas)
+```
+
+"Minimum number of instances required in synchronous replication with the primary. Undefined or 0 allow writes to complete when no standby is available."
+
+### fn spec.withPostgresGID
+
+```ts
+withPostgresGID(postgresGID)
+```
+
+"The GID of the `postgres` user inside the image, defaults to `26`"
+
+### fn spec.withPostgresUID
+
+```ts
+withPostgresUID(postgresUID)
+```
+
+"The UID of the `postgres` user inside the image, defaults to `26`"
+
+### fn spec.withPrimaryUpdateMethod
+
+```ts
+withPrimaryUpdateMethod(primaryUpdateMethod)
+```
+
+"Method to follow to upgrade the primary server during a rolling update procedure, after all replicas have been successfully updated: it can be with a switchover (`switchover`) or in-place (`restart` - default)"
+
+### fn spec.withPrimaryUpdateStrategy
+
+```ts
+withPrimaryUpdateStrategy(primaryUpdateStrategy)
+```
+
+"Deployment strategy to follow to upgrade the primary server during a rolling update procedure, after all replicas have been successfully updated: it can be automated (`unsupervised` - default) or manual (`supervised`)"
+
+### fn spec.withPriorityClassName
+
+```ts
+withPriorityClassName(priorityClassName)
+```
+
+"Name of the priority class which will be used in every generated Pod, if the PriorityClass specified does not exist, the pod will not be able to schedule. Please refer to https://kubernetes.io/docs/concepts/scheduling-eviction/pod-priority-preemption/#priorityclass for more information"
+
+### fn spec.withSchedulerName
+
+```ts
+withSchedulerName(schedulerName)
+```
+
+"If specified, the pod will be dispatched by specified Kubernetes scheduler. If not specified, the pod will be dispatched by the default scheduler. More info: https://kubernetes.io/docs/concepts/scheduling-eviction/kube-scheduler/"
+
+### fn spec.withSmartShutdownTimeout
+
+```ts
+withSmartShutdownTimeout(smartShutdownTimeout)
+```
+
+"The time in seconds that controls the window of time reserved for the smart shutdown of Postgres to complete. Make sure you reserve enough time for the operator to request a fast shutdown of Postgres (that is: `stopDelay` - `smartShutdownTimeout`)."
+
+### fn spec.withStartDelay
+
+```ts
+withStartDelay(startDelay)
+```
+
+"The time in seconds that is allowed for a PostgreSQL instance to successfully start up (default 3600). The startup probe failure threshold is derived from this value using the formula: ceiling(startDelay / 10)."
+
+### fn spec.withStopDelay
+
+```ts
+withStopDelay(stopDelay)
+```
+
+"The time in seconds that is allowed for a PostgreSQL instance to gracefully shutdown (default 1800)"
+
+### fn spec.withSwitchoverDelay
+
+```ts
+withSwitchoverDelay(switchoverDelay)
+```
+
+"The time in seconds that is allowed for a primary PostgreSQL instance to gracefully shutdown during a switchover. Default value is 3600 seconds (1 hour)."
+
+### fn spec.withTopologySpreadConstraints
+
+```ts
+withTopologySpreadConstraints(topologySpreadConstraints)
+```
+
+"TopologySpreadConstraints specifies how to spread matching pods among the given topology. More info: https://kubernetes.io/docs/concepts/scheduling-eviction/topology-spread-constraints/"
+
+### fn spec.withTopologySpreadConstraintsMixin
+
+```ts
+withTopologySpreadConstraintsMixin(topologySpreadConstraints)
+```
+
+"TopologySpreadConstraints specifies how to spread matching pods among the given topology. More info: https://kubernetes.io/docs/concepts/scheduling-eviction/topology-spread-constraints/"
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.affinity
+
+"Affinity/Anti-affinity rules for Pods"
+
+### fn spec.affinity.withEnablePodAntiAffinity
+
+```ts
+withEnablePodAntiAffinity(enablePodAntiAffinity)
+```
+
+"Activates anti-affinity for the pods. The operator will define pods anti-affinity unless this field is explicitly set to false"
+
+### fn spec.affinity.withNodeSelector
+
+```ts
+withNodeSelector(nodeSelector)
+```
+
+"NodeSelector is map of key-value pairs used to define the nodes on which the pods can run. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/"
+
+### fn spec.affinity.withNodeSelectorMixin
+
+```ts
+withNodeSelectorMixin(nodeSelector)
+```
+
+"NodeSelector is map of key-value pairs used to define the nodes on which the pods can run. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/"
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.affinity.withPodAntiAffinityType
+
+```ts
+withPodAntiAffinityType(podAntiAffinityType)
+```
+
+"PodAntiAffinityType allows the user to decide whether pod anti-affinity between cluster instance has to be considered a strong requirement during scheduling or not. Allowed values are: \"preferred\" (default if empty) or \"required\". Setting it to \"required\", could lead to instances remaining pending until new kubernetes nodes are added if all the existing nodes don't match the required pod anti-affinity rule. More info: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#inter-pod-affinity-and-anti-affinity"
+
+### fn spec.affinity.withTolerations
+
+```ts
+withTolerations(tolerations)
+```
+
+"Tolerations is a list of Tolerations that should be set for all the pods, in order to allow them to run on tainted nodes. More info: https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/"
+
+### fn spec.affinity.withTolerationsMixin
+
+```ts
+withTolerationsMixin(tolerations)
+```
+
+"Tolerations is a list of Tolerations that should be set for all the pods, in order to allow them to run on tainted nodes. More info: https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/"
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.affinity.withTopologyKey
+
+```ts
+withTopologyKey(topologyKey)
+```
+
+"TopologyKey to use for anti-affinity configuration. See k8s documentation for more info on that"
+
+## obj spec.affinity.additionalPodAffinity
+
+"AdditionalPodAffinity allows to specify pod affinity terms to be passed to all the cluster's pods."
+
+### fn spec.affinity.additionalPodAffinity.withPreferredDuringSchedulingIgnoredDuringExecution
+
+```ts
+withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution)
+```
+
+"The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding \"weight\" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred."
+
+### fn spec.affinity.additionalPodAffinity.withPreferredDuringSchedulingIgnoredDuringExecutionMixin
+
+```ts
+withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution)
+```
+
+"The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding \"weight\" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.affinity.additionalPodAffinity.withRequiredDuringSchedulingIgnoredDuringExecution
+
+```ts
+withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution)
+```
+
+"If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied."
+
+### fn spec.affinity.additionalPodAffinity.withRequiredDuringSchedulingIgnoredDuringExecutionMixin
+
+```ts
+withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution)
+```
+
+"If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied."
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.affinity.additionalPodAffinity.preferredDuringSchedulingIgnoredDuringExecution
+
+"The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding \"weight\" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred."
+
+### fn spec.affinity.additionalPodAffinity.preferredDuringSchedulingIgnoredDuringExecution.withWeight
+
+```ts
+withWeight(weight)
+```
+
+"weight associated with matching the corresponding podAffinityTerm, in the range 1-100."
+
+## obj spec.affinity.additionalPodAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm
+
+"Required. A pod affinity term, associated with the corresponding weight."
+
+### fn spec.affinity.additionalPodAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.withNamespaces
+
+```ts
+withNamespaces(namespaces)
+```
+
+"namespaces specifies a static list of namespace names that the term applies to. The term is applied to the union of the namespaces listed in this field and the ones selected by namespaceSelector. null or empty namespaces list and null namespaceSelector means \"this pod's namespace\"."
+
+### fn spec.affinity.additionalPodAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.withNamespacesMixin
+
+```ts
+withNamespacesMixin(namespaces)
+```
+
+"namespaces specifies a static list of namespace names that the term applies to. The term is applied to the union of the namespaces listed in this field and the ones selected by namespaceSelector. null or empty namespaces list and null namespaceSelector means \"this pod's namespace\"."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.affinity.additionalPodAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.withTopologyKey
+
+```ts
+withTopologyKey(topologyKey)
+```
+
+"This pod should be co-located (affinity) or not co-located (anti-affinity) with the pods matching the labelSelector in the specified namespaces, where co-located is defined as running on a node whose value of the label with key topologyKey matches that of any node on which any of the selected pods is running. Empty topologyKey is not allowed."
+
+## obj spec.affinity.additionalPodAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.labelSelector
+
+"A label query over a set of resources, in this case pods."
+
+### fn spec.affinity.additionalPodAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.labelSelector.withMatchExpressions
+
+```ts
+withMatchExpressions(matchExpressions)
+```
+
+"matchExpressions is a list of label selector requirements. The requirements are ANDed."
+
+### fn spec.affinity.additionalPodAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.labelSelector.withMatchExpressionsMixin
+
+```ts
+withMatchExpressionsMixin(matchExpressions)
+```
+
+"matchExpressions is a list of label selector requirements. The requirements are ANDed."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.affinity.additionalPodAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.labelSelector.withMatchLabels
+
+```ts
+withMatchLabels(matchLabels)
+```
+
+"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \"key\", the operator is \"In\", and the values array contains only \"value\". The requirements are ANDed."
+
+### fn spec.affinity.additionalPodAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.labelSelector.withMatchLabelsMixin
+
+```ts
+withMatchLabelsMixin(matchLabels)
+```
+
+"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \"key\", the operator is \"In\", and the values array contains only \"value\". The requirements are ANDed."
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.affinity.additionalPodAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.labelSelector.matchExpressions
+
+"matchExpressions is a list of label selector requirements. The requirements are ANDed."
+
+### fn spec.affinity.additionalPodAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.labelSelector.matchExpressions.withKey
+
+```ts
+withKey(key)
+```
+
+"key is the label key that the selector applies to."
+
+### fn spec.affinity.additionalPodAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.labelSelector.matchExpressions.withOperator
+
+```ts
+withOperator(operator)
+```
+
+"operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist."
+
+### fn spec.affinity.additionalPodAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.labelSelector.matchExpressions.withValues
+
+```ts
+withValues(values)
+```
+
+"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."
+
+### fn spec.affinity.additionalPodAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.labelSelector.matchExpressions.withValuesMixin
+
+```ts
+withValuesMixin(values)
+```
+
+"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.affinity.additionalPodAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.namespaceSelector
+
+"A label query over the set of namespaces that the term applies to. The term is applied to the union of the namespaces selected by this field and the ones listed in the namespaces field. null selector and null or empty namespaces list means \"this pod's namespace\". An empty selector ({}) matches all namespaces."
+
+### fn spec.affinity.additionalPodAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.namespaceSelector.withMatchExpressions
+
+```ts
+withMatchExpressions(matchExpressions)
+```
+
+"matchExpressions is a list of label selector requirements. The requirements are ANDed."
+
+### fn spec.affinity.additionalPodAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.namespaceSelector.withMatchExpressionsMixin
+
+```ts
+withMatchExpressionsMixin(matchExpressions)
+```
+
+"matchExpressions is a list of label selector requirements. The requirements are ANDed."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.affinity.additionalPodAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.namespaceSelector.withMatchLabels
+
+```ts
+withMatchLabels(matchLabels)
+```
+
+"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \"key\", the operator is \"In\", and the values array contains only \"value\". The requirements are ANDed."
+
+### fn spec.affinity.additionalPodAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.namespaceSelector.withMatchLabelsMixin
+
+```ts
+withMatchLabelsMixin(matchLabels)
+```
+
+"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \"key\", the operator is \"In\", and the values array contains only \"value\". The requirements are ANDed."
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.affinity.additionalPodAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.namespaceSelector.matchExpressions
+
+"matchExpressions is a list of label selector requirements. The requirements are ANDed."
+
+### fn spec.affinity.additionalPodAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.namespaceSelector.matchExpressions.withKey
+
+```ts
+withKey(key)
+```
+
+"key is the label key that the selector applies to."
+
+### fn spec.affinity.additionalPodAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.namespaceSelector.matchExpressions.withOperator
+
+```ts
+withOperator(operator)
+```
+
+"operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist."
+
+### fn spec.affinity.additionalPodAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.namespaceSelector.matchExpressions.withValues
+
+```ts
+withValues(values)
+```
+
+"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."
+
+### fn spec.affinity.additionalPodAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.namespaceSelector.matchExpressions.withValuesMixin
+
+```ts
+withValuesMixin(values)
+```
+
+"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.affinity.additionalPodAffinity.requiredDuringSchedulingIgnoredDuringExecution
+
+"If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied."
+
+### fn spec.affinity.additionalPodAffinity.requiredDuringSchedulingIgnoredDuringExecution.withNamespaces
+
+```ts
+withNamespaces(namespaces)
+```
+
+"namespaces specifies a static list of namespace names that the term applies to. The term is applied to the union of the namespaces listed in this field and the ones selected by namespaceSelector. null or empty namespaces list and null namespaceSelector means \"this pod's namespace\"."
+
+### fn spec.affinity.additionalPodAffinity.requiredDuringSchedulingIgnoredDuringExecution.withNamespacesMixin
+
+```ts
+withNamespacesMixin(namespaces)
+```
+
+"namespaces specifies a static list of namespace names that the term applies to. The term is applied to the union of the namespaces listed in this field and the ones selected by namespaceSelector. null or empty namespaces list and null namespaceSelector means \"this pod's namespace\"."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.affinity.additionalPodAffinity.requiredDuringSchedulingIgnoredDuringExecution.withTopologyKey
+
+```ts
+withTopologyKey(topologyKey)
+```
+
+"This pod should be co-located (affinity) or not co-located (anti-affinity) with the pods matching the labelSelector in the specified namespaces, where co-located is defined as running on a node whose value of the label with key topologyKey matches that of any node on which any of the selected pods is running. Empty topologyKey is not allowed."
+
+## obj spec.affinity.additionalPodAffinity.requiredDuringSchedulingIgnoredDuringExecution.labelSelector
+
+"A label query over a set of resources, in this case pods."
+
+### fn spec.affinity.additionalPodAffinity.requiredDuringSchedulingIgnoredDuringExecution.labelSelector.withMatchExpressions
+
+```ts
+withMatchExpressions(matchExpressions)
+```
+
+"matchExpressions is a list of label selector requirements. The requirements are ANDed."
+
+### fn spec.affinity.additionalPodAffinity.requiredDuringSchedulingIgnoredDuringExecution.labelSelector.withMatchExpressionsMixin
+
+```ts
+withMatchExpressionsMixin(matchExpressions)
+```
+
+"matchExpressions is a list of label selector requirements. The requirements are ANDed."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.affinity.additionalPodAffinity.requiredDuringSchedulingIgnoredDuringExecution.labelSelector.withMatchLabels
+
+```ts
+withMatchLabels(matchLabels)
+```
+
+"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \"key\", the operator is \"In\", and the values array contains only \"value\". The requirements are ANDed."
+
+### fn spec.affinity.additionalPodAffinity.requiredDuringSchedulingIgnoredDuringExecution.labelSelector.withMatchLabelsMixin
+
+```ts
+withMatchLabelsMixin(matchLabels)
+```
+
+"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \"key\", the operator is \"In\", and the values array contains only \"value\". The requirements are ANDed."
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.affinity.additionalPodAffinity.requiredDuringSchedulingIgnoredDuringExecution.labelSelector.matchExpressions
+
+"matchExpressions is a list of label selector requirements. The requirements are ANDed."
+
+### fn spec.affinity.additionalPodAffinity.requiredDuringSchedulingIgnoredDuringExecution.labelSelector.matchExpressions.withKey
+
+```ts
+withKey(key)
+```
+
+"key is the label key that the selector applies to."
+
+### fn spec.affinity.additionalPodAffinity.requiredDuringSchedulingIgnoredDuringExecution.labelSelector.matchExpressions.withOperator
+
+```ts
+withOperator(operator)
+```
+
+"operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist."
+
+### fn spec.affinity.additionalPodAffinity.requiredDuringSchedulingIgnoredDuringExecution.labelSelector.matchExpressions.withValues
+
+```ts
+withValues(values)
+```
+
+"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."
+
+### fn spec.affinity.additionalPodAffinity.requiredDuringSchedulingIgnoredDuringExecution.labelSelector.matchExpressions.withValuesMixin
+
+```ts
+withValuesMixin(values)
+```
+
+"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.affinity.additionalPodAffinity.requiredDuringSchedulingIgnoredDuringExecution.namespaceSelector
+
+"A label query over the set of namespaces that the term applies to. The term is applied to the union of the namespaces selected by this field and the ones listed in the namespaces field. null selector and null or empty namespaces list means \"this pod's namespace\". An empty selector ({}) matches all namespaces."
+
+### fn spec.affinity.additionalPodAffinity.requiredDuringSchedulingIgnoredDuringExecution.namespaceSelector.withMatchExpressions
+
+```ts
+withMatchExpressions(matchExpressions)
+```
+
+"matchExpressions is a list of label selector requirements. The requirements are ANDed."
+
+### fn spec.affinity.additionalPodAffinity.requiredDuringSchedulingIgnoredDuringExecution.namespaceSelector.withMatchExpressionsMixin
+
+```ts
+withMatchExpressionsMixin(matchExpressions)
+```
+
+"matchExpressions is a list of label selector requirements. The requirements are ANDed."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.affinity.additionalPodAffinity.requiredDuringSchedulingIgnoredDuringExecution.namespaceSelector.withMatchLabels
+
+```ts
+withMatchLabels(matchLabels)
+```
+
+"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \"key\", the operator is \"In\", and the values array contains only \"value\". The requirements are ANDed."
+
+### fn spec.affinity.additionalPodAffinity.requiredDuringSchedulingIgnoredDuringExecution.namespaceSelector.withMatchLabelsMixin
+
+```ts
+withMatchLabelsMixin(matchLabels)
+```
+
+"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \"key\", the operator is \"In\", and the values array contains only \"value\". The requirements are ANDed."
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.affinity.additionalPodAffinity.requiredDuringSchedulingIgnoredDuringExecution.namespaceSelector.matchExpressions
+
+"matchExpressions is a list of label selector requirements. The requirements are ANDed."
+
+### fn spec.affinity.additionalPodAffinity.requiredDuringSchedulingIgnoredDuringExecution.namespaceSelector.matchExpressions.withKey
+
+```ts
+withKey(key)
+```
+
+"key is the label key that the selector applies to."
+
+### fn spec.affinity.additionalPodAffinity.requiredDuringSchedulingIgnoredDuringExecution.namespaceSelector.matchExpressions.withOperator
+
+```ts
+withOperator(operator)
+```
+
+"operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist."
+
+### fn spec.affinity.additionalPodAffinity.requiredDuringSchedulingIgnoredDuringExecution.namespaceSelector.matchExpressions.withValues
+
+```ts
+withValues(values)
+```
+
+"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."
+
+### fn spec.affinity.additionalPodAffinity.requiredDuringSchedulingIgnoredDuringExecution.namespaceSelector.matchExpressions.withValuesMixin
+
+```ts
+withValuesMixin(values)
+```
+
+"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.affinity.additionalPodAntiAffinity
+
+"AdditionalPodAntiAffinity allows to specify pod anti-affinity terms to be added to the ones generated by the operator if EnablePodAntiAffinity is set to true (default) or to be used exclusively if set to false."
+
+### fn spec.affinity.additionalPodAntiAffinity.withPreferredDuringSchedulingIgnoredDuringExecution
+
+```ts
+withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution)
+```
+
+"The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding \"weight\" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred."
+
+### fn spec.affinity.additionalPodAntiAffinity.withPreferredDuringSchedulingIgnoredDuringExecutionMixin
+
+```ts
+withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution)
+```
+
+"The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding \"weight\" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.affinity.additionalPodAntiAffinity.withRequiredDuringSchedulingIgnoredDuringExecution
+
+```ts
+withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution)
+```
+
+"If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied."
+
+### fn spec.affinity.additionalPodAntiAffinity.withRequiredDuringSchedulingIgnoredDuringExecutionMixin
+
+```ts
+withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution)
+```
+
+"If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied."
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.affinity.additionalPodAntiAffinity.preferredDuringSchedulingIgnoredDuringExecution
+
+"The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding \"weight\" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred."
+
+### fn spec.affinity.additionalPodAntiAffinity.preferredDuringSchedulingIgnoredDuringExecution.withWeight
+
+```ts
+withWeight(weight)
+```
+
+"weight associated with matching the corresponding podAffinityTerm, in the range 1-100."
+
+## obj spec.affinity.additionalPodAntiAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm
+
+"Required. A pod affinity term, associated with the corresponding weight."
+
+### fn spec.affinity.additionalPodAntiAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.withNamespaces
+
+```ts
+withNamespaces(namespaces)
+```
+
+"namespaces specifies a static list of namespace names that the term applies to. The term is applied to the union of the namespaces listed in this field and the ones selected by namespaceSelector. null or empty namespaces list and null namespaceSelector means \"this pod's namespace\"."
+
+### fn spec.affinity.additionalPodAntiAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.withNamespacesMixin
+
+```ts
+withNamespacesMixin(namespaces)
+```
+
+"namespaces specifies a static list of namespace names that the term applies to. The term is applied to the union of the namespaces listed in this field and the ones selected by namespaceSelector. null or empty namespaces list and null namespaceSelector means \"this pod's namespace\"."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.affinity.additionalPodAntiAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.withTopologyKey
+
+```ts
+withTopologyKey(topologyKey)
+```
+
+"This pod should be co-located (affinity) or not co-located (anti-affinity) with the pods matching the labelSelector in the specified namespaces, where co-located is defined as running on a node whose value of the label with key topologyKey matches that of any node on which any of the selected pods is running. Empty topologyKey is not allowed."
+
+## obj spec.affinity.additionalPodAntiAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.labelSelector
+
+"A label query over a set of resources, in this case pods."
+
+### fn spec.affinity.additionalPodAntiAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.labelSelector.withMatchExpressions
+
+```ts
+withMatchExpressions(matchExpressions)
+```
+
+"matchExpressions is a list of label selector requirements. The requirements are ANDed."
+
+### fn spec.affinity.additionalPodAntiAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.labelSelector.withMatchExpressionsMixin
+
+```ts
+withMatchExpressionsMixin(matchExpressions)
+```
+
+"matchExpressions is a list of label selector requirements. The requirements are ANDed."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.affinity.additionalPodAntiAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.labelSelector.withMatchLabels
+
+```ts
+withMatchLabels(matchLabels)
+```
+
+"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \"key\", the operator is \"In\", and the values array contains only \"value\". The requirements are ANDed."
+
+### fn spec.affinity.additionalPodAntiAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.labelSelector.withMatchLabelsMixin
+
+```ts
+withMatchLabelsMixin(matchLabels)
+```
+
+"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \"key\", the operator is \"In\", and the values array contains only \"value\". The requirements are ANDed."
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.affinity.additionalPodAntiAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.labelSelector.matchExpressions
+
+"matchExpressions is a list of label selector requirements. The requirements are ANDed."
+
+### fn spec.affinity.additionalPodAntiAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.labelSelector.matchExpressions.withKey
+
+```ts
+withKey(key)
+```
+
+"key is the label key that the selector applies to."
+
+### fn spec.affinity.additionalPodAntiAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.labelSelector.matchExpressions.withOperator
+
+```ts
+withOperator(operator)
+```
+
+"operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist."
+
+### fn spec.affinity.additionalPodAntiAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.labelSelector.matchExpressions.withValues
+
+```ts
+withValues(values)
+```
+
+"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."
+
+### fn spec.affinity.additionalPodAntiAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.labelSelector.matchExpressions.withValuesMixin
+
+```ts
+withValuesMixin(values)
+```
+
+"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.affinity.additionalPodAntiAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.namespaceSelector
+
+"A label query over the set of namespaces that the term applies to. The term is applied to the union of the namespaces selected by this field and the ones listed in the namespaces field. null selector and null or empty namespaces list means \"this pod's namespace\". An empty selector ({}) matches all namespaces."
+
+### fn spec.affinity.additionalPodAntiAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.namespaceSelector.withMatchExpressions
+
+```ts
+withMatchExpressions(matchExpressions)
+```
+
+"matchExpressions is a list of label selector requirements. The requirements are ANDed."
+
+### fn spec.affinity.additionalPodAntiAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.namespaceSelector.withMatchExpressionsMixin
+
+```ts
+withMatchExpressionsMixin(matchExpressions)
+```
+
+"matchExpressions is a list of label selector requirements. The requirements are ANDed."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.affinity.additionalPodAntiAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.namespaceSelector.withMatchLabels
+
+```ts
+withMatchLabels(matchLabels)
+```
+
+"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \"key\", the operator is \"In\", and the values array contains only \"value\". The requirements are ANDed."
+
+### fn spec.affinity.additionalPodAntiAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.namespaceSelector.withMatchLabelsMixin
+
+```ts
+withMatchLabelsMixin(matchLabels)
+```
+
+"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \"key\", the operator is \"In\", and the values array contains only \"value\". The requirements are ANDed."
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.affinity.additionalPodAntiAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.namespaceSelector.matchExpressions
+
+"matchExpressions is a list of label selector requirements. The requirements are ANDed."
+
+### fn spec.affinity.additionalPodAntiAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.namespaceSelector.matchExpressions.withKey
+
+```ts
+withKey(key)
+```
+
+"key is the label key that the selector applies to."
+
+### fn spec.affinity.additionalPodAntiAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.namespaceSelector.matchExpressions.withOperator
+
+```ts
+withOperator(operator)
+```
+
+"operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist."
+
+### fn spec.affinity.additionalPodAntiAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.namespaceSelector.matchExpressions.withValues
+
+```ts
+withValues(values)
+```
+
+"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."
+
+### fn spec.affinity.additionalPodAntiAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.namespaceSelector.matchExpressions.withValuesMixin
+
+```ts
+withValuesMixin(values)
+```
+
+"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.affinity.additionalPodAntiAffinity.requiredDuringSchedulingIgnoredDuringExecution
+
+"If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied."
+
+### fn spec.affinity.additionalPodAntiAffinity.requiredDuringSchedulingIgnoredDuringExecution.withNamespaces
+
+```ts
+withNamespaces(namespaces)
+```
+
+"namespaces specifies a static list of namespace names that the term applies to. The term is applied to the union of the namespaces listed in this field and the ones selected by namespaceSelector. null or empty namespaces list and null namespaceSelector means \"this pod's namespace\"."
+
+### fn spec.affinity.additionalPodAntiAffinity.requiredDuringSchedulingIgnoredDuringExecution.withNamespacesMixin
+
+```ts
+withNamespacesMixin(namespaces)
+```
+
+"namespaces specifies a static list of namespace names that the term applies to. The term is applied to the union of the namespaces listed in this field and the ones selected by namespaceSelector. null or empty namespaces list and null namespaceSelector means \"this pod's namespace\"."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.affinity.additionalPodAntiAffinity.requiredDuringSchedulingIgnoredDuringExecution.withTopologyKey
+
+```ts
+withTopologyKey(topologyKey)
+```
+
+"This pod should be co-located (affinity) or not co-located (anti-affinity) with the pods matching the labelSelector in the specified namespaces, where co-located is defined as running on a node whose value of the label with key topologyKey matches that of any node on which any of the selected pods is running. Empty topologyKey is not allowed."
+
+## obj spec.affinity.additionalPodAntiAffinity.requiredDuringSchedulingIgnoredDuringExecution.labelSelector
+
+"A label query over a set of resources, in this case pods."
+
+### fn spec.affinity.additionalPodAntiAffinity.requiredDuringSchedulingIgnoredDuringExecution.labelSelector.withMatchExpressions
+
+```ts
+withMatchExpressions(matchExpressions)
+```
+
+"matchExpressions is a list of label selector requirements. The requirements are ANDed."
+
+### fn spec.affinity.additionalPodAntiAffinity.requiredDuringSchedulingIgnoredDuringExecution.labelSelector.withMatchExpressionsMixin
+
+```ts
+withMatchExpressionsMixin(matchExpressions)
+```
+
+"matchExpressions is a list of label selector requirements. The requirements are ANDed."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.affinity.additionalPodAntiAffinity.requiredDuringSchedulingIgnoredDuringExecution.labelSelector.withMatchLabels
+
+```ts
+withMatchLabels(matchLabels)
+```
+
+"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \"key\", the operator is \"In\", and the values array contains only \"value\". The requirements are ANDed."
+
+### fn spec.affinity.additionalPodAntiAffinity.requiredDuringSchedulingIgnoredDuringExecution.labelSelector.withMatchLabelsMixin
+
+```ts
+withMatchLabelsMixin(matchLabels)
+```
+
+"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \"key\", the operator is \"In\", and the values array contains only \"value\". The requirements are ANDed."
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.affinity.additionalPodAntiAffinity.requiredDuringSchedulingIgnoredDuringExecution.labelSelector.matchExpressions
+
+"matchExpressions is a list of label selector requirements. The requirements are ANDed."
+
+### fn spec.affinity.additionalPodAntiAffinity.requiredDuringSchedulingIgnoredDuringExecution.labelSelector.matchExpressions.withKey
+
+```ts
+withKey(key)
+```
+
+"key is the label key that the selector applies to."
+
+### fn spec.affinity.additionalPodAntiAffinity.requiredDuringSchedulingIgnoredDuringExecution.labelSelector.matchExpressions.withOperator
+
+```ts
+withOperator(operator)
+```
+
+"operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist."
+
+### fn spec.affinity.additionalPodAntiAffinity.requiredDuringSchedulingIgnoredDuringExecution.labelSelector.matchExpressions.withValues
+
+```ts
+withValues(values)
+```
+
+"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."
+
+### fn spec.affinity.additionalPodAntiAffinity.requiredDuringSchedulingIgnoredDuringExecution.labelSelector.matchExpressions.withValuesMixin
+
+```ts
+withValuesMixin(values)
+```
+
+"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.affinity.additionalPodAntiAffinity.requiredDuringSchedulingIgnoredDuringExecution.namespaceSelector
+
+"A label query over the set of namespaces that the term applies to. The term is applied to the union of the namespaces selected by this field and the ones listed in the namespaces field. null selector and null or empty namespaces list means \"this pod's namespace\". An empty selector ({}) matches all namespaces."
+
+### fn spec.affinity.additionalPodAntiAffinity.requiredDuringSchedulingIgnoredDuringExecution.namespaceSelector.withMatchExpressions
+
+```ts
+withMatchExpressions(matchExpressions)
+```
+
+"matchExpressions is a list of label selector requirements. The requirements are ANDed."
+
+### fn spec.affinity.additionalPodAntiAffinity.requiredDuringSchedulingIgnoredDuringExecution.namespaceSelector.withMatchExpressionsMixin
+
+```ts
+withMatchExpressionsMixin(matchExpressions)
+```
+
+"matchExpressions is a list of label selector requirements. The requirements are ANDed."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.affinity.additionalPodAntiAffinity.requiredDuringSchedulingIgnoredDuringExecution.namespaceSelector.withMatchLabels
+
+```ts
+withMatchLabels(matchLabels)
+```
+
+"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \"key\", the operator is \"In\", and the values array contains only \"value\". The requirements are ANDed."
+
+### fn spec.affinity.additionalPodAntiAffinity.requiredDuringSchedulingIgnoredDuringExecution.namespaceSelector.withMatchLabelsMixin
+
+```ts
+withMatchLabelsMixin(matchLabels)
+```
+
+"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \"key\", the operator is \"In\", and the values array contains only \"value\". The requirements are ANDed."
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.affinity.additionalPodAntiAffinity.requiredDuringSchedulingIgnoredDuringExecution.namespaceSelector.matchExpressions
+
+"matchExpressions is a list of label selector requirements. The requirements are ANDed."
+
+### fn spec.affinity.additionalPodAntiAffinity.requiredDuringSchedulingIgnoredDuringExecution.namespaceSelector.matchExpressions.withKey
+
+```ts
+withKey(key)
+```
+
+"key is the label key that the selector applies to."
+
+### fn spec.affinity.additionalPodAntiAffinity.requiredDuringSchedulingIgnoredDuringExecution.namespaceSelector.matchExpressions.withOperator
+
+```ts
+withOperator(operator)
+```
+
+"operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist."
+
+### fn spec.affinity.additionalPodAntiAffinity.requiredDuringSchedulingIgnoredDuringExecution.namespaceSelector.matchExpressions.withValues
+
+```ts
+withValues(values)
+```
+
+"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."
+
+### fn spec.affinity.additionalPodAntiAffinity.requiredDuringSchedulingIgnoredDuringExecution.namespaceSelector.matchExpressions.withValuesMixin
+
+```ts
+withValuesMixin(values)
+```
+
+"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.affinity.nodeAffinity
+
+"NodeAffinity describes node affinity scheduling rules for the pod. More info: https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#node-affinity"
+
+### fn spec.affinity.nodeAffinity.withPreferredDuringSchedulingIgnoredDuringExecution
+
+```ts
+withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution)
+```
+
+"The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding \"weight\" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred."
+
+### fn spec.affinity.nodeAffinity.withPreferredDuringSchedulingIgnoredDuringExecutionMixin
+
+```ts
+withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution)
+```
+
+"The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding \"weight\" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred."
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.affinity.nodeAffinity.preferredDuringSchedulingIgnoredDuringExecution
+
+"The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding \"weight\" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred."
+
+### fn spec.affinity.nodeAffinity.preferredDuringSchedulingIgnoredDuringExecution.withWeight
+
+```ts
+withWeight(weight)
+```
+
+"Weight associated with matching the corresponding nodeSelectorTerm, in the range 1-100."
+
+## obj spec.affinity.nodeAffinity.preferredDuringSchedulingIgnoredDuringExecution.preference
+
+"A node selector term, associated with the corresponding weight."
+
+### fn spec.affinity.nodeAffinity.preferredDuringSchedulingIgnoredDuringExecution.preference.withMatchExpressions
+
+```ts
+withMatchExpressions(matchExpressions)
+```
+
+"A list of node selector requirements by node's labels."
+
+### fn spec.affinity.nodeAffinity.preferredDuringSchedulingIgnoredDuringExecution.preference.withMatchExpressionsMixin
+
+```ts
+withMatchExpressionsMixin(matchExpressions)
+```
+
+"A list of node selector requirements by node's labels."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.affinity.nodeAffinity.preferredDuringSchedulingIgnoredDuringExecution.preference.withMatchFields
+
+```ts
+withMatchFields(matchFields)
+```
+
+"A list of node selector requirements by node's fields."
+
+### fn spec.affinity.nodeAffinity.preferredDuringSchedulingIgnoredDuringExecution.preference.withMatchFieldsMixin
+
+```ts
+withMatchFieldsMixin(matchFields)
+```
+
+"A list of node selector requirements by node's fields."
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.affinity.nodeAffinity.preferredDuringSchedulingIgnoredDuringExecution.preference.matchExpressions
+
+"A list of node selector requirements by node's labels."
+
+### fn spec.affinity.nodeAffinity.preferredDuringSchedulingIgnoredDuringExecution.preference.matchExpressions.withKey
+
+```ts
+withKey(key)
+```
+
+"The label key that the selector applies to."
+
+### fn spec.affinity.nodeAffinity.preferredDuringSchedulingIgnoredDuringExecution.preference.matchExpressions.withOperator
+
+```ts
+withOperator(operator)
+```
+
+"Represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists, DoesNotExist. Gt, and Lt."
+
+### fn spec.affinity.nodeAffinity.preferredDuringSchedulingIgnoredDuringExecution.preference.matchExpressions.withValues
+
+```ts
+withValues(values)
+```
+
+"An array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. If the operator is Gt or Lt, the values array must have a single element, which will be interpreted as an integer. This array is replaced during a strategic merge patch."
+
+### fn spec.affinity.nodeAffinity.preferredDuringSchedulingIgnoredDuringExecution.preference.matchExpressions.withValuesMixin
+
+```ts
+withValuesMixin(values)
+```
+
+"An array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. If the operator is Gt or Lt, the values array must have a single element, which will be interpreted as an integer. This array is replaced during a strategic merge patch."
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.affinity.nodeAffinity.preferredDuringSchedulingIgnoredDuringExecution.preference.matchFields
+
+"A list of node selector requirements by node's fields."
+
+### fn spec.affinity.nodeAffinity.preferredDuringSchedulingIgnoredDuringExecution.preference.matchFields.withKey
+
+```ts
+withKey(key)
+```
+
+"The label key that the selector applies to."
+
+### fn spec.affinity.nodeAffinity.preferredDuringSchedulingIgnoredDuringExecution.preference.matchFields.withOperator
+
+```ts
+withOperator(operator)
+```
+
+"Represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists, DoesNotExist. Gt, and Lt."
+
+### fn spec.affinity.nodeAffinity.preferredDuringSchedulingIgnoredDuringExecution.preference.matchFields.withValues
+
+```ts
+withValues(values)
+```
+
+"An array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. If the operator is Gt or Lt, the values array must have a single element, which will be interpreted as an integer. This array is replaced during a strategic merge patch."
+
+### fn spec.affinity.nodeAffinity.preferredDuringSchedulingIgnoredDuringExecution.preference.matchFields.withValuesMixin
+
+```ts
+withValuesMixin(values)
+```
+
+"An array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. If the operator is Gt or Lt, the values array must have a single element, which will be interpreted as an integer. This array is replaced during a strategic merge patch."
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.affinity.nodeAffinity.requiredDuringSchedulingIgnoredDuringExecution
+
+"If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to an update), the system may or may not try to eventually evict the pod from its node."
+
+### fn spec.affinity.nodeAffinity.requiredDuringSchedulingIgnoredDuringExecution.withNodeSelectorTerms
+
+```ts
+withNodeSelectorTerms(nodeSelectorTerms)
+```
+
+"Required. A list of node selector terms. The terms are ORed."
+
+### fn spec.affinity.nodeAffinity.requiredDuringSchedulingIgnoredDuringExecution.withNodeSelectorTermsMixin
+
+```ts
+withNodeSelectorTermsMixin(nodeSelectorTerms)
+```
+
+"Required. A list of node selector terms. The terms are ORed."
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.affinity.nodeAffinity.requiredDuringSchedulingIgnoredDuringExecution.nodeSelectorTerms
+
+"Required. A list of node selector terms. The terms are ORed."
+
+### fn spec.affinity.nodeAffinity.requiredDuringSchedulingIgnoredDuringExecution.nodeSelectorTerms.withMatchExpressions
+
+```ts
+withMatchExpressions(matchExpressions)
+```
+
+"A list of node selector requirements by node's labels."
+
+### fn spec.affinity.nodeAffinity.requiredDuringSchedulingIgnoredDuringExecution.nodeSelectorTerms.withMatchExpressionsMixin
+
+```ts
+withMatchExpressionsMixin(matchExpressions)
+```
+
+"A list of node selector requirements by node's labels."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.affinity.nodeAffinity.requiredDuringSchedulingIgnoredDuringExecution.nodeSelectorTerms.withMatchFields
+
+```ts
+withMatchFields(matchFields)
+```
+
+"A list of node selector requirements by node's fields."
+
+### fn spec.affinity.nodeAffinity.requiredDuringSchedulingIgnoredDuringExecution.nodeSelectorTerms.withMatchFieldsMixin
+
+```ts
+withMatchFieldsMixin(matchFields)
+```
+
+"A list of node selector requirements by node's fields."
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.affinity.nodeAffinity.requiredDuringSchedulingIgnoredDuringExecution.nodeSelectorTerms.matchExpressions
+
+"A list of node selector requirements by node's labels."
+
+### fn spec.affinity.nodeAffinity.requiredDuringSchedulingIgnoredDuringExecution.nodeSelectorTerms.matchExpressions.withKey
+
+```ts
+withKey(key)
+```
+
+"The label key that the selector applies to."
+
+### fn spec.affinity.nodeAffinity.requiredDuringSchedulingIgnoredDuringExecution.nodeSelectorTerms.matchExpressions.withOperator
+
+```ts
+withOperator(operator)
+```
+
+"Represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists, DoesNotExist. Gt, and Lt."
+
+### fn spec.affinity.nodeAffinity.requiredDuringSchedulingIgnoredDuringExecution.nodeSelectorTerms.matchExpressions.withValues
+
+```ts
+withValues(values)
+```
+
+"An array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. If the operator is Gt or Lt, the values array must have a single element, which will be interpreted as an integer. This array is replaced during a strategic merge patch."
+
+### fn spec.affinity.nodeAffinity.requiredDuringSchedulingIgnoredDuringExecution.nodeSelectorTerms.matchExpressions.withValuesMixin
+
+```ts
+withValuesMixin(values)
+```
+
+"An array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. If the operator is Gt or Lt, the values array must have a single element, which will be interpreted as an integer. This array is replaced during a strategic merge patch."
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.affinity.nodeAffinity.requiredDuringSchedulingIgnoredDuringExecution.nodeSelectorTerms.matchFields
+
+"A list of node selector requirements by node's fields."
+
+### fn spec.affinity.nodeAffinity.requiredDuringSchedulingIgnoredDuringExecution.nodeSelectorTerms.matchFields.withKey
+
+```ts
+withKey(key)
+```
+
+"The label key that the selector applies to."
+
+### fn spec.affinity.nodeAffinity.requiredDuringSchedulingIgnoredDuringExecution.nodeSelectorTerms.matchFields.withOperator
+
+```ts
+withOperator(operator)
+```
+
+"Represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists, DoesNotExist. Gt, and Lt."
+
+### fn spec.affinity.nodeAffinity.requiredDuringSchedulingIgnoredDuringExecution.nodeSelectorTerms.matchFields.withValues
+
+```ts
+withValues(values)
+```
+
+"An array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. If the operator is Gt or Lt, the values array must have a single element, which will be interpreted as an integer. This array is replaced during a strategic merge patch."
+
+### fn spec.affinity.nodeAffinity.requiredDuringSchedulingIgnoredDuringExecution.nodeSelectorTerms.matchFields.withValuesMixin
+
+```ts
+withValuesMixin(values)
+```
+
+"An array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. If the operator is Gt or Lt, the values array must have a single element, which will be interpreted as an integer. This array is replaced during a strategic merge patch."
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.affinity.tolerations
+
+"Tolerations is a list of Tolerations that should be set for all the pods, in order to allow them to run on tainted nodes. More info: https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/"
+
+### fn spec.affinity.tolerations.withEffect
+
+```ts
+withEffect(effect)
+```
+
+"Effect indicates the taint effect to match. Empty means match all taint effects. When specified, allowed values are NoSchedule, PreferNoSchedule and NoExecute."
+
+### fn spec.affinity.tolerations.withKey
+
+```ts
+withKey(key)
+```
+
+"Key is the taint key that the toleration applies to. Empty means match all taint keys. If the key is empty, operator must be Exists; this combination means to match all values and all keys."
+
+### fn spec.affinity.tolerations.withOperator
+
+```ts
+withOperator(operator)
+```
+
+"Operator represents a key's relationship to the value. Valid operators are Exists and Equal. Defaults to Equal. Exists is equivalent to wildcard for value, so that a pod can tolerate all taints of a particular category."
+
+### fn spec.affinity.tolerations.withTolerationSeconds
+
+```ts
+withTolerationSeconds(tolerationSeconds)
+```
+
+"TolerationSeconds represents the period of time the toleration (which must be of effect NoExecute, otherwise this field is ignored) tolerates the taint. By default, it is not set, which means tolerate the taint forever (do not evict). Zero and negative values will be treated as 0 (evict immediately) by the system."
+
+### fn spec.affinity.tolerations.withValue
+
+```ts
+withValue(value)
+```
+
+"Value is the taint value the toleration matches to. If the operator is Exists, the value should be empty, otherwise just a regular string."
+
+## obj spec.backup
+
+"The configuration to be used for backups"
+
+### fn spec.backup.withRetentionPolicy
+
+```ts
+withRetentionPolicy(retentionPolicy)
+```
+
+"RetentionPolicy is the retention policy to be used for backups and WALs (i.e. '60d'). The retention policy is expressed in the form of `XXu` where `XX` is a positive integer and `u` is in `[dwm]` - days, weeks, months. It's currently only applicable when using the BarmanObjectStore method."
+
+### fn spec.backup.withTarget
+
+```ts
+withTarget(target)
+```
+
+"The policy to decide which instance should perform backups. Available options are empty string, which will default to `prefer-standby` policy, `primary` to have backups run always on primary instances, `prefer-standby` to have backups run preferably on the most updated standby, if available."
+
+## obj spec.backup.barmanObjectStore
+
+"The configuration for the barman-cloud tool suite"
+
+### fn spec.backup.barmanObjectStore.withDestinationPath
+
+```ts
+withDestinationPath(destinationPath)
+```
+
+"The path where to store the backup (i.e. s3://bucket/path/to/folder) this path, with different destination folders, will be used for WALs and for data"
+
+### fn spec.backup.barmanObjectStore.withEndpointURL
+
+```ts
+withEndpointURL(endpointURL)
+```
+
+"Endpoint to be used to upload data to the cloud, overriding the automatic endpoint discovery"
+
+### fn spec.backup.barmanObjectStore.withHistoryTags
+
+```ts
+withHistoryTags(historyTags)
+```
+
+"HistoryTags is a list of key value pairs that will be passed to the Barman --history-tags option."
+
+### fn spec.backup.barmanObjectStore.withHistoryTagsMixin
+
+```ts
+withHistoryTagsMixin(historyTags)
+```
+
+"HistoryTags is a list of key value pairs that will be passed to the Barman --history-tags option."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.backup.barmanObjectStore.withServerName
+
+```ts
+withServerName(serverName)
+```
+
+"The server name on S3, the cluster name is used if this parameter is omitted"
+
+### fn spec.backup.barmanObjectStore.withTags
+
+```ts
+withTags(tags)
+```
+
+"Tags is a list of key value pairs that will be passed to the Barman --tags option."
+
+### fn spec.backup.barmanObjectStore.withTagsMixin
+
+```ts
+withTagsMixin(tags)
+```
+
+"Tags is a list of key value pairs that will be passed to the Barman --tags option."
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.backup.barmanObjectStore.azureCredentials
+
+"The credentials to use to upload data to Azure Blob Storage"
+
+### fn spec.backup.barmanObjectStore.azureCredentials.withInheritFromAzureAD
+
+```ts
+withInheritFromAzureAD(inheritFromAzureAD)
+```
+
+"Use the Azure AD based authentication without providing explicitly the keys."
+
+## obj spec.backup.barmanObjectStore.azureCredentials.connectionString
+
+"The connection string to be used"
+
+### fn spec.backup.barmanObjectStore.azureCredentials.connectionString.withKey
+
+```ts
+withKey(key)
+```
+
+"The key to select"
+
+### fn spec.backup.barmanObjectStore.azureCredentials.connectionString.withName
+
+```ts
+withName(name)
+```
+
+"Name of the referent."
+
+## obj spec.backup.barmanObjectStore.azureCredentials.storageAccount
+
+"The storage account where to upload data"
+
+### fn spec.backup.barmanObjectStore.azureCredentials.storageAccount.withKey
+
+```ts
+withKey(key)
+```
+
+"The key to select"
+
+### fn spec.backup.barmanObjectStore.azureCredentials.storageAccount.withName
+
+```ts
+withName(name)
+```
+
+"Name of the referent."
+
+## obj spec.backup.barmanObjectStore.azureCredentials.storageKey
+
+"The storage account key to be used in conjunction with the storage account name"
+
+### fn spec.backup.barmanObjectStore.azureCredentials.storageKey.withKey
+
+```ts
+withKey(key)
+```
+
+"The key to select"
+
+### fn spec.backup.barmanObjectStore.azureCredentials.storageKey.withName
+
+```ts
+withName(name)
+```
+
+"Name of the referent."
+
+## obj spec.backup.barmanObjectStore.azureCredentials.storageSasToken
+
+"A shared-access-signature to be used in conjunction with the storage account name"
+
+### fn spec.backup.barmanObjectStore.azureCredentials.storageSasToken.withKey
+
+```ts
+withKey(key)
+```
+
+"The key to select"
+
+### fn spec.backup.barmanObjectStore.azureCredentials.storageSasToken.withName
+
+```ts
+withName(name)
+```
+
+"Name of the referent."
+
+## obj spec.backup.barmanObjectStore.data
+
+"The configuration to be used to backup the data files When not defined, base backups files will be stored uncompressed and may be unencrypted in the object store, according to the bucket default policy."
+
+### fn spec.backup.barmanObjectStore.data.withCompression
+
+```ts
+withCompression(compression)
+```
+
+"Compress a backup file (a tar file per tablespace) while streaming it to the object store. Available options are empty string (no compression, default), `gzip`, `bzip2` or `snappy`."
+
+### fn spec.backup.barmanObjectStore.data.withEncryption
+
+```ts
+withEncryption(encryption)
+```
+
+"Whenever to force the encryption of files (if the bucket is not already configured for that). Allowed options are empty string (use the bucket policy, default), `AES256` and `aws:kms`"
+
+### fn spec.backup.barmanObjectStore.data.withImmediateCheckpoint
+
+```ts
+withImmediateCheckpoint(immediateCheckpoint)
+```
+
+"Control whether the I/O workload for the backup initial checkpoint will be limited, according to the `checkpoint_completion_target` setting on the PostgreSQL server. If set to true, an immediate checkpoint will be used, meaning PostgreSQL will complete the checkpoint as soon as possible. `false` by default."
+
+### fn spec.backup.barmanObjectStore.data.withJobs
+
+```ts
+withJobs(jobs)
+```
+
+"The number of parallel jobs to be used to upload the backup, defaults to 2"
+
+## obj spec.backup.barmanObjectStore.endpointCA
+
+"EndpointCA store the CA bundle of the barman endpoint. Useful when using self-signed certificates to avoid errors with certificate issuer and barman-cloud-wal-archive"
+
+### fn spec.backup.barmanObjectStore.endpointCA.withKey
+
+```ts
+withKey(key)
+```
+
+"The key to select"
+
+### fn spec.backup.barmanObjectStore.endpointCA.withName
+
+```ts
+withName(name)
+```
+
+"Name of the referent."
+
+## obj spec.backup.barmanObjectStore.googleCredentials
+
+"The credentials to use to upload data to Google Cloud Storage"
+
+### fn spec.backup.barmanObjectStore.googleCredentials.withGkeEnvironment
+
+```ts
+withGkeEnvironment(gkeEnvironment)
+```
+
+"If set to true, will presume that it's running inside a GKE environment, default to false."
+
+## obj spec.backup.barmanObjectStore.googleCredentials.applicationCredentials
+
+"The secret containing the Google Cloud Storage JSON file with the credentials"
+
+### fn spec.backup.barmanObjectStore.googleCredentials.applicationCredentials.withKey
+
+```ts
+withKey(key)
+```
+
+"The key to select"
+
+### fn spec.backup.barmanObjectStore.googleCredentials.applicationCredentials.withName
+
+```ts
+withName(name)
+```
+
+"Name of the referent."
+
+## obj spec.backup.barmanObjectStore.s3Credentials
+
+"The credentials to use to upload data to S3"
+
+### fn spec.backup.barmanObjectStore.s3Credentials.withInheritFromIAMRole
+
+```ts
+withInheritFromIAMRole(inheritFromIAMRole)
+```
+
+"Use the role based authentication without providing explicitly the keys."
+
+## obj spec.backup.barmanObjectStore.s3Credentials.accessKeyId
+
+"The reference to the access key id"
+
+### fn spec.backup.barmanObjectStore.s3Credentials.accessKeyId.withKey
+
+```ts
+withKey(key)
+```
+
+"The key to select"
+
+### fn spec.backup.barmanObjectStore.s3Credentials.accessKeyId.withName
+
+```ts
+withName(name)
+```
+
+"Name of the referent."
+
+## obj spec.backup.barmanObjectStore.s3Credentials.region
+
+"The reference to the secret containing the region name"
+
+### fn spec.backup.barmanObjectStore.s3Credentials.region.withKey
+
+```ts
+withKey(key)
+```
+
+"The key to select"
+
+### fn spec.backup.barmanObjectStore.s3Credentials.region.withName
+
+```ts
+withName(name)
+```
+
+"Name of the referent."
+
+## obj spec.backup.barmanObjectStore.s3Credentials.secretAccessKey
+
+"The reference to the secret access key"
+
+### fn spec.backup.barmanObjectStore.s3Credentials.secretAccessKey.withKey
+
+```ts
+withKey(key)
+```
+
+"The key to select"
+
+### fn spec.backup.barmanObjectStore.s3Credentials.secretAccessKey.withName
+
+```ts
+withName(name)
+```
+
+"Name of the referent."
+
+## obj spec.backup.barmanObjectStore.s3Credentials.sessionToken
+
+"The references to the session key"
+
+### fn spec.backup.barmanObjectStore.s3Credentials.sessionToken.withKey
+
+```ts
+withKey(key)
+```
+
+"The key to select"
+
+### fn spec.backup.barmanObjectStore.s3Credentials.sessionToken.withName
+
+```ts
+withName(name)
+```
+
+"Name of the referent."
+
+## obj spec.backup.barmanObjectStore.wal
+
+"The configuration for the backup of the WAL stream. When not defined, WAL files will be stored uncompressed and may be unencrypted in the object store, according to the bucket default policy."
+
+### fn spec.backup.barmanObjectStore.wal.withCompression
+
+```ts
+withCompression(compression)
+```
+
+"Compress a WAL file before sending it to the object store. Available options are empty string (no compression, default), `gzip`, `bzip2` or `snappy`."
+
+### fn spec.backup.barmanObjectStore.wal.withEncryption
+
+```ts
+withEncryption(encryption)
+```
+
+"Whenever to force the encryption of files (if the bucket is not already configured for that). Allowed options are empty string (use the bucket policy, default), `AES256` and `aws:kms`"
+
+### fn spec.backup.barmanObjectStore.wal.withMaxParallel
+
+```ts
+withMaxParallel(maxParallel)
+```
+
+"Number of WAL files to be either archived in parallel (when the PostgreSQL instance is archiving to a backup object store) or restored in parallel (when a PostgreSQL standby is fetching WAL files from a recovery object store). If not specified, WAL files will be processed one at a time. It accepts a positive integer as a value - with 1 being the minimum accepted value."
+
+## obj spec.backup.volumeSnapshot
+
+"VolumeSnapshot provides the configuration for the execution of volume snapshot backups."
+
+### fn spec.backup.volumeSnapshot.withAnnotations
+
+```ts
+withAnnotations(annotations)
+```
+
+"Annotations key-value pairs that will be added to .metadata.annotations snapshot resources."
+
+### fn spec.backup.volumeSnapshot.withAnnotationsMixin
+
+```ts
+withAnnotationsMixin(annotations)
+```
+
+"Annotations key-value pairs that will be added to .metadata.annotations snapshot resources."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.backup.volumeSnapshot.withClassName
+
+```ts
+withClassName(className)
+```
+
+"ClassName specifies the Snapshot Class to be used for PG_DATA PersistentVolumeClaim. It is the default class for the other types if no specific class is present"
+
+### fn spec.backup.volumeSnapshot.withLabels
+
+```ts
+withLabels(labels)
+```
+
+"Labels are key-value pairs that will be added to .metadata.labels snapshot resources."
+
+### fn spec.backup.volumeSnapshot.withLabelsMixin
+
+```ts
+withLabelsMixin(labels)
+```
+
+"Labels are key-value pairs that will be added to .metadata.labels snapshot resources."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.backup.volumeSnapshot.withSnapshotOwnerReference
+
+```ts
+withSnapshotOwnerReference(snapshotOwnerReference)
+```
+
+"SnapshotOwnerReference indicates the type of owner reference the snapshot should have"
+
+### fn spec.backup.volumeSnapshot.withWalClassName
+
+```ts
+withWalClassName(walClassName)
+```
+
+"WalClassName specifies the Snapshot Class to be used for the PG_WAL PersistentVolumeClaim."
+
+## obj spec.bootstrap
+
+"Instructions to bootstrap this cluster"
+
+## obj spec.bootstrap.initdb
+
+"Bootstrap the cluster via initdb"
+
+### fn spec.bootstrap.initdb.withDataChecksums
+
+```ts
+withDataChecksums(dataChecksums)
+```
+
+"Whether the `-k` option should be passed to initdb, enabling checksums on data pages (default: `false`)"
+
+### fn spec.bootstrap.initdb.withDatabase
+
+```ts
+withDatabase(database)
+```
+
+"Name of the database used by the application. Default: `app`."
+
+### fn spec.bootstrap.initdb.withEncoding
+
+```ts
+withEncoding(encoding)
+```
+
+"The value to be passed as option `--encoding` for initdb (default:`UTF8`)"
+
+### fn spec.bootstrap.initdb.withLocaleCType
+
+```ts
+withLocaleCType(localeCType)
+```
+
+"The value to be passed as option `--lc-ctype` for initdb (default:`C`)"
+
+### fn spec.bootstrap.initdb.withLocaleCollate
+
+```ts
+withLocaleCollate(localeCollate)
+```
+
+"The value to be passed as option `--lc-collate` for initdb (default:`C`)"
+
+### fn spec.bootstrap.initdb.withOptions
+
+```ts
+withOptions(options)
+```
+
+"The list of options that must be passed to initdb when creating the cluster. Deprecated: This could lead to inconsistent configurations, please use the explicit provided parameters instead. If defined, explicit values will be ignored."
+
+### fn spec.bootstrap.initdb.withOptionsMixin
+
+```ts
+withOptionsMixin(options)
+```
+
+"The list of options that must be passed to initdb when creating the cluster. Deprecated: This could lead to inconsistent configurations, please use the explicit provided parameters instead. If defined, explicit values will be ignored."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.bootstrap.initdb.withOwner
+
+```ts
+withOwner(owner)
+```
+
+"Name of the owner of the database in the instance to be used by applications. Defaults to the value of the `database` key."
+
+### fn spec.bootstrap.initdb.withPostInitApplicationSQL
+
+```ts
+withPostInitApplicationSQL(postInitApplicationSQL)
+```
+
+"List of SQL queries to be executed as a superuser in the application database right after is created - to be used with extreme care (by default empty)"
+
+### fn spec.bootstrap.initdb.withPostInitApplicationSQLMixin
+
+```ts
+withPostInitApplicationSQLMixin(postInitApplicationSQL)
+```
+
+"List of SQL queries to be executed as a superuser in the application database right after is created - to be used with extreme care (by default empty)"
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.bootstrap.initdb.withPostInitSQL
+
+```ts
+withPostInitSQL(postInitSQL)
+```
+
+"List of SQL queries to be executed as a superuser immediately after the cluster has been created - to be used with extreme care (by default empty)"
+
+### fn spec.bootstrap.initdb.withPostInitSQLMixin
+
+```ts
+withPostInitSQLMixin(postInitSQL)
+```
+
+"List of SQL queries to be executed as a superuser immediately after the cluster has been created - to be used with extreme care (by default empty)"
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.bootstrap.initdb.withPostInitTemplateSQL
+
+```ts
+withPostInitTemplateSQL(postInitTemplateSQL)
+```
+
+"List of SQL queries to be executed as a superuser in the `template1` after the cluster has been created - to be used with extreme care (by default empty)"
+
+### fn spec.bootstrap.initdb.withPostInitTemplateSQLMixin
+
+```ts
+withPostInitTemplateSQLMixin(postInitTemplateSQL)
+```
+
+"List of SQL queries to be executed as a superuser in the `template1` after the cluster has been created - to be used with extreme care (by default empty)"
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.bootstrap.initdb.withWalSegmentSize
+
+```ts
+withWalSegmentSize(walSegmentSize)
+```
+
+"The value in megabytes (1 to 1024) to be passed to the `--wal-segsize` option for initdb (default: empty, resulting in PostgreSQL default: 16MB)"
+
+## obj spec.bootstrap.initdb.import
+
+"Bootstraps the new cluster by importing data from an existing PostgreSQL instance using logical backup (`pg_dump` and `pg_restore`)"
+
+### fn spec.bootstrap.initdb.import.withDatabases
+
+```ts
+withDatabases(databases)
+```
+
+"The databases to import"
+
+### fn spec.bootstrap.initdb.import.withDatabasesMixin
+
+```ts
+withDatabasesMixin(databases)
+```
+
+"The databases to import"
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.bootstrap.initdb.import.withPostImportApplicationSQL
+
+```ts
+withPostImportApplicationSQL(postImportApplicationSQL)
+```
+
+"List of SQL queries to be executed as a superuser in the application database right after is imported - to be used with extreme care (by default empty). Only available in microservice type."
+
+### fn spec.bootstrap.initdb.import.withPostImportApplicationSQLMixin
+
+```ts
+withPostImportApplicationSQLMixin(postImportApplicationSQL)
+```
+
+"List of SQL queries to be executed as a superuser in the application database right after is imported - to be used with extreme care (by default empty). Only available in microservice type."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.bootstrap.initdb.import.withRoles
+
+```ts
+withRoles(roles)
+```
+
+"The roles to import"
+
+### fn spec.bootstrap.initdb.import.withRolesMixin
+
+```ts
+withRolesMixin(roles)
+```
+
+"The roles to import"
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.bootstrap.initdb.import.withSchemaOnly
+
+```ts
+withSchemaOnly(schemaOnly)
+```
+
+"When set to true, only the `pre-data` and `post-data` sections of `pg_restore` are invoked, avoiding data import. Default: `false`."
+
+### fn spec.bootstrap.initdb.import.withType
+
+```ts
+withType(type)
+```
+
+"The import type. Can be `microservice` or `monolith`."
+
+## obj spec.bootstrap.initdb.import.source
+
+"The source of the import"
+
+### fn spec.bootstrap.initdb.import.source.withExternalCluster
+
+```ts
+withExternalCluster(externalCluster)
+```
+
+"The name of the externalCluster used for import"
+
+## obj spec.bootstrap.initdb.postInitApplicationSQLRefs
+
+"PostInitApplicationSQLRefs points references to ConfigMaps or Secrets which contain SQL files, the general implementation order to these references is from all Secrets to all ConfigMaps, and inside Secrets or ConfigMaps, the implementation order is same as the order of each array (by default empty)"
+
+### fn spec.bootstrap.initdb.postInitApplicationSQLRefs.withConfigMapRefs
+
+```ts
+withConfigMapRefs(configMapRefs)
+```
+
+"ConfigMapRefs holds a list of references to ConfigMaps"
+
+### fn spec.bootstrap.initdb.postInitApplicationSQLRefs.withConfigMapRefsMixin
+
+```ts
+withConfigMapRefsMixin(configMapRefs)
+```
+
+"ConfigMapRefs holds a list of references to ConfigMaps"
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.bootstrap.initdb.postInitApplicationSQLRefs.withSecretRefs
+
+```ts
+withSecretRefs(secretRefs)
+```
+
+"SecretRefs holds a list of references to Secrets"
+
+### fn spec.bootstrap.initdb.postInitApplicationSQLRefs.withSecretRefsMixin
+
+```ts
+withSecretRefsMixin(secretRefs)
+```
+
+"SecretRefs holds a list of references to Secrets"
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.bootstrap.initdb.postInitApplicationSQLRefs.configMapRefs
+
+"ConfigMapRefs holds a list of references to ConfigMaps"
+
+### fn spec.bootstrap.initdb.postInitApplicationSQLRefs.configMapRefs.withKey
+
+```ts
+withKey(key)
+```
+
+"The key to select"
+
+### fn spec.bootstrap.initdb.postInitApplicationSQLRefs.configMapRefs.withName
+
+```ts
+withName(name)
+```
+
+"Name of the referent."
+
+## obj spec.bootstrap.initdb.postInitApplicationSQLRefs.secretRefs
+
+"SecretRefs holds a list of references to Secrets"
+
+### fn spec.bootstrap.initdb.postInitApplicationSQLRefs.secretRefs.withKey
+
+```ts
+withKey(key)
+```
+
+"The key to select"
+
+### fn spec.bootstrap.initdb.postInitApplicationSQLRefs.secretRefs.withName
+
+```ts
+withName(name)
+```
+
+"Name of the referent."
+
+## obj spec.bootstrap.initdb.secret
+
+"Name of the secret containing the initial credentials for the owner of the user database. If empty a new secret will be created from scratch"
+
+### fn spec.bootstrap.initdb.secret.withName
+
+```ts
+withName(name)
+```
+
+"Name of the referent."
+
+## obj spec.bootstrap.pg_basebackup
+
+"Bootstrap the cluster taking a physical backup of another compatible PostgreSQL instance"
+
+### fn spec.bootstrap.pg_basebackup.withDatabase
+
+```ts
+withDatabase(database)
+```
+
+"Name of the database used by the application. Default: `app`."
+
+### fn spec.bootstrap.pg_basebackup.withOwner
+
+```ts
+withOwner(owner)
+```
+
+"Name of the owner of the database in the instance to be used by applications. Defaults to the value of the `database` key."
+
+### fn spec.bootstrap.pg_basebackup.withSource
+
+```ts
+withSource(source)
+```
+
+"The name of the server of which we need to take a physical backup"
+
+## obj spec.bootstrap.pg_basebackup.secret
+
+"Name of the secret containing the initial credentials for the owner of the user database. If empty a new secret will be created from scratch"
+
+### fn spec.bootstrap.pg_basebackup.secret.withName
+
+```ts
+withName(name)
+```
+
+"Name of the referent."
+
+## obj spec.bootstrap.recovery
+
+"Bootstrap the cluster from a backup"
+
+### fn spec.bootstrap.recovery.withDatabase
+
+```ts
+withDatabase(database)
+```
+
+"Name of the database used by the application. Default: `app`."
+
+### fn spec.bootstrap.recovery.withOwner
+
+```ts
+withOwner(owner)
+```
+
+"Name of the owner of the database in the instance to be used by applications. Defaults to the value of the `database` key."
+
+### fn spec.bootstrap.recovery.withSource
+
+```ts
+withSource(source)
+```
+
+"The external cluster whose backup we will restore. This is also used as the name of the folder under which the backup is stored, so it must be set to the name of the source cluster Mutually exclusive with `backup`."
+
+## obj spec.bootstrap.recovery.backup
+
+"The backup object containing the physical base backup from which to initiate the recovery procedure. Mutually exclusive with `source` and `volumeSnapshots`."
+
+### fn spec.bootstrap.recovery.backup.withName
+
+```ts
+withName(name)
+```
+
+"Name of the referent."
+
+## obj spec.bootstrap.recovery.backup.endpointCA
+
+"EndpointCA store the CA bundle of the barman endpoint. Useful when using self-signed certificates to avoid errors with certificate issuer and barman-cloud-wal-archive."
+
+### fn spec.bootstrap.recovery.backup.endpointCA.withKey
+
+```ts
+withKey(key)
+```
+
+"The key to select"
+
+### fn spec.bootstrap.recovery.backup.endpointCA.withName
+
+```ts
+withName(name)
+```
+
+"Name of the referent."
+
+## obj spec.bootstrap.recovery.recoveryTarget
+
+"By default, the recovery process applies all the available WAL files in the archive (full recovery). However, you can also end the recovery as soon as a consistent state is reached or recover to a point-in-time (PITR) by specifying a `RecoveryTarget` object, as expected by PostgreSQL (i.e., timestamp, transaction Id, LSN, ...). More info: https://www.postgresql.org/docs/current/runtime-config-wal.html#RUNTIME-CONFIG-WAL-RECOVERY-TARGET"
+
+### fn spec.bootstrap.recovery.recoveryTarget.withBackupID
+
+```ts
+withBackupID(backupID)
+```
+
+"The ID of the backup from which to start the recovery process. If empty (default) the operator will automatically detect the backup based on targetTime or targetLSN if specified. Otherwise use the latest available backup in chronological order."
+
+### fn spec.bootstrap.recovery.recoveryTarget.withExclusive
+
+```ts
+withExclusive(exclusive)
+```
+
+"Set the target to be exclusive. If omitted, defaults to false, so that in Postgres, `recovery_target_inclusive` will be true"
+
+### fn spec.bootstrap.recovery.recoveryTarget.withTargetImmediate
+
+```ts
+withTargetImmediate(targetImmediate)
+```
+
+"End recovery as soon as a consistent state is reached"
+
+### fn spec.bootstrap.recovery.recoveryTarget.withTargetLSN
+
+```ts
+withTargetLSN(targetLSN)
+```
+
+"The target LSN (Log Sequence Number)"
+
+### fn spec.bootstrap.recovery.recoveryTarget.withTargetName
+
+```ts
+withTargetName(targetName)
+```
+
+"The target name (to be previously created with `pg_create_restore_point`)"
+
+### fn spec.bootstrap.recovery.recoveryTarget.withTargetTLI
+
+```ts
+withTargetTLI(targetTLI)
+```
+
+"The target timeline (\"latest\" or a positive integer)"
+
+### fn spec.bootstrap.recovery.recoveryTarget.withTargetTime
+
+```ts
+withTargetTime(targetTime)
+```
+
+"The target time as a timestamp in the RFC3339 standard"
+
+### fn spec.bootstrap.recovery.recoveryTarget.withTargetXID
+
+```ts
+withTargetXID(targetXID)
+```
+
+"The target transaction ID"
+
+## obj spec.bootstrap.recovery.secret
+
+"Name of the secret containing the initial credentials for the owner of the user database. If empty a new secret will be created from scratch"
+
+### fn spec.bootstrap.recovery.secret.withName
+
+```ts
+withName(name)
+```
+
+"Name of the referent."
+
+## obj spec.bootstrap.recovery.volumeSnapshots
+
+"The static PVC data source(s) from which to initiate the recovery procedure. Currently supporting `VolumeSnapshot` and `PersistentVolumeClaim` resources that map an existing PVC group, compatible with CloudNativePG, and taken with a cold backup copy on a fenced Postgres instance (limitation which will be removed in the future when online backup will be implemented). Mutually exclusive with `backup`."
+
+## obj spec.bootstrap.recovery.volumeSnapshots.storage
+
+"Configuration of the storage of the instances"
+
+### fn spec.bootstrap.recovery.volumeSnapshots.storage.withApiGroup
+
+```ts
+withApiGroup(apiGroup)
+```
+
+"APIGroup is the group for the resource being referenced. If APIGroup is not specified, the specified Kind must be in the core API group. For any other third-party types, APIGroup is required."
+
+### fn spec.bootstrap.recovery.volumeSnapshots.storage.withKind
+
+```ts
+withKind(kind)
+```
+
+"Kind is the type of resource being referenced"
+
+### fn spec.bootstrap.recovery.volumeSnapshots.storage.withName
+
+```ts
+withName(name)
+```
+
+"Name is the name of resource being referenced"
+
+## obj spec.bootstrap.recovery.volumeSnapshots.walStorage
+
+"Configuration of the storage for PostgreSQL WAL (Write-Ahead Log)"
+
+### fn spec.bootstrap.recovery.volumeSnapshots.walStorage.withApiGroup
+
+```ts
+withApiGroup(apiGroup)
+```
+
+"APIGroup is the group for the resource being referenced. If APIGroup is not specified, the specified Kind must be in the core API group. For any other third-party types, APIGroup is required."
+
+### fn spec.bootstrap.recovery.volumeSnapshots.walStorage.withKind
+
+```ts
+withKind(kind)
+```
+
+"Kind is the type of resource being referenced"
+
+### fn spec.bootstrap.recovery.volumeSnapshots.walStorage.withName
+
+```ts
+withName(name)
+```
+
+"Name is the name of resource being referenced"
+
+## obj spec.certificates
+
+"The configuration for the CA and related certificates"
+
+### fn spec.certificates.withClientCASecret
+
+```ts
+withClientCASecret(clientCASecret)
+```
+
+"The secret containing the Client CA certificate. If not defined, a new secret will be created with a self-signed CA and will be used to generate all the client certificates.
Contains:
- `ca.crt`: CA that should be used to validate the client certificates, used as `ssl_ca_file` of all the instances.
- `ca.key`: key used to generate client certificates, if ReplicationTLSSecret is provided, this can be omitted.
"
+
+### fn spec.certificates.withReplicationTLSSecret
+
+```ts
+withReplicationTLSSecret(replicationTLSSecret)
+```
+
+"The secret of type kubernetes.io/tls containing the client certificate to authenticate as the `streaming_replica` user. If not defined, ClientCASecret must provide also `ca.key`, and a new secret will be created using the provided CA."
+
+### fn spec.certificates.withServerAltDNSNames
+
+```ts
+withServerAltDNSNames(serverAltDNSNames)
+```
+
+"The list of the server alternative DNS names to be added to the generated server TLS certificates, when required."
+
+### fn spec.certificates.withServerAltDNSNamesMixin
+
+```ts
+withServerAltDNSNamesMixin(serverAltDNSNames)
+```
+
+"The list of the server alternative DNS names to be added to the generated server TLS certificates, when required."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.certificates.withServerCASecret
+
+```ts
+withServerCASecret(serverCASecret)
+```
+
+"The secret containing the Server CA certificate. If not defined, a new secret will be created with a self-signed CA and will be used to generate the TLS certificate ServerTLSSecret.
Contains:
- `ca.crt`: CA that should be used to validate the server certificate, used as `sslrootcert` in client connection strings.
- `ca.key`: key used to generate Server SSL certs, if ServerTLSSecret is provided, this can be omitted.
"
+
+### fn spec.certificates.withServerTLSSecret
+
+```ts
+withServerTLSSecret(serverTLSSecret)
+```
+
+"The secret of type kubernetes.io/tls containing the server TLS certificate and key that will be set as `ssl_cert_file` and `ssl_key_file` so that clients can connect to postgres securely. If not defined, ServerCASecret must provide also `ca.key` and a new secret will be created using the provided CA."
+
+## obj spec.env
+
+"Env follows the Env format to pass environment variables to the pods created in the cluster"
+
+### fn spec.env.withName
+
+```ts
+withName(name)
+```
+
+"Name of the environment variable. Must be a C_IDENTIFIER."
+
+### fn spec.env.withValue
+
+```ts
+withValue(value)
+```
+
+"Variable references $(VAR_NAME) are expanded using the previously defined environment variables in the container and any service environment variables. If a variable cannot be resolved, the reference in the input string will be unchanged. Double $$ are reduced to a single $, which allows for escaping the $(VAR_NAME) syntax: i.e. \"$$(VAR_NAME)\" will produce the string literal \"$(VAR_NAME)\". Escaped references will never be expanded, regardless of whether the variable exists or not. Defaults to \"\"."
+
+## obj spec.env.valueFrom
+
+"Source for the environment variable's value. Cannot be used if value is not empty."
+
+## obj spec.env.valueFrom.configMapKeyRef
+
+"Selects a key of a ConfigMap."
+
+### fn spec.env.valueFrom.configMapKeyRef.withKey
+
+```ts
+withKey(key)
+```
+
+"The key to select."
+
+### fn spec.env.valueFrom.configMapKeyRef.withName
+
+```ts
+withName(name)
+```
+
+"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"
+
+### fn spec.env.valueFrom.configMapKeyRef.withOptional
+
+```ts
+withOptional(optional)
+```
+
+"Specify whether the ConfigMap or its key must be defined"
+
+## obj spec.env.valueFrom.fieldRef
+
+"Selects a field of the pod: supports metadata.name, metadata.namespace, `metadata.labels['']`, `metadata.annotations['']`, spec.nodeName, spec.serviceAccountName, status.hostIP, status.podIP, status.podIPs."
+
+### fn spec.env.valueFrom.fieldRef.withApiVersion
+
+```ts
+withApiVersion(apiVersion)
+```
+
+"Version of the schema the FieldPath is written in terms of, defaults to \"v1\"."
+
+### fn spec.env.valueFrom.fieldRef.withFieldPath
+
+```ts
+withFieldPath(fieldPath)
+```
+
+"Path of the field to select in the specified API version."
+
+## obj spec.env.valueFrom.resourceFieldRef
+
+"Selects a resource of the container: only resources limits and requests (limits.cpu, limits.memory, limits.ephemeral-storage, requests.cpu, requests.memory and requests.ephemeral-storage) are currently supported."
+
+### fn spec.env.valueFrom.resourceFieldRef.withContainerName
+
+```ts
+withContainerName(containerName)
+```
+
+"Container name: required for volumes, optional for env vars"
+
+### fn spec.env.valueFrom.resourceFieldRef.withDivisor
+
+```ts
+withDivisor(divisor)
+```
+
+"Specifies the output format of the exposed resources, defaults to \"1\
+
+### fn spec.env.valueFrom.resourceFieldRef.withResource
+
+```ts
+withResource(resource)
+```
+
+"Required: resource to select"
+
+## obj spec.env.valueFrom.secretKeyRef
+
+"Selects a key of a secret in the pod's namespace"
+
+### fn spec.env.valueFrom.secretKeyRef.withKey
+
+```ts
+withKey(key)
+```
+
+"The key of the secret to select from. Must be a valid secret key."
+
+### fn spec.env.valueFrom.secretKeyRef.withName
+
+```ts
+withName(name)
+```
+
+"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"
+
+### fn spec.env.valueFrom.secretKeyRef.withOptional
+
+```ts
+withOptional(optional)
+```
+
+"Specify whether the Secret or its key must be defined"
+
+## obj spec.envFrom
+
+"EnvFrom follows the EnvFrom format to pass environment variables sources to the pods to be used by Env"
+
+### fn spec.envFrom.withPrefix
+
+```ts
+withPrefix(prefix)
+```
+
+"An optional identifier to prepend to each key in the ConfigMap. Must be a C_IDENTIFIER."
+
+## obj spec.envFrom.configMapRef
+
+"The ConfigMap to select from"
+
+### fn spec.envFrom.configMapRef.withName
+
+```ts
+withName(name)
+```
+
+"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"
+
+### fn spec.envFrom.configMapRef.withOptional
+
+```ts
+withOptional(optional)
+```
+
+"Specify whether the ConfigMap must be defined"
+
+## obj spec.envFrom.secretRef
+
+"The Secret to select from"
+
+### fn spec.envFrom.secretRef.withName
+
+```ts
+withName(name)
+```
+
+"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"
+
+### fn spec.envFrom.secretRef.withOptional
+
+```ts
+withOptional(optional)
+```
+
+"Specify whether the Secret must be defined"
+
+## obj spec.ephemeralVolumesSizeLimit
+
+"EphemeralVolumesSizeLimit allows the user to set the limits for the ephemeral volumes"
+
+### fn spec.ephemeralVolumesSizeLimit.withShm
+
+```ts
+withShm(shm)
+```
+
+"Shm is the size limit of the shared memory volume"
+
+### fn spec.ephemeralVolumesSizeLimit.withTemporaryData
+
+```ts
+withTemporaryData(temporaryData)
+```
+
+"TemporaryData is the size limit of the temporary data volume"
+
+## obj spec.externalClusters
+
+"The list of external clusters which are used in the configuration"
+
+### fn spec.externalClusters.withConnectionParameters
+
+```ts
+withConnectionParameters(connectionParameters)
+```
+
+"The list of connection parameters, such as dbname, host, username, etc"
+
+### fn spec.externalClusters.withConnectionParametersMixin
+
+```ts
+withConnectionParametersMixin(connectionParameters)
+```
+
+"The list of connection parameters, such as dbname, host, username, etc"
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.externalClusters.withName
+
+```ts
+withName(name)
+```
+
+"The server name, required"
+
+## obj spec.externalClusters.barmanObjectStore
+
+"The configuration for the barman-cloud tool suite"
+
+### fn spec.externalClusters.barmanObjectStore.withDestinationPath
+
+```ts
+withDestinationPath(destinationPath)
+```
+
+"The path where to store the backup (i.e. s3://bucket/path/to/folder) this path, with different destination folders, will be used for WALs and for data"
+
+### fn spec.externalClusters.barmanObjectStore.withEndpointURL
+
+```ts
+withEndpointURL(endpointURL)
+```
+
+"Endpoint to be used to upload data to the cloud, overriding the automatic endpoint discovery"
+
+### fn spec.externalClusters.barmanObjectStore.withHistoryTags
+
+```ts
+withHistoryTags(historyTags)
+```
+
+"HistoryTags is a list of key value pairs that will be passed to the Barman --history-tags option."
+
+### fn spec.externalClusters.barmanObjectStore.withHistoryTagsMixin
+
+```ts
+withHistoryTagsMixin(historyTags)
+```
+
+"HistoryTags is a list of key value pairs that will be passed to the Barman --history-tags option."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.externalClusters.barmanObjectStore.withServerName
+
+```ts
+withServerName(serverName)
+```
+
+"The server name on S3, the cluster name is used if this parameter is omitted"
+
+### fn spec.externalClusters.barmanObjectStore.withTags
+
+```ts
+withTags(tags)
+```
+
+"Tags is a list of key value pairs that will be passed to the Barman --tags option."
+
+### fn spec.externalClusters.barmanObjectStore.withTagsMixin
+
+```ts
+withTagsMixin(tags)
+```
+
+"Tags is a list of key value pairs that will be passed to the Barman --tags option."
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.externalClusters.barmanObjectStore.azureCredentials
+
+"The credentials to use to upload data to Azure Blob Storage"
+
+### fn spec.externalClusters.barmanObjectStore.azureCredentials.withInheritFromAzureAD
+
+```ts
+withInheritFromAzureAD(inheritFromAzureAD)
+```
+
+"Use the Azure AD based authentication without providing explicitly the keys."
+
+## obj spec.externalClusters.barmanObjectStore.azureCredentials.connectionString
+
+"The connection string to be used"
+
+### fn spec.externalClusters.barmanObjectStore.azureCredentials.connectionString.withKey
+
+```ts
+withKey(key)
+```
+
+"The key to select"
+
+### fn spec.externalClusters.barmanObjectStore.azureCredentials.connectionString.withName
+
+```ts
+withName(name)
+```
+
+"Name of the referent."
+
+## obj spec.externalClusters.barmanObjectStore.azureCredentials.storageAccount
+
+"The storage account where to upload data"
+
+### fn spec.externalClusters.barmanObjectStore.azureCredentials.storageAccount.withKey
+
+```ts
+withKey(key)
+```
+
+"The key to select"
+
+### fn spec.externalClusters.barmanObjectStore.azureCredentials.storageAccount.withName
+
+```ts
+withName(name)
+```
+
+"Name of the referent."
+
+## obj spec.externalClusters.barmanObjectStore.azureCredentials.storageKey
+
+"The storage account key to be used in conjunction with the storage account name"
+
+### fn spec.externalClusters.barmanObjectStore.azureCredentials.storageKey.withKey
+
+```ts
+withKey(key)
+```
+
+"The key to select"
+
+### fn spec.externalClusters.barmanObjectStore.azureCredentials.storageKey.withName
+
+```ts
+withName(name)
+```
+
+"Name of the referent."
+
+## obj spec.externalClusters.barmanObjectStore.azureCredentials.storageSasToken
+
+"A shared-access-signature to be used in conjunction with the storage account name"
+
+### fn spec.externalClusters.barmanObjectStore.azureCredentials.storageSasToken.withKey
+
+```ts
+withKey(key)
+```
+
+"The key to select"
+
+### fn spec.externalClusters.barmanObjectStore.azureCredentials.storageSasToken.withName
+
+```ts
+withName(name)
+```
+
+"Name of the referent."
+
+## obj spec.externalClusters.barmanObjectStore.data
+
+"The configuration to be used to backup the data files When not defined, base backups files will be stored uncompressed and may be unencrypted in the object store, according to the bucket default policy."
+
+### fn spec.externalClusters.barmanObjectStore.data.withCompression
+
+```ts
+withCompression(compression)
+```
+
+"Compress a backup file (a tar file per tablespace) while streaming it to the object store. Available options are empty string (no compression, default), `gzip`, `bzip2` or `snappy`."
+
+### fn spec.externalClusters.barmanObjectStore.data.withEncryption
+
+```ts
+withEncryption(encryption)
+```
+
+"Whenever to force the encryption of files (if the bucket is not already configured for that). Allowed options are empty string (use the bucket policy, default), `AES256` and `aws:kms`"
+
+### fn spec.externalClusters.barmanObjectStore.data.withImmediateCheckpoint
+
+```ts
+withImmediateCheckpoint(immediateCheckpoint)
+```
+
+"Control whether the I/O workload for the backup initial checkpoint will be limited, according to the `checkpoint_completion_target` setting on the PostgreSQL server. If set to true, an immediate checkpoint will be used, meaning PostgreSQL will complete the checkpoint as soon as possible. `false` by default."
+
+### fn spec.externalClusters.barmanObjectStore.data.withJobs
+
+```ts
+withJobs(jobs)
+```
+
+"The number of parallel jobs to be used to upload the backup, defaults to 2"
+
+## obj spec.externalClusters.barmanObjectStore.endpointCA
+
+"EndpointCA store the CA bundle of the barman endpoint. Useful when using self-signed certificates to avoid errors with certificate issuer and barman-cloud-wal-archive"
+
+### fn spec.externalClusters.barmanObjectStore.endpointCA.withKey
+
+```ts
+withKey(key)
+```
+
+"The key to select"
+
+### fn spec.externalClusters.barmanObjectStore.endpointCA.withName
+
+```ts
+withName(name)
+```
+
+"Name of the referent."
+
+## obj spec.externalClusters.barmanObjectStore.googleCredentials
+
+"The credentials to use to upload data to Google Cloud Storage"
+
+### fn spec.externalClusters.barmanObjectStore.googleCredentials.withGkeEnvironment
+
+```ts
+withGkeEnvironment(gkeEnvironment)
+```
+
+"If set to true, will presume that it's running inside a GKE environment, default to false."
+
+## obj spec.externalClusters.barmanObjectStore.googleCredentials.applicationCredentials
+
+"The secret containing the Google Cloud Storage JSON file with the credentials"
+
+### fn spec.externalClusters.barmanObjectStore.googleCredentials.applicationCredentials.withKey
+
+```ts
+withKey(key)
+```
+
+"The key to select"
+
+### fn spec.externalClusters.barmanObjectStore.googleCredentials.applicationCredentials.withName
+
+```ts
+withName(name)
+```
+
+"Name of the referent."
+
+## obj spec.externalClusters.barmanObjectStore.s3Credentials
+
+"The credentials to use to upload data to S3"
+
+### fn spec.externalClusters.barmanObjectStore.s3Credentials.withInheritFromIAMRole
+
+```ts
+withInheritFromIAMRole(inheritFromIAMRole)
+```
+
+"Use the role based authentication without providing explicitly the keys."
+
+## obj spec.externalClusters.barmanObjectStore.s3Credentials.accessKeyId
+
+"The reference to the access key id"
+
+### fn spec.externalClusters.barmanObjectStore.s3Credentials.accessKeyId.withKey
+
+```ts
+withKey(key)
+```
+
+"The key to select"
+
+### fn spec.externalClusters.barmanObjectStore.s3Credentials.accessKeyId.withName
+
+```ts
+withName(name)
+```
+
+"Name of the referent."
+
+## obj spec.externalClusters.barmanObjectStore.s3Credentials.region
+
+"The reference to the secret containing the region name"
+
+### fn spec.externalClusters.barmanObjectStore.s3Credentials.region.withKey
+
+```ts
+withKey(key)
+```
+
+"The key to select"
+
+### fn spec.externalClusters.barmanObjectStore.s3Credentials.region.withName
+
+```ts
+withName(name)
+```
+
+"Name of the referent."
+
+## obj spec.externalClusters.barmanObjectStore.s3Credentials.secretAccessKey
+
+"The reference to the secret access key"
+
+### fn spec.externalClusters.barmanObjectStore.s3Credentials.secretAccessKey.withKey
+
+```ts
+withKey(key)
+```
+
+"The key to select"
+
+### fn spec.externalClusters.barmanObjectStore.s3Credentials.secretAccessKey.withName
+
+```ts
+withName(name)
+```
+
+"Name of the referent."
+
+## obj spec.externalClusters.barmanObjectStore.s3Credentials.sessionToken
+
+"The references to the session key"
+
+### fn spec.externalClusters.barmanObjectStore.s3Credentials.sessionToken.withKey
+
+```ts
+withKey(key)
+```
+
+"The key to select"
+
+### fn spec.externalClusters.barmanObjectStore.s3Credentials.sessionToken.withName
+
+```ts
+withName(name)
+```
+
+"Name of the referent."
+
+## obj spec.externalClusters.barmanObjectStore.wal
+
+"The configuration for the backup of the WAL stream. When not defined, WAL files will be stored uncompressed and may be unencrypted in the object store, according to the bucket default policy."
+
+### fn spec.externalClusters.barmanObjectStore.wal.withCompression
+
+```ts
+withCompression(compression)
+```
+
+"Compress a WAL file before sending it to the object store. Available options are empty string (no compression, default), `gzip`, `bzip2` or `snappy`."
+
+### fn spec.externalClusters.barmanObjectStore.wal.withEncryption
+
+```ts
+withEncryption(encryption)
+```
+
+"Whenever to force the encryption of files (if the bucket is not already configured for that). Allowed options are empty string (use the bucket policy, default), `AES256` and `aws:kms`"
+
+### fn spec.externalClusters.barmanObjectStore.wal.withMaxParallel
+
+```ts
+withMaxParallel(maxParallel)
+```
+
+"Number of WAL files to be either archived in parallel (when the PostgreSQL instance is archiving to a backup object store) or restored in parallel (when a PostgreSQL standby is fetching WAL files from a recovery object store). If not specified, WAL files will be processed one at a time. It accepts a positive integer as a value - with 1 being the minimum accepted value."
+
+## obj spec.externalClusters.password
+
+"The reference to the password to be used to connect to the server"
+
+### fn spec.externalClusters.password.withKey
+
+```ts
+withKey(key)
+```
+
+"The key of the secret to select from. Must be a valid secret key."
+
+### fn spec.externalClusters.password.withName
+
+```ts
+withName(name)
+```
+
+"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"
+
+### fn spec.externalClusters.password.withOptional
+
+```ts
+withOptional(optional)
+```
+
+"Specify whether the Secret or its key must be defined"
+
+## obj spec.externalClusters.sslCert
+
+"The reference to an SSL certificate to be used to connect to this instance"
+
+### fn spec.externalClusters.sslCert.withKey
+
+```ts
+withKey(key)
+```
+
+"The key of the secret to select from. Must be a valid secret key."
+
+### fn spec.externalClusters.sslCert.withName
+
+```ts
+withName(name)
+```
+
+"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"
+
+### fn spec.externalClusters.sslCert.withOptional
+
+```ts
+withOptional(optional)
+```
+
+"Specify whether the Secret or its key must be defined"
+
+## obj spec.externalClusters.sslKey
+
+"The reference to an SSL private key to be used to connect to this instance"
+
+### fn spec.externalClusters.sslKey.withKey
+
+```ts
+withKey(key)
+```
+
+"The key of the secret to select from. Must be a valid secret key."
+
+### fn spec.externalClusters.sslKey.withName
+
+```ts
+withName(name)
+```
+
+"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"
+
+### fn spec.externalClusters.sslKey.withOptional
+
+```ts
+withOptional(optional)
+```
+
+"Specify whether the Secret or its key must be defined"
+
+## obj spec.externalClusters.sslRootCert
+
+"The reference to an SSL CA public key to be used to connect to this instance"
+
+### fn spec.externalClusters.sslRootCert.withKey
+
+```ts
+withKey(key)
+```
+
+"The key of the secret to select from. Must be a valid secret key."
+
+### fn spec.externalClusters.sslRootCert.withName
+
+```ts
+withName(name)
+```
+
+"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"
+
+### fn spec.externalClusters.sslRootCert.withOptional
+
+```ts
+withOptional(optional)
+```
+
+"Specify whether the Secret or its key must be defined"
+
+## obj spec.imagePullSecrets
+
+"The list of pull secrets to be used to pull the images"
+
+### fn spec.imagePullSecrets.withName
+
+```ts
+withName(name)
+```
+
+"Name of the referent."
+
+## obj spec.inheritedMetadata
+
+"Metadata that will be inherited by all objects related to the Cluster"
+
+### fn spec.inheritedMetadata.withAnnotations
+
+```ts
+withAnnotations(annotations)
+```
+
+
+
+### fn spec.inheritedMetadata.withAnnotationsMixin
+
+```ts
+withAnnotationsMixin(annotations)
+```
+
+
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.inheritedMetadata.withLabels
+
+```ts
+withLabels(labels)
+```
+
+
+
+### fn spec.inheritedMetadata.withLabelsMixin
+
+```ts
+withLabelsMixin(labels)
+```
+
+
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.managed
+
+"The configuration that is used by the portions of PostgreSQL that are managed by the instance manager"
+
+### fn spec.managed.withRoles
+
+```ts
+withRoles(roles)
+```
+
+"Database roles managed by the `Cluster`"
+
+### fn spec.managed.withRolesMixin
+
+```ts
+withRolesMixin(roles)
+```
+
+"Database roles managed by the `Cluster`"
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.managed.roles
+
+"Database roles managed by the `Cluster`"
+
+### fn spec.managed.roles.withBypassrls
+
+```ts
+withBypassrls(bypassrls)
+```
+
+"Whether a role bypasses every row-level security (RLS) policy. Default is `false`."
+
+### fn spec.managed.roles.withComment
+
+```ts
+withComment(comment)
+```
+
+"Description of the role"
+
+### fn spec.managed.roles.withConnectionLimit
+
+```ts
+withConnectionLimit(connectionLimit)
+```
+
+"If the role can log in, this specifies how many concurrent connections the role can make. `-1` (the default) means no limit."
+
+### fn spec.managed.roles.withCreatedb
+
+```ts
+withCreatedb(createdb)
+```
+
+"When set to `true`, the role being defined will be allowed to create new databases. Specifying `false` (default) will deny a role the ability to create databases."
+
+### fn spec.managed.roles.withCreaterole
+
+```ts
+withCreaterole(createrole)
+```
+
+"Whether the role will be permitted to create, alter, drop, comment on, change the security label for, and grant or revoke membership in other roles. Default is `false`."
+
+### fn spec.managed.roles.withDisablePassword
+
+```ts
+withDisablePassword(disablePassword)
+```
+
+"DisablePassword indicates that a role's password should be set to NULL in Postgres"
+
+### fn spec.managed.roles.withEnsure
+
+```ts
+withEnsure(ensure)
+```
+
+"Ensure the role is `present` or `absent` - defaults to \"present\
+
+### fn spec.managed.roles.withInRoles
+
+```ts
+withInRoles(inRoles)
+```
+
+"List of one or more existing roles to which this role will be immediately added as a new member. Default empty."
+
+### fn spec.managed.roles.withInRolesMixin
+
+```ts
+withInRolesMixin(inRoles)
+```
+
+"List of one or more existing roles to which this role will be immediately added as a new member. Default empty."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.managed.roles.withInherit
+
+```ts
+withInherit(inherit)
+```
+
+"Whether a role \"inherits\" the privileges of roles it is a member of. Defaults is `true`."
+
+### fn spec.managed.roles.withLogin
+
+```ts
+withLogin(login)
+```
+
+"Whether the role is allowed to log in. A role having the `login` attribute can be thought of as a user. Roles without this attribute are useful for managing database privileges, but are not users in the usual sense of the word. Default is `false`."
+
+### fn spec.managed.roles.withName
+
+```ts
+withName(name)
+```
+
+"Name of the role"
+
+### fn spec.managed.roles.withReplication
+
+```ts
+withReplication(replication)
+```
+
+"Whether a role is a replication role. A role must have this attribute (or be a superuser) in order to be able to connect to the server in replication mode (physical or logical replication) and in order to be able to create or drop replication slots. A role having the `replication` attribute is a very highly privileged role, and should only be used on roles actually used for replication. Default is `false`."
+
+### fn spec.managed.roles.withSuperuser
+
+```ts
+withSuperuser(superuser)
+```
+
+"Whether the role is a `superuser` who can override all access restrictions within the database - superuser status is dangerous and should be used only when really needed. You must yourself be a superuser to create a new superuser. Defaults is `false`."
+
+### fn spec.managed.roles.withValidUntil
+
+```ts
+withValidUntil(validUntil)
+```
+
+"Date and time after which the role's password is no longer valid. When omitted, the password will never expire (default)."
+
+## obj spec.managed.roles.passwordSecret
+
+"Secret containing the password of the role (if present) If null, the password will be ignored unless DisablePassword is set"
+
+### fn spec.managed.roles.passwordSecret.withName
+
+```ts
+withName(name)
+```
+
+"Name of the referent."
+
+## obj spec.monitoring
+
+"The configuration of the monitoring infrastructure of this cluster"
+
+### fn spec.monitoring.withCustomQueriesConfigMap
+
+```ts
+withCustomQueriesConfigMap(customQueriesConfigMap)
+```
+
+"The list of config maps containing the custom queries"
+
+### fn spec.monitoring.withCustomQueriesConfigMapMixin
+
+```ts
+withCustomQueriesConfigMapMixin(customQueriesConfigMap)
+```
+
+"The list of config maps containing the custom queries"
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.monitoring.withCustomQueriesSecret
+
+```ts
+withCustomQueriesSecret(customQueriesSecret)
+```
+
+"The list of secrets containing the custom queries"
+
+### fn spec.monitoring.withCustomQueriesSecretMixin
+
+```ts
+withCustomQueriesSecretMixin(customQueriesSecret)
+```
+
+"The list of secrets containing the custom queries"
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.monitoring.withDisableDefaultQueries
+
+```ts
+withDisableDefaultQueries(disableDefaultQueries)
+```
+
+"Whether the default queries should be injected. Set it to `true` if you don't want to inject default queries into the cluster. Default: false."
+
+### fn spec.monitoring.withEnablePodMonitor
+
+```ts
+withEnablePodMonitor(enablePodMonitor)
+```
+
+"Enable or disable the `PodMonitor`"
+
+## obj spec.monitoring.customQueriesConfigMap
+
+"The list of config maps containing the custom queries"
+
+### fn spec.monitoring.customQueriesConfigMap.withKey
+
+```ts
+withKey(key)
+```
+
+"The key to select"
+
+### fn spec.monitoring.customQueriesConfigMap.withName
+
+```ts
+withName(name)
+```
+
+"Name of the referent."
+
+## obj spec.monitoring.customQueriesSecret
+
+"The list of secrets containing the custom queries"
+
+### fn spec.monitoring.customQueriesSecret.withKey
+
+```ts
+withKey(key)
+```
+
+"The key to select"
+
+### fn spec.monitoring.customQueriesSecret.withName
+
+```ts
+withName(name)
+```
+
+"Name of the referent."
+
+## obj spec.nodeMaintenanceWindow
+
+"Define a maintenance window for the Kubernetes nodes"
+
+### fn spec.nodeMaintenanceWindow.withInProgress
+
+```ts
+withInProgress(inProgress)
+```
+
+"Is there a node maintenance activity in progress?"
+
+### fn spec.nodeMaintenanceWindow.withReusePVC
+
+```ts
+withReusePVC(reusePVC)
+```
+
+"Reuse the existing PVC (wait for the node to come up again) or not (recreate it elsewhere - when `instances` >1)"
+
+## obj spec.postgresql
+
+"Configuration of the PostgreSQL server"
+
+### fn spec.postgresql.withParameters
+
+```ts
+withParameters(parameters)
+```
+
+"PostgreSQL configuration options (postgresql.conf)"
+
+### fn spec.postgresql.withParametersMixin
+
+```ts
+withParametersMixin(parameters)
+```
+
+"PostgreSQL configuration options (postgresql.conf)"
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.postgresql.withPg_hba
+
+```ts
+withPg_hba(pg_hba)
+```
+
+"PostgreSQL Host Based Authentication rules (lines to be appended to the pg_hba.conf file)"
+
+### fn spec.postgresql.withPg_hbaMixin
+
+```ts
+withPg_hbaMixin(pg_hba)
+```
+
+"PostgreSQL Host Based Authentication rules (lines to be appended to the pg_hba.conf file)"
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.postgresql.withPromotionTimeout
+
+```ts
+withPromotionTimeout(promotionTimeout)
+```
+
+"Specifies the maximum number of seconds to wait when promoting an instance to primary. Default value is 40000000, greater than one year in seconds, big enough to simulate an infinite timeout"
+
+### fn spec.postgresql.withShared_preload_libraries
+
+```ts
+withShared_preload_libraries(shared_preload_libraries)
+```
+
+"Lists of shared preload libraries to add to the default ones"
+
+### fn spec.postgresql.withShared_preload_librariesMixin
+
+```ts
+withShared_preload_librariesMixin(shared_preload_libraries)
+```
+
+"Lists of shared preload libraries to add to the default ones"
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.postgresql.ldap
+
+"Options to specify LDAP configuration"
+
+### fn spec.postgresql.ldap.withPort
+
+```ts
+withPort(port)
+```
+
+"LDAP server port"
+
+### fn spec.postgresql.ldap.withScheme
+
+```ts
+withScheme(scheme)
+```
+
+"LDAP schema to be used, possible options are `ldap` and `ldaps`"
+
+### fn spec.postgresql.ldap.withServer
+
+```ts
+withServer(server)
+```
+
+"LDAP hostname or IP address"
+
+### fn spec.postgresql.ldap.withTls
+
+```ts
+withTls(tls)
+```
+
+"Set to 'true' to enable LDAP over TLS. 'false' is default"
+
+## obj spec.postgresql.ldap.bindAsAuth
+
+"Bind as authentication configuration"
+
+### fn spec.postgresql.ldap.bindAsAuth.withPrefix
+
+```ts
+withPrefix(prefix)
+```
+
+"Prefix for the bind authentication option"
+
+### fn spec.postgresql.ldap.bindAsAuth.withSuffix
+
+```ts
+withSuffix(suffix)
+```
+
+"Suffix for the bind authentication option"
+
+## obj spec.postgresql.ldap.bindSearchAuth
+
+"Bind+Search authentication configuration"
+
+### fn spec.postgresql.ldap.bindSearchAuth.withBaseDN
+
+```ts
+withBaseDN(baseDN)
+```
+
+"Root DN to begin the user search"
+
+### fn spec.postgresql.ldap.bindSearchAuth.withBindDN
+
+```ts
+withBindDN(bindDN)
+```
+
+"DN of the user to bind to the directory"
+
+### fn spec.postgresql.ldap.bindSearchAuth.withSearchAttribute
+
+```ts
+withSearchAttribute(searchAttribute)
+```
+
+"Attribute to match against the username"
+
+### fn spec.postgresql.ldap.bindSearchAuth.withSearchFilter
+
+```ts
+withSearchFilter(searchFilter)
+```
+
+"Search filter to use when doing the search+bind authentication"
+
+## obj spec.postgresql.ldap.bindSearchAuth.bindPassword
+
+"Secret with the password for the user to bind to the directory"
+
+### fn spec.postgresql.ldap.bindSearchAuth.bindPassword.withKey
+
+```ts
+withKey(key)
+```
+
+"The key of the secret to select from. Must be a valid secret key."
+
+### fn spec.postgresql.ldap.bindSearchAuth.bindPassword.withName
+
+```ts
+withName(name)
+```
+
+"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"
+
+### fn spec.postgresql.ldap.bindSearchAuth.bindPassword.withOptional
+
+```ts
+withOptional(optional)
+```
+
+"Specify whether the Secret or its key must be defined"
+
+## obj spec.postgresql.syncReplicaElectionConstraint
+
+"Requirements to be met by sync replicas. This will affect how the \"synchronous_standby_names\" parameter will be set up."
+
+### fn spec.postgresql.syncReplicaElectionConstraint.withEnabled
+
+```ts
+withEnabled(enabled)
+```
+
+"This flag enables the constraints for sync replicas"
+
+### fn spec.postgresql.syncReplicaElectionConstraint.withNodeLabelsAntiAffinity
+
+```ts
+withNodeLabelsAntiAffinity(nodeLabelsAntiAffinity)
+```
+
+"A list of node labels values to extract and compare to evaluate if the pods reside in the same topology or not"
+
+### fn spec.postgresql.syncReplicaElectionConstraint.withNodeLabelsAntiAffinityMixin
+
+```ts
+withNodeLabelsAntiAffinityMixin(nodeLabelsAntiAffinity)
+```
+
+"A list of node labels values to extract and compare to evaluate if the pods reside in the same topology or not"
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.projectedVolumeTemplate
+
+"Template to be used to define projected volumes, projected volumes will be mounted under `/projected` base folder"
+
+### fn spec.projectedVolumeTemplate.withDefaultMode
+
+```ts
+withDefaultMode(defaultMode)
+```
+
+"defaultMode are the mode bits used to set permissions on created files by default. Must be an octal value between 0000 and 0777 or a decimal value between 0 and 511. YAML accepts both octal and decimal values, JSON requires decimal values for mode bits. Directories within the path are not affected by this setting. This might be in conflict with other options that affect the file mode, like fsGroup, and the result can be other mode bits set."
+
+### fn spec.projectedVolumeTemplate.withSources
+
+```ts
+withSources(sources)
+```
+
+"sources is the list of volume projections"
+
+### fn spec.projectedVolumeTemplate.withSourcesMixin
+
+```ts
+withSourcesMixin(sources)
+```
+
+"sources is the list of volume projections"
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.projectedVolumeTemplate.sources
+
+"sources is the list of volume projections"
+
+## obj spec.projectedVolumeTemplate.sources.configMap
+
+"configMap information about the configMap data to project"
+
+### fn spec.projectedVolumeTemplate.sources.configMap.withItems
+
+```ts
+withItems(items)
+```
+
+"items if unspecified, each key-value pair in the Data field of the referenced ConfigMap will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the ConfigMap, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'."
+
+### fn spec.projectedVolumeTemplate.sources.configMap.withItemsMixin
+
+```ts
+withItemsMixin(items)
+```
+
+"items if unspecified, each key-value pair in the Data field of the referenced ConfigMap will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the ConfigMap, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.projectedVolumeTemplate.sources.configMap.withName
+
+```ts
+withName(name)
+```
+
+"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"
+
+### fn spec.projectedVolumeTemplate.sources.configMap.withOptional
+
+```ts
+withOptional(optional)
+```
+
+"optional specify whether the ConfigMap or its keys must be defined"
+
+## obj spec.projectedVolumeTemplate.sources.configMap.items
+
+"items if unspecified, each key-value pair in the Data field of the referenced ConfigMap will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the ConfigMap, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'."
+
+### fn spec.projectedVolumeTemplate.sources.configMap.items.withKey
+
+```ts
+withKey(key)
+```
+
+"key is the key to project."
+
+### fn spec.projectedVolumeTemplate.sources.configMap.items.withMode
+
+```ts
+withMode(mode)
+```
+
+"mode is Optional: mode bits used to set permissions on this file. Must be an octal value between 0000 and 0777 or a decimal value between 0 and 511. YAML accepts both octal and decimal values, JSON requires decimal values for mode bits. If not specified, the volume defaultMode will be used. This might be in conflict with other options that affect the file mode, like fsGroup, and the result can be other mode bits set."
+
+### fn spec.projectedVolumeTemplate.sources.configMap.items.withPath
+
+```ts
+withPath(path)
+```
+
+"path is the relative path of the file to map the key to. May not be an absolute path. May not contain the path element '..'. May not start with the string '..'."
+
+## obj spec.projectedVolumeTemplate.sources.downwardAPI
+
+"downwardAPI information about the downwardAPI data to project"
+
+### fn spec.projectedVolumeTemplate.sources.downwardAPI.withItems
+
+```ts
+withItems(items)
+```
+
+"Items is a list of DownwardAPIVolume file"
+
+### fn spec.projectedVolumeTemplate.sources.downwardAPI.withItemsMixin
+
+```ts
+withItemsMixin(items)
+```
+
+"Items is a list of DownwardAPIVolume file"
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.projectedVolumeTemplate.sources.downwardAPI.items
+
+"Items is a list of DownwardAPIVolume file"
+
+### fn spec.projectedVolumeTemplate.sources.downwardAPI.items.withMode
+
+```ts
+withMode(mode)
+```
+
+"Optional: mode bits used to set permissions on this file, must be an octal value between 0000 and 0777 or a decimal value between 0 and 511. YAML accepts both octal and decimal values, JSON requires decimal values for mode bits. If not specified, the volume defaultMode will be used. This might be in conflict with other options that affect the file mode, like fsGroup, and the result can be other mode bits set."
+
+### fn spec.projectedVolumeTemplate.sources.downwardAPI.items.withPath
+
+```ts
+withPath(path)
+```
+
+"Required: Path is the relative path name of the file to be created. Must not be absolute or contain the '..' path. Must be utf-8 encoded. The first item of the relative path must not start with '..'"
+
+## obj spec.projectedVolumeTemplate.sources.downwardAPI.items.fieldRef
+
+"Required: Selects a field of the pod: only annotations, labels, name and namespace are supported."
+
+### fn spec.projectedVolumeTemplate.sources.downwardAPI.items.fieldRef.withApiVersion
+
+```ts
+withApiVersion(apiVersion)
+```
+
+"Version of the schema the FieldPath is written in terms of, defaults to \"v1\"."
+
+### fn spec.projectedVolumeTemplate.sources.downwardAPI.items.fieldRef.withFieldPath
+
+```ts
+withFieldPath(fieldPath)
+```
+
+"Path of the field to select in the specified API version."
+
+## obj spec.projectedVolumeTemplate.sources.downwardAPI.items.resourceFieldRef
+
+"Selects a resource of the container: only resources limits and requests (limits.cpu, limits.memory, requests.cpu and requests.memory) are currently supported."
+
+### fn spec.projectedVolumeTemplate.sources.downwardAPI.items.resourceFieldRef.withContainerName
+
+```ts
+withContainerName(containerName)
+```
+
+"Container name: required for volumes, optional for env vars"
+
+### fn spec.projectedVolumeTemplate.sources.downwardAPI.items.resourceFieldRef.withDivisor
+
+```ts
+withDivisor(divisor)
+```
+
+"Specifies the output format of the exposed resources, defaults to \"1\
+
+### fn spec.projectedVolumeTemplate.sources.downwardAPI.items.resourceFieldRef.withResource
+
+```ts
+withResource(resource)
+```
+
+"Required: resource to select"
+
+## obj spec.projectedVolumeTemplate.sources.secret
+
+"secret information about the secret data to project"
+
+### fn spec.projectedVolumeTemplate.sources.secret.withItems
+
+```ts
+withItems(items)
+```
+
+"items if unspecified, each key-value pair in the Data field of the referenced Secret will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the Secret, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'."
+
+### fn spec.projectedVolumeTemplate.sources.secret.withItemsMixin
+
+```ts
+withItemsMixin(items)
+```
+
+"items if unspecified, each key-value pair in the Data field of the referenced Secret will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the Secret, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.projectedVolumeTemplate.sources.secret.withName
+
+```ts
+withName(name)
+```
+
+"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"
+
+### fn spec.projectedVolumeTemplate.sources.secret.withOptional
+
+```ts
+withOptional(optional)
+```
+
+"optional field specify whether the Secret or its key must be defined"
+
+## obj spec.projectedVolumeTemplate.sources.secret.items
+
+"items if unspecified, each key-value pair in the Data field of the referenced Secret will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the Secret, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'."
+
+### fn spec.projectedVolumeTemplate.sources.secret.items.withKey
+
+```ts
+withKey(key)
+```
+
+"key is the key to project."
+
+### fn spec.projectedVolumeTemplate.sources.secret.items.withMode
+
+```ts
+withMode(mode)
+```
+
+"mode is Optional: mode bits used to set permissions on this file. Must be an octal value between 0000 and 0777 or a decimal value between 0 and 511. YAML accepts both octal and decimal values, JSON requires decimal values for mode bits. If not specified, the volume defaultMode will be used. This might be in conflict with other options that affect the file mode, like fsGroup, and the result can be other mode bits set."
+
+### fn spec.projectedVolumeTemplate.sources.secret.items.withPath
+
+```ts
+withPath(path)
+```
+
+"path is the relative path of the file to map the key to. May not be an absolute path. May not contain the path element '..'. May not start with the string '..'."
+
+## obj spec.projectedVolumeTemplate.sources.serviceAccountToken
+
+"serviceAccountToken is information about the serviceAccountToken data to project"
+
+### fn spec.projectedVolumeTemplate.sources.serviceAccountToken.withAudience
+
+```ts
+withAudience(audience)
+```
+
+"audience is the intended audience of the token. A recipient of a token must identify itself with an identifier specified in the audience of the token, and otherwise should reject the token. The audience defaults to the identifier of the apiserver."
+
+### fn spec.projectedVolumeTemplate.sources.serviceAccountToken.withExpirationSeconds
+
+```ts
+withExpirationSeconds(expirationSeconds)
+```
+
+"expirationSeconds is the requested duration of validity of the service account token. As the token approaches expiration, the kubelet volume plugin will proactively rotate the service account token. The kubelet will start trying to rotate the token if the token is older than 80 percent of its time to live or if the token is older than 24 hours.Defaults to 1 hour and must be at least 10 minutes."
+
+### fn spec.projectedVolumeTemplate.sources.serviceAccountToken.withPath
+
+```ts
+withPath(path)
+```
+
+"path is the path relative to the mount point of the file to project the token into."
+
+## obj spec.replica
+
+"Replica cluster configuration"
+
+### fn spec.replica.withEnabled
+
+```ts
+withEnabled(enabled)
+```
+
+"If replica mode is enabled, this cluster will be a replica of an existing cluster. Replica cluster can be created from a recovery object store or via streaming through pg_basebackup. Refer to the Replica clusters page of the documentation for more information."
+
+### fn spec.replica.withSource
+
+```ts
+withSource(source)
+```
+
+"The name of the external cluster which is the replication origin"
+
+## obj spec.replicationSlots
+
+"Replication slots management configuration"
+
+### fn spec.replicationSlots.withUpdateInterval
+
+```ts
+withUpdateInterval(updateInterval)
+```
+
+"Standby will update the status of the local replication slots every `updateInterval` seconds (default 30)."
+
+## obj spec.replicationSlots.highAvailability
+
+"Replication slots for high availability configuration"
+
+### fn spec.replicationSlots.highAvailability.withEnabled
+
+```ts
+withEnabled(enabled)
+```
+
+"If enabled (default), the operator will automatically manage replication slots on the primary instance and use them in streaming replication connections with all the standby instances that are part of the HA cluster. If disabled, the operator will not take advantage of replication slots in streaming connections with the replicas. This feature also controls replication slots in replica cluster, from the designated primary to its cascading replicas."
+
+### fn spec.replicationSlots.highAvailability.withSlotPrefix
+
+```ts
+withSlotPrefix(slotPrefix)
+```
+
+"Prefix for replication slots managed by the operator for HA. It may only contain lower case letters, numbers, and the underscore character. This can only be set at creation time. By default set to `_cnpg_`."
+
+## obj spec.resources
+
+"Resources requirements of every generated Pod. Please refer to https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/ for more information."
+
+### fn spec.resources.withClaims
+
+```ts
+withClaims(claims)
+```
+
+"Claims lists the names of resources, defined in spec.resourceClaims, that are used by this container. \n This is an alpha field and requires enabling the DynamicResourceAllocation feature gate. \n This field is immutable. It can only be set for containers."
+
+### fn spec.resources.withClaimsMixin
+
+```ts
+withClaimsMixin(claims)
+```
+
+"Claims lists the names of resources, defined in spec.resourceClaims, that are used by this container. \n This is an alpha field and requires enabling the DynamicResourceAllocation feature gate. \n This field is immutable. It can only be set for containers."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.resources.withLimits
+
+```ts
+withLimits(limits)
+```
+
+"Limits describes the maximum amount of compute resources allowed. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"
+
+### fn spec.resources.withLimitsMixin
+
+```ts
+withLimitsMixin(limits)
+```
+
+"Limits describes the maximum amount of compute resources allowed. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.resources.withRequests
+
+```ts
+withRequests(requests)
+```
+
+"Requests describes the minimum amount of compute resources required. If Requests is omitted for a container, it defaults to Limits if that is explicitly specified, otherwise to an implementation-defined value. Requests cannot exceed Limits. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"
+
+### fn spec.resources.withRequestsMixin
+
+```ts
+withRequestsMixin(requests)
+```
+
+"Requests describes the minimum amount of compute resources required. If Requests is omitted for a container, it defaults to Limits if that is explicitly specified, otherwise to an implementation-defined value. Requests cannot exceed Limits. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.resources.claims
+
+"Claims lists the names of resources, defined in spec.resourceClaims, that are used by this container. \n This is an alpha field and requires enabling the DynamicResourceAllocation feature gate. \n This field is immutable. It can only be set for containers."
+
+### fn spec.resources.claims.withName
+
+```ts
+withName(name)
+```
+
+"Name must match the name of one entry in pod.spec.resourceClaims of the Pod where this field is used. It makes that resource available inside a container."
+
+## obj spec.seccompProfile
+
+"The SeccompProfile applied to every Pod and Container. Defaults to: `RuntimeDefault`"
+
+### fn spec.seccompProfile.withLocalhostProfile
+
+```ts
+withLocalhostProfile(localhostProfile)
+```
+
+"localhostProfile indicates a profile defined in a file on the node should be used. The profile must be preconfigured on the node to work. Must be a descending path, relative to the kubelet's configured seccomp profile location. Must be set if type is \"Localhost\". Must NOT be set for any other type."
+
+### fn spec.seccompProfile.withType
+
+```ts
+withType(type)
+```
+
+"type indicates which kind of seccomp profile will be applied. Valid options are: \n Localhost - a profile defined in a file on the node should be used. RuntimeDefault - the container runtime default profile should be used. Unconfined - no profile should be applied."
+
+## obj spec.serviceAccountTemplate
+
+"Configure the generation of the service account"
+
+## obj spec.serviceAccountTemplate.metadata
+
+"Metadata are the metadata to be used for the generated service account"
+
+### fn spec.serviceAccountTemplate.metadata.withAnnotations
+
+```ts
+withAnnotations(annotations)
+```
+
+"Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations"
+
+### fn spec.serviceAccountTemplate.metadata.withAnnotationsMixin
+
+```ts
+withAnnotationsMixin(annotations)
+```
+
+"Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations"
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.serviceAccountTemplate.metadata.withLabels
+
+```ts
+withLabels(labels)
+```
+
+"Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels"
+
+### fn spec.serviceAccountTemplate.metadata.withLabelsMixin
+
+```ts
+withLabelsMixin(labels)
+```
+
+"Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels"
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.storage
+
+"Configuration of the storage of the instances"
+
+### fn spec.storage.withResizeInUseVolumes
+
+```ts
+withResizeInUseVolumes(resizeInUseVolumes)
+```
+
+"Resize existent PVCs, defaults to true"
+
+### fn spec.storage.withSize
+
+```ts
+withSize(size)
+```
+
+"Size of the storage. Required if not already specified in the PVC template. Changes to this field are automatically reapplied to the created PVCs. Size cannot be decreased."
+
+### fn spec.storage.withStorageClass
+
+```ts
+withStorageClass(storageClass)
+```
+
+"StorageClass to use for database data (`PGDATA`). Applied after evaluating the PVC template, if available. If not specified, generated PVCs will be satisfied by the default storage class"
+
+## obj spec.storage.pvcTemplate
+
+"Template to be used to generate the Persistent Volume Claim"
+
+### fn spec.storage.pvcTemplate.withAccessModes
+
+```ts
+withAccessModes(accessModes)
+```
+
+"accessModes contains the desired access modes the volume should have. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#access-modes-1"
+
+### fn spec.storage.pvcTemplate.withAccessModesMixin
+
+```ts
+withAccessModesMixin(accessModes)
+```
+
+"accessModes contains the desired access modes the volume should have. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#access-modes-1"
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.storage.pvcTemplate.withStorageClassName
+
+```ts
+withStorageClassName(storageClassName)
+```
+
+"storageClassName is the name of the StorageClass required by the claim. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#class-1"
+
+### fn spec.storage.pvcTemplate.withVolumeMode
+
+```ts
+withVolumeMode(volumeMode)
+```
+
+"volumeMode defines what type of volume is required by the claim. Value of Filesystem is implied when not included in claim spec."
+
+### fn spec.storage.pvcTemplate.withVolumeName
+
+```ts
+withVolumeName(volumeName)
+```
+
+"volumeName is the binding reference to the PersistentVolume backing this claim."
+
+## obj spec.storage.pvcTemplate.dataSource
+
+"dataSource field can be used to specify either: * An existing VolumeSnapshot object (snapshot.storage.k8s.io/VolumeSnapshot) * An existing PVC (PersistentVolumeClaim) If the provisioner or an external controller can support the specified data source, it will create a new volume based on the contents of the specified data source. When the AnyVolumeDataSource feature gate is enabled, dataSource contents will be copied to dataSourceRef, and dataSourceRef contents will be copied to dataSource when dataSourceRef.namespace is not specified. If the namespace is specified, then dataSourceRef will not be copied to dataSource."
+
+### fn spec.storage.pvcTemplate.dataSource.withApiGroup
+
+```ts
+withApiGroup(apiGroup)
+```
+
+"APIGroup is the group for the resource being referenced. If APIGroup is not specified, the specified Kind must be in the core API group. For any other third-party types, APIGroup is required."
+
+### fn spec.storage.pvcTemplate.dataSource.withKind
+
+```ts
+withKind(kind)
+```
+
+"Kind is the type of resource being referenced"
+
+### fn spec.storage.pvcTemplate.dataSource.withName
+
+```ts
+withName(name)
+```
+
+"Name is the name of resource being referenced"
+
+## obj spec.storage.pvcTemplate.dataSourceRef
+
+"dataSourceRef specifies the object from which to populate the volume with data, if a non-empty volume is desired. This may be any object from a non-empty API group (non core object) or a PersistentVolumeClaim object. When this field is specified, volume binding will only succeed if the type of the specified object matches some installed volume populator or dynamic provisioner. This field will replace the functionality of the dataSource field and as such if both fields are non-empty, they must have the same value. For backwards compatibility, when namespace isn't specified in dataSourceRef, both fields (dataSource and dataSourceRef) will be set to the same value automatically if one of them is empty and the other is non-empty. When namespace is specified in dataSourceRef, dataSource isn't set to the same value and must be empty. There are three important differences between dataSource and dataSourceRef: * While dataSource only allows two specific types of objects, dataSourceRef allows any non-core object, as well as PersistentVolumeClaim objects. * While dataSource ignores disallowed values (dropping them), dataSourceRef preserves all values, and generates an error if a disallowed value is specified. * While dataSource only allows local objects, dataSourceRef allows objects in any namespaces. (Beta) Using this field requires the AnyVolumeDataSource feature gate to be enabled. (Alpha) Using the namespace field of dataSourceRef requires the CrossNamespaceVolumeDataSource feature gate to be enabled."
+
+### fn spec.storage.pvcTemplate.dataSourceRef.withApiGroup
+
+```ts
+withApiGroup(apiGroup)
+```
+
+"APIGroup is the group for the resource being referenced. If APIGroup is not specified, the specified Kind must be in the core API group. For any other third-party types, APIGroup is required."
+
+### fn spec.storage.pvcTemplate.dataSourceRef.withKind
+
+```ts
+withKind(kind)
+```
+
+"Kind is the type of resource being referenced"
+
+### fn spec.storage.pvcTemplate.dataSourceRef.withName
+
+```ts
+withName(name)
+```
+
+"Name is the name of resource being referenced"
+
+### fn spec.storage.pvcTemplate.dataSourceRef.withNamespace
+
+```ts
+withNamespace(namespace)
+```
+
+"Namespace is the namespace of resource being referenced Note that when a namespace is specified, a gateway.networking.k8s.io/ReferenceGrant object is required in the referent namespace to allow that namespace's owner to accept the reference. See the ReferenceGrant documentation for details. (Alpha) This field requires the CrossNamespaceVolumeDataSource feature gate to be enabled."
+
+## obj spec.storage.pvcTemplate.resources
+
+"resources represents the minimum resources the volume should have. If RecoverVolumeExpansionFailure feature is enabled users are allowed to specify resource requirements that are lower than previous value but must still be higher than capacity recorded in the status field of the claim. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#resources"
+
+### fn spec.storage.pvcTemplate.resources.withClaims
+
+```ts
+withClaims(claims)
+```
+
+"Claims lists the names of resources, defined in spec.resourceClaims, that are used by this container. \n This is an alpha field and requires enabling the DynamicResourceAllocation feature gate. \n This field is immutable. It can only be set for containers."
+
+### fn spec.storage.pvcTemplate.resources.withClaimsMixin
+
+```ts
+withClaimsMixin(claims)
+```
+
+"Claims lists the names of resources, defined in spec.resourceClaims, that are used by this container. \n This is an alpha field and requires enabling the DynamicResourceAllocation feature gate. \n This field is immutable. It can only be set for containers."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.storage.pvcTemplate.resources.withLimits
+
+```ts
+withLimits(limits)
+```
+
+"Limits describes the maximum amount of compute resources allowed. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"
+
+### fn spec.storage.pvcTemplate.resources.withLimitsMixin
+
+```ts
+withLimitsMixin(limits)
+```
+
+"Limits describes the maximum amount of compute resources allowed. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.storage.pvcTemplate.resources.withRequests
+
+```ts
+withRequests(requests)
+```
+
+"Requests describes the minimum amount of compute resources required. If Requests is omitted for a container, it defaults to Limits if that is explicitly specified, otherwise to an implementation-defined value. Requests cannot exceed Limits. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"
+
+### fn spec.storage.pvcTemplate.resources.withRequestsMixin
+
+```ts
+withRequestsMixin(requests)
+```
+
+"Requests describes the minimum amount of compute resources required. If Requests is omitted for a container, it defaults to Limits if that is explicitly specified, otherwise to an implementation-defined value. Requests cannot exceed Limits. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.storage.pvcTemplate.resources.claims
+
+"Claims lists the names of resources, defined in spec.resourceClaims, that are used by this container. \n This is an alpha field and requires enabling the DynamicResourceAllocation feature gate. \n This field is immutable. It can only be set for containers."
+
+### fn spec.storage.pvcTemplate.resources.claims.withName
+
+```ts
+withName(name)
+```
+
+"Name must match the name of one entry in pod.spec.resourceClaims of the Pod where this field is used. It makes that resource available inside a container."
+
+## obj spec.storage.pvcTemplate.selector
+
+"selector is a label query over volumes to consider for binding."
+
+### fn spec.storage.pvcTemplate.selector.withMatchExpressions
+
+```ts
+withMatchExpressions(matchExpressions)
+```
+
+"matchExpressions is a list of label selector requirements. The requirements are ANDed."
+
+### fn spec.storage.pvcTemplate.selector.withMatchExpressionsMixin
+
+```ts
+withMatchExpressionsMixin(matchExpressions)
+```
+
+"matchExpressions is a list of label selector requirements. The requirements are ANDed."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.storage.pvcTemplate.selector.withMatchLabels
+
+```ts
+withMatchLabels(matchLabels)
+```
+
+"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \"key\", the operator is \"In\", and the values array contains only \"value\". The requirements are ANDed."
+
+### fn spec.storage.pvcTemplate.selector.withMatchLabelsMixin
+
+```ts
+withMatchLabelsMixin(matchLabels)
+```
+
+"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \"key\", the operator is \"In\", and the values array contains only \"value\". The requirements are ANDed."
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.storage.pvcTemplate.selector.matchExpressions
+
+"matchExpressions is a list of label selector requirements. The requirements are ANDed."
+
+### fn spec.storage.pvcTemplate.selector.matchExpressions.withKey
+
+```ts
+withKey(key)
+```
+
+"key is the label key that the selector applies to."
+
+### fn spec.storage.pvcTemplate.selector.matchExpressions.withOperator
+
+```ts
+withOperator(operator)
+```
+
+"operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist."
+
+### fn spec.storage.pvcTemplate.selector.matchExpressions.withValues
+
+```ts
+withValues(values)
+```
+
+"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."
+
+### fn spec.storage.pvcTemplate.selector.matchExpressions.withValuesMixin
+
+```ts
+withValuesMixin(values)
+```
+
+"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.superuserSecret
+
+"The secret containing the superuser password. If not defined a new secret will be created with a randomly generated password"
+
+### fn spec.superuserSecret.withName
+
+```ts
+withName(name)
+```
+
+"Name of the referent."
+
+## obj spec.topologySpreadConstraints
+
+"TopologySpreadConstraints specifies how to spread matching pods among the given topology. More info: https://kubernetes.io/docs/concepts/scheduling-eviction/topology-spread-constraints/"
+
+### fn spec.topologySpreadConstraints.withMatchLabelKeys
+
+```ts
+withMatchLabelKeys(matchLabelKeys)
+```
+
+"MatchLabelKeys is a set of pod label keys to select the pods over which spreading will be calculated. The keys are used to lookup values from the incoming pod labels, those key-value labels are ANDed with labelSelector to select the group of existing pods over which spreading will be calculated for the incoming pod. The same key is forbidden to exist in both MatchLabelKeys and LabelSelector. MatchLabelKeys cannot be set when LabelSelector isn't set. Keys that don't exist in the incoming pod labels will be ignored. A null or empty list means only match against labelSelector. \n This is a beta field and requires the MatchLabelKeysInPodTopologySpread feature gate to be enabled (enabled by default)."
+
+### fn spec.topologySpreadConstraints.withMatchLabelKeysMixin
+
+```ts
+withMatchLabelKeysMixin(matchLabelKeys)
+```
+
+"MatchLabelKeys is a set of pod label keys to select the pods over which spreading will be calculated. The keys are used to lookup values from the incoming pod labels, those key-value labels are ANDed with labelSelector to select the group of existing pods over which spreading will be calculated for the incoming pod. The same key is forbidden to exist in both MatchLabelKeys and LabelSelector. MatchLabelKeys cannot be set when LabelSelector isn't set. Keys that don't exist in the incoming pod labels will be ignored. A null or empty list means only match against labelSelector. \n This is a beta field and requires the MatchLabelKeysInPodTopologySpread feature gate to be enabled (enabled by default)."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.topologySpreadConstraints.withMaxSkew
+
+```ts
+withMaxSkew(maxSkew)
+```
+
+"MaxSkew describes the degree to which pods may be unevenly distributed. When `whenUnsatisfiable=DoNotSchedule`, it is the maximum permitted difference between the number of matching pods in the target topology and the global minimum. The global minimum is the minimum number of matching pods in an eligible domain or zero if the number of eligible domains is less than MinDomains. For example, in a 3-zone cluster, MaxSkew is set to 1, and pods with the same labelSelector spread as 2/2/1: In this case, the global minimum is 1. | zone1 | zone2 | zone3 | | P P | P P | P | - if MaxSkew is 1, incoming pod can only be scheduled to zone3 to become 2/2/2; scheduling it onto zone1(zone2) would make the ActualSkew(3-1) on zone1(zone2) violate MaxSkew(1). - if MaxSkew is 2, incoming pod can be scheduled onto any zone. When `whenUnsatisfiable=ScheduleAnyway`, it is used to give higher precedence to topologies that satisfy it. It's a required field. Default value is 1 and 0 is not allowed."
+
+### fn spec.topologySpreadConstraints.withMinDomains
+
+```ts
+withMinDomains(minDomains)
+```
+
+"MinDomains indicates a minimum number of eligible domains. When the number of eligible domains with matching topology keys is less than minDomains, Pod Topology Spread treats \"global minimum\" as 0, and then the calculation of Skew is performed. And when the number of eligible domains with matching topology keys equals or greater than minDomains, this value has no effect on scheduling. As a result, when the number of eligible domains is less than minDomains, scheduler won't schedule more than maxSkew Pods to those domains. If value is nil, the constraint behaves as if MinDomains is equal to 1. Valid values are integers greater than 0. When value is not nil, WhenUnsatisfiable must be DoNotSchedule. \n For example, in a 3-zone cluster, MaxSkew is set to 2, MinDomains is set to 5 and pods with the same labelSelector spread as 2/2/2: | zone1 | zone2 | zone3 | | P P | P P | P P | The number of domains is less than 5(MinDomains), so \"global minimum\" is treated as 0. In this situation, new pod with the same labelSelector cannot be scheduled, because computed skew will be 3(3 - 0) if new Pod is scheduled to any of the three zones, it will violate MaxSkew. \n This is a beta field and requires the MinDomainsInPodTopologySpread feature gate to be enabled (enabled by default)."
+
+### fn spec.topologySpreadConstraints.withNodeAffinityPolicy
+
+```ts
+withNodeAffinityPolicy(nodeAffinityPolicy)
+```
+
+"NodeAffinityPolicy indicates how we will treat Pod's nodeAffinity/nodeSelector when calculating pod topology spread skew. Options are: - Honor: only nodes matching nodeAffinity/nodeSelector are included in the calculations. - Ignore: nodeAffinity/nodeSelector are ignored. All nodes are included in the calculations. \n If this value is nil, the behavior is equivalent to the Honor policy. This is a beta-level feature default enabled by the NodeInclusionPolicyInPodTopologySpread feature flag."
+
+### fn spec.topologySpreadConstraints.withNodeTaintsPolicy
+
+```ts
+withNodeTaintsPolicy(nodeTaintsPolicy)
+```
+
+"NodeTaintsPolicy indicates how we will treat node taints when calculating pod topology spread skew. Options are: - Honor: nodes without taints, along with tainted nodes for which the incoming pod has a toleration, are included. - Ignore: node taints are ignored. All nodes are included. \n If this value is nil, the behavior is equivalent to the Ignore policy. This is a beta-level feature default enabled by the NodeInclusionPolicyInPodTopologySpread feature flag."
+
+### fn spec.topologySpreadConstraints.withTopologyKey
+
+```ts
+withTopologyKey(topologyKey)
+```
+
+"TopologyKey is the key of node labels. Nodes that have a label with this key and identical values are considered to be in the same topology. We consider each as a \"bucket\", and try to put balanced number of pods into each bucket. We define a domain as a particular instance of a topology. Also, we define an eligible domain as a domain whose nodes meet the requirements of nodeAffinityPolicy and nodeTaintsPolicy. e.g. If TopologyKey is \"kubernetes.io/hostname\", each Node is a domain of that topology. And, if TopologyKey is \"topology.kubernetes.io/zone\", each zone is a domain of that topology. It's a required field."
+
+### fn spec.topologySpreadConstraints.withWhenUnsatisfiable
+
+```ts
+withWhenUnsatisfiable(whenUnsatisfiable)
+```
+
+"WhenUnsatisfiable indicates how to deal with a pod if it doesn't satisfy the spread constraint. - DoNotSchedule (default) tells the scheduler not to schedule it. - ScheduleAnyway tells the scheduler to schedule the pod in any location, but giving higher precedence to topologies that would help reduce the skew. A constraint is considered \"Unsatisfiable\" for an incoming pod if and only if every possible node assignment for that pod would violate \"MaxSkew\" on some topology. For example, in a 3-zone cluster, MaxSkew is set to 1, and pods with the same labelSelector spread as 3/1/1: | zone1 | zone2 | zone3 | | P P P | P | P | If WhenUnsatisfiable is set to DoNotSchedule, incoming pod can only be scheduled to zone2(zone3) to become 3/2/1(3/1/2) as ActualSkew(2-1) on zone2(zone3) satisfies MaxSkew(1). In other words, the cluster can still be imbalanced, but scheduler won't make it *more* imbalanced. It's a required field."
+
+## obj spec.topologySpreadConstraints.labelSelector
+
+"LabelSelector is used to find matching pods. Pods that match this label selector are counted to determine the number of pods in their corresponding topology domain."
+
+### fn spec.topologySpreadConstraints.labelSelector.withMatchExpressions
+
+```ts
+withMatchExpressions(matchExpressions)
+```
+
+"matchExpressions is a list of label selector requirements. The requirements are ANDed."
+
+### fn spec.topologySpreadConstraints.labelSelector.withMatchExpressionsMixin
+
+```ts
+withMatchExpressionsMixin(matchExpressions)
+```
+
+"matchExpressions is a list of label selector requirements. The requirements are ANDed."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.topologySpreadConstraints.labelSelector.withMatchLabels
+
+```ts
+withMatchLabels(matchLabels)
+```
+
+"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \"key\", the operator is \"In\", and the values array contains only \"value\". The requirements are ANDed."
+
+### fn spec.topologySpreadConstraints.labelSelector.withMatchLabelsMixin
+
+```ts
+withMatchLabelsMixin(matchLabels)
+```
+
+"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \"key\", the operator is \"In\", and the values array contains only \"value\". The requirements are ANDed."
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.topologySpreadConstraints.labelSelector.matchExpressions
+
+"matchExpressions is a list of label selector requirements. The requirements are ANDed."
+
+### fn spec.topologySpreadConstraints.labelSelector.matchExpressions.withKey
+
+```ts
+withKey(key)
+```
+
+"key is the label key that the selector applies to."
+
+### fn spec.topologySpreadConstraints.labelSelector.matchExpressions.withOperator
+
+```ts
+withOperator(operator)
+```
+
+"operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist."
+
+### fn spec.topologySpreadConstraints.labelSelector.matchExpressions.withValues
+
+```ts
+withValues(values)
+```
+
+"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."
+
+### fn spec.topologySpreadConstraints.labelSelector.matchExpressions.withValuesMixin
+
+```ts
+withValuesMixin(values)
+```
+
+"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.walStorage
+
+"Configuration of the storage for PostgreSQL WAL (Write-Ahead Log)"
+
+### fn spec.walStorage.withResizeInUseVolumes
+
+```ts
+withResizeInUseVolumes(resizeInUseVolumes)
+```
+
+"Resize existent PVCs, defaults to true"
+
+### fn spec.walStorage.withSize
+
+```ts
+withSize(size)
+```
+
+"Size of the storage. Required if not already specified in the PVC template. Changes to this field are automatically reapplied to the created PVCs. Size cannot be decreased."
+
+### fn spec.walStorage.withStorageClass
+
+```ts
+withStorageClass(storageClass)
+```
+
+"StorageClass to use for database data (`PGDATA`). Applied after evaluating the PVC template, if available. If not specified, generated PVCs will be satisfied by the default storage class"
+
+## obj spec.walStorage.pvcTemplate
+
+"Template to be used to generate the Persistent Volume Claim"
+
+### fn spec.walStorage.pvcTemplate.withAccessModes
+
+```ts
+withAccessModes(accessModes)
+```
+
+"accessModes contains the desired access modes the volume should have. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#access-modes-1"
+
+### fn spec.walStorage.pvcTemplate.withAccessModesMixin
+
+```ts
+withAccessModesMixin(accessModes)
+```
+
+"accessModes contains the desired access modes the volume should have. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#access-modes-1"
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.walStorage.pvcTemplate.withStorageClassName
+
+```ts
+withStorageClassName(storageClassName)
+```
+
+"storageClassName is the name of the StorageClass required by the claim. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#class-1"
+
+### fn spec.walStorage.pvcTemplate.withVolumeMode
+
+```ts
+withVolumeMode(volumeMode)
+```
+
+"volumeMode defines what type of volume is required by the claim. Value of Filesystem is implied when not included in claim spec."
+
+### fn spec.walStorage.pvcTemplate.withVolumeName
+
+```ts
+withVolumeName(volumeName)
+```
+
+"volumeName is the binding reference to the PersistentVolume backing this claim."
+
+## obj spec.walStorage.pvcTemplate.dataSource
+
+"dataSource field can be used to specify either: * An existing VolumeSnapshot object (snapshot.storage.k8s.io/VolumeSnapshot) * An existing PVC (PersistentVolumeClaim) If the provisioner or an external controller can support the specified data source, it will create a new volume based on the contents of the specified data source. When the AnyVolumeDataSource feature gate is enabled, dataSource contents will be copied to dataSourceRef, and dataSourceRef contents will be copied to dataSource when dataSourceRef.namespace is not specified. If the namespace is specified, then dataSourceRef will not be copied to dataSource."
+
+### fn spec.walStorage.pvcTemplate.dataSource.withApiGroup
+
+```ts
+withApiGroup(apiGroup)
+```
+
+"APIGroup is the group for the resource being referenced. If APIGroup is not specified, the specified Kind must be in the core API group. For any other third-party types, APIGroup is required."
+
+### fn spec.walStorage.pvcTemplate.dataSource.withKind
+
+```ts
+withKind(kind)
+```
+
+"Kind is the type of resource being referenced"
+
+### fn spec.walStorage.pvcTemplate.dataSource.withName
+
+```ts
+withName(name)
+```
+
+"Name is the name of resource being referenced"
+
+## obj spec.walStorage.pvcTemplate.dataSourceRef
+
+"dataSourceRef specifies the object from which to populate the volume with data, if a non-empty volume is desired. This may be any object from a non-empty API group (non core object) or a PersistentVolumeClaim object. When this field is specified, volume binding will only succeed if the type of the specified object matches some installed volume populator or dynamic provisioner. This field will replace the functionality of the dataSource field and as such if both fields are non-empty, they must have the same value. For backwards compatibility, when namespace isn't specified in dataSourceRef, both fields (dataSource and dataSourceRef) will be set to the same value automatically if one of them is empty and the other is non-empty. When namespace is specified in dataSourceRef, dataSource isn't set to the same value and must be empty. There are three important differences between dataSource and dataSourceRef: * While dataSource only allows two specific types of objects, dataSourceRef allows any non-core object, as well as PersistentVolumeClaim objects. * While dataSource ignores disallowed values (dropping them), dataSourceRef preserves all values, and generates an error if a disallowed value is specified. * While dataSource only allows local objects, dataSourceRef allows objects in any namespaces. (Beta) Using this field requires the AnyVolumeDataSource feature gate to be enabled. (Alpha) Using the namespace field of dataSourceRef requires the CrossNamespaceVolumeDataSource feature gate to be enabled."
+
+### fn spec.walStorage.pvcTemplate.dataSourceRef.withApiGroup
+
+```ts
+withApiGroup(apiGroup)
+```
+
+"APIGroup is the group for the resource being referenced. If APIGroup is not specified, the specified Kind must be in the core API group. For any other third-party types, APIGroup is required."
+
+### fn spec.walStorage.pvcTemplate.dataSourceRef.withKind
+
+```ts
+withKind(kind)
+```
+
+"Kind is the type of resource being referenced"
+
+### fn spec.walStorage.pvcTemplate.dataSourceRef.withName
+
+```ts
+withName(name)
+```
+
+"Name is the name of resource being referenced"
+
+### fn spec.walStorage.pvcTemplate.dataSourceRef.withNamespace
+
+```ts
+withNamespace(namespace)
+```
+
+"Namespace is the namespace of resource being referenced Note that when a namespace is specified, a gateway.networking.k8s.io/ReferenceGrant object is required in the referent namespace to allow that namespace's owner to accept the reference. See the ReferenceGrant documentation for details. (Alpha) This field requires the CrossNamespaceVolumeDataSource feature gate to be enabled."
+
+## obj spec.walStorage.pvcTemplate.resources
+
+"resources represents the minimum resources the volume should have. If RecoverVolumeExpansionFailure feature is enabled users are allowed to specify resource requirements that are lower than previous value but must still be higher than capacity recorded in the status field of the claim. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#resources"
+
+### fn spec.walStorage.pvcTemplate.resources.withClaims
+
+```ts
+withClaims(claims)
+```
+
+"Claims lists the names of resources, defined in spec.resourceClaims, that are used by this container. \n This is an alpha field and requires enabling the DynamicResourceAllocation feature gate. \n This field is immutable. It can only be set for containers."
+
+### fn spec.walStorage.pvcTemplate.resources.withClaimsMixin
+
+```ts
+withClaimsMixin(claims)
+```
+
+"Claims lists the names of resources, defined in spec.resourceClaims, that are used by this container. \n This is an alpha field and requires enabling the DynamicResourceAllocation feature gate. \n This field is immutable. It can only be set for containers."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.walStorage.pvcTemplate.resources.withLimits
+
+```ts
+withLimits(limits)
+```
+
+"Limits describes the maximum amount of compute resources allowed. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"
+
+### fn spec.walStorage.pvcTemplate.resources.withLimitsMixin
+
+```ts
+withLimitsMixin(limits)
+```
+
+"Limits describes the maximum amount of compute resources allowed. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.walStorage.pvcTemplate.resources.withRequests
+
+```ts
+withRequests(requests)
+```
+
+"Requests describes the minimum amount of compute resources required. If Requests is omitted for a container, it defaults to Limits if that is explicitly specified, otherwise to an implementation-defined value. Requests cannot exceed Limits. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"
+
+### fn spec.walStorage.pvcTemplate.resources.withRequestsMixin
+
+```ts
+withRequestsMixin(requests)
+```
+
+"Requests describes the minimum amount of compute resources required. If Requests is omitted for a container, it defaults to Limits if that is explicitly specified, otherwise to an implementation-defined value. Requests cannot exceed Limits. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.walStorage.pvcTemplate.resources.claims
+
+"Claims lists the names of resources, defined in spec.resourceClaims, that are used by this container. \n This is an alpha field and requires enabling the DynamicResourceAllocation feature gate. \n This field is immutable. It can only be set for containers."
+
+### fn spec.walStorage.pvcTemplate.resources.claims.withName
+
+```ts
+withName(name)
+```
+
+"Name must match the name of one entry in pod.spec.resourceClaims of the Pod where this field is used. It makes that resource available inside a container."
+
+## obj spec.walStorage.pvcTemplate.selector
+
+"selector is a label query over volumes to consider for binding."
+
+### fn spec.walStorage.pvcTemplate.selector.withMatchExpressions
+
+```ts
+withMatchExpressions(matchExpressions)
+```
+
+"matchExpressions is a list of label selector requirements. The requirements are ANDed."
+
+### fn spec.walStorage.pvcTemplate.selector.withMatchExpressionsMixin
+
+```ts
+withMatchExpressionsMixin(matchExpressions)
+```
+
+"matchExpressions is a list of label selector requirements. The requirements are ANDed."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.walStorage.pvcTemplate.selector.withMatchLabels
+
+```ts
+withMatchLabels(matchLabels)
+```
+
+"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \"key\", the operator is \"In\", and the values array contains only \"value\". The requirements are ANDed."
+
+### fn spec.walStorage.pvcTemplate.selector.withMatchLabelsMixin
+
+```ts
+withMatchLabelsMixin(matchLabels)
+```
+
+"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \"key\", the operator is \"In\", and the values array contains only \"value\". The requirements are ANDed."
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.walStorage.pvcTemplate.selector.matchExpressions
+
+"matchExpressions is a list of label selector requirements. The requirements are ANDed."
+
+### fn spec.walStorage.pvcTemplate.selector.matchExpressions.withKey
+
+```ts
+withKey(key)
+```
+
+"key is the label key that the selector applies to."
+
+### fn spec.walStorage.pvcTemplate.selector.matchExpressions.withOperator
+
+```ts
+withOperator(operator)
+```
+
+"operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist."
+
+### fn spec.walStorage.pvcTemplate.selector.matchExpressions.withValues
+
+```ts
+withValues(values)
+```
+
+"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."
+
+### fn spec.walStorage.pvcTemplate.selector.matchExpressions.withValuesMixin
+
+```ts
+withValuesMixin(values)
+```
+
+"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."
+
+**Note:** This function appends passed data to existing values
\ No newline at end of file
diff --git a/docs/1.21.0/postgresql/v1/index.md b/docs/1.21.0/postgresql/v1/index.md
new file mode 100644
index 0000000..8e77206
--- /dev/null
+++ b/docs/1.21.0/postgresql/v1/index.md
@@ -0,0 +1,12 @@
+---
+permalink: /1.21.0/postgresql/v1/
+---
+
+# postgresql.v1
+
+
+
+* [backup](backup.md)
+* [cluster](cluster.md)
+* [pooler](pooler.md)
+* [scheduledBackup](scheduledBackup.md)
\ No newline at end of file
diff --git a/docs/1.21.0/postgresql/v1/pooler.md b/docs/1.21.0/postgresql/v1/pooler.md
new file mode 100644
index 0000000..cf16c98
--- /dev/null
+++ b/docs/1.21.0/postgresql/v1/pooler.md
@@ -0,0 +1,10494 @@
+---
+permalink: /1.21.0/postgresql/v1/pooler/
+---
+
+# postgresql.v1.pooler
+
+"Pooler is the Schema for the poolers API"
+
+## Index
+
+* [`fn new(name)`](#fn-new)
+* [`obj metadata`](#obj-metadata)
+ * [`fn withAnnotations(annotations)`](#fn-metadatawithannotations)
+ * [`fn withAnnotationsMixin(annotations)`](#fn-metadatawithannotationsmixin)
+ * [`fn withClusterName(clusterName)`](#fn-metadatawithclustername)
+ * [`fn withCreationTimestamp(creationTimestamp)`](#fn-metadatawithcreationtimestamp)
+ * [`fn withDeletionGracePeriodSeconds(deletionGracePeriodSeconds)`](#fn-metadatawithdeletiongraceperiodseconds)
+ * [`fn withDeletionTimestamp(deletionTimestamp)`](#fn-metadatawithdeletiontimestamp)
+ * [`fn withFinalizers(finalizers)`](#fn-metadatawithfinalizers)
+ * [`fn withFinalizersMixin(finalizers)`](#fn-metadatawithfinalizersmixin)
+ * [`fn withGenerateName(generateName)`](#fn-metadatawithgeneratename)
+ * [`fn withGeneration(generation)`](#fn-metadatawithgeneration)
+ * [`fn withLabels(labels)`](#fn-metadatawithlabels)
+ * [`fn withLabelsMixin(labels)`](#fn-metadatawithlabelsmixin)
+ * [`fn withName(name)`](#fn-metadatawithname)
+ * [`fn withNamespace(namespace)`](#fn-metadatawithnamespace)
+ * [`fn withOwnerReferences(ownerReferences)`](#fn-metadatawithownerreferences)
+ * [`fn withOwnerReferencesMixin(ownerReferences)`](#fn-metadatawithownerreferencesmixin)
+ * [`fn withResourceVersion(resourceVersion)`](#fn-metadatawithresourceversion)
+ * [`fn withSelfLink(selfLink)`](#fn-metadatawithselflink)
+ * [`fn withUid(uid)`](#fn-metadatawithuid)
+* [`obj spec`](#obj-spec)
+ * [`fn withInstances(instances)`](#fn-specwithinstances)
+ * [`fn withType(type)`](#fn-specwithtype)
+ * [`obj spec.cluster`](#obj-speccluster)
+ * [`fn withName(name)`](#fn-specclusterwithname)
+ * [`obj spec.deploymentStrategy`](#obj-specdeploymentstrategy)
+ * [`fn withType(type)`](#fn-specdeploymentstrategywithtype)
+ * [`obj spec.deploymentStrategy.rollingUpdate`](#obj-specdeploymentstrategyrollingupdate)
+ * [`fn withMaxSurge(maxSurge)`](#fn-specdeploymentstrategyrollingupdatewithmaxsurge)
+ * [`fn withMaxUnavailable(maxUnavailable)`](#fn-specdeploymentstrategyrollingupdatewithmaxunavailable)
+ * [`obj spec.monitoring`](#obj-specmonitoring)
+ * [`fn withEnablePodMonitor(enablePodMonitor)`](#fn-specmonitoringwithenablepodmonitor)
+ * [`obj spec.pgbouncer`](#obj-specpgbouncer)
+ * [`fn withAuthQuery(authQuery)`](#fn-specpgbouncerwithauthquery)
+ * [`fn withParameters(parameters)`](#fn-specpgbouncerwithparameters)
+ * [`fn withParametersMixin(parameters)`](#fn-specpgbouncerwithparametersmixin)
+ * [`fn withPaused(paused)`](#fn-specpgbouncerwithpaused)
+ * [`fn withPg_hba(pg_hba)`](#fn-specpgbouncerwithpg_hba)
+ * [`fn withPg_hbaMixin(pg_hba)`](#fn-specpgbouncerwithpg_hbamixin)
+ * [`fn withPoolMode(poolMode)`](#fn-specpgbouncerwithpoolmode)
+ * [`obj spec.pgbouncer.authQuerySecret`](#obj-specpgbouncerauthquerysecret)
+ * [`fn withName(name)`](#fn-specpgbouncerauthquerysecretwithname)
+ * [`obj spec.template`](#obj-spectemplate)
+ * [`obj spec.template.metadata`](#obj-spectemplatemetadata)
+ * [`fn withAnnotations(annotations)`](#fn-spectemplatemetadatawithannotations)
+ * [`fn withAnnotationsMixin(annotations)`](#fn-spectemplatemetadatawithannotationsmixin)
+ * [`fn withLabels(labels)`](#fn-spectemplatemetadatawithlabels)
+ * [`fn withLabelsMixin(labels)`](#fn-spectemplatemetadatawithlabelsmixin)
+ * [`obj spec.template.spec`](#obj-spectemplatespec)
+ * [`fn withActiveDeadlineSeconds(activeDeadlineSeconds)`](#fn-spectemplatespecwithactivedeadlineseconds)
+ * [`fn withAutomountServiceAccountToken(automountServiceAccountToken)`](#fn-spectemplatespecwithautomountserviceaccounttoken)
+ * [`fn withContainers(containers)`](#fn-spectemplatespecwithcontainers)
+ * [`fn withContainersMixin(containers)`](#fn-spectemplatespecwithcontainersmixin)
+ * [`fn withDnsPolicy(dnsPolicy)`](#fn-spectemplatespecwithdnspolicy)
+ * [`fn withEnableServiceLinks(enableServiceLinks)`](#fn-spectemplatespecwithenableservicelinks)
+ * [`fn withEphemeralContainers(ephemeralContainers)`](#fn-spectemplatespecwithephemeralcontainers)
+ * [`fn withEphemeralContainersMixin(ephemeralContainers)`](#fn-spectemplatespecwithephemeralcontainersmixin)
+ * [`fn withHostAliases(hostAliases)`](#fn-spectemplatespecwithhostaliases)
+ * [`fn withHostAliasesMixin(hostAliases)`](#fn-spectemplatespecwithhostaliasesmixin)
+ * [`fn withHostIPC(hostIPC)`](#fn-spectemplatespecwithhostipc)
+ * [`fn withHostNetwork(hostNetwork)`](#fn-spectemplatespecwithhostnetwork)
+ * [`fn withHostPID(hostPID)`](#fn-spectemplatespecwithhostpid)
+ * [`fn withHostUsers(hostUsers)`](#fn-spectemplatespecwithhostusers)
+ * [`fn withHostname(hostname)`](#fn-spectemplatespecwithhostname)
+ * [`fn withImagePullSecrets(imagePullSecrets)`](#fn-spectemplatespecwithimagepullsecrets)
+ * [`fn withImagePullSecretsMixin(imagePullSecrets)`](#fn-spectemplatespecwithimagepullsecretsmixin)
+ * [`fn withInitContainers(initContainers)`](#fn-spectemplatespecwithinitcontainers)
+ * [`fn withInitContainersMixin(initContainers)`](#fn-spectemplatespecwithinitcontainersmixin)
+ * [`fn withNodeName(nodeName)`](#fn-spectemplatespecwithnodename)
+ * [`fn withNodeSelector(nodeSelector)`](#fn-spectemplatespecwithnodeselector)
+ * [`fn withNodeSelectorMixin(nodeSelector)`](#fn-spectemplatespecwithnodeselectormixin)
+ * [`fn withOverhead(overhead)`](#fn-spectemplatespecwithoverhead)
+ * [`fn withOverheadMixin(overhead)`](#fn-spectemplatespecwithoverheadmixin)
+ * [`fn withPreemptionPolicy(preemptionPolicy)`](#fn-spectemplatespecwithpreemptionpolicy)
+ * [`fn withPriority(priority)`](#fn-spectemplatespecwithpriority)
+ * [`fn withPriorityClassName(priorityClassName)`](#fn-spectemplatespecwithpriorityclassname)
+ * [`fn withReadinessGates(readinessGates)`](#fn-spectemplatespecwithreadinessgates)
+ * [`fn withReadinessGatesMixin(readinessGates)`](#fn-spectemplatespecwithreadinessgatesmixin)
+ * [`fn withResourceClaims(resourceClaims)`](#fn-spectemplatespecwithresourceclaims)
+ * [`fn withResourceClaimsMixin(resourceClaims)`](#fn-spectemplatespecwithresourceclaimsmixin)
+ * [`fn withRestartPolicy(restartPolicy)`](#fn-spectemplatespecwithrestartpolicy)
+ * [`fn withRuntimeClassName(runtimeClassName)`](#fn-spectemplatespecwithruntimeclassname)
+ * [`fn withSchedulerName(schedulerName)`](#fn-spectemplatespecwithschedulername)
+ * [`fn withSchedulingGates(schedulingGates)`](#fn-spectemplatespecwithschedulinggates)
+ * [`fn withSchedulingGatesMixin(schedulingGates)`](#fn-spectemplatespecwithschedulinggatesmixin)
+ * [`fn withServiceAccount(serviceAccount)`](#fn-spectemplatespecwithserviceaccount)
+ * [`fn withServiceAccountName(serviceAccountName)`](#fn-spectemplatespecwithserviceaccountname)
+ * [`fn withSetHostnameAsFQDN(setHostnameAsFQDN)`](#fn-spectemplatespecwithsethostnameasfqdn)
+ * [`fn withShareProcessNamespace(shareProcessNamespace)`](#fn-spectemplatespecwithshareprocessnamespace)
+ * [`fn withSubdomain(subdomain)`](#fn-spectemplatespecwithsubdomain)
+ * [`fn withTerminationGracePeriodSeconds(terminationGracePeriodSeconds)`](#fn-spectemplatespecwithterminationgraceperiodseconds)
+ * [`fn withTolerations(tolerations)`](#fn-spectemplatespecwithtolerations)
+ * [`fn withTolerationsMixin(tolerations)`](#fn-spectemplatespecwithtolerationsmixin)
+ * [`fn withTopologySpreadConstraints(topologySpreadConstraints)`](#fn-spectemplatespecwithtopologyspreadconstraints)
+ * [`fn withTopologySpreadConstraintsMixin(topologySpreadConstraints)`](#fn-spectemplatespecwithtopologyspreadconstraintsmixin)
+ * [`fn withVolumes(volumes)`](#fn-spectemplatespecwithvolumes)
+ * [`fn withVolumesMixin(volumes)`](#fn-spectemplatespecwithvolumesmixin)
+ * [`obj spec.template.spec.affinity`](#obj-spectemplatespecaffinity)
+ * [`obj spec.template.spec.affinity.nodeAffinity`](#obj-spectemplatespecaffinitynodeaffinity)
+ * [`fn withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution)`](#fn-spectemplatespecaffinitynodeaffinitywithpreferredduringschedulingignoredduringexecution)
+ * [`fn withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution)`](#fn-spectemplatespecaffinitynodeaffinitywithpreferredduringschedulingignoredduringexecutionmixin)
+ * [`obj spec.template.spec.affinity.nodeAffinity.preferredDuringSchedulingIgnoredDuringExecution`](#obj-spectemplatespecaffinitynodeaffinitypreferredduringschedulingignoredduringexecution)
+ * [`fn withWeight(weight)`](#fn-spectemplatespecaffinitynodeaffinitypreferredduringschedulingignoredduringexecutionwithweight)
+ * [`obj spec.template.spec.affinity.nodeAffinity.preferredDuringSchedulingIgnoredDuringExecution.preference`](#obj-spectemplatespecaffinitynodeaffinitypreferredduringschedulingignoredduringexecutionpreference)
+ * [`fn withMatchExpressions(matchExpressions)`](#fn-spectemplatespecaffinitynodeaffinitypreferredduringschedulingignoredduringexecutionpreferencewithmatchexpressions)
+ * [`fn withMatchExpressionsMixin(matchExpressions)`](#fn-spectemplatespecaffinitynodeaffinitypreferredduringschedulingignoredduringexecutionpreferencewithmatchexpressionsmixin)
+ * [`fn withMatchFields(matchFields)`](#fn-spectemplatespecaffinitynodeaffinitypreferredduringschedulingignoredduringexecutionpreferencewithmatchfields)
+ * [`fn withMatchFieldsMixin(matchFields)`](#fn-spectemplatespecaffinitynodeaffinitypreferredduringschedulingignoredduringexecutionpreferencewithmatchfieldsmixin)
+ * [`obj spec.template.spec.affinity.nodeAffinity.preferredDuringSchedulingIgnoredDuringExecution.preference.matchExpressions`](#obj-spectemplatespecaffinitynodeaffinitypreferredduringschedulingignoredduringexecutionpreferencematchexpressions)
+ * [`fn withKey(key)`](#fn-spectemplatespecaffinitynodeaffinitypreferredduringschedulingignoredduringexecutionpreferencematchexpressionswithkey)
+ * [`fn withOperator(operator)`](#fn-spectemplatespecaffinitynodeaffinitypreferredduringschedulingignoredduringexecutionpreferencematchexpressionswithoperator)
+ * [`fn withValues(values)`](#fn-spectemplatespecaffinitynodeaffinitypreferredduringschedulingignoredduringexecutionpreferencematchexpressionswithvalues)
+ * [`fn withValuesMixin(values)`](#fn-spectemplatespecaffinitynodeaffinitypreferredduringschedulingignoredduringexecutionpreferencematchexpressionswithvaluesmixin)
+ * [`obj spec.template.spec.affinity.nodeAffinity.preferredDuringSchedulingIgnoredDuringExecution.preference.matchFields`](#obj-spectemplatespecaffinitynodeaffinitypreferredduringschedulingignoredduringexecutionpreferencematchfields)
+ * [`fn withKey(key)`](#fn-spectemplatespecaffinitynodeaffinitypreferredduringschedulingignoredduringexecutionpreferencematchfieldswithkey)
+ * [`fn withOperator(operator)`](#fn-spectemplatespecaffinitynodeaffinitypreferredduringschedulingignoredduringexecutionpreferencematchfieldswithoperator)
+ * [`fn withValues(values)`](#fn-spectemplatespecaffinitynodeaffinitypreferredduringschedulingignoredduringexecutionpreferencematchfieldswithvalues)
+ * [`fn withValuesMixin(values)`](#fn-spectemplatespecaffinitynodeaffinitypreferredduringschedulingignoredduringexecutionpreferencematchfieldswithvaluesmixin)
+ * [`obj spec.template.spec.affinity.nodeAffinity.requiredDuringSchedulingIgnoredDuringExecution`](#obj-spectemplatespecaffinitynodeaffinityrequiredduringschedulingignoredduringexecution)
+ * [`fn withNodeSelectorTerms(nodeSelectorTerms)`](#fn-spectemplatespecaffinitynodeaffinityrequiredduringschedulingignoredduringexecutionwithnodeselectorterms)
+ * [`fn withNodeSelectorTermsMixin(nodeSelectorTerms)`](#fn-spectemplatespecaffinitynodeaffinityrequiredduringschedulingignoredduringexecutionwithnodeselectortermsmixin)
+ * [`obj spec.template.spec.affinity.nodeAffinity.requiredDuringSchedulingIgnoredDuringExecution.nodeSelectorTerms`](#obj-spectemplatespecaffinitynodeaffinityrequiredduringschedulingignoredduringexecutionnodeselectorterms)
+ * [`fn withMatchExpressions(matchExpressions)`](#fn-spectemplatespecaffinitynodeaffinityrequiredduringschedulingignoredduringexecutionnodeselectortermswithmatchexpressions)
+ * [`fn withMatchExpressionsMixin(matchExpressions)`](#fn-spectemplatespecaffinitynodeaffinityrequiredduringschedulingignoredduringexecutionnodeselectortermswithmatchexpressionsmixin)
+ * [`fn withMatchFields(matchFields)`](#fn-spectemplatespecaffinitynodeaffinityrequiredduringschedulingignoredduringexecutionnodeselectortermswithmatchfields)
+ * [`fn withMatchFieldsMixin(matchFields)`](#fn-spectemplatespecaffinitynodeaffinityrequiredduringschedulingignoredduringexecutionnodeselectortermswithmatchfieldsmixin)
+ * [`obj spec.template.spec.affinity.nodeAffinity.requiredDuringSchedulingIgnoredDuringExecution.nodeSelectorTerms.matchExpressions`](#obj-spectemplatespecaffinitynodeaffinityrequiredduringschedulingignoredduringexecutionnodeselectortermsmatchexpressions)
+ * [`fn withKey(key)`](#fn-spectemplatespecaffinitynodeaffinityrequiredduringschedulingignoredduringexecutionnodeselectortermsmatchexpressionswithkey)
+ * [`fn withOperator(operator)`](#fn-spectemplatespecaffinitynodeaffinityrequiredduringschedulingignoredduringexecutionnodeselectortermsmatchexpressionswithoperator)
+ * [`fn withValues(values)`](#fn-spectemplatespecaffinitynodeaffinityrequiredduringschedulingignoredduringexecutionnodeselectortermsmatchexpressionswithvalues)
+ * [`fn withValuesMixin(values)`](#fn-spectemplatespecaffinitynodeaffinityrequiredduringschedulingignoredduringexecutionnodeselectortermsmatchexpressionswithvaluesmixin)
+ * [`obj spec.template.spec.affinity.nodeAffinity.requiredDuringSchedulingIgnoredDuringExecution.nodeSelectorTerms.matchFields`](#obj-spectemplatespecaffinitynodeaffinityrequiredduringschedulingignoredduringexecutionnodeselectortermsmatchfields)
+ * [`fn withKey(key)`](#fn-spectemplatespecaffinitynodeaffinityrequiredduringschedulingignoredduringexecutionnodeselectortermsmatchfieldswithkey)
+ * [`fn withOperator(operator)`](#fn-spectemplatespecaffinitynodeaffinityrequiredduringschedulingignoredduringexecutionnodeselectortermsmatchfieldswithoperator)
+ * [`fn withValues(values)`](#fn-spectemplatespecaffinitynodeaffinityrequiredduringschedulingignoredduringexecutionnodeselectortermsmatchfieldswithvalues)
+ * [`fn withValuesMixin(values)`](#fn-spectemplatespecaffinitynodeaffinityrequiredduringschedulingignoredduringexecutionnodeselectortermsmatchfieldswithvaluesmixin)
+ * [`obj spec.template.spec.affinity.podAffinity`](#obj-spectemplatespecaffinitypodaffinity)
+ * [`fn withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution)`](#fn-spectemplatespecaffinitypodaffinitywithpreferredduringschedulingignoredduringexecution)
+ * [`fn withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution)`](#fn-spectemplatespecaffinitypodaffinitywithpreferredduringschedulingignoredduringexecutionmixin)
+ * [`fn withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution)`](#fn-spectemplatespecaffinitypodaffinitywithrequiredduringschedulingignoredduringexecution)
+ * [`fn withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution)`](#fn-spectemplatespecaffinitypodaffinitywithrequiredduringschedulingignoredduringexecutionmixin)
+ * [`obj spec.template.spec.affinity.podAffinity.preferredDuringSchedulingIgnoredDuringExecution`](#obj-spectemplatespecaffinitypodaffinitypreferredduringschedulingignoredduringexecution)
+ * [`fn withWeight(weight)`](#fn-spectemplatespecaffinitypodaffinitypreferredduringschedulingignoredduringexecutionwithweight)
+ * [`obj spec.template.spec.affinity.podAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm`](#obj-spectemplatespecaffinitypodaffinitypreferredduringschedulingignoredduringexecutionpodaffinityterm)
+ * [`fn withNamespaces(namespaces)`](#fn-spectemplatespecaffinitypodaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermwithnamespaces)
+ * [`fn withNamespacesMixin(namespaces)`](#fn-spectemplatespecaffinitypodaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermwithnamespacesmixin)
+ * [`fn withTopologyKey(topologyKey)`](#fn-spectemplatespecaffinitypodaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermwithtopologykey)
+ * [`obj spec.template.spec.affinity.podAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.labelSelector`](#obj-spectemplatespecaffinitypodaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermlabelselector)
+ * [`fn withMatchExpressions(matchExpressions)`](#fn-spectemplatespecaffinitypodaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermlabelselectorwithmatchexpressions)
+ * [`fn withMatchExpressionsMixin(matchExpressions)`](#fn-spectemplatespecaffinitypodaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermlabelselectorwithmatchexpressionsmixin)
+ * [`fn withMatchLabels(matchLabels)`](#fn-spectemplatespecaffinitypodaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermlabelselectorwithmatchlabels)
+ * [`fn withMatchLabelsMixin(matchLabels)`](#fn-spectemplatespecaffinitypodaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermlabelselectorwithmatchlabelsmixin)
+ * [`obj spec.template.spec.affinity.podAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.labelSelector.matchExpressions`](#obj-spectemplatespecaffinitypodaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermlabelselectormatchexpressions)
+ * [`fn withKey(key)`](#fn-spectemplatespecaffinitypodaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermlabelselectormatchexpressionswithkey)
+ * [`fn withOperator(operator)`](#fn-spectemplatespecaffinitypodaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermlabelselectormatchexpressionswithoperator)
+ * [`fn withValues(values)`](#fn-spectemplatespecaffinitypodaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermlabelselectormatchexpressionswithvalues)
+ * [`fn withValuesMixin(values)`](#fn-spectemplatespecaffinitypodaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermlabelselectormatchexpressionswithvaluesmixin)
+ * [`obj spec.template.spec.affinity.podAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.namespaceSelector`](#obj-spectemplatespecaffinitypodaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermnamespaceselector)
+ * [`fn withMatchExpressions(matchExpressions)`](#fn-spectemplatespecaffinitypodaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermnamespaceselectorwithmatchexpressions)
+ * [`fn withMatchExpressionsMixin(matchExpressions)`](#fn-spectemplatespecaffinitypodaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermnamespaceselectorwithmatchexpressionsmixin)
+ * [`fn withMatchLabels(matchLabels)`](#fn-spectemplatespecaffinitypodaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermnamespaceselectorwithmatchlabels)
+ * [`fn withMatchLabelsMixin(matchLabels)`](#fn-spectemplatespecaffinitypodaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermnamespaceselectorwithmatchlabelsmixin)
+ * [`obj spec.template.spec.affinity.podAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.namespaceSelector.matchExpressions`](#obj-spectemplatespecaffinitypodaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermnamespaceselectormatchexpressions)
+ * [`fn withKey(key)`](#fn-spectemplatespecaffinitypodaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermnamespaceselectormatchexpressionswithkey)
+ * [`fn withOperator(operator)`](#fn-spectemplatespecaffinitypodaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermnamespaceselectormatchexpressionswithoperator)
+ * [`fn withValues(values)`](#fn-spectemplatespecaffinitypodaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermnamespaceselectormatchexpressionswithvalues)
+ * [`fn withValuesMixin(values)`](#fn-spectemplatespecaffinitypodaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermnamespaceselectormatchexpressionswithvaluesmixin)
+ * [`obj spec.template.spec.affinity.podAffinity.requiredDuringSchedulingIgnoredDuringExecution`](#obj-spectemplatespecaffinitypodaffinityrequiredduringschedulingignoredduringexecution)
+ * [`fn withNamespaces(namespaces)`](#fn-spectemplatespecaffinitypodaffinityrequiredduringschedulingignoredduringexecutionwithnamespaces)
+ * [`fn withNamespacesMixin(namespaces)`](#fn-spectemplatespecaffinitypodaffinityrequiredduringschedulingignoredduringexecutionwithnamespacesmixin)
+ * [`fn withTopologyKey(topologyKey)`](#fn-spectemplatespecaffinitypodaffinityrequiredduringschedulingignoredduringexecutionwithtopologykey)
+ * [`obj spec.template.spec.affinity.podAffinity.requiredDuringSchedulingIgnoredDuringExecution.labelSelector`](#obj-spectemplatespecaffinitypodaffinityrequiredduringschedulingignoredduringexecutionlabelselector)
+ * [`fn withMatchExpressions(matchExpressions)`](#fn-spectemplatespecaffinitypodaffinityrequiredduringschedulingignoredduringexecutionlabelselectorwithmatchexpressions)
+ * [`fn withMatchExpressionsMixin(matchExpressions)`](#fn-spectemplatespecaffinitypodaffinityrequiredduringschedulingignoredduringexecutionlabelselectorwithmatchexpressionsmixin)
+ * [`fn withMatchLabels(matchLabels)`](#fn-spectemplatespecaffinitypodaffinityrequiredduringschedulingignoredduringexecutionlabelselectorwithmatchlabels)
+ * [`fn withMatchLabelsMixin(matchLabels)`](#fn-spectemplatespecaffinitypodaffinityrequiredduringschedulingignoredduringexecutionlabelselectorwithmatchlabelsmixin)
+ * [`obj spec.template.spec.affinity.podAffinity.requiredDuringSchedulingIgnoredDuringExecution.labelSelector.matchExpressions`](#obj-spectemplatespecaffinitypodaffinityrequiredduringschedulingignoredduringexecutionlabelselectormatchexpressions)
+ * [`fn withKey(key)`](#fn-spectemplatespecaffinitypodaffinityrequiredduringschedulingignoredduringexecutionlabelselectormatchexpressionswithkey)
+ * [`fn withOperator(operator)`](#fn-spectemplatespecaffinitypodaffinityrequiredduringschedulingignoredduringexecutionlabelselectormatchexpressionswithoperator)
+ * [`fn withValues(values)`](#fn-spectemplatespecaffinitypodaffinityrequiredduringschedulingignoredduringexecutionlabelselectormatchexpressionswithvalues)
+ * [`fn withValuesMixin(values)`](#fn-spectemplatespecaffinitypodaffinityrequiredduringschedulingignoredduringexecutionlabelselectormatchexpressionswithvaluesmixin)
+ * [`obj spec.template.spec.affinity.podAffinity.requiredDuringSchedulingIgnoredDuringExecution.namespaceSelector`](#obj-spectemplatespecaffinitypodaffinityrequiredduringschedulingignoredduringexecutionnamespaceselector)
+ * [`fn withMatchExpressions(matchExpressions)`](#fn-spectemplatespecaffinitypodaffinityrequiredduringschedulingignoredduringexecutionnamespaceselectorwithmatchexpressions)
+ * [`fn withMatchExpressionsMixin(matchExpressions)`](#fn-spectemplatespecaffinitypodaffinityrequiredduringschedulingignoredduringexecutionnamespaceselectorwithmatchexpressionsmixin)
+ * [`fn withMatchLabels(matchLabels)`](#fn-spectemplatespecaffinitypodaffinityrequiredduringschedulingignoredduringexecutionnamespaceselectorwithmatchlabels)
+ * [`fn withMatchLabelsMixin(matchLabels)`](#fn-spectemplatespecaffinitypodaffinityrequiredduringschedulingignoredduringexecutionnamespaceselectorwithmatchlabelsmixin)
+ * [`obj spec.template.spec.affinity.podAffinity.requiredDuringSchedulingIgnoredDuringExecution.namespaceSelector.matchExpressions`](#obj-spectemplatespecaffinitypodaffinityrequiredduringschedulingignoredduringexecutionnamespaceselectormatchexpressions)
+ * [`fn withKey(key)`](#fn-spectemplatespecaffinitypodaffinityrequiredduringschedulingignoredduringexecutionnamespaceselectormatchexpressionswithkey)
+ * [`fn withOperator(operator)`](#fn-spectemplatespecaffinitypodaffinityrequiredduringschedulingignoredduringexecutionnamespaceselectormatchexpressionswithoperator)
+ * [`fn withValues(values)`](#fn-spectemplatespecaffinitypodaffinityrequiredduringschedulingignoredduringexecutionnamespaceselectormatchexpressionswithvalues)
+ * [`fn withValuesMixin(values)`](#fn-spectemplatespecaffinitypodaffinityrequiredduringschedulingignoredduringexecutionnamespaceselectormatchexpressionswithvaluesmixin)
+ * [`obj spec.template.spec.affinity.podAntiAffinity`](#obj-spectemplatespecaffinitypodantiaffinity)
+ * [`fn withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution)`](#fn-spectemplatespecaffinitypodantiaffinitywithpreferredduringschedulingignoredduringexecution)
+ * [`fn withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution)`](#fn-spectemplatespecaffinitypodantiaffinitywithpreferredduringschedulingignoredduringexecutionmixin)
+ * [`fn withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution)`](#fn-spectemplatespecaffinitypodantiaffinitywithrequiredduringschedulingignoredduringexecution)
+ * [`fn withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution)`](#fn-spectemplatespecaffinitypodantiaffinitywithrequiredduringschedulingignoredduringexecutionmixin)
+ * [`obj spec.template.spec.affinity.podAntiAffinity.preferredDuringSchedulingIgnoredDuringExecution`](#obj-spectemplatespecaffinitypodantiaffinitypreferredduringschedulingignoredduringexecution)
+ * [`fn withWeight(weight)`](#fn-spectemplatespecaffinitypodantiaffinitypreferredduringschedulingignoredduringexecutionwithweight)
+ * [`obj spec.template.spec.affinity.podAntiAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm`](#obj-spectemplatespecaffinitypodantiaffinitypreferredduringschedulingignoredduringexecutionpodaffinityterm)
+ * [`fn withNamespaces(namespaces)`](#fn-spectemplatespecaffinitypodantiaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermwithnamespaces)
+ * [`fn withNamespacesMixin(namespaces)`](#fn-spectemplatespecaffinitypodantiaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermwithnamespacesmixin)
+ * [`fn withTopologyKey(topologyKey)`](#fn-spectemplatespecaffinitypodantiaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermwithtopologykey)
+ * [`obj spec.template.spec.affinity.podAntiAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.labelSelector`](#obj-spectemplatespecaffinitypodantiaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermlabelselector)
+ * [`fn withMatchExpressions(matchExpressions)`](#fn-spectemplatespecaffinitypodantiaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermlabelselectorwithmatchexpressions)
+ * [`fn withMatchExpressionsMixin(matchExpressions)`](#fn-spectemplatespecaffinitypodantiaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermlabelselectorwithmatchexpressionsmixin)
+ * [`fn withMatchLabels(matchLabels)`](#fn-spectemplatespecaffinitypodantiaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermlabelselectorwithmatchlabels)
+ * [`fn withMatchLabelsMixin(matchLabels)`](#fn-spectemplatespecaffinitypodantiaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermlabelselectorwithmatchlabelsmixin)
+ * [`obj spec.template.spec.affinity.podAntiAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.labelSelector.matchExpressions`](#obj-spectemplatespecaffinitypodantiaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermlabelselectormatchexpressions)
+ * [`fn withKey(key)`](#fn-spectemplatespecaffinitypodantiaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermlabelselectormatchexpressionswithkey)
+ * [`fn withOperator(operator)`](#fn-spectemplatespecaffinitypodantiaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermlabelselectormatchexpressionswithoperator)
+ * [`fn withValues(values)`](#fn-spectemplatespecaffinitypodantiaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermlabelselectormatchexpressionswithvalues)
+ * [`fn withValuesMixin(values)`](#fn-spectemplatespecaffinitypodantiaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermlabelselectormatchexpressionswithvaluesmixin)
+ * [`obj spec.template.spec.affinity.podAntiAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.namespaceSelector`](#obj-spectemplatespecaffinitypodantiaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermnamespaceselector)
+ * [`fn withMatchExpressions(matchExpressions)`](#fn-spectemplatespecaffinitypodantiaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermnamespaceselectorwithmatchexpressions)
+ * [`fn withMatchExpressionsMixin(matchExpressions)`](#fn-spectemplatespecaffinitypodantiaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermnamespaceselectorwithmatchexpressionsmixin)
+ * [`fn withMatchLabels(matchLabels)`](#fn-spectemplatespecaffinitypodantiaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermnamespaceselectorwithmatchlabels)
+ * [`fn withMatchLabelsMixin(matchLabels)`](#fn-spectemplatespecaffinitypodantiaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermnamespaceselectorwithmatchlabelsmixin)
+ * [`obj spec.template.spec.affinity.podAntiAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.namespaceSelector.matchExpressions`](#obj-spectemplatespecaffinitypodantiaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermnamespaceselectormatchexpressions)
+ * [`fn withKey(key)`](#fn-spectemplatespecaffinitypodantiaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermnamespaceselectormatchexpressionswithkey)
+ * [`fn withOperator(operator)`](#fn-spectemplatespecaffinitypodantiaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermnamespaceselectormatchexpressionswithoperator)
+ * [`fn withValues(values)`](#fn-spectemplatespecaffinitypodantiaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermnamespaceselectormatchexpressionswithvalues)
+ * [`fn withValuesMixin(values)`](#fn-spectemplatespecaffinitypodantiaffinitypreferredduringschedulingignoredduringexecutionpodaffinitytermnamespaceselectormatchexpressionswithvaluesmixin)
+ * [`obj spec.template.spec.affinity.podAntiAffinity.requiredDuringSchedulingIgnoredDuringExecution`](#obj-spectemplatespecaffinitypodantiaffinityrequiredduringschedulingignoredduringexecution)
+ * [`fn withNamespaces(namespaces)`](#fn-spectemplatespecaffinitypodantiaffinityrequiredduringschedulingignoredduringexecutionwithnamespaces)
+ * [`fn withNamespacesMixin(namespaces)`](#fn-spectemplatespecaffinitypodantiaffinityrequiredduringschedulingignoredduringexecutionwithnamespacesmixin)
+ * [`fn withTopologyKey(topologyKey)`](#fn-spectemplatespecaffinitypodantiaffinityrequiredduringschedulingignoredduringexecutionwithtopologykey)
+ * [`obj spec.template.spec.affinity.podAntiAffinity.requiredDuringSchedulingIgnoredDuringExecution.labelSelector`](#obj-spectemplatespecaffinitypodantiaffinityrequiredduringschedulingignoredduringexecutionlabelselector)
+ * [`fn withMatchExpressions(matchExpressions)`](#fn-spectemplatespecaffinitypodantiaffinityrequiredduringschedulingignoredduringexecutionlabelselectorwithmatchexpressions)
+ * [`fn withMatchExpressionsMixin(matchExpressions)`](#fn-spectemplatespecaffinitypodantiaffinityrequiredduringschedulingignoredduringexecutionlabelselectorwithmatchexpressionsmixin)
+ * [`fn withMatchLabels(matchLabels)`](#fn-spectemplatespecaffinitypodantiaffinityrequiredduringschedulingignoredduringexecutionlabelselectorwithmatchlabels)
+ * [`fn withMatchLabelsMixin(matchLabels)`](#fn-spectemplatespecaffinitypodantiaffinityrequiredduringschedulingignoredduringexecutionlabelselectorwithmatchlabelsmixin)
+ * [`obj spec.template.spec.affinity.podAntiAffinity.requiredDuringSchedulingIgnoredDuringExecution.labelSelector.matchExpressions`](#obj-spectemplatespecaffinitypodantiaffinityrequiredduringschedulingignoredduringexecutionlabelselectormatchexpressions)
+ * [`fn withKey(key)`](#fn-spectemplatespecaffinitypodantiaffinityrequiredduringschedulingignoredduringexecutionlabelselectormatchexpressionswithkey)
+ * [`fn withOperator(operator)`](#fn-spectemplatespecaffinitypodantiaffinityrequiredduringschedulingignoredduringexecutionlabelselectormatchexpressionswithoperator)
+ * [`fn withValues(values)`](#fn-spectemplatespecaffinitypodantiaffinityrequiredduringschedulingignoredduringexecutionlabelselectormatchexpressionswithvalues)
+ * [`fn withValuesMixin(values)`](#fn-spectemplatespecaffinitypodantiaffinityrequiredduringschedulingignoredduringexecutionlabelselectormatchexpressionswithvaluesmixin)
+ * [`obj spec.template.spec.affinity.podAntiAffinity.requiredDuringSchedulingIgnoredDuringExecution.namespaceSelector`](#obj-spectemplatespecaffinitypodantiaffinityrequiredduringschedulingignoredduringexecutionnamespaceselector)
+ * [`fn withMatchExpressions(matchExpressions)`](#fn-spectemplatespecaffinitypodantiaffinityrequiredduringschedulingignoredduringexecutionnamespaceselectorwithmatchexpressions)
+ * [`fn withMatchExpressionsMixin(matchExpressions)`](#fn-spectemplatespecaffinitypodantiaffinityrequiredduringschedulingignoredduringexecutionnamespaceselectorwithmatchexpressionsmixin)
+ * [`fn withMatchLabels(matchLabels)`](#fn-spectemplatespecaffinitypodantiaffinityrequiredduringschedulingignoredduringexecutionnamespaceselectorwithmatchlabels)
+ * [`fn withMatchLabelsMixin(matchLabels)`](#fn-spectemplatespecaffinitypodantiaffinityrequiredduringschedulingignoredduringexecutionnamespaceselectorwithmatchlabelsmixin)
+ * [`obj spec.template.spec.affinity.podAntiAffinity.requiredDuringSchedulingIgnoredDuringExecution.namespaceSelector.matchExpressions`](#obj-spectemplatespecaffinitypodantiaffinityrequiredduringschedulingignoredduringexecutionnamespaceselectormatchexpressions)
+ * [`fn withKey(key)`](#fn-spectemplatespecaffinitypodantiaffinityrequiredduringschedulingignoredduringexecutionnamespaceselectormatchexpressionswithkey)
+ * [`fn withOperator(operator)`](#fn-spectemplatespecaffinitypodantiaffinityrequiredduringschedulingignoredduringexecutionnamespaceselectormatchexpressionswithoperator)
+ * [`fn withValues(values)`](#fn-spectemplatespecaffinitypodantiaffinityrequiredduringschedulingignoredduringexecutionnamespaceselectormatchexpressionswithvalues)
+ * [`fn withValuesMixin(values)`](#fn-spectemplatespecaffinitypodantiaffinityrequiredduringschedulingignoredduringexecutionnamespaceselectormatchexpressionswithvaluesmixin)
+ * [`obj spec.template.spec.containers`](#obj-spectemplatespeccontainers)
+ * [`fn withArgs(args)`](#fn-spectemplatespeccontainerswithargs)
+ * [`fn withArgsMixin(args)`](#fn-spectemplatespeccontainerswithargsmixin)
+ * [`fn withCommand(command)`](#fn-spectemplatespeccontainerswithcommand)
+ * [`fn withCommandMixin(command)`](#fn-spectemplatespeccontainerswithcommandmixin)
+ * [`fn withEnv(env)`](#fn-spectemplatespeccontainerswithenv)
+ * [`fn withEnvFrom(envFrom)`](#fn-spectemplatespeccontainerswithenvfrom)
+ * [`fn withEnvFromMixin(envFrom)`](#fn-spectemplatespeccontainerswithenvfrommixin)
+ * [`fn withEnvMixin(env)`](#fn-spectemplatespeccontainerswithenvmixin)
+ * [`fn withImage(image)`](#fn-spectemplatespeccontainerswithimage)
+ * [`fn withImagePullPolicy(imagePullPolicy)`](#fn-spectemplatespeccontainerswithimagepullpolicy)
+ * [`fn withName(name)`](#fn-spectemplatespeccontainerswithname)
+ * [`fn withPorts(ports)`](#fn-spectemplatespeccontainerswithports)
+ * [`fn withPortsMixin(ports)`](#fn-spectemplatespeccontainerswithportsmixin)
+ * [`fn withResizePolicy(resizePolicy)`](#fn-spectemplatespeccontainerswithresizepolicy)
+ * [`fn withResizePolicyMixin(resizePolicy)`](#fn-spectemplatespeccontainerswithresizepolicymixin)
+ * [`fn withRestartPolicy(restartPolicy)`](#fn-spectemplatespeccontainerswithrestartpolicy)
+ * [`fn withStdin(stdin)`](#fn-spectemplatespeccontainerswithstdin)
+ * [`fn withStdinOnce(stdinOnce)`](#fn-spectemplatespeccontainerswithstdinonce)
+ * [`fn withTerminationMessagePath(terminationMessagePath)`](#fn-spectemplatespeccontainerswithterminationmessagepath)
+ * [`fn withTerminationMessagePolicy(terminationMessagePolicy)`](#fn-spectemplatespeccontainerswithterminationmessagepolicy)
+ * [`fn withTty(tty)`](#fn-spectemplatespeccontainerswithtty)
+ * [`fn withVolumeDevices(volumeDevices)`](#fn-spectemplatespeccontainerswithvolumedevices)
+ * [`fn withVolumeDevicesMixin(volumeDevices)`](#fn-spectemplatespeccontainerswithvolumedevicesmixin)
+ * [`fn withVolumeMounts(volumeMounts)`](#fn-spectemplatespeccontainerswithvolumemounts)
+ * [`fn withVolumeMountsMixin(volumeMounts)`](#fn-spectemplatespeccontainerswithvolumemountsmixin)
+ * [`fn withWorkingDir(workingDir)`](#fn-spectemplatespeccontainerswithworkingdir)
+ * [`obj spec.template.spec.containers.env`](#obj-spectemplatespeccontainersenv)
+ * [`fn withName(name)`](#fn-spectemplatespeccontainersenvwithname)
+ * [`fn withValue(value)`](#fn-spectemplatespeccontainersenvwithvalue)
+ * [`obj spec.template.spec.containers.env.valueFrom`](#obj-spectemplatespeccontainersenvvaluefrom)
+ * [`obj spec.template.spec.containers.env.valueFrom.configMapKeyRef`](#obj-spectemplatespeccontainersenvvaluefromconfigmapkeyref)
+ * [`fn withKey(key)`](#fn-spectemplatespeccontainersenvvaluefromconfigmapkeyrefwithkey)
+ * [`fn withName(name)`](#fn-spectemplatespeccontainersenvvaluefromconfigmapkeyrefwithname)
+ * [`fn withOptional(optional)`](#fn-spectemplatespeccontainersenvvaluefromconfigmapkeyrefwithoptional)
+ * [`obj spec.template.spec.containers.env.valueFrom.fieldRef`](#obj-spectemplatespeccontainersenvvaluefromfieldref)
+ * [`fn withApiVersion(apiVersion)`](#fn-spectemplatespeccontainersenvvaluefromfieldrefwithapiversion)
+ * [`fn withFieldPath(fieldPath)`](#fn-spectemplatespeccontainersenvvaluefromfieldrefwithfieldpath)
+ * [`obj spec.template.spec.containers.env.valueFrom.resourceFieldRef`](#obj-spectemplatespeccontainersenvvaluefromresourcefieldref)
+ * [`fn withContainerName(containerName)`](#fn-spectemplatespeccontainersenvvaluefromresourcefieldrefwithcontainername)
+ * [`fn withDivisor(divisor)`](#fn-spectemplatespeccontainersenvvaluefromresourcefieldrefwithdivisor)
+ * [`fn withResource(resource)`](#fn-spectemplatespeccontainersenvvaluefromresourcefieldrefwithresource)
+ * [`obj spec.template.spec.containers.env.valueFrom.secretKeyRef`](#obj-spectemplatespeccontainersenvvaluefromsecretkeyref)
+ * [`fn withKey(key)`](#fn-spectemplatespeccontainersenvvaluefromsecretkeyrefwithkey)
+ * [`fn withName(name)`](#fn-spectemplatespeccontainersenvvaluefromsecretkeyrefwithname)
+ * [`fn withOptional(optional)`](#fn-spectemplatespeccontainersenvvaluefromsecretkeyrefwithoptional)
+ * [`obj spec.template.spec.containers.envFrom`](#obj-spectemplatespeccontainersenvfrom)
+ * [`fn withPrefix(prefix)`](#fn-spectemplatespeccontainersenvfromwithprefix)
+ * [`obj spec.template.spec.containers.envFrom.configMapRef`](#obj-spectemplatespeccontainersenvfromconfigmapref)
+ * [`fn withName(name)`](#fn-spectemplatespeccontainersenvfromconfigmaprefwithname)
+ * [`fn withOptional(optional)`](#fn-spectemplatespeccontainersenvfromconfigmaprefwithoptional)
+ * [`obj spec.template.spec.containers.envFrom.secretRef`](#obj-spectemplatespeccontainersenvfromsecretref)
+ * [`fn withName(name)`](#fn-spectemplatespeccontainersenvfromsecretrefwithname)
+ * [`fn withOptional(optional)`](#fn-spectemplatespeccontainersenvfromsecretrefwithoptional)
+ * [`obj spec.template.spec.containers.lifecycle`](#obj-spectemplatespeccontainerslifecycle)
+ * [`obj spec.template.spec.containers.lifecycle.postStart`](#obj-spectemplatespeccontainerslifecyclepoststart)
+ * [`obj spec.template.spec.containers.lifecycle.postStart.exec`](#obj-spectemplatespeccontainerslifecyclepoststartexec)
+ * [`fn withCommand(command)`](#fn-spectemplatespeccontainerslifecyclepoststartexecwithcommand)
+ * [`fn withCommandMixin(command)`](#fn-spectemplatespeccontainerslifecyclepoststartexecwithcommandmixin)
+ * [`obj spec.template.spec.containers.lifecycle.postStart.httpGet`](#obj-spectemplatespeccontainerslifecyclepoststarthttpget)
+ * [`fn withHost(host)`](#fn-spectemplatespeccontainerslifecyclepoststarthttpgetwithhost)
+ * [`fn withHttpHeaders(httpHeaders)`](#fn-spectemplatespeccontainerslifecyclepoststarthttpgetwithhttpheaders)
+ * [`fn withHttpHeadersMixin(httpHeaders)`](#fn-spectemplatespeccontainerslifecyclepoststarthttpgetwithhttpheadersmixin)
+ * [`fn withPath(path)`](#fn-spectemplatespeccontainerslifecyclepoststarthttpgetwithpath)
+ * [`fn withPort(port)`](#fn-spectemplatespeccontainerslifecyclepoststarthttpgetwithport)
+ * [`fn withScheme(scheme)`](#fn-spectemplatespeccontainerslifecyclepoststarthttpgetwithscheme)
+ * [`obj spec.template.spec.containers.lifecycle.postStart.httpGet.httpHeaders`](#obj-spectemplatespeccontainerslifecyclepoststarthttpgethttpheaders)
+ * [`fn withName(name)`](#fn-spectemplatespeccontainerslifecyclepoststarthttpgethttpheaderswithname)
+ * [`fn withValue(value)`](#fn-spectemplatespeccontainerslifecyclepoststarthttpgethttpheaderswithvalue)
+ * [`obj spec.template.spec.containers.lifecycle.postStart.tcpSocket`](#obj-spectemplatespeccontainerslifecyclepoststarttcpsocket)
+ * [`fn withHost(host)`](#fn-spectemplatespeccontainerslifecyclepoststarttcpsocketwithhost)
+ * [`fn withPort(port)`](#fn-spectemplatespeccontainerslifecyclepoststarttcpsocketwithport)
+ * [`obj spec.template.spec.containers.lifecycle.preStop`](#obj-spectemplatespeccontainerslifecycleprestop)
+ * [`obj spec.template.spec.containers.lifecycle.preStop.exec`](#obj-spectemplatespeccontainerslifecycleprestopexec)
+ * [`fn withCommand(command)`](#fn-spectemplatespeccontainerslifecycleprestopexecwithcommand)
+ * [`fn withCommandMixin(command)`](#fn-spectemplatespeccontainerslifecycleprestopexecwithcommandmixin)
+ * [`obj spec.template.spec.containers.lifecycle.preStop.httpGet`](#obj-spectemplatespeccontainerslifecycleprestophttpget)
+ * [`fn withHost(host)`](#fn-spectemplatespeccontainerslifecycleprestophttpgetwithhost)
+ * [`fn withHttpHeaders(httpHeaders)`](#fn-spectemplatespeccontainerslifecycleprestophttpgetwithhttpheaders)
+ * [`fn withHttpHeadersMixin(httpHeaders)`](#fn-spectemplatespeccontainerslifecycleprestophttpgetwithhttpheadersmixin)
+ * [`fn withPath(path)`](#fn-spectemplatespeccontainerslifecycleprestophttpgetwithpath)
+ * [`fn withPort(port)`](#fn-spectemplatespeccontainerslifecycleprestophttpgetwithport)
+ * [`fn withScheme(scheme)`](#fn-spectemplatespeccontainerslifecycleprestophttpgetwithscheme)
+ * [`obj spec.template.spec.containers.lifecycle.preStop.httpGet.httpHeaders`](#obj-spectemplatespeccontainerslifecycleprestophttpgethttpheaders)
+ * [`fn withName(name)`](#fn-spectemplatespeccontainerslifecycleprestophttpgethttpheaderswithname)
+ * [`fn withValue(value)`](#fn-spectemplatespeccontainerslifecycleprestophttpgethttpheaderswithvalue)
+ * [`obj spec.template.spec.containers.lifecycle.preStop.tcpSocket`](#obj-spectemplatespeccontainerslifecycleprestoptcpsocket)
+ * [`fn withHost(host)`](#fn-spectemplatespeccontainerslifecycleprestoptcpsocketwithhost)
+ * [`fn withPort(port)`](#fn-spectemplatespeccontainerslifecycleprestoptcpsocketwithport)
+ * [`obj spec.template.spec.containers.livenessProbe`](#obj-spectemplatespeccontainerslivenessprobe)
+ * [`fn withFailureThreshold(failureThreshold)`](#fn-spectemplatespeccontainerslivenessprobewithfailurethreshold)
+ * [`fn withInitialDelaySeconds(initialDelaySeconds)`](#fn-spectemplatespeccontainerslivenessprobewithinitialdelayseconds)
+ * [`fn withPeriodSeconds(periodSeconds)`](#fn-spectemplatespeccontainerslivenessprobewithperiodseconds)
+ * [`fn withSuccessThreshold(successThreshold)`](#fn-spectemplatespeccontainerslivenessprobewithsuccessthreshold)
+ * [`fn withTerminationGracePeriodSeconds(terminationGracePeriodSeconds)`](#fn-spectemplatespeccontainerslivenessprobewithterminationgraceperiodseconds)
+ * [`fn withTimeoutSeconds(timeoutSeconds)`](#fn-spectemplatespeccontainerslivenessprobewithtimeoutseconds)
+ * [`obj spec.template.spec.containers.livenessProbe.exec`](#obj-spectemplatespeccontainerslivenessprobeexec)
+ * [`fn withCommand(command)`](#fn-spectemplatespeccontainerslivenessprobeexecwithcommand)
+ * [`fn withCommandMixin(command)`](#fn-spectemplatespeccontainerslivenessprobeexecwithcommandmixin)
+ * [`obj spec.template.spec.containers.livenessProbe.grpc`](#obj-spectemplatespeccontainerslivenessprobegrpc)
+ * [`fn withPort(port)`](#fn-spectemplatespeccontainerslivenessprobegrpcwithport)
+ * [`fn withService(service)`](#fn-spectemplatespeccontainerslivenessprobegrpcwithservice)
+ * [`obj spec.template.spec.containers.livenessProbe.httpGet`](#obj-spectemplatespeccontainerslivenessprobehttpget)
+ * [`fn withHost(host)`](#fn-spectemplatespeccontainerslivenessprobehttpgetwithhost)
+ * [`fn withHttpHeaders(httpHeaders)`](#fn-spectemplatespeccontainerslivenessprobehttpgetwithhttpheaders)
+ * [`fn withHttpHeadersMixin(httpHeaders)`](#fn-spectemplatespeccontainerslivenessprobehttpgetwithhttpheadersmixin)
+ * [`fn withPath(path)`](#fn-spectemplatespeccontainerslivenessprobehttpgetwithpath)
+ * [`fn withPort(port)`](#fn-spectemplatespeccontainerslivenessprobehttpgetwithport)
+ * [`fn withScheme(scheme)`](#fn-spectemplatespeccontainerslivenessprobehttpgetwithscheme)
+ * [`obj spec.template.spec.containers.livenessProbe.httpGet.httpHeaders`](#obj-spectemplatespeccontainerslivenessprobehttpgethttpheaders)
+ * [`fn withName(name)`](#fn-spectemplatespeccontainerslivenessprobehttpgethttpheaderswithname)
+ * [`fn withValue(value)`](#fn-spectemplatespeccontainerslivenessprobehttpgethttpheaderswithvalue)
+ * [`obj spec.template.spec.containers.livenessProbe.tcpSocket`](#obj-spectemplatespeccontainerslivenessprobetcpsocket)
+ * [`fn withHost(host)`](#fn-spectemplatespeccontainerslivenessprobetcpsocketwithhost)
+ * [`fn withPort(port)`](#fn-spectemplatespeccontainerslivenessprobetcpsocketwithport)
+ * [`obj spec.template.spec.containers.ports`](#obj-spectemplatespeccontainersports)
+ * [`fn withContainerPort(containerPort)`](#fn-spectemplatespeccontainersportswithcontainerport)
+ * [`fn withHostIP(hostIP)`](#fn-spectemplatespeccontainersportswithhostip)
+ * [`fn withHostPort(hostPort)`](#fn-spectemplatespeccontainersportswithhostport)
+ * [`fn withName(name)`](#fn-spectemplatespeccontainersportswithname)
+ * [`fn withProtocol(protocol)`](#fn-spectemplatespeccontainersportswithprotocol)
+ * [`obj spec.template.spec.containers.readinessProbe`](#obj-spectemplatespeccontainersreadinessprobe)
+ * [`fn withFailureThreshold(failureThreshold)`](#fn-spectemplatespeccontainersreadinessprobewithfailurethreshold)
+ * [`fn withInitialDelaySeconds(initialDelaySeconds)`](#fn-spectemplatespeccontainersreadinessprobewithinitialdelayseconds)
+ * [`fn withPeriodSeconds(periodSeconds)`](#fn-spectemplatespeccontainersreadinessprobewithperiodseconds)
+ * [`fn withSuccessThreshold(successThreshold)`](#fn-spectemplatespeccontainersreadinessprobewithsuccessthreshold)
+ * [`fn withTerminationGracePeriodSeconds(terminationGracePeriodSeconds)`](#fn-spectemplatespeccontainersreadinessprobewithterminationgraceperiodseconds)
+ * [`fn withTimeoutSeconds(timeoutSeconds)`](#fn-spectemplatespeccontainersreadinessprobewithtimeoutseconds)
+ * [`obj spec.template.spec.containers.readinessProbe.exec`](#obj-spectemplatespeccontainersreadinessprobeexec)
+ * [`fn withCommand(command)`](#fn-spectemplatespeccontainersreadinessprobeexecwithcommand)
+ * [`fn withCommandMixin(command)`](#fn-spectemplatespeccontainersreadinessprobeexecwithcommandmixin)
+ * [`obj spec.template.spec.containers.readinessProbe.grpc`](#obj-spectemplatespeccontainersreadinessprobegrpc)
+ * [`fn withPort(port)`](#fn-spectemplatespeccontainersreadinessprobegrpcwithport)
+ * [`fn withService(service)`](#fn-spectemplatespeccontainersreadinessprobegrpcwithservice)
+ * [`obj spec.template.spec.containers.readinessProbe.httpGet`](#obj-spectemplatespeccontainersreadinessprobehttpget)
+ * [`fn withHost(host)`](#fn-spectemplatespeccontainersreadinessprobehttpgetwithhost)
+ * [`fn withHttpHeaders(httpHeaders)`](#fn-spectemplatespeccontainersreadinessprobehttpgetwithhttpheaders)
+ * [`fn withHttpHeadersMixin(httpHeaders)`](#fn-spectemplatespeccontainersreadinessprobehttpgetwithhttpheadersmixin)
+ * [`fn withPath(path)`](#fn-spectemplatespeccontainersreadinessprobehttpgetwithpath)
+ * [`fn withPort(port)`](#fn-spectemplatespeccontainersreadinessprobehttpgetwithport)
+ * [`fn withScheme(scheme)`](#fn-spectemplatespeccontainersreadinessprobehttpgetwithscheme)
+ * [`obj spec.template.spec.containers.readinessProbe.httpGet.httpHeaders`](#obj-spectemplatespeccontainersreadinessprobehttpgethttpheaders)
+ * [`fn withName(name)`](#fn-spectemplatespeccontainersreadinessprobehttpgethttpheaderswithname)
+ * [`fn withValue(value)`](#fn-spectemplatespeccontainersreadinessprobehttpgethttpheaderswithvalue)
+ * [`obj spec.template.spec.containers.readinessProbe.tcpSocket`](#obj-spectemplatespeccontainersreadinessprobetcpsocket)
+ * [`fn withHost(host)`](#fn-spectemplatespeccontainersreadinessprobetcpsocketwithhost)
+ * [`fn withPort(port)`](#fn-spectemplatespeccontainersreadinessprobetcpsocketwithport)
+ * [`obj spec.template.spec.containers.resizePolicy`](#obj-spectemplatespeccontainersresizepolicy)
+ * [`fn withResourceName(resourceName)`](#fn-spectemplatespeccontainersresizepolicywithresourcename)
+ * [`fn withRestartPolicy(restartPolicy)`](#fn-spectemplatespeccontainersresizepolicywithrestartpolicy)
+ * [`obj spec.template.spec.containers.resources`](#obj-spectemplatespeccontainersresources)
+ * [`fn withClaims(claims)`](#fn-spectemplatespeccontainersresourceswithclaims)
+ * [`fn withClaimsMixin(claims)`](#fn-spectemplatespeccontainersresourceswithclaimsmixin)
+ * [`fn withLimits(limits)`](#fn-spectemplatespeccontainersresourceswithlimits)
+ * [`fn withLimitsMixin(limits)`](#fn-spectemplatespeccontainersresourceswithlimitsmixin)
+ * [`fn withRequests(requests)`](#fn-spectemplatespeccontainersresourceswithrequests)
+ * [`fn withRequestsMixin(requests)`](#fn-spectemplatespeccontainersresourceswithrequestsmixin)
+ * [`obj spec.template.spec.containers.resources.claims`](#obj-spectemplatespeccontainersresourcesclaims)
+ * [`fn withName(name)`](#fn-spectemplatespeccontainersresourcesclaimswithname)
+ * [`obj spec.template.spec.containers.securityContext`](#obj-spectemplatespeccontainerssecuritycontext)
+ * [`fn withAllowPrivilegeEscalation(allowPrivilegeEscalation)`](#fn-spectemplatespeccontainerssecuritycontextwithallowprivilegeescalation)
+ * [`fn withPrivileged(privileged)`](#fn-spectemplatespeccontainerssecuritycontextwithprivileged)
+ * [`fn withProcMount(procMount)`](#fn-spectemplatespeccontainerssecuritycontextwithprocmount)
+ * [`fn withReadOnlyRootFilesystem(readOnlyRootFilesystem)`](#fn-spectemplatespeccontainerssecuritycontextwithreadonlyrootfilesystem)
+ * [`fn withRunAsGroup(runAsGroup)`](#fn-spectemplatespeccontainerssecuritycontextwithrunasgroup)
+ * [`fn withRunAsNonRoot(runAsNonRoot)`](#fn-spectemplatespeccontainerssecuritycontextwithrunasnonroot)
+ * [`fn withRunAsUser(runAsUser)`](#fn-spectemplatespeccontainerssecuritycontextwithrunasuser)
+ * [`obj spec.template.spec.containers.securityContext.capabilities`](#obj-spectemplatespeccontainerssecuritycontextcapabilities)
+ * [`fn withAdd(add)`](#fn-spectemplatespeccontainerssecuritycontextcapabilitieswithadd)
+ * [`fn withAddMixin(add)`](#fn-spectemplatespeccontainerssecuritycontextcapabilitieswithaddmixin)
+ * [`fn withDrop(drop)`](#fn-spectemplatespeccontainerssecuritycontextcapabilitieswithdrop)
+ * [`fn withDropMixin(drop)`](#fn-spectemplatespeccontainerssecuritycontextcapabilitieswithdropmixin)
+ * [`obj spec.template.spec.containers.securityContext.seLinuxOptions`](#obj-spectemplatespeccontainerssecuritycontextselinuxoptions)
+ * [`fn withLevel(level)`](#fn-spectemplatespeccontainerssecuritycontextselinuxoptionswithlevel)
+ * [`fn withRole(role)`](#fn-spectemplatespeccontainerssecuritycontextselinuxoptionswithrole)
+ * [`fn withType(type)`](#fn-spectemplatespeccontainerssecuritycontextselinuxoptionswithtype)
+ * [`fn withUser(user)`](#fn-spectemplatespeccontainerssecuritycontextselinuxoptionswithuser)
+ * [`obj spec.template.spec.containers.securityContext.seccompProfile`](#obj-spectemplatespeccontainerssecuritycontextseccompprofile)
+ * [`fn withLocalhostProfile(localhostProfile)`](#fn-spectemplatespeccontainerssecuritycontextseccompprofilewithlocalhostprofile)
+ * [`fn withType(type)`](#fn-spectemplatespeccontainerssecuritycontextseccompprofilewithtype)
+ * [`obj spec.template.spec.containers.securityContext.windowsOptions`](#obj-spectemplatespeccontainerssecuritycontextwindowsoptions)
+ * [`fn withGmsaCredentialSpec(gmsaCredentialSpec)`](#fn-spectemplatespeccontainerssecuritycontextwindowsoptionswithgmsacredentialspec)
+ * [`fn withGmsaCredentialSpecName(gmsaCredentialSpecName)`](#fn-spectemplatespeccontainerssecuritycontextwindowsoptionswithgmsacredentialspecname)
+ * [`fn withHostProcess(hostProcess)`](#fn-spectemplatespeccontainerssecuritycontextwindowsoptionswithhostprocess)
+ * [`fn withRunAsUserName(runAsUserName)`](#fn-spectemplatespeccontainerssecuritycontextwindowsoptionswithrunasusername)
+ * [`obj spec.template.spec.containers.startupProbe`](#obj-spectemplatespeccontainersstartupprobe)
+ * [`fn withFailureThreshold(failureThreshold)`](#fn-spectemplatespeccontainersstartupprobewithfailurethreshold)
+ * [`fn withInitialDelaySeconds(initialDelaySeconds)`](#fn-spectemplatespeccontainersstartupprobewithinitialdelayseconds)
+ * [`fn withPeriodSeconds(periodSeconds)`](#fn-spectemplatespeccontainersstartupprobewithperiodseconds)
+ * [`fn withSuccessThreshold(successThreshold)`](#fn-spectemplatespeccontainersstartupprobewithsuccessthreshold)
+ * [`fn withTerminationGracePeriodSeconds(terminationGracePeriodSeconds)`](#fn-spectemplatespeccontainersstartupprobewithterminationgraceperiodseconds)
+ * [`fn withTimeoutSeconds(timeoutSeconds)`](#fn-spectemplatespeccontainersstartupprobewithtimeoutseconds)
+ * [`obj spec.template.spec.containers.startupProbe.exec`](#obj-spectemplatespeccontainersstartupprobeexec)
+ * [`fn withCommand(command)`](#fn-spectemplatespeccontainersstartupprobeexecwithcommand)
+ * [`fn withCommandMixin(command)`](#fn-spectemplatespeccontainersstartupprobeexecwithcommandmixin)
+ * [`obj spec.template.spec.containers.startupProbe.grpc`](#obj-spectemplatespeccontainersstartupprobegrpc)
+ * [`fn withPort(port)`](#fn-spectemplatespeccontainersstartupprobegrpcwithport)
+ * [`fn withService(service)`](#fn-spectemplatespeccontainersstartupprobegrpcwithservice)
+ * [`obj spec.template.spec.containers.startupProbe.httpGet`](#obj-spectemplatespeccontainersstartupprobehttpget)
+ * [`fn withHost(host)`](#fn-spectemplatespeccontainersstartupprobehttpgetwithhost)
+ * [`fn withHttpHeaders(httpHeaders)`](#fn-spectemplatespeccontainersstartupprobehttpgetwithhttpheaders)
+ * [`fn withHttpHeadersMixin(httpHeaders)`](#fn-spectemplatespeccontainersstartupprobehttpgetwithhttpheadersmixin)
+ * [`fn withPath(path)`](#fn-spectemplatespeccontainersstartupprobehttpgetwithpath)
+ * [`fn withPort(port)`](#fn-spectemplatespeccontainersstartupprobehttpgetwithport)
+ * [`fn withScheme(scheme)`](#fn-spectemplatespeccontainersstartupprobehttpgetwithscheme)
+ * [`obj spec.template.spec.containers.startupProbe.httpGet.httpHeaders`](#obj-spectemplatespeccontainersstartupprobehttpgethttpheaders)
+ * [`fn withName(name)`](#fn-spectemplatespeccontainersstartupprobehttpgethttpheaderswithname)
+ * [`fn withValue(value)`](#fn-spectemplatespeccontainersstartupprobehttpgethttpheaderswithvalue)
+ * [`obj spec.template.spec.containers.startupProbe.tcpSocket`](#obj-spectemplatespeccontainersstartupprobetcpsocket)
+ * [`fn withHost(host)`](#fn-spectemplatespeccontainersstartupprobetcpsocketwithhost)
+ * [`fn withPort(port)`](#fn-spectemplatespeccontainersstartupprobetcpsocketwithport)
+ * [`obj spec.template.spec.containers.volumeDevices`](#obj-spectemplatespeccontainersvolumedevices)
+ * [`fn withDevicePath(devicePath)`](#fn-spectemplatespeccontainersvolumedeviceswithdevicepath)
+ * [`fn withName(name)`](#fn-spectemplatespeccontainersvolumedeviceswithname)
+ * [`obj spec.template.spec.containers.volumeMounts`](#obj-spectemplatespeccontainersvolumemounts)
+ * [`fn withMountPath(mountPath)`](#fn-spectemplatespeccontainersvolumemountswithmountpath)
+ * [`fn withMountPropagation(mountPropagation)`](#fn-spectemplatespeccontainersvolumemountswithmountpropagation)
+ * [`fn withName(name)`](#fn-spectemplatespeccontainersvolumemountswithname)
+ * [`fn withReadOnly(readOnly)`](#fn-spectemplatespeccontainersvolumemountswithreadonly)
+ * [`fn withSubPath(subPath)`](#fn-spectemplatespeccontainersvolumemountswithsubpath)
+ * [`fn withSubPathExpr(subPathExpr)`](#fn-spectemplatespeccontainersvolumemountswithsubpathexpr)
+ * [`obj spec.template.spec.dnsConfig`](#obj-spectemplatespecdnsconfig)
+ * [`fn withNameservers(nameservers)`](#fn-spectemplatespecdnsconfigwithnameservers)
+ * [`fn withNameserversMixin(nameservers)`](#fn-spectemplatespecdnsconfigwithnameserversmixin)
+ * [`fn withOptions(options)`](#fn-spectemplatespecdnsconfigwithoptions)
+ * [`fn withOptionsMixin(options)`](#fn-spectemplatespecdnsconfigwithoptionsmixin)
+ * [`fn withSearches(searches)`](#fn-spectemplatespecdnsconfigwithsearches)
+ * [`fn withSearchesMixin(searches)`](#fn-spectemplatespecdnsconfigwithsearchesmixin)
+ * [`obj spec.template.spec.dnsConfig.options`](#obj-spectemplatespecdnsconfigoptions)
+ * [`fn withName(name)`](#fn-spectemplatespecdnsconfigoptionswithname)
+ * [`fn withValue(value)`](#fn-spectemplatespecdnsconfigoptionswithvalue)
+ * [`obj spec.template.spec.ephemeralContainers`](#obj-spectemplatespecephemeralcontainers)
+ * [`fn withArgs(args)`](#fn-spectemplatespecephemeralcontainerswithargs)
+ * [`fn withArgsMixin(args)`](#fn-spectemplatespecephemeralcontainerswithargsmixin)
+ * [`fn withCommand(command)`](#fn-spectemplatespecephemeralcontainerswithcommand)
+ * [`fn withCommandMixin(command)`](#fn-spectemplatespecephemeralcontainerswithcommandmixin)
+ * [`fn withEnv(env)`](#fn-spectemplatespecephemeralcontainerswithenv)
+ * [`fn withEnvFrom(envFrom)`](#fn-spectemplatespecephemeralcontainerswithenvfrom)
+ * [`fn withEnvFromMixin(envFrom)`](#fn-spectemplatespecephemeralcontainerswithenvfrommixin)
+ * [`fn withEnvMixin(env)`](#fn-spectemplatespecephemeralcontainerswithenvmixin)
+ * [`fn withImage(image)`](#fn-spectemplatespecephemeralcontainerswithimage)
+ * [`fn withImagePullPolicy(imagePullPolicy)`](#fn-spectemplatespecephemeralcontainerswithimagepullpolicy)
+ * [`fn withName(name)`](#fn-spectemplatespecephemeralcontainerswithname)
+ * [`fn withPorts(ports)`](#fn-spectemplatespecephemeralcontainerswithports)
+ * [`fn withPortsMixin(ports)`](#fn-spectemplatespecephemeralcontainerswithportsmixin)
+ * [`fn withResizePolicy(resizePolicy)`](#fn-spectemplatespecephemeralcontainerswithresizepolicy)
+ * [`fn withResizePolicyMixin(resizePolicy)`](#fn-spectemplatespecephemeralcontainerswithresizepolicymixin)
+ * [`fn withRestartPolicy(restartPolicy)`](#fn-spectemplatespecephemeralcontainerswithrestartpolicy)
+ * [`fn withStdin(stdin)`](#fn-spectemplatespecephemeralcontainerswithstdin)
+ * [`fn withStdinOnce(stdinOnce)`](#fn-spectemplatespecephemeralcontainerswithstdinonce)
+ * [`fn withTargetContainerName(targetContainerName)`](#fn-spectemplatespecephemeralcontainerswithtargetcontainername)
+ * [`fn withTerminationMessagePath(terminationMessagePath)`](#fn-spectemplatespecephemeralcontainerswithterminationmessagepath)
+ * [`fn withTerminationMessagePolicy(terminationMessagePolicy)`](#fn-spectemplatespecephemeralcontainerswithterminationmessagepolicy)
+ * [`fn withTty(tty)`](#fn-spectemplatespecephemeralcontainerswithtty)
+ * [`fn withVolumeDevices(volumeDevices)`](#fn-spectemplatespecephemeralcontainerswithvolumedevices)
+ * [`fn withVolumeDevicesMixin(volumeDevices)`](#fn-spectemplatespecephemeralcontainerswithvolumedevicesmixin)
+ * [`fn withVolumeMounts(volumeMounts)`](#fn-spectemplatespecephemeralcontainerswithvolumemounts)
+ * [`fn withVolumeMountsMixin(volumeMounts)`](#fn-spectemplatespecephemeralcontainerswithvolumemountsmixin)
+ * [`fn withWorkingDir(workingDir)`](#fn-spectemplatespecephemeralcontainerswithworkingdir)
+ * [`obj spec.template.spec.ephemeralContainers.env`](#obj-spectemplatespecephemeralcontainersenv)
+ * [`fn withName(name)`](#fn-spectemplatespecephemeralcontainersenvwithname)
+ * [`fn withValue(value)`](#fn-spectemplatespecephemeralcontainersenvwithvalue)
+ * [`obj spec.template.spec.ephemeralContainers.env.valueFrom`](#obj-spectemplatespecephemeralcontainersenvvaluefrom)
+ * [`obj spec.template.spec.ephemeralContainers.env.valueFrom.configMapKeyRef`](#obj-spectemplatespecephemeralcontainersenvvaluefromconfigmapkeyref)
+ * [`fn withKey(key)`](#fn-spectemplatespecephemeralcontainersenvvaluefromconfigmapkeyrefwithkey)
+ * [`fn withName(name)`](#fn-spectemplatespecephemeralcontainersenvvaluefromconfigmapkeyrefwithname)
+ * [`fn withOptional(optional)`](#fn-spectemplatespecephemeralcontainersenvvaluefromconfigmapkeyrefwithoptional)
+ * [`obj spec.template.spec.ephemeralContainers.env.valueFrom.fieldRef`](#obj-spectemplatespecephemeralcontainersenvvaluefromfieldref)
+ * [`fn withApiVersion(apiVersion)`](#fn-spectemplatespecephemeralcontainersenvvaluefromfieldrefwithapiversion)
+ * [`fn withFieldPath(fieldPath)`](#fn-spectemplatespecephemeralcontainersenvvaluefromfieldrefwithfieldpath)
+ * [`obj spec.template.spec.ephemeralContainers.env.valueFrom.resourceFieldRef`](#obj-spectemplatespecephemeralcontainersenvvaluefromresourcefieldref)
+ * [`fn withContainerName(containerName)`](#fn-spectemplatespecephemeralcontainersenvvaluefromresourcefieldrefwithcontainername)
+ * [`fn withDivisor(divisor)`](#fn-spectemplatespecephemeralcontainersenvvaluefromresourcefieldrefwithdivisor)
+ * [`fn withResource(resource)`](#fn-spectemplatespecephemeralcontainersenvvaluefromresourcefieldrefwithresource)
+ * [`obj spec.template.spec.ephemeralContainers.env.valueFrom.secretKeyRef`](#obj-spectemplatespecephemeralcontainersenvvaluefromsecretkeyref)
+ * [`fn withKey(key)`](#fn-spectemplatespecephemeralcontainersenvvaluefromsecretkeyrefwithkey)
+ * [`fn withName(name)`](#fn-spectemplatespecephemeralcontainersenvvaluefromsecretkeyrefwithname)
+ * [`fn withOptional(optional)`](#fn-spectemplatespecephemeralcontainersenvvaluefromsecretkeyrefwithoptional)
+ * [`obj spec.template.spec.ephemeralContainers.envFrom`](#obj-spectemplatespecephemeralcontainersenvfrom)
+ * [`fn withPrefix(prefix)`](#fn-spectemplatespecephemeralcontainersenvfromwithprefix)
+ * [`obj spec.template.spec.ephemeralContainers.envFrom.configMapRef`](#obj-spectemplatespecephemeralcontainersenvfromconfigmapref)
+ * [`fn withName(name)`](#fn-spectemplatespecephemeralcontainersenvfromconfigmaprefwithname)
+ * [`fn withOptional(optional)`](#fn-spectemplatespecephemeralcontainersenvfromconfigmaprefwithoptional)
+ * [`obj spec.template.spec.ephemeralContainers.envFrom.secretRef`](#obj-spectemplatespecephemeralcontainersenvfromsecretref)
+ * [`fn withName(name)`](#fn-spectemplatespecephemeralcontainersenvfromsecretrefwithname)
+ * [`fn withOptional(optional)`](#fn-spectemplatespecephemeralcontainersenvfromsecretrefwithoptional)
+ * [`obj spec.template.spec.ephemeralContainers.lifecycle`](#obj-spectemplatespecephemeralcontainerslifecycle)
+ * [`obj spec.template.spec.ephemeralContainers.lifecycle.postStart`](#obj-spectemplatespecephemeralcontainerslifecyclepoststart)
+ * [`obj spec.template.spec.ephemeralContainers.lifecycle.postStart.exec`](#obj-spectemplatespecephemeralcontainerslifecyclepoststartexec)
+ * [`fn withCommand(command)`](#fn-spectemplatespecephemeralcontainerslifecyclepoststartexecwithcommand)
+ * [`fn withCommandMixin(command)`](#fn-spectemplatespecephemeralcontainerslifecyclepoststartexecwithcommandmixin)
+ * [`obj spec.template.spec.ephemeralContainers.lifecycle.postStart.httpGet`](#obj-spectemplatespecephemeralcontainerslifecyclepoststarthttpget)
+ * [`fn withHost(host)`](#fn-spectemplatespecephemeralcontainerslifecyclepoststarthttpgetwithhost)
+ * [`fn withHttpHeaders(httpHeaders)`](#fn-spectemplatespecephemeralcontainerslifecyclepoststarthttpgetwithhttpheaders)
+ * [`fn withHttpHeadersMixin(httpHeaders)`](#fn-spectemplatespecephemeralcontainerslifecyclepoststarthttpgetwithhttpheadersmixin)
+ * [`fn withPath(path)`](#fn-spectemplatespecephemeralcontainerslifecyclepoststarthttpgetwithpath)
+ * [`fn withPort(port)`](#fn-spectemplatespecephemeralcontainerslifecyclepoststarthttpgetwithport)
+ * [`fn withScheme(scheme)`](#fn-spectemplatespecephemeralcontainerslifecyclepoststarthttpgetwithscheme)
+ * [`obj spec.template.spec.ephemeralContainers.lifecycle.postStart.httpGet.httpHeaders`](#obj-spectemplatespecephemeralcontainerslifecyclepoststarthttpgethttpheaders)
+ * [`fn withName(name)`](#fn-spectemplatespecephemeralcontainerslifecyclepoststarthttpgethttpheaderswithname)
+ * [`fn withValue(value)`](#fn-spectemplatespecephemeralcontainerslifecyclepoststarthttpgethttpheaderswithvalue)
+ * [`obj spec.template.spec.ephemeralContainers.lifecycle.postStart.tcpSocket`](#obj-spectemplatespecephemeralcontainerslifecyclepoststarttcpsocket)
+ * [`fn withHost(host)`](#fn-spectemplatespecephemeralcontainerslifecyclepoststarttcpsocketwithhost)
+ * [`fn withPort(port)`](#fn-spectemplatespecephemeralcontainerslifecyclepoststarttcpsocketwithport)
+ * [`obj spec.template.spec.ephemeralContainers.lifecycle.preStop`](#obj-spectemplatespecephemeralcontainerslifecycleprestop)
+ * [`obj spec.template.spec.ephemeralContainers.lifecycle.preStop.exec`](#obj-spectemplatespecephemeralcontainerslifecycleprestopexec)
+ * [`fn withCommand(command)`](#fn-spectemplatespecephemeralcontainerslifecycleprestopexecwithcommand)
+ * [`fn withCommandMixin(command)`](#fn-spectemplatespecephemeralcontainerslifecycleprestopexecwithcommandmixin)
+ * [`obj spec.template.spec.ephemeralContainers.lifecycle.preStop.httpGet`](#obj-spectemplatespecephemeralcontainerslifecycleprestophttpget)
+ * [`fn withHost(host)`](#fn-spectemplatespecephemeralcontainerslifecycleprestophttpgetwithhost)
+ * [`fn withHttpHeaders(httpHeaders)`](#fn-spectemplatespecephemeralcontainerslifecycleprestophttpgetwithhttpheaders)
+ * [`fn withHttpHeadersMixin(httpHeaders)`](#fn-spectemplatespecephemeralcontainerslifecycleprestophttpgetwithhttpheadersmixin)
+ * [`fn withPath(path)`](#fn-spectemplatespecephemeralcontainerslifecycleprestophttpgetwithpath)
+ * [`fn withPort(port)`](#fn-spectemplatespecephemeralcontainerslifecycleprestophttpgetwithport)
+ * [`fn withScheme(scheme)`](#fn-spectemplatespecephemeralcontainerslifecycleprestophttpgetwithscheme)
+ * [`obj spec.template.spec.ephemeralContainers.lifecycle.preStop.httpGet.httpHeaders`](#obj-spectemplatespecephemeralcontainerslifecycleprestophttpgethttpheaders)
+ * [`fn withName(name)`](#fn-spectemplatespecephemeralcontainerslifecycleprestophttpgethttpheaderswithname)
+ * [`fn withValue(value)`](#fn-spectemplatespecephemeralcontainerslifecycleprestophttpgethttpheaderswithvalue)
+ * [`obj spec.template.spec.ephemeralContainers.lifecycle.preStop.tcpSocket`](#obj-spectemplatespecephemeralcontainerslifecycleprestoptcpsocket)
+ * [`fn withHost(host)`](#fn-spectemplatespecephemeralcontainerslifecycleprestoptcpsocketwithhost)
+ * [`fn withPort(port)`](#fn-spectemplatespecephemeralcontainerslifecycleprestoptcpsocketwithport)
+ * [`obj spec.template.spec.ephemeralContainers.livenessProbe`](#obj-spectemplatespecephemeralcontainerslivenessprobe)
+ * [`fn withFailureThreshold(failureThreshold)`](#fn-spectemplatespecephemeralcontainerslivenessprobewithfailurethreshold)
+ * [`fn withInitialDelaySeconds(initialDelaySeconds)`](#fn-spectemplatespecephemeralcontainerslivenessprobewithinitialdelayseconds)
+ * [`fn withPeriodSeconds(periodSeconds)`](#fn-spectemplatespecephemeralcontainerslivenessprobewithperiodseconds)
+ * [`fn withSuccessThreshold(successThreshold)`](#fn-spectemplatespecephemeralcontainerslivenessprobewithsuccessthreshold)
+ * [`fn withTerminationGracePeriodSeconds(terminationGracePeriodSeconds)`](#fn-spectemplatespecephemeralcontainerslivenessprobewithterminationgraceperiodseconds)
+ * [`fn withTimeoutSeconds(timeoutSeconds)`](#fn-spectemplatespecephemeralcontainerslivenessprobewithtimeoutseconds)
+ * [`obj spec.template.spec.ephemeralContainers.livenessProbe.exec`](#obj-spectemplatespecephemeralcontainerslivenessprobeexec)
+ * [`fn withCommand(command)`](#fn-spectemplatespecephemeralcontainerslivenessprobeexecwithcommand)
+ * [`fn withCommandMixin(command)`](#fn-spectemplatespecephemeralcontainerslivenessprobeexecwithcommandmixin)
+ * [`obj spec.template.spec.ephemeralContainers.livenessProbe.grpc`](#obj-spectemplatespecephemeralcontainerslivenessprobegrpc)
+ * [`fn withPort(port)`](#fn-spectemplatespecephemeralcontainerslivenessprobegrpcwithport)
+ * [`fn withService(service)`](#fn-spectemplatespecephemeralcontainerslivenessprobegrpcwithservice)
+ * [`obj spec.template.spec.ephemeralContainers.livenessProbe.httpGet`](#obj-spectemplatespecephemeralcontainerslivenessprobehttpget)
+ * [`fn withHost(host)`](#fn-spectemplatespecephemeralcontainerslivenessprobehttpgetwithhost)
+ * [`fn withHttpHeaders(httpHeaders)`](#fn-spectemplatespecephemeralcontainerslivenessprobehttpgetwithhttpheaders)
+ * [`fn withHttpHeadersMixin(httpHeaders)`](#fn-spectemplatespecephemeralcontainerslivenessprobehttpgetwithhttpheadersmixin)
+ * [`fn withPath(path)`](#fn-spectemplatespecephemeralcontainerslivenessprobehttpgetwithpath)
+ * [`fn withPort(port)`](#fn-spectemplatespecephemeralcontainerslivenessprobehttpgetwithport)
+ * [`fn withScheme(scheme)`](#fn-spectemplatespecephemeralcontainerslivenessprobehttpgetwithscheme)
+ * [`obj spec.template.spec.ephemeralContainers.livenessProbe.httpGet.httpHeaders`](#obj-spectemplatespecephemeralcontainerslivenessprobehttpgethttpheaders)
+ * [`fn withName(name)`](#fn-spectemplatespecephemeralcontainerslivenessprobehttpgethttpheaderswithname)
+ * [`fn withValue(value)`](#fn-spectemplatespecephemeralcontainerslivenessprobehttpgethttpheaderswithvalue)
+ * [`obj spec.template.spec.ephemeralContainers.livenessProbe.tcpSocket`](#obj-spectemplatespecephemeralcontainerslivenessprobetcpsocket)
+ * [`fn withHost(host)`](#fn-spectemplatespecephemeralcontainerslivenessprobetcpsocketwithhost)
+ * [`fn withPort(port)`](#fn-spectemplatespecephemeralcontainerslivenessprobetcpsocketwithport)
+ * [`obj spec.template.spec.ephemeralContainers.ports`](#obj-spectemplatespecephemeralcontainersports)
+ * [`fn withContainerPort(containerPort)`](#fn-spectemplatespecephemeralcontainersportswithcontainerport)
+ * [`fn withHostIP(hostIP)`](#fn-spectemplatespecephemeralcontainersportswithhostip)
+ * [`fn withHostPort(hostPort)`](#fn-spectemplatespecephemeralcontainersportswithhostport)
+ * [`fn withName(name)`](#fn-spectemplatespecephemeralcontainersportswithname)
+ * [`fn withProtocol(protocol)`](#fn-spectemplatespecephemeralcontainersportswithprotocol)
+ * [`obj spec.template.spec.ephemeralContainers.readinessProbe`](#obj-spectemplatespecephemeralcontainersreadinessprobe)
+ * [`fn withFailureThreshold(failureThreshold)`](#fn-spectemplatespecephemeralcontainersreadinessprobewithfailurethreshold)
+ * [`fn withInitialDelaySeconds(initialDelaySeconds)`](#fn-spectemplatespecephemeralcontainersreadinessprobewithinitialdelayseconds)
+ * [`fn withPeriodSeconds(periodSeconds)`](#fn-spectemplatespecephemeralcontainersreadinessprobewithperiodseconds)
+ * [`fn withSuccessThreshold(successThreshold)`](#fn-spectemplatespecephemeralcontainersreadinessprobewithsuccessthreshold)
+ * [`fn withTerminationGracePeriodSeconds(terminationGracePeriodSeconds)`](#fn-spectemplatespecephemeralcontainersreadinessprobewithterminationgraceperiodseconds)
+ * [`fn withTimeoutSeconds(timeoutSeconds)`](#fn-spectemplatespecephemeralcontainersreadinessprobewithtimeoutseconds)
+ * [`obj spec.template.spec.ephemeralContainers.readinessProbe.exec`](#obj-spectemplatespecephemeralcontainersreadinessprobeexec)
+ * [`fn withCommand(command)`](#fn-spectemplatespecephemeralcontainersreadinessprobeexecwithcommand)
+ * [`fn withCommandMixin(command)`](#fn-spectemplatespecephemeralcontainersreadinessprobeexecwithcommandmixin)
+ * [`obj spec.template.spec.ephemeralContainers.readinessProbe.grpc`](#obj-spectemplatespecephemeralcontainersreadinessprobegrpc)
+ * [`fn withPort(port)`](#fn-spectemplatespecephemeralcontainersreadinessprobegrpcwithport)
+ * [`fn withService(service)`](#fn-spectemplatespecephemeralcontainersreadinessprobegrpcwithservice)
+ * [`obj spec.template.spec.ephemeralContainers.readinessProbe.httpGet`](#obj-spectemplatespecephemeralcontainersreadinessprobehttpget)
+ * [`fn withHost(host)`](#fn-spectemplatespecephemeralcontainersreadinessprobehttpgetwithhost)
+ * [`fn withHttpHeaders(httpHeaders)`](#fn-spectemplatespecephemeralcontainersreadinessprobehttpgetwithhttpheaders)
+ * [`fn withHttpHeadersMixin(httpHeaders)`](#fn-spectemplatespecephemeralcontainersreadinessprobehttpgetwithhttpheadersmixin)
+ * [`fn withPath(path)`](#fn-spectemplatespecephemeralcontainersreadinessprobehttpgetwithpath)
+ * [`fn withPort(port)`](#fn-spectemplatespecephemeralcontainersreadinessprobehttpgetwithport)
+ * [`fn withScheme(scheme)`](#fn-spectemplatespecephemeralcontainersreadinessprobehttpgetwithscheme)
+ * [`obj spec.template.spec.ephemeralContainers.readinessProbe.httpGet.httpHeaders`](#obj-spectemplatespecephemeralcontainersreadinessprobehttpgethttpheaders)
+ * [`fn withName(name)`](#fn-spectemplatespecephemeralcontainersreadinessprobehttpgethttpheaderswithname)
+ * [`fn withValue(value)`](#fn-spectemplatespecephemeralcontainersreadinessprobehttpgethttpheaderswithvalue)
+ * [`obj spec.template.spec.ephemeralContainers.readinessProbe.tcpSocket`](#obj-spectemplatespecephemeralcontainersreadinessprobetcpsocket)
+ * [`fn withHost(host)`](#fn-spectemplatespecephemeralcontainersreadinessprobetcpsocketwithhost)
+ * [`fn withPort(port)`](#fn-spectemplatespecephemeralcontainersreadinessprobetcpsocketwithport)
+ * [`obj spec.template.spec.ephemeralContainers.resizePolicy`](#obj-spectemplatespecephemeralcontainersresizepolicy)
+ * [`fn withResourceName(resourceName)`](#fn-spectemplatespecephemeralcontainersresizepolicywithresourcename)
+ * [`fn withRestartPolicy(restartPolicy)`](#fn-spectemplatespecephemeralcontainersresizepolicywithrestartpolicy)
+ * [`obj spec.template.spec.ephemeralContainers.resources`](#obj-spectemplatespecephemeralcontainersresources)
+ * [`fn withClaims(claims)`](#fn-spectemplatespecephemeralcontainersresourceswithclaims)
+ * [`fn withClaimsMixin(claims)`](#fn-spectemplatespecephemeralcontainersresourceswithclaimsmixin)
+ * [`fn withLimits(limits)`](#fn-spectemplatespecephemeralcontainersresourceswithlimits)
+ * [`fn withLimitsMixin(limits)`](#fn-spectemplatespecephemeralcontainersresourceswithlimitsmixin)
+ * [`fn withRequests(requests)`](#fn-spectemplatespecephemeralcontainersresourceswithrequests)
+ * [`fn withRequestsMixin(requests)`](#fn-spectemplatespecephemeralcontainersresourceswithrequestsmixin)
+ * [`obj spec.template.spec.ephemeralContainers.resources.claims`](#obj-spectemplatespecephemeralcontainersresourcesclaims)
+ * [`fn withName(name)`](#fn-spectemplatespecephemeralcontainersresourcesclaimswithname)
+ * [`obj spec.template.spec.ephemeralContainers.securityContext`](#obj-spectemplatespecephemeralcontainerssecuritycontext)
+ * [`fn withAllowPrivilegeEscalation(allowPrivilegeEscalation)`](#fn-spectemplatespecephemeralcontainerssecuritycontextwithallowprivilegeescalation)
+ * [`fn withPrivileged(privileged)`](#fn-spectemplatespecephemeralcontainerssecuritycontextwithprivileged)
+ * [`fn withProcMount(procMount)`](#fn-spectemplatespecephemeralcontainerssecuritycontextwithprocmount)
+ * [`fn withReadOnlyRootFilesystem(readOnlyRootFilesystem)`](#fn-spectemplatespecephemeralcontainerssecuritycontextwithreadonlyrootfilesystem)
+ * [`fn withRunAsGroup(runAsGroup)`](#fn-spectemplatespecephemeralcontainerssecuritycontextwithrunasgroup)
+ * [`fn withRunAsNonRoot(runAsNonRoot)`](#fn-spectemplatespecephemeralcontainerssecuritycontextwithrunasnonroot)
+ * [`fn withRunAsUser(runAsUser)`](#fn-spectemplatespecephemeralcontainerssecuritycontextwithrunasuser)
+ * [`obj spec.template.spec.ephemeralContainers.securityContext.capabilities`](#obj-spectemplatespecephemeralcontainerssecuritycontextcapabilities)
+ * [`fn withAdd(add)`](#fn-spectemplatespecephemeralcontainerssecuritycontextcapabilitieswithadd)
+ * [`fn withAddMixin(add)`](#fn-spectemplatespecephemeralcontainerssecuritycontextcapabilitieswithaddmixin)
+ * [`fn withDrop(drop)`](#fn-spectemplatespecephemeralcontainerssecuritycontextcapabilitieswithdrop)
+ * [`fn withDropMixin(drop)`](#fn-spectemplatespecephemeralcontainerssecuritycontextcapabilitieswithdropmixin)
+ * [`obj spec.template.spec.ephemeralContainers.securityContext.seLinuxOptions`](#obj-spectemplatespecephemeralcontainerssecuritycontextselinuxoptions)
+ * [`fn withLevel(level)`](#fn-spectemplatespecephemeralcontainerssecuritycontextselinuxoptionswithlevel)
+ * [`fn withRole(role)`](#fn-spectemplatespecephemeralcontainerssecuritycontextselinuxoptionswithrole)
+ * [`fn withType(type)`](#fn-spectemplatespecephemeralcontainerssecuritycontextselinuxoptionswithtype)
+ * [`fn withUser(user)`](#fn-spectemplatespecephemeralcontainerssecuritycontextselinuxoptionswithuser)
+ * [`obj spec.template.spec.ephemeralContainers.securityContext.seccompProfile`](#obj-spectemplatespecephemeralcontainerssecuritycontextseccompprofile)
+ * [`fn withLocalhostProfile(localhostProfile)`](#fn-spectemplatespecephemeralcontainerssecuritycontextseccompprofilewithlocalhostprofile)
+ * [`fn withType(type)`](#fn-spectemplatespecephemeralcontainerssecuritycontextseccompprofilewithtype)
+ * [`obj spec.template.spec.ephemeralContainers.securityContext.windowsOptions`](#obj-spectemplatespecephemeralcontainerssecuritycontextwindowsoptions)
+ * [`fn withGmsaCredentialSpec(gmsaCredentialSpec)`](#fn-spectemplatespecephemeralcontainerssecuritycontextwindowsoptionswithgmsacredentialspec)
+ * [`fn withGmsaCredentialSpecName(gmsaCredentialSpecName)`](#fn-spectemplatespecephemeralcontainerssecuritycontextwindowsoptionswithgmsacredentialspecname)
+ * [`fn withHostProcess(hostProcess)`](#fn-spectemplatespecephemeralcontainerssecuritycontextwindowsoptionswithhostprocess)
+ * [`fn withRunAsUserName(runAsUserName)`](#fn-spectemplatespecephemeralcontainerssecuritycontextwindowsoptionswithrunasusername)
+ * [`obj spec.template.spec.ephemeralContainers.startupProbe`](#obj-spectemplatespecephemeralcontainersstartupprobe)
+ * [`fn withFailureThreshold(failureThreshold)`](#fn-spectemplatespecephemeralcontainersstartupprobewithfailurethreshold)
+ * [`fn withInitialDelaySeconds(initialDelaySeconds)`](#fn-spectemplatespecephemeralcontainersstartupprobewithinitialdelayseconds)
+ * [`fn withPeriodSeconds(periodSeconds)`](#fn-spectemplatespecephemeralcontainersstartupprobewithperiodseconds)
+ * [`fn withSuccessThreshold(successThreshold)`](#fn-spectemplatespecephemeralcontainersstartupprobewithsuccessthreshold)
+ * [`fn withTerminationGracePeriodSeconds(terminationGracePeriodSeconds)`](#fn-spectemplatespecephemeralcontainersstartupprobewithterminationgraceperiodseconds)
+ * [`fn withTimeoutSeconds(timeoutSeconds)`](#fn-spectemplatespecephemeralcontainersstartupprobewithtimeoutseconds)
+ * [`obj spec.template.spec.ephemeralContainers.startupProbe.exec`](#obj-spectemplatespecephemeralcontainersstartupprobeexec)
+ * [`fn withCommand(command)`](#fn-spectemplatespecephemeralcontainersstartupprobeexecwithcommand)
+ * [`fn withCommandMixin(command)`](#fn-spectemplatespecephemeralcontainersstartupprobeexecwithcommandmixin)
+ * [`obj spec.template.spec.ephemeralContainers.startupProbe.grpc`](#obj-spectemplatespecephemeralcontainersstartupprobegrpc)
+ * [`fn withPort(port)`](#fn-spectemplatespecephemeralcontainersstartupprobegrpcwithport)
+ * [`fn withService(service)`](#fn-spectemplatespecephemeralcontainersstartupprobegrpcwithservice)
+ * [`obj spec.template.spec.ephemeralContainers.startupProbe.httpGet`](#obj-spectemplatespecephemeralcontainersstartupprobehttpget)
+ * [`fn withHost(host)`](#fn-spectemplatespecephemeralcontainersstartupprobehttpgetwithhost)
+ * [`fn withHttpHeaders(httpHeaders)`](#fn-spectemplatespecephemeralcontainersstartupprobehttpgetwithhttpheaders)
+ * [`fn withHttpHeadersMixin(httpHeaders)`](#fn-spectemplatespecephemeralcontainersstartupprobehttpgetwithhttpheadersmixin)
+ * [`fn withPath(path)`](#fn-spectemplatespecephemeralcontainersstartupprobehttpgetwithpath)
+ * [`fn withPort(port)`](#fn-spectemplatespecephemeralcontainersstartupprobehttpgetwithport)
+ * [`fn withScheme(scheme)`](#fn-spectemplatespecephemeralcontainersstartupprobehttpgetwithscheme)
+ * [`obj spec.template.spec.ephemeralContainers.startupProbe.httpGet.httpHeaders`](#obj-spectemplatespecephemeralcontainersstartupprobehttpgethttpheaders)
+ * [`fn withName(name)`](#fn-spectemplatespecephemeralcontainersstartupprobehttpgethttpheaderswithname)
+ * [`fn withValue(value)`](#fn-spectemplatespecephemeralcontainersstartupprobehttpgethttpheaderswithvalue)
+ * [`obj spec.template.spec.ephemeralContainers.startupProbe.tcpSocket`](#obj-spectemplatespecephemeralcontainersstartupprobetcpsocket)
+ * [`fn withHost(host)`](#fn-spectemplatespecephemeralcontainersstartupprobetcpsocketwithhost)
+ * [`fn withPort(port)`](#fn-spectemplatespecephemeralcontainersstartupprobetcpsocketwithport)
+ * [`obj spec.template.spec.ephemeralContainers.volumeDevices`](#obj-spectemplatespecephemeralcontainersvolumedevices)
+ * [`fn withDevicePath(devicePath)`](#fn-spectemplatespecephemeralcontainersvolumedeviceswithdevicepath)
+ * [`fn withName(name)`](#fn-spectemplatespecephemeralcontainersvolumedeviceswithname)
+ * [`obj spec.template.spec.ephemeralContainers.volumeMounts`](#obj-spectemplatespecephemeralcontainersvolumemounts)
+ * [`fn withMountPath(mountPath)`](#fn-spectemplatespecephemeralcontainersvolumemountswithmountpath)
+ * [`fn withMountPropagation(mountPropagation)`](#fn-spectemplatespecephemeralcontainersvolumemountswithmountpropagation)
+ * [`fn withName(name)`](#fn-spectemplatespecephemeralcontainersvolumemountswithname)
+ * [`fn withReadOnly(readOnly)`](#fn-spectemplatespecephemeralcontainersvolumemountswithreadonly)
+ * [`fn withSubPath(subPath)`](#fn-spectemplatespecephemeralcontainersvolumemountswithsubpath)
+ * [`fn withSubPathExpr(subPathExpr)`](#fn-spectemplatespecephemeralcontainersvolumemountswithsubpathexpr)
+ * [`obj spec.template.spec.hostAliases`](#obj-spectemplatespechostaliases)
+ * [`fn withHostnames(hostnames)`](#fn-spectemplatespechostaliaseswithhostnames)
+ * [`fn withHostnamesMixin(hostnames)`](#fn-spectemplatespechostaliaseswithhostnamesmixin)
+ * [`fn withIp(ip)`](#fn-spectemplatespechostaliaseswithip)
+ * [`obj spec.template.spec.imagePullSecrets`](#obj-spectemplatespecimagepullsecrets)
+ * [`fn withName(name)`](#fn-spectemplatespecimagepullsecretswithname)
+ * [`obj spec.template.spec.initContainers`](#obj-spectemplatespecinitcontainers)
+ * [`fn withArgs(args)`](#fn-spectemplatespecinitcontainerswithargs)
+ * [`fn withArgsMixin(args)`](#fn-spectemplatespecinitcontainerswithargsmixin)
+ * [`fn withCommand(command)`](#fn-spectemplatespecinitcontainerswithcommand)
+ * [`fn withCommandMixin(command)`](#fn-spectemplatespecinitcontainerswithcommandmixin)
+ * [`fn withEnv(env)`](#fn-spectemplatespecinitcontainerswithenv)
+ * [`fn withEnvFrom(envFrom)`](#fn-spectemplatespecinitcontainerswithenvfrom)
+ * [`fn withEnvFromMixin(envFrom)`](#fn-spectemplatespecinitcontainerswithenvfrommixin)
+ * [`fn withEnvMixin(env)`](#fn-spectemplatespecinitcontainerswithenvmixin)
+ * [`fn withImage(image)`](#fn-spectemplatespecinitcontainerswithimage)
+ * [`fn withImagePullPolicy(imagePullPolicy)`](#fn-spectemplatespecinitcontainerswithimagepullpolicy)
+ * [`fn withName(name)`](#fn-spectemplatespecinitcontainerswithname)
+ * [`fn withPorts(ports)`](#fn-spectemplatespecinitcontainerswithports)
+ * [`fn withPortsMixin(ports)`](#fn-spectemplatespecinitcontainerswithportsmixin)
+ * [`fn withResizePolicy(resizePolicy)`](#fn-spectemplatespecinitcontainerswithresizepolicy)
+ * [`fn withResizePolicyMixin(resizePolicy)`](#fn-spectemplatespecinitcontainerswithresizepolicymixin)
+ * [`fn withRestartPolicy(restartPolicy)`](#fn-spectemplatespecinitcontainerswithrestartpolicy)
+ * [`fn withStdin(stdin)`](#fn-spectemplatespecinitcontainerswithstdin)
+ * [`fn withStdinOnce(stdinOnce)`](#fn-spectemplatespecinitcontainerswithstdinonce)
+ * [`fn withTerminationMessagePath(terminationMessagePath)`](#fn-spectemplatespecinitcontainerswithterminationmessagepath)
+ * [`fn withTerminationMessagePolicy(terminationMessagePolicy)`](#fn-spectemplatespecinitcontainerswithterminationmessagepolicy)
+ * [`fn withTty(tty)`](#fn-spectemplatespecinitcontainerswithtty)
+ * [`fn withVolumeDevices(volumeDevices)`](#fn-spectemplatespecinitcontainerswithvolumedevices)
+ * [`fn withVolumeDevicesMixin(volumeDevices)`](#fn-spectemplatespecinitcontainerswithvolumedevicesmixin)
+ * [`fn withVolumeMounts(volumeMounts)`](#fn-spectemplatespecinitcontainerswithvolumemounts)
+ * [`fn withVolumeMountsMixin(volumeMounts)`](#fn-spectemplatespecinitcontainerswithvolumemountsmixin)
+ * [`fn withWorkingDir(workingDir)`](#fn-spectemplatespecinitcontainerswithworkingdir)
+ * [`obj spec.template.spec.initContainers.env`](#obj-spectemplatespecinitcontainersenv)
+ * [`fn withName(name)`](#fn-spectemplatespecinitcontainersenvwithname)
+ * [`fn withValue(value)`](#fn-spectemplatespecinitcontainersenvwithvalue)
+ * [`obj spec.template.spec.initContainers.env.valueFrom`](#obj-spectemplatespecinitcontainersenvvaluefrom)
+ * [`obj spec.template.spec.initContainers.env.valueFrom.configMapKeyRef`](#obj-spectemplatespecinitcontainersenvvaluefromconfigmapkeyref)
+ * [`fn withKey(key)`](#fn-spectemplatespecinitcontainersenvvaluefromconfigmapkeyrefwithkey)
+ * [`fn withName(name)`](#fn-spectemplatespecinitcontainersenvvaluefromconfigmapkeyrefwithname)
+ * [`fn withOptional(optional)`](#fn-spectemplatespecinitcontainersenvvaluefromconfigmapkeyrefwithoptional)
+ * [`obj spec.template.spec.initContainers.env.valueFrom.fieldRef`](#obj-spectemplatespecinitcontainersenvvaluefromfieldref)
+ * [`fn withApiVersion(apiVersion)`](#fn-spectemplatespecinitcontainersenvvaluefromfieldrefwithapiversion)
+ * [`fn withFieldPath(fieldPath)`](#fn-spectemplatespecinitcontainersenvvaluefromfieldrefwithfieldpath)
+ * [`obj spec.template.spec.initContainers.env.valueFrom.resourceFieldRef`](#obj-spectemplatespecinitcontainersenvvaluefromresourcefieldref)
+ * [`fn withContainerName(containerName)`](#fn-spectemplatespecinitcontainersenvvaluefromresourcefieldrefwithcontainername)
+ * [`fn withDivisor(divisor)`](#fn-spectemplatespecinitcontainersenvvaluefromresourcefieldrefwithdivisor)
+ * [`fn withResource(resource)`](#fn-spectemplatespecinitcontainersenvvaluefromresourcefieldrefwithresource)
+ * [`obj spec.template.spec.initContainers.env.valueFrom.secretKeyRef`](#obj-spectemplatespecinitcontainersenvvaluefromsecretkeyref)
+ * [`fn withKey(key)`](#fn-spectemplatespecinitcontainersenvvaluefromsecretkeyrefwithkey)
+ * [`fn withName(name)`](#fn-spectemplatespecinitcontainersenvvaluefromsecretkeyrefwithname)
+ * [`fn withOptional(optional)`](#fn-spectemplatespecinitcontainersenvvaluefromsecretkeyrefwithoptional)
+ * [`obj spec.template.spec.initContainers.envFrom`](#obj-spectemplatespecinitcontainersenvfrom)
+ * [`fn withPrefix(prefix)`](#fn-spectemplatespecinitcontainersenvfromwithprefix)
+ * [`obj spec.template.spec.initContainers.envFrom.configMapRef`](#obj-spectemplatespecinitcontainersenvfromconfigmapref)
+ * [`fn withName(name)`](#fn-spectemplatespecinitcontainersenvfromconfigmaprefwithname)
+ * [`fn withOptional(optional)`](#fn-spectemplatespecinitcontainersenvfromconfigmaprefwithoptional)
+ * [`obj spec.template.spec.initContainers.envFrom.secretRef`](#obj-spectemplatespecinitcontainersenvfromsecretref)
+ * [`fn withName(name)`](#fn-spectemplatespecinitcontainersenvfromsecretrefwithname)
+ * [`fn withOptional(optional)`](#fn-spectemplatespecinitcontainersenvfromsecretrefwithoptional)
+ * [`obj spec.template.spec.initContainers.lifecycle`](#obj-spectemplatespecinitcontainerslifecycle)
+ * [`obj spec.template.spec.initContainers.lifecycle.postStart`](#obj-spectemplatespecinitcontainerslifecyclepoststart)
+ * [`obj spec.template.spec.initContainers.lifecycle.postStart.exec`](#obj-spectemplatespecinitcontainerslifecyclepoststartexec)
+ * [`fn withCommand(command)`](#fn-spectemplatespecinitcontainerslifecyclepoststartexecwithcommand)
+ * [`fn withCommandMixin(command)`](#fn-spectemplatespecinitcontainerslifecyclepoststartexecwithcommandmixin)
+ * [`obj spec.template.spec.initContainers.lifecycle.postStart.httpGet`](#obj-spectemplatespecinitcontainerslifecyclepoststarthttpget)
+ * [`fn withHost(host)`](#fn-spectemplatespecinitcontainerslifecyclepoststarthttpgetwithhost)
+ * [`fn withHttpHeaders(httpHeaders)`](#fn-spectemplatespecinitcontainerslifecyclepoststarthttpgetwithhttpheaders)
+ * [`fn withHttpHeadersMixin(httpHeaders)`](#fn-spectemplatespecinitcontainerslifecyclepoststarthttpgetwithhttpheadersmixin)
+ * [`fn withPath(path)`](#fn-spectemplatespecinitcontainerslifecyclepoststarthttpgetwithpath)
+ * [`fn withPort(port)`](#fn-spectemplatespecinitcontainerslifecyclepoststarthttpgetwithport)
+ * [`fn withScheme(scheme)`](#fn-spectemplatespecinitcontainerslifecyclepoststarthttpgetwithscheme)
+ * [`obj spec.template.spec.initContainers.lifecycle.postStart.httpGet.httpHeaders`](#obj-spectemplatespecinitcontainerslifecyclepoststarthttpgethttpheaders)
+ * [`fn withName(name)`](#fn-spectemplatespecinitcontainerslifecyclepoststarthttpgethttpheaderswithname)
+ * [`fn withValue(value)`](#fn-spectemplatespecinitcontainerslifecyclepoststarthttpgethttpheaderswithvalue)
+ * [`obj spec.template.spec.initContainers.lifecycle.postStart.tcpSocket`](#obj-spectemplatespecinitcontainerslifecyclepoststarttcpsocket)
+ * [`fn withHost(host)`](#fn-spectemplatespecinitcontainerslifecyclepoststarttcpsocketwithhost)
+ * [`fn withPort(port)`](#fn-spectemplatespecinitcontainerslifecyclepoststarttcpsocketwithport)
+ * [`obj spec.template.spec.initContainers.lifecycle.preStop`](#obj-spectemplatespecinitcontainerslifecycleprestop)
+ * [`obj spec.template.spec.initContainers.lifecycle.preStop.exec`](#obj-spectemplatespecinitcontainerslifecycleprestopexec)
+ * [`fn withCommand(command)`](#fn-spectemplatespecinitcontainerslifecycleprestopexecwithcommand)
+ * [`fn withCommandMixin(command)`](#fn-spectemplatespecinitcontainerslifecycleprestopexecwithcommandmixin)
+ * [`obj spec.template.spec.initContainers.lifecycle.preStop.httpGet`](#obj-spectemplatespecinitcontainerslifecycleprestophttpget)
+ * [`fn withHost(host)`](#fn-spectemplatespecinitcontainerslifecycleprestophttpgetwithhost)
+ * [`fn withHttpHeaders(httpHeaders)`](#fn-spectemplatespecinitcontainerslifecycleprestophttpgetwithhttpheaders)
+ * [`fn withHttpHeadersMixin(httpHeaders)`](#fn-spectemplatespecinitcontainerslifecycleprestophttpgetwithhttpheadersmixin)
+ * [`fn withPath(path)`](#fn-spectemplatespecinitcontainerslifecycleprestophttpgetwithpath)
+ * [`fn withPort(port)`](#fn-spectemplatespecinitcontainerslifecycleprestophttpgetwithport)
+ * [`fn withScheme(scheme)`](#fn-spectemplatespecinitcontainerslifecycleprestophttpgetwithscheme)
+ * [`obj spec.template.spec.initContainers.lifecycle.preStop.httpGet.httpHeaders`](#obj-spectemplatespecinitcontainerslifecycleprestophttpgethttpheaders)
+ * [`fn withName(name)`](#fn-spectemplatespecinitcontainerslifecycleprestophttpgethttpheaderswithname)
+ * [`fn withValue(value)`](#fn-spectemplatespecinitcontainerslifecycleprestophttpgethttpheaderswithvalue)
+ * [`obj spec.template.spec.initContainers.lifecycle.preStop.tcpSocket`](#obj-spectemplatespecinitcontainerslifecycleprestoptcpsocket)
+ * [`fn withHost(host)`](#fn-spectemplatespecinitcontainerslifecycleprestoptcpsocketwithhost)
+ * [`fn withPort(port)`](#fn-spectemplatespecinitcontainerslifecycleprestoptcpsocketwithport)
+ * [`obj spec.template.spec.initContainers.livenessProbe`](#obj-spectemplatespecinitcontainerslivenessprobe)
+ * [`fn withFailureThreshold(failureThreshold)`](#fn-spectemplatespecinitcontainerslivenessprobewithfailurethreshold)
+ * [`fn withInitialDelaySeconds(initialDelaySeconds)`](#fn-spectemplatespecinitcontainerslivenessprobewithinitialdelayseconds)
+ * [`fn withPeriodSeconds(periodSeconds)`](#fn-spectemplatespecinitcontainerslivenessprobewithperiodseconds)
+ * [`fn withSuccessThreshold(successThreshold)`](#fn-spectemplatespecinitcontainerslivenessprobewithsuccessthreshold)
+ * [`fn withTerminationGracePeriodSeconds(terminationGracePeriodSeconds)`](#fn-spectemplatespecinitcontainerslivenessprobewithterminationgraceperiodseconds)
+ * [`fn withTimeoutSeconds(timeoutSeconds)`](#fn-spectemplatespecinitcontainerslivenessprobewithtimeoutseconds)
+ * [`obj spec.template.spec.initContainers.livenessProbe.exec`](#obj-spectemplatespecinitcontainerslivenessprobeexec)
+ * [`fn withCommand(command)`](#fn-spectemplatespecinitcontainerslivenessprobeexecwithcommand)
+ * [`fn withCommandMixin(command)`](#fn-spectemplatespecinitcontainerslivenessprobeexecwithcommandmixin)
+ * [`obj spec.template.spec.initContainers.livenessProbe.grpc`](#obj-spectemplatespecinitcontainerslivenessprobegrpc)
+ * [`fn withPort(port)`](#fn-spectemplatespecinitcontainerslivenessprobegrpcwithport)
+ * [`fn withService(service)`](#fn-spectemplatespecinitcontainerslivenessprobegrpcwithservice)
+ * [`obj spec.template.spec.initContainers.livenessProbe.httpGet`](#obj-spectemplatespecinitcontainerslivenessprobehttpget)
+ * [`fn withHost(host)`](#fn-spectemplatespecinitcontainerslivenessprobehttpgetwithhost)
+ * [`fn withHttpHeaders(httpHeaders)`](#fn-spectemplatespecinitcontainerslivenessprobehttpgetwithhttpheaders)
+ * [`fn withHttpHeadersMixin(httpHeaders)`](#fn-spectemplatespecinitcontainerslivenessprobehttpgetwithhttpheadersmixin)
+ * [`fn withPath(path)`](#fn-spectemplatespecinitcontainerslivenessprobehttpgetwithpath)
+ * [`fn withPort(port)`](#fn-spectemplatespecinitcontainerslivenessprobehttpgetwithport)
+ * [`fn withScheme(scheme)`](#fn-spectemplatespecinitcontainerslivenessprobehttpgetwithscheme)
+ * [`obj spec.template.spec.initContainers.livenessProbe.httpGet.httpHeaders`](#obj-spectemplatespecinitcontainerslivenessprobehttpgethttpheaders)
+ * [`fn withName(name)`](#fn-spectemplatespecinitcontainerslivenessprobehttpgethttpheaderswithname)
+ * [`fn withValue(value)`](#fn-spectemplatespecinitcontainerslivenessprobehttpgethttpheaderswithvalue)
+ * [`obj spec.template.spec.initContainers.livenessProbe.tcpSocket`](#obj-spectemplatespecinitcontainerslivenessprobetcpsocket)
+ * [`fn withHost(host)`](#fn-spectemplatespecinitcontainerslivenessprobetcpsocketwithhost)
+ * [`fn withPort(port)`](#fn-spectemplatespecinitcontainerslivenessprobetcpsocketwithport)
+ * [`obj spec.template.spec.initContainers.ports`](#obj-spectemplatespecinitcontainersports)
+ * [`fn withContainerPort(containerPort)`](#fn-spectemplatespecinitcontainersportswithcontainerport)
+ * [`fn withHostIP(hostIP)`](#fn-spectemplatespecinitcontainersportswithhostip)
+ * [`fn withHostPort(hostPort)`](#fn-spectemplatespecinitcontainersportswithhostport)
+ * [`fn withName(name)`](#fn-spectemplatespecinitcontainersportswithname)
+ * [`fn withProtocol(protocol)`](#fn-spectemplatespecinitcontainersportswithprotocol)
+ * [`obj spec.template.spec.initContainers.readinessProbe`](#obj-spectemplatespecinitcontainersreadinessprobe)
+ * [`fn withFailureThreshold(failureThreshold)`](#fn-spectemplatespecinitcontainersreadinessprobewithfailurethreshold)
+ * [`fn withInitialDelaySeconds(initialDelaySeconds)`](#fn-spectemplatespecinitcontainersreadinessprobewithinitialdelayseconds)
+ * [`fn withPeriodSeconds(periodSeconds)`](#fn-spectemplatespecinitcontainersreadinessprobewithperiodseconds)
+ * [`fn withSuccessThreshold(successThreshold)`](#fn-spectemplatespecinitcontainersreadinessprobewithsuccessthreshold)
+ * [`fn withTerminationGracePeriodSeconds(terminationGracePeriodSeconds)`](#fn-spectemplatespecinitcontainersreadinessprobewithterminationgraceperiodseconds)
+ * [`fn withTimeoutSeconds(timeoutSeconds)`](#fn-spectemplatespecinitcontainersreadinessprobewithtimeoutseconds)
+ * [`obj spec.template.spec.initContainers.readinessProbe.exec`](#obj-spectemplatespecinitcontainersreadinessprobeexec)
+ * [`fn withCommand(command)`](#fn-spectemplatespecinitcontainersreadinessprobeexecwithcommand)
+ * [`fn withCommandMixin(command)`](#fn-spectemplatespecinitcontainersreadinessprobeexecwithcommandmixin)
+ * [`obj spec.template.spec.initContainers.readinessProbe.grpc`](#obj-spectemplatespecinitcontainersreadinessprobegrpc)
+ * [`fn withPort(port)`](#fn-spectemplatespecinitcontainersreadinessprobegrpcwithport)
+ * [`fn withService(service)`](#fn-spectemplatespecinitcontainersreadinessprobegrpcwithservice)
+ * [`obj spec.template.spec.initContainers.readinessProbe.httpGet`](#obj-spectemplatespecinitcontainersreadinessprobehttpget)
+ * [`fn withHost(host)`](#fn-spectemplatespecinitcontainersreadinessprobehttpgetwithhost)
+ * [`fn withHttpHeaders(httpHeaders)`](#fn-spectemplatespecinitcontainersreadinessprobehttpgetwithhttpheaders)
+ * [`fn withHttpHeadersMixin(httpHeaders)`](#fn-spectemplatespecinitcontainersreadinessprobehttpgetwithhttpheadersmixin)
+ * [`fn withPath(path)`](#fn-spectemplatespecinitcontainersreadinessprobehttpgetwithpath)
+ * [`fn withPort(port)`](#fn-spectemplatespecinitcontainersreadinessprobehttpgetwithport)
+ * [`fn withScheme(scheme)`](#fn-spectemplatespecinitcontainersreadinessprobehttpgetwithscheme)
+ * [`obj spec.template.spec.initContainers.readinessProbe.httpGet.httpHeaders`](#obj-spectemplatespecinitcontainersreadinessprobehttpgethttpheaders)
+ * [`fn withName(name)`](#fn-spectemplatespecinitcontainersreadinessprobehttpgethttpheaderswithname)
+ * [`fn withValue(value)`](#fn-spectemplatespecinitcontainersreadinessprobehttpgethttpheaderswithvalue)
+ * [`obj spec.template.spec.initContainers.readinessProbe.tcpSocket`](#obj-spectemplatespecinitcontainersreadinessprobetcpsocket)
+ * [`fn withHost(host)`](#fn-spectemplatespecinitcontainersreadinessprobetcpsocketwithhost)
+ * [`fn withPort(port)`](#fn-spectemplatespecinitcontainersreadinessprobetcpsocketwithport)
+ * [`obj spec.template.spec.initContainers.resizePolicy`](#obj-spectemplatespecinitcontainersresizepolicy)
+ * [`fn withResourceName(resourceName)`](#fn-spectemplatespecinitcontainersresizepolicywithresourcename)
+ * [`fn withRestartPolicy(restartPolicy)`](#fn-spectemplatespecinitcontainersresizepolicywithrestartpolicy)
+ * [`obj spec.template.spec.initContainers.resources`](#obj-spectemplatespecinitcontainersresources)
+ * [`fn withClaims(claims)`](#fn-spectemplatespecinitcontainersresourceswithclaims)
+ * [`fn withClaimsMixin(claims)`](#fn-spectemplatespecinitcontainersresourceswithclaimsmixin)
+ * [`fn withLimits(limits)`](#fn-spectemplatespecinitcontainersresourceswithlimits)
+ * [`fn withLimitsMixin(limits)`](#fn-spectemplatespecinitcontainersresourceswithlimitsmixin)
+ * [`fn withRequests(requests)`](#fn-spectemplatespecinitcontainersresourceswithrequests)
+ * [`fn withRequestsMixin(requests)`](#fn-spectemplatespecinitcontainersresourceswithrequestsmixin)
+ * [`obj spec.template.spec.initContainers.resources.claims`](#obj-spectemplatespecinitcontainersresourcesclaims)
+ * [`fn withName(name)`](#fn-spectemplatespecinitcontainersresourcesclaimswithname)
+ * [`obj spec.template.spec.initContainers.securityContext`](#obj-spectemplatespecinitcontainerssecuritycontext)
+ * [`fn withAllowPrivilegeEscalation(allowPrivilegeEscalation)`](#fn-spectemplatespecinitcontainerssecuritycontextwithallowprivilegeescalation)
+ * [`fn withPrivileged(privileged)`](#fn-spectemplatespecinitcontainerssecuritycontextwithprivileged)
+ * [`fn withProcMount(procMount)`](#fn-spectemplatespecinitcontainerssecuritycontextwithprocmount)
+ * [`fn withReadOnlyRootFilesystem(readOnlyRootFilesystem)`](#fn-spectemplatespecinitcontainerssecuritycontextwithreadonlyrootfilesystem)
+ * [`fn withRunAsGroup(runAsGroup)`](#fn-spectemplatespecinitcontainerssecuritycontextwithrunasgroup)
+ * [`fn withRunAsNonRoot(runAsNonRoot)`](#fn-spectemplatespecinitcontainerssecuritycontextwithrunasnonroot)
+ * [`fn withRunAsUser(runAsUser)`](#fn-spectemplatespecinitcontainerssecuritycontextwithrunasuser)
+ * [`obj spec.template.spec.initContainers.securityContext.capabilities`](#obj-spectemplatespecinitcontainerssecuritycontextcapabilities)
+ * [`fn withAdd(add)`](#fn-spectemplatespecinitcontainerssecuritycontextcapabilitieswithadd)
+ * [`fn withAddMixin(add)`](#fn-spectemplatespecinitcontainerssecuritycontextcapabilitieswithaddmixin)
+ * [`fn withDrop(drop)`](#fn-spectemplatespecinitcontainerssecuritycontextcapabilitieswithdrop)
+ * [`fn withDropMixin(drop)`](#fn-spectemplatespecinitcontainerssecuritycontextcapabilitieswithdropmixin)
+ * [`obj spec.template.spec.initContainers.securityContext.seLinuxOptions`](#obj-spectemplatespecinitcontainerssecuritycontextselinuxoptions)
+ * [`fn withLevel(level)`](#fn-spectemplatespecinitcontainerssecuritycontextselinuxoptionswithlevel)
+ * [`fn withRole(role)`](#fn-spectemplatespecinitcontainerssecuritycontextselinuxoptionswithrole)
+ * [`fn withType(type)`](#fn-spectemplatespecinitcontainerssecuritycontextselinuxoptionswithtype)
+ * [`fn withUser(user)`](#fn-spectemplatespecinitcontainerssecuritycontextselinuxoptionswithuser)
+ * [`obj spec.template.spec.initContainers.securityContext.seccompProfile`](#obj-spectemplatespecinitcontainerssecuritycontextseccompprofile)
+ * [`fn withLocalhostProfile(localhostProfile)`](#fn-spectemplatespecinitcontainerssecuritycontextseccompprofilewithlocalhostprofile)
+ * [`fn withType(type)`](#fn-spectemplatespecinitcontainerssecuritycontextseccompprofilewithtype)
+ * [`obj spec.template.spec.initContainers.securityContext.windowsOptions`](#obj-spectemplatespecinitcontainerssecuritycontextwindowsoptions)
+ * [`fn withGmsaCredentialSpec(gmsaCredentialSpec)`](#fn-spectemplatespecinitcontainerssecuritycontextwindowsoptionswithgmsacredentialspec)
+ * [`fn withGmsaCredentialSpecName(gmsaCredentialSpecName)`](#fn-spectemplatespecinitcontainerssecuritycontextwindowsoptionswithgmsacredentialspecname)
+ * [`fn withHostProcess(hostProcess)`](#fn-spectemplatespecinitcontainerssecuritycontextwindowsoptionswithhostprocess)
+ * [`fn withRunAsUserName(runAsUserName)`](#fn-spectemplatespecinitcontainerssecuritycontextwindowsoptionswithrunasusername)
+ * [`obj spec.template.spec.initContainers.startupProbe`](#obj-spectemplatespecinitcontainersstartupprobe)
+ * [`fn withFailureThreshold(failureThreshold)`](#fn-spectemplatespecinitcontainersstartupprobewithfailurethreshold)
+ * [`fn withInitialDelaySeconds(initialDelaySeconds)`](#fn-spectemplatespecinitcontainersstartupprobewithinitialdelayseconds)
+ * [`fn withPeriodSeconds(periodSeconds)`](#fn-spectemplatespecinitcontainersstartupprobewithperiodseconds)
+ * [`fn withSuccessThreshold(successThreshold)`](#fn-spectemplatespecinitcontainersstartupprobewithsuccessthreshold)
+ * [`fn withTerminationGracePeriodSeconds(terminationGracePeriodSeconds)`](#fn-spectemplatespecinitcontainersstartupprobewithterminationgraceperiodseconds)
+ * [`fn withTimeoutSeconds(timeoutSeconds)`](#fn-spectemplatespecinitcontainersstartupprobewithtimeoutseconds)
+ * [`obj spec.template.spec.initContainers.startupProbe.exec`](#obj-spectemplatespecinitcontainersstartupprobeexec)
+ * [`fn withCommand(command)`](#fn-spectemplatespecinitcontainersstartupprobeexecwithcommand)
+ * [`fn withCommandMixin(command)`](#fn-spectemplatespecinitcontainersstartupprobeexecwithcommandmixin)
+ * [`obj spec.template.spec.initContainers.startupProbe.grpc`](#obj-spectemplatespecinitcontainersstartupprobegrpc)
+ * [`fn withPort(port)`](#fn-spectemplatespecinitcontainersstartupprobegrpcwithport)
+ * [`fn withService(service)`](#fn-spectemplatespecinitcontainersstartupprobegrpcwithservice)
+ * [`obj spec.template.spec.initContainers.startupProbe.httpGet`](#obj-spectemplatespecinitcontainersstartupprobehttpget)
+ * [`fn withHost(host)`](#fn-spectemplatespecinitcontainersstartupprobehttpgetwithhost)
+ * [`fn withHttpHeaders(httpHeaders)`](#fn-spectemplatespecinitcontainersstartupprobehttpgetwithhttpheaders)
+ * [`fn withHttpHeadersMixin(httpHeaders)`](#fn-spectemplatespecinitcontainersstartupprobehttpgetwithhttpheadersmixin)
+ * [`fn withPath(path)`](#fn-spectemplatespecinitcontainersstartupprobehttpgetwithpath)
+ * [`fn withPort(port)`](#fn-spectemplatespecinitcontainersstartupprobehttpgetwithport)
+ * [`fn withScheme(scheme)`](#fn-spectemplatespecinitcontainersstartupprobehttpgetwithscheme)
+ * [`obj spec.template.spec.initContainers.startupProbe.httpGet.httpHeaders`](#obj-spectemplatespecinitcontainersstartupprobehttpgethttpheaders)
+ * [`fn withName(name)`](#fn-spectemplatespecinitcontainersstartupprobehttpgethttpheaderswithname)
+ * [`fn withValue(value)`](#fn-spectemplatespecinitcontainersstartupprobehttpgethttpheaderswithvalue)
+ * [`obj spec.template.spec.initContainers.startupProbe.tcpSocket`](#obj-spectemplatespecinitcontainersstartupprobetcpsocket)
+ * [`fn withHost(host)`](#fn-spectemplatespecinitcontainersstartupprobetcpsocketwithhost)
+ * [`fn withPort(port)`](#fn-spectemplatespecinitcontainersstartupprobetcpsocketwithport)
+ * [`obj spec.template.spec.initContainers.volumeDevices`](#obj-spectemplatespecinitcontainersvolumedevices)
+ * [`fn withDevicePath(devicePath)`](#fn-spectemplatespecinitcontainersvolumedeviceswithdevicepath)
+ * [`fn withName(name)`](#fn-spectemplatespecinitcontainersvolumedeviceswithname)
+ * [`obj spec.template.spec.initContainers.volumeMounts`](#obj-spectemplatespecinitcontainersvolumemounts)
+ * [`fn withMountPath(mountPath)`](#fn-spectemplatespecinitcontainersvolumemountswithmountpath)
+ * [`fn withMountPropagation(mountPropagation)`](#fn-spectemplatespecinitcontainersvolumemountswithmountpropagation)
+ * [`fn withName(name)`](#fn-spectemplatespecinitcontainersvolumemountswithname)
+ * [`fn withReadOnly(readOnly)`](#fn-spectemplatespecinitcontainersvolumemountswithreadonly)
+ * [`fn withSubPath(subPath)`](#fn-spectemplatespecinitcontainersvolumemountswithsubpath)
+ * [`fn withSubPathExpr(subPathExpr)`](#fn-spectemplatespecinitcontainersvolumemountswithsubpathexpr)
+ * [`obj spec.template.spec.os`](#obj-spectemplatespecos)
+ * [`fn withName(name)`](#fn-spectemplatespecoswithname)
+ * [`obj spec.template.spec.readinessGates`](#obj-spectemplatespecreadinessgates)
+ * [`fn withConditionType(conditionType)`](#fn-spectemplatespecreadinessgateswithconditiontype)
+ * [`obj spec.template.spec.resourceClaims`](#obj-spectemplatespecresourceclaims)
+ * [`fn withName(name)`](#fn-spectemplatespecresourceclaimswithname)
+ * [`obj spec.template.spec.resourceClaims.source`](#obj-spectemplatespecresourceclaimssource)
+ * [`fn withResourceClaimName(resourceClaimName)`](#fn-spectemplatespecresourceclaimssourcewithresourceclaimname)
+ * [`fn withResourceClaimTemplateName(resourceClaimTemplateName)`](#fn-spectemplatespecresourceclaimssourcewithresourceclaimtemplatename)
+ * [`obj spec.template.spec.schedulingGates`](#obj-spectemplatespecschedulinggates)
+ * [`fn withName(name)`](#fn-spectemplatespecschedulinggateswithname)
+ * [`obj spec.template.spec.securityContext`](#obj-spectemplatespecsecuritycontext)
+ * [`fn withFsGroup(fsGroup)`](#fn-spectemplatespecsecuritycontextwithfsgroup)
+ * [`fn withFsGroupChangePolicy(fsGroupChangePolicy)`](#fn-spectemplatespecsecuritycontextwithfsgroupchangepolicy)
+ * [`fn withRunAsGroup(runAsGroup)`](#fn-spectemplatespecsecuritycontextwithrunasgroup)
+ * [`fn withRunAsNonRoot(runAsNonRoot)`](#fn-spectemplatespecsecuritycontextwithrunasnonroot)
+ * [`fn withRunAsUser(runAsUser)`](#fn-spectemplatespecsecuritycontextwithrunasuser)
+ * [`fn withSupplementalGroups(supplementalGroups)`](#fn-spectemplatespecsecuritycontextwithsupplementalgroups)
+ * [`fn withSupplementalGroupsMixin(supplementalGroups)`](#fn-spectemplatespecsecuritycontextwithsupplementalgroupsmixin)
+ * [`fn withSysctls(sysctls)`](#fn-spectemplatespecsecuritycontextwithsysctls)
+ * [`fn withSysctlsMixin(sysctls)`](#fn-spectemplatespecsecuritycontextwithsysctlsmixin)
+ * [`obj spec.template.spec.securityContext.seLinuxOptions`](#obj-spectemplatespecsecuritycontextselinuxoptions)
+ * [`fn withLevel(level)`](#fn-spectemplatespecsecuritycontextselinuxoptionswithlevel)
+ * [`fn withRole(role)`](#fn-spectemplatespecsecuritycontextselinuxoptionswithrole)
+ * [`fn withType(type)`](#fn-spectemplatespecsecuritycontextselinuxoptionswithtype)
+ * [`fn withUser(user)`](#fn-spectemplatespecsecuritycontextselinuxoptionswithuser)
+ * [`obj spec.template.spec.securityContext.seccompProfile`](#obj-spectemplatespecsecuritycontextseccompprofile)
+ * [`fn withLocalhostProfile(localhostProfile)`](#fn-spectemplatespecsecuritycontextseccompprofilewithlocalhostprofile)
+ * [`fn withType(type)`](#fn-spectemplatespecsecuritycontextseccompprofilewithtype)
+ * [`obj spec.template.spec.securityContext.sysctls`](#obj-spectemplatespecsecuritycontextsysctls)
+ * [`fn withName(name)`](#fn-spectemplatespecsecuritycontextsysctlswithname)
+ * [`fn withValue(value)`](#fn-spectemplatespecsecuritycontextsysctlswithvalue)
+ * [`obj spec.template.spec.securityContext.windowsOptions`](#obj-spectemplatespecsecuritycontextwindowsoptions)
+ * [`fn withGmsaCredentialSpec(gmsaCredentialSpec)`](#fn-spectemplatespecsecuritycontextwindowsoptionswithgmsacredentialspec)
+ * [`fn withGmsaCredentialSpecName(gmsaCredentialSpecName)`](#fn-spectemplatespecsecuritycontextwindowsoptionswithgmsacredentialspecname)
+ * [`fn withHostProcess(hostProcess)`](#fn-spectemplatespecsecuritycontextwindowsoptionswithhostprocess)
+ * [`fn withRunAsUserName(runAsUserName)`](#fn-spectemplatespecsecuritycontextwindowsoptionswithrunasusername)
+ * [`obj spec.template.spec.tolerations`](#obj-spectemplatespectolerations)
+ * [`fn withEffect(effect)`](#fn-spectemplatespectolerationswitheffect)
+ * [`fn withKey(key)`](#fn-spectemplatespectolerationswithkey)
+ * [`fn withOperator(operator)`](#fn-spectemplatespectolerationswithoperator)
+ * [`fn withTolerationSeconds(tolerationSeconds)`](#fn-spectemplatespectolerationswithtolerationseconds)
+ * [`fn withValue(value)`](#fn-spectemplatespectolerationswithvalue)
+ * [`obj spec.template.spec.topologySpreadConstraints`](#obj-spectemplatespectopologyspreadconstraints)
+ * [`fn withMatchLabelKeys(matchLabelKeys)`](#fn-spectemplatespectopologyspreadconstraintswithmatchlabelkeys)
+ * [`fn withMatchLabelKeysMixin(matchLabelKeys)`](#fn-spectemplatespectopologyspreadconstraintswithmatchlabelkeysmixin)
+ * [`fn withMaxSkew(maxSkew)`](#fn-spectemplatespectopologyspreadconstraintswithmaxskew)
+ * [`fn withMinDomains(minDomains)`](#fn-spectemplatespectopologyspreadconstraintswithmindomains)
+ * [`fn withNodeAffinityPolicy(nodeAffinityPolicy)`](#fn-spectemplatespectopologyspreadconstraintswithnodeaffinitypolicy)
+ * [`fn withNodeTaintsPolicy(nodeTaintsPolicy)`](#fn-spectemplatespectopologyspreadconstraintswithnodetaintspolicy)
+ * [`fn withTopologyKey(topologyKey)`](#fn-spectemplatespectopologyspreadconstraintswithtopologykey)
+ * [`fn withWhenUnsatisfiable(whenUnsatisfiable)`](#fn-spectemplatespectopologyspreadconstraintswithwhenunsatisfiable)
+ * [`obj spec.template.spec.topologySpreadConstraints.labelSelector`](#obj-spectemplatespectopologyspreadconstraintslabelselector)
+ * [`fn withMatchExpressions(matchExpressions)`](#fn-spectemplatespectopologyspreadconstraintslabelselectorwithmatchexpressions)
+ * [`fn withMatchExpressionsMixin(matchExpressions)`](#fn-spectemplatespectopologyspreadconstraintslabelselectorwithmatchexpressionsmixin)
+ * [`fn withMatchLabels(matchLabels)`](#fn-spectemplatespectopologyspreadconstraintslabelselectorwithmatchlabels)
+ * [`fn withMatchLabelsMixin(matchLabels)`](#fn-spectemplatespectopologyspreadconstraintslabelselectorwithmatchlabelsmixin)
+ * [`obj spec.template.spec.topologySpreadConstraints.labelSelector.matchExpressions`](#obj-spectemplatespectopologyspreadconstraintslabelselectormatchexpressions)
+ * [`fn withKey(key)`](#fn-spectemplatespectopologyspreadconstraintslabelselectormatchexpressionswithkey)
+ * [`fn withOperator(operator)`](#fn-spectemplatespectopologyspreadconstraintslabelselectormatchexpressionswithoperator)
+ * [`fn withValues(values)`](#fn-spectemplatespectopologyspreadconstraintslabelselectormatchexpressionswithvalues)
+ * [`fn withValuesMixin(values)`](#fn-spectemplatespectopologyspreadconstraintslabelselectormatchexpressionswithvaluesmixin)
+ * [`obj spec.template.spec.volumes`](#obj-spectemplatespecvolumes)
+ * [`fn withName(name)`](#fn-spectemplatespecvolumeswithname)
+ * [`obj spec.template.spec.volumes.awsElasticBlockStore`](#obj-spectemplatespecvolumesawselasticblockstore)
+ * [`fn withFsType(fsType)`](#fn-spectemplatespecvolumesawselasticblockstorewithfstype)
+ * [`fn withPartition(partition)`](#fn-spectemplatespecvolumesawselasticblockstorewithpartition)
+ * [`fn withReadOnly(readOnly)`](#fn-spectemplatespecvolumesawselasticblockstorewithreadonly)
+ * [`fn withVolumeID(volumeID)`](#fn-spectemplatespecvolumesawselasticblockstorewithvolumeid)
+ * [`obj spec.template.spec.volumes.azureDisk`](#obj-spectemplatespecvolumesazuredisk)
+ * [`fn withCachingMode(cachingMode)`](#fn-spectemplatespecvolumesazurediskwithcachingmode)
+ * [`fn withDiskName(diskName)`](#fn-spectemplatespecvolumesazurediskwithdiskname)
+ * [`fn withDiskURI(diskURI)`](#fn-spectemplatespecvolumesazurediskwithdiskuri)
+ * [`fn withFsType(fsType)`](#fn-spectemplatespecvolumesazurediskwithfstype)
+ * [`fn withKind(kind)`](#fn-spectemplatespecvolumesazurediskwithkind)
+ * [`fn withReadOnly(readOnly)`](#fn-spectemplatespecvolumesazurediskwithreadonly)
+ * [`obj spec.template.spec.volumes.azureFile`](#obj-spectemplatespecvolumesazurefile)
+ * [`fn withReadOnly(readOnly)`](#fn-spectemplatespecvolumesazurefilewithreadonly)
+ * [`fn withSecretName(secretName)`](#fn-spectemplatespecvolumesazurefilewithsecretname)
+ * [`fn withShareName(shareName)`](#fn-spectemplatespecvolumesazurefilewithsharename)
+ * [`obj spec.template.spec.volumes.cephfs`](#obj-spectemplatespecvolumescephfs)
+ * [`fn withMonitors(monitors)`](#fn-spectemplatespecvolumescephfswithmonitors)
+ * [`fn withMonitorsMixin(monitors)`](#fn-spectemplatespecvolumescephfswithmonitorsmixin)
+ * [`fn withPath(path)`](#fn-spectemplatespecvolumescephfswithpath)
+ * [`fn withReadOnly(readOnly)`](#fn-spectemplatespecvolumescephfswithreadonly)
+ * [`fn withSecretFile(secretFile)`](#fn-spectemplatespecvolumescephfswithsecretfile)
+ * [`fn withUser(user)`](#fn-spectemplatespecvolumescephfswithuser)
+ * [`obj spec.template.spec.volumes.cephfs.secretRef`](#obj-spectemplatespecvolumescephfssecretref)
+ * [`fn withName(name)`](#fn-spectemplatespecvolumescephfssecretrefwithname)
+ * [`obj spec.template.spec.volumes.cinder`](#obj-spectemplatespecvolumescinder)
+ * [`fn withFsType(fsType)`](#fn-spectemplatespecvolumescinderwithfstype)
+ * [`fn withReadOnly(readOnly)`](#fn-spectemplatespecvolumescinderwithreadonly)
+ * [`fn withVolumeID(volumeID)`](#fn-spectemplatespecvolumescinderwithvolumeid)
+ * [`obj spec.template.spec.volumes.cinder.secretRef`](#obj-spectemplatespecvolumescindersecretref)
+ * [`fn withName(name)`](#fn-spectemplatespecvolumescindersecretrefwithname)
+ * [`obj spec.template.spec.volumes.configMap`](#obj-spectemplatespecvolumesconfigmap)
+ * [`fn withDefaultMode(defaultMode)`](#fn-spectemplatespecvolumesconfigmapwithdefaultmode)
+ * [`fn withItems(items)`](#fn-spectemplatespecvolumesconfigmapwithitems)
+ * [`fn withItemsMixin(items)`](#fn-spectemplatespecvolumesconfigmapwithitemsmixin)
+ * [`fn withName(name)`](#fn-spectemplatespecvolumesconfigmapwithname)
+ * [`fn withOptional(optional)`](#fn-spectemplatespecvolumesconfigmapwithoptional)
+ * [`obj spec.template.spec.volumes.configMap.items`](#obj-spectemplatespecvolumesconfigmapitems)
+ * [`fn withKey(key)`](#fn-spectemplatespecvolumesconfigmapitemswithkey)
+ * [`fn withMode(mode)`](#fn-spectemplatespecvolumesconfigmapitemswithmode)
+ * [`fn withPath(path)`](#fn-spectemplatespecvolumesconfigmapitemswithpath)
+ * [`obj spec.template.spec.volumes.csi`](#obj-spectemplatespecvolumescsi)
+ * [`fn withDriver(driver)`](#fn-spectemplatespecvolumescsiwithdriver)
+ * [`fn withFsType(fsType)`](#fn-spectemplatespecvolumescsiwithfstype)
+ * [`fn withReadOnly(readOnly)`](#fn-spectemplatespecvolumescsiwithreadonly)
+ * [`fn withVolumeAttributes(volumeAttributes)`](#fn-spectemplatespecvolumescsiwithvolumeattributes)
+ * [`fn withVolumeAttributesMixin(volumeAttributes)`](#fn-spectemplatespecvolumescsiwithvolumeattributesmixin)
+ * [`obj spec.template.spec.volumes.csi.nodePublishSecretRef`](#obj-spectemplatespecvolumescsinodepublishsecretref)
+ * [`fn withName(name)`](#fn-spectemplatespecvolumescsinodepublishsecretrefwithname)
+ * [`obj spec.template.spec.volumes.downwardAPI`](#obj-spectemplatespecvolumesdownwardapi)
+ * [`fn withDefaultMode(defaultMode)`](#fn-spectemplatespecvolumesdownwardapiwithdefaultmode)
+ * [`fn withItems(items)`](#fn-spectemplatespecvolumesdownwardapiwithitems)
+ * [`fn withItemsMixin(items)`](#fn-spectemplatespecvolumesdownwardapiwithitemsmixin)
+ * [`obj spec.template.spec.volumes.downwardAPI.items`](#obj-spectemplatespecvolumesdownwardapiitems)
+ * [`fn withMode(mode)`](#fn-spectemplatespecvolumesdownwardapiitemswithmode)
+ * [`fn withPath(path)`](#fn-spectemplatespecvolumesdownwardapiitemswithpath)
+ * [`obj spec.template.spec.volumes.downwardAPI.items.fieldRef`](#obj-spectemplatespecvolumesdownwardapiitemsfieldref)
+ * [`fn withApiVersion(apiVersion)`](#fn-spectemplatespecvolumesdownwardapiitemsfieldrefwithapiversion)
+ * [`fn withFieldPath(fieldPath)`](#fn-spectemplatespecvolumesdownwardapiitemsfieldrefwithfieldpath)
+ * [`obj spec.template.spec.volumes.downwardAPI.items.resourceFieldRef`](#obj-spectemplatespecvolumesdownwardapiitemsresourcefieldref)
+ * [`fn withContainerName(containerName)`](#fn-spectemplatespecvolumesdownwardapiitemsresourcefieldrefwithcontainername)
+ * [`fn withDivisor(divisor)`](#fn-spectemplatespecvolumesdownwardapiitemsresourcefieldrefwithdivisor)
+ * [`fn withResource(resource)`](#fn-spectemplatespecvolumesdownwardapiitemsresourcefieldrefwithresource)
+ * [`obj spec.template.spec.volumes.emptyDir`](#obj-spectemplatespecvolumesemptydir)
+ * [`fn withMedium(medium)`](#fn-spectemplatespecvolumesemptydirwithmedium)
+ * [`fn withSizeLimit(sizeLimit)`](#fn-spectemplatespecvolumesemptydirwithsizelimit)
+ * [`obj spec.template.spec.volumes.ephemeral`](#obj-spectemplatespecvolumesephemeral)
+ * [`obj spec.template.spec.volumes.ephemeral.volumeClaimTemplate`](#obj-spectemplatespecvolumesephemeralvolumeclaimtemplate)
+ * [`fn withMetadata(metadata)`](#fn-spectemplatespecvolumesephemeralvolumeclaimtemplatewithmetadata)
+ * [`fn withMetadataMixin(metadata)`](#fn-spectemplatespecvolumesephemeralvolumeclaimtemplatewithmetadatamixin)
+ * [`obj spec.template.spec.volumes.ephemeral.volumeClaimTemplate.spec`](#obj-spectemplatespecvolumesephemeralvolumeclaimtemplatespec)
+ * [`fn withAccessModes(accessModes)`](#fn-spectemplatespecvolumesephemeralvolumeclaimtemplatespecwithaccessmodes)
+ * [`fn withAccessModesMixin(accessModes)`](#fn-spectemplatespecvolumesephemeralvolumeclaimtemplatespecwithaccessmodesmixin)
+ * [`fn withStorageClassName(storageClassName)`](#fn-spectemplatespecvolumesephemeralvolumeclaimtemplatespecwithstorageclassname)
+ * [`fn withVolumeMode(volumeMode)`](#fn-spectemplatespecvolumesephemeralvolumeclaimtemplatespecwithvolumemode)
+ * [`fn withVolumeName(volumeName)`](#fn-spectemplatespecvolumesephemeralvolumeclaimtemplatespecwithvolumename)
+ * [`obj spec.template.spec.volumes.ephemeral.volumeClaimTemplate.spec.dataSource`](#obj-spectemplatespecvolumesephemeralvolumeclaimtemplatespecdatasource)
+ * [`fn withApiGroup(apiGroup)`](#fn-spectemplatespecvolumesephemeralvolumeclaimtemplatespecdatasourcewithapigroup)
+ * [`fn withKind(kind)`](#fn-spectemplatespecvolumesephemeralvolumeclaimtemplatespecdatasourcewithkind)
+ * [`fn withName(name)`](#fn-spectemplatespecvolumesephemeralvolumeclaimtemplatespecdatasourcewithname)
+ * [`obj spec.template.spec.volumes.ephemeral.volumeClaimTemplate.spec.dataSourceRef`](#obj-spectemplatespecvolumesephemeralvolumeclaimtemplatespecdatasourceref)
+ * [`fn withApiGroup(apiGroup)`](#fn-spectemplatespecvolumesephemeralvolumeclaimtemplatespecdatasourcerefwithapigroup)
+ * [`fn withKind(kind)`](#fn-spectemplatespecvolumesephemeralvolumeclaimtemplatespecdatasourcerefwithkind)
+ * [`fn withName(name)`](#fn-spectemplatespecvolumesephemeralvolumeclaimtemplatespecdatasourcerefwithname)
+ * [`fn withNamespace(namespace)`](#fn-spectemplatespecvolumesephemeralvolumeclaimtemplatespecdatasourcerefwithnamespace)
+ * [`obj spec.template.spec.volumes.ephemeral.volumeClaimTemplate.spec.resources`](#obj-spectemplatespecvolumesephemeralvolumeclaimtemplatespecresources)
+ * [`fn withClaims(claims)`](#fn-spectemplatespecvolumesephemeralvolumeclaimtemplatespecresourceswithclaims)
+ * [`fn withClaimsMixin(claims)`](#fn-spectemplatespecvolumesephemeralvolumeclaimtemplatespecresourceswithclaimsmixin)
+ * [`fn withLimits(limits)`](#fn-spectemplatespecvolumesephemeralvolumeclaimtemplatespecresourceswithlimits)
+ * [`fn withLimitsMixin(limits)`](#fn-spectemplatespecvolumesephemeralvolumeclaimtemplatespecresourceswithlimitsmixin)
+ * [`fn withRequests(requests)`](#fn-spectemplatespecvolumesephemeralvolumeclaimtemplatespecresourceswithrequests)
+ * [`fn withRequestsMixin(requests)`](#fn-spectemplatespecvolumesephemeralvolumeclaimtemplatespecresourceswithrequestsmixin)
+ * [`obj spec.template.spec.volumes.ephemeral.volumeClaimTemplate.spec.resources.claims`](#obj-spectemplatespecvolumesephemeralvolumeclaimtemplatespecresourcesclaims)
+ * [`fn withName(name)`](#fn-spectemplatespecvolumesephemeralvolumeclaimtemplatespecresourcesclaimswithname)
+ * [`obj spec.template.spec.volumes.ephemeral.volumeClaimTemplate.spec.selector`](#obj-spectemplatespecvolumesephemeralvolumeclaimtemplatespecselector)
+ * [`fn withMatchExpressions(matchExpressions)`](#fn-spectemplatespecvolumesephemeralvolumeclaimtemplatespecselectorwithmatchexpressions)
+ * [`fn withMatchExpressionsMixin(matchExpressions)`](#fn-spectemplatespecvolumesephemeralvolumeclaimtemplatespecselectorwithmatchexpressionsmixin)
+ * [`fn withMatchLabels(matchLabels)`](#fn-spectemplatespecvolumesephemeralvolumeclaimtemplatespecselectorwithmatchlabels)
+ * [`fn withMatchLabelsMixin(matchLabels)`](#fn-spectemplatespecvolumesephemeralvolumeclaimtemplatespecselectorwithmatchlabelsmixin)
+ * [`obj spec.template.spec.volumes.ephemeral.volumeClaimTemplate.spec.selector.matchExpressions`](#obj-spectemplatespecvolumesephemeralvolumeclaimtemplatespecselectormatchexpressions)
+ * [`fn withKey(key)`](#fn-spectemplatespecvolumesephemeralvolumeclaimtemplatespecselectormatchexpressionswithkey)
+ * [`fn withOperator(operator)`](#fn-spectemplatespecvolumesephemeralvolumeclaimtemplatespecselectormatchexpressionswithoperator)
+ * [`fn withValues(values)`](#fn-spectemplatespecvolumesephemeralvolumeclaimtemplatespecselectormatchexpressionswithvalues)
+ * [`fn withValuesMixin(values)`](#fn-spectemplatespecvolumesephemeralvolumeclaimtemplatespecselectormatchexpressionswithvaluesmixin)
+ * [`obj spec.template.spec.volumes.fc`](#obj-spectemplatespecvolumesfc)
+ * [`fn withFsType(fsType)`](#fn-spectemplatespecvolumesfcwithfstype)
+ * [`fn withLun(lun)`](#fn-spectemplatespecvolumesfcwithlun)
+ * [`fn withReadOnly(readOnly)`](#fn-spectemplatespecvolumesfcwithreadonly)
+ * [`fn withTargetWWNs(targetWWNs)`](#fn-spectemplatespecvolumesfcwithtargetwwns)
+ * [`fn withTargetWWNsMixin(targetWWNs)`](#fn-spectemplatespecvolumesfcwithtargetwwnsmixin)
+ * [`fn withWwids(wwids)`](#fn-spectemplatespecvolumesfcwithwwids)
+ * [`fn withWwidsMixin(wwids)`](#fn-spectemplatespecvolumesfcwithwwidsmixin)
+ * [`obj spec.template.spec.volumes.flexVolume`](#obj-spectemplatespecvolumesflexvolume)
+ * [`fn withDriver(driver)`](#fn-spectemplatespecvolumesflexvolumewithdriver)
+ * [`fn withFsType(fsType)`](#fn-spectemplatespecvolumesflexvolumewithfstype)
+ * [`fn withOptions(options)`](#fn-spectemplatespecvolumesflexvolumewithoptions)
+ * [`fn withOptionsMixin(options)`](#fn-spectemplatespecvolumesflexvolumewithoptionsmixin)
+ * [`fn withReadOnly(readOnly)`](#fn-spectemplatespecvolumesflexvolumewithreadonly)
+ * [`obj spec.template.spec.volumes.flexVolume.secretRef`](#obj-spectemplatespecvolumesflexvolumesecretref)
+ * [`fn withName(name)`](#fn-spectemplatespecvolumesflexvolumesecretrefwithname)
+ * [`obj spec.template.spec.volumes.flocker`](#obj-spectemplatespecvolumesflocker)
+ * [`fn withDatasetName(datasetName)`](#fn-spectemplatespecvolumesflockerwithdatasetname)
+ * [`fn withDatasetUUID(datasetUUID)`](#fn-spectemplatespecvolumesflockerwithdatasetuuid)
+ * [`obj spec.template.spec.volumes.gcePersistentDisk`](#obj-spectemplatespecvolumesgcepersistentdisk)
+ * [`fn withFsType(fsType)`](#fn-spectemplatespecvolumesgcepersistentdiskwithfstype)
+ * [`fn withPartition(partition)`](#fn-spectemplatespecvolumesgcepersistentdiskwithpartition)
+ * [`fn withPdName(pdName)`](#fn-spectemplatespecvolumesgcepersistentdiskwithpdname)
+ * [`fn withReadOnly(readOnly)`](#fn-spectemplatespecvolumesgcepersistentdiskwithreadonly)
+ * [`obj spec.template.spec.volumes.gitRepo`](#obj-spectemplatespecvolumesgitrepo)
+ * [`fn withDirectory(directory)`](#fn-spectemplatespecvolumesgitrepowithdirectory)
+ * [`fn withRepository(repository)`](#fn-spectemplatespecvolumesgitrepowithrepository)
+ * [`fn withRevision(revision)`](#fn-spectemplatespecvolumesgitrepowithrevision)
+ * [`obj spec.template.spec.volumes.glusterfs`](#obj-spectemplatespecvolumesglusterfs)
+ * [`fn withEndpoints(endpoints)`](#fn-spectemplatespecvolumesglusterfswithendpoints)
+ * [`fn withPath(path)`](#fn-spectemplatespecvolumesglusterfswithpath)
+ * [`fn withReadOnly(readOnly)`](#fn-spectemplatespecvolumesglusterfswithreadonly)
+ * [`obj spec.template.spec.volumes.hostPath`](#obj-spectemplatespecvolumeshostpath)
+ * [`fn withPath(path)`](#fn-spectemplatespecvolumeshostpathwithpath)
+ * [`fn withType(type)`](#fn-spectemplatespecvolumeshostpathwithtype)
+ * [`obj spec.template.spec.volumes.iscsi`](#obj-spectemplatespecvolumesiscsi)
+ * [`fn withChapAuthDiscovery(chapAuthDiscovery)`](#fn-spectemplatespecvolumesiscsiwithchapauthdiscovery)
+ * [`fn withChapAuthSession(chapAuthSession)`](#fn-spectemplatespecvolumesiscsiwithchapauthsession)
+ * [`fn withFsType(fsType)`](#fn-spectemplatespecvolumesiscsiwithfstype)
+ * [`fn withInitiatorName(initiatorName)`](#fn-spectemplatespecvolumesiscsiwithinitiatorname)
+ * [`fn withIqn(iqn)`](#fn-spectemplatespecvolumesiscsiwithiqn)
+ * [`fn withIscsiInterface(iscsiInterface)`](#fn-spectemplatespecvolumesiscsiwithiscsiinterface)
+ * [`fn withLun(lun)`](#fn-spectemplatespecvolumesiscsiwithlun)
+ * [`fn withPortals(portals)`](#fn-spectemplatespecvolumesiscsiwithportals)
+ * [`fn withPortalsMixin(portals)`](#fn-spectemplatespecvolumesiscsiwithportalsmixin)
+ * [`fn withReadOnly(readOnly)`](#fn-spectemplatespecvolumesiscsiwithreadonly)
+ * [`fn withTargetPortal(targetPortal)`](#fn-spectemplatespecvolumesiscsiwithtargetportal)
+ * [`obj spec.template.spec.volumes.iscsi.secretRef`](#obj-spectemplatespecvolumesiscsisecretref)
+ * [`fn withName(name)`](#fn-spectemplatespecvolumesiscsisecretrefwithname)
+ * [`obj spec.template.spec.volumes.nfs`](#obj-spectemplatespecvolumesnfs)
+ * [`fn withPath(path)`](#fn-spectemplatespecvolumesnfswithpath)
+ * [`fn withReadOnly(readOnly)`](#fn-spectemplatespecvolumesnfswithreadonly)
+ * [`fn withServer(server)`](#fn-spectemplatespecvolumesnfswithserver)
+ * [`obj spec.template.spec.volumes.persistentVolumeClaim`](#obj-spectemplatespecvolumespersistentvolumeclaim)
+ * [`fn withClaimName(claimName)`](#fn-spectemplatespecvolumespersistentvolumeclaimwithclaimname)
+ * [`fn withReadOnly(readOnly)`](#fn-spectemplatespecvolumespersistentvolumeclaimwithreadonly)
+ * [`obj spec.template.spec.volumes.photonPersistentDisk`](#obj-spectemplatespecvolumesphotonpersistentdisk)
+ * [`fn withFsType(fsType)`](#fn-spectemplatespecvolumesphotonpersistentdiskwithfstype)
+ * [`fn withPdID(pdID)`](#fn-spectemplatespecvolumesphotonpersistentdiskwithpdid)
+ * [`obj spec.template.spec.volumes.portworxVolume`](#obj-spectemplatespecvolumesportworxvolume)
+ * [`fn withFsType(fsType)`](#fn-spectemplatespecvolumesportworxvolumewithfstype)
+ * [`fn withReadOnly(readOnly)`](#fn-spectemplatespecvolumesportworxvolumewithreadonly)
+ * [`fn withVolumeID(volumeID)`](#fn-spectemplatespecvolumesportworxvolumewithvolumeid)
+ * [`obj spec.template.spec.volumes.projected`](#obj-spectemplatespecvolumesprojected)
+ * [`fn withDefaultMode(defaultMode)`](#fn-spectemplatespecvolumesprojectedwithdefaultmode)
+ * [`fn withSources(sources)`](#fn-spectemplatespecvolumesprojectedwithsources)
+ * [`fn withSourcesMixin(sources)`](#fn-spectemplatespecvolumesprojectedwithsourcesmixin)
+ * [`obj spec.template.spec.volumes.projected.sources`](#obj-spectemplatespecvolumesprojectedsources)
+ * [`obj spec.template.spec.volumes.projected.sources.configMap`](#obj-spectemplatespecvolumesprojectedsourcesconfigmap)
+ * [`fn withItems(items)`](#fn-spectemplatespecvolumesprojectedsourcesconfigmapwithitems)
+ * [`fn withItemsMixin(items)`](#fn-spectemplatespecvolumesprojectedsourcesconfigmapwithitemsmixin)
+ * [`fn withName(name)`](#fn-spectemplatespecvolumesprojectedsourcesconfigmapwithname)
+ * [`fn withOptional(optional)`](#fn-spectemplatespecvolumesprojectedsourcesconfigmapwithoptional)
+ * [`obj spec.template.spec.volumes.projected.sources.configMap.items`](#obj-spectemplatespecvolumesprojectedsourcesconfigmapitems)
+ * [`fn withKey(key)`](#fn-spectemplatespecvolumesprojectedsourcesconfigmapitemswithkey)
+ * [`fn withMode(mode)`](#fn-spectemplatespecvolumesprojectedsourcesconfigmapitemswithmode)
+ * [`fn withPath(path)`](#fn-spectemplatespecvolumesprojectedsourcesconfigmapitemswithpath)
+ * [`obj spec.template.spec.volumes.projected.sources.downwardAPI`](#obj-spectemplatespecvolumesprojectedsourcesdownwardapi)
+ * [`fn withItems(items)`](#fn-spectemplatespecvolumesprojectedsourcesdownwardapiwithitems)
+ * [`fn withItemsMixin(items)`](#fn-spectemplatespecvolumesprojectedsourcesdownwardapiwithitemsmixin)
+ * [`obj spec.template.spec.volumes.projected.sources.downwardAPI.items`](#obj-spectemplatespecvolumesprojectedsourcesdownwardapiitems)
+ * [`fn withMode(mode)`](#fn-spectemplatespecvolumesprojectedsourcesdownwardapiitemswithmode)
+ * [`fn withPath(path)`](#fn-spectemplatespecvolumesprojectedsourcesdownwardapiitemswithpath)
+ * [`obj spec.template.spec.volumes.projected.sources.downwardAPI.items.fieldRef`](#obj-spectemplatespecvolumesprojectedsourcesdownwardapiitemsfieldref)
+ * [`fn withApiVersion(apiVersion)`](#fn-spectemplatespecvolumesprojectedsourcesdownwardapiitemsfieldrefwithapiversion)
+ * [`fn withFieldPath(fieldPath)`](#fn-spectemplatespecvolumesprojectedsourcesdownwardapiitemsfieldrefwithfieldpath)
+ * [`obj spec.template.spec.volumes.projected.sources.downwardAPI.items.resourceFieldRef`](#obj-spectemplatespecvolumesprojectedsourcesdownwardapiitemsresourcefieldref)
+ * [`fn withContainerName(containerName)`](#fn-spectemplatespecvolumesprojectedsourcesdownwardapiitemsresourcefieldrefwithcontainername)
+ * [`fn withDivisor(divisor)`](#fn-spectemplatespecvolumesprojectedsourcesdownwardapiitemsresourcefieldrefwithdivisor)
+ * [`fn withResource(resource)`](#fn-spectemplatespecvolumesprojectedsourcesdownwardapiitemsresourcefieldrefwithresource)
+ * [`obj spec.template.spec.volumes.projected.sources.secret`](#obj-spectemplatespecvolumesprojectedsourcessecret)
+ * [`fn withItems(items)`](#fn-spectemplatespecvolumesprojectedsourcessecretwithitems)
+ * [`fn withItemsMixin(items)`](#fn-spectemplatespecvolumesprojectedsourcessecretwithitemsmixin)
+ * [`fn withName(name)`](#fn-spectemplatespecvolumesprojectedsourcessecretwithname)
+ * [`fn withOptional(optional)`](#fn-spectemplatespecvolumesprojectedsourcessecretwithoptional)
+ * [`obj spec.template.spec.volumes.projected.sources.secret.items`](#obj-spectemplatespecvolumesprojectedsourcessecretitems)
+ * [`fn withKey(key)`](#fn-spectemplatespecvolumesprojectedsourcessecretitemswithkey)
+ * [`fn withMode(mode)`](#fn-spectemplatespecvolumesprojectedsourcessecretitemswithmode)
+ * [`fn withPath(path)`](#fn-spectemplatespecvolumesprojectedsourcessecretitemswithpath)
+ * [`obj spec.template.spec.volumes.projected.sources.serviceAccountToken`](#obj-spectemplatespecvolumesprojectedsourcesserviceaccounttoken)
+ * [`fn withAudience(audience)`](#fn-spectemplatespecvolumesprojectedsourcesserviceaccounttokenwithaudience)
+ * [`fn withExpirationSeconds(expirationSeconds)`](#fn-spectemplatespecvolumesprojectedsourcesserviceaccounttokenwithexpirationseconds)
+ * [`fn withPath(path)`](#fn-spectemplatespecvolumesprojectedsourcesserviceaccounttokenwithpath)
+ * [`obj spec.template.spec.volumes.quobyte`](#obj-spectemplatespecvolumesquobyte)
+ * [`fn withGroup(group)`](#fn-spectemplatespecvolumesquobytewithgroup)
+ * [`fn withReadOnly(readOnly)`](#fn-spectemplatespecvolumesquobytewithreadonly)
+ * [`fn withRegistry(registry)`](#fn-spectemplatespecvolumesquobytewithregistry)
+ * [`fn withTenant(tenant)`](#fn-spectemplatespecvolumesquobytewithtenant)
+ * [`fn withUser(user)`](#fn-spectemplatespecvolumesquobytewithuser)
+ * [`fn withVolume(volume)`](#fn-spectemplatespecvolumesquobytewithvolume)
+ * [`obj spec.template.spec.volumes.rbd`](#obj-spectemplatespecvolumesrbd)
+ * [`fn withFsType(fsType)`](#fn-spectemplatespecvolumesrbdwithfstype)
+ * [`fn withImage(image)`](#fn-spectemplatespecvolumesrbdwithimage)
+ * [`fn withKeyring(keyring)`](#fn-spectemplatespecvolumesrbdwithkeyring)
+ * [`fn withMonitors(monitors)`](#fn-spectemplatespecvolumesrbdwithmonitors)
+ * [`fn withMonitorsMixin(monitors)`](#fn-spectemplatespecvolumesrbdwithmonitorsmixin)
+ * [`fn withPool(pool)`](#fn-spectemplatespecvolumesrbdwithpool)
+ * [`fn withReadOnly(readOnly)`](#fn-spectemplatespecvolumesrbdwithreadonly)
+ * [`fn withUser(user)`](#fn-spectemplatespecvolumesrbdwithuser)
+ * [`obj spec.template.spec.volumes.rbd.secretRef`](#obj-spectemplatespecvolumesrbdsecretref)
+ * [`fn withName(name)`](#fn-spectemplatespecvolumesrbdsecretrefwithname)
+ * [`obj spec.template.spec.volumes.scaleIO`](#obj-spectemplatespecvolumesscaleio)
+ * [`fn withFsType(fsType)`](#fn-spectemplatespecvolumesscaleiowithfstype)
+ * [`fn withGateway(gateway)`](#fn-spectemplatespecvolumesscaleiowithgateway)
+ * [`fn withProtectionDomain(protectionDomain)`](#fn-spectemplatespecvolumesscaleiowithprotectiondomain)
+ * [`fn withReadOnly(readOnly)`](#fn-spectemplatespecvolumesscaleiowithreadonly)
+ * [`fn withSslEnabled(sslEnabled)`](#fn-spectemplatespecvolumesscaleiowithsslenabled)
+ * [`fn withStorageMode(storageMode)`](#fn-spectemplatespecvolumesscaleiowithstoragemode)
+ * [`fn withStoragePool(storagePool)`](#fn-spectemplatespecvolumesscaleiowithstoragepool)
+ * [`fn withSystem(system)`](#fn-spectemplatespecvolumesscaleiowithsystem)
+ * [`fn withVolumeName(volumeName)`](#fn-spectemplatespecvolumesscaleiowithvolumename)
+ * [`obj spec.template.spec.volumes.scaleIO.secretRef`](#obj-spectemplatespecvolumesscaleiosecretref)
+ * [`fn withName(name)`](#fn-spectemplatespecvolumesscaleiosecretrefwithname)
+ * [`obj spec.template.spec.volumes.secret`](#obj-spectemplatespecvolumessecret)
+ * [`fn withDefaultMode(defaultMode)`](#fn-spectemplatespecvolumessecretwithdefaultmode)
+ * [`fn withItems(items)`](#fn-spectemplatespecvolumessecretwithitems)
+ * [`fn withItemsMixin(items)`](#fn-spectemplatespecvolumessecretwithitemsmixin)
+ * [`fn withOptional(optional)`](#fn-spectemplatespecvolumessecretwithoptional)
+ * [`fn withSecretName(secretName)`](#fn-spectemplatespecvolumessecretwithsecretname)
+ * [`obj spec.template.spec.volumes.secret.items`](#obj-spectemplatespecvolumessecretitems)
+ * [`fn withKey(key)`](#fn-spectemplatespecvolumessecretitemswithkey)
+ * [`fn withMode(mode)`](#fn-spectemplatespecvolumessecretitemswithmode)
+ * [`fn withPath(path)`](#fn-spectemplatespecvolumessecretitemswithpath)
+ * [`obj spec.template.spec.volumes.storageos`](#obj-spectemplatespecvolumesstorageos)
+ * [`fn withFsType(fsType)`](#fn-spectemplatespecvolumesstorageoswithfstype)
+ * [`fn withReadOnly(readOnly)`](#fn-spectemplatespecvolumesstorageoswithreadonly)
+ * [`fn withVolumeName(volumeName)`](#fn-spectemplatespecvolumesstorageoswithvolumename)
+ * [`fn withVolumeNamespace(volumeNamespace)`](#fn-spectemplatespecvolumesstorageoswithvolumenamespace)
+ * [`obj spec.template.spec.volumes.storageos.secretRef`](#obj-spectemplatespecvolumesstorageossecretref)
+ * [`fn withName(name)`](#fn-spectemplatespecvolumesstorageossecretrefwithname)
+ * [`obj spec.template.spec.volumes.vsphereVolume`](#obj-spectemplatespecvolumesvspherevolume)
+ * [`fn withFsType(fsType)`](#fn-spectemplatespecvolumesvspherevolumewithfstype)
+ * [`fn withStoragePolicyID(storagePolicyID)`](#fn-spectemplatespecvolumesvspherevolumewithstoragepolicyid)
+ * [`fn withStoragePolicyName(storagePolicyName)`](#fn-spectemplatespecvolumesvspherevolumewithstoragepolicyname)
+ * [`fn withVolumePath(volumePath)`](#fn-spectemplatespecvolumesvspherevolumewithvolumepath)
+
+## Fields
+
+### fn new
+
+```ts
+new(name)
+```
+
+new returns an instance of Pooler
+
+## obj metadata
+
+"ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create."
+
+### fn metadata.withAnnotations
+
+```ts
+withAnnotations(annotations)
+```
+
+"Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations"
+
+### fn metadata.withAnnotationsMixin
+
+```ts
+withAnnotationsMixin(annotations)
+```
+
+"Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations"
+
+**Note:** This function appends passed data to existing values
+
+### fn metadata.withClusterName
+
+```ts
+withClusterName(clusterName)
+```
+
+"The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request."
+
+### fn metadata.withCreationTimestamp
+
+```ts
+withCreationTimestamp(creationTimestamp)
+```
+
+"Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers."
+
+### fn metadata.withDeletionGracePeriodSeconds
+
+```ts
+withDeletionGracePeriodSeconds(deletionGracePeriodSeconds)
+```
+
+"Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only."
+
+### fn metadata.withDeletionTimestamp
+
+```ts
+withDeletionTimestamp(deletionTimestamp)
+```
+
+"Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers."
+
+### fn metadata.withFinalizers
+
+```ts
+withFinalizers(finalizers)
+```
+
+"Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list."
+
+### fn metadata.withFinalizersMixin
+
+```ts
+withFinalizersMixin(finalizers)
+```
+
+"Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list."
+
+**Note:** This function appends passed data to existing values
+
+### fn metadata.withGenerateName
+
+```ts
+withGenerateName(generateName)
+```
+
+"GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency"
+
+### fn metadata.withGeneration
+
+```ts
+withGeneration(generation)
+```
+
+"A sequence number representing a specific generation of the desired state. Populated by the system. Read-only."
+
+### fn metadata.withLabels
+
+```ts
+withLabels(labels)
+```
+
+"Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels"
+
+### fn metadata.withLabelsMixin
+
+```ts
+withLabelsMixin(labels)
+```
+
+"Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels"
+
+**Note:** This function appends passed data to existing values
+
+### fn metadata.withName
+
+```ts
+withName(name)
+```
+
+"Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names"
+
+### fn metadata.withNamespace
+
+```ts
+withNamespace(namespace)
+```
+
+"Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the \"default\" namespace, but \"default\" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces"
+
+### fn metadata.withOwnerReferences
+
+```ts
+withOwnerReferences(ownerReferences)
+```
+
+"List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller."
+
+### fn metadata.withOwnerReferencesMixin
+
+```ts
+withOwnerReferencesMixin(ownerReferences)
+```
+
+"List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller."
+
+**Note:** This function appends passed data to existing values
+
+### fn metadata.withResourceVersion
+
+```ts
+withResourceVersion(resourceVersion)
+```
+
+"An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency"
+
+### fn metadata.withSelfLink
+
+```ts
+withSelfLink(selfLink)
+```
+
+"SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release."
+
+### fn metadata.withUid
+
+```ts
+withUid(uid)
+```
+
+"UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids"
+
+## obj spec
+
+"Specification of the desired behavior of the Pooler. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#spec-and-status"
+
+### fn spec.withInstances
+
+```ts
+withInstances(instances)
+```
+
+"The number of replicas we want. Default: 1."
+
+### fn spec.withType
+
+```ts
+withType(type)
+```
+
+"Type of service to forward traffic to. Default: `rw`."
+
+## obj spec.cluster
+
+"This is the cluster reference on which the Pooler will work. Pooler name should never match with any cluster name within the same namespace."
+
+### fn spec.cluster.withName
+
+```ts
+withName(name)
+```
+
+"Name of the referent."
+
+## obj spec.deploymentStrategy
+
+"The deployment strategy to use for pgbouncer to replace existing pods with new ones"
+
+### fn spec.deploymentStrategy.withType
+
+```ts
+withType(type)
+```
+
+"Type of deployment. Can be \"Recreate\" or \"RollingUpdate\". Default is RollingUpdate."
+
+## obj spec.deploymentStrategy.rollingUpdate
+
+"Rolling update config params. Present only if DeploymentStrategyType = RollingUpdate. --- TODO: Update this to follow our convention for oneOf, whatever we decide it to be."
+
+### fn spec.deploymentStrategy.rollingUpdate.withMaxSurge
+
+```ts
+withMaxSurge(maxSurge)
+```
+
+"The maximum number of pods that can be scheduled above the desired number of pods. Value can be an absolute number (ex: 5) or a percentage of desired pods (ex: 10%). This can not be 0 if MaxUnavailable is 0. Absolute number is calculated from percentage by rounding up. Defaults to 25%. Example: when this is set to 30%, the new ReplicaSet can be scaled up immediately when the rolling update starts, such that the total number of old and new pods do not exceed 130% of desired pods. Once old pods have been killed, new ReplicaSet can be scaled up further, ensuring that total number of pods running at any time during the update is at most 130% of desired pods."
+
+### fn spec.deploymentStrategy.rollingUpdate.withMaxUnavailable
+
+```ts
+withMaxUnavailable(maxUnavailable)
+```
+
+"The maximum number of pods that can be unavailable during the update. Value can be an absolute number (ex: 5) or a percentage of desired pods (ex: 10%). Absolute number is calculated from percentage by rounding down. This can not be 0 if MaxSurge is 0. Defaults to 25%. Example: when this is set to 30%, the old ReplicaSet can be scaled down to 70% of desired pods immediately when the rolling update starts. Once new pods are ready, old ReplicaSet can be scaled down further, followed by scaling up the new ReplicaSet, ensuring that the total number of pods available at all times during the update is at least 70% of desired pods."
+
+## obj spec.monitoring
+
+"The configuration of the monitoring infrastructure of this pooler."
+
+### fn spec.monitoring.withEnablePodMonitor
+
+```ts
+withEnablePodMonitor(enablePodMonitor)
+```
+
+"Enable or disable the `PodMonitor`"
+
+## obj spec.pgbouncer
+
+"The PgBouncer configuration"
+
+### fn spec.pgbouncer.withAuthQuery
+
+```ts
+withAuthQuery(authQuery)
+```
+
+"The query that will be used to download the hash of the password of a certain user. Default: \"SELECT usename, passwd FROM user_search($1)\". In case it is specified, also an AuthQuerySecret has to be specified and no automatic CNPG Cluster integration will be triggered."
+
+### fn spec.pgbouncer.withParameters
+
+```ts
+withParameters(parameters)
+```
+
+"Additional parameters to be passed to PgBouncer - please check the CNPG documentation for a list of options you can configure"
+
+### fn spec.pgbouncer.withParametersMixin
+
+```ts
+withParametersMixin(parameters)
+```
+
+"Additional parameters to be passed to PgBouncer - please check the CNPG documentation for a list of options you can configure"
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.pgbouncer.withPaused
+
+```ts
+withPaused(paused)
+```
+
+"When set to `true`, PgBouncer will disconnect from the PostgreSQL server, first waiting for all queries to complete, and pause all new client connections until this value is set to `false` (default). Internally, the operator calls PgBouncer's `PAUSE` and `RESUME` commands."
+
+### fn spec.pgbouncer.withPg_hba
+
+```ts
+withPg_hba(pg_hba)
+```
+
+"PostgreSQL Host Based Authentication rules (lines to be appended to the pg_hba.conf file)"
+
+### fn spec.pgbouncer.withPg_hbaMixin
+
+```ts
+withPg_hbaMixin(pg_hba)
+```
+
+"PostgreSQL Host Based Authentication rules (lines to be appended to the pg_hba.conf file)"
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.pgbouncer.withPoolMode
+
+```ts
+withPoolMode(poolMode)
+```
+
+"The pool mode. Default: `session`."
+
+## obj spec.pgbouncer.authQuerySecret
+
+"The credentials of the user that need to be used for the authentication query. In case it is specified, also an AuthQuery (e.g. \"SELECT usename, passwd FROM pg_shadow WHERE usename=$1\") has to be specified and no automatic CNPG Cluster integration will be triggered."
+
+### fn spec.pgbouncer.authQuerySecret.withName
+
+```ts
+withName(name)
+```
+
+"Name of the referent."
+
+## obj spec.template
+
+"The template of the Pod to be created"
+
+## obj spec.template.metadata
+
+"Standard object's metadata. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#metadata"
+
+### fn spec.template.metadata.withAnnotations
+
+```ts
+withAnnotations(annotations)
+```
+
+"Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations"
+
+### fn spec.template.metadata.withAnnotationsMixin
+
+```ts
+withAnnotationsMixin(annotations)
+```
+
+"Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations"
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.metadata.withLabels
+
+```ts
+withLabels(labels)
+```
+
+"Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels"
+
+### fn spec.template.metadata.withLabelsMixin
+
+```ts
+withLabelsMixin(labels)
+```
+
+"Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels"
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.template.spec
+
+"Specification of the desired behavior of the pod. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#spec-and-status"
+
+### fn spec.template.spec.withActiveDeadlineSeconds
+
+```ts
+withActiveDeadlineSeconds(activeDeadlineSeconds)
+```
+
+"Optional duration in seconds the pod may be active on the node relative to StartTime before the system will actively try to mark it failed and kill associated containers. Value must be a positive integer."
+
+### fn spec.template.spec.withAutomountServiceAccountToken
+
+```ts
+withAutomountServiceAccountToken(automountServiceAccountToken)
+```
+
+"AutomountServiceAccountToken indicates whether a service account token should be automatically mounted."
+
+### fn spec.template.spec.withContainers
+
+```ts
+withContainers(containers)
+```
+
+"List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated."
+
+### fn spec.template.spec.withContainersMixin
+
+```ts
+withContainersMixin(containers)
+```
+
+"List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.withDnsPolicy
+
+```ts
+withDnsPolicy(dnsPolicy)
+```
+
+"Set DNS policy for the pod. Defaults to \"ClusterFirst\". Valid values are 'ClusterFirstWithHostNet', 'ClusterFirst', 'Default' or 'None'. DNS parameters given in DNSConfig will be merged with the policy selected with DNSPolicy. To have DNS options set along with hostNetwork, you have to specify DNS policy explicitly to 'ClusterFirstWithHostNet'."
+
+### fn spec.template.spec.withEnableServiceLinks
+
+```ts
+withEnableServiceLinks(enableServiceLinks)
+```
+
+"EnableServiceLinks indicates whether information about services should be injected into pod's environment variables, matching the syntax of Docker links. Optional: Defaults to true."
+
+### fn spec.template.spec.withEphemeralContainers
+
+```ts
+withEphemeralContainers(ephemeralContainers)
+```
+
+"List of ephemeral containers run in this pod. Ephemeral containers may be run in an existing pod to perform user-initiated actions such as debugging. This list cannot be specified when creating a pod, and it cannot be modified by updating the pod spec. In order to add an ephemeral container to an existing pod, use the pod's ephemeralcontainers subresource."
+
+### fn spec.template.spec.withEphemeralContainersMixin
+
+```ts
+withEphemeralContainersMixin(ephemeralContainers)
+```
+
+"List of ephemeral containers run in this pod. Ephemeral containers may be run in an existing pod to perform user-initiated actions such as debugging. This list cannot be specified when creating a pod, and it cannot be modified by updating the pod spec. In order to add an ephemeral container to an existing pod, use the pod's ephemeralcontainers subresource."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.withHostAliases
+
+```ts
+withHostAliases(hostAliases)
+```
+
+"HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods."
+
+### fn spec.template.spec.withHostAliasesMixin
+
+```ts
+withHostAliasesMixin(hostAliases)
+```
+
+"HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.withHostIPC
+
+```ts
+withHostIPC(hostIPC)
+```
+
+"Use the host's ipc namespace. Optional: Default to false."
+
+### fn spec.template.spec.withHostNetwork
+
+```ts
+withHostNetwork(hostNetwork)
+```
+
+"Host networking requested for this pod. Use the host's network namespace. If this option is set, the ports that will be used must be specified. Default to false."
+
+### fn spec.template.spec.withHostPID
+
+```ts
+withHostPID(hostPID)
+```
+
+"Use the host's pid namespace. Optional: Default to false."
+
+### fn spec.template.spec.withHostUsers
+
+```ts
+withHostUsers(hostUsers)
+```
+
+"Use the host's user namespace. Optional: Default to true. If set to true or not present, the pod will be run in the host user namespace, useful for when the pod needs a feature only available to the host user namespace, such as loading a kernel module with CAP_SYS_MODULE. When set to false, a new userns is created for the pod. Setting false is useful for mitigating container breakout vulnerabilities even allowing users to run their containers as root without actually having root privileges on the host. This field is alpha-level and is only honored by servers that enable the UserNamespacesSupport feature."
+
+### fn spec.template.spec.withHostname
+
+```ts
+withHostname(hostname)
+```
+
+"Specifies the hostname of the Pod If not specified, the pod's hostname will be set to a system-defined value."
+
+### fn spec.template.spec.withImagePullSecrets
+
+```ts
+withImagePullSecrets(imagePullSecrets)
+```
+
+"ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod"
+
+### fn spec.template.spec.withImagePullSecretsMixin
+
+```ts
+withImagePullSecretsMixin(imagePullSecrets)
+```
+
+"ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod"
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.withInitContainers
+
+```ts
+withInitContainers(initContainers)
+```
+
+"List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, Liveness probes, or Startup probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/"
+
+### fn spec.template.spec.withInitContainersMixin
+
+```ts
+withInitContainersMixin(initContainers)
+```
+
+"List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, Liveness probes, or Startup probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/"
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.withNodeName
+
+```ts
+withNodeName(nodeName)
+```
+
+"NodeName is a request to schedule this pod onto a specific node. If it is non-empty, the scheduler simply schedules this pod onto that node, assuming that it fits resource requirements."
+
+### fn spec.template.spec.withNodeSelector
+
+```ts
+withNodeSelector(nodeSelector)
+```
+
+"NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/"
+
+### fn spec.template.spec.withNodeSelectorMixin
+
+```ts
+withNodeSelectorMixin(nodeSelector)
+```
+
+"NodeSelector is a selector which must be true for the pod to fit on a node. Selector which must match a node's labels for the pod to be scheduled on that node. More info: https://kubernetes.io/docs/concepts/configuration/assign-pod-node/"
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.withOverhead
+
+```ts
+withOverhead(overhead)
+```
+
+"Overhead represents the resource overhead associated with running a pod for a given RuntimeClass. This field will be autopopulated at admission time by the RuntimeClass admission controller. If the RuntimeClass admission controller is enabled, overhead must not be set in Pod create requests. The RuntimeClass admission controller will reject Pod create requests which have the overhead already set. If RuntimeClass is configured and selected in the PodSpec, Overhead will be set to the value defined in the corresponding RuntimeClass, otherwise it will remain unset and treated as zero. More info: https://git.k8s.io/enhancements/keps/sig-node/688-pod-overhead/README.md"
+
+### fn spec.template.spec.withOverheadMixin
+
+```ts
+withOverheadMixin(overhead)
+```
+
+"Overhead represents the resource overhead associated with running a pod for a given RuntimeClass. This field will be autopopulated at admission time by the RuntimeClass admission controller. If the RuntimeClass admission controller is enabled, overhead must not be set in Pod create requests. The RuntimeClass admission controller will reject Pod create requests which have the overhead already set. If RuntimeClass is configured and selected in the PodSpec, Overhead will be set to the value defined in the corresponding RuntimeClass, otherwise it will remain unset and treated as zero. More info: https://git.k8s.io/enhancements/keps/sig-node/688-pod-overhead/README.md"
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.withPreemptionPolicy
+
+```ts
+withPreemptionPolicy(preemptionPolicy)
+```
+
+"PreemptionPolicy is the Policy for preempting pods with lower priority. One of Never, PreemptLowerPriority. Defaults to PreemptLowerPriority if unset."
+
+### fn spec.template.spec.withPriority
+
+```ts
+withPriority(priority)
+```
+
+"The priority value. Various system components use this field to find the priority of the pod. When Priority Admission Controller is enabled, it prevents users from setting this field. The admission controller populates this field from PriorityClassName. The higher the value, the higher the priority."
+
+### fn spec.template.spec.withPriorityClassName
+
+```ts
+withPriorityClassName(priorityClassName)
+```
+
+"If specified, indicates the pod's priority. \"system-node-critical\" and \"system-cluster-critical\" are two special keywords which indicate the highest priorities with the former being the highest priority. Any other name must be defined by creating a PriorityClass object with that name. If not specified, the pod priority will be default or zero if there is no default."
+
+### fn spec.template.spec.withReadinessGates
+
+```ts
+withReadinessGates(readinessGates)
+```
+
+"If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to \"True\" More info: https://git.k8s.io/enhancements/keps/sig-network/580-pod-readiness-gates"
+
+### fn spec.template.spec.withReadinessGatesMixin
+
+```ts
+withReadinessGatesMixin(readinessGates)
+```
+
+"If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to \"True\" More info: https://git.k8s.io/enhancements/keps/sig-network/580-pod-readiness-gates"
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.withResourceClaims
+
+```ts
+withResourceClaims(resourceClaims)
+```
+
+"ResourceClaims defines which ResourceClaims must be allocated and reserved before the Pod is allowed to start. The resources will be made available to those containers which consume them by name. \n This is an alpha field and requires enabling the DynamicResourceAllocation feature gate. \n This field is immutable."
+
+### fn spec.template.spec.withResourceClaimsMixin
+
+```ts
+withResourceClaimsMixin(resourceClaims)
+```
+
+"ResourceClaims defines which ResourceClaims must be allocated and reserved before the Pod is allowed to start. The resources will be made available to those containers which consume them by name. \n This is an alpha field and requires enabling the DynamicResourceAllocation feature gate. \n This field is immutable."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.withRestartPolicy
+
+```ts
+withRestartPolicy(restartPolicy)
+```
+
+"Restart policy for all containers within the pod. One of Always, OnFailure, Never. In some contexts, only a subset of those values may be permitted. Default to Always. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle/#restart-policy"
+
+### fn spec.template.spec.withRuntimeClassName
+
+```ts
+withRuntimeClassName(runtimeClassName)
+```
+
+"RuntimeClassName refers to a RuntimeClass object in the node.k8s.io group, which should be used to run this pod. If no RuntimeClass resource matches the named class, the pod will not be run. If unset or empty, the \"legacy\" RuntimeClass will be used, which is an implicit class with an empty definition that uses the default runtime handler. More info: https://git.k8s.io/enhancements/keps/sig-node/585-runtime-class"
+
+### fn spec.template.spec.withSchedulerName
+
+```ts
+withSchedulerName(schedulerName)
+```
+
+"If specified, the pod will be dispatched by specified scheduler. If not specified, the pod will be dispatched by default scheduler."
+
+### fn spec.template.spec.withSchedulingGates
+
+```ts
+withSchedulingGates(schedulingGates)
+```
+
+"SchedulingGates is an opaque list of values that if specified will block scheduling the pod. If schedulingGates is not empty, the pod will stay in the SchedulingGated state and the scheduler will not attempt to schedule the pod. \n SchedulingGates can only be set at pod creation time, and be removed only afterwards. \n This is a beta feature enabled by the PodSchedulingReadiness feature gate."
+
+### fn spec.template.spec.withSchedulingGatesMixin
+
+```ts
+withSchedulingGatesMixin(schedulingGates)
+```
+
+"SchedulingGates is an opaque list of values that if specified will block scheduling the pod. If schedulingGates is not empty, the pod will stay in the SchedulingGated state and the scheduler will not attempt to schedule the pod. \n SchedulingGates can only be set at pod creation time, and be removed only afterwards. \n This is a beta feature enabled by the PodSchedulingReadiness feature gate."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.withServiceAccount
+
+```ts
+withServiceAccount(serviceAccount)
+```
+
+"DeprecatedServiceAccount is a depreciated alias for ServiceAccountName. Deprecated: Use serviceAccountName instead."
+
+### fn spec.template.spec.withServiceAccountName
+
+```ts
+withServiceAccountName(serviceAccountName)
+```
+
+"ServiceAccountName is the name of the ServiceAccount to use to run this pod. More info: https://kubernetes.io/docs/tasks/configure-pod-container/configure-service-account/"
+
+### fn spec.template.spec.withSetHostnameAsFQDN
+
+```ts
+withSetHostnameAsFQDN(setHostnameAsFQDN)
+```
+
+"If true the pod's hostname will be configured as the pod's FQDN, rather than the leaf name (the default). In Linux containers, this means setting the FQDN in the hostname field of the kernel (the nodename field of struct utsname). In Windows containers, this means setting the registry value of hostname for the registry key HKEY_LOCAL_MACHINE\\\\SYSTEM\\\\CurrentControlSet\\\\Services\\\\Tcpip\\\\Parameters to FQDN. If a pod does not have FQDN, this has no effect. Default to false."
+
+### fn spec.template.spec.withShareProcessNamespace
+
+```ts
+withShareProcessNamespace(shareProcessNamespace)
+```
+
+"Share a single process namespace between all of the containers in a pod. When this is set containers will be able to view and signal processes from other containers in the same pod, and the first process in each container will not be assigned PID 1. HostPID and ShareProcessNamespace cannot both be set. Optional: Default to false."
+
+### fn spec.template.spec.withSubdomain
+
+```ts
+withSubdomain(subdomain)
+```
+
+"If specified, the fully qualified Pod hostname will be \"...svc.\". If not specified, the pod will not have a domainname at all."
+
+### fn spec.template.spec.withTerminationGracePeriodSeconds
+
+```ts
+withTerminationGracePeriodSeconds(terminationGracePeriodSeconds)
+```
+
+"Optional duration in seconds the pod needs to terminate gracefully. May be decreased in delete request. Value must be non-negative integer. The value zero indicates stop immediately via the kill signal (no opportunity to shut down). If this value is nil, the default grace period will be used instead. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. Defaults to 30 seconds."
+
+### fn spec.template.spec.withTolerations
+
+```ts
+withTolerations(tolerations)
+```
+
+"If specified, the pod's tolerations."
+
+### fn spec.template.spec.withTolerationsMixin
+
+```ts
+withTolerationsMixin(tolerations)
+```
+
+"If specified, the pod's tolerations."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.withTopologySpreadConstraints
+
+```ts
+withTopologySpreadConstraints(topologySpreadConstraints)
+```
+
+"TopologySpreadConstraints describes how a group of pods ought to spread across topology domains. Scheduler will schedule pods in a way which abides by the constraints. All topologySpreadConstraints are ANDed."
+
+### fn spec.template.spec.withTopologySpreadConstraintsMixin
+
+```ts
+withTopologySpreadConstraintsMixin(topologySpreadConstraints)
+```
+
+"TopologySpreadConstraints describes how a group of pods ought to spread across topology domains. Scheduler will schedule pods in a way which abides by the constraints. All topologySpreadConstraints are ANDed."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.withVolumes
+
+```ts
+withVolumes(volumes)
+```
+
+"List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes"
+
+### fn spec.template.spec.withVolumesMixin
+
+```ts
+withVolumesMixin(volumes)
+```
+
+"List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes"
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.template.spec.affinity
+
+"If specified, the pod's scheduling constraints"
+
+## obj spec.template.spec.affinity.nodeAffinity
+
+"Describes node affinity scheduling rules for the pod."
+
+### fn spec.template.spec.affinity.nodeAffinity.withPreferredDuringSchedulingIgnoredDuringExecution
+
+```ts
+withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution)
+```
+
+"The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding \"weight\" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred."
+
+### fn spec.template.spec.affinity.nodeAffinity.withPreferredDuringSchedulingIgnoredDuringExecutionMixin
+
+```ts
+withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution)
+```
+
+"The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding \"weight\" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred."
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.template.spec.affinity.nodeAffinity.preferredDuringSchedulingIgnoredDuringExecution
+
+"The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding \"weight\" to the sum if the node matches the corresponding matchExpressions; the node(s) with the highest sum are the most preferred."
+
+### fn spec.template.spec.affinity.nodeAffinity.preferredDuringSchedulingIgnoredDuringExecution.withWeight
+
+```ts
+withWeight(weight)
+```
+
+"Weight associated with matching the corresponding nodeSelectorTerm, in the range 1-100."
+
+## obj spec.template.spec.affinity.nodeAffinity.preferredDuringSchedulingIgnoredDuringExecution.preference
+
+"A node selector term, associated with the corresponding weight."
+
+### fn spec.template.spec.affinity.nodeAffinity.preferredDuringSchedulingIgnoredDuringExecution.preference.withMatchExpressions
+
+```ts
+withMatchExpressions(matchExpressions)
+```
+
+"A list of node selector requirements by node's labels."
+
+### fn spec.template.spec.affinity.nodeAffinity.preferredDuringSchedulingIgnoredDuringExecution.preference.withMatchExpressionsMixin
+
+```ts
+withMatchExpressionsMixin(matchExpressions)
+```
+
+"A list of node selector requirements by node's labels."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.affinity.nodeAffinity.preferredDuringSchedulingIgnoredDuringExecution.preference.withMatchFields
+
+```ts
+withMatchFields(matchFields)
+```
+
+"A list of node selector requirements by node's fields."
+
+### fn spec.template.spec.affinity.nodeAffinity.preferredDuringSchedulingIgnoredDuringExecution.preference.withMatchFieldsMixin
+
+```ts
+withMatchFieldsMixin(matchFields)
+```
+
+"A list of node selector requirements by node's fields."
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.template.spec.affinity.nodeAffinity.preferredDuringSchedulingIgnoredDuringExecution.preference.matchExpressions
+
+"A list of node selector requirements by node's labels."
+
+### fn spec.template.spec.affinity.nodeAffinity.preferredDuringSchedulingIgnoredDuringExecution.preference.matchExpressions.withKey
+
+```ts
+withKey(key)
+```
+
+"The label key that the selector applies to."
+
+### fn spec.template.spec.affinity.nodeAffinity.preferredDuringSchedulingIgnoredDuringExecution.preference.matchExpressions.withOperator
+
+```ts
+withOperator(operator)
+```
+
+"Represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists, DoesNotExist. Gt, and Lt."
+
+### fn spec.template.spec.affinity.nodeAffinity.preferredDuringSchedulingIgnoredDuringExecution.preference.matchExpressions.withValues
+
+```ts
+withValues(values)
+```
+
+"An array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. If the operator is Gt or Lt, the values array must have a single element, which will be interpreted as an integer. This array is replaced during a strategic merge patch."
+
+### fn spec.template.spec.affinity.nodeAffinity.preferredDuringSchedulingIgnoredDuringExecution.preference.matchExpressions.withValuesMixin
+
+```ts
+withValuesMixin(values)
+```
+
+"An array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. If the operator is Gt or Lt, the values array must have a single element, which will be interpreted as an integer. This array is replaced during a strategic merge patch."
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.template.spec.affinity.nodeAffinity.preferredDuringSchedulingIgnoredDuringExecution.preference.matchFields
+
+"A list of node selector requirements by node's fields."
+
+### fn spec.template.spec.affinity.nodeAffinity.preferredDuringSchedulingIgnoredDuringExecution.preference.matchFields.withKey
+
+```ts
+withKey(key)
+```
+
+"The label key that the selector applies to."
+
+### fn spec.template.spec.affinity.nodeAffinity.preferredDuringSchedulingIgnoredDuringExecution.preference.matchFields.withOperator
+
+```ts
+withOperator(operator)
+```
+
+"Represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists, DoesNotExist. Gt, and Lt."
+
+### fn spec.template.spec.affinity.nodeAffinity.preferredDuringSchedulingIgnoredDuringExecution.preference.matchFields.withValues
+
+```ts
+withValues(values)
+```
+
+"An array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. If the operator is Gt or Lt, the values array must have a single element, which will be interpreted as an integer. This array is replaced during a strategic merge patch."
+
+### fn spec.template.spec.affinity.nodeAffinity.preferredDuringSchedulingIgnoredDuringExecution.preference.matchFields.withValuesMixin
+
+```ts
+withValuesMixin(values)
+```
+
+"An array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. If the operator is Gt or Lt, the values array must have a single element, which will be interpreted as an integer. This array is replaced during a strategic merge patch."
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.template.spec.affinity.nodeAffinity.requiredDuringSchedulingIgnoredDuringExecution
+
+"If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to an update), the system may or may not try to eventually evict the pod from its node."
+
+### fn spec.template.spec.affinity.nodeAffinity.requiredDuringSchedulingIgnoredDuringExecution.withNodeSelectorTerms
+
+```ts
+withNodeSelectorTerms(nodeSelectorTerms)
+```
+
+"Required. A list of node selector terms. The terms are ORed."
+
+### fn spec.template.spec.affinity.nodeAffinity.requiredDuringSchedulingIgnoredDuringExecution.withNodeSelectorTermsMixin
+
+```ts
+withNodeSelectorTermsMixin(nodeSelectorTerms)
+```
+
+"Required. A list of node selector terms. The terms are ORed."
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.template.spec.affinity.nodeAffinity.requiredDuringSchedulingIgnoredDuringExecution.nodeSelectorTerms
+
+"Required. A list of node selector terms. The terms are ORed."
+
+### fn spec.template.spec.affinity.nodeAffinity.requiredDuringSchedulingIgnoredDuringExecution.nodeSelectorTerms.withMatchExpressions
+
+```ts
+withMatchExpressions(matchExpressions)
+```
+
+"A list of node selector requirements by node's labels."
+
+### fn spec.template.spec.affinity.nodeAffinity.requiredDuringSchedulingIgnoredDuringExecution.nodeSelectorTerms.withMatchExpressionsMixin
+
+```ts
+withMatchExpressionsMixin(matchExpressions)
+```
+
+"A list of node selector requirements by node's labels."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.affinity.nodeAffinity.requiredDuringSchedulingIgnoredDuringExecution.nodeSelectorTerms.withMatchFields
+
+```ts
+withMatchFields(matchFields)
+```
+
+"A list of node selector requirements by node's fields."
+
+### fn spec.template.spec.affinity.nodeAffinity.requiredDuringSchedulingIgnoredDuringExecution.nodeSelectorTerms.withMatchFieldsMixin
+
+```ts
+withMatchFieldsMixin(matchFields)
+```
+
+"A list of node selector requirements by node's fields."
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.template.spec.affinity.nodeAffinity.requiredDuringSchedulingIgnoredDuringExecution.nodeSelectorTerms.matchExpressions
+
+"A list of node selector requirements by node's labels."
+
+### fn spec.template.spec.affinity.nodeAffinity.requiredDuringSchedulingIgnoredDuringExecution.nodeSelectorTerms.matchExpressions.withKey
+
+```ts
+withKey(key)
+```
+
+"The label key that the selector applies to."
+
+### fn spec.template.spec.affinity.nodeAffinity.requiredDuringSchedulingIgnoredDuringExecution.nodeSelectorTerms.matchExpressions.withOperator
+
+```ts
+withOperator(operator)
+```
+
+"Represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists, DoesNotExist. Gt, and Lt."
+
+### fn spec.template.spec.affinity.nodeAffinity.requiredDuringSchedulingIgnoredDuringExecution.nodeSelectorTerms.matchExpressions.withValues
+
+```ts
+withValues(values)
+```
+
+"An array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. If the operator is Gt or Lt, the values array must have a single element, which will be interpreted as an integer. This array is replaced during a strategic merge patch."
+
+### fn spec.template.spec.affinity.nodeAffinity.requiredDuringSchedulingIgnoredDuringExecution.nodeSelectorTerms.matchExpressions.withValuesMixin
+
+```ts
+withValuesMixin(values)
+```
+
+"An array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. If the operator is Gt or Lt, the values array must have a single element, which will be interpreted as an integer. This array is replaced during a strategic merge patch."
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.template.spec.affinity.nodeAffinity.requiredDuringSchedulingIgnoredDuringExecution.nodeSelectorTerms.matchFields
+
+"A list of node selector requirements by node's fields."
+
+### fn spec.template.spec.affinity.nodeAffinity.requiredDuringSchedulingIgnoredDuringExecution.nodeSelectorTerms.matchFields.withKey
+
+```ts
+withKey(key)
+```
+
+"The label key that the selector applies to."
+
+### fn spec.template.spec.affinity.nodeAffinity.requiredDuringSchedulingIgnoredDuringExecution.nodeSelectorTerms.matchFields.withOperator
+
+```ts
+withOperator(operator)
+```
+
+"Represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists, DoesNotExist. Gt, and Lt."
+
+### fn spec.template.spec.affinity.nodeAffinity.requiredDuringSchedulingIgnoredDuringExecution.nodeSelectorTerms.matchFields.withValues
+
+```ts
+withValues(values)
+```
+
+"An array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. If the operator is Gt or Lt, the values array must have a single element, which will be interpreted as an integer. This array is replaced during a strategic merge patch."
+
+### fn spec.template.spec.affinity.nodeAffinity.requiredDuringSchedulingIgnoredDuringExecution.nodeSelectorTerms.matchFields.withValuesMixin
+
+```ts
+withValuesMixin(values)
+```
+
+"An array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. If the operator is Gt or Lt, the values array must have a single element, which will be interpreted as an integer. This array is replaced during a strategic merge patch."
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.template.spec.affinity.podAffinity
+
+"Describes pod affinity scheduling rules (e.g. co-locate this pod in the same node, zone, etc. as some other pod(s))."
+
+### fn spec.template.spec.affinity.podAffinity.withPreferredDuringSchedulingIgnoredDuringExecution
+
+```ts
+withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution)
+```
+
+"The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding \"weight\" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred."
+
+### fn spec.template.spec.affinity.podAffinity.withPreferredDuringSchedulingIgnoredDuringExecutionMixin
+
+```ts
+withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution)
+```
+
+"The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding \"weight\" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.affinity.podAffinity.withRequiredDuringSchedulingIgnoredDuringExecution
+
+```ts
+withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution)
+```
+
+"If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied."
+
+### fn spec.template.spec.affinity.podAffinity.withRequiredDuringSchedulingIgnoredDuringExecutionMixin
+
+```ts
+withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution)
+```
+
+"If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied."
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.template.spec.affinity.podAffinity.preferredDuringSchedulingIgnoredDuringExecution
+
+"The scheduler will prefer to schedule pods to nodes that satisfy the affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding \"weight\" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred."
+
+### fn spec.template.spec.affinity.podAffinity.preferredDuringSchedulingIgnoredDuringExecution.withWeight
+
+```ts
+withWeight(weight)
+```
+
+"weight associated with matching the corresponding podAffinityTerm, in the range 1-100."
+
+## obj spec.template.spec.affinity.podAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm
+
+"Required. A pod affinity term, associated with the corresponding weight."
+
+### fn spec.template.spec.affinity.podAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.withNamespaces
+
+```ts
+withNamespaces(namespaces)
+```
+
+"namespaces specifies a static list of namespace names that the term applies to. The term is applied to the union of the namespaces listed in this field and the ones selected by namespaceSelector. null or empty namespaces list and null namespaceSelector means \"this pod's namespace\"."
+
+### fn spec.template.spec.affinity.podAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.withNamespacesMixin
+
+```ts
+withNamespacesMixin(namespaces)
+```
+
+"namespaces specifies a static list of namespace names that the term applies to. The term is applied to the union of the namespaces listed in this field and the ones selected by namespaceSelector. null or empty namespaces list and null namespaceSelector means \"this pod's namespace\"."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.affinity.podAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.withTopologyKey
+
+```ts
+withTopologyKey(topologyKey)
+```
+
+"This pod should be co-located (affinity) or not co-located (anti-affinity) with the pods matching the labelSelector in the specified namespaces, where co-located is defined as running on a node whose value of the label with key topologyKey matches that of any node on which any of the selected pods is running. Empty topologyKey is not allowed."
+
+## obj spec.template.spec.affinity.podAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.labelSelector
+
+"A label query over a set of resources, in this case pods."
+
+### fn spec.template.spec.affinity.podAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.labelSelector.withMatchExpressions
+
+```ts
+withMatchExpressions(matchExpressions)
+```
+
+"matchExpressions is a list of label selector requirements. The requirements are ANDed."
+
+### fn spec.template.spec.affinity.podAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.labelSelector.withMatchExpressionsMixin
+
+```ts
+withMatchExpressionsMixin(matchExpressions)
+```
+
+"matchExpressions is a list of label selector requirements. The requirements are ANDed."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.affinity.podAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.labelSelector.withMatchLabels
+
+```ts
+withMatchLabels(matchLabels)
+```
+
+"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \"key\", the operator is \"In\", and the values array contains only \"value\". The requirements are ANDed."
+
+### fn spec.template.spec.affinity.podAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.labelSelector.withMatchLabelsMixin
+
+```ts
+withMatchLabelsMixin(matchLabels)
+```
+
+"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \"key\", the operator is \"In\", and the values array contains only \"value\". The requirements are ANDed."
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.template.spec.affinity.podAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.labelSelector.matchExpressions
+
+"matchExpressions is a list of label selector requirements. The requirements are ANDed."
+
+### fn spec.template.spec.affinity.podAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.labelSelector.matchExpressions.withKey
+
+```ts
+withKey(key)
+```
+
+"key is the label key that the selector applies to."
+
+### fn spec.template.spec.affinity.podAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.labelSelector.matchExpressions.withOperator
+
+```ts
+withOperator(operator)
+```
+
+"operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist."
+
+### fn spec.template.spec.affinity.podAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.labelSelector.matchExpressions.withValues
+
+```ts
+withValues(values)
+```
+
+"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."
+
+### fn spec.template.spec.affinity.podAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.labelSelector.matchExpressions.withValuesMixin
+
+```ts
+withValuesMixin(values)
+```
+
+"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.template.spec.affinity.podAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.namespaceSelector
+
+"A label query over the set of namespaces that the term applies to. The term is applied to the union of the namespaces selected by this field and the ones listed in the namespaces field. null selector and null or empty namespaces list means \"this pod's namespace\". An empty selector ({}) matches all namespaces."
+
+### fn spec.template.spec.affinity.podAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.namespaceSelector.withMatchExpressions
+
+```ts
+withMatchExpressions(matchExpressions)
+```
+
+"matchExpressions is a list of label selector requirements. The requirements are ANDed."
+
+### fn spec.template.spec.affinity.podAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.namespaceSelector.withMatchExpressionsMixin
+
+```ts
+withMatchExpressionsMixin(matchExpressions)
+```
+
+"matchExpressions is a list of label selector requirements. The requirements are ANDed."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.affinity.podAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.namespaceSelector.withMatchLabels
+
+```ts
+withMatchLabels(matchLabels)
+```
+
+"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \"key\", the operator is \"In\", and the values array contains only \"value\". The requirements are ANDed."
+
+### fn spec.template.spec.affinity.podAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.namespaceSelector.withMatchLabelsMixin
+
+```ts
+withMatchLabelsMixin(matchLabels)
+```
+
+"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \"key\", the operator is \"In\", and the values array contains only \"value\". The requirements are ANDed."
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.template.spec.affinity.podAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.namespaceSelector.matchExpressions
+
+"matchExpressions is a list of label selector requirements. The requirements are ANDed."
+
+### fn spec.template.spec.affinity.podAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.namespaceSelector.matchExpressions.withKey
+
+```ts
+withKey(key)
+```
+
+"key is the label key that the selector applies to."
+
+### fn spec.template.spec.affinity.podAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.namespaceSelector.matchExpressions.withOperator
+
+```ts
+withOperator(operator)
+```
+
+"operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist."
+
+### fn spec.template.spec.affinity.podAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.namespaceSelector.matchExpressions.withValues
+
+```ts
+withValues(values)
+```
+
+"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."
+
+### fn spec.template.spec.affinity.podAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.namespaceSelector.matchExpressions.withValuesMixin
+
+```ts
+withValuesMixin(values)
+```
+
+"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.template.spec.affinity.podAffinity.requiredDuringSchedulingIgnoredDuringExecution
+
+"If the affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied."
+
+### fn spec.template.spec.affinity.podAffinity.requiredDuringSchedulingIgnoredDuringExecution.withNamespaces
+
+```ts
+withNamespaces(namespaces)
+```
+
+"namespaces specifies a static list of namespace names that the term applies to. The term is applied to the union of the namespaces listed in this field and the ones selected by namespaceSelector. null or empty namespaces list and null namespaceSelector means \"this pod's namespace\"."
+
+### fn spec.template.spec.affinity.podAffinity.requiredDuringSchedulingIgnoredDuringExecution.withNamespacesMixin
+
+```ts
+withNamespacesMixin(namespaces)
+```
+
+"namespaces specifies a static list of namespace names that the term applies to. The term is applied to the union of the namespaces listed in this field and the ones selected by namespaceSelector. null or empty namespaces list and null namespaceSelector means \"this pod's namespace\"."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.affinity.podAffinity.requiredDuringSchedulingIgnoredDuringExecution.withTopologyKey
+
+```ts
+withTopologyKey(topologyKey)
+```
+
+"This pod should be co-located (affinity) or not co-located (anti-affinity) with the pods matching the labelSelector in the specified namespaces, where co-located is defined as running on a node whose value of the label with key topologyKey matches that of any node on which any of the selected pods is running. Empty topologyKey is not allowed."
+
+## obj spec.template.spec.affinity.podAffinity.requiredDuringSchedulingIgnoredDuringExecution.labelSelector
+
+"A label query over a set of resources, in this case pods."
+
+### fn spec.template.spec.affinity.podAffinity.requiredDuringSchedulingIgnoredDuringExecution.labelSelector.withMatchExpressions
+
+```ts
+withMatchExpressions(matchExpressions)
+```
+
+"matchExpressions is a list of label selector requirements. The requirements are ANDed."
+
+### fn spec.template.spec.affinity.podAffinity.requiredDuringSchedulingIgnoredDuringExecution.labelSelector.withMatchExpressionsMixin
+
+```ts
+withMatchExpressionsMixin(matchExpressions)
+```
+
+"matchExpressions is a list of label selector requirements. The requirements are ANDed."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.affinity.podAffinity.requiredDuringSchedulingIgnoredDuringExecution.labelSelector.withMatchLabels
+
+```ts
+withMatchLabels(matchLabels)
+```
+
+"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \"key\", the operator is \"In\", and the values array contains only \"value\". The requirements are ANDed."
+
+### fn spec.template.spec.affinity.podAffinity.requiredDuringSchedulingIgnoredDuringExecution.labelSelector.withMatchLabelsMixin
+
+```ts
+withMatchLabelsMixin(matchLabels)
+```
+
+"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \"key\", the operator is \"In\", and the values array contains only \"value\". The requirements are ANDed."
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.template.spec.affinity.podAffinity.requiredDuringSchedulingIgnoredDuringExecution.labelSelector.matchExpressions
+
+"matchExpressions is a list of label selector requirements. The requirements are ANDed."
+
+### fn spec.template.spec.affinity.podAffinity.requiredDuringSchedulingIgnoredDuringExecution.labelSelector.matchExpressions.withKey
+
+```ts
+withKey(key)
+```
+
+"key is the label key that the selector applies to."
+
+### fn spec.template.spec.affinity.podAffinity.requiredDuringSchedulingIgnoredDuringExecution.labelSelector.matchExpressions.withOperator
+
+```ts
+withOperator(operator)
+```
+
+"operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist."
+
+### fn spec.template.spec.affinity.podAffinity.requiredDuringSchedulingIgnoredDuringExecution.labelSelector.matchExpressions.withValues
+
+```ts
+withValues(values)
+```
+
+"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."
+
+### fn spec.template.spec.affinity.podAffinity.requiredDuringSchedulingIgnoredDuringExecution.labelSelector.matchExpressions.withValuesMixin
+
+```ts
+withValuesMixin(values)
+```
+
+"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.template.spec.affinity.podAffinity.requiredDuringSchedulingIgnoredDuringExecution.namespaceSelector
+
+"A label query over the set of namespaces that the term applies to. The term is applied to the union of the namespaces selected by this field and the ones listed in the namespaces field. null selector and null or empty namespaces list means \"this pod's namespace\". An empty selector ({}) matches all namespaces."
+
+### fn spec.template.spec.affinity.podAffinity.requiredDuringSchedulingIgnoredDuringExecution.namespaceSelector.withMatchExpressions
+
+```ts
+withMatchExpressions(matchExpressions)
+```
+
+"matchExpressions is a list of label selector requirements. The requirements are ANDed."
+
+### fn spec.template.spec.affinity.podAffinity.requiredDuringSchedulingIgnoredDuringExecution.namespaceSelector.withMatchExpressionsMixin
+
+```ts
+withMatchExpressionsMixin(matchExpressions)
+```
+
+"matchExpressions is a list of label selector requirements. The requirements are ANDed."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.affinity.podAffinity.requiredDuringSchedulingIgnoredDuringExecution.namespaceSelector.withMatchLabels
+
+```ts
+withMatchLabels(matchLabels)
+```
+
+"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \"key\", the operator is \"In\", and the values array contains only \"value\". The requirements are ANDed."
+
+### fn spec.template.spec.affinity.podAffinity.requiredDuringSchedulingIgnoredDuringExecution.namespaceSelector.withMatchLabelsMixin
+
+```ts
+withMatchLabelsMixin(matchLabels)
+```
+
+"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \"key\", the operator is \"In\", and the values array contains only \"value\". The requirements are ANDed."
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.template.spec.affinity.podAffinity.requiredDuringSchedulingIgnoredDuringExecution.namespaceSelector.matchExpressions
+
+"matchExpressions is a list of label selector requirements. The requirements are ANDed."
+
+### fn spec.template.spec.affinity.podAffinity.requiredDuringSchedulingIgnoredDuringExecution.namespaceSelector.matchExpressions.withKey
+
+```ts
+withKey(key)
+```
+
+"key is the label key that the selector applies to."
+
+### fn spec.template.spec.affinity.podAffinity.requiredDuringSchedulingIgnoredDuringExecution.namespaceSelector.matchExpressions.withOperator
+
+```ts
+withOperator(operator)
+```
+
+"operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist."
+
+### fn spec.template.spec.affinity.podAffinity.requiredDuringSchedulingIgnoredDuringExecution.namespaceSelector.matchExpressions.withValues
+
+```ts
+withValues(values)
+```
+
+"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."
+
+### fn spec.template.spec.affinity.podAffinity.requiredDuringSchedulingIgnoredDuringExecution.namespaceSelector.matchExpressions.withValuesMixin
+
+```ts
+withValuesMixin(values)
+```
+
+"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.template.spec.affinity.podAntiAffinity
+
+"Describes pod anti-affinity scheduling rules (e.g. avoid putting this pod in the same node, zone, etc. as some other pod(s))."
+
+### fn spec.template.spec.affinity.podAntiAffinity.withPreferredDuringSchedulingIgnoredDuringExecution
+
+```ts
+withPreferredDuringSchedulingIgnoredDuringExecution(preferredDuringSchedulingIgnoredDuringExecution)
+```
+
+"The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding \"weight\" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred."
+
+### fn spec.template.spec.affinity.podAntiAffinity.withPreferredDuringSchedulingIgnoredDuringExecutionMixin
+
+```ts
+withPreferredDuringSchedulingIgnoredDuringExecutionMixin(preferredDuringSchedulingIgnoredDuringExecution)
+```
+
+"The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding \"weight\" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.affinity.podAntiAffinity.withRequiredDuringSchedulingIgnoredDuringExecution
+
+```ts
+withRequiredDuringSchedulingIgnoredDuringExecution(requiredDuringSchedulingIgnoredDuringExecution)
+```
+
+"If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied."
+
+### fn spec.template.spec.affinity.podAntiAffinity.withRequiredDuringSchedulingIgnoredDuringExecutionMixin
+
+```ts
+withRequiredDuringSchedulingIgnoredDuringExecutionMixin(requiredDuringSchedulingIgnoredDuringExecution)
+```
+
+"If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied."
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.template.spec.affinity.podAntiAffinity.preferredDuringSchedulingIgnoredDuringExecution
+
+"The scheduler will prefer to schedule pods to nodes that satisfy the anti-affinity expressions specified by this field, but it may choose a node that violates one or more of the expressions. The node that is most preferred is the one with the greatest sum of weights, i.e. for each node that meets all of the scheduling requirements (resource request, requiredDuringScheduling anti-affinity expressions, etc.), compute a sum by iterating through the elements of this field and adding \"weight\" to the sum if the node has pods which matches the corresponding podAffinityTerm; the node(s) with the highest sum are the most preferred."
+
+### fn spec.template.spec.affinity.podAntiAffinity.preferredDuringSchedulingIgnoredDuringExecution.withWeight
+
+```ts
+withWeight(weight)
+```
+
+"weight associated with matching the corresponding podAffinityTerm, in the range 1-100."
+
+## obj spec.template.spec.affinity.podAntiAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm
+
+"Required. A pod affinity term, associated with the corresponding weight."
+
+### fn spec.template.spec.affinity.podAntiAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.withNamespaces
+
+```ts
+withNamespaces(namespaces)
+```
+
+"namespaces specifies a static list of namespace names that the term applies to. The term is applied to the union of the namespaces listed in this field and the ones selected by namespaceSelector. null or empty namespaces list and null namespaceSelector means \"this pod's namespace\"."
+
+### fn spec.template.spec.affinity.podAntiAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.withNamespacesMixin
+
+```ts
+withNamespacesMixin(namespaces)
+```
+
+"namespaces specifies a static list of namespace names that the term applies to. The term is applied to the union of the namespaces listed in this field and the ones selected by namespaceSelector. null or empty namespaces list and null namespaceSelector means \"this pod's namespace\"."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.affinity.podAntiAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.withTopologyKey
+
+```ts
+withTopologyKey(topologyKey)
+```
+
+"This pod should be co-located (affinity) or not co-located (anti-affinity) with the pods matching the labelSelector in the specified namespaces, where co-located is defined as running on a node whose value of the label with key topologyKey matches that of any node on which any of the selected pods is running. Empty topologyKey is not allowed."
+
+## obj spec.template.spec.affinity.podAntiAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.labelSelector
+
+"A label query over a set of resources, in this case pods."
+
+### fn spec.template.spec.affinity.podAntiAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.labelSelector.withMatchExpressions
+
+```ts
+withMatchExpressions(matchExpressions)
+```
+
+"matchExpressions is a list of label selector requirements. The requirements are ANDed."
+
+### fn spec.template.spec.affinity.podAntiAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.labelSelector.withMatchExpressionsMixin
+
+```ts
+withMatchExpressionsMixin(matchExpressions)
+```
+
+"matchExpressions is a list of label selector requirements. The requirements are ANDed."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.affinity.podAntiAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.labelSelector.withMatchLabels
+
+```ts
+withMatchLabels(matchLabels)
+```
+
+"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \"key\", the operator is \"In\", and the values array contains only \"value\". The requirements are ANDed."
+
+### fn spec.template.spec.affinity.podAntiAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.labelSelector.withMatchLabelsMixin
+
+```ts
+withMatchLabelsMixin(matchLabels)
+```
+
+"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \"key\", the operator is \"In\", and the values array contains only \"value\". The requirements are ANDed."
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.template.spec.affinity.podAntiAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.labelSelector.matchExpressions
+
+"matchExpressions is a list of label selector requirements. The requirements are ANDed."
+
+### fn spec.template.spec.affinity.podAntiAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.labelSelector.matchExpressions.withKey
+
+```ts
+withKey(key)
+```
+
+"key is the label key that the selector applies to."
+
+### fn spec.template.spec.affinity.podAntiAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.labelSelector.matchExpressions.withOperator
+
+```ts
+withOperator(operator)
+```
+
+"operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist."
+
+### fn spec.template.spec.affinity.podAntiAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.labelSelector.matchExpressions.withValues
+
+```ts
+withValues(values)
+```
+
+"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."
+
+### fn spec.template.spec.affinity.podAntiAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.labelSelector.matchExpressions.withValuesMixin
+
+```ts
+withValuesMixin(values)
+```
+
+"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.template.spec.affinity.podAntiAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.namespaceSelector
+
+"A label query over the set of namespaces that the term applies to. The term is applied to the union of the namespaces selected by this field and the ones listed in the namespaces field. null selector and null or empty namespaces list means \"this pod's namespace\". An empty selector ({}) matches all namespaces."
+
+### fn spec.template.spec.affinity.podAntiAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.namespaceSelector.withMatchExpressions
+
+```ts
+withMatchExpressions(matchExpressions)
+```
+
+"matchExpressions is a list of label selector requirements. The requirements are ANDed."
+
+### fn spec.template.spec.affinity.podAntiAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.namespaceSelector.withMatchExpressionsMixin
+
+```ts
+withMatchExpressionsMixin(matchExpressions)
+```
+
+"matchExpressions is a list of label selector requirements. The requirements are ANDed."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.affinity.podAntiAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.namespaceSelector.withMatchLabels
+
+```ts
+withMatchLabels(matchLabels)
+```
+
+"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \"key\", the operator is \"In\", and the values array contains only \"value\". The requirements are ANDed."
+
+### fn spec.template.spec.affinity.podAntiAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.namespaceSelector.withMatchLabelsMixin
+
+```ts
+withMatchLabelsMixin(matchLabels)
+```
+
+"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \"key\", the operator is \"In\", and the values array contains only \"value\". The requirements are ANDed."
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.template.spec.affinity.podAntiAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.namespaceSelector.matchExpressions
+
+"matchExpressions is a list of label selector requirements. The requirements are ANDed."
+
+### fn spec.template.spec.affinity.podAntiAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.namespaceSelector.matchExpressions.withKey
+
+```ts
+withKey(key)
+```
+
+"key is the label key that the selector applies to."
+
+### fn spec.template.spec.affinity.podAntiAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.namespaceSelector.matchExpressions.withOperator
+
+```ts
+withOperator(operator)
+```
+
+"operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist."
+
+### fn spec.template.spec.affinity.podAntiAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.namespaceSelector.matchExpressions.withValues
+
+```ts
+withValues(values)
+```
+
+"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."
+
+### fn spec.template.spec.affinity.podAntiAffinity.preferredDuringSchedulingIgnoredDuringExecution.podAffinityTerm.namespaceSelector.matchExpressions.withValuesMixin
+
+```ts
+withValuesMixin(values)
+```
+
+"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.template.spec.affinity.podAntiAffinity.requiredDuringSchedulingIgnoredDuringExecution
+
+"If the anti-affinity requirements specified by this field are not met at scheduling time, the pod will not be scheduled onto the node. If the anti-affinity requirements specified by this field cease to be met at some point during pod execution (e.g. due to a pod label update), the system may or may not try to eventually evict the pod from its node. When there are multiple elements, the lists of nodes corresponding to each podAffinityTerm are intersected, i.e. all terms must be satisfied."
+
+### fn spec.template.spec.affinity.podAntiAffinity.requiredDuringSchedulingIgnoredDuringExecution.withNamespaces
+
+```ts
+withNamespaces(namespaces)
+```
+
+"namespaces specifies a static list of namespace names that the term applies to. The term is applied to the union of the namespaces listed in this field and the ones selected by namespaceSelector. null or empty namespaces list and null namespaceSelector means \"this pod's namespace\"."
+
+### fn spec.template.spec.affinity.podAntiAffinity.requiredDuringSchedulingIgnoredDuringExecution.withNamespacesMixin
+
+```ts
+withNamespacesMixin(namespaces)
+```
+
+"namespaces specifies a static list of namespace names that the term applies to. The term is applied to the union of the namespaces listed in this field and the ones selected by namespaceSelector. null or empty namespaces list and null namespaceSelector means \"this pod's namespace\"."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.affinity.podAntiAffinity.requiredDuringSchedulingIgnoredDuringExecution.withTopologyKey
+
+```ts
+withTopologyKey(topologyKey)
+```
+
+"This pod should be co-located (affinity) or not co-located (anti-affinity) with the pods matching the labelSelector in the specified namespaces, where co-located is defined as running on a node whose value of the label with key topologyKey matches that of any node on which any of the selected pods is running. Empty topologyKey is not allowed."
+
+## obj spec.template.spec.affinity.podAntiAffinity.requiredDuringSchedulingIgnoredDuringExecution.labelSelector
+
+"A label query over a set of resources, in this case pods."
+
+### fn spec.template.spec.affinity.podAntiAffinity.requiredDuringSchedulingIgnoredDuringExecution.labelSelector.withMatchExpressions
+
+```ts
+withMatchExpressions(matchExpressions)
+```
+
+"matchExpressions is a list of label selector requirements. The requirements are ANDed."
+
+### fn spec.template.spec.affinity.podAntiAffinity.requiredDuringSchedulingIgnoredDuringExecution.labelSelector.withMatchExpressionsMixin
+
+```ts
+withMatchExpressionsMixin(matchExpressions)
+```
+
+"matchExpressions is a list of label selector requirements. The requirements are ANDed."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.affinity.podAntiAffinity.requiredDuringSchedulingIgnoredDuringExecution.labelSelector.withMatchLabels
+
+```ts
+withMatchLabels(matchLabels)
+```
+
+"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \"key\", the operator is \"In\", and the values array contains only \"value\". The requirements are ANDed."
+
+### fn spec.template.spec.affinity.podAntiAffinity.requiredDuringSchedulingIgnoredDuringExecution.labelSelector.withMatchLabelsMixin
+
+```ts
+withMatchLabelsMixin(matchLabels)
+```
+
+"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \"key\", the operator is \"In\", and the values array contains only \"value\". The requirements are ANDed."
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.template.spec.affinity.podAntiAffinity.requiredDuringSchedulingIgnoredDuringExecution.labelSelector.matchExpressions
+
+"matchExpressions is a list of label selector requirements. The requirements are ANDed."
+
+### fn spec.template.spec.affinity.podAntiAffinity.requiredDuringSchedulingIgnoredDuringExecution.labelSelector.matchExpressions.withKey
+
+```ts
+withKey(key)
+```
+
+"key is the label key that the selector applies to."
+
+### fn spec.template.spec.affinity.podAntiAffinity.requiredDuringSchedulingIgnoredDuringExecution.labelSelector.matchExpressions.withOperator
+
+```ts
+withOperator(operator)
+```
+
+"operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist."
+
+### fn spec.template.spec.affinity.podAntiAffinity.requiredDuringSchedulingIgnoredDuringExecution.labelSelector.matchExpressions.withValues
+
+```ts
+withValues(values)
+```
+
+"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."
+
+### fn spec.template.spec.affinity.podAntiAffinity.requiredDuringSchedulingIgnoredDuringExecution.labelSelector.matchExpressions.withValuesMixin
+
+```ts
+withValuesMixin(values)
+```
+
+"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.template.spec.affinity.podAntiAffinity.requiredDuringSchedulingIgnoredDuringExecution.namespaceSelector
+
+"A label query over the set of namespaces that the term applies to. The term is applied to the union of the namespaces selected by this field and the ones listed in the namespaces field. null selector and null or empty namespaces list means \"this pod's namespace\". An empty selector ({}) matches all namespaces."
+
+### fn spec.template.spec.affinity.podAntiAffinity.requiredDuringSchedulingIgnoredDuringExecution.namespaceSelector.withMatchExpressions
+
+```ts
+withMatchExpressions(matchExpressions)
+```
+
+"matchExpressions is a list of label selector requirements. The requirements are ANDed."
+
+### fn spec.template.spec.affinity.podAntiAffinity.requiredDuringSchedulingIgnoredDuringExecution.namespaceSelector.withMatchExpressionsMixin
+
+```ts
+withMatchExpressionsMixin(matchExpressions)
+```
+
+"matchExpressions is a list of label selector requirements. The requirements are ANDed."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.affinity.podAntiAffinity.requiredDuringSchedulingIgnoredDuringExecution.namespaceSelector.withMatchLabels
+
+```ts
+withMatchLabels(matchLabels)
+```
+
+"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \"key\", the operator is \"In\", and the values array contains only \"value\". The requirements are ANDed."
+
+### fn spec.template.spec.affinity.podAntiAffinity.requiredDuringSchedulingIgnoredDuringExecution.namespaceSelector.withMatchLabelsMixin
+
+```ts
+withMatchLabelsMixin(matchLabels)
+```
+
+"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \"key\", the operator is \"In\", and the values array contains only \"value\". The requirements are ANDed."
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.template.spec.affinity.podAntiAffinity.requiredDuringSchedulingIgnoredDuringExecution.namespaceSelector.matchExpressions
+
+"matchExpressions is a list of label selector requirements. The requirements are ANDed."
+
+### fn spec.template.spec.affinity.podAntiAffinity.requiredDuringSchedulingIgnoredDuringExecution.namespaceSelector.matchExpressions.withKey
+
+```ts
+withKey(key)
+```
+
+"key is the label key that the selector applies to."
+
+### fn spec.template.spec.affinity.podAntiAffinity.requiredDuringSchedulingIgnoredDuringExecution.namespaceSelector.matchExpressions.withOperator
+
+```ts
+withOperator(operator)
+```
+
+"operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist."
+
+### fn spec.template.spec.affinity.podAntiAffinity.requiredDuringSchedulingIgnoredDuringExecution.namespaceSelector.matchExpressions.withValues
+
+```ts
+withValues(values)
+```
+
+"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."
+
+### fn spec.template.spec.affinity.podAntiAffinity.requiredDuringSchedulingIgnoredDuringExecution.namespaceSelector.matchExpressions.withValuesMixin
+
+```ts
+withValuesMixin(values)
+```
+
+"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.template.spec.containers
+
+"List of containers belonging to the pod. Containers cannot currently be added or removed. There must be at least one container in a Pod. Cannot be updated."
+
+### fn spec.template.spec.containers.withArgs
+
+```ts
+withArgs(args)
+```
+
+"Arguments to the entrypoint. The container image's CMD is used if this is not provided. Variable references $(VAR_NAME) are expanded using the container's environment. If a variable cannot be resolved, the reference in the input string will be unchanged. Double $$ are reduced to a single $, which allows for escaping the $(VAR_NAME) syntax: i.e. \"$$(VAR_NAME)\" will produce the string literal \"$(VAR_NAME)\". Escaped references will never be expanded, regardless of whether the variable exists or not. Cannot be updated. More info: https://kubernetes.io/docs/tasks/inject-data-application/define-command-argument-container/#running-a-command-in-a-shell"
+
+### fn spec.template.spec.containers.withArgsMixin
+
+```ts
+withArgsMixin(args)
+```
+
+"Arguments to the entrypoint. The container image's CMD is used if this is not provided. Variable references $(VAR_NAME) are expanded using the container's environment. If a variable cannot be resolved, the reference in the input string will be unchanged. Double $$ are reduced to a single $, which allows for escaping the $(VAR_NAME) syntax: i.e. \"$$(VAR_NAME)\" will produce the string literal \"$(VAR_NAME)\". Escaped references will never be expanded, regardless of whether the variable exists or not. Cannot be updated. More info: https://kubernetes.io/docs/tasks/inject-data-application/define-command-argument-container/#running-a-command-in-a-shell"
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.containers.withCommand
+
+```ts
+withCommand(command)
+```
+
+"Entrypoint array. Not executed within a shell. The container image's ENTRYPOINT is used if this is not provided. Variable references $(VAR_NAME) are expanded using the container's environment. If a variable cannot be resolved, the reference in the input string will be unchanged. Double $$ are reduced to a single $, which allows for escaping the $(VAR_NAME) syntax: i.e. \"$$(VAR_NAME)\" will produce the string literal \"$(VAR_NAME)\". Escaped references will never be expanded, regardless of whether the variable exists or not. Cannot be updated. More info: https://kubernetes.io/docs/tasks/inject-data-application/define-command-argument-container/#running-a-command-in-a-shell"
+
+### fn spec.template.spec.containers.withCommandMixin
+
+```ts
+withCommandMixin(command)
+```
+
+"Entrypoint array. Not executed within a shell. The container image's ENTRYPOINT is used if this is not provided. Variable references $(VAR_NAME) are expanded using the container's environment. If a variable cannot be resolved, the reference in the input string will be unchanged. Double $$ are reduced to a single $, which allows for escaping the $(VAR_NAME) syntax: i.e. \"$$(VAR_NAME)\" will produce the string literal \"$(VAR_NAME)\". Escaped references will never be expanded, regardless of whether the variable exists or not. Cannot be updated. More info: https://kubernetes.io/docs/tasks/inject-data-application/define-command-argument-container/#running-a-command-in-a-shell"
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.containers.withEnv
+
+```ts
+withEnv(env)
+```
+
+"List of environment variables to set in the container. Cannot be updated."
+
+### fn spec.template.spec.containers.withEnvFrom
+
+```ts
+withEnvFrom(envFrom)
+```
+
+"List of sources to populate environment variables in the container. The keys defined within a source must be a C_IDENTIFIER. All invalid keys will be reported as an event when the container is starting. When a key exists in multiple sources, the value associated with the last source will take precedence. Values defined by an Env with a duplicate key will take precedence. Cannot be updated."
+
+### fn spec.template.spec.containers.withEnvFromMixin
+
+```ts
+withEnvFromMixin(envFrom)
+```
+
+"List of sources to populate environment variables in the container. The keys defined within a source must be a C_IDENTIFIER. All invalid keys will be reported as an event when the container is starting. When a key exists in multiple sources, the value associated with the last source will take precedence. Values defined by an Env with a duplicate key will take precedence. Cannot be updated."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.containers.withEnvMixin
+
+```ts
+withEnvMixin(env)
+```
+
+"List of environment variables to set in the container. Cannot be updated."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.containers.withImage
+
+```ts
+withImage(image)
+```
+
+"Container image name. More info: https://kubernetes.io/docs/concepts/containers/images This field is optional to allow higher level config management to default or override container images in workload controllers like Deployments and StatefulSets."
+
+### fn spec.template.spec.containers.withImagePullPolicy
+
+```ts
+withImagePullPolicy(imagePullPolicy)
+```
+
+"Image pull policy. One of Always, Never, IfNotPresent. Defaults to Always if :latest tag is specified, or IfNotPresent otherwise. Cannot be updated. More info: https://kubernetes.io/docs/concepts/containers/images#updating-images"
+
+### fn spec.template.spec.containers.withName
+
+```ts
+withName(name)
+```
+
+"Name of the container specified as a DNS_LABEL. Each container in a pod must have a unique name (DNS_LABEL). Cannot be updated."
+
+### fn spec.template.spec.containers.withPorts
+
+```ts
+withPorts(ports)
+```
+
+"List of ports to expose from the container. Not specifying a port here DOES NOT prevent that port from being exposed. Any port which is listening on the default \"0.0.0.0\" address inside a container will be accessible from the network. Modifying this array with strategic merge patch may corrupt the data. For more information See https://github.com/kubernetes/kubernetes/issues/108255. Cannot be updated."
+
+### fn spec.template.spec.containers.withPortsMixin
+
+```ts
+withPortsMixin(ports)
+```
+
+"List of ports to expose from the container. Not specifying a port here DOES NOT prevent that port from being exposed. Any port which is listening on the default \"0.0.0.0\" address inside a container will be accessible from the network. Modifying this array with strategic merge patch may corrupt the data. For more information See https://github.com/kubernetes/kubernetes/issues/108255. Cannot be updated."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.containers.withResizePolicy
+
+```ts
+withResizePolicy(resizePolicy)
+```
+
+"Resources resize policy for the container."
+
+### fn spec.template.spec.containers.withResizePolicyMixin
+
+```ts
+withResizePolicyMixin(resizePolicy)
+```
+
+"Resources resize policy for the container."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.containers.withRestartPolicy
+
+```ts
+withRestartPolicy(restartPolicy)
+```
+
+"RestartPolicy defines the restart behavior of individual containers in a pod. This field may only be set for init containers, and the only allowed value is \"Always\". For non-init containers or when this field is not specified, the restart behavior is defined by the Pod's restart policy and the container type. Setting the RestartPolicy as \"Always\" for the init container will have the following effect: this init container will be continually restarted on exit until all regular containers have terminated. Once all regular containers have completed, all init containers with restartPolicy \"Always\" will be shut down. This lifecycle differs from normal init containers and is often referred to as a \"sidecar\" container. Although this init container still starts in the init container sequence, it does not wait for the container to complete before proceeding to the next init container. Instead, the next init container starts immediately after this init container is started, or after any startupProbe has successfully completed."
+
+### fn spec.template.spec.containers.withStdin
+
+```ts
+withStdin(stdin)
+```
+
+"Whether this container should allocate a buffer for stdin in the container runtime. If this is not set, reads from stdin in the container will always result in EOF. Default is false."
+
+### fn spec.template.spec.containers.withStdinOnce
+
+```ts
+withStdinOnce(stdinOnce)
+```
+
+"Whether the container runtime should close the stdin channel after it has been opened by a single attach. When stdin is true the stdin stream will remain open across multiple attach sessions. If stdinOnce is set to true, stdin is opened on container start, is empty until the first client attaches to stdin, and then remains open and accepts data until the client disconnects, at which time stdin is closed and remains closed until the container is restarted. If this flag is false, a container processes that reads from stdin will never receive an EOF. Default is false"
+
+### fn spec.template.spec.containers.withTerminationMessagePath
+
+```ts
+withTerminationMessagePath(terminationMessagePath)
+```
+
+"Optional: Path at which the file to which the container's termination message will be written is mounted into the container's filesystem. Message written is intended to be brief final status, such as an assertion failure message. Will be truncated by the node if greater than 4096 bytes. The total message length across all containers will be limited to 12kb. Defaults to /dev/termination-log. Cannot be updated."
+
+### fn spec.template.spec.containers.withTerminationMessagePolicy
+
+```ts
+withTerminationMessagePolicy(terminationMessagePolicy)
+```
+
+"Indicate how the termination message should be populated. File will use the contents of terminationMessagePath to populate the container status message on both success and failure. FallbackToLogsOnError will use the last chunk of container log output if the termination message file is empty and the container exited with an error. The log output is limited to 2048 bytes or 80 lines, whichever is smaller. Defaults to File. Cannot be updated."
+
+### fn spec.template.spec.containers.withTty
+
+```ts
+withTty(tty)
+```
+
+"Whether this container should allocate a TTY for itself, also requires 'stdin' to be true. Default is false."
+
+### fn spec.template.spec.containers.withVolumeDevices
+
+```ts
+withVolumeDevices(volumeDevices)
+```
+
+"volumeDevices is the list of block devices to be used by the container."
+
+### fn spec.template.spec.containers.withVolumeDevicesMixin
+
+```ts
+withVolumeDevicesMixin(volumeDevices)
+```
+
+"volumeDevices is the list of block devices to be used by the container."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.containers.withVolumeMounts
+
+```ts
+withVolumeMounts(volumeMounts)
+```
+
+"Pod volumes to mount into the container's filesystem. Cannot be updated."
+
+### fn spec.template.spec.containers.withVolumeMountsMixin
+
+```ts
+withVolumeMountsMixin(volumeMounts)
+```
+
+"Pod volumes to mount into the container's filesystem. Cannot be updated."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.containers.withWorkingDir
+
+```ts
+withWorkingDir(workingDir)
+```
+
+"Container's working directory. If not specified, the container runtime's default will be used, which might be configured in the container image. Cannot be updated."
+
+## obj spec.template.spec.containers.env
+
+"List of environment variables to set in the container. Cannot be updated."
+
+### fn spec.template.spec.containers.env.withName
+
+```ts
+withName(name)
+```
+
+"Name of the environment variable. Must be a C_IDENTIFIER."
+
+### fn spec.template.spec.containers.env.withValue
+
+```ts
+withValue(value)
+```
+
+"Variable references $(VAR_NAME) are expanded using the previously defined environment variables in the container and any service environment variables. If a variable cannot be resolved, the reference in the input string will be unchanged. Double $$ are reduced to a single $, which allows for escaping the $(VAR_NAME) syntax: i.e. \"$$(VAR_NAME)\" will produce the string literal \"$(VAR_NAME)\". Escaped references will never be expanded, regardless of whether the variable exists or not. Defaults to \"\"."
+
+## obj spec.template.spec.containers.env.valueFrom
+
+"Source for the environment variable's value. Cannot be used if value is not empty."
+
+## obj spec.template.spec.containers.env.valueFrom.configMapKeyRef
+
+"Selects a key of a ConfigMap."
+
+### fn spec.template.spec.containers.env.valueFrom.configMapKeyRef.withKey
+
+```ts
+withKey(key)
+```
+
+"The key to select."
+
+### fn spec.template.spec.containers.env.valueFrom.configMapKeyRef.withName
+
+```ts
+withName(name)
+```
+
+"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"
+
+### fn spec.template.spec.containers.env.valueFrom.configMapKeyRef.withOptional
+
+```ts
+withOptional(optional)
+```
+
+"Specify whether the ConfigMap or its key must be defined"
+
+## obj spec.template.spec.containers.env.valueFrom.fieldRef
+
+"Selects a field of the pod: supports metadata.name, metadata.namespace, `metadata.labels['']`, `metadata.annotations['']`, spec.nodeName, spec.serviceAccountName, status.hostIP, status.podIP, status.podIPs."
+
+### fn spec.template.spec.containers.env.valueFrom.fieldRef.withApiVersion
+
+```ts
+withApiVersion(apiVersion)
+```
+
+"Version of the schema the FieldPath is written in terms of, defaults to \"v1\"."
+
+### fn spec.template.spec.containers.env.valueFrom.fieldRef.withFieldPath
+
+```ts
+withFieldPath(fieldPath)
+```
+
+"Path of the field to select in the specified API version."
+
+## obj spec.template.spec.containers.env.valueFrom.resourceFieldRef
+
+"Selects a resource of the container: only resources limits and requests (limits.cpu, limits.memory, limits.ephemeral-storage, requests.cpu, requests.memory and requests.ephemeral-storage) are currently supported."
+
+### fn spec.template.spec.containers.env.valueFrom.resourceFieldRef.withContainerName
+
+```ts
+withContainerName(containerName)
+```
+
+"Container name: required for volumes, optional for env vars"
+
+### fn spec.template.spec.containers.env.valueFrom.resourceFieldRef.withDivisor
+
+```ts
+withDivisor(divisor)
+```
+
+"Specifies the output format of the exposed resources, defaults to \"1\
+
+### fn spec.template.spec.containers.env.valueFrom.resourceFieldRef.withResource
+
+```ts
+withResource(resource)
+```
+
+"Required: resource to select"
+
+## obj spec.template.spec.containers.env.valueFrom.secretKeyRef
+
+"Selects a key of a secret in the pod's namespace"
+
+### fn spec.template.spec.containers.env.valueFrom.secretKeyRef.withKey
+
+```ts
+withKey(key)
+```
+
+"The key of the secret to select from. Must be a valid secret key."
+
+### fn spec.template.spec.containers.env.valueFrom.secretKeyRef.withName
+
+```ts
+withName(name)
+```
+
+"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"
+
+### fn spec.template.spec.containers.env.valueFrom.secretKeyRef.withOptional
+
+```ts
+withOptional(optional)
+```
+
+"Specify whether the Secret or its key must be defined"
+
+## obj spec.template.spec.containers.envFrom
+
+"List of sources to populate environment variables in the container. The keys defined within a source must be a C_IDENTIFIER. All invalid keys will be reported as an event when the container is starting. When a key exists in multiple sources, the value associated with the last source will take precedence. Values defined by an Env with a duplicate key will take precedence. Cannot be updated."
+
+### fn spec.template.spec.containers.envFrom.withPrefix
+
+```ts
+withPrefix(prefix)
+```
+
+"An optional identifier to prepend to each key in the ConfigMap. Must be a C_IDENTIFIER."
+
+## obj spec.template.spec.containers.envFrom.configMapRef
+
+"The ConfigMap to select from"
+
+### fn spec.template.spec.containers.envFrom.configMapRef.withName
+
+```ts
+withName(name)
+```
+
+"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"
+
+### fn spec.template.spec.containers.envFrom.configMapRef.withOptional
+
+```ts
+withOptional(optional)
+```
+
+"Specify whether the ConfigMap must be defined"
+
+## obj spec.template.spec.containers.envFrom.secretRef
+
+"The Secret to select from"
+
+### fn spec.template.spec.containers.envFrom.secretRef.withName
+
+```ts
+withName(name)
+```
+
+"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"
+
+### fn spec.template.spec.containers.envFrom.secretRef.withOptional
+
+```ts
+withOptional(optional)
+```
+
+"Specify whether the Secret must be defined"
+
+## obj spec.template.spec.containers.lifecycle
+
+"Actions that the management system should take in response to container lifecycle events. Cannot be updated."
+
+## obj spec.template.spec.containers.lifecycle.postStart
+
+"PostStart is called immediately after a container is created. If the handler fails, the container is terminated and restarted according to its restart policy. Other management of the container blocks until the hook completes. More info: https://kubernetes.io/docs/concepts/containers/container-lifecycle-hooks/#container-hooks"
+
+## obj spec.template.spec.containers.lifecycle.postStart.exec
+
+"Exec specifies the action to take."
+
+### fn spec.template.spec.containers.lifecycle.postStart.exec.withCommand
+
+```ts
+withCommand(command)
+```
+
+"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy."
+
+### fn spec.template.spec.containers.lifecycle.postStart.exec.withCommandMixin
+
+```ts
+withCommandMixin(command)
+```
+
+"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy."
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.template.spec.containers.lifecycle.postStart.httpGet
+
+"HTTPGet specifies the http request to perform."
+
+### fn spec.template.spec.containers.lifecycle.postStart.httpGet.withHost
+
+```ts
+withHost(host)
+```
+
+"Host name to connect to, defaults to the pod IP. You probably want to set \"Host\" in httpHeaders instead."
+
+### fn spec.template.spec.containers.lifecycle.postStart.httpGet.withHttpHeaders
+
+```ts
+withHttpHeaders(httpHeaders)
+```
+
+"Custom headers to set in the request. HTTP allows repeated headers."
+
+### fn spec.template.spec.containers.lifecycle.postStart.httpGet.withHttpHeadersMixin
+
+```ts
+withHttpHeadersMixin(httpHeaders)
+```
+
+"Custom headers to set in the request. HTTP allows repeated headers."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.containers.lifecycle.postStart.httpGet.withPath
+
+```ts
+withPath(path)
+```
+
+"Path to access on the HTTP server."
+
+### fn spec.template.spec.containers.lifecycle.postStart.httpGet.withPort
+
+```ts
+withPort(port)
+```
+
+"Name or number of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."
+
+### fn spec.template.spec.containers.lifecycle.postStart.httpGet.withScheme
+
+```ts
+withScheme(scheme)
+```
+
+"Scheme to use for connecting to the host. Defaults to HTTP."
+
+## obj spec.template.spec.containers.lifecycle.postStart.httpGet.httpHeaders
+
+"Custom headers to set in the request. HTTP allows repeated headers."
+
+### fn spec.template.spec.containers.lifecycle.postStart.httpGet.httpHeaders.withName
+
+```ts
+withName(name)
+```
+
+"The header field name. This will be canonicalized upon output, so case-variant names will be understood as the same header."
+
+### fn spec.template.spec.containers.lifecycle.postStart.httpGet.httpHeaders.withValue
+
+```ts
+withValue(value)
+```
+
+"The header field value"
+
+## obj spec.template.spec.containers.lifecycle.postStart.tcpSocket
+
+"Deprecated. TCPSocket is NOT supported as a LifecycleHandler and kept for the backward compatibility. There are no validation of this field and lifecycle hooks will fail in runtime when tcp handler is specified."
+
+### fn spec.template.spec.containers.lifecycle.postStart.tcpSocket.withHost
+
+```ts
+withHost(host)
+```
+
+"Optional: Host name to connect to, defaults to the pod IP."
+
+### fn spec.template.spec.containers.lifecycle.postStart.tcpSocket.withPort
+
+```ts
+withPort(port)
+```
+
+"Number or name of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."
+
+## obj spec.template.spec.containers.lifecycle.preStop
+
+"PreStop is called immediately before a container is terminated due to an API request or management event such as liveness/startup probe failure, preemption, resource contention, etc. The handler is not called if the container crashes or exits. The Pod's termination grace period countdown begins before the PreStop hook is executed. Regardless of the outcome of the handler, the container will eventually terminate within the Pod's termination grace period (unless delayed by finalizers). Other management of the container blocks until the hook completes or until the termination grace period is reached. More info: https://kubernetes.io/docs/concepts/containers/container-lifecycle-hooks/#container-hooks"
+
+## obj spec.template.spec.containers.lifecycle.preStop.exec
+
+"Exec specifies the action to take."
+
+### fn spec.template.spec.containers.lifecycle.preStop.exec.withCommand
+
+```ts
+withCommand(command)
+```
+
+"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy."
+
+### fn spec.template.spec.containers.lifecycle.preStop.exec.withCommandMixin
+
+```ts
+withCommandMixin(command)
+```
+
+"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy."
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.template.spec.containers.lifecycle.preStop.httpGet
+
+"HTTPGet specifies the http request to perform."
+
+### fn spec.template.spec.containers.lifecycle.preStop.httpGet.withHost
+
+```ts
+withHost(host)
+```
+
+"Host name to connect to, defaults to the pod IP. You probably want to set \"Host\" in httpHeaders instead."
+
+### fn spec.template.spec.containers.lifecycle.preStop.httpGet.withHttpHeaders
+
+```ts
+withHttpHeaders(httpHeaders)
+```
+
+"Custom headers to set in the request. HTTP allows repeated headers."
+
+### fn spec.template.spec.containers.lifecycle.preStop.httpGet.withHttpHeadersMixin
+
+```ts
+withHttpHeadersMixin(httpHeaders)
+```
+
+"Custom headers to set in the request. HTTP allows repeated headers."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.containers.lifecycle.preStop.httpGet.withPath
+
+```ts
+withPath(path)
+```
+
+"Path to access on the HTTP server."
+
+### fn spec.template.spec.containers.lifecycle.preStop.httpGet.withPort
+
+```ts
+withPort(port)
+```
+
+"Name or number of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."
+
+### fn spec.template.spec.containers.lifecycle.preStop.httpGet.withScheme
+
+```ts
+withScheme(scheme)
+```
+
+"Scheme to use for connecting to the host. Defaults to HTTP."
+
+## obj spec.template.spec.containers.lifecycle.preStop.httpGet.httpHeaders
+
+"Custom headers to set in the request. HTTP allows repeated headers."
+
+### fn spec.template.spec.containers.lifecycle.preStop.httpGet.httpHeaders.withName
+
+```ts
+withName(name)
+```
+
+"The header field name. This will be canonicalized upon output, so case-variant names will be understood as the same header."
+
+### fn spec.template.spec.containers.lifecycle.preStop.httpGet.httpHeaders.withValue
+
+```ts
+withValue(value)
+```
+
+"The header field value"
+
+## obj spec.template.spec.containers.lifecycle.preStop.tcpSocket
+
+"Deprecated. TCPSocket is NOT supported as a LifecycleHandler and kept for the backward compatibility. There are no validation of this field and lifecycle hooks will fail in runtime when tcp handler is specified."
+
+### fn spec.template.spec.containers.lifecycle.preStop.tcpSocket.withHost
+
+```ts
+withHost(host)
+```
+
+"Optional: Host name to connect to, defaults to the pod IP."
+
+### fn spec.template.spec.containers.lifecycle.preStop.tcpSocket.withPort
+
+```ts
+withPort(port)
+```
+
+"Number or name of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."
+
+## obj spec.template.spec.containers.livenessProbe
+
+"Periodic probe of container liveness. Container will be restarted if the probe fails. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes"
+
+### fn spec.template.spec.containers.livenessProbe.withFailureThreshold
+
+```ts
+withFailureThreshold(failureThreshold)
+```
+
+"Minimum consecutive failures for the probe to be considered failed after having succeeded. Defaults to 3. Minimum value is 1."
+
+### fn spec.template.spec.containers.livenessProbe.withInitialDelaySeconds
+
+```ts
+withInitialDelaySeconds(initialDelaySeconds)
+```
+
+"Number of seconds after the container has started before liveness probes are initiated. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes"
+
+### fn spec.template.spec.containers.livenessProbe.withPeriodSeconds
+
+```ts
+withPeriodSeconds(periodSeconds)
+```
+
+"How often (in seconds) to perform the probe. Default to 10 seconds. Minimum value is 1."
+
+### fn spec.template.spec.containers.livenessProbe.withSuccessThreshold
+
+```ts
+withSuccessThreshold(successThreshold)
+```
+
+"Minimum consecutive successes for the probe to be considered successful after having failed. Defaults to 1. Must be 1 for liveness and startup. Minimum value is 1."
+
+### fn spec.template.spec.containers.livenessProbe.withTerminationGracePeriodSeconds
+
+```ts
+withTerminationGracePeriodSeconds(terminationGracePeriodSeconds)
+```
+
+"Optional duration in seconds the pod needs to terminate gracefully upon probe failure. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. If this value is nil, the pod's terminationGracePeriodSeconds will be used. Otherwise, this value overrides the value provided by the pod spec. Value must be non-negative integer. The value zero indicates stop immediately via the kill signal (no opportunity to shut down). This is a beta field and requires enabling ProbeTerminationGracePeriod feature gate. Minimum value is 1. spec.terminationGracePeriodSeconds is used if unset."
+
+### fn spec.template.spec.containers.livenessProbe.withTimeoutSeconds
+
+```ts
+withTimeoutSeconds(timeoutSeconds)
+```
+
+"Number of seconds after which the probe times out. Defaults to 1 second. Minimum value is 1. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes"
+
+## obj spec.template.spec.containers.livenessProbe.exec
+
+"Exec specifies the action to take."
+
+### fn spec.template.spec.containers.livenessProbe.exec.withCommand
+
+```ts
+withCommand(command)
+```
+
+"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy."
+
+### fn spec.template.spec.containers.livenessProbe.exec.withCommandMixin
+
+```ts
+withCommandMixin(command)
+```
+
+"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy."
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.template.spec.containers.livenessProbe.grpc
+
+"GRPC specifies an action involving a GRPC port."
+
+### fn spec.template.spec.containers.livenessProbe.grpc.withPort
+
+```ts
+withPort(port)
+```
+
+"Port number of the gRPC service. Number must be in the range 1 to 65535."
+
+### fn spec.template.spec.containers.livenessProbe.grpc.withService
+
+```ts
+withService(service)
+```
+
+"Service is the name of the service to place in the gRPC HealthCheckRequest (see https://github.com/grpc/grpc/blob/master/doc/health-checking.md). \n If this is not specified, the default behavior is defined by gRPC."
+
+## obj spec.template.spec.containers.livenessProbe.httpGet
+
+"HTTPGet specifies the http request to perform."
+
+### fn spec.template.spec.containers.livenessProbe.httpGet.withHost
+
+```ts
+withHost(host)
+```
+
+"Host name to connect to, defaults to the pod IP. You probably want to set \"Host\" in httpHeaders instead."
+
+### fn spec.template.spec.containers.livenessProbe.httpGet.withHttpHeaders
+
+```ts
+withHttpHeaders(httpHeaders)
+```
+
+"Custom headers to set in the request. HTTP allows repeated headers."
+
+### fn spec.template.spec.containers.livenessProbe.httpGet.withHttpHeadersMixin
+
+```ts
+withHttpHeadersMixin(httpHeaders)
+```
+
+"Custom headers to set in the request. HTTP allows repeated headers."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.containers.livenessProbe.httpGet.withPath
+
+```ts
+withPath(path)
+```
+
+"Path to access on the HTTP server."
+
+### fn spec.template.spec.containers.livenessProbe.httpGet.withPort
+
+```ts
+withPort(port)
+```
+
+"Name or number of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."
+
+### fn spec.template.spec.containers.livenessProbe.httpGet.withScheme
+
+```ts
+withScheme(scheme)
+```
+
+"Scheme to use for connecting to the host. Defaults to HTTP."
+
+## obj spec.template.spec.containers.livenessProbe.httpGet.httpHeaders
+
+"Custom headers to set in the request. HTTP allows repeated headers."
+
+### fn spec.template.spec.containers.livenessProbe.httpGet.httpHeaders.withName
+
+```ts
+withName(name)
+```
+
+"The header field name. This will be canonicalized upon output, so case-variant names will be understood as the same header."
+
+### fn spec.template.spec.containers.livenessProbe.httpGet.httpHeaders.withValue
+
+```ts
+withValue(value)
+```
+
+"The header field value"
+
+## obj spec.template.spec.containers.livenessProbe.tcpSocket
+
+"TCPSocket specifies an action involving a TCP port."
+
+### fn spec.template.spec.containers.livenessProbe.tcpSocket.withHost
+
+```ts
+withHost(host)
+```
+
+"Optional: Host name to connect to, defaults to the pod IP."
+
+### fn spec.template.spec.containers.livenessProbe.tcpSocket.withPort
+
+```ts
+withPort(port)
+```
+
+"Number or name of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."
+
+## obj spec.template.spec.containers.ports
+
+"List of ports to expose from the container. Not specifying a port here DOES NOT prevent that port from being exposed. Any port which is listening on the default \"0.0.0.0\" address inside a container will be accessible from the network. Modifying this array with strategic merge patch may corrupt the data. For more information See https://github.com/kubernetes/kubernetes/issues/108255. Cannot be updated."
+
+### fn spec.template.spec.containers.ports.withContainerPort
+
+```ts
+withContainerPort(containerPort)
+```
+
+"Number of port to expose on the pod's IP address. This must be a valid port number, 0 < x < 65536."
+
+### fn spec.template.spec.containers.ports.withHostIP
+
+```ts
+withHostIP(hostIP)
+```
+
+"What host IP to bind the external port to."
+
+### fn spec.template.spec.containers.ports.withHostPort
+
+```ts
+withHostPort(hostPort)
+```
+
+"Number of port to expose on the host. If specified, this must be a valid port number, 0 < x < 65536. If HostNetwork is specified, this must match ContainerPort. Most containers do not need this."
+
+### fn spec.template.spec.containers.ports.withName
+
+```ts
+withName(name)
+```
+
+"If specified, this must be an IANA_SVC_NAME and unique within the pod. Each named port in a pod must have a unique name. Name for the port that can be referred to by services."
+
+### fn spec.template.spec.containers.ports.withProtocol
+
+```ts
+withProtocol(protocol)
+```
+
+"Protocol for port. Must be UDP, TCP, or SCTP. Defaults to \"TCP\"."
+
+## obj spec.template.spec.containers.readinessProbe
+
+"Periodic probe of container service readiness. Container will be removed from service endpoints if the probe fails. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes"
+
+### fn spec.template.spec.containers.readinessProbe.withFailureThreshold
+
+```ts
+withFailureThreshold(failureThreshold)
+```
+
+"Minimum consecutive failures for the probe to be considered failed after having succeeded. Defaults to 3. Minimum value is 1."
+
+### fn spec.template.spec.containers.readinessProbe.withInitialDelaySeconds
+
+```ts
+withInitialDelaySeconds(initialDelaySeconds)
+```
+
+"Number of seconds after the container has started before liveness probes are initiated. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes"
+
+### fn spec.template.spec.containers.readinessProbe.withPeriodSeconds
+
+```ts
+withPeriodSeconds(periodSeconds)
+```
+
+"How often (in seconds) to perform the probe. Default to 10 seconds. Minimum value is 1."
+
+### fn spec.template.spec.containers.readinessProbe.withSuccessThreshold
+
+```ts
+withSuccessThreshold(successThreshold)
+```
+
+"Minimum consecutive successes for the probe to be considered successful after having failed. Defaults to 1. Must be 1 for liveness and startup. Minimum value is 1."
+
+### fn spec.template.spec.containers.readinessProbe.withTerminationGracePeriodSeconds
+
+```ts
+withTerminationGracePeriodSeconds(terminationGracePeriodSeconds)
+```
+
+"Optional duration in seconds the pod needs to terminate gracefully upon probe failure. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. If this value is nil, the pod's terminationGracePeriodSeconds will be used. Otherwise, this value overrides the value provided by the pod spec. Value must be non-negative integer. The value zero indicates stop immediately via the kill signal (no opportunity to shut down). This is a beta field and requires enabling ProbeTerminationGracePeriod feature gate. Minimum value is 1. spec.terminationGracePeriodSeconds is used if unset."
+
+### fn spec.template.spec.containers.readinessProbe.withTimeoutSeconds
+
+```ts
+withTimeoutSeconds(timeoutSeconds)
+```
+
+"Number of seconds after which the probe times out. Defaults to 1 second. Minimum value is 1. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes"
+
+## obj spec.template.spec.containers.readinessProbe.exec
+
+"Exec specifies the action to take."
+
+### fn spec.template.spec.containers.readinessProbe.exec.withCommand
+
+```ts
+withCommand(command)
+```
+
+"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy."
+
+### fn spec.template.spec.containers.readinessProbe.exec.withCommandMixin
+
+```ts
+withCommandMixin(command)
+```
+
+"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy."
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.template.spec.containers.readinessProbe.grpc
+
+"GRPC specifies an action involving a GRPC port."
+
+### fn spec.template.spec.containers.readinessProbe.grpc.withPort
+
+```ts
+withPort(port)
+```
+
+"Port number of the gRPC service. Number must be in the range 1 to 65535."
+
+### fn spec.template.spec.containers.readinessProbe.grpc.withService
+
+```ts
+withService(service)
+```
+
+"Service is the name of the service to place in the gRPC HealthCheckRequest (see https://github.com/grpc/grpc/blob/master/doc/health-checking.md). \n If this is not specified, the default behavior is defined by gRPC."
+
+## obj spec.template.spec.containers.readinessProbe.httpGet
+
+"HTTPGet specifies the http request to perform."
+
+### fn spec.template.spec.containers.readinessProbe.httpGet.withHost
+
+```ts
+withHost(host)
+```
+
+"Host name to connect to, defaults to the pod IP. You probably want to set \"Host\" in httpHeaders instead."
+
+### fn spec.template.spec.containers.readinessProbe.httpGet.withHttpHeaders
+
+```ts
+withHttpHeaders(httpHeaders)
+```
+
+"Custom headers to set in the request. HTTP allows repeated headers."
+
+### fn spec.template.spec.containers.readinessProbe.httpGet.withHttpHeadersMixin
+
+```ts
+withHttpHeadersMixin(httpHeaders)
+```
+
+"Custom headers to set in the request. HTTP allows repeated headers."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.containers.readinessProbe.httpGet.withPath
+
+```ts
+withPath(path)
+```
+
+"Path to access on the HTTP server."
+
+### fn spec.template.spec.containers.readinessProbe.httpGet.withPort
+
+```ts
+withPort(port)
+```
+
+"Name or number of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."
+
+### fn spec.template.spec.containers.readinessProbe.httpGet.withScheme
+
+```ts
+withScheme(scheme)
+```
+
+"Scheme to use for connecting to the host. Defaults to HTTP."
+
+## obj spec.template.spec.containers.readinessProbe.httpGet.httpHeaders
+
+"Custom headers to set in the request. HTTP allows repeated headers."
+
+### fn spec.template.spec.containers.readinessProbe.httpGet.httpHeaders.withName
+
+```ts
+withName(name)
+```
+
+"The header field name. This will be canonicalized upon output, so case-variant names will be understood as the same header."
+
+### fn spec.template.spec.containers.readinessProbe.httpGet.httpHeaders.withValue
+
+```ts
+withValue(value)
+```
+
+"The header field value"
+
+## obj spec.template.spec.containers.readinessProbe.tcpSocket
+
+"TCPSocket specifies an action involving a TCP port."
+
+### fn spec.template.spec.containers.readinessProbe.tcpSocket.withHost
+
+```ts
+withHost(host)
+```
+
+"Optional: Host name to connect to, defaults to the pod IP."
+
+### fn spec.template.spec.containers.readinessProbe.tcpSocket.withPort
+
+```ts
+withPort(port)
+```
+
+"Number or name of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."
+
+## obj spec.template.spec.containers.resizePolicy
+
+"Resources resize policy for the container."
+
+### fn spec.template.spec.containers.resizePolicy.withResourceName
+
+```ts
+withResourceName(resourceName)
+```
+
+"Name of the resource to which this resource resize policy applies. Supported values: cpu, memory."
+
+### fn spec.template.spec.containers.resizePolicy.withRestartPolicy
+
+```ts
+withRestartPolicy(restartPolicy)
+```
+
+"Restart policy to apply when specified resource is resized. If not specified, it defaults to NotRequired."
+
+## obj spec.template.spec.containers.resources
+
+"Compute Resources required by this container. Cannot be updated. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"
+
+### fn spec.template.spec.containers.resources.withClaims
+
+```ts
+withClaims(claims)
+```
+
+"Claims lists the names of resources, defined in spec.resourceClaims, that are used by this container. \n This is an alpha field and requires enabling the DynamicResourceAllocation feature gate. \n This field is immutable. It can only be set for containers."
+
+### fn spec.template.spec.containers.resources.withClaimsMixin
+
+```ts
+withClaimsMixin(claims)
+```
+
+"Claims lists the names of resources, defined in spec.resourceClaims, that are used by this container. \n This is an alpha field and requires enabling the DynamicResourceAllocation feature gate. \n This field is immutable. It can only be set for containers."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.containers.resources.withLimits
+
+```ts
+withLimits(limits)
+```
+
+"Limits describes the maximum amount of compute resources allowed. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"
+
+### fn spec.template.spec.containers.resources.withLimitsMixin
+
+```ts
+withLimitsMixin(limits)
+```
+
+"Limits describes the maximum amount of compute resources allowed. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.containers.resources.withRequests
+
+```ts
+withRequests(requests)
+```
+
+"Requests describes the minimum amount of compute resources required. If Requests is omitted for a container, it defaults to Limits if that is explicitly specified, otherwise to an implementation-defined value. Requests cannot exceed Limits. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"
+
+### fn spec.template.spec.containers.resources.withRequestsMixin
+
+```ts
+withRequestsMixin(requests)
+```
+
+"Requests describes the minimum amount of compute resources required. If Requests is omitted for a container, it defaults to Limits if that is explicitly specified, otherwise to an implementation-defined value. Requests cannot exceed Limits. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.template.spec.containers.resources.claims
+
+"Claims lists the names of resources, defined in spec.resourceClaims, that are used by this container. \n This is an alpha field and requires enabling the DynamicResourceAllocation feature gate. \n This field is immutable. It can only be set for containers."
+
+### fn spec.template.spec.containers.resources.claims.withName
+
+```ts
+withName(name)
+```
+
+"Name must match the name of one entry in pod.spec.resourceClaims of the Pod where this field is used. It makes that resource available inside a container."
+
+## obj spec.template.spec.containers.securityContext
+
+"SecurityContext defines the security options the container should be run with. If set, the fields of SecurityContext override the equivalent fields of PodSecurityContext. More info: https://kubernetes.io/docs/tasks/configure-pod-container/security-context/"
+
+### fn spec.template.spec.containers.securityContext.withAllowPrivilegeEscalation
+
+```ts
+withAllowPrivilegeEscalation(allowPrivilegeEscalation)
+```
+
+"AllowPrivilegeEscalation controls whether a process can gain more privileges than its parent process. This bool directly controls if the no_new_privs flag will be set on the container process. AllowPrivilegeEscalation is true always when the container is: 1) run as Privileged 2) has CAP_SYS_ADMIN Note that this field cannot be set when spec.os.name is windows."
+
+### fn spec.template.spec.containers.securityContext.withPrivileged
+
+```ts
+withPrivileged(privileged)
+```
+
+"Run container in privileged mode. Processes in privileged containers are essentially equivalent to root on the host. Defaults to false. Note that this field cannot be set when spec.os.name is windows."
+
+### fn spec.template.spec.containers.securityContext.withProcMount
+
+```ts
+withProcMount(procMount)
+```
+
+"procMount denotes the type of proc mount to use for the containers. The default is DefaultProcMount which uses the container runtime defaults for readonly paths and masked paths. This requires the ProcMountType feature flag to be enabled. Note that this field cannot be set when spec.os.name is windows."
+
+### fn spec.template.spec.containers.securityContext.withReadOnlyRootFilesystem
+
+```ts
+withReadOnlyRootFilesystem(readOnlyRootFilesystem)
+```
+
+"Whether this container has a read-only root filesystem. Default is false. Note that this field cannot be set when spec.os.name is windows."
+
+### fn spec.template.spec.containers.securityContext.withRunAsGroup
+
+```ts
+withRunAsGroup(runAsGroup)
+```
+
+"The GID to run the entrypoint of the container process. Uses runtime default if unset. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. Note that this field cannot be set when spec.os.name is windows."
+
+### fn spec.template.spec.containers.securityContext.withRunAsNonRoot
+
+```ts
+withRunAsNonRoot(runAsNonRoot)
+```
+
+"Indicates that the container must run as a non-root user. If true, the Kubelet will validate the image at runtime to ensure that it does not run as UID 0 (root) and fail to start the container if it does. If unset or false, no such validation will be performed. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence."
+
+### fn spec.template.spec.containers.securityContext.withRunAsUser
+
+```ts
+withRunAsUser(runAsUser)
+```
+
+"The UID to run the entrypoint of the container process. Defaults to user specified in image metadata if unspecified. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. Note that this field cannot be set when spec.os.name is windows."
+
+## obj spec.template.spec.containers.securityContext.capabilities
+
+"The capabilities to add/drop when running containers. Defaults to the default set of capabilities granted by the container runtime. Note that this field cannot be set when spec.os.name is windows."
+
+### fn spec.template.spec.containers.securityContext.capabilities.withAdd
+
+```ts
+withAdd(add)
+```
+
+"Added capabilities"
+
+### fn spec.template.spec.containers.securityContext.capabilities.withAddMixin
+
+```ts
+withAddMixin(add)
+```
+
+"Added capabilities"
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.containers.securityContext.capabilities.withDrop
+
+```ts
+withDrop(drop)
+```
+
+"Removed capabilities"
+
+### fn spec.template.spec.containers.securityContext.capabilities.withDropMixin
+
+```ts
+withDropMixin(drop)
+```
+
+"Removed capabilities"
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.template.spec.containers.securityContext.seLinuxOptions
+
+"The SELinux context to be applied to the container. If unspecified, the container runtime will allocate a random SELinux context for each container. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. Note that this field cannot be set when spec.os.name is windows."
+
+### fn spec.template.spec.containers.securityContext.seLinuxOptions.withLevel
+
+```ts
+withLevel(level)
+```
+
+"Level is SELinux level label that applies to the container."
+
+### fn spec.template.spec.containers.securityContext.seLinuxOptions.withRole
+
+```ts
+withRole(role)
+```
+
+"Role is a SELinux role label that applies to the container."
+
+### fn spec.template.spec.containers.securityContext.seLinuxOptions.withType
+
+```ts
+withType(type)
+```
+
+"Type is a SELinux type label that applies to the container."
+
+### fn spec.template.spec.containers.securityContext.seLinuxOptions.withUser
+
+```ts
+withUser(user)
+```
+
+"User is a SELinux user label that applies to the container."
+
+## obj spec.template.spec.containers.securityContext.seccompProfile
+
+"The seccomp options to use by this container. If seccomp options are provided at both the pod & container level, the container options override the pod options. Note that this field cannot be set when spec.os.name is windows."
+
+### fn spec.template.spec.containers.securityContext.seccompProfile.withLocalhostProfile
+
+```ts
+withLocalhostProfile(localhostProfile)
+```
+
+"localhostProfile indicates a profile defined in a file on the node should be used. The profile must be preconfigured on the node to work. Must be a descending path, relative to the kubelet's configured seccomp profile location. Must be set if type is \"Localhost\". Must NOT be set for any other type."
+
+### fn spec.template.spec.containers.securityContext.seccompProfile.withType
+
+```ts
+withType(type)
+```
+
+"type indicates which kind of seccomp profile will be applied. Valid options are: \n Localhost - a profile defined in a file on the node should be used. RuntimeDefault - the container runtime default profile should be used. Unconfined - no profile should be applied."
+
+## obj spec.template.spec.containers.securityContext.windowsOptions
+
+"The Windows specific settings applied to all containers. If unspecified, the options from the PodSecurityContext will be used. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. Note that this field cannot be set when spec.os.name is linux."
+
+### fn spec.template.spec.containers.securityContext.windowsOptions.withGmsaCredentialSpec
+
+```ts
+withGmsaCredentialSpec(gmsaCredentialSpec)
+```
+
+"GMSACredentialSpec is where the GMSA admission webhook (https://github.com/kubernetes-sigs/windows-gmsa) inlines the contents of the GMSA credential spec named by the GMSACredentialSpecName field."
+
+### fn spec.template.spec.containers.securityContext.windowsOptions.withGmsaCredentialSpecName
+
+```ts
+withGmsaCredentialSpecName(gmsaCredentialSpecName)
+```
+
+"GMSACredentialSpecName is the name of the GMSA credential spec to use."
+
+### fn spec.template.spec.containers.securityContext.windowsOptions.withHostProcess
+
+```ts
+withHostProcess(hostProcess)
+```
+
+"HostProcess determines if a container should be run as a 'Host Process' container. All of a Pod's containers must have the same effective HostProcess value (it is not allowed to have a mix of HostProcess containers and non-HostProcess containers). In addition, if HostProcess is true then HostNetwork must also be set to true."
+
+### fn spec.template.spec.containers.securityContext.windowsOptions.withRunAsUserName
+
+```ts
+withRunAsUserName(runAsUserName)
+```
+
+"The UserName in Windows to run the entrypoint of the container process. Defaults to the user specified in image metadata if unspecified. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence."
+
+## obj spec.template.spec.containers.startupProbe
+
+"StartupProbe indicates that the Pod has successfully initialized. If specified, no other probes are executed until this completes successfully. If this probe fails, the Pod will be restarted, just as if the livenessProbe failed. This can be used to provide different probe parameters at the beginning of a Pod's lifecycle, when it might take a long time to load data or warm a cache, than during steady-state operation. This cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes"
+
+### fn spec.template.spec.containers.startupProbe.withFailureThreshold
+
+```ts
+withFailureThreshold(failureThreshold)
+```
+
+"Minimum consecutive failures for the probe to be considered failed after having succeeded. Defaults to 3. Minimum value is 1."
+
+### fn spec.template.spec.containers.startupProbe.withInitialDelaySeconds
+
+```ts
+withInitialDelaySeconds(initialDelaySeconds)
+```
+
+"Number of seconds after the container has started before liveness probes are initiated. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes"
+
+### fn spec.template.spec.containers.startupProbe.withPeriodSeconds
+
+```ts
+withPeriodSeconds(periodSeconds)
+```
+
+"How often (in seconds) to perform the probe. Default to 10 seconds. Minimum value is 1."
+
+### fn spec.template.spec.containers.startupProbe.withSuccessThreshold
+
+```ts
+withSuccessThreshold(successThreshold)
+```
+
+"Minimum consecutive successes for the probe to be considered successful after having failed. Defaults to 1. Must be 1 for liveness and startup. Minimum value is 1."
+
+### fn spec.template.spec.containers.startupProbe.withTerminationGracePeriodSeconds
+
+```ts
+withTerminationGracePeriodSeconds(terminationGracePeriodSeconds)
+```
+
+"Optional duration in seconds the pod needs to terminate gracefully upon probe failure. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. If this value is nil, the pod's terminationGracePeriodSeconds will be used. Otherwise, this value overrides the value provided by the pod spec. Value must be non-negative integer. The value zero indicates stop immediately via the kill signal (no opportunity to shut down). This is a beta field and requires enabling ProbeTerminationGracePeriod feature gate. Minimum value is 1. spec.terminationGracePeriodSeconds is used if unset."
+
+### fn spec.template.spec.containers.startupProbe.withTimeoutSeconds
+
+```ts
+withTimeoutSeconds(timeoutSeconds)
+```
+
+"Number of seconds after which the probe times out. Defaults to 1 second. Minimum value is 1. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes"
+
+## obj spec.template.spec.containers.startupProbe.exec
+
+"Exec specifies the action to take."
+
+### fn spec.template.spec.containers.startupProbe.exec.withCommand
+
+```ts
+withCommand(command)
+```
+
+"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy."
+
+### fn spec.template.spec.containers.startupProbe.exec.withCommandMixin
+
+```ts
+withCommandMixin(command)
+```
+
+"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy."
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.template.spec.containers.startupProbe.grpc
+
+"GRPC specifies an action involving a GRPC port."
+
+### fn spec.template.spec.containers.startupProbe.grpc.withPort
+
+```ts
+withPort(port)
+```
+
+"Port number of the gRPC service. Number must be in the range 1 to 65535."
+
+### fn spec.template.spec.containers.startupProbe.grpc.withService
+
+```ts
+withService(service)
+```
+
+"Service is the name of the service to place in the gRPC HealthCheckRequest (see https://github.com/grpc/grpc/blob/master/doc/health-checking.md). \n If this is not specified, the default behavior is defined by gRPC."
+
+## obj spec.template.spec.containers.startupProbe.httpGet
+
+"HTTPGet specifies the http request to perform."
+
+### fn spec.template.spec.containers.startupProbe.httpGet.withHost
+
+```ts
+withHost(host)
+```
+
+"Host name to connect to, defaults to the pod IP. You probably want to set \"Host\" in httpHeaders instead."
+
+### fn spec.template.spec.containers.startupProbe.httpGet.withHttpHeaders
+
+```ts
+withHttpHeaders(httpHeaders)
+```
+
+"Custom headers to set in the request. HTTP allows repeated headers."
+
+### fn spec.template.spec.containers.startupProbe.httpGet.withHttpHeadersMixin
+
+```ts
+withHttpHeadersMixin(httpHeaders)
+```
+
+"Custom headers to set in the request. HTTP allows repeated headers."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.containers.startupProbe.httpGet.withPath
+
+```ts
+withPath(path)
+```
+
+"Path to access on the HTTP server."
+
+### fn spec.template.spec.containers.startupProbe.httpGet.withPort
+
+```ts
+withPort(port)
+```
+
+"Name or number of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."
+
+### fn spec.template.spec.containers.startupProbe.httpGet.withScheme
+
+```ts
+withScheme(scheme)
+```
+
+"Scheme to use for connecting to the host. Defaults to HTTP."
+
+## obj spec.template.spec.containers.startupProbe.httpGet.httpHeaders
+
+"Custom headers to set in the request. HTTP allows repeated headers."
+
+### fn spec.template.spec.containers.startupProbe.httpGet.httpHeaders.withName
+
+```ts
+withName(name)
+```
+
+"The header field name. This will be canonicalized upon output, so case-variant names will be understood as the same header."
+
+### fn spec.template.spec.containers.startupProbe.httpGet.httpHeaders.withValue
+
+```ts
+withValue(value)
+```
+
+"The header field value"
+
+## obj spec.template.spec.containers.startupProbe.tcpSocket
+
+"TCPSocket specifies an action involving a TCP port."
+
+### fn spec.template.spec.containers.startupProbe.tcpSocket.withHost
+
+```ts
+withHost(host)
+```
+
+"Optional: Host name to connect to, defaults to the pod IP."
+
+### fn spec.template.spec.containers.startupProbe.tcpSocket.withPort
+
+```ts
+withPort(port)
+```
+
+"Number or name of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."
+
+## obj spec.template.spec.containers.volumeDevices
+
+"volumeDevices is the list of block devices to be used by the container."
+
+### fn spec.template.spec.containers.volumeDevices.withDevicePath
+
+```ts
+withDevicePath(devicePath)
+```
+
+"devicePath is the path inside of the container that the device will be mapped to."
+
+### fn spec.template.spec.containers.volumeDevices.withName
+
+```ts
+withName(name)
+```
+
+"name must match the name of a persistentVolumeClaim in the pod"
+
+## obj spec.template.spec.containers.volumeMounts
+
+"Pod volumes to mount into the container's filesystem. Cannot be updated."
+
+### fn spec.template.spec.containers.volumeMounts.withMountPath
+
+```ts
+withMountPath(mountPath)
+```
+
+"Path within the container at which the volume should be mounted. Must not contain ':'."
+
+### fn spec.template.spec.containers.volumeMounts.withMountPropagation
+
+```ts
+withMountPropagation(mountPropagation)
+```
+
+"mountPropagation determines how mounts are propagated from the host to container and the other way around. When not set, MountPropagationNone is used. This field is beta in 1.10."
+
+### fn spec.template.spec.containers.volumeMounts.withName
+
+```ts
+withName(name)
+```
+
+"This must match the Name of a Volume."
+
+### fn spec.template.spec.containers.volumeMounts.withReadOnly
+
+```ts
+withReadOnly(readOnly)
+```
+
+"Mounted read-only if true, read-write otherwise (false or unspecified). Defaults to false."
+
+### fn spec.template.spec.containers.volumeMounts.withSubPath
+
+```ts
+withSubPath(subPath)
+```
+
+"Path within the volume from which the container's volume should be mounted. Defaults to \"\" (volume's root)."
+
+### fn spec.template.spec.containers.volumeMounts.withSubPathExpr
+
+```ts
+withSubPathExpr(subPathExpr)
+```
+
+"Expanded path within the volume from which the container's volume should be mounted. Behaves similarly to SubPath but environment variable references $(VAR_NAME) are expanded using the container's environment. Defaults to \"\" (volume's root). SubPathExpr and SubPath are mutually exclusive."
+
+## obj spec.template.spec.dnsConfig
+
+"Specifies the DNS parameters of a pod. Parameters specified here will be merged to the generated DNS configuration based on DNSPolicy."
+
+### fn spec.template.spec.dnsConfig.withNameservers
+
+```ts
+withNameservers(nameservers)
+```
+
+"A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed."
+
+### fn spec.template.spec.dnsConfig.withNameserversMixin
+
+```ts
+withNameserversMixin(nameservers)
+```
+
+"A list of DNS name server IP addresses. This will be appended to the base nameservers generated from DNSPolicy. Duplicated nameservers will be removed."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.dnsConfig.withOptions
+
+```ts
+withOptions(options)
+```
+
+"A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy."
+
+### fn spec.template.spec.dnsConfig.withOptionsMixin
+
+```ts
+withOptionsMixin(options)
+```
+
+"A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.dnsConfig.withSearches
+
+```ts
+withSearches(searches)
+```
+
+"A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed."
+
+### fn spec.template.spec.dnsConfig.withSearchesMixin
+
+```ts
+withSearchesMixin(searches)
+```
+
+"A list of DNS search domains for host-name lookup. This will be appended to the base search paths generated from DNSPolicy. Duplicated search paths will be removed."
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.template.spec.dnsConfig.options
+
+"A list of DNS resolver options. This will be merged with the base options generated from DNSPolicy. Duplicated entries will be removed. Resolution options given in Options will override those that appear in the base DNSPolicy."
+
+### fn spec.template.spec.dnsConfig.options.withName
+
+```ts
+withName(name)
+```
+
+"Required."
+
+### fn spec.template.spec.dnsConfig.options.withValue
+
+```ts
+withValue(value)
+```
+
+
+
+## obj spec.template.spec.ephemeralContainers
+
+"List of ephemeral containers run in this pod. Ephemeral containers may be run in an existing pod to perform user-initiated actions such as debugging. This list cannot be specified when creating a pod, and it cannot be modified by updating the pod spec. In order to add an ephemeral container to an existing pod, use the pod's ephemeralcontainers subresource."
+
+### fn spec.template.spec.ephemeralContainers.withArgs
+
+```ts
+withArgs(args)
+```
+
+"Arguments to the entrypoint. The image's CMD is used if this is not provided. Variable references $(VAR_NAME) are expanded using the container's environment. If a variable cannot be resolved, the reference in the input string will be unchanged. Double $$ are reduced to a single $, which allows for escaping the $(VAR_NAME) syntax: i.e. \"$$(VAR_NAME)\" will produce the string literal \"$(VAR_NAME)\". Escaped references will never be expanded, regardless of whether the variable exists or not. Cannot be updated. More info: https://kubernetes.io/docs/tasks/inject-data-application/define-command-argument-container/#running-a-command-in-a-shell"
+
+### fn spec.template.spec.ephemeralContainers.withArgsMixin
+
+```ts
+withArgsMixin(args)
+```
+
+"Arguments to the entrypoint. The image's CMD is used if this is not provided. Variable references $(VAR_NAME) are expanded using the container's environment. If a variable cannot be resolved, the reference in the input string will be unchanged. Double $$ are reduced to a single $, which allows for escaping the $(VAR_NAME) syntax: i.e. \"$$(VAR_NAME)\" will produce the string literal \"$(VAR_NAME)\". Escaped references will never be expanded, regardless of whether the variable exists or not. Cannot be updated. More info: https://kubernetes.io/docs/tasks/inject-data-application/define-command-argument-container/#running-a-command-in-a-shell"
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.ephemeralContainers.withCommand
+
+```ts
+withCommand(command)
+```
+
+"Entrypoint array. Not executed within a shell. The image's ENTRYPOINT is used if this is not provided. Variable references $(VAR_NAME) are expanded using the container's environment. If a variable cannot be resolved, the reference in the input string will be unchanged. Double $$ are reduced to a single $, which allows for escaping the $(VAR_NAME) syntax: i.e. \"$$(VAR_NAME)\" will produce the string literal \"$(VAR_NAME)\". Escaped references will never be expanded, regardless of whether the variable exists or not. Cannot be updated. More info: https://kubernetes.io/docs/tasks/inject-data-application/define-command-argument-container/#running-a-command-in-a-shell"
+
+### fn spec.template.spec.ephemeralContainers.withCommandMixin
+
+```ts
+withCommandMixin(command)
+```
+
+"Entrypoint array. Not executed within a shell. The image's ENTRYPOINT is used if this is not provided. Variable references $(VAR_NAME) are expanded using the container's environment. If a variable cannot be resolved, the reference in the input string will be unchanged. Double $$ are reduced to a single $, which allows for escaping the $(VAR_NAME) syntax: i.e. \"$$(VAR_NAME)\" will produce the string literal \"$(VAR_NAME)\". Escaped references will never be expanded, regardless of whether the variable exists or not. Cannot be updated. More info: https://kubernetes.io/docs/tasks/inject-data-application/define-command-argument-container/#running-a-command-in-a-shell"
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.ephemeralContainers.withEnv
+
+```ts
+withEnv(env)
+```
+
+"List of environment variables to set in the container. Cannot be updated."
+
+### fn spec.template.spec.ephemeralContainers.withEnvFrom
+
+```ts
+withEnvFrom(envFrom)
+```
+
+"List of sources to populate environment variables in the container. The keys defined within a source must be a C_IDENTIFIER. All invalid keys will be reported as an event when the container is starting. When a key exists in multiple sources, the value associated with the last source will take precedence. Values defined by an Env with a duplicate key will take precedence. Cannot be updated."
+
+### fn spec.template.spec.ephemeralContainers.withEnvFromMixin
+
+```ts
+withEnvFromMixin(envFrom)
+```
+
+"List of sources to populate environment variables in the container. The keys defined within a source must be a C_IDENTIFIER. All invalid keys will be reported as an event when the container is starting. When a key exists in multiple sources, the value associated with the last source will take precedence. Values defined by an Env with a duplicate key will take precedence. Cannot be updated."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.ephemeralContainers.withEnvMixin
+
+```ts
+withEnvMixin(env)
+```
+
+"List of environment variables to set in the container. Cannot be updated."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.ephemeralContainers.withImage
+
+```ts
+withImage(image)
+```
+
+"Container image name. More info: https://kubernetes.io/docs/concepts/containers/images"
+
+### fn spec.template.spec.ephemeralContainers.withImagePullPolicy
+
+```ts
+withImagePullPolicy(imagePullPolicy)
+```
+
+"Image pull policy. One of Always, Never, IfNotPresent. Defaults to Always if :latest tag is specified, or IfNotPresent otherwise. Cannot be updated. More info: https://kubernetes.io/docs/concepts/containers/images#updating-images"
+
+### fn spec.template.spec.ephemeralContainers.withName
+
+```ts
+withName(name)
+```
+
+"Name of the ephemeral container specified as a DNS_LABEL. This name must be unique among all containers, init containers and ephemeral containers."
+
+### fn spec.template.spec.ephemeralContainers.withPorts
+
+```ts
+withPorts(ports)
+```
+
+"Ports are not allowed for ephemeral containers."
+
+### fn spec.template.spec.ephemeralContainers.withPortsMixin
+
+```ts
+withPortsMixin(ports)
+```
+
+"Ports are not allowed for ephemeral containers."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.ephemeralContainers.withResizePolicy
+
+```ts
+withResizePolicy(resizePolicy)
+```
+
+"Resources resize policy for the container."
+
+### fn spec.template.spec.ephemeralContainers.withResizePolicyMixin
+
+```ts
+withResizePolicyMixin(resizePolicy)
+```
+
+"Resources resize policy for the container."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.ephemeralContainers.withRestartPolicy
+
+```ts
+withRestartPolicy(restartPolicy)
+```
+
+"Restart policy for the container to manage the restart behavior of each container within a pod. This may only be set for init containers. You cannot set this field on ephemeral containers."
+
+### fn spec.template.spec.ephemeralContainers.withStdin
+
+```ts
+withStdin(stdin)
+```
+
+"Whether this container should allocate a buffer for stdin in the container runtime. If this is not set, reads from stdin in the container will always result in EOF. Default is false."
+
+### fn spec.template.spec.ephemeralContainers.withStdinOnce
+
+```ts
+withStdinOnce(stdinOnce)
+```
+
+"Whether the container runtime should close the stdin channel after it has been opened by a single attach. When stdin is true the stdin stream will remain open across multiple attach sessions. If stdinOnce is set to true, stdin is opened on container start, is empty until the first client attaches to stdin, and then remains open and accepts data until the client disconnects, at which time stdin is closed and remains closed until the container is restarted. If this flag is false, a container processes that reads from stdin will never receive an EOF. Default is false"
+
+### fn spec.template.spec.ephemeralContainers.withTargetContainerName
+
+```ts
+withTargetContainerName(targetContainerName)
+```
+
+"If set, the name of the container from PodSpec that this ephemeral container targets. The ephemeral container will be run in the namespaces (IPC, PID, etc) of this container. If not set then the ephemeral container uses the namespaces configured in the Pod spec. \n The container runtime must implement support for this feature. If the runtime does not support namespace targeting then the result of setting this field is undefined."
+
+### fn spec.template.spec.ephemeralContainers.withTerminationMessagePath
+
+```ts
+withTerminationMessagePath(terminationMessagePath)
+```
+
+"Optional: Path at which the file to which the container's termination message will be written is mounted into the container's filesystem. Message written is intended to be brief final status, such as an assertion failure message. Will be truncated by the node if greater than 4096 bytes. The total message length across all containers will be limited to 12kb. Defaults to /dev/termination-log. Cannot be updated."
+
+### fn spec.template.spec.ephemeralContainers.withTerminationMessagePolicy
+
+```ts
+withTerminationMessagePolicy(terminationMessagePolicy)
+```
+
+"Indicate how the termination message should be populated. File will use the contents of terminationMessagePath to populate the container status message on both success and failure. FallbackToLogsOnError will use the last chunk of container log output if the termination message file is empty and the container exited with an error. The log output is limited to 2048 bytes or 80 lines, whichever is smaller. Defaults to File. Cannot be updated."
+
+### fn spec.template.spec.ephemeralContainers.withTty
+
+```ts
+withTty(tty)
+```
+
+"Whether this container should allocate a TTY for itself, also requires 'stdin' to be true. Default is false."
+
+### fn spec.template.spec.ephemeralContainers.withVolumeDevices
+
+```ts
+withVolumeDevices(volumeDevices)
+```
+
+"volumeDevices is the list of block devices to be used by the container."
+
+### fn spec.template.spec.ephemeralContainers.withVolumeDevicesMixin
+
+```ts
+withVolumeDevicesMixin(volumeDevices)
+```
+
+"volumeDevices is the list of block devices to be used by the container."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.ephemeralContainers.withVolumeMounts
+
+```ts
+withVolumeMounts(volumeMounts)
+```
+
+"Pod volumes to mount into the container's filesystem. Subpath mounts are not allowed for ephemeral containers. Cannot be updated."
+
+### fn spec.template.spec.ephemeralContainers.withVolumeMountsMixin
+
+```ts
+withVolumeMountsMixin(volumeMounts)
+```
+
+"Pod volumes to mount into the container's filesystem. Subpath mounts are not allowed for ephemeral containers. Cannot be updated."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.ephemeralContainers.withWorkingDir
+
+```ts
+withWorkingDir(workingDir)
+```
+
+"Container's working directory. If not specified, the container runtime's default will be used, which might be configured in the container image. Cannot be updated."
+
+## obj spec.template.spec.ephemeralContainers.env
+
+"List of environment variables to set in the container. Cannot be updated."
+
+### fn spec.template.spec.ephemeralContainers.env.withName
+
+```ts
+withName(name)
+```
+
+"Name of the environment variable. Must be a C_IDENTIFIER."
+
+### fn spec.template.spec.ephemeralContainers.env.withValue
+
+```ts
+withValue(value)
+```
+
+"Variable references $(VAR_NAME) are expanded using the previously defined environment variables in the container and any service environment variables. If a variable cannot be resolved, the reference in the input string will be unchanged. Double $$ are reduced to a single $, which allows for escaping the $(VAR_NAME) syntax: i.e. \"$$(VAR_NAME)\" will produce the string literal \"$(VAR_NAME)\". Escaped references will never be expanded, regardless of whether the variable exists or not. Defaults to \"\"."
+
+## obj spec.template.spec.ephemeralContainers.env.valueFrom
+
+"Source for the environment variable's value. Cannot be used if value is not empty."
+
+## obj spec.template.spec.ephemeralContainers.env.valueFrom.configMapKeyRef
+
+"Selects a key of a ConfigMap."
+
+### fn spec.template.spec.ephemeralContainers.env.valueFrom.configMapKeyRef.withKey
+
+```ts
+withKey(key)
+```
+
+"The key to select."
+
+### fn spec.template.spec.ephemeralContainers.env.valueFrom.configMapKeyRef.withName
+
+```ts
+withName(name)
+```
+
+"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"
+
+### fn spec.template.spec.ephemeralContainers.env.valueFrom.configMapKeyRef.withOptional
+
+```ts
+withOptional(optional)
+```
+
+"Specify whether the ConfigMap or its key must be defined"
+
+## obj spec.template.spec.ephemeralContainers.env.valueFrom.fieldRef
+
+"Selects a field of the pod: supports metadata.name, metadata.namespace, `metadata.labels['']`, `metadata.annotations['']`, spec.nodeName, spec.serviceAccountName, status.hostIP, status.podIP, status.podIPs."
+
+### fn spec.template.spec.ephemeralContainers.env.valueFrom.fieldRef.withApiVersion
+
+```ts
+withApiVersion(apiVersion)
+```
+
+"Version of the schema the FieldPath is written in terms of, defaults to \"v1\"."
+
+### fn spec.template.spec.ephemeralContainers.env.valueFrom.fieldRef.withFieldPath
+
+```ts
+withFieldPath(fieldPath)
+```
+
+"Path of the field to select in the specified API version."
+
+## obj spec.template.spec.ephemeralContainers.env.valueFrom.resourceFieldRef
+
+"Selects a resource of the container: only resources limits and requests (limits.cpu, limits.memory, limits.ephemeral-storage, requests.cpu, requests.memory and requests.ephemeral-storage) are currently supported."
+
+### fn spec.template.spec.ephemeralContainers.env.valueFrom.resourceFieldRef.withContainerName
+
+```ts
+withContainerName(containerName)
+```
+
+"Container name: required for volumes, optional for env vars"
+
+### fn spec.template.spec.ephemeralContainers.env.valueFrom.resourceFieldRef.withDivisor
+
+```ts
+withDivisor(divisor)
+```
+
+"Specifies the output format of the exposed resources, defaults to \"1\
+
+### fn spec.template.spec.ephemeralContainers.env.valueFrom.resourceFieldRef.withResource
+
+```ts
+withResource(resource)
+```
+
+"Required: resource to select"
+
+## obj spec.template.spec.ephemeralContainers.env.valueFrom.secretKeyRef
+
+"Selects a key of a secret in the pod's namespace"
+
+### fn spec.template.spec.ephemeralContainers.env.valueFrom.secretKeyRef.withKey
+
+```ts
+withKey(key)
+```
+
+"The key of the secret to select from. Must be a valid secret key."
+
+### fn spec.template.spec.ephemeralContainers.env.valueFrom.secretKeyRef.withName
+
+```ts
+withName(name)
+```
+
+"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"
+
+### fn spec.template.spec.ephemeralContainers.env.valueFrom.secretKeyRef.withOptional
+
+```ts
+withOptional(optional)
+```
+
+"Specify whether the Secret or its key must be defined"
+
+## obj spec.template.spec.ephemeralContainers.envFrom
+
+"List of sources to populate environment variables in the container. The keys defined within a source must be a C_IDENTIFIER. All invalid keys will be reported as an event when the container is starting. When a key exists in multiple sources, the value associated with the last source will take precedence. Values defined by an Env with a duplicate key will take precedence. Cannot be updated."
+
+### fn spec.template.spec.ephemeralContainers.envFrom.withPrefix
+
+```ts
+withPrefix(prefix)
+```
+
+"An optional identifier to prepend to each key in the ConfigMap. Must be a C_IDENTIFIER."
+
+## obj spec.template.spec.ephemeralContainers.envFrom.configMapRef
+
+"The ConfigMap to select from"
+
+### fn spec.template.spec.ephemeralContainers.envFrom.configMapRef.withName
+
+```ts
+withName(name)
+```
+
+"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"
+
+### fn spec.template.spec.ephemeralContainers.envFrom.configMapRef.withOptional
+
+```ts
+withOptional(optional)
+```
+
+"Specify whether the ConfigMap must be defined"
+
+## obj spec.template.spec.ephemeralContainers.envFrom.secretRef
+
+"The Secret to select from"
+
+### fn spec.template.spec.ephemeralContainers.envFrom.secretRef.withName
+
+```ts
+withName(name)
+```
+
+"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"
+
+### fn spec.template.spec.ephemeralContainers.envFrom.secretRef.withOptional
+
+```ts
+withOptional(optional)
+```
+
+"Specify whether the Secret must be defined"
+
+## obj spec.template.spec.ephemeralContainers.lifecycle
+
+"Lifecycle is not allowed for ephemeral containers."
+
+## obj spec.template.spec.ephemeralContainers.lifecycle.postStart
+
+"PostStart is called immediately after a container is created. If the handler fails, the container is terminated and restarted according to its restart policy. Other management of the container blocks until the hook completes. More info: https://kubernetes.io/docs/concepts/containers/container-lifecycle-hooks/#container-hooks"
+
+## obj spec.template.spec.ephemeralContainers.lifecycle.postStart.exec
+
+"Exec specifies the action to take."
+
+### fn spec.template.spec.ephemeralContainers.lifecycle.postStart.exec.withCommand
+
+```ts
+withCommand(command)
+```
+
+"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy."
+
+### fn spec.template.spec.ephemeralContainers.lifecycle.postStart.exec.withCommandMixin
+
+```ts
+withCommandMixin(command)
+```
+
+"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy."
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.template.spec.ephemeralContainers.lifecycle.postStart.httpGet
+
+"HTTPGet specifies the http request to perform."
+
+### fn spec.template.spec.ephemeralContainers.lifecycle.postStart.httpGet.withHost
+
+```ts
+withHost(host)
+```
+
+"Host name to connect to, defaults to the pod IP. You probably want to set \"Host\" in httpHeaders instead."
+
+### fn spec.template.spec.ephemeralContainers.lifecycle.postStart.httpGet.withHttpHeaders
+
+```ts
+withHttpHeaders(httpHeaders)
+```
+
+"Custom headers to set in the request. HTTP allows repeated headers."
+
+### fn spec.template.spec.ephemeralContainers.lifecycle.postStart.httpGet.withHttpHeadersMixin
+
+```ts
+withHttpHeadersMixin(httpHeaders)
+```
+
+"Custom headers to set in the request. HTTP allows repeated headers."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.ephemeralContainers.lifecycle.postStart.httpGet.withPath
+
+```ts
+withPath(path)
+```
+
+"Path to access on the HTTP server."
+
+### fn spec.template.spec.ephemeralContainers.lifecycle.postStart.httpGet.withPort
+
+```ts
+withPort(port)
+```
+
+"Name or number of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."
+
+### fn spec.template.spec.ephemeralContainers.lifecycle.postStart.httpGet.withScheme
+
+```ts
+withScheme(scheme)
+```
+
+"Scheme to use for connecting to the host. Defaults to HTTP."
+
+## obj spec.template.spec.ephemeralContainers.lifecycle.postStart.httpGet.httpHeaders
+
+"Custom headers to set in the request. HTTP allows repeated headers."
+
+### fn spec.template.spec.ephemeralContainers.lifecycle.postStart.httpGet.httpHeaders.withName
+
+```ts
+withName(name)
+```
+
+"The header field name. This will be canonicalized upon output, so case-variant names will be understood as the same header."
+
+### fn spec.template.spec.ephemeralContainers.lifecycle.postStart.httpGet.httpHeaders.withValue
+
+```ts
+withValue(value)
+```
+
+"The header field value"
+
+## obj spec.template.spec.ephemeralContainers.lifecycle.postStart.tcpSocket
+
+"Deprecated. TCPSocket is NOT supported as a LifecycleHandler and kept for the backward compatibility. There are no validation of this field and lifecycle hooks will fail in runtime when tcp handler is specified."
+
+### fn spec.template.spec.ephemeralContainers.lifecycle.postStart.tcpSocket.withHost
+
+```ts
+withHost(host)
+```
+
+"Optional: Host name to connect to, defaults to the pod IP."
+
+### fn spec.template.spec.ephemeralContainers.lifecycle.postStart.tcpSocket.withPort
+
+```ts
+withPort(port)
+```
+
+"Number or name of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."
+
+## obj spec.template.spec.ephemeralContainers.lifecycle.preStop
+
+"PreStop is called immediately before a container is terminated due to an API request or management event such as liveness/startup probe failure, preemption, resource contention, etc. The handler is not called if the container crashes or exits. The Pod's termination grace period countdown begins before the PreStop hook is executed. Regardless of the outcome of the handler, the container will eventually terminate within the Pod's termination grace period (unless delayed by finalizers). Other management of the container blocks until the hook completes or until the termination grace period is reached. More info: https://kubernetes.io/docs/concepts/containers/container-lifecycle-hooks/#container-hooks"
+
+## obj spec.template.spec.ephemeralContainers.lifecycle.preStop.exec
+
+"Exec specifies the action to take."
+
+### fn spec.template.spec.ephemeralContainers.lifecycle.preStop.exec.withCommand
+
+```ts
+withCommand(command)
+```
+
+"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy."
+
+### fn spec.template.spec.ephemeralContainers.lifecycle.preStop.exec.withCommandMixin
+
+```ts
+withCommandMixin(command)
+```
+
+"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy."
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.template.spec.ephemeralContainers.lifecycle.preStop.httpGet
+
+"HTTPGet specifies the http request to perform."
+
+### fn spec.template.spec.ephemeralContainers.lifecycle.preStop.httpGet.withHost
+
+```ts
+withHost(host)
+```
+
+"Host name to connect to, defaults to the pod IP. You probably want to set \"Host\" in httpHeaders instead."
+
+### fn spec.template.spec.ephemeralContainers.lifecycle.preStop.httpGet.withHttpHeaders
+
+```ts
+withHttpHeaders(httpHeaders)
+```
+
+"Custom headers to set in the request. HTTP allows repeated headers."
+
+### fn spec.template.spec.ephemeralContainers.lifecycle.preStop.httpGet.withHttpHeadersMixin
+
+```ts
+withHttpHeadersMixin(httpHeaders)
+```
+
+"Custom headers to set in the request. HTTP allows repeated headers."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.ephemeralContainers.lifecycle.preStop.httpGet.withPath
+
+```ts
+withPath(path)
+```
+
+"Path to access on the HTTP server."
+
+### fn spec.template.spec.ephemeralContainers.lifecycle.preStop.httpGet.withPort
+
+```ts
+withPort(port)
+```
+
+"Name or number of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."
+
+### fn spec.template.spec.ephemeralContainers.lifecycle.preStop.httpGet.withScheme
+
+```ts
+withScheme(scheme)
+```
+
+"Scheme to use for connecting to the host. Defaults to HTTP."
+
+## obj spec.template.spec.ephemeralContainers.lifecycle.preStop.httpGet.httpHeaders
+
+"Custom headers to set in the request. HTTP allows repeated headers."
+
+### fn spec.template.spec.ephemeralContainers.lifecycle.preStop.httpGet.httpHeaders.withName
+
+```ts
+withName(name)
+```
+
+"The header field name. This will be canonicalized upon output, so case-variant names will be understood as the same header."
+
+### fn spec.template.spec.ephemeralContainers.lifecycle.preStop.httpGet.httpHeaders.withValue
+
+```ts
+withValue(value)
+```
+
+"The header field value"
+
+## obj spec.template.spec.ephemeralContainers.lifecycle.preStop.tcpSocket
+
+"Deprecated. TCPSocket is NOT supported as a LifecycleHandler and kept for the backward compatibility. There are no validation of this field and lifecycle hooks will fail in runtime when tcp handler is specified."
+
+### fn spec.template.spec.ephemeralContainers.lifecycle.preStop.tcpSocket.withHost
+
+```ts
+withHost(host)
+```
+
+"Optional: Host name to connect to, defaults to the pod IP."
+
+### fn spec.template.spec.ephemeralContainers.lifecycle.preStop.tcpSocket.withPort
+
+```ts
+withPort(port)
+```
+
+"Number or name of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."
+
+## obj spec.template.spec.ephemeralContainers.livenessProbe
+
+"Probes are not allowed for ephemeral containers."
+
+### fn spec.template.spec.ephemeralContainers.livenessProbe.withFailureThreshold
+
+```ts
+withFailureThreshold(failureThreshold)
+```
+
+"Minimum consecutive failures for the probe to be considered failed after having succeeded. Defaults to 3. Minimum value is 1."
+
+### fn spec.template.spec.ephemeralContainers.livenessProbe.withInitialDelaySeconds
+
+```ts
+withInitialDelaySeconds(initialDelaySeconds)
+```
+
+"Number of seconds after the container has started before liveness probes are initiated. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes"
+
+### fn spec.template.spec.ephemeralContainers.livenessProbe.withPeriodSeconds
+
+```ts
+withPeriodSeconds(periodSeconds)
+```
+
+"How often (in seconds) to perform the probe. Default to 10 seconds. Minimum value is 1."
+
+### fn spec.template.spec.ephemeralContainers.livenessProbe.withSuccessThreshold
+
+```ts
+withSuccessThreshold(successThreshold)
+```
+
+"Minimum consecutive successes for the probe to be considered successful after having failed. Defaults to 1. Must be 1 for liveness and startup. Minimum value is 1."
+
+### fn spec.template.spec.ephemeralContainers.livenessProbe.withTerminationGracePeriodSeconds
+
+```ts
+withTerminationGracePeriodSeconds(terminationGracePeriodSeconds)
+```
+
+"Optional duration in seconds the pod needs to terminate gracefully upon probe failure. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. If this value is nil, the pod's terminationGracePeriodSeconds will be used. Otherwise, this value overrides the value provided by the pod spec. Value must be non-negative integer. The value zero indicates stop immediately via the kill signal (no opportunity to shut down). This is a beta field and requires enabling ProbeTerminationGracePeriod feature gate. Minimum value is 1. spec.terminationGracePeriodSeconds is used if unset."
+
+### fn spec.template.spec.ephemeralContainers.livenessProbe.withTimeoutSeconds
+
+```ts
+withTimeoutSeconds(timeoutSeconds)
+```
+
+"Number of seconds after which the probe times out. Defaults to 1 second. Minimum value is 1. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes"
+
+## obj spec.template.spec.ephemeralContainers.livenessProbe.exec
+
+"Exec specifies the action to take."
+
+### fn spec.template.spec.ephemeralContainers.livenessProbe.exec.withCommand
+
+```ts
+withCommand(command)
+```
+
+"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy."
+
+### fn spec.template.spec.ephemeralContainers.livenessProbe.exec.withCommandMixin
+
+```ts
+withCommandMixin(command)
+```
+
+"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy."
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.template.spec.ephemeralContainers.livenessProbe.grpc
+
+"GRPC specifies an action involving a GRPC port."
+
+### fn spec.template.spec.ephemeralContainers.livenessProbe.grpc.withPort
+
+```ts
+withPort(port)
+```
+
+"Port number of the gRPC service. Number must be in the range 1 to 65535."
+
+### fn spec.template.spec.ephemeralContainers.livenessProbe.grpc.withService
+
+```ts
+withService(service)
+```
+
+"Service is the name of the service to place in the gRPC HealthCheckRequest (see https://github.com/grpc/grpc/blob/master/doc/health-checking.md). \n If this is not specified, the default behavior is defined by gRPC."
+
+## obj spec.template.spec.ephemeralContainers.livenessProbe.httpGet
+
+"HTTPGet specifies the http request to perform."
+
+### fn spec.template.spec.ephemeralContainers.livenessProbe.httpGet.withHost
+
+```ts
+withHost(host)
+```
+
+"Host name to connect to, defaults to the pod IP. You probably want to set \"Host\" in httpHeaders instead."
+
+### fn spec.template.spec.ephemeralContainers.livenessProbe.httpGet.withHttpHeaders
+
+```ts
+withHttpHeaders(httpHeaders)
+```
+
+"Custom headers to set in the request. HTTP allows repeated headers."
+
+### fn spec.template.spec.ephemeralContainers.livenessProbe.httpGet.withHttpHeadersMixin
+
+```ts
+withHttpHeadersMixin(httpHeaders)
+```
+
+"Custom headers to set in the request. HTTP allows repeated headers."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.ephemeralContainers.livenessProbe.httpGet.withPath
+
+```ts
+withPath(path)
+```
+
+"Path to access on the HTTP server."
+
+### fn spec.template.spec.ephemeralContainers.livenessProbe.httpGet.withPort
+
+```ts
+withPort(port)
+```
+
+"Name or number of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."
+
+### fn spec.template.spec.ephemeralContainers.livenessProbe.httpGet.withScheme
+
+```ts
+withScheme(scheme)
+```
+
+"Scheme to use for connecting to the host. Defaults to HTTP."
+
+## obj spec.template.spec.ephemeralContainers.livenessProbe.httpGet.httpHeaders
+
+"Custom headers to set in the request. HTTP allows repeated headers."
+
+### fn spec.template.spec.ephemeralContainers.livenessProbe.httpGet.httpHeaders.withName
+
+```ts
+withName(name)
+```
+
+"The header field name. This will be canonicalized upon output, so case-variant names will be understood as the same header."
+
+### fn spec.template.spec.ephemeralContainers.livenessProbe.httpGet.httpHeaders.withValue
+
+```ts
+withValue(value)
+```
+
+"The header field value"
+
+## obj spec.template.spec.ephemeralContainers.livenessProbe.tcpSocket
+
+"TCPSocket specifies an action involving a TCP port."
+
+### fn spec.template.spec.ephemeralContainers.livenessProbe.tcpSocket.withHost
+
+```ts
+withHost(host)
+```
+
+"Optional: Host name to connect to, defaults to the pod IP."
+
+### fn spec.template.spec.ephemeralContainers.livenessProbe.tcpSocket.withPort
+
+```ts
+withPort(port)
+```
+
+"Number or name of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."
+
+## obj spec.template.spec.ephemeralContainers.ports
+
+"Ports are not allowed for ephemeral containers."
+
+### fn spec.template.spec.ephemeralContainers.ports.withContainerPort
+
+```ts
+withContainerPort(containerPort)
+```
+
+"Number of port to expose on the pod's IP address. This must be a valid port number, 0 < x < 65536."
+
+### fn spec.template.spec.ephemeralContainers.ports.withHostIP
+
+```ts
+withHostIP(hostIP)
+```
+
+"What host IP to bind the external port to."
+
+### fn spec.template.spec.ephemeralContainers.ports.withHostPort
+
+```ts
+withHostPort(hostPort)
+```
+
+"Number of port to expose on the host. If specified, this must be a valid port number, 0 < x < 65536. If HostNetwork is specified, this must match ContainerPort. Most containers do not need this."
+
+### fn spec.template.spec.ephemeralContainers.ports.withName
+
+```ts
+withName(name)
+```
+
+"If specified, this must be an IANA_SVC_NAME and unique within the pod. Each named port in a pod must have a unique name. Name for the port that can be referred to by services."
+
+### fn spec.template.spec.ephemeralContainers.ports.withProtocol
+
+```ts
+withProtocol(protocol)
+```
+
+"Protocol for port. Must be UDP, TCP, or SCTP. Defaults to \"TCP\"."
+
+## obj spec.template.spec.ephemeralContainers.readinessProbe
+
+"Probes are not allowed for ephemeral containers."
+
+### fn spec.template.spec.ephemeralContainers.readinessProbe.withFailureThreshold
+
+```ts
+withFailureThreshold(failureThreshold)
+```
+
+"Minimum consecutive failures for the probe to be considered failed after having succeeded. Defaults to 3. Minimum value is 1."
+
+### fn spec.template.spec.ephemeralContainers.readinessProbe.withInitialDelaySeconds
+
+```ts
+withInitialDelaySeconds(initialDelaySeconds)
+```
+
+"Number of seconds after the container has started before liveness probes are initiated. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes"
+
+### fn spec.template.spec.ephemeralContainers.readinessProbe.withPeriodSeconds
+
+```ts
+withPeriodSeconds(periodSeconds)
+```
+
+"How often (in seconds) to perform the probe. Default to 10 seconds. Minimum value is 1."
+
+### fn spec.template.spec.ephemeralContainers.readinessProbe.withSuccessThreshold
+
+```ts
+withSuccessThreshold(successThreshold)
+```
+
+"Minimum consecutive successes for the probe to be considered successful after having failed. Defaults to 1. Must be 1 for liveness and startup. Minimum value is 1."
+
+### fn spec.template.spec.ephemeralContainers.readinessProbe.withTerminationGracePeriodSeconds
+
+```ts
+withTerminationGracePeriodSeconds(terminationGracePeriodSeconds)
+```
+
+"Optional duration in seconds the pod needs to terminate gracefully upon probe failure. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. If this value is nil, the pod's terminationGracePeriodSeconds will be used. Otherwise, this value overrides the value provided by the pod spec. Value must be non-negative integer. The value zero indicates stop immediately via the kill signal (no opportunity to shut down). This is a beta field and requires enabling ProbeTerminationGracePeriod feature gate. Minimum value is 1. spec.terminationGracePeriodSeconds is used if unset."
+
+### fn spec.template.spec.ephemeralContainers.readinessProbe.withTimeoutSeconds
+
+```ts
+withTimeoutSeconds(timeoutSeconds)
+```
+
+"Number of seconds after which the probe times out. Defaults to 1 second. Minimum value is 1. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes"
+
+## obj spec.template.spec.ephemeralContainers.readinessProbe.exec
+
+"Exec specifies the action to take."
+
+### fn spec.template.spec.ephemeralContainers.readinessProbe.exec.withCommand
+
+```ts
+withCommand(command)
+```
+
+"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy."
+
+### fn spec.template.spec.ephemeralContainers.readinessProbe.exec.withCommandMixin
+
+```ts
+withCommandMixin(command)
+```
+
+"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy."
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.template.spec.ephemeralContainers.readinessProbe.grpc
+
+"GRPC specifies an action involving a GRPC port."
+
+### fn spec.template.spec.ephemeralContainers.readinessProbe.grpc.withPort
+
+```ts
+withPort(port)
+```
+
+"Port number of the gRPC service. Number must be in the range 1 to 65535."
+
+### fn spec.template.spec.ephemeralContainers.readinessProbe.grpc.withService
+
+```ts
+withService(service)
+```
+
+"Service is the name of the service to place in the gRPC HealthCheckRequest (see https://github.com/grpc/grpc/blob/master/doc/health-checking.md). \n If this is not specified, the default behavior is defined by gRPC."
+
+## obj spec.template.spec.ephemeralContainers.readinessProbe.httpGet
+
+"HTTPGet specifies the http request to perform."
+
+### fn spec.template.spec.ephemeralContainers.readinessProbe.httpGet.withHost
+
+```ts
+withHost(host)
+```
+
+"Host name to connect to, defaults to the pod IP. You probably want to set \"Host\" in httpHeaders instead."
+
+### fn spec.template.spec.ephemeralContainers.readinessProbe.httpGet.withHttpHeaders
+
+```ts
+withHttpHeaders(httpHeaders)
+```
+
+"Custom headers to set in the request. HTTP allows repeated headers."
+
+### fn spec.template.spec.ephemeralContainers.readinessProbe.httpGet.withHttpHeadersMixin
+
+```ts
+withHttpHeadersMixin(httpHeaders)
+```
+
+"Custom headers to set in the request. HTTP allows repeated headers."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.ephemeralContainers.readinessProbe.httpGet.withPath
+
+```ts
+withPath(path)
+```
+
+"Path to access on the HTTP server."
+
+### fn spec.template.spec.ephemeralContainers.readinessProbe.httpGet.withPort
+
+```ts
+withPort(port)
+```
+
+"Name or number of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."
+
+### fn spec.template.spec.ephemeralContainers.readinessProbe.httpGet.withScheme
+
+```ts
+withScheme(scheme)
+```
+
+"Scheme to use for connecting to the host. Defaults to HTTP."
+
+## obj spec.template.spec.ephemeralContainers.readinessProbe.httpGet.httpHeaders
+
+"Custom headers to set in the request. HTTP allows repeated headers."
+
+### fn spec.template.spec.ephemeralContainers.readinessProbe.httpGet.httpHeaders.withName
+
+```ts
+withName(name)
+```
+
+"The header field name. This will be canonicalized upon output, so case-variant names will be understood as the same header."
+
+### fn spec.template.spec.ephemeralContainers.readinessProbe.httpGet.httpHeaders.withValue
+
+```ts
+withValue(value)
+```
+
+"The header field value"
+
+## obj spec.template.spec.ephemeralContainers.readinessProbe.tcpSocket
+
+"TCPSocket specifies an action involving a TCP port."
+
+### fn spec.template.spec.ephemeralContainers.readinessProbe.tcpSocket.withHost
+
+```ts
+withHost(host)
+```
+
+"Optional: Host name to connect to, defaults to the pod IP."
+
+### fn spec.template.spec.ephemeralContainers.readinessProbe.tcpSocket.withPort
+
+```ts
+withPort(port)
+```
+
+"Number or name of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."
+
+## obj spec.template.spec.ephemeralContainers.resizePolicy
+
+"Resources resize policy for the container."
+
+### fn spec.template.spec.ephemeralContainers.resizePolicy.withResourceName
+
+```ts
+withResourceName(resourceName)
+```
+
+"Name of the resource to which this resource resize policy applies. Supported values: cpu, memory."
+
+### fn spec.template.spec.ephemeralContainers.resizePolicy.withRestartPolicy
+
+```ts
+withRestartPolicy(restartPolicy)
+```
+
+"Restart policy to apply when specified resource is resized. If not specified, it defaults to NotRequired."
+
+## obj spec.template.spec.ephemeralContainers.resources
+
+"Resources are not allowed for ephemeral containers. Ephemeral containers use spare resources already allocated to the pod."
+
+### fn spec.template.spec.ephemeralContainers.resources.withClaims
+
+```ts
+withClaims(claims)
+```
+
+"Claims lists the names of resources, defined in spec.resourceClaims, that are used by this container. \n This is an alpha field and requires enabling the DynamicResourceAllocation feature gate. \n This field is immutable. It can only be set for containers."
+
+### fn spec.template.spec.ephemeralContainers.resources.withClaimsMixin
+
+```ts
+withClaimsMixin(claims)
+```
+
+"Claims lists the names of resources, defined in spec.resourceClaims, that are used by this container. \n This is an alpha field and requires enabling the DynamicResourceAllocation feature gate. \n This field is immutable. It can only be set for containers."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.ephemeralContainers.resources.withLimits
+
+```ts
+withLimits(limits)
+```
+
+"Limits describes the maximum amount of compute resources allowed. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"
+
+### fn spec.template.spec.ephemeralContainers.resources.withLimitsMixin
+
+```ts
+withLimitsMixin(limits)
+```
+
+"Limits describes the maximum amount of compute resources allowed. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.ephemeralContainers.resources.withRequests
+
+```ts
+withRequests(requests)
+```
+
+"Requests describes the minimum amount of compute resources required. If Requests is omitted for a container, it defaults to Limits if that is explicitly specified, otherwise to an implementation-defined value. Requests cannot exceed Limits. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"
+
+### fn spec.template.spec.ephemeralContainers.resources.withRequestsMixin
+
+```ts
+withRequestsMixin(requests)
+```
+
+"Requests describes the minimum amount of compute resources required. If Requests is omitted for a container, it defaults to Limits if that is explicitly specified, otherwise to an implementation-defined value. Requests cannot exceed Limits. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.template.spec.ephemeralContainers.resources.claims
+
+"Claims lists the names of resources, defined in spec.resourceClaims, that are used by this container. \n This is an alpha field and requires enabling the DynamicResourceAllocation feature gate. \n This field is immutable. It can only be set for containers."
+
+### fn spec.template.spec.ephemeralContainers.resources.claims.withName
+
+```ts
+withName(name)
+```
+
+"Name must match the name of one entry in pod.spec.resourceClaims of the Pod where this field is used. It makes that resource available inside a container."
+
+## obj spec.template.spec.ephemeralContainers.securityContext
+
+"Optional: SecurityContext defines the security options the ephemeral container should be run with. If set, the fields of SecurityContext override the equivalent fields of PodSecurityContext."
+
+### fn spec.template.spec.ephemeralContainers.securityContext.withAllowPrivilegeEscalation
+
+```ts
+withAllowPrivilegeEscalation(allowPrivilegeEscalation)
+```
+
+"AllowPrivilegeEscalation controls whether a process can gain more privileges than its parent process. This bool directly controls if the no_new_privs flag will be set on the container process. AllowPrivilegeEscalation is true always when the container is: 1) run as Privileged 2) has CAP_SYS_ADMIN Note that this field cannot be set when spec.os.name is windows."
+
+### fn spec.template.spec.ephemeralContainers.securityContext.withPrivileged
+
+```ts
+withPrivileged(privileged)
+```
+
+"Run container in privileged mode. Processes in privileged containers are essentially equivalent to root on the host. Defaults to false. Note that this field cannot be set when spec.os.name is windows."
+
+### fn spec.template.spec.ephemeralContainers.securityContext.withProcMount
+
+```ts
+withProcMount(procMount)
+```
+
+"procMount denotes the type of proc mount to use for the containers. The default is DefaultProcMount which uses the container runtime defaults for readonly paths and masked paths. This requires the ProcMountType feature flag to be enabled. Note that this field cannot be set when spec.os.name is windows."
+
+### fn spec.template.spec.ephemeralContainers.securityContext.withReadOnlyRootFilesystem
+
+```ts
+withReadOnlyRootFilesystem(readOnlyRootFilesystem)
+```
+
+"Whether this container has a read-only root filesystem. Default is false. Note that this field cannot be set when spec.os.name is windows."
+
+### fn spec.template.spec.ephemeralContainers.securityContext.withRunAsGroup
+
+```ts
+withRunAsGroup(runAsGroup)
+```
+
+"The GID to run the entrypoint of the container process. Uses runtime default if unset. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. Note that this field cannot be set when spec.os.name is windows."
+
+### fn spec.template.spec.ephemeralContainers.securityContext.withRunAsNonRoot
+
+```ts
+withRunAsNonRoot(runAsNonRoot)
+```
+
+"Indicates that the container must run as a non-root user. If true, the Kubelet will validate the image at runtime to ensure that it does not run as UID 0 (root) and fail to start the container if it does. If unset or false, no such validation will be performed. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence."
+
+### fn spec.template.spec.ephemeralContainers.securityContext.withRunAsUser
+
+```ts
+withRunAsUser(runAsUser)
+```
+
+"The UID to run the entrypoint of the container process. Defaults to user specified in image metadata if unspecified. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. Note that this field cannot be set when spec.os.name is windows."
+
+## obj spec.template.spec.ephemeralContainers.securityContext.capabilities
+
+"The capabilities to add/drop when running containers. Defaults to the default set of capabilities granted by the container runtime. Note that this field cannot be set when spec.os.name is windows."
+
+### fn spec.template.spec.ephemeralContainers.securityContext.capabilities.withAdd
+
+```ts
+withAdd(add)
+```
+
+"Added capabilities"
+
+### fn spec.template.spec.ephemeralContainers.securityContext.capabilities.withAddMixin
+
+```ts
+withAddMixin(add)
+```
+
+"Added capabilities"
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.ephemeralContainers.securityContext.capabilities.withDrop
+
+```ts
+withDrop(drop)
+```
+
+"Removed capabilities"
+
+### fn spec.template.spec.ephemeralContainers.securityContext.capabilities.withDropMixin
+
+```ts
+withDropMixin(drop)
+```
+
+"Removed capabilities"
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.template.spec.ephemeralContainers.securityContext.seLinuxOptions
+
+"The SELinux context to be applied to the container. If unspecified, the container runtime will allocate a random SELinux context for each container. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. Note that this field cannot be set when spec.os.name is windows."
+
+### fn spec.template.spec.ephemeralContainers.securityContext.seLinuxOptions.withLevel
+
+```ts
+withLevel(level)
+```
+
+"Level is SELinux level label that applies to the container."
+
+### fn spec.template.spec.ephemeralContainers.securityContext.seLinuxOptions.withRole
+
+```ts
+withRole(role)
+```
+
+"Role is a SELinux role label that applies to the container."
+
+### fn spec.template.spec.ephemeralContainers.securityContext.seLinuxOptions.withType
+
+```ts
+withType(type)
+```
+
+"Type is a SELinux type label that applies to the container."
+
+### fn spec.template.spec.ephemeralContainers.securityContext.seLinuxOptions.withUser
+
+```ts
+withUser(user)
+```
+
+"User is a SELinux user label that applies to the container."
+
+## obj spec.template.spec.ephemeralContainers.securityContext.seccompProfile
+
+"The seccomp options to use by this container. If seccomp options are provided at both the pod & container level, the container options override the pod options. Note that this field cannot be set when spec.os.name is windows."
+
+### fn spec.template.spec.ephemeralContainers.securityContext.seccompProfile.withLocalhostProfile
+
+```ts
+withLocalhostProfile(localhostProfile)
+```
+
+"localhostProfile indicates a profile defined in a file on the node should be used. The profile must be preconfigured on the node to work. Must be a descending path, relative to the kubelet's configured seccomp profile location. Must be set if type is \"Localhost\". Must NOT be set for any other type."
+
+### fn spec.template.spec.ephemeralContainers.securityContext.seccompProfile.withType
+
+```ts
+withType(type)
+```
+
+"type indicates which kind of seccomp profile will be applied. Valid options are: \n Localhost - a profile defined in a file on the node should be used. RuntimeDefault - the container runtime default profile should be used. Unconfined - no profile should be applied."
+
+## obj spec.template.spec.ephemeralContainers.securityContext.windowsOptions
+
+"The Windows specific settings applied to all containers. If unspecified, the options from the PodSecurityContext will be used. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. Note that this field cannot be set when spec.os.name is linux."
+
+### fn spec.template.spec.ephemeralContainers.securityContext.windowsOptions.withGmsaCredentialSpec
+
+```ts
+withGmsaCredentialSpec(gmsaCredentialSpec)
+```
+
+"GMSACredentialSpec is where the GMSA admission webhook (https://github.com/kubernetes-sigs/windows-gmsa) inlines the contents of the GMSA credential spec named by the GMSACredentialSpecName field."
+
+### fn spec.template.spec.ephemeralContainers.securityContext.windowsOptions.withGmsaCredentialSpecName
+
+```ts
+withGmsaCredentialSpecName(gmsaCredentialSpecName)
+```
+
+"GMSACredentialSpecName is the name of the GMSA credential spec to use."
+
+### fn spec.template.spec.ephemeralContainers.securityContext.windowsOptions.withHostProcess
+
+```ts
+withHostProcess(hostProcess)
+```
+
+"HostProcess determines if a container should be run as a 'Host Process' container. All of a Pod's containers must have the same effective HostProcess value (it is not allowed to have a mix of HostProcess containers and non-HostProcess containers). In addition, if HostProcess is true then HostNetwork must also be set to true."
+
+### fn spec.template.spec.ephemeralContainers.securityContext.windowsOptions.withRunAsUserName
+
+```ts
+withRunAsUserName(runAsUserName)
+```
+
+"The UserName in Windows to run the entrypoint of the container process. Defaults to the user specified in image metadata if unspecified. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence."
+
+## obj spec.template.spec.ephemeralContainers.startupProbe
+
+"Probes are not allowed for ephemeral containers."
+
+### fn spec.template.spec.ephemeralContainers.startupProbe.withFailureThreshold
+
+```ts
+withFailureThreshold(failureThreshold)
+```
+
+"Minimum consecutive failures for the probe to be considered failed after having succeeded. Defaults to 3. Minimum value is 1."
+
+### fn spec.template.spec.ephemeralContainers.startupProbe.withInitialDelaySeconds
+
+```ts
+withInitialDelaySeconds(initialDelaySeconds)
+```
+
+"Number of seconds after the container has started before liveness probes are initiated. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes"
+
+### fn spec.template.spec.ephemeralContainers.startupProbe.withPeriodSeconds
+
+```ts
+withPeriodSeconds(periodSeconds)
+```
+
+"How often (in seconds) to perform the probe. Default to 10 seconds. Minimum value is 1."
+
+### fn spec.template.spec.ephemeralContainers.startupProbe.withSuccessThreshold
+
+```ts
+withSuccessThreshold(successThreshold)
+```
+
+"Minimum consecutive successes for the probe to be considered successful after having failed. Defaults to 1. Must be 1 for liveness and startup. Minimum value is 1."
+
+### fn spec.template.spec.ephemeralContainers.startupProbe.withTerminationGracePeriodSeconds
+
+```ts
+withTerminationGracePeriodSeconds(terminationGracePeriodSeconds)
+```
+
+"Optional duration in seconds the pod needs to terminate gracefully upon probe failure. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. If this value is nil, the pod's terminationGracePeriodSeconds will be used. Otherwise, this value overrides the value provided by the pod spec. Value must be non-negative integer. The value zero indicates stop immediately via the kill signal (no opportunity to shut down). This is a beta field and requires enabling ProbeTerminationGracePeriod feature gate. Minimum value is 1. spec.terminationGracePeriodSeconds is used if unset."
+
+### fn spec.template.spec.ephemeralContainers.startupProbe.withTimeoutSeconds
+
+```ts
+withTimeoutSeconds(timeoutSeconds)
+```
+
+"Number of seconds after which the probe times out. Defaults to 1 second. Minimum value is 1. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes"
+
+## obj spec.template.spec.ephemeralContainers.startupProbe.exec
+
+"Exec specifies the action to take."
+
+### fn spec.template.spec.ephemeralContainers.startupProbe.exec.withCommand
+
+```ts
+withCommand(command)
+```
+
+"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy."
+
+### fn spec.template.spec.ephemeralContainers.startupProbe.exec.withCommandMixin
+
+```ts
+withCommandMixin(command)
+```
+
+"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy."
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.template.spec.ephemeralContainers.startupProbe.grpc
+
+"GRPC specifies an action involving a GRPC port."
+
+### fn spec.template.spec.ephemeralContainers.startupProbe.grpc.withPort
+
+```ts
+withPort(port)
+```
+
+"Port number of the gRPC service. Number must be in the range 1 to 65535."
+
+### fn spec.template.spec.ephemeralContainers.startupProbe.grpc.withService
+
+```ts
+withService(service)
+```
+
+"Service is the name of the service to place in the gRPC HealthCheckRequest (see https://github.com/grpc/grpc/blob/master/doc/health-checking.md). \n If this is not specified, the default behavior is defined by gRPC."
+
+## obj spec.template.spec.ephemeralContainers.startupProbe.httpGet
+
+"HTTPGet specifies the http request to perform."
+
+### fn spec.template.spec.ephemeralContainers.startupProbe.httpGet.withHost
+
+```ts
+withHost(host)
+```
+
+"Host name to connect to, defaults to the pod IP. You probably want to set \"Host\" in httpHeaders instead."
+
+### fn spec.template.spec.ephemeralContainers.startupProbe.httpGet.withHttpHeaders
+
+```ts
+withHttpHeaders(httpHeaders)
+```
+
+"Custom headers to set in the request. HTTP allows repeated headers."
+
+### fn spec.template.spec.ephemeralContainers.startupProbe.httpGet.withHttpHeadersMixin
+
+```ts
+withHttpHeadersMixin(httpHeaders)
+```
+
+"Custom headers to set in the request. HTTP allows repeated headers."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.ephemeralContainers.startupProbe.httpGet.withPath
+
+```ts
+withPath(path)
+```
+
+"Path to access on the HTTP server."
+
+### fn spec.template.spec.ephemeralContainers.startupProbe.httpGet.withPort
+
+```ts
+withPort(port)
+```
+
+"Name or number of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."
+
+### fn spec.template.spec.ephemeralContainers.startupProbe.httpGet.withScheme
+
+```ts
+withScheme(scheme)
+```
+
+"Scheme to use for connecting to the host. Defaults to HTTP."
+
+## obj spec.template.spec.ephemeralContainers.startupProbe.httpGet.httpHeaders
+
+"Custom headers to set in the request. HTTP allows repeated headers."
+
+### fn spec.template.spec.ephemeralContainers.startupProbe.httpGet.httpHeaders.withName
+
+```ts
+withName(name)
+```
+
+"The header field name. This will be canonicalized upon output, so case-variant names will be understood as the same header."
+
+### fn spec.template.spec.ephemeralContainers.startupProbe.httpGet.httpHeaders.withValue
+
+```ts
+withValue(value)
+```
+
+"The header field value"
+
+## obj spec.template.spec.ephemeralContainers.startupProbe.tcpSocket
+
+"TCPSocket specifies an action involving a TCP port."
+
+### fn spec.template.spec.ephemeralContainers.startupProbe.tcpSocket.withHost
+
+```ts
+withHost(host)
+```
+
+"Optional: Host name to connect to, defaults to the pod IP."
+
+### fn spec.template.spec.ephemeralContainers.startupProbe.tcpSocket.withPort
+
+```ts
+withPort(port)
+```
+
+"Number or name of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."
+
+## obj spec.template.spec.ephemeralContainers.volumeDevices
+
+"volumeDevices is the list of block devices to be used by the container."
+
+### fn spec.template.spec.ephemeralContainers.volumeDevices.withDevicePath
+
+```ts
+withDevicePath(devicePath)
+```
+
+"devicePath is the path inside of the container that the device will be mapped to."
+
+### fn spec.template.spec.ephemeralContainers.volumeDevices.withName
+
+```ts
+withName(name)
+```
+
+"name must match the name of a persistentVolumeClaim in the pod"
+
+## obj spec.template.spec.ephemeralContainers.volumeMounts
+
+"Pod volumes to mount into the container's filesystem. Subpath mounts are not allowed for ephemeral containers. Cannot be updated."
+
+### fn spec.template.spec.ephemeralContainers.volumeMounts.withMountPath
+
+```ts
+withMountPath(mountPath)
+```
+
+"Path within the container at which the volume should be mounted. Must not contain ':'."
+
+### fn spec.template.spec.ephemeralContainers.volumeMounts.withMountPropagation
+
+```ts
+withMountPropagation(mountPropagation)
+```
+
+"mountPropagation determines how mounts are propagated from the host to container and the other way around. When not set, MountPropagationNone is used. This field is beta in 1.10."
+
+### fn spec.template.spec.ephemeralContainers.volumeMounts.withName
+
+```ts
+withName(name)
+```
+
+"This must match the Name of a Volume."
+
+### fn spec.template.spec.ephemeralContainers.volumeMounts.withReadOnly
+
+```ts
+withReadOnly(readOnly)
+```
+
+"Mounted read-only if true, read-write otherwise (false or unspecified). Defaults to false."
+
+### fn spec.template.spec.ephemeralContainers.volumeMounts.withSubPath
+
+```ts
+withSubPath(subPath)
+```
+
+"Path within the volume from which the container's volume should be mounted. Defaults to \"\" (volume's root)."
+
+### fn spec.template.spec.ephemeralContainers.volumeMounts.withSubPathExpr
+
+```ts
+withSubPathExpr(subPathExpr)
+```
+
+"Expanded path within the volume from which the container's volume should be mounted. Behaves similarly to SubPath but environment variable references $(VAR_NAME) are expanded using the container's environment. Defaults to \"\" (volume's root). SubPathExpr and SubPath are mutually exclusive."
+
+## obj spec.template.spec.hostAliases
+
+"HostAliases is an optional list of hosts and IPs that will be injected into the pod's hosts file if specified. This is only valid for non-hostNetwork pods."
+
+### fn spec.template.spec.hostAliases.withHostnames
+
+```ts
+withHostnames(hostnames)
+```
+
+"Hostnames for the above IP address."
+
+### fn spec.template.spec.hostAliases.withHostnamesMixin
+
+```ts
+withHostnamesMixin(hostnames)
+```
+
+"Hostnames for the above IP address."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.hostAliases.withIp
+
+```ts
+withIp(ip)
+```
+
+"IP address of the host file entry."
+
+## obj spec.template.spec.imagePullSecrets
+
+"ImagePullSecrets is an optional list of references to secrets in the same namespace to use for pulling any of the images used by this PodSpec. If specified, these secrets will be passed to individual puller implementations for them to use. More info: https://kubernetes.io/docs/concepts/containers/images#specifying-imagepullsecrets-on-a-pod"
+
+### fn spec.template.spec.imagePullSecrets.withName
+
+```ts
+withName(name)
+```
+
+"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"
+
+## obj spec.template.spec.initContainers
+
+"List of initialization containers belonging to the pod. Init containers are executed in order prior to containers being started. If any init container fails, the pod is considered to have failed and is handled according to its restartPolicy. The name for an init container or normal container must be unique among all containers. Init containers may not have Lifecycle actions, Readiness probes, Liveness probes, or Startup probes. The resourceRequirements of an init container are taken into account during scheduling by finding the highest request/limit for each resource type, and then using the max of of that value or the sum of the normal containers. Limits are applied to init containers in a similar fashion. Init containers cannot currently be added or removed. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/init-containers/"
+
+### fn spec.template.spec.initContainers.withArgs
+
+```ts
+withArgs(args)
+```
+
+"Arguments to the entrypoint. The container image's CMD is used if this is not provided. Variable references $(VAR_NAME) are expanded using the container's environment. If a variable cannot be resolved, the reference in the input string will be unchanged. Double $$ are reduced to a single $, which allows for escaping the $(VAR_NAME) syntax: i.e. \"$$(VAR_NAME)\" will produce the string literal \"$(VAR_NAME)\". Escaped references will never be expanded, regardless of whether the variable exists or not. Cannot be updated. More info: https://kubernetes.io/docs/tasks/inject-data-application/define-command-argument-container/#running-a-command-in-a-shell"
+
+### fn spec.template.spec.initContainers.withArgsMixin
+
+```ts
+withArgsMixin(args)
+```
+
+"Arguments to the entrypoint. The container image's CMD is used if this is not provided. Variable references $(VAR_NAME) are expanded using the container's environment. If a variable cannot be resolved, the reference in the input string will be unchanged. Double $$ are reduced to a single $, which allows for escaping the $(VAR_NAME) syntax: i.e. \"$$(VAR_NAME)\" will produce the string literal \"$(VAR_NAME)\". Escaped references will never be expanded, regardless of whether the variable exists or not. Cannot be updated. More info: https://kubernetes.io/docs/tasks/inject-data-application/define-command-argument-container/#running-a-command-in-a-shell"
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.initContainers.withCommand
+
+```ts
+withCommand(command)
+```
+
+"Entrypoint array. Not executed within a shell. The container image's ENTRYPOINT is used if this is not provided. Variable references $(VAR_NAME) are expanded using the container's environment. If a variable cannot be resolved, the reference in the input string will be unchanged. Double $$ are reduced to a single $, which allows for escaping the $(VAR_NAME) syntax: i.e. \"$$(VAR_NAME)\" will produce the string literal \"$(VAR_NAME)\". Escaped references will never be expanded, regardless of whether the variable exists or not. Cannot be updated. More info: https://kubernetes.io/docs/tasks/inject-data-application/define-command-argument-container/#running-a-command-in-a-shell"
+
+### fn spec.template.spec.initContainers.withCommandMixin
+
+```ts
+withCommandMixin(command)
+```
+
+"Entrypoint array. Not executed within a shell. The container image's ENTRYPOINT is used if this is not provided. Variable references $(VAR_NAME) are expanded using the container's environment. If a variable cannot be resolved, the reference in the input string will be unchanged. Double $$ are reduced to a single $, which allows for escaping the $(VAR_NAME) syntax: i.e. \"$$(VAR_NAME)\" will produce the string literal \"$(VAR_NAME)\". Escaped references will never be expanded, regardless of whether the variable exists or not. Cannot be updated. More info: https://kubernetes.io/docs/tasks/inject-data-application/define-command-argument-container/#running-a-command-in-a-shell"
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.initContainers.withEnv
+
+```ts
+withEnv(env)
+```
+
+"List of environment variables to set in the container. Cannot be updated."
+
+### fn spec.template.spec.initContainers.withEnvFrom
+
+```ts
+withEnvFrom(envFrom)
+```
+
+"List of sources to populate environment variables in the container. The keys defined within a source must be a C_IDENTIFIER. All invalid keys will be reported as an event when the container is starting. When a key exists in multiple sources, the value associated with the last source will take precedence. Values defined by an Env with a duplicate key will take precedence. Cannot be updated."
+
+### fn spec.template.spec.initContainers.withEnvFromMixin
+
+```ts
+withEnvFromMixin(envFrom)
+```
+
+"List of sources to populate environment variables in the container. The keys defined within a source must be a C_IDENTIFIER. All invalid keys will be reported as an event when the container is starting. When a key exists in multiple sources, the value associated with the last source will take precedence. Values defined by an Env with a duplicate key will take precedence. Cannot be updated."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.initContainers.withEnvMixin
+
+```ts
+withEnvMixin(env)
+```
+
+"List of environment variables to set in the container. Cannot be updated."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.initContainers.withImage
+
+```ts
+withImage(image)
+```
+
+"Container image name. More info: https://kubernetes.io/docs/concepts/containers/images This field is optional to allow higher level config management to default or override container images in workload controllers like Deployments and StatefulSets."
+
+### fn spec.template.spec.initContainers.withImagePullPolicy
+
+```ts
+withImagePullPolicy(imagePullPolicy)
+```
+
+"Image pull policy. One of Always, Never, IfNotPresent. Defaults to Always if :latest tag is specified, or IfNotPresent otherwise. Cannot be updated. More info: https://kubernetes.io/docs/concepts/containers/images#updating-images"
+
+### fn spec.template.spec.initContainers.withName
+
+```ts
+withName(name)
+```
+
+"Name of the container specified as a DNS_LABEL. Each container in a pod must have a unique name (DNS_LABEL). Cannot be updated."
+
+### fn spec.template.spec.initContainers.withPorts
+
+```ts
+withPorts(ports)
+```
+
+"List of ports to expose from the container. Not specifying a port here DOES NOT prevent that port from being exposed. Any port which is listening on the default \"0.0.0.0\" address inside a container will be accessible from the network. Modifying this array with strategic merge patch may corrupt the data. For more information See https://github.com/kubernetes/kubernetes/issues/108255. Cannot be updated."
+
+### fn spec.template.spec.initContainers.withPortsMixin
+
+```ts
+withPortsMixin(ports)
+```
+
+"List of ports to expose from the container. Not specifying a port here DOES NOT prevent that port from being exposed. Any port which is listening on the default \"0.0.0.0\" address inside a container will be accessible from the network. Modifying this array with strategic merge patch may corrupt the data. For more information See https://github.com/kubernetes/kubernetes/issues/108255. Cannot be updated."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.initContainers.withResizePolicy
+
+```ts
+withResizePolicy(resizePolicy)
+```
+
+"Resources resize policy for the container."
+
+### fn spec.template.spec.initContainers.withResizePolicyMixin
+
+```ts
+withResizePolicyMixin(resizePolicy)
+```
+
+"Resources resize policy for the container."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.initContainers.withRestartPolicy
+
+```ts
+withRestartPolicy(restartPolicy)
+```
+
+"RestartPolicy defines the restart behavior of individual containers in a pod. This field may only be set for init containers, and the only allowed value is \"Always\". For non-init containers or when this field is not specified, the restart behavior is defined by the Pod's restart policy and the container type. Setting the RestartPolicy as \"Always\" for the init container will have the following effect: this init container will be continually restarted on exit until all regular containers have terminated. Once all regular containers have completed, all init containers with restartPolicy \"Always\" will be shut down. This lifecycle differs from normal init containers and is often referred to as a \"sidecar\" container. Although this init container still starts in the init container sequence, it does not wait for the container to complete before proceeding to the next init container. Instead, the next init container starts immediately after this init container is started, or after any startupProbe has successfully completed."
+
+### fn spec.template.spec.initContainers.withStdin
+
+```ts
+withStdin(stdin)
+```
+
+"Whether this container should allocate a buffer for stdin in the container runtime. If this is not set, reads from stdin in the container will always result in EOF. Default is false."
+
+### fn spec.template.spec.initContainers.withStdinOnce
+
+```ts
+withStdinOnce(stdinOnce)
+```
+
+"Whether the container runtime should close the stdin channel after it has been opened by a single attach. When stdin is true the stdin stream will remain open across multiple attach sessions. If stdinOnce is set to true, stdin is opened on container start, is empty until the first client attaches to stdin, and then remains open and accepts data until the client disconnects, at which time stdin is closed and remains closed until the container is restarted. If this flag is false, a container processes that reads from stdin will never receive an EOF. Default is false"
+
+### fn spec.template.spec.initContainers.withTerminationMessagePath
+
+```ts
+withTerminationMessagePath(terminationMessagePath)
+```
+
+"Optional: Path at which the file to which the container's termination message will be written is mounted into the container's filesystem. Message written is intended to be brief final status, such as an assertion failure message. Will be truncated by the node if greater than 4096 bytes. The total message length across all containers will be limited to 12kb. Defaults to /dev/termination-log. Cannot be updated."
+
+### fn spec.template.spec.initContainers.withTerminationMessagePolicy
+
+```ts
+withTerminationMessagePolicy(terminationMessagePolicy)
+```
+
+"Indicate how the termination message should be populated. File will use the contents of terminationMessagePath to populate the container status message on both success and failure. FallbackToLogsOnError will use the last chunk of container log output if the termination message file is empty and the container exited with an error. The log output is limited to 2048 bytes or 80 lines, whichever is smaller. Defaults to File. Cannot be updated."
+
+### fn spec.template.spec.initContainers.withTty
+
+```ts
+withTty(tty)
+```
+
+"Whether this container should allocate a TTY for itself, also requires 'stdin' to be true. Default is false."
+
+### fn spec.template.spec.initContainers.withVolumeDevices
+
+```ts
+withVolumeDevices(volumeDevices)
+```
+
+"volumeDevices is the list of block devices to be used by the container."
+
+### fn spec.template.spec.initContainers.withVolumeDevicesMixin
+
+```ts
+withVolumeDevicesMixin(volumeDevices)
+```
+
+"volumeDevices is the list of block devices to be used by the container."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.initContainers.withVolumeMounts
+
+```ts
+withVolumeMounts(volumeMounts)
+```
+
+"Pod volumes to mount into the container's filesystem. Cannot be updated."
+
+### fn spec.template.spec.initContainers.withVolumeMountsMixin
+
+```ts
+withVolumeMountsMixin(volumeMounts)
+```
+
+"Pod volumes to mount into the container's filesystem. Cannot be updated."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.initContainers.withWorkingDir
+
+```ts
+withWorkingDir(workingDir)
+```
+
+"Container's working directory. If not specified, the container runtime's default will be used, which might be configured in the container image. Cannot be updated."
+
+## obj spec.template.spec.initContainers.env
+
+"List of environment variables to set in the container. Cannot be updated."
+
+### fn spec.template.spec.initContainers.env.withName
+
+```ts
+withName(name)
+```
+
+"Name of the environment variable. Must be a C_IDENTIFIER."
+
+### fn spec.template.spec.initContainers.env.withValue
+
+```ts
+withValue(value)
+```
+
+"Variable references $(VAR_NAME) are expanded using the previously defined environment variables in the container and any service environment variables. If a variable cannot be resolved, the reference in the input string will be unchanged. Double $$ are reduced to a single $, which allows for escaping the $(VAR_NAME) syntax: i.e. \"$$(VAR_NAME)\" will produce the string literal \"$(VAR_NAME)\". Escaped references will never be expanded, regardless of whether the variable exists or not. Defaults to \"\"."
+
+## obj spec.template.spec.initContainers.env.valueFrom
+
+"Source for the environment variable's value. Cannot be used if value is not empty."
+
+## obj spec.template.spec.initContainers.env.valueFrom.configMapKeyRef
+
+"Selects a key of a ConfigMap."
+
+### fn spec.template.spec.initContainers.env.valueFrom.configMapKeyRef.withKey
+
+```ts
+withKey(key)
+```
+
+"The key to select."
+
+### fn spec.template.spec.initContainers.env.valueFrom.configMapKeyRef.withName
+
+```ts
+withName(name)
+```
+
+"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"
+
+### fn spec.template.spec.initContainers.env.valueFrom.configMapKeyRef.withOptional
+
+```ts
+withOptional(optional)
+```
+
+"Specify whether the ConfigMap or its key must be defined"
+
+## obj spec.template.spec.initContainers.env.valueFrom.fieldRef
+
+"Selects a field of the pod: supports metadata.name, metadata.namespace, `metadata.labels['']`, `metadata.annotations['']`, spec.nodeName, spec.serviceAccountName, status.hostIP, status.podIP, status.podIPs."
+
+### fn spec.template.spec.initContainers.env.valueFrom.fieldRef.withApiVersion
+
+```ts
+withApiVersion(apiVersion)
+```
+
+"Version of the schema the FieldPath is written in terms of, defaults to \"v1\"."
+
+### fn spec.template.spec.initContainers.env.valueFrom.fieldRef.withFieldPath
+
+```ts
+withFieldPath(fieldPath)
+```
+
+"Path of the field to select in the specified API version."
+
+## obj spec.template.spec.initContainers.env.valueFrom.resourceFieldRef
+
+"Selects a resource of the container: only resources limits and requests (limits.cpu, limits.memory, limits.ephemeral-storage, requests.cpu, requests.memory and requests.ephemeral-storage) are currently supported."
+
+### fn spec.template.spec.initContainers.env.valueFrom.resourceFieldRef.withContainerName
+
+```ts
+withContainerName(containerName)
+```
+
+"Container name: required for volumes, optional for env vars"
+
+### fn spec.template.spec.initContainers.env.valueFrom.resourceFieldRef.withDivisor
+
+```ts
+withDivisor(divisor)
+```
+
+"Specifies the output format of the exposed resources, defaults to \"1\
+
+### fn spec.template.spec.initContainers.env.valueFrom.resourceFieldRef.withResource
+
+```ts
+withResource(resource)
+```
+
+"Required: resource to select"
+
+## obj spec.template.spec.initContainers.env.valueFrom.secretKeyRef
+
+"Selects a key of a secret in the pod's namespace"
+
+### fn spec.template.spec.initContainers.env.valueFrom.secretKeyRef.withKey
+
+```ts
+withKey(key)
+```
+
+"The key of the secret to select from. Must be a valid secret key."
+
+### fn spec.template.spec.initContainers.env.valueFrom.secretKeyRef.withName
+
+```ts
+withName(name)
+```
+
+"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"
+
+### fn spec.template.spec.initContainers.env.valueFrom.secretKeyRef.withOptional
+
+```ts
+withOptional(optional)
+```
+
+"Specify whether the Secret or its key must be defined"
+
+## obj spec.template.spec.initContainers.envFrom
+
+"List of sources to populate environment variables in the container. The keys defined within a source must be a C_IDENTIFIER. All invalid keys will be reported as an event when the container is starting. When a key exists in multiple sources, the value associated with the last source will take precedence. Values defined by an Env with a duplicate key will take precedence. Cannot be updated."
+
+### fn spec.template.spec.initContainers.envFrom.withPrefix
+
+```ts
+withPrefix(prefix)
+```
+
+"An optional identifier to prepend to each key in the ConfigMap. Must be a C_IDENTIFIER."
+
+## obj spec.template.spec.initContainers.envFrom.configMapRef
+
+"The ConfigMap to select from"
+
+### fn spec.template.spec.initContainers.envFrom.configMapRef.withName
+
+```ts
+withName(name)
+```
+
+"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"
+
+### fn spec.template.spec.initContainers.envFrom.configMapRef.withOptional
+
+```ts
+withOptional(optional)
+```
+
+"Specify whether the ConfigMap must be defined"
+
+## obj spec.template.spec.initContainers.envFrom.secretRef
+
+"The Secret to select from"
+
+### fn spec.template.spec.initContainers.envFrom.secretRef.withName
+
+```ts
+withName(name)
+```
+
+"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"
+
+### fn spec.template.spec.initContainers.envFrom.secretRef.withOptional
+
+```ts
+withOptional(optional)
+```
+
+"Specify whether the Secret must be defined"
+
+## obj spec.template.spec.initContainers.lifecycle
+
+"Actions that the management system should take in response to container lifecycle events. Cannot be updated."
+
+## obj spec.template.spec.initContainers.lifecycle.postStart
+
+"PostStart is called immediately after a container is created. If the handler fails, the container is terminated and restarted according to its restart policy. Other management of the container blocks until the hook completes. More info: https://kubernetes.io/docs/concepts/containers/container-lifecycle-hooks/#container-hooks"
+
+## obj spec.template.spec.initContainers.lifecycle.postStart.exec
+
+"Exec specifies the action to take."
+
+### fn spec.template.spec.initContainers.lifecycle.postStart.exec.withCommand
+
+```ts
+withCommand(command)
+```
+
+"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy."
+
+### fn spec.template.spec.initContainers.lifecycle.postStart.exec.withCommandMixin
+
+```ts
+withCommandMixin(command)
+```
+
+"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy."
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.template.spec.initContainers.lifecycle.postStart.httpGet
+
+"HTTPGet specifies the http request to perform."
+
+### fn spec.template.spec.initContainers.lifecycle.postStart.httpGet.withHost
+
+```ts
+withHost(host)
+```
+
+"Host name to connect to, defaults to the pod IP. You probably want to set \"Host\" in httpHeaders instead."
+
+### fn spec.template.spec.initContainers.lifecycle.postStart.httpGet.withHttpHeaders
+
+```ts
+withHttpHeaders(httpHeaders)
+```
+
+"Custom headers to set in the request. HTTP allows repeated headers."
+
+### fn spec.template.spec.initContainers.lifecycle.postStart.httpGet.withHttpHeadersMixin
+
+```ts
+withHttpHeadersMixin(httpHeaders)
+```
+
+"Custom headers to set in the request. HTTP allows repeated headers."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.initContainers.lifecycle.postStart.httpGet.withPath
+
+```ts
+withPath(path)
+```
+
+"Path to access on the HTTP server."
+
+### fn spec.template.spec.initContainers.lifecycle.postStart.httpGet.withPort
+
+```ts
+withPort(port)
+```
+
+"Name or number of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."
+
+### fn spec.template.spec.initContainers.lifecycle.postStart.httpGet.withScheme
+
+```ts
+withScheme(scheme)
+```
+
+"Scheme to use for connecting to the host. Defaults to HTTP."
+
+## obj spec.template.spec.initContainers.lifecycle.postStart.httpGet.httpHeaders
+
+"Custom headers to set in the request. HTTP allows repeated headers."
+
+### fn spec.template.spec.initContainers.lifecycle.postStart.httpGet.httpHeaders.withName
+
+```ts
+withName(name)
+```
+
+"The header field name. This will be canonicalized upon output, so case-variant names will be understood as the same header."
+
+### fn spec.template.spec.initContainers.lifecycle.postStart.httpGet.httpHeaders.withValue
+
+```ts
+withValue(value)
+```
+
+"The header field value"
+
+## obj spec.template.spec.initContainers.lifecycle.postStart.tcpSocket
+
+"Deprecated. TCPSocket is NOT supported as a LifecycleHandler and kept for the backward compatibility. There are no validation of this field and lifecycle hooks will fail in runtime when tcp handler is specified."
+
+### fn spec.template.spec.initContainers.lifecycle.postStart.tcpSocket.withHost
+
+```ts
+withHost(host)
+```
+
+"Optional: Host name to connect to, defaults to the pod IP."
+
+### fn spec.template.spec.initContainers.lifecycle.postStart.tcpSocket.withPort
+
+```ts
+withPort(port)
+```
+
+"Number or name of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."
+
+## obj spec.template.spec.initContainers.lifecycle.preStop
+
+"PreStop is called immediately before a container is terminated due to an API request or management event such as liveness/startup probe failure, preemption, resource contention, etc. The handler is not called if the container crashes or exits. The Pod's termination grace period countdown begins before the PreStop hook is executed. Regardless of the outcome of the handler, the container will eventually terminate within the Pod's termination grace period (unless delayed by finalizers). Other management of the container blocks until the hook completes or until the termination grace period is reached. More info: https://kubernetes.io/docs/concepts/containers/container-lifecycle-hooks/#container-hooks"
+
+## obj spec.template.spec.initContainers.lifecycle.preStop.exec
+
+"Exec specifies the action to take."
+
+### fn spec.template.spec.initContainers.lifecycle.preStop.exec.withCommand
+
+```ts
+withCommand(command)
+```
+
+"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy."
+
+### fn spec.template.spec.initContainers.lifecycle.preStop.exec.withCommandMixin
+
+```ts
+withCommandMixin(command)
+```
+
+"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy."
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.template.spec.initContainers.lifecycle.preStop.httpGet
+
+"HTTPGet specifies the http request to perform."
+
+### fn spec.template.spec.initContainers.lifecycle.preStop.httpGet.withHost
+
+```ts
+withHost(host)
+```
+
+"Host name to connect to, defaults to the pod IP. You probably want to set \"Host\" in httpHeaders instead."
+
+### fn spec.template.spec.initContainers.lifecycle.preStop.httpGet.withHttpHeaders
+
+```ts
+withHttpHeaders(httpHeaders)
+```
+
+"Custom headers to set in the request. HTTP allows repeated headers."
+
+### fn spec.template.spec.initContainers.lifecycle.preStop.httpGet.withHttpHeadersMixin
+
+```ts
+withHttpHeadersMixin(httpHeaders)
+```
+
+"Custom headers to set in the request. HTTP allows repeated headers."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.initContainers.lifecycle.preStop.httpGet.withPath
+
+```ts
+withPath(path)
+```
+
+"Path to access on the HTTP server."
+
+### fn spec.template.spec.initContainers.lifecycle.preStop.httpGet.withPort
+
+```ts
+withPort(port)
+```
+
+"Name or number of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."
+
+### fn spec.template.spec.initContainers.lifecycle.preStop.httpGet.withScheme
+
+```ts
+withScheme(scheme)
+```
+
+"Scheme to use for connecting to the host. Defaults to HTTP."
+
+## obj spec.template.spec.initContainers.lifecycle.preStop.httpGet.httpHeaders
+
+"Custom headers to set in the request. HTTP allows repeated headers."
+
+### fn spec.template.spec.initContainers.lifecycle.preStop.httpGet.httpHeaders.withName
+
+```ts
+withName(name)
+```
+
+"The header field name. This will be canonicalized upon output, so case-variant names will be understood as the same header."
+
+### fn spec.template.spec.initContainers.lifecycle.preStop.httpGet.httpHeaders.withValue
+
+```ts
+withValue(value)
+```
+
+"The header field value"
+
+## obj spec.template.spec.initContainers.lifecycle.preStop.tcpSocket
+
+"Deprecated. TCPSocket is NOT supported as a LifecycleHandler and kept for the backward compatibility. There are no validation of this field and lifecycle hooks will fail in runtime when tcp handler is specified."
+
+### fn spec.template.spec.initContainers.lifecycle.preStop.tcpSocket.withHost
+
+```ts
+withHost(host)
+```
+
+"Optional: Host name to connect to, defaults to the pod IP."
+
+### fn spec.template.spec.initContainers.lifecycle.preStop.tcpSocket.withPort
+
+```ts
+withPort(port)
+```
+
+"Number or name of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."
+
+## obj spec.template.spec.initContainers.livenessProbe
+
+"Periodic probe of container liveness. Container will be restarted if the probe fails. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes"
+
+### fn spec.template.spec.initContainers.livenessProbe.withFailureThreshold
+
+```ts
+withFailureThreshold(failureThreshold)
+```
+
+"Minimum consecutive failures for the probe to be considered failed after having succeeded. Defaults to 3. Minimum value is 1."
+
+### fn spec.template.spec.initContainers.livenessProbe.withInitialDelaySeconds
+
+```ts
+withInitialDelaySeconds(initialDelaySeconds)
+```
+
+"Number of seconds after the container has started before liveness probes are initiated. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes"
+
+### fn spec.template.spec.initContainers.livenessProbe.withPeriodSeconds
+
+```ts
+withPeriodSeconds(periodSeconds)
+```
+
+"How often (in seconds) to perform the probe. Default to 10 seconds. Minimum value is 1."
+
+### fn spec.template.spec.initContainers.livenessProbe.withSuccessThreshold
+
+```ts
+withSuccessThreshold(successThreshold)
+```
+
+"Minimum consecutive successes for the probe to be considered successful after having failed. Defaults to 1. Must be 1 for liveness and startup. Minimum value is 1."
+
+### fn spec.template.spec.initContainers.livenessProbe.withTerminationGracePeriodSeconds
+
+```ts
+withTerminationGracePeriodSeconds(terminationGracePeriodSeconds)
+```
+
+"Optional duration in seconds the pod needs to terminate gracefully upon probe failure. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. If this value is nil, the pod's terminationGracePeriodSeconds will be used. Otherwise, this value overrides the value provided by the pod spec. Value must be non-negative integer. The value zero indicates stop immediately via the kill signal (no opportunity to shut down). This is a beta field and requires enabling ProbeTerminationGracePeriod feature gate. Minimum value is 1. spec.terminationGracePeriodSeconds is used if unset."
+
+### fn spec.template.spec.initContainers.livenessProbe.withTimeoutSeconds
+
+```ts
+withTimeoutSeconds(timeoutSeconds)
+```
+
+"Number of seconds after which the probe times out. Defaults to 1 second. Minimum value is 1. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes"
+
+## obj spec.template.spec.initContainers.livenessProbe.exec
+
+"Exec specifies the action to take."
+
+### fn spec.template.spec.initContainers.livenessProbe.exec.withCommand
+
+```ts
+withCommand(command)
+```
+
+"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy."
+
+### fn spec.template.spec.initContainers.livenessProbe.exec.withCommandMixin
+
+```ts
+withCommandMixin(command)
+```
+
+"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy."
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.template.spec.initContainers.livenessProbe.grpc
+
+"GRPC specifies an action involving a GRPC port."
+
+### fn spec.template.spec.initContainers.livenessProbe.grpc.withPort
+
+```ts
+withPort(port)
+```
+
+"Port number of the gRPC service. Number must be in the range 1 to 65535."
+
+### fn spec.template.spec.initContainers.livenessProbe.grpc.withService
+
+```ts
+withService(service)
+```
+
+"Service is the name of the service to place in the gRPC HealthCheckRequest (see https://github.com/grpc/grpc/blob/master/doc/health-checking.md). \n If this is not specified, the default behavior is defined by gRPC."
+
+## obj spec.template.spec.initContainers.livenessProbe.httpGet
+
+"HTTPGet specifies the http request to perform."
+
+### fn spec.template.spec.initContainers.livenessProbe.httpGet.withHost
+
+```ts
+withHost(host)
+```
+
+"Host name to connect to, defaults to the pod IP. You probably want to set \"Host\" in httpHeaders instead."
+
+### fn spec.template.spec.initContainers.livenessProbe.httpGet.withHttpHeaders
+
+```ts
+withHttpHeaders(httpHeaders)
+```
+
+"Custom headers to set in the request. HTTP allows repeated headers."
+
+### fn spec.template.spec.initContainers.livenessProbe.httpGet.withHttpHeadersMixin
+
+```ts
+withHttpHeadersMixin(httpHeaders)
+```
+
+"Custom headers to set in the request. HTTP allows repeated headers."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.initContainers.livenessProbe.httpGet.withPath
+
+```ts
+withPath(path)
+```
+
+"Path to access on the HTTP server."
+
+### fn spec.template.spec.initContainers.livenessProbe.httpGet.withPort
+
+```ts
+withPort(port)
+```
+
+"Name or number of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."
+
+### fn spec.template.spec.initContainers.livenessProbe.httpGet.withScheme
+
+```ts
+withScheme(scheme)
+```
+
+"Scheme to use for connecting to the host. Defaults to HTTP."
+
+## obj spec.template.spec.initContainers.livenessProbe.httpGet.httpHeaders
+
+"Custom headers to set in the request. HTTP allows repeated headers."
+
+### fn spec.template.spec.initContainers.livenessProbe.httpGet.httpHeaders.withName
+
+```ts
+withName(name)
+```
+
+"The header field name. This will be canonicalized upon output, so case-variant names will be understood as the same header."
+
+### fn spec.template.spec.initContainers.livenessProbe.httpGet.httpHeaders.withValue
+
+```ts
+withValue(value)
+```
+
+"The header field value"
+
+## obj spec.template.spec.initContainers.livenessProbe.tcpSocket
+
+"TCPSocket specifies an action involving a TCP port."
+
+### fn spec.template.spec.initContainers.livenessProbe.tcpSocket.withHost
+
+```ts
+withHost(host)
+```
+
+"Optional: Host name to connect to, defaults to the pod IP."
+
+### fn spec.template.spec.initContainers.livenessProbe.tcpSocket.withPort
+
+```ts
+withPort(port)
+```
+
+"Number or name of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."
+
+## obj spec.template.spec.initContainers.ports
+
+"List of ports to expose from the container. Not specifying a port here DOES NOT prevent that port from being exposed. Any port which is listening on the default \"0.0.0.0\" address inside a container will be accessible from the network. Modifying this array with strategic merge patch may corrupt the data. For more information See https://github.com/kubernetes/kubernetes/issues/108255. Cannot be updated."
+
+### fn spec.template.spec.initContainers.ports.withContainerPort
+
+```ts
+withContainerPort(containerPort)
+```
+
+"Number of port to expose on the pod's IP address. This must be a valid port number, 0 < x < 65536."
+
+### fn spec.template.spec.initContainers.ports.withHostIP
+
+```ts
+withHostIP(hostIP)
+```
+
+"What host IP to bind the external port to."
+
+### fn spec.template.spec.initContainers.ports.withHostPort
+
+```ts
+withHostPort(hostPort)
+```
+
+"Number of port to expose on the host. If specified, this must be a valid port number, 0 < x < 65536. If HostNetwork is specified, this must match ContainerPort. Most containers do not need this."
+
+### fn spec.template.spec.initContainers.ports.withName
+
+```ts
+withName(name)
+```
+
+"If specified, this must be an IANA_SVC_NAME and unique within the pod. Each named port in a pod must have a unique name. Name for the port that can be referred to by services."
+
+### fn spec.template.spec.initContainers.ports.withProtocol
+
+```ts
+withProtocol(protocol)
+```
+
+"Protocol for port. Must be UDP, TCP, or SCTP. Defaults to \"TCP\"."
+
+## obj spec.template.spec.initContainers.readinessProbe
+
+"Periodic probe of container service readiness. Container will be removed from service endpoints if the probe fails. Cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes"
+
+### fn spec.template.spec.initContainers.readinessProbe.withFailureThreshold
+
+```ts
+withFailureThreshold(failureThreshold)
+```
+
+"Minimum consecutive failures for the probe to be considered failed after having succeeded. Defaults to 3. Minimum value is 1."
+
+### fn spec.template.spec.initContainers.readinessProbe.withInitialDelaySeconds
+
+```ts
+withInitialDelaySeconds(initialDelaySeconds)
+```
+
+"Number of seconds after the container has started before liveness probes are initiated. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes"
+
+### fn spec.template.spec.initContainers.readinessProbe.withPeriodSeconds
+
+```ts
+withPeriodSeconds(periodSeconds)
+```
+
+"How often (in seconds) to perform the probe. Default to 10 seconds. Minimum value is 1."
+
+### fn spec.template.spec.initContainers.readinessProbe.withSuccessThreshold
+
+```ts
+withSuccessThreshold(successThreshold)
+```
+
+"Minimum consecutive successes for the probe to be considered successful after having failed. Defaults to 1. Must be 1 for liveness and startup. Minimum value is 1."
+
+### fn spec.template.spec.initContainers.readinessProbe.withTerminationGracePeriodSeconds
+
+```ts
+withTerminationGracePeriodSeconds(terminationGracePeriodSeconds)
+```
+
+"Optional duration in seconds the pod needs to terminate gracefully upon probe failure. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. If this value is nil, the pod's terminationGracePeriodSeconds will be used. Otherwise, this value overrides the value provided by the pod spec. Value must be non-negative integer. The value zero indicates stop immediately via the kill signal (no opportunity to shut down). This is a beta field and requires enabling ProbeTerminationGracePeriod feature gate. Minimum value is 1. spec.terminationGracePeriodSeconds is used if unset."
+
+### fn spec.template.spec.initContainers.readinessProbe.withTimeoutSeconds
+
+```ts
+withTimeoutSeconds(timeoutSeconds)
+```
+
+"Number of seconds after which the probe times out. Defaults to 1 second. Minimum value is 1. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes"
+
+## obj spec.template.spec.initContainers.readinessProbe.exec
+
+"Exec specifies the action to take."
+
+### fn spec.template.spec.initContainers.readinessProbe.exec.withCommand
+
+```ts
+withCommand(command)
+```
+
+"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy."
+
+### fn spec.template.spec.initContainers.readinessProbe.exec.withCommandMixin
+
+```ts
+withCommandMixin(command)
+```
+
+"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy."
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.template.spec.initContainers.readinessProbe.grpc
+
+"GRPC specifies an action involving a GRPC port."
+
+### fn spec.template.spec.initContainers.readinessProbe.grpc.withPort
+
+```ts
+withPort(port)
+```
+
+"Port number of the gRPC service. Number must be in the range 1 to 65535."
+
+### fn spec.template.spec.initContainers.readinessProbe.grpc.withService
+
+```ts
+withService(service)
+```
+
+"Service is the name of the service to place in the gRPC HealthCheckRequest (see https://github.com/grpc/grpc/blob/master/doc/health-checking.md). \n If this is not specified, the default behavior is defined by gRPC."
+
+## obj spec.template.spec.initContainers.readinessProbe.httpGet
+
+"HTTPGet specifies the http request to perform."
+
+### fn spec.template.spec.initContainers.readinessProbe.httpGet.withHost
+
+```ts
+withHost(host)
+```
+
+"Host name to connect to, defaults to the pod IP. You probably want to set \"Host\" in httpHeaders instead."
+
+### fn spec.template.spec.initContainers.readinessProbe.httpGet.withHttpHeaders
+
+```ts
+withHttpHeaders(httpHeaders)
+```
+
+"Custom headers to set in the request. HTTP allows repeated headers."
+
+### fn spec.template.spec.initContainers.readinessProbe.httpGet.withHttpHeadersMixin
+
+```ts
+withHttpHeadersMixin(httpHeaders)
+```
+
+"Custom headers to set in the request. HTTP allows repeated headers."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.initContainers.readinessProbe.httpGet.withPath
+
+```ts
+withPath(path)
+```
+
+"Path to access on the HTTP server."
+
+### fn spec.template.spec.initContainers.readinessProbe.httpGet.withPort
+
+```ts
+withPort(port)
+```
+
+"Name or number of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."
+
+### fn spec.template.spec.initContainers.readinessProbe.httpGet.withScheme
+
+```ts
+withScheme(scheme)
+```
+
+"Scheme to use for connecting to the host. Defaults to HTTP."
+
+## obj spec.template.spec.initContainers.readinessProbe.httpGet.httpHeaders
+
+"Custom headers to set in the request. HTTP allows repeated headers."
+
+### fn spec.template.spec.initContainers.readinessProbe.httpGet.httpHeaders.withName
+
+```ts
+withName(name)
+```
+
+"The header field name. This will be canonicalized upon output, so case-variant names will be understood as the same header."
+
+### fn spec.template.spec.initContainers.readinessProbe.httpGet.httpHeaders.withValue
+
+```ts
+withValue(value)
+```
+
+"The header field value"
+
+## obj spec.template.spec.initContainers.readinessProbe.tcpSocket
+
+"TCPSocket specifies an action involving a TCP port."
+
+### fn spec.template.spec.initContainers.readinessProbe.tcpSocket.withHost
+
+```ts
+withHost(host)
+```
+
+"Optional: Host name to connect to, defaults to the pod IP."
+
+### fn spec.template.spec.initContainers.readinessProbe.tcpSocket.withPort
+
+```ts
+withPort(port)
+```
+
+"Number or name of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."
+
+## obj spec.template.spec.initContainers.resizePolicy
+
+"Resources resize policy for the container."
+
+### fn spec.template.spec.initContainers.resizePolicy.withResourceName
+
+```ts
+withResourceName(resourceName)
+```
+
+"Name of the resource to which this resource resize policy applies. Supported values: cpu, memory."
+
+### fn spec.template.spec.initContainers.resizePolicy.withRestartPolicy
+
+```ts
+withRestartPolicy(restartPolicy)
+```
+
+"Restart policy to apply when specified resource is resized. If not specified, it defaults to NotRequired."
+
+## obj spec.template.spec.initContainers.resources
+
+"Compute Resources required by this container. Cannot be updated. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"
+
+### fn spec.template.spec.initContainers.resources.withClaims
+
+```ts
+withClaims(claims)
+```
+
+"Claims lists the names of resources, defined in spec.resourceClaims, that are used by this container. \n This is an alpha field and requires enabling the DynamicResourceAllocation feature gate. \n This field is immutable. It can only be set for containers."
+
+### fn spec.template.spec.initContainers.resources.withClaimsMixin
+
+```ts
+withClaimsMixin(claims)
+```
+
+"Claims lists the names of resources, defined in spec.resourceClaims, that are used by this container. \n This is an alpha field and requires enabling the DynamicResourceAllocation feature gate. \n This field is immutable. It can only be set for containers."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.initContainers.resources.withLimits
+
+```ts
+withLimits(limits)
+```
+
+"Limits describes the maximum amount of compute resources allowed. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"
+
+### fn spec.template.spec.initContainers.resources.withLimitsMixin
+
+```ts
+withLimitsMixin(limits)
+```
+
+"Limits describes the maximum amount of compute resources allowed. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.initContainers.resources.withRequests
+
+```ts
+withRequests(requests)
+```
+
+"Requests describes the minimum amount of compute resources required. If Requests is omitted for a container, it defaults to Limits if that is explicitly specified, otherwise to an implementation-defined value. Requests cannot exceed Limits. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"
+
+### fn spec.template.spec.initContainers.resources.withRequestsMixin
+
+```ts
+withRequestsMixin(requests)
+```
+
+"Requests describes the minimum amount of compute resources required. If Requests is omitted for a container, it defaults to Limits if that is explicitly specified, otherwise to an implementation-defined value. Requests cannot exceed Limits. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.template.spec.initContainers.resources.claims
+
+"Claims lists the names of resources, defined in spec.resourceClaims, that are used by this container. \n This is an alpha field and requires enabling the DynamicResourceAllocation feature gate. \n This field is immutable. It can only be set for containers."
+
+### fn spec.template.spec.initContainers.resources.claims.withName
+
+```ts
+withName(name)
+```
+
+"Name must match the name of one entry in pod.spec.resourceClaims of the Pod where this field is used. It makes that resource available inside a container."
+
+## obj spec.template.spec.initContainers.securityContext
+
+"SecurityContext defines the security options the container should be run with. If set, the fields of SecurityContext override the equivalent fields of PodSecurityContext. More info: https://kubernetes.io/docs/tasks/configure-pod-container/security-context/"
+
+### fn spec.template.spec.initContainers.securityContext.withAllowPrivilegeEscalation
+
+```ts
+withAllowPrivilegeEscalation(allowPrivilegeEscalation)
+```
+
+"AllowPrivilegeEscalation controls whether a process can gain more privileges than its parent process. This bool directly controls if the no_new_privs flag will be set on the container process. AllowPrivilegeEscalation is true always when the container is: 1) run as Privileged 2) has CAP_SYS_ADMIN Note that this field cannot be set when spec.os.name is windows."
+
+### fn spec.template.spec.initContainers.securityContext.withPrivileged
+
+```ts
+withPrivileged(privileged)
+```
+
+"Run container in privileged mode. Processes in privileged containers are essentially equivalent to root on the host. Defaults to false. Note that this field cannot be set when spec.os.name is windows."
+
+### fn spec.template.spec.initContainers.securityContext.withProcMount
+
+```ts
+withProcMount(procMount)
+```
+
+"procMount denotes the type of proc mount to use for the containers. The default is DefaultProcMount which uses the container runtime defaults for readonly paths and masked paths. This requires the ProcMountType feature flag to be enabled. Note that this field cannot be set when spec.os.name is windows."
+
+### fn spec.template.spec.initContainers.securityContext.withReadOnlyRootFilesystem
+
+```ts
+withReadOnlyRootFilesystem(readOnlyRootFilesystem)
+```
+
+"Whether this container has a read-only root filesystem. Default is false. Note that this field cannot be set when spec.os.name is windows."
+
+### fn spec.template.spec.initContainers.securityContext.withRunAsGroup
+
+```ts
+withRunAsGroup(runAsGroup)
+```
+
+"The GID to run the entrypoint of the container process. Uses runtime default if unset. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. Note that this field cannot be set when spec.os.name is windows."
+
+### fn spec.template.spec.initContainers.securityContext.withRunAsNonRoot
+
+```ts
+withRunAsNonRoot(runAsNonRoot)
+```
+
+"Indicates that the container must run as a non-root user. If true, the Kubelet will validate the image at runtime to ensure that it does not run as UID 0 (root) and fail to start the container if it does. If unset or false, no such validation will be performed. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence."
+
+### fn spec.template.spec.initContainers.securityContext.withRunAsUser
+
+```ts
+withRunAsUser(runAsUser)
+```
+
+"The UID to run the entrypoint of the container process. Defaults to user specified in image metadata if unspecified. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. Note that this field cannot be set when spec.os.name is windows."
+
+## obj spec.template.spec.initContainers.securityContext.capabilities
+
+"The capabilities to add/drop when running containers. Defaults to the default set of capabilities granted by the container runtime. Note that this field cannot be set when spec.os.name is windows."
+
+### fn spec.template.spec.initContainers.securityContext.capabilities.withAdd
+
+```ts
+withAdd(add)
+```
+
+"Added capabilities"
+
+### fn spec.template.spec.initContainers.securityContext.capabilities.withAddMixin
+
+```ts
+withAddMixin(add)
+```
+
+"Added capabilities"
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.initContainers.securityContext.capabilities.withDrop
+
+```ts
+withDrop(drop)
+```
+
+"Removed capabilities"
+
+### fn spec.template.spec.initContainers.securityContext.capabilities.withDropMixin
+
+```ts
+withDropMixin(drop)
+```
+
+"Removed capabilities"
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.template.spec.initContainers.securityContext.seLinuxOptions
+
+"The SELinux context to be applied to the container. If unspecified, the container runtime will allocate a random SELinux context for each container. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. Note that this field cannot be set when spec.os.name is windows."
+
+### fn spec.template.spec.initContainers.securityContext.seLinuxOptions.withLevel
+
+```ts
+withLevel(level)
+```
+
+"Level is SELinux level label that applies to the container."
+
+### fn spec.template.spec.initContainers.securityContext.seLinuxOptions.withRole
+
+```ts
+withRole(role)
+```
+
+"Role is a SELinux role label that applies to the container."
+
+### fn spec.template.spec.initContainers.securityContext.seLinuxOptions.withType
+
+```ts
+withType(type)
+```
+
+"Type is a SELinux type label that applies to the container."
+
+### fn spec.template.spec.initContainers.securityContext.seLinuxOptions.withUser
+
+```ts
+withUser(user)
+```
+
+"User is a SELinux user label that applies to the container."
+
+## obj spec.template.spec.initContainers.securityContext.seccompProfile
+
+"The seccomp options to use by this container. If seccomp options are provided at both the pod & container level, the container options override the pod options. Note that this field cannot be set when spec.os.name is windows."
+
+### fn spec.template.spec.initContainers.securityContext.seccompProfile.withLocalhostProfile
+
+```ts
+withLocalhostProfile(localhostProfile)
+```
+
+"localhostProfile indicates a profile defined in a file on the node should be used. The profile must be preconfigured on the node to work. Must be a descending path, relative to the kubelet's configured seccomp profile location. Must be set if type is \"Localhost\". Must NOT be set for any other type."
+
+### fn spec.template.spec.initContainers.securityContext.seccompProfile.withType
+
+```ts
+withType(type)
+```
+
+"type indicates which kind of seccomp profile will be applied. Valid options are: \n Localhost - a profile defined in a file on the node should be used. RuntimeDefault - the container runtime default profile should be used. Unconfined - no profile should be applied."
+
+## obj spec.template.spec.initContainers.securityContext.windowsOptions
+
+"The Windows specific settings applied to all containers. If unspecified, the options from the PodSecurityContext will be used. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. Note that this field cannot be set when spec.os.name is linux."
+
+### fn spec.template.spec.initContainers.securityContext.windowsOptions.withGmsaCredentialSpec
+
+```ts
+withGmsaCredentialSpec(gmsaCredentialSpec)
+```
+
+"GMSACredentialSpec is where the GMSA admission webhook (https://github.com/kubernetes-sigs/windows-gmsa) inlines the contents of the GMSA credential spec named by the GMSACredentialSpecName field."
+
+### fn spec.template.spec.initContainers.securityContext.windowsOptions.withGmsaCredentialSpecName
+
+```ts
+withGmsaCredentialSpecName(gmsaCredentialSpecName)
+```
+
+"GMSACredentialSpecName is the name of the GMSA credential spec to use."
+
+### fn spec.template.spec.initContainers.securityContext.windowsOptions.withHostProcess
+
+```ts
+withHostProcess(hostProcess)
+```
+
+"HostProcess determines if a container should be run as a 'Host Process' container. All of a Pod's containers must have the same effective HostProcess value (it is not allowed to have a mix of HostProcess containers and non-HostProcess containers). In addition, if HostProcess is true then HostNetwork must also be set to true."
+
+### fn spec.template.spec.initContainers.securityContext.windowsOptions.withRunAsUserName
+
+```ts
+withRunAsUserName(runAsUserName)
+```
+
+"The UserName in Windows to run the entrypoint of the container process. Defaults to the user specified in image metadata if unspecified. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence."
+
+## obj spec.template.spec.initContainers.startupProbe
+
+"StartupProbe indicates that the Pod has successfully initialized. If specified, no other probes are executed until this completes successfully. If this probe fails, the Pod will be restarted, just as if the livenessProbe failed. This can be used to provide different probe parameters at the beginning of a Pod's lifecycle, when it might take a long time to load data or warm a cache, than during steady-state operation. This cannot be updated. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes"
+
+### fn spec.template.spec.initContainers.startupProbe.withFailureThreshold
+
+```ts
+withFailureThreshold(failureThreshold)
+```
+
+"Minimum consecutive failures for the probe to be considered failed after having succeeded. Defaults to 3. Minimum value is 1."
+
+### fn spec.template.spec.initContainers.startupProbe.withInitialDelaySeconds
+
+```ts
+withInitialDelaySeconds(initialDelaySeconds)
+```
+
+"Number of seconds after the container has started before liveness probes are initiated. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes"
+
+### fn spec.template.spec.initContainers.startupProbe.withPeriodSeconds
+
+```ts
+withPeriodSeconds(periodSeconds)
+```
+
+"How often (in seconds) to perform the probe. Default to 10 seconds. Minimum value is 1."
+
+### fn spec.template.spec.initContainers.startupProbe.withSuccessThreshold
+
+```ts
+withSuccessThreshold(successThreshold)
+```
+
+"Minimum consecutive successes for the probe to be considered successful after having failed. Defaults to 1. Must be 1 for liveness and startup. Minimum value is 1."
+
+### fn spec.template.spec.initContainers.startupProbe.withTerminationGracePeriodSeconds
+
+```ts
+withTerminationGracePeriodSeconds(terminationGracePeriodSeconds)
+```
+
+"Optional duration in seconds the pod needs to terminate gracefully upon probe failure. The grace period is the duration in seconds after the processes running in the pod are sent a termination signal and the time when the processes are forcibly halted with a kill signal. Set this value longer than the expected cleanup time for your process. If this value is nil, the pod's terminationGracePeriodSeconds will be used. Otherwise, this value overrides the value provided by the pod spec. Value must be non-negative integer. The value zero indicates stop immediately via the kill signal (no opportunity to shut down). This is a beta field and requires enabling ProbeTerminationGracePeriod feature gate. Minimum value is 1. spec.terminationGracePeriodSeconds is used if unset."
+
+### fn spec.template.spec.initContainers.startupProbe.withTimeoutSeconds
+
+```ts
+withTimeoutSeconds(timeoutSeconds)
+```
+
+"Number of seconds after which the probe times out. Defaults to 1 second. Minimum value is 1. More info: https://kubernetes.io/docs/concepts/workloads/pods/pod-lifecycle#container-probes"
+
+## obj spec.template.spec.initContainers.startupProbe.exec
+
+"Exec specifies the action to take."
+
+### fn spec.template.spec.initContainers.startupProbe.exec.withCommand
+
+```ts
+withCommand(command)
+```
+
+"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy."
+
+### fn spec.template.spec.initContainers.startupProbe.exec.withCommandMixin
+
+```ts
+withCommandMixin(command)
+```
+
+"Command is the command line to execute inside the container, the working directory for the command is root ('/') in the container's filesystem. The command is simply exec'd, it is not run inside a shell, so traditional shell instructions ('|', etc) won't work. To use a shell, you need to explicitly call out to that shell. Exit status of 0 is treated as live/healthy and non-zero is unhealthy."
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.template.spec.initContainers.startupProbe.grpc
+
+"GRPC specifies an action involving a GRPC port."
+
+### fn spec.template.spec.initContainers.startupProbe.grpc.withPort
+
+```ts
+withPort(port)
+```
+
+"Port number of the gRPC service. Number must be in the range 1 to 65535."
+
+### fn spec.template.spec.initContainers.startupProbe.grpc.withService
+
+```ts
+withService(service)
+```
+
+"Service is the name of the service to place in the gRPC HealthCheckRequest (see https://github.com/grpc/grpc/blob/master/doc/health-checking.md). \n If this is not specified, the default behavior is defined by gRPC."
+
+## obj spec.template.spec.initContainers.startupProbe.httpGet
+
+"HTTPGet specifies the http request to perform."
+
+### fn spec.template.spec.initContainers.startupProbe.httpGet.withHost
+
+```ts
+withHost(host)
+```
+
+"Host name to connect to, defaults to the pod IP. You probably want to set \"Host\" in httpHeaders instead."
+
+### fn spec.template.spec.initContainers.startupProbe.httpGet.withHttpHeaders
+
+```ts
+withHttpHeaders(httpHeaders)
+```
+
+"Custom headers to set in the request. HTTP allows repeated headers."
+
+### fn spec.template.spec.initContainers.startupProbe.httpGet.withHttpHeadersMixin
+
+```ts
+withHttpHeadersMixin(httpHeaders)
+```
+
+"Custom headers to set in the request. HTTP allows repeated headers."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.initContainers.startupProbe.httpGet.withPath
+
+```ts
+withPath(path)
+```
+
+"Path to access on the HTTP server."
+
+### fn spec.template.spec.initContainers.startupProbe.httpGet.withPort
+
+```ts
+withPort(port)
+```
+
+"Name or number of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."
+
+### fn spec.template.spec.initContainers.startupProbe.httpGet.withScheme
+
+```ts
+withScheme(scheme)
+```
+
+"Scheme to use for connecting to the host. Defaults to HTTP."
+
+## obj spec.template.spec.initContainers.startupProbe.httpGet.httpHeaders
+
+"Custom headers to set in the request. HTTP allows repeated headers."
+
+### fn spec.template.spec.initContainers.startupProbe.httpGet.httpHeaders.withName
+
+```ts
+withName(name)
+```
+
+"The header field name. This will be canonicalized upon output, so case-variant names will be understood as the same header."
+
+### fn spec.template.spec.initContainers.startupProbe.httpGet.httpHeaders.withValue
+
+```ts
+withValue(value)
+```
+
+"The header field value"
+
+## obj spec.template.spec.initContainers.startupProbe.tcpSocket
+
+"TCPSocket specifies an action involving a TCP port."
+
+### fn spec.template.spec.initContainers.startupProbe.tcpSocket.withHost
+
+```ts
+withHost(host)
+```
+
+"Optional: Host name to connect to, defaults to the pod IP."
+
+### fn spec.template.spec.initContainers.startupProbe.tcpSocket.withPort
+
+```ts
+withPort(port)
+```
+
+"Number or name of the port to access on the container. Number must be in the range 1 to 65535. Name must be an IANA_SVC_NAME."
+
+## obj spec.template.spec.initContainers.volumeDevices
+
+"volumeDevices is the list of block devices to be used by the container."
+
+### fn spec.template.spec.initContainers.volumeDevices.withDevicePath
+
+```ts
+withDevicePath(devicePath)
+```
+
+"devicePath is the path inside of the container that the device will be mapped to."
+
+### fn spec.template.spec.initContainers.volumeDevices.withName
+
+```ts
+withName(name)
+```
+
+"name must match the name of a persistentVolumeClaim in the pod"
+
+## obj spec.template.spec.initContainers.volumeMounts
+
+"Pod volumes to mount into the container's filesystem. Cannot be updated."
+
+### fn spec.template.spec.initContainers.volumeMounts.withMountPath
+
+```ts
+withMountPath(mountPath)
+```
+
+"Path within the container at which the volume should be mounted. Must not contain ':'."
+
+### fn spec.template.spec.initContainers.volumeMounts.withMountPropagation
+
+```ts
+withMountPropagation(mountPropagation)
+```
+
+"mountPropagation determines how mounts are propagated from the host to container and the other way around. When not set, MountPropagationNone is used. This field is beta in 1.10."
+
+### fn spec.template.spec.initContainers.volumeMounts.withName
+
+```ts
+withName(name)
+```
+
+"This must match the Name of a Volume."
+
+### fn spec.template.spec.initContainers.volumeMounts.withReadOnly
+
+```ts
+withReadOnly(readOnly)
+```
+
+"Mounted read-only if true, read-write otherwise (false or unspecified). Defaults to false."
+
+### fn spec.template.spec.initContainers.volumeMounts.withSubPath
+
+```ts
+withSubPath(subPath)
+```
+
+"Path within the volume from which the container's volume should be mounted. Defaults to \"\" (volume's root)."
+
+### fn spec.template.spec.initContainers.volumeMounts.withSubPathExpr
+
+```ts
+withSubPathExpr(subPathExpr)
+```
+
+"Expanded path within the volume from which the container's volume should be mounted. Behaves similarly to SubPath but environment variable references $(VAR_NAME) are expanded using the container's environment. Defaults to \"\" (volume's root). SubPathExpr and SubPath are mutually exclusive."
+
+## obj spec.template.spec.os
+
+"Specifies the OS of the containers in the pod. Some pod and container fields are restricted if this is set. \n If the OS field is set to linux, the following fields must be unset: -securityContext.windowsOptions \n If the OS field is set to windows, following fields must be unset: - spec.hostPID - spec.hostIPC - spec.hostUsers - spec.securityContext.seLinuxOptions - spec.securityContext.seccompProfile - spec.securityContext.fsGroup - spec.securityContext.fsGroupChangePolicy - spec.securityContext.sysctls - spec.shareProcessNamespace - spec.securityContext.runAsUser - spec.securityContext.runAsGroup - spec.securityContext.supplementalGroups - spec.containers[*].securityContext.seLinuxOptions - spec.containers[*].securityContext.seccompProfile - spec.containers[*].securityContext.capabilities - spec.containers[*].securityContext.readOnlyRootFilesystem - spec.containers[*].securityContext.privileged - spec.containers[*].securityContext.allowPrivilegeEscalation - spec.containers[*].securityContext.procMount - spec.containers[*].securityContext.runAsUser - spec.containers[*].securityContext.runAsGroup"
+
+### fn spec.template.spec.os.withName
+
+```ts
+withName(name)
+```
+
+"Name is the name of the operating system. The currently supported values are linux and windows. Additional value may be defined in future and can be one of: https://github.com/opencontainers/runtime-spec/blob/master/config.md#platform-specific-configuration Clients should expect to handle additional values and treat unrecognized values in this field as os: null"
+
+## obj spec.template.spec.readinessGates
+
+"If specified, all readiness gates will be evaluated for pod readiness. A pod is ready when all its containers are ready AND all conditions specified in the readiness gates have status equal to \"True\" More info: https://git.k8s.io/enhancements/keps/sig-network/580-pod-readiness-gates"
+
+### fn spec.template.spec.readinessGates.withConditionType
+
+```ts
+withConditionType(conditionType)
+```
+
+"ConditionType refers to a condition in the pod's condition list with matching type."
+
+## obj spec.template.spec.resourceClaims
+
+"ResourceClaims defines which ResourceClaims must be allocated and reserved before the Pod is allowed to start. The resources will be made available to those containers which consume them by name. \n This is an alpha field and requires enabling the DynamicResourceAllocation feature gate. \n This field is immutable."
+
+### fn spec.template.spec.resourceClaims.withName
+
+```ts
+withName(name)
+```
+
+"Name uniquely identifies this resource claim inside the pod. This must be a DNS_LABEL."
+
+## obj spec.template.spec.resourceClaims.source
+
+"Source describes where to find the ResourceClaim."
+
+### fn spec.template.spec.resourceClaims.source.withResourceClaimName
+
+```ts
+withResourceClaimName(resourceClaimName)
+```
+
+"ResourceClaimName is the name of a ResourceClaim object in the same namespace as this pod."
+
+### fn spec.template.spec.resourceClaims.source.withResourceClaimTemplateName
+
+```ts
+withResourceClaimTemplateName(resourceClaimTemplateName)
+```
+
+"ResourceClaimTemplateName is the name of a ResourceClaimTemplate object in the same namespace as this pod. \n The template will be used to create a new ResourceClaim, which will be bound to this pod. When this pod is deleted, the ResourceClaim will also be deleted. The pod name and resource name, along with a generated component, will be used to form a unique name for the ResourceClaim, which will be recorded in pod.status.resourceClaimStatuses. \n This field is immutable and no changes will be made to the corresponding ResourceClaim by the control plane after creating the ResourceClaim."
+
+## obj spec.template.spec.schedulingGates
+
+"SchedulingGates is an opaque list of values that if specified will block scheduling the pod. If schedulingGates is not empty, the pod will stay in the SchedulingGated state and the scheduler will not attempt to schedule the pod. \n SchedulingGates can only be set at pod creation time, and be removed only afterwards. \n This is a beta feature enabled by the PodSchedulingReadiness feature gate."
+
+### fn spec.template.spec.schedulingGates.withName
+
+```ts
+withName(name)
+```
+
+"Name of the scheduling gate. Each scheduling gate must have a unique name field."
+
+## obj spec.template.spec.securityContext
+
+"SecurityContext holds pod-level security attributes and common container settings. Optional: Defaults to empty. See type description for default values of each field."
+
+### fn spec.template.spec.securityContext.withFsGroup
+
+```ts
+withFsGroup(fsGroup)
+```
+
+"A special supplemental group that applies to all containers in a pod. Some volume types allow the Kubelet to change the ownership of that volume to be owned by the pod: \n 1. The owning GID will be the FSGroup 2. The setgid bit is set (new files created in the volume will be owned by FSGroup) 3. The permission bits are OR'd with rw-rw---- \n If unset, the Kubelet will not modify the ownership and permissions of any volume. Note that this field cannot be set when spec.os.name is windows."
+
+### fn spec.template.spec.securityContext.withFsGroupChangePolicy
+
+```ts
+withFsGroupChangePolicy(fsGroupChangePolicy)
+```
+
+"fsGroupChangePolicy defines behavior of changing ownership and permission of the volume before being exposed inside Pod. This field will only apply to volume types which support fsGroup based ownership(and permissions). It will have no effect on ephemeral volume types such as: secret, configmaps and emptydir. Valid values are \"OnRootMismatch\" and \"Always\". If not specified, \"Always\" is used. Note that this field cannot be set when spec.os.name is windows."
+
+### fn spec.template.spec.securityContext.withRunAsGroup
+
+```ts
+withRunAsGroup(runAsGroup)
+```
+
+"The GID to run the entrypoint of the container process. Uses runtime default if unset. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. Note that this field cannot be set when spec.os.name is windows."
+
+### fn spec.template.spec.securityContext.withRunAsNonRoot
+
+```ts
+withRunAsNonRoot(runAsNonRoot)
+```
+
+"Indicates that the container must run as a non-root user. If true, the Kubelet will validate the image at runtime to ensure that it does not run as UID 0 (root) and fail to start the container if it does. If unset or false, no such validation will be performed. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence."
+
+### fn spec.template.spec.securityContext.withRunAsUser
+
+```ts
+withRunAsUser(runAsUser)
+```
+
+"The UID to run the entrypoint of the container process. Defaults to user specified in image metadata if unspecified. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. Note that this field cannot be set when spec.os.name is windows."
+
+### fn spec.template.spec.securityContext.withSupplementalGroups
+
+```ts
+withSupplementalGroups(supplementalGroups)
+```
+
+"A list of groups applied to the first process run in each container, in addition to the container's primary GID, the fsGroup (if specified), and group memberships defined in the container image for the uid of the container process. If unspecified, no additional groups are added to any container. Note that group memberships defined in the container image for the uid of the container process are still effective, even if they are not included in this list. Note that this field cannot be set when spec.os.name is windows."
+
+### fn spec.template.spec.securityContext.withSupplementalGroupsMixin
+
+```ts
+withSupplementalGroupsMixin(supplementalGroups)
+```
+
+"A list of groups applied to the first process run in each container, in addition to the container's primary GID, the fsGroup (if specified), and group memberships defined in the container image for the uid of the container process. If unspecified, no additional groups are added to any container. Note that group memberships defined in the container image for the uid of the container process are still effective, even if they are not included in this list. Note that this field cannot be set when spec.os.name is windows."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.securityContext.withSysctls
+
+```ts
+withSysctls(sysctls)
+```
+
+"Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch. Note that this field cannot be set when spec.os.name is windows."
+
+### fn spec.template.spec.securityContext.withSysctlsMixin
+
+```ts
+withSysctlsMixin(sysctls)
+```
+
+"Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch. Note that this field cannot be set when spec.os.name is windows."
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.template.spec.securityContext.seLinuxOptions
+
+"The SELinux context to be applied to all containers. If unspecified, the container runtime will allocate a random SELinux context for each container. May also be set in SecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence for that container. Note that this field cannot be set when spec.os.name is windows."
+
+### fn spec.template.spec.securityContext.seLinuxOptions.withLevel
+
+```ts
+withLevel(level)
+```
+
+"Level is SELinux level label that applies to the container."
+
+### fn spec.template.spec.securityContext.seLinuxOptions.withRole
+
+```ts
+withRole(role)
+```
+
+"Role is a SELinux role label that applies to the container."
+
+### fn spec.template.spec.securityContext.seLinuxOptions.withType
+
+```ts
+withType(type)
+```
+
+"Type is a SELinux type label that applies to the container."
+
+### fn spec.template.spec.securityContext.seLinuxOptions.withUser
+
+```ts
+withUser(user)
+```
+
+"User is a SELinux user label that applies to the container."
+
+## obj spec.template.spec.securityContext.seccompProfile
+
+"The seccomp options to use by the containers in this pod. Note that this field cannot be set when spec.os.name is windows."
+
+### fn spec.template.spec.securityContext.seccompProfile.withLocalhostProfile
+
+```ts
+withLocalhostProfile(localhostProfile)
+```
+
+"localhostProfile indicates a profile defined in a file on the node should be used. The profile must be preconfigured on the node to work. Must be a descending path, relative to the kubelet's configured seccomp profile location. Must be set if type is \"Localhost\". Must NOT be set for any other type."
+
+### fn spec.template.spec.securityContext.seccompProfile.withType
+
+```ts
+withType(type)
+```
+
+"type indicates which kind of seccomp profile will be applied. Valid options are: \n Localhost - a profile defined in a file on the node should be used. RuntimeDefault - the container runtime default profile should be used. Unconfined - no profile should be applied."
+
+## obj spec.template.spec.securityContext.sysctls
+
+"Sysctls hold a list of namespaced sysctls used for the pod. Pods with unsupported sysctls (by the container runtime) might fail to launch. Note that this field cannot be set when spec.os.name is windows."
+
+### fn spec.template.spec.securityContext.sysctls.withName
+
+```ts
+withName(name)
+```
+
+"Name of a property to set"
+
+### fn spec.template.spec.securityContext.sysctls.withValue
+
+```ts
+withValue(value)
+```
+
+"Value of a property to set"
+
+## obj spec.template.spec.securityContext.windowsOptions
+
+"The Windows specific settings applied to all containers. If unspecified, the options within a container's SecurityContext will be used. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence. Note that this field cannot be set when spec.os.name is linux."
+
+### fn spec.template.spec.securityContext.windowsOptions.withGmsaCredentialSpec
+
+```ts
+withGmsaCredentialSpec(gmsaCredentialSpec)
+```
+
+"GMSACredentialSpec is where the GMSA admission webhook (https://github.com/kubernetes-sigs/windows-gmsa) inlines the contents of the GMSA credential spec named by the GMSACredentialSpecName field."
+
+### fn spec.template.spec.securityContext.windowsOptions.withGmsaCredentialSpecName
+
+```ts
+withGmsaCredentialSpecName(gmsaCredentialSpecName)
+```
+
+"GMSACredentialSpecName is the name of the GMSA credential spec to use."
+
+### fn spec.template.spec.securityContext.windowsOptions.withHostProcess
+
+```ts
+withHostProcess(hostProcess)
+```
+
+"HostProcess determines if a container should be run as a 'Host Process' container. All of a Pod's containers must have the same effective HostProcess value (it is not allowed to have a mix of HostProcess containers and non-HostProcess containers). In addition, if HostProcess is true then HostNetwork must also be set to true."
+
+### fn spec.template.spec.securityContext.windowsOptions.withRunAsUserName
+
+```ts
+withRunAsUserName(runAsUserName)
+```
+
+"The UserName in Windows to run the entrypoint of the container process. Defaults to the user specified in image metadata if unspecified. May also be set in PodSecurityContext. If set in both SecurityContext and PodSecurityContext, the value specified in SecurityContext takes precedence."
+
+## obj spec.template.spec.tolerations
+
+"If specified, the pod's tolerations."
+
+### fn spec.template.spec.tolerations.withEffect
+
+```ts
+withEffect(effect)
+```
+
+"Effect indicates the taint effect to match. Empty means match all taint effects. When specified, allowed values are NoSchedule, PreferNoSchedule and NoExecute."
+
+### fn spec.template.spec.tolerations.withKey
+
+```ts
+withKey(key)
+```
+
+"Key is the taint key that the toleration applies to. Empty means match all taint keys. If the key is empty, operator must be Exists; this combination means to match all values and all keys."
+
+### fn spec.template.spec.tolerations.withOperator
+
+```ts
+withOperator(operator)
+```
+
+"Operator represents a key's relationship to the value. Valid operators are Exists and Equal. Defaults to Equal. Exists is equivalent to wildcard for value, so that a pod can tolerate all taints of a particular category."
+
+### fn spec.template.spec.tolerations.withTolerationSeconds
+
+```ts
+withTolerationSeconds(tolerationSeconds)
+```
+
+"TolerationSeconds represents the period of time the toleration (which must be of effect NoExecute, otherwise this field is ignored) tolerates the taint. By default, it is not set, which means tolerate the taint forever (do not evict). Zero and negative values will be treated as 0 (evict immediately) by the system."
+
+### fn spec.template.spec.tolerations.withValue
+
+```ts
+withValue(value)
+```
+
+"Value is the taint value the toleration matches to. If the operator is Exists, the value should be empty, otherwise just a regular string."
+
+## obj spec.template.spec.topologySpreadConstraints
+
+"TopologySpreadConstraints describes how a group of pods ought to spread across topology domains. Scheduler will schedule pods in a way which abides by the constraints. All topologySpreadConstraints are ANDed."
+
+### fn spec.template.spec.topologySpreadConstraints.withMatchLabelKeys
+
+```ts
+withMatchLabelKeys(matchLabelKeys)
+```
+
+"MatchLabelKeys is a set of pod label keys to select the pods over which spreading will be calculated. The keys are used to lookup values from the incoming pod labels, those key-value labels are ANDed with labelSelector to select the group of existing pods over which spreading will be calculated for the incoming pod. The same key is forbidden to exist in both MatchLabelKeys and LabelSelector. MatchLabelKeys cannot be set when LabelSelector isn't set. Keys that don't exist in the incoming pod labels will be ignored. A null or empty list means only match against labelSelector. \n This is a beta field and requires the MatchLabelKeysInPodTopologySpread feature gate to be enabled (enabled by default)."
+
+### fn spec.template.spec.topologySpreadConstraints.withMatchLabelKeysMixin
+
+```ts
+withMatchLabelKeysMixin(matchLabelKeys)
+```
+
+"MatchLabelKeys is a set of pod label keys to select the pods over which spreading will be calculated. The keys are used to lookup values from the incoming pod labels, those key-value labels are ANDed with labelSelector to select the group of existing pods over which spreading will be calculated for the incoming pod. The same key is forbidden to exist in both MatchLabelKeys and LabelSelector. MatchLabelKeys cannot be set when LabelSelector isn't set. Keys that don't exist in the incoming pod labels will be ignored. A null or empty list means only match against labelSelector. \n This is a beta field and requires the MatchLabelKeysInPodTopologySpread feature gate to be enabled (enabled by default)."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.topologySpreadConstraints.withMaxSkew
+
+```ts
+withMaxSkew(maxSkew)
+```
+
+"MaxSkew describes the degree to which pods may be unevenly distributed. When `whenUnsatisfiable=DoNotSchedule`, it is the maximum permitted difference between the number of matching pods in the target topology and the global minimum. The global minimum is the minimum number of matching pods in an eligible domain or zero if the number of eligible domains is less than MinDomains. For example, in a 3-zone cluster, MaxSkew is set to 1, and pods with the same labelSelector spread as 2/2/1: In this case, the global minimum is 1. | zone1 | zone2 | zone3 | | P P | P P | P | - if MaxSkew is 1, incoming pod can only be scheduled to zone3 to become 2/2/2; scheduling it onto zone1(zone2) would make the ActualSkew(3-1) on zone1(zone2) violate MaxSkew(1). - if MaxSkew is 2, incoming pod can be scheduled onto any zone. When `whenUnsatisfiable=ScheduleAnyway`, it is used to give higher precedence to topologies that satisfy it. It's a required field. Default value is 1 and 0 is not allowed."
+
+### fn spec.template.spec.topologySpreadConstraints.withMinDomains
+
+```ts
+withMinDomains(minDomains)
+```
+
+"MinDomains indicates a minimum number of eligible domains. When the number of eligible domains with matching topology keys is less than minDomains, Pod Topology Spread treats \"global minimum\" as 0, and then the calculation of Skew is performed. And when the number of eligible domains with matching topology keys equals or greater than minDomains, this value has no effect on scheduling. As a result, when the number of eligible domains is less than minDomains, scheduler won't schedule more than maxSkew Pods to those domains. If value is nil, the constraint behaves as if MinDomains is equal to 1. Valid values are integers greater than 0. When value is not nil, WhenUnsatisfiable must be DoNotSchedule. \n For example, in a 3-zone cluster, MaxSkew is set to 2, MinDomains is set to 5 and pods with the same labelSelector spread as 2/2/2: | zone1 | zone2 | zone3 | | P P | P P | P P | The number of domains is less than 5(MinDomains), so \"global minimum\" is treated as 0. In this situation, new pod with the same labelSelector cannot be scheduled, because computed skew will be 3(3 - 0) if new Pod is scheduled to any of the three zones, it will violate MaxSkew. \n This is a beta field and requires the MinDomainsInPodTopologySpread feature gate to be enabled (enabled by default)."
+
+### fn spec.template.spec.topologySpreadConstraints.withNodeAffinityPolicy
+
+```ts
+withNodeAffinityPolicy(nodeAffinityPolicy)
+```
+
+"NodeAffinityPolicy indicates how we will treat Pod's nodeAffinity/nodeSelector when calculating pod topology spread skew. Options are: - Honor: only nodes matching nodeAffinity/nodeSelector are included in the calculations. - Ignore: nodeAffinity/nodeSelector are ignored. All nodes are included in the calculations. \n If this value is nil, the behavior is equivalent to the Honor policy. This is a beta-level feature default enabled by the NodeInclusionPolicyInPodTopologySpread feature flag."
+
+### fn spec.template.spec.topologySpreadConstraints.withNodeTaintsPolicy
+
+```ts
+withNodeTaintsPolicy(nodeTaintsPolicy)
+```
+
+"NodeTaintsPolicy indicates how we will treat node taints when calculating pod topology spread skew. Options are: - Honor: nodes without taints, along with tainted nodes for which the incoming pod has a toleration, are included. - Ignore: node taints are ignored. All nodes are included. \n If this value is nil, the behavior is equivalent to the Ignore policy. This is a beta-level feature default enabled by the NodeInclusionPolicyInPodTopologySpread feature flag."
+
+### fn spec.template.spec.topologySpreadConstraints.withTopologyKey
+
+```ts
+withTopologyKey(topologyKey)
+```
+
+"TopologyKey is the key of node labels. Nodes that have a label with this key and identical values are considered to be in the same topology. We consider each as a \"bucket\", and try to put balanced number of pods into each bucket. We define a domain as a particular instance of a topology. Also, we define an eligible domain as a domain whose nodes meet the requirements of nodeAffinityPolicy and nodeTaintsPolicy. e.g. If TopologyKey is \"kubernetes.io/hostname\", each Node is a domain of that topology. And, if TopologyKey is \"topology.kubernetes.io/zone\", each zone is a domain of that topology. It's a required field."
+
+### fn spec.template.spec.topologySpreadConstraints.withWhenUnsatisfiable
+
+```ts
+withWhenUnsatisfiable(whenUnsatisfiable)
+```
+
+"WhenUnsatisfiable indicates how to deal with a pod if it doesn't satisfy the spread constraint. - DoNotSchedule (default) tells the scheduler not to schedule it. - ScheduleAnyway tells the scheduler to schedule the pod in any location, but giving higher precedence to topologies that would help reduce the skew. A constraint is considered \"Unsatisfiable\" for an incoming pod if and only if every possible node assignment for that pod would violate \"MaxSkew\" on some topology. For example, in a 3-zone cluster, MaxSkew is set to 1, and pods with the same labelSelector spread as 3/1/1: | zone1 | zone2 | zone3 | | P P P | P | P | If WhenUnsatisfiable is set to DoNotSchedule, incoming pod can only be scheduled to zone2(zone3) to become 3/2/1(3/1/2) as ActualSkew(2-1) on zone2(zone3) satisfies MaxSkew(1). In other words, the cluster can still be imbalanced, but scheduler won't make it *more* imbalanced. It's a required field."
+
+## obj spec.template.spec.topologySpreadConstraints.labelSelector
+
+"LabelSelector is used to find matching pods. Pods that match this label selector are counted to determine the number of pods in their corresponding topology domain."
+
+### fn spec.template.spec.topologySpreadConstraints.labelSelector.withMatchExpressions
+
+```ts
+withMatchExpressions(matchExpressions)
+```
+
+"matchExpressions is a list of label selector requirements. The requirements are ANDed."
+
+### fn spec.template.spec.topologySpreadConstraints.labelSelector.withMatchExpressionsMixin
+
+```ts
+withMatchExpressionsMixin(matchExpressions)
+```
+
+"matchExpressions is a list of label selector requirements. The requirements are ANDed."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.topologySpreadConstraints.labelSelector.withMatchLabels
+
+```ts
+withMatchLabels(matchLabels)
+```
+
+"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \"key\", the operator is \"In\", and the values array contains only \"value\". The requirements are ANDed."
+
+### fn spec.template.spec.topologySpreadConstraints.labelSelector.withMatchLabelsMixin
+
+```ts
+withMatchLabelsMixin(matchLabels)
+```
+
+"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \"key\", the operator is \"In\", and the values array contains only \"value\". The requirements are ANDed."
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.template.spec.topologySpreadConstraints.labelSelector.matchExpressions
+
+"matchExpressions is a list of label selector requirements. The requirements are ANDed."
+
+### fn spec.template.spec.topologySpreadConstraints.labelSelector.matchExpressions.withKey
+
+```ts
+withKey(key)
+```
+
+"key is the label key that the selector applies to."
+
+### fn spec.template.spec.topologySpreadConstraints.labelSelector.matchExpressions.withOperator
+
+```ts
+withOperator(operator)
+```
+
+"operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist."
+
+### fn spec.template.spec.topologySpreadConstraints.labelSelector.matchExpressions.withValues
+
+```ts
+withValues(values)
+```
+
+"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."
+
+### fn spec.template.spec.topologySpreadConstraints.labelSelector.matchExpressions.withValuesMixin
+
+```ts
+withValuesMixin(values)
+```
+
+"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.template.spec.volumes
+
+"List of volumes that can be mounted by containers belonging to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes"
+
+### fn spec.template.spec.volumes.withName
+
+```ts
+withName(name)
+```
+
+"name of the volume. Must be a DNS_LABEL and unique within the pod. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names"
+
+## obj spec.template.spec.volumes.awsElasticBlockStore
+
+"awsElasticBlockStore represents an AWS Disk resource that is attached to a kubelet's host machine and then exposed to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes#awselasticblockstore"
+
+### fn spec.template.spec.volumes.awsElasticBlockStore.withFsType
+
+```ts
+withFsType(fsType)
+```
+
+"fsType is the filesystem type of the volume that you want to mount. Tip: Ensure that the filesystem type is supported by the host operating system. Examples: \"ext4\", \"xfs\", \"ntfs\". Implicitly inferred to be \"ext4\" if unspecified. More info: https://kubernetes.io/docs/concepts/storage/volumes#awselasticblockstore TODO: how do we prevent errors in the filesystem from compromising the machine"
+
+### fn spec.template.spec.volumes.awsElasticBlockStore.withPartition
+
+```ts
+withPartition(partition)
+```
+
+"partition is the partition in the volume that you want to mount. If omitted, the default is to mount by volume name. Examples: For volume /dev/sda1, you specify the partition as \"1\". Similarly, the volume partition for /dev/sda is \"0\" (or you can leave the property empty)."
+
+### fn spec.template.spec.volumes.awsElasticBlockStore.withReadOnly
+
+```ts
+withReadOnly(readOnly)
+```
+
+"readOnly value true will force the readOnly setting in VolumeMounts. More info: https://kubernetes.io/docs/concepts/storage/volumes#awselasticblockstore"
+
+### fn spec.template.spec.volumes.awsElasticBlockStore.withVolumeID
+
+```ts
+withVolumeID(volumeID)
+```
+
+"volumeID is unique ID of the persistent disk resource in AWS (Amazon EBS volume). More info: https://kubernetes.io/docs/concepts/storage/volumes#awselasticblockstore"
+
+## obj spec.template.spec.volumes.azureDisk
+
+"azureDisk represents an Azure Data Disk mount on the host and bind mount to the pod."
+
+### fn spec.template.spec.volumes.azureDisk.withCachingMode
+
+```ts
+withCachingMode(cachingMode)
+```
+
+"cachingMode is the Host Caching mode: None, Read Only, Read Write."
+
+### fn spec.template.spec.volumes.azureDisk.withDiskName
+
+```ts
+withDiskName(diskName)
+```
+
+"diskName is the Name of the data disk in the blob storage"
+
+### fn spec.template.spec.volumes.azureDisk.withDiskURI
+
+```ts
+withDiskURI(diskURI)
+```
+
+"diskURI is the URI of data disk in the blob storage"
+
+### fn spec.template.spec.volumes.azureDisk.withFsType
+
+```ts
+withFsType(fsType)
+```
+
+"fsType is Filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. \"ext4\", \"xfs\", \"ntfs\". Implicitly inferred to be \"ext4\" if unspecified."
+
+### fn spec.template.spec.volumes.azureDisk.withKind
+
+```ts
+withKind(kind)
+```
+
+"kind expected values are Shared: multiple blob disks per storage account Dedicated: single blob disk per storage account Managed: azure managed data disk (only in managed availability set). defaults to shared"
+
+### fn spec.template.spec.volumes.azureDisk.withReadOnly
+
+```ts
+withReadOnly(readOnly)
+```
+
+"readOnly Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts."
+
+## obj spec.template.spec.volumes.azureFile
+
+"azureFile represents an Azure File Service mount on the host and bind mount to the pod."
+
+### fn spec.template.spec.volumes.azureFile.withReadOnly
+
+```ts
+withReadOnly(readOnly)
+```
+
+"readOnly defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts."
+
+### fn spec.template.spec.volumes.azureFile.withSecretName
+
+```ts
+withSecretName(secretName)
+```
+
+"secretName is the name of secret that contains Azure Storage Account Name and Key"
+
+### fn spec.template.spec.volumes.azureFile.withShareName
+
+```ts
+withShareName(shareName)
+```
+
+"shareName is the azure share Name"
+
+## obj spec.template.spec.volumes.cephfs
+
+"cephFS represents a Ceph FS mount on the host that shares a pod's lifetime"
+
+### fn spec.template.spec.volumes.cephfs.withMonitors
+
+```ts
+withMonitors(monitors)
+```
+
+"monitors is Required: Monitors is a collection of Ceph monitors More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it"
+
+### fn spec.template.spec.volumes.cephfs.withMonitorsMixin
+
+```ts
+withMonitorsMixin(monitors)
+```
+
+"monitors is Required: Monitors is a collection of Ceph monitors More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it"
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.volumes.cephfs.withPath
+
+```ts
+withPath(path)
+```
+
+"path is Optional: Used as the mounted root, rather than the full Ceph tree, default is /"
+
+### fn spec.template.spec.volumes.cephfs.withReadOnly
+
+```ts
+withReadOnly(readOnly)
+```
+
+"readOnly is Optional: Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts. More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it"
+
+### fn spec.template.spec.volumes.cephfs.withSecretFile
+
+```ts
+withSecretFile(secretFile)
+```
+
+"secretFile is Optional: SecretFile is the path to key ring for User, default is /etc/ceph/user.secret More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it"
+
+### fn spec.template.spec.volumes.cephfs.withUser
+
+```ts
+withUser(user)
+```
+
+"user is optional: User is the rados user name, default is admin More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it"
+
+## obj spec.template.spec.volumes.cephfs.secretRef
+
+"secretRef is Optional: SecretRef is reference to the authentication secret for User, default is empty. More info: https://examples.k8s.io/volumes/cephfs/README.md#how-to-use-it"
+
+### fn spec.template.spec.volumes.cephfs.secretRef.withName
+
+```ts
+withName(name)
+```
+
+"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"
+
+## obj spec.template.spec.volumes.cinder
+
+"cinder represents a cinder volume attached and mounted on kubelets host machine. More info: https://examples.k8s.io/mysql-cinder-pd/README.md"
+
+### fn spec.template.spec.volumes.cinder.withFsType
+
+```ts
+withFsType(fsType)
+```
+
+"fsType is the filesystem type to mount. Must be a filesystem type supported by the host operating system. Examples: \"ext4\", \"xfs\", \"ntfs\". Implicitly inferred to be \"ext4\" if unspecified. More info: https://examples.k8s.io/mysql-cinder-pd/README.md"
+
+### fn spec.template.spec.volumes.cinder.withReadOnly
+
+```ts
+withReadOnly(readOnly)
+```
+
+"readOnly defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts. More info: https://examples.k8s.io/mysql-cinder-pd/README.md"
+
+### fn spec.template.spec.volumes.cinder.withVolumeID
+
+```ts
+withVolumeID(volumeID)
+```
+
+"volumeID used to identify the volume in cinder. More info: https://examples.k8s.io/mysql-cinder-pd/README.md"
+
+## obj spec.template.spec.volumes.cinder.secretRef
+
+"secretRef is optional: points to a secret object containing parameters used to connect to OpenStack."
+
+### fn spec.template.spec.volumes.cinder.secretRef.withName
+
+```ts
+withName(name)
+```
+
+"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"
+
+## obj spec.template.spec.volumes.configMap
+
+"configMap represents a configMap that should populate this volume"
+
+### fn spec.template.spec.volumes.configMap.withDefaultMode
+
+```ts
+withDefaultMode(defaultMode)
+```
+
+"defaultMode is optional: mode bits used to set permissions on created files by default. Must be an octal value between 0000 and 0777 or a decimal value between 0 and 511. YAML accepts both octal and decimal values, JSON requires decimal values for mode bits. Defaults to 0644. Directories within the path are not affected by this setting. This might be in conflict with other options that affect the file mode, like fsGroup, and the result can be other mode bits set."
+
+### fn spec.template.spec.volumes.configMap.withItems
+
+```ts
+withItems(items)
+```
+
+"items if unspecified, each key-value pair in the Data field of the referenced ConfigMap will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the ConfigMap, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'."
+
+### fn spec.template.spec.volumes.configMap.withItemsMixin
+
+```ts
+withItemsMixin(items)
+```
+
+"items if unspecified, each key-value pair in the Data field of the referenced ConfigMap will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the ConfigMap, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.volumes.configMap.withName
+
+```ts
+withName(name)
+```
+
+"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"
+
+### fn spec.template.spec.volumes.configMap.withOptional
+
+```ts
+withOptional(optional)
+```
+
+"optional specify whether the ConfigMap or its keys must be defined"
+
+## obj spec.template.spec.volumes.configMap.items
+
+"items if unspecified, each key-value pair in the Data field of the referenced ConfigMap will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the ConfigMap, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'."
+
+### fn spec.template.spec.volumes.configMap.items.withKey
+
+```ts
+withKey(key)
+```
+
+"key is the key to project."
+
+### fn spec.template.spec.volumes.configMap.items.withMode
+
+```ts
+withMode(mode)
+```
+
+"mode is Optional: mode bits used to set permissions on this file. Must be an octal value between 0000 and 0777 or a decimal value between 0 and 511. YAML accepts both octal and decimal values, JSON requires decimal values for mode bits. If not specified, the volume defaultMode will be used. This might be in conflict with other options that affect the file mode, like fsGroup, and the result can be other mode bits set."
+
+### fn spec.template.spec.volumes.configMap.items.withPath
+
+```ts
+withPath(path)
+```
+
+"path is the relative path of the file to map the key to. May not be an absolute path. May not contain the path element '..'. May not start with the string '..'."
+
+## obj spec.template.spec.volumes.csi
+
+"csi (Container Storage Interface) represents ephemeral storage that is handled by certain external CSI drivers (Beta feature)."
+
+### fn spec.template.spec.volumes.csi.withDriver
+
+```ts
+withDriver(driver)
+```
+
+"driver is the name of the CSI driver that handles this volume. Consult with your admin for the correct name as registered in the cluster."
+
+### fn spec.template.spec.volumes.csi.withFsType
+
+```ts
+withFsType(fsType)
+```
+
+"fsType to mount. Ex. \"ext4\", \"xfs\", \"ntfs\". If not provided, the empty value is passed to the associated CSI driver which will determine the default filesystem to apply."
+
+### fn spec.template.spec.volumes.csi.withReadOnly
+
+```ts
+withReadOnly(readOnly)
+```
+
+"readOnly specifies a read-only configuration for the volume. Defaults to false (read/write)."
+
+### fn spec.template.spec.volumes.csi.withVolumeAttributes
+
+```ts
+withVolumeAttributes(volumeAttributes)
+```
+
+"volumeAttributes stores driver-specific properties that are passed to the CSI driver. Consult your driver's documentation for supported values."
+
+### fn spec.template.spec.volumes.csi.withVolumeAttributesMixin
+
+```ts
+withVolumeAttributesMixin(volumeAttributes)
+```
+
+"volumeAttributes stores driver-specific properties that are passed to the CSI driver. Consult your driver's documentation for supported values."
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.template.spec.volumes.csi.nodePublishSecretRef
+
+"nodePublishSecretRef is a reference to the secret object containing sensitive information to pass to the CSI driver to complete the CSI NodePublishVolume and NodeUnpublishVolume calls. This field is optional, and may be empty if no secret is required. If the secret object contains more than one secret, all secret references are passed."
+
+### fn spec.template.spec.volumes.csi.nodePublishSecretRef.withName
+
+```ts
+withName(name)
+```
+
+"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"
+
+## obj spec.template.spec.volumes.downwardAPI
+
+"downwardAPI represents downward API about the pod that should populate this volume"
+
+### fn spec.template.spec.volumes.downwardAPI.withDefaultMode
+
+```ts
+withDefaultMode(defaultMode)
+```
+
+"Optional: mode bits to use on created files by default. Must be a Optional: mode bits used to set permissions on created files by default. Must be an octal value between 0000 and 0777 or a decimal value between 0 and 511. YAML accepts both octal and decimal values, JSON requires decimal values for mode bits. Defaults to 0644. Directories within the path are not affected by this setting. This might be in conflict with other options that affect the file mode, like fsGroup, and the result can be other mode bits set."
+
+### fn spec.template.spec.volumes.downwardAPI.withItems
+
+```ts
+withItems(items)
+```
+
+"Items is a list of downward API volume file"
+
+### fn spec.template.spec.volumes.downwardAPI.withItemsMixin
+
+```ts
+withItemsMixin(items)
+```
+
+"Items is a list of downward API volume file"
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.template.spec.volumes.downwardAPI.items
+
+"Items is a list of downward API volume file"
+
+### fn spec.template.spec.volumes.downwardAPI.items.withMode
+
+```ts
+withMode(mode)
+```
+
+"Optional: mode bits used to set permissions on this file, must be an octal value between 0000 and 0777 or a decimal value between 0 and 511. YAML accepts both octal and decimal values, JSON requires decimal values for mode bits. If not specified, the volume defaultMode will be used. This might be in conflict with other options that affect the file mode, like fsGroup, and the result can be other mode bits set."
+
+### fn spec.template.spec.volumes.downwardAPI.items.withPath
+
+```ts
+withPath(path)
+```
+
+"Required: Path is the relative path name of the file to be created. Must not be absolute or contain the '..' path. Must be utf-8 encoded. The first item of the relative path must not start with '..'"
+
+## obj spec.template.spec.volumes.downwardAPI.items.fieldRef
+
+"Required: Selects a field of the pod: only annotations, labels, name and namespace are supported."
+
+### fn spec.template.spec.volumes.downwardAPI.items.fieldRef.withApiVersion
+
+```ts
+withApiVersion(apiVersion)
+```
+
+"Version of the schema the FieldPath is written in terms of, defaults to \"v1\"."
+
+### fn spec.template.spec.volumes.downwardAPI.items.fieldRef.withFieldPath
+
+```ts
+withFieldPath(fieldPath)
+```
+
+"Path of the field to select in the specified API version."
+
+## obj spec.template.spec.volumes.downwardAPI.items.resourceFieldRef
+
+"Selects a resource of the container: only resources limits and requests (limits.cpu, limits.memory, requests.cpu and requests.memory) are currently supported."
+
+### fn spec.template.spec.volumes.downwardAPI.items.resourceFieldRef.withContainerName
+
+```ts
+withContainerName(containerName)
+```
+
+"Container name: required for volumes, optional for env vars"
+
+### fn spec.template.spec.volumes.downwardAPI.items.resourceFieldRef.withDivisor
+
+```ts
+withDivisor(divisor)
+```
+
+"Specifies the output format of the exposed resources, defaults to \"1\
+
+### fn spec.template.spec.volumes.downwardAPI.items.resourceFieldRef.withResource
+
+```ts
+withResource(resource)
+```
+
+"Required: resource to select"
+
+## obj spec.template.spec.volumes.emptyDir
+
+"emptyDir represents a temporary directory that shares a pod's lifetime. More info: https://kubernetes.io/docs/concepts/storage/volumes#emptydir"
+
+### fn spec.template.spec.volumes.emptyDir.withMedium
+
+```ts
+withMedium(medium)
+```
+
+"medium represents what type of storage medium should back this directory. The default is \"\" which means to use the node's default medium. Must be an empty string (default) or Memory. More info: https://kubernetes.io/docs/concepts/storage/volumes#emptydir"
+
+### fn spec.template.spec.volumes.emptyDir.withSizeLimit
+
+```ts
+withSizeLimit(sizeLimit)
+```
+
+"sizeLimit is the total amount of local storage required for this EmptyDir volume. The size limit is also applicable for memory medium. The maximum usage on memory medium EmptyDir would be the minimum value between the SizeLimit specified here and the sum of memory limits of all containers in a pod. The default is nil which means that the limit is undefined. More info: https://kubernetes.io/docs/concepts/storage/volumes#emptydir"
+
+## obj spec.template.spec.volumes.ephemeral
+
+"ephemeral represents a volume that is handled by a cluster storage driver. The volume's lifecycle is tied to the pod that defines it - it will be created before the pod starts, and deleted when the pod is removed. \n Use this if: a) the volume is only needed while the pod runs, b) features of normal volumes like restoring from snapshot or capacity tracking are needed, c) the storage driver is specified through a storage class, and d) the storage driver supports dynamic volume provisioning through a PersistentVolumeClaim (see EphemeralVolumeSource for more information on the connection between this volume type and PersistentVolumeClaim). \n Use PersistentVolumeClaim or one of the vendor-specific APIs for volumes that persist for longer than the lifecycle of an individual pod. \n Use CSI for light-weight local ephemeral volumes if the CSI driver is meant to be used that way - see the documentation of the driver for more information. \n A pod can use both types of ephemeral volumes and persistent volumes at the same time."
+
+## obj spec.template.spec.volumes.ephemeral.volumeClaimTemplate
+
+"Will be used to create a stand-alone PVC to provision the volume. The pod in which this EphemeralVolumeSource is embedded will be the owner of the PVC, i.e. the PVC will be deleted together with the pod. The name of the PVC will be `-` where `` is the name from the `PodSpec.Volumes` array entry. Pod validation will reject the pod if the concatenated name is not valid for a PVC (for example, too long). \n An existing PVC with that name that is not owned by the pod will *not* be used for the pod to avoid using an unrelated volume by mistake. Starting the pod is then blocked until the unrelated PVC is removed. If such a pre-created PVC is meant to be used by the pod, the PVC has to updated with an owner reference to the pod once the pod exists. Normally this should not be necessary, but it may be useful when manually reconstructing a broken cluster. \n This field is read-only and no changes will be made by Kubernetes to the PVC after it has been created. \n Required, must not be nil."
+
+### fn spec.template.spec.volumes.ephemeral.volumeClaimTemplate.withMetadata
+
+```ts
+withMetadata(metadata)
+```
+
+"May contain labels and annotations that will be copied into the PVC when creating it. No other fields are allowed and will be rejected during validation."
+
+### fn spec.template.spec.volumes.ephemeral.volumeClaimTemplate.withMetadataMixin
+
+```ts
+withMetadataMixin(metadata)
+```
+
+"May contain labels and annotations that will be copied into the PVC when creating it. No other fields are allowed and will be rejected during validation."
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.template.spec.volumes.ephemeral.volumeClaimTemplate.spec
+
+"The specification for the PersistentVolumeClaim. The entire content is copied unchanged into the PVC that gets created from this template. The same fields as in a PersistentVolumeClaim are also valid here."
+
+### fn spec.template.spec.volumes.ephemeral.volumeClaimTemplate.spec.withAccessModes
+
+```ts
+withAccessModes(accessModes)
+```
+
+"accessModes contains the desired access modes the volume should have. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#access-modes-1"
+
+### fn spec.template.spec.volumes.ephemeral.volumeClaimTemplate.spec.withAccessModesMixin
+
+```ts
+withAccessModesMixin(accessModes)
+```
+
+"accessModes contains the desired access modes the volume should have. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#access-modes-1"
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.volumes.ephemeral.volumeClaimTemplate.spec.withStorageClassName
+
+```ts
+withStorageClassName(storageClassName)
+```
+
+"storageClassName is the name of the StorageClass required by the claim. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#class-1"
+
+### fn spec.template.spec.volumes.ephemeral.volumeClaimTemplate.spec.withVolumeMode
+
+```ts
+withVolumeMode(volumeMode)
+```
+
+"volumeMode defines what type of volume is required by the claim. Value of Filesystem is implied when not included in claim spec."
+
+### fn spec.template.spec.volumes.ephemeral.volumeClaimTemplate.spec.withVolumeName
+
+```ts
+withVolumeName(volumeName)
+```
+
+"volumeName is the binding reference to the PersistentVolume backing this claim."
+
+## obj spec.template.spec.volumes.ephemeral.volumeClaimTemplate.spec.dataSource
+
+"dataSource field can be used to specify either: * An existing VolumeSnapshot object (snapshot.storage.k8s.io/VolumeSnapshot) * An existing PVC (PersistentVolumeClaim) If the provisioner or an external controller can support the specified data source, it will create a new volume based on the contents of the specified data source. When the AnyVolumeDataSource feature gate is enabled, dataSource contents will be copied to dataSourceRef, and dataSourceRef contents will be copied to dataSource when dataSourceRef.namespace is not specified. If the namespace is specified, then dataSourceRef will not be copied to dataSource."
+
+### fn spec.template.spec.volumes.ephemeral.volumeClaimTemplate.spec.dataSource.withApiGroup
+
+```ts
+withApiGroup(apiGroup)
+```
+
+"APIGroup is the group for the resource being referenced. If APIGroup is not specified, the specified Kind must be in the core API group. For any other third-party types, APIGroup is required."
+
+### fn spec.template.spec.volumes.ephemeral.volumeClaimTemplate.spec.dataSource.withKind
+
+```ts
+withKind(kind)
+```
+
+"Kind is the type of resource being referenced"
+
+### fn spec.template.spec.volumes.ephemeral.volumeClaimTemplate.spec.dataSource.withName
+
+```ts
+withName(name)
+```
+
+"Name is the name of resource being referenced"
+
+## obj spec.template.spec.volumes.ephemeral.volumeClaimTemplate.spec.dataSourceRef
+
+"dataSourceRef specifies the object from which to populate the volume with data, if a non-empty volume is desired. This may be any object from a non-empty API group (non core object) or a PersistentVolumeClaim object. When this field is specified, volume binding will only succeed if the type of the specified object matches some installed volume populator or dynamic provisioner. This field will replace the functionality of the dataSource field and as such if both fields are non-empty, they must have the same value. For backwards compatibility, when namespace isn't specified in dataSourceRef, both fields (dataSource and dataSourceRef) will be set to the same value automatically if one of them is empty and the other is non-empty. When namespace is specified in dataSourceRef, dataSource isn't set to the same value and must be empty. There are three important differences between dataSource and dataSourceRef: * While dataSource only allows two specific types of objects, dataSourceRef allows any non-core object, as well as PersistentVolumeClaim objects. * While dataSource ignores disallowed values (dropping them), dataSourceRef preserves all values, and generates an error if a disallowed value is specified. * While dataSource only allows local objects, dataSourceRef allows objects in any namespaces. (Beta) Using this field requires the AnyVolumeDataSource feature gate to be enabled. (Alpha) Using the namespace field of dataSourceRef requires the CrossNamespaceVolumeDataSource feature gate to be enabled."
+
+### fn spec.template.spec.volumes.ephemeral.volumeClaimTemplate.spec.dataSourceRef.withApiGroup
+
+```ts
+withApiGroup(apiGroup)
+```
+
+"APIGroup is the group for the resource being referenced. If APIGroup is not specified, the specified Kind must be in the core API group. For any other third-party types, APIGroup is required."
+
+### fn spec.template.spec.volumes.ephemeral.volumeClaimTemplate.spec.dataSourceRef.withKind
+
+```ts
+withKind(kind)
+```
+
+"Kind is the type of resource being referenced"
+
+### fn spec.template.spec.volumes.ephemeral.volumeClaimTemplate.spec.dataSourceRef.withName
+
+```ts
+withName(name)
+```
+
+"Name is the name of resource being referenced"
+
+### fn spec.template.spec.volumes.ephemeral.volumeClaimTemplate.spec.dataSourceRef.withNamespace
+
+```ts
+withNamespace(namespace)
+```
+
+"Namespace is the namespace of resource being referenced Note that when a namespace is specified, a gateway.networking.k8s.io/ReferenceGrant object is required in the referent namespace to allow that namespace's owner to accept the reference. See the ReferenceGrant documentation for details. (Alpha) This field requires the CrossNamespaceVolumeDataSource feature gate to be enabled."
+
+## obj spec.template.spec.volumes.ephemeral.volumeClaimTemplate.spec.resources
+
+"resources represents the minimum resources the volume should have. If RecoverVolumeExpansionFailure feature is enabled users are allowed to specify resource requirements that are lower than previous value but must still be higher than capacity recorded in the status field of the claim. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#resources"
+
+### fn spec.template.spec.volumes.ephemeral.volumeClaimTemplate.spec.resources.withClaims
+
+```ts
+withClaims(claims)
+```
+
+"Claims lists the names of resources, defined in spec.resourceClaims, that are used by this container. \n This is an alpha field and requires enabling the DynamicResourceAllocation feature gate. \n This field is immutable. It can only be set for containers."
+
+### fn spec.template.spec.volumes.ephemeral.volumeClaimTemplate.spec.resources.withClaimsMixin
+
+```ts
+withClaimsMixin(claims)
+```
+
+"Claims lists the names of resources, defined in spec.resourceClaims, that are used by this container. \n This is an alpha field and requires enabling the DynamicResourceAllocation feature gate. \n This field is immutable. It can only be set for containers."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.volumes.ephemeral.volumeClaimTemplate.spec.resources.withLimits
+
+```ts
+withLimits(limits)
+```
+
+"Limits describes the maximum amount of compute resources allowed. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"
+
+### fn spec.template.spec.volumes.ephemeral.volumeClaimTemplate.spec.resources.withLimitsMixin
+
+```ts
+withLimitsMixin(limits)
+```
+
+"Limits describes the maximum amount of compute resources allowed. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.volumes.ephemeral.volumeClaimTemplate.spec.resources.withRequests
+
+```ts
+withRequests(requests)
+```
+
+"Requests describes the minimum amount of compute resources required. If Requests is omitted for a container, it defaults to Limits if that is explicitly specified, otherwise to an implementation-defined value. Requests cannot exceed Limits. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"
+
+### fn spec.template.spec.volumes.ephemeral.volumeClaimTemplate.spec.resources.withRequestsMixin
+
+```ts
+withRequestsMixin(requests)
+```
+
+"Requests describes the minimum amount of compute resources required. If Requests is omitted for a container, it defaults to Limits if that is explicitly specified, otherwise to an implementation-defined value. Requests cannot exceed Limits. More info: https://kubernetes.io/docs/concepts/configuration/manage-resources-containers/"
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.template.spec.volumes.ephemeral.volumeClaimTemplate.spec.resources.claims
+
+"Claims lists the names of resources, defined in spec.resourceClaims, that are used by this container. \n This is an alpha field and requires enabling the DynamicResourceAllocation feature gate. \n This field is immutable. It can only be set for containers."
+
+### fn spec.template.spec.volumes.ephemeral.volumeClaimTemplate.spec.resources.claims.withName
+
+```ts
+withName(name)
+```
+
+"Name must match the name of one entry in pod.spec.resourceClaims of the Pod where this field is used. It makes that resource available inside a container."
+
+## obj spec.template.spec.volumes.ephemeral.volumeClaimTemplate.spec.selector
+
+"selector is a label query over volumes to consider for binding."
+
+### fn spec.template.spec.volumes.ephemeral.volumeClaimTemplate.spec.selector.withMatchExpressions
+
+```ts
+withMatchExpressions(matchExpressions)
+```
+
+"matchExpressions is a list of label selector requirements. The requirements are ANDed."
+
+### fn spec.template.spec.volumes.ephemeral.volumeClaimTemplate.spec.selector.withMatchExpressionsMixin
+
+```ts
+withMatchExpressionsMixin(matchExpressions)
+```
+
+"matchExpressions is a list of label selector requirements. The requirements are ANDed."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.volumes.ephemeral.volumeClaimTemplate.spec.selector.withMatchLabels
+
+```ts
+withMatchLabels(matchLabels)
+```
+
+"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \"key\", the operator is \"In\", and the values array contains only \"value\". The requirements are ANDed."
+
+### fn spec.template.spec.volumes.ephemeral.volumeClaimTemplate.spec.selector.withMatchLabelsMixin
+
+```ts
+withMatchLabelsMixin(matchLabels)
+```
+
+"matchLabels is a map of {key,value} pairs. A single {key,value} in the matchLabels map is equivalent to an element of matchExpressions, whose key field is \"key\", the operator is \"In\", and the values array contains only \"value\". The requirements are ANDed."
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.template.spec.volumes.ephemeral.volumeClaimTemplate.spec.selector.matchExpressions
+
+"matchExpressions is a list of label selector requirements. The requirements are ANDed."
+
+### fn spec.template.spec.volumes.ephemeral.volumeClaimTemplate.spec.selector.matchExpressions.withKey
+
+```ts
+withKey(key)
+```
+
+"key is the label key that the selector applies to."
+
+### fn spec.template.spec.volumes.ephemeral.volumeClaimTemplate.spec.selector.matchExpressions.withOperator
+
+```ts
+withOperator(operator)
+```
+
+"operator represents a key's relationship to a set of values. Valid operators are In, NotIn, Exists and DoesNotExist."
+
+### fn spec.template.spec.volumes.ephemeral.volumeClaimTemplate.spec.selector.matchExpressions.withValues
+
+```ts
+withValues(values)
+```
+
+"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."
+
+### fn spec.template.spec.volumes.ephemeral.volumeClaimTemplate.spec.selector.matchExpressions.withValuesMixin
+
+```ts
+withValuesMixin(values)
+```
+
+"values is an array of string values. If the operator is In or NotIn, the values array must be non-empty. If the operator is Exists or DoesNotExist, the values array must be empty. This array is replaced during a strategic merge patch."
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.template.spec.volumes.fc
+
+"fc represents a Fibre Channel resource that is attached to a kubelet's host machine and then exposed to the pod."
+
+### fn spec.template.spec.volumes.fc.withFsType
+
+```ts
+withFsType(fsType)
+```
+
+"fsType is the filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. \"ext4\", \"xfs\", \"ntfs\". Implicitly inferred to be \"ext4\" if unspecified. TODO: how do we prevent errors in the filesystem from compromising the machine"
+
+### fn spec.template.spec.volumes.fc.withLun
+
+```ts
+withLun(lun)
+```
+
+"lun is Optional: FC target lun number"
+
+### fn spec.template.spec.volumes.fc.withReadOnly
+
+```ts
+withReadOnly(readOnly)
+```
+
+"readOnly is Optional: Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts."
+
+### fn spec.template.spec.volumes.fc.withTargetWWNs
+
+```ts
+withTargetWWNs(targetWWNs)
+```
+
+"targetWWNs is Optional: FC target worldwide names (WWNs)"
+
+### fn spec.template.spec.volumes.fc.withTargetWWNsMixin
+
+```ts
+withTargetWWNsMixin(targetWWNs)
+```
+
+"targetWWNs is Optional: FC target worldwide names (WWNs)"
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.volumes.fc.withWwids
+
+```ts
+withWwids(wwids)
+```
+
+"wwids Optional: FC volume world wide identifiers (wwids) Either wwids or combination of targetWWNs and lun must be set, but not both simultaneously."
+
+### fn spec.template.spec.volumes.fc.withWwidsMixin
+
+```ts
+withWwidsMixin(wwids)
+```
+
+"wwids Optional: FC volume world wide identifiers (wwids) Either wwids or combination of targetWWNs and lun must be set, but not both simultaneously."
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.template.spec.volumes.flexVolume
+
+"flexVolume represents a generic volume resource that is provisioned/attached using an exec based plugin."
+
+### fn spec.template.spec.volumes.flexVolume.withDriver
+
+```ts
+withDriver(driver)
+```
+
+"driver is the name of the driver to use for this volume."
+
+### fn spec.template.spec.volumes.flexVolume.withFsType
+
+```ts
+withFsType(fsType)
+```
+
+"fsType is the filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. \"ext4\", \"xfs\", \"ntfs\". The default filesystem depends on FlexVolume script."
+
+### fn spec.template.spec.volumes.flexVolume.withOptions
+
+```ts
+withOptions(options)
+```
+
+"options is Optional: this field holds extra command options if any."
+
+### fn spec.template.spec.volumes.flexVolume.withOptionsMixin
+
+```ts
+withOptionsMixin(options)
+```
+
+"options is Optional: this field holds extra command options if any."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.volumes.flexVolume.withReadOnly
+
+```ts
+withReadOnly(readOnly)
+```
+
+"readOnly is Optional: defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts."
+
+## obj spec.template.spec.volumes.flexVolume.secretRef
+
+"secretRef is Optional: secretRef is reference to the secret object containing sensitive information to pass to the plugin scripts. This may be empty if no secret object is specified. If the secret object contains more than one secret, all secrets are passed to the plugin scripts."
+
+### fn spec.template.spec.volumes.flexVolume.secretRef.withName
+
+```ts
+withName(name)
+```
+
+"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"
+
+## obj spec.template.spec.volumes.flocker
+
+"flocker represents a Flocker volume attached to a kubelet's host machine. This depends on the Flocker control service being running"
+
+### fn spec.template.spec.volumes.flocker.withDatasetName
+
+```ts
+withDatasetName(datasetName)
+```
+
+"datasetName is Name of the dataset stored as metadata -> name on the dataset for Flocker should be considered as deprecated"
+
+### fn spec.template.spec.volumes.flocker.withDatasetUUID
+
+```ts
+withDatasetUUID(datasetUUID)
+```
+
+"datasetUUID is the UUID of the dataset. This is unique identifier of a Flocker dataset"
+
+## obj spec.template.spec.volumes.gcePersistentDisk
+
+"gcePersistentDisk represents a GCE Disk resource that is attached to a kubelet's host machine and then exposed to the pod. More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk"
+
+### fn spec.template.spec.volumes.gcePersistentDisk.withFsType
+
+```ts
+withFsType(fsType)
+```
+
+"fsType is filesystem type of the volume that you want to mount. Tip: Ensure that the filesystem type is supported by the host operating system. Examples: \"ext4\", \"xfs\", \"ntfs\". Implicitly inferred to be \"ext4\" if unspecified. More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk TODO: how do we prevent errors in the filesystem from compromising the machine"
+
+### fn spec.template.spec.volumes.gcePersistentDisk.withPartition
+
+```ts
+withPartition(partition)
+```
+
+"partition is the partition in the volume that you want to mount. If omitted, the default is to mount by volume name. Examples: For volume /dev/sda1, you specify the partition as \"1\". Similarly, the volume partition for /dev/sda is \"0\" (or you can leave the property empty). More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk"
+
+### fn spec.template.spec.volumes.gcePersistentDisk.withPdName
+
+```ts
+withPdName(pdName)
+```
+
+"pdName is unique name of the PD resource in GCE. Used to identify the disk in GCE. More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk"
+
+### fn spec.template.spec.volumes.gcePersistentDisk.withReadOnly
+
+```ts
+withReadOnly(readOnly)
+```
+
+"readOnly here will force the ReadOnly setting in VolumeMounts. Defaults to false. More info: https://kubernetes.io/docs/concepts/storage/volumes#gcepersistentdisk"
+
+## obj spec.template.spec.volumes.gitRepo
+
+"gitRepo represents a git repository at a particular revision. DEPRECATED: GitRepo is deprecated. To provision a container with a git repo, mount an EmptyDir into an InitContainer that clones the repo using git, then mount the EmptyDir into the Pod's container."
+
+### fn spec.template.spec.volumes.gitRepo.withDirectory
+
+```ts
+withDirectory(directory)
+```
+
+"directory is the target directory name. Must not contain or start with '..'. If '.' is supplied, the volume directory will be the git repository. Otherwise, if specified, the volume will contain the git repository in the subdirectory with the given name."
+
+### fn spec.template.spec.volumes.gitRepo.withRepository
+
+```ts
+withRepository(repository)
+```
+
+"repository is the URL"
+
+### fn spec.template.spec.volumes.gitRepo.withRevision
+
+```ts
+withRevision(revision)
+```
+
+"revision is the commit hash for the specified revision."
+
+## obj spec.template.spec.volumes.glusterfs
+
+"glusterfs represents a Glusterfs mount on the host that shares a pod's lifetime. More info: https://examples.k8s.io/volumes/glusterfs/README.md"
+
+### fn spec.template.spec.volumes.glusterfs.withEndpoints
+
+```ts
+withEndpoints(endpoints)
+```
+
+"endpoints is the endpoint name that details Glusterfs topology. More info: https://examples.k8s.io/volumes/glusterfs/README.md#create-a-pod"
+
+### fn spec.template.spec.volumes.glusterfs.withPath
+
+```ts
+withPath(path)
+```
+
+"path is the Glusterfs volume path. More info: https://examples.k8s.io/volumes/glusterfs/README.md#create-a-pod"
+
+### fn spec.template.spec.volumes.glusterfs.withReadOnly
+
+```ts
+withReadOnly(readOnly)
+```
+
+"readOnly here will force the Glusterfs volume to be mounted with read-only permissions. Defaults to false. More info: https://examples.k8s.io/volumes/glusterfs/README.md#create-a-pod"
+
+## obj spec.template.spec.volumes.hostPath
+
+"hostPath represents a pre-existing file or directory on the host machine that is directly exposed to the container. This is generally used for system agents or other privileged things that are allowed to see the host machine. Most containers will NOT need this. More info: https://kubernetes.io/docs/concepts/storage/volumes#hostpath --- TODO(jonesdl) We need to restrict who can use host directory mounts and who can/can not mount host directories as read/write."
+
+### fn spec.template.spec.volumes.hostPath.withPath
+
+```ts
+withPath(path)
+```
+
+"path of the directory on the host. If the path is a symlink, it will follow the link to the real path. More info: https://kubernetes.io/docs/concepts/storage/volumes#hostpath"
+
+### fn spec.template.spec.volumes.hostPath.withType
+
+```ts
+withType(type)
+```
+
+"type for HostPath Volume Defaults to \"\" More info: https://kubernetes.io/docs/concepts/storage/volumes#hostpath"
+
+## obj spec.template.spec.volumes.iscsi
+
+"iscsi represents an ISCSI Disk resource that is attached to a kubelet's host machine and then exposed to the pod. More info: https://examples.k8s.io/volumes/iscsi/README.md"
+
+### fn spec.template.spec.volumes.iscsi.withChapAuthDiscovery
+
+```ts
+withChapAuthDiscovery(chapAuthDiscovery)
+```
+
+"chapAuthDiscovery defines whether support iSCSI Discovery CHAP authentication"
+
+### fn spec.template.spec.volumes.iscsi.withChapAuthSession
+
+```ts
+withChapAuthSession(chapAuthSession)
+```
+
+"chapAuthSession defines whether support iSCSI Session CHAP authentication"
+
+### fn spec.template.spec.volumes.iscsi.withFsType
+
+```ts
+withFsType(fsType)
+```
+
+"fsType is the filesystem type of the volume that you want to mount. Tip: Ensure that the filesystem type is supported by the host operating system. Examples: \"ext4\", \"xfs\", \"ntfs\". Implicitly inferred to be \"ext4\" if unspecified. More info: https://kubernetes.io/docs/concepts/storage/volumes#iscsi TODO: how do we prevent errors in the filesystem from compromising the machine"
+
+### fn spec.template.spec.volumes.iscsi.withInitiatorName
+
+```ts
+withInitiatorName(initiatorName)
+```
+
+"initiatorName is the custom iSCSI Initiator Name. If initiatorName is specified with iscsiInterface simultaneously, new iSCSI interface : will be created for the connection."
+
+### fn spec.template.spec.volumes.iscsi.withIqn
+
+```ts
+withIqn(iqn)
+```
+
+"iqn is the target iSCSI Qualified Name."
+
+### fn spec.template.spec.volumes.iscsi.withIscsiInterface
+
+```ts
+withIscsiInterface(iscsiInterface)
+```
+
+"iscsiInterface is the interface Name that uses an iSCSI transport. Defaults to 'default' (tcp)."
+
+### fn spec.template.spec.volumes.iscsi.withLun
+
+```ts
+withLun(lun)
+```
+
+"lun represents iSCSI Target Lun number."
+
+### fn spec.template.spec.volumes.iscsi.withPortals
+
+```ts
+withPortals(portals)
+```
+
+"portals is the iSCSI Target Portal List. The portal is either an IP or ip_addr:port if the port is other than default (typically TCP ports 860 and 3260)."
+
+### fn spec.template.spec.volumes.iscsi.withPortalsMixin
+
+```ts
+withPortalsMixin(portals)
+```
+
+"portals is the iSCSI Target Portal List. The portal is either an IP or ip_addr:port if the port is other than default (typically TCP ports 860 and 3260)."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.volumes.iscsi.withReadOnly
+
+```ts
+withReadOnly(readOnly)
+```
+
+"readOnly here will force the ReadOnly setting in VolumeMounts. Defaults to false."
+
+### fn spec.template.spec.volumes.iscsi.withTargetPortal
+
+```ts
+withTargetPortal(targetPortal)
+```
+
+"targetPortal is iSCSI Target Portal. The Portal is either an IP or ip_addr:port if the port is other than default (typically TCP ports 860 and 3260)."
+
+## obj spec.template.spec.volumes.iscsi.secretRef
+
+"secretRef is the CHAP Secret for iSCSI target and initiator authentication"
+
+### fn spec.template.spec.volumes.iscsi.secretRef.withName
+
+```ts
+withName(name)
+```
+
+"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"
+
+## obj spec.template.spec.volumes.nfs
+
+"nfs represents an NFS mount on the host that shares a pod's lifetime More info: https://kubernetes.io/docs/concepts/storage/volumes#nfs"
+
+### fn spec.template.spec.volumes.nfs.withPath
+
+```ts
+withPath(path)
+```
+
+"path that is exported by the NFS server. More info: https://kubernetes.io/docs/concepts/storage/volumes#nfs"
+
+### fn spec.template.spec.volumes.nfs.withReadOnly
+
+```ts
+withReadOnly(readOnly)
+```
+
+"readOnly here will force the NFS export to be mounted with read-only permissions. Defaults to false. More info: https://kubernetes.io/docs/concepts/storage/volumes#nfs"
+
+### fn spec.template.spec.volumes.nfs.withServer
+
+```ts
+withServer(server)
+```
+
+"server is the hostname or IP address of the NFS server. More info: https://kubernetes.io/docs/concepts/storage/volumes#nfs"
+
+## obj spec.template.spec.volumes.persistentVolumeClaim
+
+"persistentVolumeClaimVolumeSource represents a reference to a PersistentVolumeClaim in the same namespace. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#persistentvolumeclaims"
+
+### fn spec.template.spec.volumes.persistentVolumeClaim.withClaimName
+
+```ts
+withClaimName(claimName)
+```
+
+"claimName is the name of a PersistentVolumeClaim in the same namespace as the pod using this volume. More info: https://kubernetes.io/docs/concepts/storage/persistent-volumes#persistentvolumeclaims"
+
+### fn spec.template.spec.volumes.persistentVolumeClaim.withReadOnly
+
+```ts
+withReadOnly(readOnly)
+```
+
+"readOnly Will force the ReadOnly setting in VolumeMounts. Default false."
+
+## obj spec.template.spec.volumes.photonPersistentDisk
+
+"photonPersistentDisk represents a PhotonController persistent disk attached and mounted on kubelets host machine"
+
+### fn spec.template.spec.volumes.photonPersistentDisk.withFsType
+
+```ts
+withFsType(fsType)
+```
+
+"fsType is the filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. \"ext4\", \"xfs\", \"ntfs\". Implicitly inferred to be \"ext4\" if unspecified."
+
+### fn spec.template.spec.volumes.photonPersistentDisk.withPdID
+
+```ts
+withPdID(pdID)
+```
+
+"pdID is the ID that identifies Photon Controller persistent disk"
+
+## obj spec.template.spec.volumes.portworxVolume
+
+"portworxVolume represents a portworx volume attached and mounted on kubelets host machine"
+
+### fn spec.template.spec.volumes.portworxVolume.withFsType
+
+```ts
+withFsType(fsType)
+```
+
+"fSType represents the filesystem type to mount Must be a filesystem type supported by the host operating system. Ex. \"ext4\", \"xfs\". Implicitly inferred to be \"ext4\" if unspecified."
+
+### fn spec.template.spec.volumes.portworxVolume.withReadOnly
+
+```ts
+withReadOnly(readOnly)
+```
+
+"readOnly defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts."
+
+### fn spec.template.spec.volumes.portworxVolume.withVolumeID
+
+```ts
+withVolumeID(volumeID)
+```
+
+"volumeID uniquely identifies a Portworx volume"
+
+## obj spec.template.spec.volumes.projected
+
+"projected items for all in one resources secrets, configmaps, and downward API"
+
+### fn spec.template.spec.volumes.projected.withDefaultMode
+
+```ts
+withDefaultMode(defaultMode)
+```
+
+"defaultMode are the mode bits used to set permissions on created files by default. Must be an octal value between 0000 and 0777 or a decimal value between 0 and 511. YAML accepts both octal and decimal values, JSON requires decimal values for mode bits. Directories within the path are not affected by this setting. This might be in conflict with other options that affect the file mode, like fsGroup, and the result can be other mode bits set."
+
+### fn spec.template.spec.volumes.projected.withSources
+
+```ts
+withSources(sources)
+```
+
+"sources is the list of volume projections"
+
+### fn spec.template.spec.volumes.projected.withSourcesMixin
+
+```ts
+withSourcesMixin(sources)
+```
+
+"sources is the list of volume projections"
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.template.spec.volumes.projected.sources
+
+"sources is the list of volume projections"
+
+## obj spec.template.spec.volumes.projected.sources.configMap
+
+"configMap information about the configMap data to project"
+
+### fn spec.template.spec.volumes.projected.sources.configMap.withItems
+
+```ts
+withItems(items)
+```
+
+"items if unspecified, each key-value pair in the Data field of the referenced ConfigMap will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the ConfigMap, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'."
+
+### fn spec.template.spec.volumes.projected.sources.configMap.withItemsMixin
+
+```ts
+withItemsMixin(items)
+```
+
+"items if unspecified, each key-value pair in the Data field of the referenced ConfigMap will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the ConfigMap, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.volumes.projected.sources.configMap.withName
+
+```ts
+withName(name)
+```
+
+"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"
+
+### fn spec.template.spec.volumes.projected.sources.configMap.withOptional
+
+```ts
+withOptional(optional)
+```
+
+"optional specify whether the ConfigMap or its keys must be defined"
+
+## obj spec.template.spec.volumes.projected.sources.configMap.items
+
+"items if unspecified, each key-value pair in the Data field of the referenced ConfigMap will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the ConfigMap, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'."
+
+### fn spec.template.spec.volumes.projected.sources.configMap.items.withKey
+
+```ts
+withKey(key)
+```
+
+"key is the key to project."
+
+### fn spec.template.spec.volumes.projected.sources.configMap.items.withMode
+
+```ts
+withMode(mode)
+```
+
+"mode is Optional: mode bits used to set permissions on this file. Must be an octal value between 0000 and 0777 or a decimal value between 0 and 511. YAML accepts both octal and decimal values, JSON requires decimal values for mode bits. If not specified, the volume defaultMode will be used. This might be in conflict with other options that affect the file mode, like fsGroup, and the result can be other mode bits set."
+
+### fn spec.template.spec.volumes.projected.sources.configMap.items.withPath
+
+```ts
+withPath(path)
+```
+
+"path is the relative path of the file to map the key to. May not be an absolute path. May not contain the path element '..'. May not start with the string '..'."
+
+## obj spec.template.spec.volumes.projected.sources.downwardAPI
+
+"downwardAPI information about the downwardAPI data to project"
+
+### fn spec.template.spec.volumes.projected.sources.downwardAPI.withItems
+
+```ts
+withItems(items)
+```
+
+"Items is a list of DownwardAPIVolume file"
+
+### fn spec.template.spec.volumes.projected.sources.downwardAPI.withItemsMixin
+
+```ts
+withItemsMixin(items)
+```
+
+"Items is a list of DownwardAPIVolume file"
+
+**Note:** This function appends passed data to existing values
+
+## obj spec.template.spec.volumes.projected.sources.downwardAPI.items
+
+"Items is a list of DownwardAPIVolume file"
+
+### fn spec.template.spec.volumes.projected.sources.downwardAPI.items.withMode
+
+```ts
+withMode(mode)
+```
+
+"Optional: mode bits used to set permissions on this file, must be an octal value between 0000 and 0777 or a decimal value between 0 and 511. YAML accepts both octal and decimal values, JSON requires decimal values for mode bits. If not specified, the volume defaultMode will be used. This might be in conflict with other options that affect the file mode, like fsGroup, and the result can be other mode bits set."
+
+### fn spec.template.spec.volumes.projected.sources.downwardAPI.items.withPath
+
+```ts
+withPath(path)
+```
+
+"Required: Path is the relative path name of the file to be created. Must not be absolute or contain the '..' path. Must be utf-8 encoded. The first item of the relative path must not start with '..'"
+
+## obj spec.template.spec.volumes.projected.sources.downwardAPI.items.fieldRef
+
+"Required: Selects a field of the pod: only annotations, labels, name and namespace are supported."
+
+### fn spec.template.spec.volumes.projected.sources.downwardAPI.items.fieldRef.withApiVersion
+
+```ts
+withApiVersion(apiVersion)
+```
+
+"Version of the schema the FieldPath is written in terms of, defaults to \"v1\"."
+
+### fn spec.template.spec.volumes.projected.sources.downwardAPI.items.fieldRef.withFieldPath
+
+```ts
+withFieldPath(fieldPath)
+```
+
+"Path of the field to select in the specified API version."
+
+## obj spec.template.spec.volumes.projected.sources.downwardAPI.items.resourceFieldRef
+
+"Selects a resource of the container: only resources limits and requests (limits.cpu, limits.memory, requests.cpu and requests.memory) are currently supported."
+
+### fn spec.template.spec.volumes.projected.sources.downwardAPI.items.resourceFieldRef.withContainerName
+
+```ts
+withContainerName(containerName)
+```
+
+"Container name: required for volumes, optional for env vars"
+
+### fn spec.template.spec.volumes.projected.sources.downwardAPI.items.resourceFieldRef.withDivisor
+
+```ts
+withDivisor(divisor)
+```
+
+"Specifies the output format of the exposed resources, defaults to \"1\
+
+### fn spec.template.spec.volumes.projected.sources.downwardAPI.items.resourceFieldRef.withResource
+
+```ts
+withResource(resource)
+```
+
+"Required: resource to select"
+
+## obj spec.template.spec.volumes.projected.sources.secret
+
+"secret information about the secret data to project"
+
+### fn spec.template.spec.volumes.projected.sources.secret.withItems
+
+```ts
+withItems(items)
+```
+
+"items if unspecified, each key-value pair in the Data field of the referenced Secret will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the Secret, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'."
+
+### fn spec.template.spec.volumes.projected.sources.secret.withItemsMixin
+
+```ts
+withItemsMixin(items)
+```
+
+"items if unspecified, each key-value pair in the Data field of the referenced Secret will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the Secret, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.volumes.projected.sources.secret.withName
+
+```ts
+withName(name)
+```
+
+"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"
+
+### fn spec.template.spec.volumes.projected.sources.secret.withOptional
+
+```ts
+withOptional(optional)
+```
+
+"optional field specify whether the Secret or its key must be defined"
+
+## obj spec.template.spec.volumes.projected.sources.secret.items
+
+"items if unspecified, each key-value pair in the Data field of the referenced Secret will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the Secret, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'."
+
+### fn spec.template.spec.volumes.projected.sources.secret.items.withKey
+
+```ts
+withKey(key)
+```
+
+"key is the key to project."
+
+### fn spec.template.spec.volumes.projected.sources.secret.items.withMode
+
+```ts
+withMode(mode)
+```
+
+"mode is Optional: mode bits used to set permissions on this file. Must be an octal value between 0000 and 0777 or a decimal value between 0 and 511. YAML accepts both octal and decimal values, JSON requires decimal values for mode bits. If not specified, the volume defaultMode will be used. This might be in conflict with other options that affect the file mode, like fsGroup, and the result can be other mode bits set."
+
+### fn spec.template.spec.volumes.projected.sources.secret.items.withPath
+
+```ts
+withPath(path)
+```
+
+"path is the relative path of the file to map the key to. May not be an absolute path. May not contain the path element '..'. May not start with the string '..'."
+
+## obj spec.template.spec.volumes.projected.sources.serviceAccountToken
+
+"serviceAccountToken is information about the serviceAccountToken data to project"
+
+### fn spec.template.spec.volumes.projected.sources.serviceAccountToken.withAudience
+
+```ts
+withAudience(audience)
+```
+
+"audience is the intended audience of the token. A recipient of a token must identify itself with an identifier specified in the audience of the token, and otherwise should reject the token. The audience defaults to the identifier of the apiserver."
+
+### fn spec.template.spec.volumes.projected.sources.serviceAccountToken.withExpirationSeconds
+
+```ts
+withExpirationSeconds(expirationSeconds)
+```
+
+"expirationSeconds is the requested duration of validity of the service account token. As the token approaches expiration, the kubelet volume plugin will proactively rotate the service account token. The kubelet will start trying to rotate the token if the token is older than 80 percent of its time to live or if the token is older than 24 hours.Defaults to 1 hour and must be at least 10 minutes."
+
+### fn spec.template.spec.volumes.projected.sources.serviceAccountToken.withPath
+
+```ts
+withPath(path)
+```
+
+"path is the path relative to the mount point of the file to project the token into."
+
+## obj spec.template.spec.volumes.quobyte
+
+"quobyte represents a Quobyte mount on the host that shares a pod's lifetime"
+
+### fn spec.template.spec.volumes.quobyte.withGroup
+
+```ts
+withGroup(group)
+```
+
+"group to map volume access to Default is no group"
+
+### fn spec.template.spec.volumes.quobyte.withReadOnly
+
+```ts
+withReadOnly(readOnly)
+```
+
+"readOnly here will force the Quobyte volume to be mounted with read-only permissions. Defaults to false."
+
+### fn spec.template.spec.volumes.quobyte.withRegistry
+
+```ts
+withRegistry(registry)
+```
+
+"registry represents a single or multiple Quobyte Registry services specified as a string as host:port pair (multiple entries are separated with commas) which acts as the central registry for volumes"
+
+### fn spec.template.spec.volumes.quobyte.withTenant
+
+```ts
+withTenant(tenant)
+```
+
+"tenant owning the given Quobyte volume in the Backend Used with dynamically provisioned Quobyte volumes, value is set by the plugin"
+
+### fn spec.template.spec.volumes.quobyte.withUser
+
+```ts
+withUser(user)
+```
+
+"user to map volume access to Defaults to serivceaccount user"
+
+### fn spec.template.spec.volumes.quobyte.withVolume
+
+```ts
+withVolume(volume)
+```
+
+"volume is a string that references an already created Quobyte volume by name."
+
+## obj spec.template.spec.volumes.rbd
+
+"rbd represents a Rados Block Device mount on the host that shares a pod's lifetime. More info: https://examples.k8s.io/volumes/rbd/README.md"
+
+### fn spec.template.spec.volumes.rbd.withFsType
+
+```ts
+withFsType(fsType)
+```
+
+"fsType is the filesystem type of the volume that you want to mount. Tip: Ensure that the filesystem type is supported by the host operating system. Examples: \"ext4\", \"xfs\", \"ntfs\". Implicitly inferred to be \"ext4\" if unspecified. More info: https://kubernetes.io/docs/concepts/storage/volumes#rbd TODO: how do we prevent errors in the filesystem from compromising the machine"
+
+### fn spec.template.spec.volumes.rbd.withImage
+
+```ts
+withImage(image)
+```
+
+"image is the rados image name. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it"
+
+### fn spec.template.spec.volumes.rbd.withKeyring
+
+```ts
+withKeyring(keyring)
+```
+
+"keyring is the path to key ring for RBDUser. Default is /etc/ceph/keyring. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it"
+
+### fn spec.template.spec.volumes.rbd.withMonitors
+
+```ts
+withMonitors(monitors)
+```
+
+"monitors is a collection of Ceph monitors. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it"
+
+### fn spec.template.spec.volumes.rbd.withMonitorsMixin
+
+```ts
+withMonitorsMixin(monitors)
+```
+
+"monitors is a collection of Ceph monitors. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it"
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.volumes.rbd.withPool
+
+```ts
+withPool(pool)
+```
+
+"pool is the rados pool name. Default is rbd. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it"
+
+### fn spec.template.spec.volumes.rbd.withReadOnly
+
+```ts
+withReadOnly(readOnly)
+```
+
+"readOnly here will force the ReadOnly setting in VolumeMounts. Defaults to false. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it"
+
+### fn spec.template.spec.volumes.rbd.withUser
+
+```ts
+withUser(user)
+```
+
+"user is the rados user name. Default is admin. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it"
+
+## obj spec.template.spec.volumes.rbd.secretRef
+
+"secretRef is name of the authentication secret for RBDUser. If provided overrides keyring. Default is nil. More info: https://examples.k8s.io/volumes/rbd/README.md#how-to-use-it"
+
+### fn spec.template.spec.volumes.rbd.secretRef.withName
+
+```ts
+withName(name)
+```
+
+"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"
+
+## obj spec.template.spec.volumes.scaleIO
+
+"scaleIO represents a ScaleIO persistent volume attached and mounted on Kubernetes nodes."
+
+### fn spec.template.spec.volumes.scaleIO.withFsType
+
+```ts
+withFsType(fsType)
+```
+
+"fsType is the filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. \"ext4\", \"xfs\", \"ntfs\". Default is \"xfs\"."
+
+### fn spec.template.spec.volumes.scaleIO.withGateway
+
+```ts
+withGateway(gateway)
+```
+
+"gateway is the host address of the ScaleIO API Gateway."
+
+### fn spec.template.spec.volumes.scaleIO.withProtectionDomain
+
+```ts
+withProtectionDomain(protectionDomain)
+```
+
+"protectionDomain is the name of the ScaleIO Protection Domain for the configured storage."
+
+### fn spec.template.spec.volumes.scaleIO.withReadOnly
+
+```ts
+withReadOnly(readOnly)
+```
+
+"readOnly Defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts."
+
+### fn spec.template.spec.volumes.scaleIO.withSslEnabled
+
+```ts
+withSslEnabled(sslEnabled)
+```
+
+"sslEnabled Flag enable/disable SSL communication with Gateway, default false"
+
+### fn spec.template.spec.volumes.scaleIO.withStorageMode
+
+```ts
+withStorageMode(storageMode)
+```
+
+"storageMode indicates whether the storage for a volume should be ThickProvisioned or ThinProvisioned. Default is ThinProvisioned."
+
+### fn spec.template.spec.volumes.scaleIO.withStoragePool
+
+```ts
+withStoragePool(storagePool)
+```
+
+"storagePool is the ScaleIO Storage Pool associated with the protection domain."
+
+### fn spec.template.spec.volumes.scaleIO.withSystem
+
+```ts
+withSystem(system)
+```
+
+"system is the name of the storage system as configured in ScaleIO."
+
+### fn spec.template.spec.volumes.scaleIO.withVolumeName
+
+```ts
+withVolumeName(volumeName)
+```
+
+"volumeName is the name of a volume already created in the ScaleIO system that is associated with this volume source."
+
+## obj spec.template.spec.volumes.scaleIO.secretRef
+
+"secretRef references to the secret for ScaleIO user and other sensitive information. If this is not provided, Login operation will fail."
+
+### fn spec.template.spec.volumes.scaleIO.secretRef.withName
+
+```ts
+withName(name)
+```
+
+"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"
+
+## obj spec.template.spec.volumes.secret
+
+"secret represents a secret that should populate this volume. More info: https://kubernetes.io/docs/concepts/storage/volumes#secret"
+
+### fn spec.template.spec.volumes.secret.withDefaultMode
+
+```ts
+withDefaultMode(defaultMode)
+```
+
+"defaultMode is Optional: mode bits used to set permissions on created files by default. Must be an octal value between 0000 and 0777 or a decimal value between 0 and 511. YAML accepts both octal and decimal values, JSON requires decimal values for mode bits. Defaults to 0644. Directories within the path are not affected by this setting. This might be in conflict with other options that affect the file mode, like fsGroup, and the result can be other mode bits set."
+
+### fn spec.template.spec.volumes.secret.withItems
+
+```ts
+withItems(items)
+```
+
+"items If unspecified, each key-value pair in the Data field of the referenced Secret will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the Secret, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'."
+
+### fn spec.template.spec.volumes.secret.withItemsMixin
+
+```ts
+withItemsMixin(items)
+```
+
+"items If unspecified, each key-value pair in the Data field of the referenced Secret will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the Secret, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'."
+
+**Note:** This function appends passed data to existing values
+
+### fn spec.template.spec.volumes.secret.withOptional
+
+```ts
+withOptional(optional)
+```
+
+"optional field specify whether the Secret or its keys must be defined"
+
+### fn spec.template.spec.volumes.secret.withSecretName
+
+```ts
+withSecretName(secretName)
+```
+
+"secretName is the name of the secret in the pod's namespace to use. More info: https://kubernetes.io/docs/concepts/storage/volumes#secret"
+
+## obj spec.template.spec.volumes.secret.items
+
+"items If unspecified, each key-value pair in the Data field of the referenced Secret will be projected into the volume as a file whose name is the key and content is the value. If specified, the listed keys will be projected into the specified paths, and unlisted keys will not be present. If a key is specified which is not present in the Secret, the volume setup will error unless it is marked optional. Paths must be relative and may not contain the '..' path or start with '..'."
+
+### fn spec.template.spec.volumes.secret.items.withKey
+
+```ts
+withKey(key)
+```
+
+"key is the key to project."
+
+### fn spec.template.spec.volumes.secret.items.withMode
+
+```ts
+withMode(mode)
+```
+
+"mode is Optional: mode bits used to set permissions on this file. Must be an octal value between 0000 and 0777 or a decimal value between 0 and 511. YAML accepts both octal and decimal values, JSON requires decimal values for mode bits. If not specified, the volume defaultMode will be used. This might be in conflict with other options that affect the file mode, like fsGroup, and the result can be other mode bits set."
+
+### fn spec.template.spec.volumes.secret.items.withPath
+
+```ts
+withPath(path)
+```
+
+"path is the relative path of the file to map the key to. May not be an absolute path. May not contain the path element '..'. May not start with the string '..'."
+
+## obj spec.template.spec.volumes.storageos
+
+"storageOS represents a StorageOS volume attached and mounted on Kubernetes nodes."
+
+### fn spec.template.spec.volumes.storageos.withFsType
+
+```ts
+withFsType(fsType)
+```
+
+"fsType is the filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. \"ext4\", \"xfs\", \"ntfs\". Implicitly inferred to be \"ext4\" if unspecified."
+
+### fn spec.template.spec.volumes.storageos.withReadOnly
+
+```ts
+withReadOnly(readOnly)
+```
+
+"readOnly defaults to false (read/write). ReadOnly here will force the ReadOnly setting in VolumeMounts."
+
+### fn spec.template.spec.volumes.storageos.withVolumeName
+
+```ts
+withVolumeName(volumeName)
+```
+
+"volumeName is the human-readable name of the StorageOS volume. Volume names are only unique within a namespace."
+
+### fn spec.template.spec.volumes.storageos.withVolumeNamespace
+
+```ts
+withVolumeNamespace(volumeNamespace)
+```
+
+"volumeNamespace specifies the scope of the volume within StorageOS. If no namespace is specified then the Pod's namespace will be used. This allows the Kubernetes name scoping to be mirrored within StorageOS for tighter integration. Set VolumeName to any name to override the default behaviour. Set to \"default\" if you are not using namespaces within StorageOS. Namespaces that do not pre-exist within StorageOS will be created."
+
+## obj spec.template.spec.volumes.storageos.secretRef
+
+"secretRef specifies the secret to use for obtaining the StorageOS API credentials. If not specified, default values will be attempted."
+
+### fn spec.template.spec.volumes.storageos.secretRef.withName
+
+```ts
+withName(name)
+```
+
+"Name of the referent. More info: https://kubernetes.io/docs/concepts/overview/working-with-objects/names/#names TODO: Add other useful fields. apiVersion, kind, uid?"
+
+## obj spec.template.spec.volumes.vsphereVolume
+
+"vsphereVolume represents a vSphere volume attached and mounted on kubelets host machine"
+
+### fn spec.template.spec.volumes.vsphereVolume.withFsType
+
+```ts
+withFsType(fsType)
+```
+
+"fsType is filesystem type to mount. Must be a filesystem type supported by the host operating system. Ex. \"ext4\", \"xfs\", \"ntfs\". Implicitly inferred to be \"ext4\" if unspecified."
+
+### fn spec.template.spec.volumes.vsphereVolume.withStoragePolicyID
+
+```ts
+withStoragePolicyID(storagePolicyID)
+```
+
+"storagePolicyID is the storage Policy Based Management (SPBM) profile ID associated with the StoragePolicyName."
+
+### fn spec.template.spec.volumes.vsphereVolume.withStoragePolicyName
+
+```ts
+withStoragePolicyName(storagePolicyName)
+```
+
+"storagePolicyName is the storage Policy Based Management (SPBM) profile name."
+
+### fn spec.template.spec.volumes.vsphereVolume.withVolumePath
+
+```ts
+withVolumePath(volumePath)
+```
+
+"volumePath is the path that identifies vSphere volume vmdk"
\ No newline at end of file
diff --git a/docs/1.21.0/postgresql/v1/scheduledBackup.md b/docs/1.21.0/postgresql/v1/scheduledBackup.md
new file mode 100644
index 0000000..9ab6def
--- /dev/null
+++ b/docs/1.21.0/postgresql/v1/scheduledBackup.md
@@ -0,0 +1,278 @@
+---
+permalink: /1.21.0/postgresql/v1/scheduledBackup/
+---
+
+# postgresql.v1.scheduledBackup
+
+"ScheduledBackup is the Schema for the scheduledbackups API"
+
+## Index
+
+* [`fn new(name)`](#fn-new)
+* [`obj metadata`](#obj-metadata)
+ * [`fn withAnnotations(annotations)`](#fn-metadatawithannotations)
+ * [`fn withAnnotationsMixin(annotations)`](#fn-metadatawithannotationsmixin)
+ * [`fn withClusterName(clusterName)`](#fn-metadatawithclustername)
+ * [`fn withCreationTimestamp(creationTimestamp)`](#fn-metadatawithcreationtimestamp)
+ * [`fn withDeletionGracePeriodSeconds(deletionGracePeriodSeconds)`](#fn-metadatawithdeletiongraceperiodseconds)
+ * [`fn withDeletionTimestamp(deletionTimestamp)`](#fn-metadatawithdeletiontimestamp)
+ * [`fn withFinalizers(finalizers)`](#fn-metadatawithfinalizers)
+ * [`fn withFinalizersMixin(finalizers)`](#fn-metadatawithfinalizersmixin)
+ * [`fn withGenerateName(generateName)`](#fn-metadatawithgeneratename)
+ * [`fn withGeneration(generation)`](#fn-metadatawithgeneration)
+ * [`fn withLabels(labels)`](#fn-metadatawithlabels)
+ * [`fn withLabelsMixin(labels)`](#fn-metadatawithlabelsmixin)
+ * [`fn withName(name)`](#fn-metadatawithname)
+ * [`fn withNamespace(namespace)`](#fn-metadatawithnamespace)
+ * [`fn withOwnerReferences(ownerReferences)`](#fn-metadatawithownerreferences)
+ * [`fn withOwnerReferencesMixin(ownerReferences)`](#fn-metadatawithownerreferencesmixin)
+ * [`fn withResourceVersion(resourceVersion)`](#fn-metadatawithresourceversion)
+ * [`fn withSelfLink(selfLink)`](#fn-metadatawithselflink)
+ * [`fn withUid(uid)`](#fn-metadatawithuid)
+* [`obj spec`](#obj-spec)
+ * [`fn withBackupOwnerReference(backupOwnerReference)`](#fn-specwithbackupownerreference)
+ * [`fn withImmediate(immediate)`](#fn-specwithimmediate)
+ * [`fn withMethod(method)`](#fn-specwithmethod)
+ * [`fn withSchedule(schedule)`](#fn-specwithschedule)
+ * [`fn withSuspend(suspend)`](#fn-specwithsuspend)
+ * [`fn withTarget(target)`](#fn-specwithtarget)
+ * [`obj spec.cluster`](#obj-speccluster)
+ * [`fn withName(name)`](#fn-specclusterwithname)
+
+## Fields
+
+### fn new
+
+```ts
+new(name)
+```
+
+new returns an instance of ScheduledBackup
+
+## obj metadata
+
+"ObjectMeta is metadata that all persisted resources must have, which includes all objects users must create."
+
+### fn metadata.withAnnotations
+
+```ts
+withAnnotations(annotations)
+```
+
+"Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations"
+
+### fn metadata.withAnnotationsMixin
+
+```ts
+withAnnotationsMixin(annotations)
+```
+
+"Annotations is an unstructured key value map stored with a resource that may be set by external tools to store and retrieve arbitrary metadata. They are not queryable and should be preserved when modifying objects. More info: http://kubernetes.io/docs/user-guide/annotations"
+
+**Note:** This function appends passed data to existing values
+
+### fn metadata.withClusterName
+
+```ts
+withClusterName(clusterName)
+```
+
+"The name of the cluster which the object belongs to. This is used to distinguish resources with same name and namespace in different clusters. This field is not set anywhere right now and apiserver is going to ignore it if set in create or update request."
+
+### fn metadata.withCreationTimestamp
+
+```ts
+withCreationTimestamp(creationTimestamp)
+```
+
+"Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers."
+
+### fn metadata.withDeletionGracePeriodSeconds
+
+```ts
+withDeletionGracePeriodSeconds(deletionGracePeriodSeconds)
+```
+
+"Number of seconds allowed for this object to gracefully terminate before it will be removed from the system. Only set when deletionTimestamp is also set. May only be shortened. Read-only."
+
+### fn metadata.withDeletionTimestamp
+
+```ts
+withDeletionTimestamp(deletionTimestamp)
+```
+
+"Time is a wrapper around time.Time which supports correct marshaling to YAML and JSON. Wrappers are provided for many of the factory methods that the time package offers."
+
+### fn metadata.withFinalizers
+
+```ts
+withFinalizers(finalizers)
+```
+
+"Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list."
+
+### fn metadata.withFinalizersMixin
+
+```ts
+withFinalizersMixin(finalizers)
+```
+
+"Must be empty before the object is deleted from the registry. Each entry is an identifier for the responsible component that will remove the entry from the list. If the deletionTimestamp of the object is non-nil, entries in this list can only be removed. Finalizers may be processed and removed in any order. Order is NOT enforced because it introduces significant risk of stuck finalizers. finalizers is a shared field, any actor with permission can reorder it. If the finalizer list is processed in order, then this can lead to a situation in which the component responsible for the first finalizer in the list is waiting for a signal (field value, external system, or other) produced by a component responsible for a finalizer later in the list, resulting in a deadlock. Without enforced ordering finalizers are free to order amongst themselves and are not vulnerable to ordering changes in the list."
+
+**Note:** This function appends passed data to existing values
+
+### fn metadata.withGenerateName
+
+```ts
+withGenerateName(generateName)
+```
+
+"GenerateName is an optional prefix, used by the server, to generate a unique name ONLY IF the Name field has not been provided. If this field is used, the name returned to the client will be different than the name passed. This value will also be combined with a unique suffix. The provided value has the same validation rules as the Name field, and may be truncated by the length of the suffix required to make the value unique on the server.\n\nIf this field is specified and the generated name exists, the server will NOT return a 409 - instead, it will either return 201 Created or 500 with Reason ServerTimeout indicating a unique name could not be found in the time allotted, and the client should retry (optionally after the time indicated in the Retry-After header).\n\nApplied only if Name is not specified. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#idempotency"
+
+### fn metadata.withGeneration
+
+```ts
+withGeneration(generation)
+```
+
+"A sequence number representing a specific generation of the desired state. Populated by the system. Read-only."
+
+### fn metadata.withLabels
+
+```ts
+withLabels(labels)
+```
+
+"Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels"
+
+### fn metadata.withLabelsMixin
+
+```ts
+withLabelsMixin(labels)
+```
+
+"Map of string keys and values that can be used to organize and categorize (scope and select) objects. May match selectors of replication controllers and services. More info: http://kubernetes.io/docs/user-guide/labels"
+
+**Note:** This function appends passed data to existing values
+
+### fn metadata.withName
+
+```ts
+withName(name)
+```
+
+"Name must be unique within a namespace. Is required when creating resources, although some resources may allow a client to request the generation of an appropriate name automatically. Name is primarily intended for creation idempotence and configuration definition. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/identifiers#names"
+
+### fn metadata.withNamespace
+
+```ts
+withNamespace(namespace)
+```
+
+"Namespace defines the space within which each name must be unique. An empty namespace is equivalent to the \"default\" namespace, but \"default\" is the canonical representation. Not all objects are required to be scoped to a namespace - the value of this field for those objects will be empty.\n\nMust be a DNS_LABEL. Cannot be updated. More info: http://kubernetes.io/docs/user-guide/namespaces"
+
+### fn metadata.withOwnerReferences
+
+```ts
+withOwnerReferences(ownerReferences)
+```
+
+"List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller."
+
+### fn metadata.withOwnerReferencesMixin
+
+```ts
+withOwnerReferencesMixin(ownerReferences)
+```
+
+"List of objects depended by this object. If ALL objects in the list have been deleted, this object will be garbage collected. If this object is managed by a controller, then an entry in this list will point to this controller, with the controller field set to true. There cannot be more than one managing controller."
+
+**Note:** This function appends passed data to existing values
+
+### fn metadata.withResourceVersion
+
+```ts
+withResourceVersion(resourceVersion)
+```
+
+"An opaque value that represents the internal version of this object that can be used by clients to determine when objects have changed. May be used for optimistic concurrency, change detection, and the watch operation on a resource or set of resources. Clients must treat these values as opaque and passed unmodified back to the server. They may only be valid for a particular resource or set of resources.\n\nPopulated by the system. Read-only. Value must be treated as opaque by clients and . More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#concurrency-control-and-consistency"
+
+### fn metadata.withSelfLink
+
+```ts
+withSelfLink(selfLink)
+```
+
+"SelfLink is a URL representing this object. Populated by the system. Read-only.\n\nDEPRECATED Kubernetes will stop propagating this field in 1.20 release and the field is planned to be removed in 1.21 release."
+
+### fn metadata.withUid
+
+```ts
+withUid(uid)
+```
+
+"UID is the unique in time and space value for this object. It is typically generated by the server on successful creation of a resource and is not allowed to change on PUT operations.\n\nPopulated by the system. Read-only. More info: http://kubernetes.io/docs/user-guide/identifiers#uids"
+
+## obj spec
+
+"Specification of the desired behavior of the ScheduledBackup. More info: https://git.k8s.io/community/contributors/devel/sig-architecture/api-conventions.md#spec-and-status"
+
+### fn spec.withBackupOwnerReference
+
+```ts
+withBackupOwnerReference(backupOwnerReference)
+```
+
+"Indicates which ownerReference should be put inside the created backup resources.
- none: no owner reference for created backup objects (same behavior as before the field was introduced)
- self: sets the Scheduled backup object as owner of the backup
- cluster: set the cluster as owner of the backup
"
+
+### fn spec.withImmediate
+
+```ts
+withImmediate(immediate)
+```
+
+"If the first backup has to be immediately start after creation or not"
+
+### fn spec.withMethod
+
+```ts
+withMethod(method)
+```
+
+"The backup method to be used, possible options are `barmanObjectStore` and `volumeSnapshot`. Defaults to: `barmanObjectStore`."
+
+### fn spec.withSchedule
+
+```ts
+withSchedule(schedule)
+```
+
+"The schedule does not follow the same format used in Kubernetes CronJobs as it includes an additional seconds specifier, see https://pkg.go.dev/github.com/robfig/cron#hdr-CRON_Expression_Format"
+
+### fn spec.withSuspend
+
+```ts
+withSuspend(suspend)
+```
+
+"If this backup is suspended or not"
+
+### fn spec.withTarget
+
+```ts
+withTarget(target)
+```
+
+"The policy to decide which instance should perform this backup. If empty, it defaults to `cluster.spec.backup.target`. Available options are empty string, `primary` and `prefer-standby`. `primary` to have backups run always on primary instances, `prefer-standby` to have backups run preferably on the most updated standby, if available."
+
+## obj spec.cluster
+
+"The cluster to backup"
+
+### fn spec.cluster.withName
+
+```ts
+withName(name)
+```
+
+"Name of the referent."
\ No newline at end of file
diff --git a/docs/README.md b/docs/README.md
index 20e4393..2762d6c 100644
--- a/docs/README.md
+++ b/docs/README.md
@@ -2,6 +2,7 @@
This library is generated with [`k8s`](https://github.com/jsonnet-libs/k8s).
+- [1.21.0](1.21.0/README.md)
- [1.20.2](1.20.2/README.md)
- [1.19.4](1.19.4/README.md)
- [1.18.5](1.18.5/README.md)