From 094d24c01104b3a9610bfaa4d1769216c7c6a4c2 Mon Sep 17 00:00:00 2001 From: jenkins-x-bot Date: Tue, 2 Jul 2024 09:50:10 +0000 Subject: [PATCH] chore: release 1.19.0 --- changelog.md | 6 + charts/lighthouse/Chart.yaml | 4 +- charts/lighthouse/README.md | 308 ++++++++++++------------ charts/lighthouse/lighthouse-1.19.0.tgz | Bin 0 -> 15826 bytes charts/lighthouse/values.yaml | 151 +----------- 5 files changed, 166 insertions(+), 303 deletions(-) create mode 100644 changelog.md create mode 100644 charts/lighthouse/lighthouse-1.19.0.tgz diff --git a/changelog.md b/changelog.md new file mode 100644 index 000000000..49ff78832 --- /dev/null +++ b/changelog.md @@ -0,0 +1,6 @@ + +## Changes in version 1.19.0 + +### New Features + +* skip running status (running/pending) if last state (JordanGoasdoue) diff --git a/charts/lighthouse/Chart.yaml b/charts/lighthouse/Chart.yaml index edc18364e..a8b53103b 100644 --- a/charts/lighthouse/Chart.yaml +++ b/charts/lighthouse/Chart.yaml @@ -3,6 +3,6 @@ description: | This chart bootstraps installation of [Lighthouse](https://github.com/jenkins-x/lighthouse). icon: https://raw.githubusercontent.com/jenkins-x/jenkins-x-website/master/images/logo/jenkinsx-icon-color.svg name: lighthouse -version: 0.1.0-SNAPSHOT +version: 1.19.0 home: https://github.com/jenkins-x/lighthouse - +appVersion: 1.19.0 diff --git a/charts/lighthouse/README.md b/charts/lighthouse/README.md index ed4700eef..48370b1c7 100644 --- a/charts/lighthouse/README.md +++ b/charts/lighthouse/README.md @@ -42,157 +42,161 @@ helm uninstall my-lighthouse --namespace lighthouse ## Values -| Key | Type | Description | Default | -| --------------------------------------------------- | ------ | ---------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------------- | ---------------------------------------------------------------------------------------- | -| `cluster.crds.create` | bool | Create custom resource definitions | `true` | -| `configMaps.config` | string | Raw `config.yaml` content | `nil` | -| `configMaps.configUpdater` | object | Settings used to configure the `config-updater` plugin | `{"orgAndRepo":"","path":""}` | -| `configMaps.create` | bool | Enables creation of `config.yaml` and `plugins.yaml` config maps | `false` | -| `configMaps.plugins` | string | Raw `plugins.yaml` content | `nil` | -| `engines.jenkins` | bool | Enables the Jenkins engine | `false` | -| `engines.jx` | bool | Enables the jx engine | `true` | -| `engines.tekton` | bool | Enables the tekton engine | `false` | -| `env` | object | Environment variables | `{"JX_DEFAULT_IMAGE":""}` | -| `externalPlugins[0].name` | string | | `"cd-indicators"` | -| `externalPlugins[0].requiredResources[0].kind` | string | | `"Service"` | -| `externalPlugins[0].requiredResources[0].name` | string | | `"cd-indicators"` | -| `externalPlugins[0].requiredResources[0].namespace` | string | | `"jx"` | -| `externalPlugins[1].name` | string | | `"lighthouse-webui-plugin"` | -| `externalPlugins[1].requiredResources[0].kind` | string | | `"Service"` | -| `externalPlugins[1].requiredResources[0].name` | string | | `"lighthouse-webui-plugin"` | -| `externalPlugins[1].requiredResources[0].namespace` | string | | `"jx"` | -| `foghorn.affinity` | object | [Affinity rules](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity) applied to the foghorn pods | `{}` | -| `foghorn.image.pullPolicy` | string | Template for computing the foghorn controller docker image pull policy | `"{{ .Values.image.pullPolicy }}"` | -| `foghorn.image.repository` | string | Template for computing the foghorn controller docker image repository | `"{{ .Values.image.parentRepository }}/lighthouse-foghorn"` | -| `foghorn.image.tag` | string | Template for computing the foghorn controller docker image tag | `"{{ .Values.image.tag }}"` | -| `foghorn.logLevel` | string | The logging level: trace, debug, info, warn, panic, fatal | `"info"` | -| `foghorn.nodeSelector` | object | [Node selector](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#nodeselector) applied to the foghorn pods | `{}` | -| `foghorn.replicaCount` | int | Number of replicas | `1` | -| `foghorn.resources.limits` | object | Resource limits applied to the foghorn pods | `{"cpu":"100m","memory":"256Mi"}` | -| `foghorn.resources.requests` | object | Resource requests applied to the foghorn pods | `{"cpu":"80m","memory":"128Mi"}` | -| `foghorn.terminationGracePeriodSeconds` | int | Termination grace period for foghorn pods | `180` | -| `foghorn.tolerations` | list | [Tolerations](https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/) applied to the foghorn pods | `[]` | -| `gcJobs.backoffLimit` | int | Set the backoff limit for failed cronJobs | `6` | -| `gcJobs.concurrencyPolicy` | string | Drives the job's concurrency policy | `"Forbid"` | -| `gcJobs.failedJobsHistoryLimit` | int | Drives the failed jobs history limit | `1` | -| `gcJobs.image.pullPolicy` | string | Template for computing the gc job docker image pull policy | `"{{ .Values.image.pullPolicy }}"` | -| `gcJobs.image.repository` | string | Template for computing the gc job docker image repository | `"{{ .Values.image.parentRepository }}/lighthouse-gc-jobs"` | -| `gcJobs.image.tag` | string | Template for computing the gc job docker image tag | `"{{ .Values.image.tag }}"` | -| `gcJobs.logLevel` | string | The logging level: trace, debug, info, warn, panic, fatal | `"info"` | -| `gcJobs.maxAge` | string | Max age from which `LighthouseJob`s will be deleted | `"168h"` | -| `gcJobs.schedule` | string | Cron expression to periodically delete `LighthouseJob`s | `"0/30 * * * *"` | -| `gcJobs.successfulJobsHistoryLimit` | int | Drives the successful jobs history limit | `3` | -| `git.kind` | string | Git SCM provider (`github`, `gitlab`, `stash`) | `"github"` | -| `git.server` | string | Git server URL | `""` | -| `githubApp.enabled` | bool | Enables GitHub app authentication | `false` | -| `githubApp.username` | string | GitHub app user name | `"jenkins-x[bot]"` | -| `hmacSecretName` | string | Existing hmac secret to use for webhooks | `""` | -| `hmacToken` | string | Secret used for webhooks | `""` | -| `hmacTokenEnabled` | bool | Enables the use of a hmac token. This should always be enabled if possible - though some git providers don't support it such as bitbucket cloud | `true` | -| `image.parentRepository` | string | Docker registry to pull images from | `"ghcr.io/jenkins-x"` | -| `image.pullPolicy` | string | Image pull policy | `"IfNotPresent"` | -| `image.tag` | string | Docker images tag the following tag is latest on the main branch, it's a specific version on a git tag | `"latest"` | -| `jenkinscontroller.affinity` | object | [Affinity rules](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity) applied to the tekton controller pods | `{}` | -| `jenkinscontroller.image.pullPolicy` | string | Template for computing the tekton controller docker image pull policy | `"{{ .Values.image.pullPolicy }}"` | -| `jenkinscontroller.image.repository` | string | Template for computing the Jenkins controller docker image repository | `"{{ .Values.image.parentRepository }}/lighthouse-jenkins-controller"` | -| `jenkinscontroller.image.tag` | string | Template for computing the tekton controller docker image tag | `"{{ .Values.image.tag }}"` | -| `jenkinscontroller.jenkinsToken` | string | The token for authenticating the Jenkins user | `nil` | -| `jenkinscontroller.jenkinsURL` | string | The URL of the Jenkins instance | `nil` | -| `jenkinscontroller.jenkinsUser` | string | The username for the Jenkins user | `nil` | -| `jenkinscontroller.logLevel` | string | The logging level: trace, debug, info, warn, panic, fatal | `"info"` | -| `jenkinscontroller.nodeSelector` | object | [Node selector](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#nodeselector) applied to the tekton controller pods | `{}` | -| `jenkinscontroller.podAnnotations` | object | Annotations applied to the tekton controller pods | `{}` | -| `jenkinscontroller.resources.limits` | object | Resource limits applied to the tekton controller pods | `{"cpu":"100m","memory":"256Mi"}` | -| `jenkinscontroller.resources.requests` | object | Resource requests applied to the tekton controller pods | `{"cpu":"80m","memory":"128Mi"}` | -| `jenkinscontroller.service` | object | Service settings for the tekton controller | `{"annotations":{}}` | -| `jenkinscontroller.terminationGracePeriodSeconds` | int | Termination grace period for tekton controller pods | `180` | -| `jenkinscontroller.tolerations` | list | [Tolerations](https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/) applied to the tekton controller pods | `[]` | -| `keeper.affinity` | object | [Affinity rules](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity) applied to the keeper pods | `{}` | -| `keeper.datadog.enabled` | string | Enables datadog | `"true"` | -| `keeper.env` | object | Lets you define keeper specific environment variables | `{}` | -| `keeper.image.pullPolicy` | string | Template for computing the keeper controller docker image pull policy | `"{{ .Values.image.pullPolicy }}"` | -| `keeper.image.repository` | string | Template for computing the keeper controller docker image repository | `"{{ .Values.image.parentRepository }}/lighthouse-keeper"` | -| `keeper.image.tag` | string | Template for computing the keeper controller docker image tag | `"{{ .Values.image.tag }}"` | -| `keeper.livenessProbe` | object | Liveness probe configuration | `{"initialDelaySeconds":120,"periodSeconds":10,"successThreshold":1,"timeoutSeconds":1}` | -| `keeper.logLevel` | string | The logging level: trace, debug, info, warn, panic, fatal | `"info"` | -| `keeper.nodeSelector` | object | [Node selector](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#nodeselector) applied to the keeper pods | `{}` | -| `keeper.podAnnotations` | object | Annotations applied to the keeper pods | `{}` | -| `keeper.probe` | object | Liveness and readiness probes settings | `{"path":"/"}` | -| `keeper.readinessProbe` | object | Readiness probe configuration | `{"periodSeconds":10,"successThreshold":1,"timeoutSeconds":1}` | -| `keeper.replicaCount` | int | Number of replicas | `1` | -| `keeper.resources.limits` | object | Resource limits applied to the keeper pods | `{"cpu":"400m","memory":"512Mi"}` | -| `keeper.resources.requests` | object | Resource requests applied to the keeper pods | `{"cpu":"100m","memory":"128Mi"}` | -| `keeper.service` | object | Service settings for the keeper controller | `{"externalPort":80,"internalPort":8888,"type":"ClusterIP"}` | -| `keeper.statusContextLabel` | string | Label used to report status to git provider | `"Lighthouse Merge Status"` | -| `keeper.terminationGracePeriodSeconds` | int | Termination grace period for keeper pods | `30` | -| `keeper.tolerations` | list | [Tolerations](https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/) applied to the keeper pods | `[]` | -| `lighthouseJobNamespace` | string | Namespace where `LighthouseJob`s and `Pod`s are created | Deployment namespace | -| `logFormat` | string | Log format either json or stackdriver | `"json"` | -| `logService` | string | The name of the service registered with logging | `""` | -| `logStackSkip` | string | Comma separated stack frames to skip from the log | `""` | -| `oauthSecretName` | string | Existing Git token secret | `""` | -| `oauthToken` | string | Git token (used when GitHub app authentication is not enabled) | `""` | -| `poller.affinity` | object | [Affinity rules](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity) applied to the poller pods | `{}` | -| `poller.contextMatchPattern` | string | Regex pattern to use to match commit status context | `""` | -| `poller.datadog.enabled` | string | Enables datadog | `"true"` | -| `poller.enabled` | bool | Whether to enable or disable the poller component | `false` | -| `poller.env` | object | Lets you define poller specific environment variables | `{"POLL_HOOK_ENDPOINT":"http://hook/hook/poll","POLL_PERIOD":"20s"}` | -| `poller.image.pullPolicy` | string | Template for computing the poller controller docker image pull policy | `"{{ .Values.image.pullPolicy }}"` | -| `poller.image.repository` | string | Template for computing the poller controller docker image repository | `"{{ .Values.image.parentRepository }}/lighthouse-poller"` | -| `poller.image.tag` | string | Template for computing the poller controller docker image tag | `"{{ .Values.image.tag }}"` | -| `poller.internalPort` | int | | `8888` | -| `poller.livenessProbe` | object | Liveness probe configuration | `{"initialDelaySeconds":120,"periodSeconds":10,"successThreshold":1,"timeoutSeconds":1}` | -| `poller.logLevel` | string | The logging level: trace, debug, info, warn, panic, fatal | `"info"` | -| `poller.nodeSelector` | object | [Node selector](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#nodeselector) applied to the poller pods | `{}` | -| `poller.podAnnotations` | object | Annotations applied to the poller pods | `{}` | -| `poller.probe` | object | Liveness and readiness probes settings | `{"path":"/"}` | -| `poller.readinessProbe` | object | Readiness probe configuration | `{"periodSeconds":10,"successThreshold":1,"timeoutSeconds":1}` | -| `poller.replicaCount` | int | Number of replicas | `1` | -| `poller.requireReleaseSuccess` | bool | Keep polling releases until the most recent commit status is successful | `false` | -| `poller.resources.limits` | object | Resource limits applied to the poller pods | `{"cpu":"400m","memory":"512Mi"}` | -| `poller.resources.requests` | object | Resource requests applied to the poller pods | `{"cpu":"100m","memory":"128Mi"}` | -| `poller.terminationGracePeriodSeconds` | int | Termination grace period for poller pods | `30` | -| `poller.tolerations` | list | [Tolerations](https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/) applied to the poller pods | `[]` | -| `scope` | string | set scope to either `cluster` or `namespace` for permissions | `cluster` | -| `tektoncontroller.affinity` | object | [Affinity rules](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity) applied to the tekton controller pods | `{}` | -| `tektoncontroller.dashboardTemplate` | string | Go template expression for URLs in the dashboard if not using Tekton dashboard | `""` | -| `tektoncontroller.dashboardURL` | string | the dashboard URL (e.g. Tekton dashboard) | `""` | -| `tektoncontroller.image.pullPolicy` | string | Template for computing the tekton controller docker image pull policy | `"{{ .Values.image.pullPolicy }}"` | -| `tektoncontroller.image.repository` | string | Template for computing the tekton controller docker image repository | `"{{ .Values.image.parentRepository }}/lighthouse-tekton-controller"` | -| `tektoncontroller.image.tag` | string | Template for computing the tekton controller docker image tag | `"{{ .Values.image.tag }}"` | -| `tektoncontroller.logLevel` | string | The logging level: trace, debug, info, warn, panic, fatal | `"info"` | -| `tektoncontroller.nodeSelector` | object | [Node selector](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#nodeselector) applied to the tekton controller pods | `{}` | -| `tektoncontroller.podAnnotations` | object | Annotations applied to the tekton controller pods | `{}` | -| `tektoncontroller.replicaCount` | int | Number of replicas | `1` | -| `tektoncontroller.resources.limits` | object | Resource limits applied to the tekton controller pods | `{"cpu":"100m","memory":"256Mi"}` | -| `tektoncontroller.resources.requests` | object | Resource requests applied to the tekton controller pods | `{"cpu":"80m","memory":"128Mi"}` | -| `tektoncontroller.service` | object | Service settings for the tekton controller | `{"annotations":{}}` | -| `tektoncontroller.terminationGracePeriodSeconds` | int | Termination grace period for tekton controller pods | `180` | -| `tektoncontroller.tolerations` | list | [Tolerations](https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/) applied to the tekton controller pods | `[]` | -| `user` | string | Git user name (used when GitHub app authentication is not enabled) | `""` | -| `webhooks.affinity` | object | [Affinity rules](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity) applied to the webhooks pods | `{}` | -| `webhooks.customDeploymentTriggerCommand` | string | deployments can configure the ability to allow custom lighthouse triggers using their own unique chat prefix, for example extending the default `/test` trigger prefix let them specify `customDeploymentTriggerPrefix: foo` which means they can also use their own custom trigger /foo mycoolthing | `""` | -| `webhooks.image.pullPolicy` | string | Template for computing the webhooks controller docker image pull policy | `"{{ .Values.image.pullPolicy }}"` | -| `webhooks.image.repository` | string | Template for computing the webhooks controller docker image repository | `"{{ .Values.image.parentRepository }}/lighthouse-webhooks"` | -| `webhooks.image.tag` | string | Template for computing the webhooks controller docker image tag | `"{{ .Values.image.tag }}"` | -| `webhooks.ingress.annotations` | object | Webhooks ingress annotations | `{}` | -| `webhooks.ingress.enabled` | bool | Enable webhooks ingress | `false` | -| `webhooks.ingress.hosts` | list | Webhooks ingress host names | `[]` | -| `webhooks.ingress.ingressClassName` | string | Webhooks ingress ingressClassName | `nil` | -| `webhooks.labels` | object | allow optional labels to be added to the webhook deployment | `{}` | -| `webhooks.livenessProbe` | object | Liveness probe configuration | `{"initialDelaySeconds":60,"periodSeconds":10,"successThreshold":1,"timeoutSeconds":1}` | -| `webhooks.logLevel` | string | The logging level: trace, debug, info, warn, panic, fatal | `"info"` | -| `webhooks.nodeSelector` | object | [Node selector](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#nodeselector) applied to the webhooks pods | `{}` | -| `webhooks.podAnnotations` | object | Annotations applied to the webhooks pods | `{}` | -| `webhooks.podLabels` | object | | `{}` | -| `webhooks.probe` | object | Liveness and readiness probes settings | `{"path":"/"}` | -| `webhooks.readinessProbe` | object | Readiness probe configuration | `{"periodSeconds":10,"successThreshold":1,"timeoutSeconds":1}` | -| `webhooks.replicaCount` | int | Number of replicas | `1` | -| `webhooks.resources.limits` | object | Resource limits applied to the webhooks pods | `{"cpu":"100m","memory":"512Mi"}` | -| `webhooks.resources.requests` | object | Resource requests applied to the webhooks pods | `{"cpu":"80m","memory":"128Mi"}` | -| `webhooks.service` | object | Service settings for the webhooks controller | `{"annotations":{},"externalPort":80,"internalPort":8080,"type":"ClusterIP"}` | -| `webhooks.serviceName` | string | Allows overriding the service name, this is here for compatibility reasons, regular users should clear this out | `"hook"` | -| `webhooks.terminationGracePeriodSeconds` | int | Termination grace period for webhooks pods | `180` | -| `webhooks.tolerations` | list | [Tolerations](https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/) applied to the webhooks pods | `[]` | +| Key | Type | Description | Default | +|-----|------|-------------|---------| +| `cluster.crds.create` | bool | Create custom resource definitions | `true` | +| `configMaps.config` | string | Raw `config.yaml` content | `nil` | +| `configMaps.configUpdater` | object | Settings used to configure the `config-updater` plugin | `{"orgAndRepo":"","path":""}` | +| `configMaps.create` | bool | Enables creation of `config.yaml` and `plugins.yaml` config maps | `false` | +| `configMaps.plugins` | string | Raw `plugins.yaml` content | `nil` | +| `engines.jenkins` | bool | Enables the Jenkins engine | `false` | +| `engines.jx` | bool | Enables the jx engine | `true` | +| `engines.tekton` | bool | Enables the tekton engine | `false` | +| `env` | object | Environment variables | `{"JX_DEFAULT_IMAGE":""}` | +| `externalPlugins[0].name` | string | | `"cd-indicators"` | +| `externalPlugins[0].requiredResources[0].kind` | string | | `"Service"` | +| `externalPlugins[0].requiredResources[0].name` | string | | `"cd-indicators"` | +| `externalPlugins[0].requiredResources[0].namespace` | string | | `"jx"` | +| `externalPlugins[1].name` | string | | `"lighthouse-webui-plugin"` | +| `externalPlugins[1].requiredResources[0].kind` | string | | `"Service"` | +| `externalPlugins[1].requiredResources[0].name` | string | | `"lighthouse-webui-plugin"` | +| `externalPlugins[1].requiredResources[0].namespace` | string | | `"jx"` | +| `foghorn.affinity` | object | [Affinity rules](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity) applied to the foghorn pods | `{}` | +| `foghorn.image.pullPolicy` | string | Template for computing the foghorn controller docker image pull policy | `"{{ .Values.image.pullPolicy }}"` | +| `foghorn.image.repository` | string | Template for computing the foghorn controller docker image repository | `"{{ .Values.image.parentRepository }}/lighthouse-foghorn"` | +| `foghorn.image.tag` | string | Template for computing the foghorn controller docker image tag | `"{{ .Values.image.tag }}"` | +| `foghorn.logLevel` | string | | `"info"` | +| `foghorn.nodeSelector` | object | [Node selector](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#nodeselector) applied to the foghorn pods | `{}` | +| `foghorn.replicaCount` | int | Number of replicas | `1` | +| `foghorn.resources.limits` | object | Resource limits applied to the foghorn pods | `{"cpu":"100m","memory":"256Mi"}` | +| `foghorn.resources.requests` | object | Resource requests applied to the foghorn pods | `{"cpu":"80m","memory":"128Mi"}` | +| `foghorn.terminationGracePeriodSeconds` | int | Termination grace period for foghorn pods | `180` | +| `foghorn.tolerations` | list | [Tolerations](https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/) applied to the foghorn pods | `[]` | +| `gcJobs.backoffLimit` | int | Drives the job's backoff limit | `6` | +| `gcJobs.concurrencyPolicy` | string | Drives the job's concurrency policy | `"Forbid"` | +| `gcJobs.failedJobsHistoryLimit` | int | Drives the failed jobs history limit | `1` | +| `gcJobs.image.pullPolicy` | string | Template for computing the gc job docker image pull policy | `"{{ .Values.image.pullPolicy }}"` | +| `gcJobs.image.repository` | string | Template for computing the gc job docker image repository | `"{{ .Values.image.parentRepository }}/lighthouse-gc-jobs"` | +| `gcJobs.image.tag` | string | Template for computing the gc job docker image tag | `"{{ .Values.image.tag }}"` | +| `gcJobs.logLevel` | string | | `"info"` | +| `gcJobs.maxAge` | string | Max age from which `LighthouseJob`s will be deleted | `"168h"` | +| `gcJobs.schedule` | string | Cron expression to periodically delete `LighthouseJob`s | `"0/30 * * * *"` | +| `gcJobs.successfulJobsHistoryLimit` | int | Drives the successful jobs history limit | `3` | +| `git.kind` | string | Git SCM provider (`github`, `gitlab`, `stash`) | `"github"` | +| `git.server` | string | Git server URL | `""` | +| `githubApp.enabled` | bool | Enables GitHub app authentication | `false` | +| `githubApp.username` | string | GitHub app user name | `"jenkins-x[bot]"` | +| `hmacSecretName` | string | Existing hmac secret to use for webhooks | `""` | +| `hmacToken` | string | Secret used for webhooks | `""` | +| `hmacTokenEnabled` | bool | Enables the use of a hmac token. This should always be enabled if possible - though some git providers don't support it such as bitbucket cloud | `true` | +| `hmacTokenVolumeMount` | object | Mount hmac token as a volume instead of using an environment variable Secret reference | `{"enabled":false}` | +| `image.parentRepository` | string | Docker registry to pull images from | `"ghcr.io/jenkins-x"` | +| `image.pullPolicy` | string | Image pull policy | `"IfNotPresent"` | +| `image.tag` | string | Docker images tag the following tag is latest on the main branch, it's a specific version on a git tag | `"latest"` | +| `jenkinscontroller.affinity` | object | [Affinity rules](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity) applied to the tekton controller pods | `{}` | +| `jenkinscontroller.image.pullPolicy` | string | Template for computing the tekton controller docker image pull policy | `"{{ .Values.image.pullPolicy }}"` | +| `jenkinscontroller.image.repository` | string | Template for computing the Jenkins controller docker image repository | `"{{ .Values.image.parentRepository }}/lighthouse-jenkins-controller"` | +| `jenkinscontroller.image.tag` | string | Template for computing the tekton controller docker image tag | `"{{ .Values.image.tag }}"` | +| `jenkinscontroller.jenkinsToken` | string | The token for authenticating the Jenkins user | `nil` | +| `jenkinscontroller.jenkinsURL` | string | The URL of the Jenkins instance | `nil` | +| `jenkinscontroller.jenkinsUser` | string | The username for the Jenkins user | `nil` | +| `jenkinscontroller.logLevel` | string | | `"info"` | +| `jenkinscontroller.nodeSelector` | object | [Node selector](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#nodeselector) applied to the tekton controller pods | `{}` | +| `jenkinscontroller.podAnnotations` | object | Annotations applied to the tekton controller pods | `{}` | +| `jenkinscontroller.resources.limits` | object | Resource limits applied to the tekton controller pods | `{"cpu":"100m","memory":"256Mi"}` | +| `jenkinscontroller.resources.requests` | object | Resource requests applied to the tekton controller pods | `{"cpu":"80m","memory":"128Mi"}` | +| `jenkinscontroller.service` | object | Service settings for the tekton controller | `{"annotations":{}}` | +| `jenkinscontroller.terminationGracePeriodSeconds` | int | Termination grace period for tekton controller pods | `180` | +| `jenkinscontroller.tolerations` | list | [Tolerations](https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/) applied to the tekton controller pods | `[]` | +| `keeper.affinity` | object | [Affinity rules](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity) applied to the keeper pods | `{}` | +| `keeper.datadog.enabled` | string | Enables datadog | `"true"` | +| `keeper.env` | object | Lets you define keeper specific environment variables | `{}` | +| `keeper.image.pullPolicy` | string | Template for computing the keeper controller docker image pull policy | `"{{ .Values.image.pullPolicy }}"` | +| `keeper.image.repository` | string | Template for computing the keeper controller docker image repository | `"{{ .Values.image.parentRepository }}/lighthouse-keeper"` | +| `keeper.image.tag` | string | Template for computing the keeper controller docker image tag | `"{{ .Values.image.tag }}"` | +| `keeper.livenessProbe` | object | Liveness probe configuration | `{"initialDelaySeconds":120,"periodSeconds":10,"successThreshold":1,"timeoutSeconds":1}` | +| `keeper.logLevel` | string | | `"info"` | +| `keeper.nodeSelector` | object | [Node selector](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#nodeselector) applied to the keeper pods | `{}` | +| `keeper.podAnnotations` | object | Annotations applied to the keeper pods | `{}` | +| `keeper.probe` | object | Liveness and readiness probes settings | `{"path":"/"}` | +| `keeper.readinessProbe` | object | Readiness probe configuration | `{"periodSeconds":10,"successThreshold":1,"timeoutSeconds":1}` | +| `keeper.replicaCount` | int | Number of replicas | `1` | +| `keeper.resources.limits` | object | Resource limits applied to the keeper pods | `{"cpu":"400m","memory":"512Mi"}` | +| `keeper.resources.requests` | object | Resource requests applied to the keeper pods | `{"cpu":"100m","memory":"128Mi"}` | +| `keeper.service` | object | Service settings for the keeper controller | `{"externalPort":80,"internalPort":8888,"type":"ClusterIP"}` | +| `keeper.statusContextLabel` | string | Label used to report status to git provider | `"Lighthouse Merge Status"` | +| `keeper.terminationGracePeriodSeconds` | int | Termination grace period for keeper pods | `30` | +| `keeper.tolerations` | list | [Tolerations](https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/) applied to the keeper pods | `[]` | +| `lighthouseJobNamespace` | string | Namespace where `LighthouseJob`s and `Pod`s are created | Deployment namespace | +| `logFormat` | string | Log format either json or stackdriver | `"json"` | +| `logService` | string | The name of the service registered with logging | `""` | +| `logStackSkip` | string | Comma separated stack frames to skip from the log | `""` | +| `oauthSecretName` | string | Existing Git token secret | `""` | +| `oauthToken` | string | Git token (used when GitHub app authentication is not enabled) | `""` | +| `oauthTokenVolumeMount` | object | Mount Git token as a volume instead of using an environment variable Secret reference (used when GitHub app authentication is not enabled) | `{"enabled":false}` | +| `poller.affinity` | object | [Affinity rules](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity) applied to the poller pods | `{}` | +| `poller.contextMatchPattern` | string | Regex pattern to use to match commit status context | `""` | +| `poller.datadog.enabled` | string | Enables datadog | `"true"` | +| `poller.enabled` | bool | Whether to enable or disable the poller component | `false` | +| `poller.env` | object | Lets you define poller specific environment variables | `{"POLL_HOOK_ENDPOINT":"http://hook/hook/poll","POLL_PERIOD":"20s"}` | +| `poller.image.pullPolicy` | string | Template for computing the poller controller docker image pull policy | `"{{ .Values.image.pullPolicy }}"` | +| `poller.image.repository` | string | Template for computing the poller controller docker image repository | `"{{ .Values.image.parentRepository }}/lighthouse-poller"` | +| `poller.image.tag` | string | Template for computing the poller controller docker image tag | `"{{ .Values.image.tag }}"` | +| `poller.internalPort` | int | | `8888` | +| `poller.livenessProbe` | object | Liveness probe configuration | `{"initialDelaySeconds":120,"periodSeconds":10,"successThreshold":1,"timeoutSeconds":1}` | +| `poller.logLevel` | string | | `"info"` | +| `poller.nodeSelector` | object | [Node selector](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#nodeselector) applied to the poller pods | `{}` | +| `poller.podAnnotations` | object | Annotations applied to the poller pods | `{}` | +| `poller.probe` | object | Liveness and readiness probes settings | `{"path":"/"}` | +| `poller.readinessProbe` | object | Readiness probe configuration | `{"periodSeconds":10,"successThreshold":1,"timeoutSeconds":1}` | +| `poller.replicaCount` | int | Number of replicas | `1` | +| `poller.requireReleaseSuccess` | bool | Keep polling releases until the most recent commit status is successful | `false` | +| `poller.resources.limits` | object | Resource limits applied to the poller pods | `{"cpu":"400m","memory":"512Mi"}` | +| `poller.resources.requests` | object | Resource requests applied to the poller pods | `{"cpu":"100m","memory":"128Mi"}` | +| `poller.terminationGracePeriodSeconds` | int | Termination grace period for poller pods | `30` | +| `poller.tolerations` | list | [Tolerations](https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/) applied to the poller pods | `[]` | +| `scope` | string | limit permissions to namespace privileges | `"cluster"` | +| `tektoncontroller.affinity` | object | [Affinity rules](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity) applied to the tekton controller pods | `{}` | +| `tektoncontroller.dashboardTemplate` | string | Go template expression for URLs in the dashboard if not using Tekton dashboard | `""` | +| `tektoncontroller.dashboardURL` | string | the dashboard URL (e.g. Tekton dashboard) | `""` | +| `tektoncontroller.image.pullPolicy` | string | Template for computing the tekton controller docker image pull policy | `"{{ .Values.image.pullPolicy }}"` | +| `tektoncontroller.image.repository` | string | Template for computing the tekton controller docker image repository | `"{{ .Values.image.parentRepository }}/lighthouse-tekton-controller"` | +| `tektoncontroller.image.tag` | string | Template for computing the tekton controller docker image tag | `"{{ .Values.image.tag }}"` | +| `tektoncontroller.logLevel` | string | | `"info"` | +| `tektoncontroller.nodeSelector` | object | [Node selector](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#nodeselector) applied to the tekton controller pods | `{}` | +| `tektoncontroller.podAnnotations` | object | Annotations applied to the tekton controller pods | `{}` | +| `tektoncontroller.replicaCount` | int | Number of replicas | `1` | +| `tektoncontroller.resources.limits` | object | Resource limits applied to the tekton controller pods | `{"cpu":"100m","memory":"256Mi"}` | +| `tektoncontroller.resources.requests` | object | Resource requests applied to the tekton controller pods | `{"cpu":"80m","memory":"128Mi"}` | +| `tektoncontroller.service` | object | Service settings for the tekton controller | `{"annotations":{}}` | +| `tektoncontroller.terminationGracePeriodSeconds` | int | Termination grace period for tekton controller pods | `180` | +| `tektoncontroller.tolerations` | list | [Tolerations](https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/) applied to the tekton controller pods | `[]` | +| `user` | string | Git user name (used when GitHub app authentication is not enabled) | `""` | +| `webhooks.affinity` | object | [Affinity rules](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity) applied to the webhooks pods | `{}` | +| `webhooks.customDeploymentTriggerCommand` | string | deployments can configure the ability to allow custom lighthouse triggers using their own unique chat prefix, for example extending the default `/test` trigger prefix let them specify `customDeploymentTriggerPrefix: foo` which means they can also use their own custom trigger /foo mycoolthing | `""` | +| `webhooks.image.pullPolicy` | string | Template for computing the webhooks controller docker image pull policy | `"{{ .Values.image.pullPolicy }}"` | +| `webhooks.image.repository` | string | Template for computing the webhooks controller docker image repository | `"{{ .Values.image.parentRepository }}/lighthouse-webhooks"` | +| `webhooks.image.tag` | string | Template for computing the webhooks controller docker image tag | `"{{ .Values.image.tag }}"` | +| `webhooks.ingress.annotations` | object | Webhooks ingress annotations | `{}` | +| `webhooks.ingress.enabled` | bool | Enable webhooks ingress | `false` | +| `webhooks.ingress.hosts` | list | Webhooks ingress host names | `[]` | +| `webhooks.ingress.ingressClassName` | string | Webhooks ingress ingressClassName | `nil` | +| `webhooks.ingress.tls.enabled` | bool | Enable webhooks ingress tls | `false` | +| `webhooks.ingress.tls.secretName` | string | Specify webhooks ingress tls secretName | `""` | +| `webhooks.labels` | object | allow optional labels to be added to the webhook deployment | `{}` | +| `webhooks.livenessProbe` | object | Liveness probe configuration | `{"initialDelaySeconds":60,"periodSeconds":10,"successThreshold":1,"timeoutSeconds":1}` | +| `webhooks.logLevel` | string | | `"info"` | +| `webhooks.nodeSelector` | object | [Node selector](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#nodeselector) applied to the webhooks pods | `{}` | +| `webhooks.podAnnotations` | object | Annotations applied to the webhooks pods | `{}` | +| `webhooks.podLabels` | object | | `{}` | +| `webhooks.probe` | object | Liveness and readiness probes settings | `{"path":"/"}` | +| `webhooks.readinessProbe` | object | Readiness probe configuration | `{"periodSeconds":10,"successThreshold":1,"timeoutSeconds":1}` | +| `webhooks.replicaCount` | int | Number of replicas | `1` | +| `webhooks.resources.limits` | object | Resource limits applied to the webhooks pods | `{"cpu":"100m","memory":"512Mi"}` | +| `webhooks.resources.requests` | object | Resource requests applied to the webhooks pods | `{"cpu":"80m","memory":"128Mi"}` | +| `webhooks.service` | object | Service settings for the webhooks controller | `{"annotations":{},"externalPort":80,"internalPort":8080,"type":"ClusterIP"}` | +| `webhooks.serviceName` | string | Allows overriding the service name, this is here for compatibility reasons, regular users should clear this out | `"hook"` | +| `webhooks.terminationGracePeriodSeconds` | int | Termination grace period for webhooks pods | `180` | +| `webhooks.tolerations` | list | [Tolerations](https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/) applied to the webhooks pods | `[]` | You can look directly at the [values.yaml](./values.yaml) file to look at the options and their default values. diff --git a/charts/lighthouse/lighthouse-1.19.0.tgz b/charts/lighthouse/lighthouse-1.19.0.tgz new file mode 100644 index 0000000000000000000000000000000000000000..b74c773d067c9aa35457ed0343a1d24805b4f9d4 GIT binary patch literal 15826 zcmV;@JuSi?iwG0|00000|0w_~VMtOiV@ORlOnEsqVl!4SWK%V1T2nbTPgYhoO;>Dc zVQyr3R8em|NM&qo0PKDFbK5x5=>Dv~0-v&5N#>b4eCAWtyn0^8lVrw;{d~;sSE(r* zA|VPfMQ{L6ws(^Ex4*(c5+p&9IxIUgLaH2*1iBknqtWPY1lS+*F^L)K9*rU8od-Ax z{-Bp3ySwuLUawdDzqi-l`&)l^f46tAyT8}(|E<^G+1cCs8|bZ=4r_`ThxBi~ zrEwK2_njmd;a?GDn1ll`?YCeQrH}ni|GzuEmWP;2am3}jmzQJAfGb7{7!tx6r!Zmw zhm6A@fFcthBk=ho9ag`+8gm}8LAUE;K8}YSmrS}}Pr zi57MRO-XJF?>eeP%n)@+$Wh1(h0|Zo9U3ysQFj6vN3@G4&_}Er5T7KO?i~T&aY;aE zhfV!f2nBHl{%N79p1Ly^NXGo1!XQTM5g5Si`MM8N@2c~WQ$(J7({ zJpFWae$oQ^VF23g)(c}o{Y-|R-~_P9Dr7A2_B&65BTuO*c&0Fo>N&`#Ob=ON+0iAIwd zb5x(x+{ptdP#0kn84=pVIjB~FcbO(aw+1Ubk&dV zG2=M&4O~oSXcaP-2>g`<@dSM&aVR@k{m7s|1|XQqJaTnH&=U(&%mgwR0u)X$C7~P@ zQ%JF(87Opsq7kAfbQf$-D~*HlfOrAM6X>R`t&vNNBSOI)8jcCMWvw)qZUgaL(Td`e zZa~0^fK>)nYSR?{ad~|6F_KWo2tWm0Hg5+=Ie?8x9C!c*ckqFMAxdfp@CZbNF)SL@ z0b-u{W5CD+i5ZZx6$U&KzT|+#QA8*Qf=k5iSoGZx^I`1XA`aYu#9pgN;{b3P=bF$u z9x{qoGayOS(jsEQGZMBH zm3h7N+uFrKorOXf(FW55=rs`JQ2N%=maw%7fBE)%LNg|Qyd^4 zv6d_{0Btvj1>NnI{vn#ur5+Ri3x26zcf`An;_i}c^~;SJClf#sBQbRm@X!c{Sjkq~R7oloGlHMv%jx1yVQz z&=$GcN}iC2Ium$*=*yY$5#9sQt#Z!Zjj=l}3r%-82n4hCP=JKA1b`|(0R84&`Q{Zu&1F6K)cu7>46{Azg!JQ5C_N;Ki^{} z=G2MgC1TzP=BO4Tl?i+yLk7kwo8mbw0PFw;pr5Nx?7E1tQ5@7@$}-xqRRbD;9TTHV zLN}&j?Rdy$o`L8K8NOt|$d%2-QqckUiO?bTO!!0S-jdPCRtE)IXGkeG3k<-13xJd{ z8ekQ_pgG=ni6)WYOj4q7$s~%U$RSYquD}L7gT;ZYumnvFZrT3h577Bl?KUfT@wacr zKHu?OM|4BmMxmrZOmt4Pd2XBMx0Y7PPzhoZDQ*pXB|#;I zK!9C%l!#TI;>i$EA%EyEig9O|k{Opm;6fY-rKJ2$&~zr5#szXo=rPrVX$Ig^DJaM+ zIms17#1p7yLK}eon_h85r-aNorc=x=2{|P}fav25@B~@gPIUbkucdab0bfAhc*xgM zvDEm_Q5L`<3gq~KLV5y3Dkxz9^aoDF5(hmGdE&Q09q>|pfEfNV`vCm$O^83lJ29UP zd=c>u!;o+}^dw&M5k^sfOPJD8n=f1juI2b-J9j8*!~mHhO0kzTZX)ap(PIm6vH!vh zNWDnX6OiK}4lsWJ6hTHpwk1T_IDk~@2+|Fk8z4v(2oiIxkI@7SfS}1rpW3CWpA2!) z?M&jX==%(?nvDEG2>nM&rvGpz-=)$ap?m=3^aTKi`Iq`NEVUsuCAR|O(qB4q<&!;@ z$az4&s_8YL^C{Q>xKTU+{a$aPpI?9pd`N^yFd^a(nh?EhKveAbBw)EIfdP0a1avnb zA$keiiDGa=>}RY4Bc@?1Km)MX-~NcJ>n4yRR;XS2vbKIn&Rh-Oq!mnv=x@J~M3~co zlA#fdb@mBK3lwKM_yNQuQV~gt?7}h=zy@CQG^=aamUgnkeF|s|9r&`VN|6 zSCHO?jA1`?BH}p$MfZjHn}B^?H^fY;%4`pgaVT9V18BnR{@gAv@4jn6|DaMD1eM5>tYFZF_Y#3`IiF_#z9bpXG}EvECCl^ z)Xra)l^Jyd1HglnXm7@5%I@_hHKc- zX-tc6=n??S?dLR<#kX?_8FTLD61?l!%xPS0E5-F&xjItk+R5dSH{QtA(Xiw`t|Yc+ zqjPSly)dHcg!V+H-QL?z1?w{Qx#a2+1#^ql2?5VTh%3~`D#IldKRXewKzvDpOWD}d z32;RoSdia1y5%IyX5JbWha$?32ibT?AoZ@!PXtD};wJBac=-x-{0_L31+!1DTL744 z2jJz)8Vo5HavboE0G>P=M-75MU7eo@*``1V9?7G3B}x>DF=uD!dQZ|+zP3{ znVfgwwX#InoQr0elt)`L6)Gn~)t+64`en3+__P}r zElb#*$Ffyf4eJ6{s>HKf$tvqpTGPrL=;;=>YW7u9Q&Y(L#toD1K_=r=QhFWqTer4V4CKFdO@(GP)2?gIG&AEQ6_Cpu_CGb zI^>GPAge4`R8?cWl0_92&qlN;6ZojIMH$`qC|s0l@f}MS7FP6qiWlXIJhgl=i=rnJ zFv>(diiB~CP_!1YLLurf4tdOuNXXGWmsTF~kXinaHus#!rG|2#vWgd`wYsQK3B~}l zGlNI)5m6t33z;j2d|~B+F2CI3uYfdM@pnoKzNTT$RbC@q$0Dy#wQc0qujVy$&b6B7 zP&Zz}YbaQH3D;E3wQlE!OvPfY0c*?f6T}&KAh9yTPeG)n>NW=pCS50{MTkupPN~&a zWo`84%^Tb5kAO4>e;FKI(E=KhsEJx6oi$cT zOM_@6Y317wh11U2>B-6U+41>@)3-UxDzW?|V|$4B9YUcjCK{8X06~V703e$!5h^5n z1LOc~_gDqe^W&4_!;51Rc@v}>ngxVHR#n9-ng9P@9ba5FMOm(z#%lpK6Z!ko(_gNS zKfOIW{qX5hqLr=)iJ*%@FCsXUPTLV-JTvFxBx}(e5Eb`i(CvzR>W@Hb4i{H#R6j!Q zj?W+$Qm&F*d~`45G5JEfiE{!bBA=Ks6P)gWH3;R7Rs@#X5T?9Wn!=x}e!5U%jllQ| zLQ#3_#39Fla&Z+^rpOfoHjf`O!`Ciz+03KJmVuJg9xbfsiR!gGHBx&C_gqqY5&!&B zd%}%YEVY~4r9?zX-T|pUNz^G*jPAe$af)3A5O-S@d##kz?_oob_Ug0T5JZV}&B7ak z=sBs~qk3I?^h_1((QK`H^yI4aXr}%gg}xZ9YDb_SI$!RKu|dAv>yIq}zTE3j3uY0m z7cLajZ_}dc*%aqfl=`r+03bO-%#f=W2?#a$KEg=XAbU2#NF)H|8y&gqM&*=Iux?a* zA?4K1^egopOkl)XS%`^@a1)ctDtH6DD?rxa$<)F+!hY(dkb|I6ryxqCXc+pv9fg@T z-H(puh&)5~4jwwK)+GT`MDgfBt{Pwrr$`9ulp}8o&^R2Z?T=WmTpj?&Ndu?+D7Md9 zYVTxW-M)PTKH>_R6M#01B9SvS0&b_4@o<9Lpp$9B zP*^X&eOomVWvk5sF(~u|+6uaob$IoJSxi3clod%+Ywri`@AmDR7@ofJ`%= zeqtM)B5UF-mmhO+5M(5n6g>50eS~D?|JP#_M2NBuk5&;1w1)rhZ0~Fr`TxQ0!G`~@ zBmMD5_XjY=lL26egAoo8e~8eJ6VWX07!AM=T{#E3KeUeTBO!V5LJ&F`47k(M`J6;4 zS>$?-{`_}YcJO1WXG}A;)0v$sdaA-;pcFzxamYuY{SW5+LljbRtEEG%0{&HOLqZqq z@5rBcauJV4_#U*Kj0`cO9mQ0-KeTi}IhYb9S3K~03)5I_aeG@*BufGeP3Qo8Z5d_>{!P6)_cz;&tPNI-H&Fn?TS<*>urL#KE3Oh~x0B9`h<9y7^EFzpqG+kz ztT1e94^wfViZRaOp*%%S+A^ZZe6>i0D7pJWhRavV4yRGYVm7|^AxC%cA?ZkY!Nava z`!&NFaqOY=2wRyHT&viP=^y|yda+X;joA=O4#Y7s~P6eAN_cA$|v-c^2nGXF!G)5-A`b;95X9z)4ghFwlx z%5dqSe`3k)5ho&pecH#C2U?CYlIx6pEVq#-9*aEgAIWe#H@mG`L{$? zHT`QXtgcWj<#cZ>0)5hki6D<%bm+RGu|75I{6xBkSZwtYSWe@?qmuHCklCZ4)QKf8 z2~l=1Gc$+21;d$P=0CHzVey}X0>$I-H#rw?r$ti3{CqpL-+j2e{^i4`w{7Di z0P>#3yhlQMd7lZj0O;cQ{MY02MxgXveHJr?@@05NJLD0*zxwC(;n~^s#qrVk@#Xc~ z4;8q(Dkv?}w|&AW>dM$DtX&w$$jz3+W9bUe- zB2MdLNJ*4>=Z;H8V^W@KsJCGS5T9xc)*0+sF(@w2W+GBl6VC52o4>^+jvT-_mA zWlH^Fsn*Z`zCJm9cm31p`NzY{TIP`5YFEK%PEOyQUtO#UoQ5Vi*N3Cy2GA03@vqhc z@k%7uC&#}YpVYxm7K7A3vI)`GzaF1od^r773EaNLv4gxkJo@GO;ur1AUegrmJ%0up zA4B#FdPuT?HVUWhWQ_gsOAU|4gQJeN3ILBh7MhNPI_j5PTT3IP#+%!+P|Wrv*Er;bttB%Z)B7r= zAm{&Nvd8M}TqTCoUn<)`CN4@#LN;BRvc@yqi<-fIYQ@N3B+bD#$|BfSu({ROmwlmtlchG2x>Q4&Wp zaI^r(sB4Ua#ltViE>mL;7|3SaMng8oZ z#teJRi5~|@&d6li>#MfgrQ2Yf?6PepAJr+IcDHr3geQJa5gqCxKH~EK05kRD4sv%~ zMwbot#7aT2nF0k_wl8E@K&4F8et!Sdwc4bn$vXeFKaS&mS;zd}+3D>R=l}M`{$oAK zoa#d%0%QgtN0fv=lVRQT&-r*+j6bB2L*iU??-&|4-<9%JVs7A)(2XfYq5EJQ#}{0} zN|_YG5yS!VM2`2EaY7$X@C0)k0DDdav>3E2E@Y7I?zz+w~v zEQTS#Y8+x~X;Z_kt>P?3aV4#`f>dQo0_yEa><|jaF_!Zkvqgc`N+i(&yZWBg(kZ6{ zx??gy-7ohEqbg*v+l^?78<~yd@|LVd3@kHjM{+Q#Ohcq-0G0cgxd%#`hhGW+}_8H)Cgh%wPG>U_vU)2{K+ z33TAInMHMX{)7W$WcEUqe9NO_X$6Pn7Pz6Z#0kNnhuW zHe!nTgT85?P~#?HWiv=M=3BOA&!WCGVW`!)H8fp8YgzpbeATDKs)T4p9Qtn&^a32B zv}(J(UQg1;A{3_4f1i2@{jSqY`EOnN|Lv0gZ+my6|65CXJpF%}v#ebGUq-Q*_P-40 zsk>!uY4-eIm;S$`|Jgg(-pv1Xq({*ITPFTS{l8KFZ_;y2Gx`78^#8^9pZnW8yBq$$ zj(|6f>o`u>x3{*NcnRV(dU=Kub7zc~MQH~Rneq$U=;DkhKBmbNGfGBC@Hg1y3XRdyJI3IRGsnJ#Qr0sd;g`+^FCy81C<)&^2{l zPiOaPo&V)fa7(#@*UbODey`~Ny}i@h%>T8d`XDZ=@DQIB<9SWK;xkDodyV%Oa|*Ai z=(6tNHQ+(<+%AD6dbx0bS9`JhR}h+}`Cwkh@CLLtbP;a=$vpJSqEHV0!P?-Jc_>$P z-@bVs__|pUc`W#P9&hpnm;KlHl{YyLEEGt1c}xRQB#D{@+J|5Y1MH<&sa&0(0L;=1 z|1~T6r+y=fR3l`F&c-Nm$grUEqSgM~cDz!aa7H*lKX#{(CW~1&?>`g~%9+jSJ|KSn zJUJbt$)&+>dbN)Ge`Yt>a<_bs4zXG8b^5;Mu~xgj=JVBZZqpw&Y~T=U*2Hk9Y@72X zKWbHdb5m@Vd!>ogz~|J!nZqH`2BNu#UbzWtef4xZZF`cpJj11J@DoNoVxN^ zf#y8#=7MWF@)baD{Pg}Tsp5d{mP5kv33mB*2#<}?#Ob9Fl zJPKB;i9iDzfbWBEZPIE|Q~tk)rdueGG6{C_D6fj@zq^k+?t=l>Gs z0JZ1;y@R6sx4XTO|JIV`JO7*Obib&g`~DC3WlgG1_$!YqnjHH-S^qbm|FU%!VgPmX z|KOl>|HFQ7qyJe;T6q4;;qc7P{Yx?F%J`FU42yoRrE_yrM{el9cU5*nYDrGT71Pkp zf$~VoIAnC>L+<#HL+T^0P8E$)$A}V6ToMex=8eWvRe4u z@G6Pi0#O0kv|90!ujAi4E}3X<(h=Ni0A7Cnqpg%y?ST<}wY}BWoVh(nE{6X0+shfj zdU);qRok+8vX*Vy0<>k@wg9fWadRejGjGaidQVQx zEjg>bBd7TdISby8GxK(w)!&WN;AWhq@5PyYD^A4)EVHg(d6fHbYH!0SRj2aqmBu!r zvy3+_Kr%_JY_C~wySfNfcDvPOtyUXbmNv61ZDLhAk4!ORybG?g3Qm}NLZRc9F za&3CW3ks+C>6)C>mu9NA7QJ_dgmC8z89FGOI#WpV7TW6mXc&|(h@NqkOuFopZ<6U% z8_urya*#Q1ky+&(GL0_@S?Ku03{P&!?aa> zRW_~aw>GL(4~da$#c?-Bw_I)tt7udiIv$!<(Hh|h1vo_Mr>rbwx2%+}{D#J%wNj~h zvZY>Th2!k3%B-?Rl6y+SgXJW1Yt5@{kFAYRvU{$pOiGBQZiV`oRkhb`)%0f;)N8E9 zn&5>hEv(|60-ttb6~$($n{!X3qa5Z2rttYya2ZE!zKVZ*S~>){>U7 z?ki7xyNyOOo4ZOKt*v;f7TPVSh4RLD)raunySDzRF{iUxK%`ayiCJGYB6MFq?Kds| z=Cczh+Xood2#fT;b^d2BhZMovh~SVj$EDum*#GaA{J;CV2M3$^zm8Oxv97XS?YPuq z9VO(N^ZE!zaEJrUF=Cy=vk#heu+DqPKpRF_jP_8yO81NU2nuw5!U1BPcZh?w+Q!)j zTE{A~JJY@#ec9#Q`jmKS$OW*9kEFkyKes$nGyl~!V2e(Co#xj6Zuk2Ajr_ls^jzlu z7kOjQN+*6b$iMWA&nN5rmuHcy3jka<|M&Jw`~Usjz0LezN76IfjPP4DhFZo4Fn=We z{d)k)`c!!UW|jlb)!VO}tn)t#Vqg0IP1bV%=l*u-{?DC*&HBHVBnD;oht_ErJZPmh z;9~^ZdI+?^kz7_k!U&KM;E-_`1Y3X+z{ik-I|MuuzT|*R5xt|BBY?Tc8Am>a9_qAy z=*k7v5@*~^Ag7YXD?)X!@nlMVU{?yzYg;i*WOWeolz0o5(uJQ1TgrLzR9982AJl`K zoCEN+W#|wrMJ&3K)WEg~NjCu{QJ5Mt zNf#JA!pIXm7Xg?ap}YVNAk^=QtSKBB0`{7rV5AY{Z2DZ7-Gu0l5v32;5q58{<4{xO zWuIj$T&ts(G*F}e9~>o@1}5tsvJOZ2Q+aw(Wr5!|EPf|2%EeM_2;$qfu6nCl+b{`J z15$ll2yU($OMsJ+Ubd~4H4%+gu+U4CR@b^{+OTq}q+V;Dhn3~!wVpyCjDtXaC3n)b zD1qIusR{KDL+>0##Axt@hF0@-G2ndrHYgk(J9=1@+QY1Dp+R1N<9j%X0tCoNgNc>B z8W<4@Ajn*ubpB*;5A?G-o|NO%e8U*_L*s=Dx$)%b`%2Z*7zL5}NCMAzlwDvx`+Q

VNwO2b=hR>q!NPL-o9A?Ys&H@#u-}d?@3c=^!2@(X_t@2XR}QHg^!$rK=poBkPS2 zjU2=)AkT6TkJ2#m4IRW)ff@(#$~u$}8tqwE=^$?FtvMaU?I7lH5Rc{xb>7fHyo#(U z2l1%t+6{|?cvQM?V;%?bXt6ss=Du{_KJHxQsr6CJ2Cf36`(ya5Jb;Wy_+(wtE}= z^Pk5H5K-0~uRK761_9YXk(A3_5d2O2NDQ#$lR1e6sM~Wy8`3SWC!3O_eZScyp8Y)35TPnS7 zR6^P-y`INlsRB@Q)1@YsON9a+WaA-$)N^7Q{5Wgir#Y|APiDbMI;9z~%cKjN*bq$+ zr`Tn)%!rKpwF)*6O%4v{I67=_c)0xW;p}r$EBsPZThCDEIGJpEHrecSa=u9FiYd)L zgDf0A)*U~V4|c>4cYGTX{e z0%mRi7CIA{1@Zgg0!`=tGDiV5^S|FO&;NtX{r~GpbIkwsOnh|`P&fxD1Ypj|NcoV; zslH$7`WJ$u12G$*(r~}&sX=m-kenQp5LLw;F5tArDQj6{DHp1ii!E%>wJLh@5s(e! z2GV>RWI>F-AO81OPYwD1Qb$3x=Rf_T{VHl>HJ?N6m;GE zKR77If8E4?TThzn{Ab<((d#|;DcPfz&^HT2UDN^@DxO`Hy)xePkqVzFHJ*kRJBbo) zD5uR3(0R4C&9$|t{;rEz-JIjv=YQT-TpPVV*8Z=z-`_5-|NHx!`0r~;P5cjLMSsl2 zNi4GiyVt?<2)PEDMS@H!6f;Cq9+%qq>%Ek7|9MFGZ*5CMCgo~kLOzlBW1atZXgDV1 zmaQlP(Cqo&-#OST%m3RO`;WDxRhV}q!)5(u9X8CGqop%3&7%miM}R42s=Qyq9{k1( zpIh(;O8FOZZ&^IiApm7PaOeiHhrlaoj{yGONe;t1|AY*k$*_ac)7S{b+b9R$G>|k}I)cvG@A2wkvb>Xhgh|CY_q)o9S(o50PY+Hf>%qQ_1;*BUQny zCBJc8qMR*flgLL4g*=+e$j3fVKy%}sY5-M1hSCy^!VDF1&mCx}Q2x_K-ChTlkI|P2#Whji7NM;^m|r%& zR~62$0^*`k=PJN09Coe(-s1)Jd*lvo&=D(&|5pWX`L1Xj18`<2BLiGHn9(*^Nid^E z;f&@EXtdgpMsox;S}?59x&s?E4{fwqaHB^FZ`5e0l!Nj-a*(4r21gxjPZ8><`ADdv zesuwl6zt~S2q~Yc^+G5e0B4G$M+>HuqF8-6rL5ws38-Y%|Ia@E!J$tPTm0l#rJD2K zo!;(#@&1SXP5htrq$&k<((|aShM0W`Xo&b7p|?2n3paq8Cr!3Yc{V*ik|E--877%C zr^_%)WLKExvxY`Egb%89mOmJ4daN?jK&7hhpw&}%*E}cEeag^eDix#=qzJVC_mBNf z|BcgY7y99sc!+d68JqEz*fvRwM*(E4v|g0xD&{U3)c}wNgvmDDmEgvNam^SVARZKU z^Ozmz?MJe{iCsl*V>^WIEegH78yFsezb`u4^XsVGl~e}yP1es>$$?;T5d8f$_9Nhh^SIWRPLQ@7uIL3HUQnin@ec{@?jDMXo5mD-Y(R&Z*HzR%ZkIa zX+Aw#Q|M2rzGD7*=fG9k%^~c>y)7#b1sNlltRC!QV?@Vgh1O> zw)O2q($B?y%t<|kAdt!7SiKLW&CRpu+9}~GwheIDk(7lB|Vx86ta5=^dusD*e z4zMu_0^kQ^s9eTy=x+gv0?6?c$-^PzJq*3p3lJh-GFiC#<(lZ1MeF?d@a@Or&cy5ZgioU2 z8E{x*{-@Nf8prJUzq{AlE9(EY5Bi(^-&)cOaFULIR!dHADUN|5A)IjvBL;BDI1BEean;SuZCI`GC6vUBFp@&{) zXf?aL!$FX=fzet3hk%cfp=HWclHiqOY7k5w9OJtVNFejQZna*3_kynJwxUlH030Vl zmM?8M3z|r6t)0Q%Sv_pob|-|~g3m(~kh|Yry^wwk>@^q>s@sHdF;WB~IE6TX!vJY7 z111d|`50jeTtX>g5s@cPz(dTXSY}eAg@(E2<1LAP_q z-jx}Urkx2HizkA@`8rfn86fLM%+8#HM@`kVac6raAfzJeRAZtu9%V;!X?y#ZeZTL{2ca3o&<O#v5xvFfH6)*t9}LN@Gi+BbqzPt(?o2>zJi+&2US3>0#q&;Z2x7tF4 z5x>4!2q-p*ve0EprK~w+zD{D?-zan}tL6y%U?SM;S8y|e0kgE2E>mu@60mZcB@vzU zRJIT-TIFZ;6{y@rQc@WA6~({Y8zqfi&xez$D6C^mRiU2ddiOY-VoJh^sOJ26y!*yrAFa-*ATQ30jxdJYaf>_Sc`Ygd=1==T^? zh-ip7l4q9G>N|ZGa0hZ5W2Ty zG&&Len^pTQ#nVKVA;Xsp=xk~n2srzBFoM#UQWUxm7A9E&XJj`y;7^1OvDeOn9zh%+ zPyBq3ncz6pw5qHE2Lof3Iit9r2fHX&_gw*h6`VXr$o!1lM#z%P+jkw&c2)xC;N3RL z8uDw6vKhP9nJ4i6(9a7YAK^U^SVxphz}*4*BE2p3B=r9Nn{ffU zo_cxsM^s4Y_Yp-*9>Z~>`2u!fkZrz8I0cSgcc%w_Q2&a!lrQus4jK*jEPKU}-$}sh zAVax2yu%z^9DM{4B~$Dn3SQkPJ<-jUED=EYi*d-tH?Jk>>U-OOs>1%1;Ob~VSLY|P zoKYh0?`qisy~F%{JQQpUz?hFw$c1#THtxBtP&uSzk&Fb64CbR_PTmwlRBC8>FU}mK z<9p2H*cN%T@to-Kn5n^^oE7G(nmVNBKq(ArD_&ppxU!K%?+f%~1fU{I`j&R|fy9`^ zfd^o42OpRi;)$;X9)XB3hKB(H4iE#=9|J}v2>6(%1AzgLgfDq=t^q`U#_kwE28Ngq zWA_$u;07f23fo(o6K-xvmv0Pu1xTp9qoqm_K~K~vW>UK(Ho@iXZPJJojgVMDkXrk7 z9oqvrMv~>;&i3+)lVVd8`IvF~Am_5+=dvhMEJP0eu}j6uoQ|J%9*2D;`CwfjK8!vQ zent^PAhQEIyh#32|`7`o#vfcZ<&%1rwp zfvGkfAR&OV-xJoXU?k+|YFcg&dwyFutuit83MEmwT%AP)fHJZb$O2X5%`X-ts;fuZ z-zB3z17BB#IRuBArfn!glnK#q%KZ3JAeEd*5M!=4h1QfyNKTPhiRJ42q=Z8(gk+S1 zDAn7Vdo_W|3d`)JS^`uo<>a7Y9&($>SmkohY;(NO+|S4P%+7$q3Bt!^NPUotkZ;hkVB97=H?<3;Dg+!Vf~|RiG|l5^tXFk?OY_a_MqR}Y8Qi{wFmvJc0Qn4d(bZot!4|k&iu?)@0uCC z*g9PYw7M5z=w%W6RG+eJ8It{}&%nFw0@QpIy27fOgJ>His~QW7oCKsr7Z+5DEv{vO z-GX$o*WWgTffC-uM3M^JMad{Hh-ln%NgrjFq7)B?<$8PYrnl8L?a|wVH*em&*=qBL z2(<_8^hTr-CDI%3h=d&7a~YjKH=X61)PzHDSjvGaBVL&58#(h0Q=X5A`UqUeyai^k z3VEo|{5!n@iYFI*3IwLoPGIV0KUvSRv$*qjV?-_bTxr5`Mm+J8BJDouNfVXd3b(D005f!x)P0em`q?BdSaE#=^7NUpS^{ekN z*Sm48WGz7`@8GjCujK5Iw%XM?WX$B{Bw9jzU4db?{xU_c76Os?r*omfXigIqhh) zph~oc#)kWt1iK-aWXrn%AcT~B-efCz-ppVTmKg$JO}>OFURo zAj!RiN;tn%^kiR60kN8r;@Msv!`1l-c!fHCrv(1>Z1~UWD7u2W`NgXm>XsC*N|&)z3?MS!594q2s^(uUH?tlJWD^Zs86LY)!h zqefX~VBzkJk~wo$hXiC{yKIm9ZZPH&O(KKZM&&qW?bQppRxd#bX2$4^Dx)v6&Yt^i!SBGRB^{_RKk;=u*B9p!$i z*lpw$f)?D1?J_Fy&)i%TBl$JkRM?h0+qPn>-ZY~p-c_1r*upjD_`{)Kk_?b3q7*enXD6G)WO;BxGBFB0mlwl>&U~!s7-AQU!v<>J(E@RANV~ z*eI2{Fxwy5u3meheNl-@wYazbhs08%4dhycDRO!60HBnV!DkhaN5%#D?Nz61JieB( z$pmy+QWq$P=1F97UGy1*D%+$-NJjiOwNMyR`;a1d8xb6G!J#c7*PUsegRHHB)N3V5V&`;6~c)qlGHkJcpH+t&>3MAc&u;%y8xj% zG$n+m_HB{?@JeuN95U=70K*3%??s4mjM$dgvS789IZNNZ&cWpb_)4(?IE6`WaCH6_ z4AB@)F^R=W3^@pi(u80kQpO?jwt$CNgn6D&Ew@N5O(&_}m8xDw!~n=0U&kz(19O28 zurH~@fcV|*UjM(lz5VWXuj^9JVVuS;k0}fs)hgW*Fw9u210fdz72bw2QfXHTFrnX- zIa&9=``ZSPxw=hm!PUr6vnDg3{AW59)WvK#l=6>#wy2QKfv8u-bW%XZgvxDhI*uI> zjE6GdvrEL7Of5RSmURhVivKr^3tu1qx5WQR-W*OD6We{~1yb65nP^X#%ME4d|Q znuXcdi`G;3{oiWl{}&mVWC_7fr&{@MyAc0(duL~FbN|Cy(j&`%)kA0__-zEgzlPxV zGZ|K{NniB0@?xYfQma~pRKbQq-i)*u|DW$dpj!EFx5WQ@JDd9-){@pC|1G}=JeS$^ coeFuIv`L$^NtX2g0{{U3|K5%M#{l{P0KAvJJpcdz literal 0 HcmV?d00001 diff --git a/charts/lighthouse/values.yaml b/charts/lighthouse/values.yaml index 54f43949c..ef8ed8e44 100644 --- a/charts/lighthouse/values.yaml +++ b/charts/lighthouse/values.yaml @@ -1,80 +1,57 @@ git: # git.kind -- Git SCM provider (`github`, `gitlab`, `stash`) kind: github - # git.server -- Git server URL server: "" - # lighthouseJobNamespace -- Namespace where `LighthouseJob`s and `Pod`s are created # @default -- Deployment namespace lighthouseJobNamespace: "" - githubApp: # githubApp.enabled -- Enables GitHub app authentication enabled: false - # githubApp.username -- GitHub app user name - username: "jenkins-x[bot]" - + username: "jenkins-x[bot]" # user -- Git user name (used when GitHub app authentication is not enabled) user: "" - # oauthToken -- Git token (used when GitHub app authentication is not enabled) oauthToken: "" - # oauthSecretName -- Existing Git token secret oauthSecretName: "" - # oauthTokenVolumeMount -- Mount Git token as a volume instead of using an environment variable Secret reference (used when GitHub app authentication is not enabled) oauthTokenVolumeMount: enabled: false - # hmacToken -- Secret used for webhooks hmacToken: "" - # hmacSecretName -- Existing hmac secret to use for webhooks hmacSecretName: "" - # hmacTokenEnabled -- Enables the use of a hmac token. This should always be enabled if possible - though some git providers don't support it such as bitbucket cloud hmacTokenEnabled: true - # hmacTokenVolumeMount -- Mount hmac token as a volume instead of using an environment variable Secret reference hmacTokenVolumeMount: enabled: false - # logFormat -- Log format either json or stackdriver logFormat: "json" - # logService -- The name of the service registered with logging logService: "" - # logStackSkip -- Comma separated stack frames to skip from the log logStackSkip: "" - # scope -- limit permissions to namespace privileges scope: "cluster" - cluster: crds: # cluster.crds.create -- Create custom resource definitions create: true - image: # image.parentRepository -- Docker registry to pull images from parentRepository: ghcr.io/jenkins-x - # image.tag -- Docker images tag # the following tag is latest on the main branch, it's a specific version on a git tag - tag: latest - + tag: 1.19.0 # image.pullPolicy -- Image pull policy pullPolicy: IfNotPresent - # env -- Environment variables env: JX_DEFAULT_IMAGE: "" - - externalPlugins: - name: cd-indicators requiredResources: @@ -86,392 +63,287 @@ externalPlugins: - kind: Service namespace: jx name: lighthouse-webui-plugin - gcJobs: # logLevel -- The logging level: trace, debug, info, warn, error, fatal logLevel: "info" - # gcJobs.maxAge -- Max age from which `LighthouseJob`s will be deleted maxAge: 168h - # gcJobs.schedule -- Cron expression to periodically delete `LighthouseJob`s schedule: "0/30 * * * *" - # gcJobs.failedJobsHistoryLimit -- Drives the failed jobs history limit failedJobsHistoryLimit: 1 - # gcJobs.successfulJobsHistoryLimit -- Drives the successful jobs history limit successfulJobsHistoryLimit: 3 - # gcJobs.concurrencyPolicy -- Drives the job's concurrency policy concurrencyPolicy: Forbid - # gcJobs.backoffLimit -- Drives the job's backoff limit backoffLimit: 6 - image: # gcJobs.image.repository -- Template for computing the gc job docker image repository repository: "{{ .Values.image.parentRepository }}/lighthouse-gc-jobs" - # gcJobs.image.tag -- Template for computing the gc job docker image tag tag: "{{ .Values.image.tag }}" - # gcJobs.image.pullPolicy -- Template for computing the gc job docker image pull policy pullPolicy: "{{ .Values.image.pullPolicy }}" - webhooks: # logLevel -- The logging level: trace, debug, info, warn, error, fatal logLevel: "info" - # webhooks.replicaCount -- Number of replicas replicaCount: 1 - # webhooks.terminationGracePeriodSeconds -- Termination grace period for webhooks pods terminationGracePeriodSeconds: 180 - image: # webhooks.image.repository -- Template for computing the webhooks controller docker image repository repository: "{{ .Values.image.parentRepository }}/lighthouse-webhooks" - # webhooks.image.tag -- Template for computing the webhooks controller docker image tag tag: "{{ .Values.image.tag }}" - # webhooks.image.pullPolicy -- Template for computing the webhooks controller docker image pull policy pullPolicy: "{{ .Values.image.pullPolicy }}" - - # webhooks.labels -- allow optional labels to be added to the webhook deployment labels: {} podLabels: {} - # webhooks.podAnnotations -- Annotations applied to the webhooks pods podAnnotations: {} - # webhooks.serviceName -- Allows overriding the service name, this is here for compatibility reasons, regular users should clear this out serviceName: hook - # webhooks.service -- Service settings for the webhooks controller service: type: ClusterIP externalPort: 80 internalPort: 8080 annotations: {} - resources: # webhooks.resources.limits -- Resource limits applied to the webhooks pods limits: cpu: 100m # may require more memory to perform the initial 'git clone' cmd for big repositories memory: 512Mi - # webhooks.resources.requests -- Resource requests applied to the webhooks pods requests: cpu: 80m memory: 128Mi - # webhooks.probe -- Liveness and readiness probes settings probe: path: / - # webhooks.livenessProbe -- Liveness probe configuration livenessProbe: initialDelaySeconds: 60 periodSeconds: 10 successThreshold: 1 timeoutSeconds: 1 - # webhooks.readinessProbe -- Readiness probe configuration readinessProbe: periodSeconds: 10 successThreshold: 1 timeoutSeconds: 1 - # webhooks.nodeSelector -- [Node selector](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#nodeselector) applied to the webhooks pods nodeSelector: {} - # webhooks.affinity -- [Affinity rules](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity) applied to the webhooks pods affinity: {} - # webhooks.tolerations -- [Tolerations](https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/) applied to the webhooks pods tolerations: [] - ingress: # webhooks.ingress.enabled -- Enable webhooks ingress enabled: false - # webhooks.ingress.annotations -- Webhooks ingress annotations annotations: {} - # webhooks.ingress.ingressClassName -- Webhooks ingress ingressClassName ingressClassName: null - # webhooks.ingress.hosts -- Webhooks ingress host names hosts: [] - tls: # webhooks.ingress.tls.enabled -- Enable webhooks ingress tls enabled: false # webhooks.ingress.tls.secretName -- Specify webhooks ingress tls secretName secretName: "" - # webhooks.customDeploymentTriggerCommand -- deployments can configure the ability to allow custom lighthouse triggers # using their own unique chat prefix, for example extending the default `/test` trigger prefix let them specify # `customDeploymentTriggerPrefix: foo` which means they can also use their own custom trigger /foo mycoolthing customDeploymentTriggerCommand: "" - foghorn: # logLevel -- The logging level: trace, debug, info, warn, error, fatal logLevel: "info" - # foghorn.replicaCount -- Number of replicas replicaCount: 1 - # foghorn.terminationGracePeriodSeconds -- Termination grace period for foghorn pods terminationGracePeriodSeconds: 180 - image: # foghorn.image.repository -- Template for computing the foghorn controller docker image repository repository: "{{ .Values.image.parentRepository }}/lighthouse-foghorn" - # foghorn.image.tag -- Template for computing the foghorn controller docker image tag tag: "{{ .Values.image.tag }}" - # foghorn.image.pullPolicy -- Template for computing the foghorn controller docker image pull policy pullPolicy: "{{ .Values.image.pullPolicy }}" - resources: # foghorn.resources.limits -- Resource limits applied to the foghorn pods limits: cpu: 100m memory: 256Mi - # foghorn.resources.requests -- Resource requests applied to the foghorn pods requests: cpu: 80m memory: 128Mi - # foghorn.nodeSelector -- [Node selector](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#nodeselector) applied to the foghorn pods nodeSelector: {} - # foghorn.affinity -- [Affinity rules](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity) applied to the foghorn pods affinity: {} - # foghorn.tolerations -- [Tolerations](https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/) applied to the foghorn pods tolerations: [] - - tektoncontroller: # logLevel -- The logging level: trace, debug, info, warn, error, fatal logLevel: "info" - # tektoncontroller.dashboardURL -- the dashboard URL (e.g. Tekton dashboard) dashboardURL: '' # tektoncontroller.dashboardTemplate -- Go template expression for URLs in the dashboard if not using Tekton dashboard dashboardTemplate: '' - # tektoncontroller.replicaCount -- Number of replicas replicaCount: 1 - # tektoncontroller.terminationGracePeriodSeconds -- Termination grace period for tekton controller pods terminationGracePeriodSeconds: 180 - image: # tektoncontroller.image.repository -- Template for computing the tekton controller docker image repository repository: "{{ .Values.image.parentRepository }}/lighthouse-tekton-controller" - # tektoncontroller.image.tag -- Template for computing the tekton controller docker image tag tag: "{{ .Values.image.tag }}" - # tektoncontroller.image.pullPolicy -- Template for computing the tekton controller docker image pull policy pullPolicy: "{{ .Values.image.pullPolicy }}" - # tektoncontroller.podAnnotations -- Annotations applied to the tekton controller pods podAnnotations: {} - # tektoncontroller.nodeSelector -- [Node selector](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#nodeselector) applied to the tekton controller pods nodeSelector: {} - # tektoncontroller.affinity -- [Affinity rules](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity) applied to the tekton controller pods affinity: {} - # tektoncontroller.tolerations -- [Tolerations](https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/) applied to the tekton controller pods tolerations: [] - resources: # tektoncontroller.resources.limits -- Resource limits applied to the tekton controller pods limits: cpu: 100m memory: 256Mi - # tektoncontroller.resources.requests -- Resource requests applied to the tekton controller pods requests: cpu: 80m memory: 128Mi - # tektoncontroller.service -- Service settings for the tekton controller service: annotations: {} - jenkinscontroller: # logLevel -- The logging level: trace, debug, info, warn, error, fatal logLevel: "info" - # jenkinscontroller.jenkinsURL -- The URL of the Jenkins instance jenkinsURL: - # jenkinscontroller.jenkinsUser -- The username for the Jenkins user jenkinsUser: - # jenkinscontroller.jenkinsToken -- The token for authenticating the Jenkins user jenkinsToken: - # jenkinscontroller.terminationGracePeriodSeconds -- Termination grace period for tekton controller pods terminationGracePeriodSeconds: 180 - image: # jenkinscontroller.image.repository -- Template for computing the Jenkins controller docker image repository repository: "{{ .Values.image.parentRepository }}/lighthouse-jenkins-controller" - # jenkinscontroller.image.tag -- Template for computing the tekton controller docker image tag tag: "{{ .Values.image.tag }}" - # jenkinscontroller.image.pullPolicy -- Template for computing the tekton controller docker image pull policy pullPolicy: "{{ .Values.image.pullPolicy }}" - # jenkinscontroller.podAnnotations -- Annotations applied to the tekton controller pods podAnnotations: {} - # jenkinscontroller.nodeSelector -- [Node selector](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#nodeselector) applied to the tekton controller pods nodeSelector: {} - # jenkinscontroller.affinity -- [Affinity rules](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity) applied to the tekton controller pods affinity: {} - # jenkinscontroller.tolerations -- [Tolerations](https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/) applied to the tekton controller pods tolerations: [] - resources: # jenkinscontroller.resources.limits -- Resource limits applied to the tekton controller pods limits: cpu: 100m memory: 256Mi - # jenkinscontroller.resources.requests -- Resource requests applied to the tekton controller pods requests: cpu: 80m memory: 128Mi - # jenkinscontroller.service -- Service settings for the tekton controller service: annotations: {} - keeper: # logLevel -- The logging level: trace, debug, info, warn, error, fatal logLevel: "info" - # keeper.statusContextLabel -- Label used to report status to git provider statusContextLabel: "Lighthouse Merge Status" - # keeper.replicaCount -- Number of replicas replicaCount: 1 - # keeper.terminationGracePeriodSeconds -- Termination grace period for keeper pods terminationGracePeriodSeconds: 30 - image: # keeper.image.repository -- Template for computing the keeper controller docker image repository repository: "{{ .Values.image.parentRepository }}/lighthouse-keeper" - # keeper.image.tag -- Template for computing the keeper controller docker image tag tag: "{{ .Values.image.tag }}" - # keeper.image.pullPolicy -- Template for computing the keeper controller docker image pull policy pullPolicy: "{{ .Values.image.pullPolicy }}" - # keeper.podAnnotations -- Annotations applied to the keeper pods podAnnotations: {} - # keeper.env -- Lets you define keeper specific environment variables env: {} - # keeper.service -- Service settings for the keeper controller service: type: ClusterIP externalPort: 80 internalPort: 8888 - resources: # keeper.resources.limits -- Resource limits applied to the keeper pods limits: cpu: 400m memory: 512Mi - # keeper.resources.requests -- Resource requests applied to the keeper pods requests: cpu: 100m memory: 128Mi - # keeper.probe -- Liveness and readiness probes settings probe: path: / - # keeper.livenessProbe -- Liveness probe configuration livenessProbe: initialDelaySeconds: 120 periodSeconds: 10 successThreshold: 1 timeoutSeconds: 1 - # keeper.readinessProbe -- Readiness probe configuration readinessProbe: periodSeconds: 10 successThreshold: 1 timeoutSeconds: 1 - datadog: # keeper.datadog.enabled -- Enables datadog enabled: "true" - # keeper.nodeSelector -- [Node selector](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#nodeselector) applied to the keeper pods nodeSelector: {} - # keeper.affinity -- [Affinity rules](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity) applied to the keeper pods affinity: {} - # keeper.tolerations -- [Tolerations](https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/) applied to the keeper pods tolerations: [] - poller: # logLevel -- The logging level: trace, debug, info, warn, error, fatal logLevel: "info" - # poller.enabled -- Whether to enable or disable the poller component enabled: false - # poller.replicaCount -- Number of replicas replicaCount: 1 - # poller.terminationGracePeriodSeconds -- Termination grace period for poller pods terminationGracePeriodSeconds: 30 - image: # poller.image.repository -- Template for computing the poller controller docker image repository repository: "{{ .Values.image.parentRepository }}/lighthouse-poller" - # poller.image.tag -- Template for computing the poller controller docker image tag tag: "{{ .Values.image.tag }}" - # poller.image.pullPolicy -- Template for computing the poller controller docker image pull policy pullPolicy: "{{ .Values.image.pullPolicy }}" - # poller.podAnnotations -- Annotations applied to the poller pods podAnnotations: {} - # poller.env -- Lets you define poller specific environment variables env: # poller.env.POLL_PERIOD the default time period between polling releases and pull requests POLL_PERIOD: 20s - # poller.env.POLL_RELEASE_PERIOD the time period between polling releases # POLL_RELEASE_PERIOD: 20s @@ -480,77 +352,58 @@ poller: # poller.env.POLL_HOOK_ENDPOINT the hook service endpoint to post webhooks to POLL_HOOK_ENDPOINT: http://hook/hook/poll - # poller.contextMatchPattern -- Regex pattern to use to match commit status context contextMatchPattern: "" - # poller.requireReleaseSuccess -- Keep polling releases until the most recent commit status is successful requireReleaseSuccess: false - resources: # poller.resources.limits -- Resource limits applied to the poller pods limits: cpu: 400m memory: 512Mi - # poller.resources.requests -- Resource requests applied to the poller pods requests: cpu: 100m memory: 128Mi - # poller.probe -- Liveness and readiness probes settings probe: path: / - # keeper.internalPort -- The internal port used to view metrics etc internalPort: 8888 - # poller.livenessProbe -- Liveness probe configuration livenessProbe: initialDelaySeconds: 120 periodSeconds: 10 successThreshold: 1 timeoutSeconds: 1 - # poller.readinessProbe -- Readiness probe configuration readinessProbe: periodSeconds: 10 successThreshold: 1 timeoutSeconds: 1 - datadog: # poller.datadog.enabled -- Enables datadog enabled: "true" - # poller.nodeSelector -- [Node selector](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#nodeselector) applied to the poller pods nodeSelector: {} - # poller.affinity -- [Affinity rules](https://kubernetes.io/docs/concepts/scheduling-eviction/assign-pod-node/#affinity-and-anti-affinity) applied to the poller pods affinity: {} - # poller.tolerations -- [Tolerations](https://kubernetes.io/docs/concepts/scheduling-eviction/taint-and-toleration/) applied to the poller pods tolerations: [] - engines: # engines.jx -- Enables the jx engine jx: true - # engines.tekton -- Enables the tekton engine tekton: false - # engines.jenkins -- Enables the Jenkins engine jenkins: false - configMaps: # configMaps.create -- Enables creation of `config.yaml` and `plugins.yaml` config maps create: false - # configMaps.config -- Raw `config.yaml` content config: null - # configMaps.plugins -- Raw `plugins.yaml` content plugins: null - # configMaps.configUpdater -- Settings used to configure the `config-updater` plugin configUpdater: orgAndRepo: ""