This chart enables the creation of a Selenium Grid Server in Kubernetes.
- Selenium-Grid Helm Chart
- Contents
- Introduction
- Installing the chart
- Enable Selenium Grid Autoscaling
- Updating Selenium-Grid release
- Uninstalling Selenium Grid release
- Ingress Configuration
- Configuration
- Configuration global
- Configuration of Nodes
- Configuration extra scripts mount to container
- Configuration of video recorder and video uploader
- Configuration of Secure Communication (HTTPS)
- Configuration of tracing observability
- Configuration of Selenium Grid chart
- Configuration of KEDA
- Configuration of Ingress NGINX Controller
- Configuration of Jaeger
- Configuration for Selenium-Hub
- Configuration for isolated components
We offer a Helm chart to simplify the deployment of Selenium Grid Docker images to Kubernetes.
- Chart changes are tracked in CHANGELOG.
- Sanity/Regression tests for the chart features are tracked in TESTING.
- There are some reference values file that used to test and deploy Selenium Grid chart. You can find them in tests/charts/refValues and tests/charts/ci.
If you want to install the latest master version of Selenium Grid onto your cluster you can do that by using the helm charts repository located at https://www.selenium.dev/docker-selenium.
# Add docker-selenium helm repository
helm repo add docker-selenium https://www.selenium.dev/docker-selenium
# Update charts from docker-selenium repo
helm repo update
# List all versions present in the docker-selenium repo
helm search repo docker-selenium --versions
# Install basic grid latest version
helm install selenium-grid docker-selenium/selenium-grid
# Or install full grid (Router, Distributor, EventBus, SessionMap and SessionQueue components separated)
helm install selenium-grid docker-selenium/selenium-grid --set isolateComponents=true
# Or install specified version
helm install selenium-grid docker-selenium/selenium-grid --version <version>
# In both cases grid exposed by default using ingress. You may want to set hostname for the grid. Default hostname is selenium-grid.local.
helm install selenium-grid --set ingress.hostname=selenium-grid.k8s.local docker-selenium/chart/selenium-grid/.
# Verify ingress configuration via kubectl get ingress
# Notes: In case you want to set hostname is selenium-grid.local. You need to add the IP and hostname to the local host file in `/etc/hosts`
sudo -- sh -c -e "echo \"$(hostname -i) selenium-grid.local\" >> /etc/hosts"
Nightly chart is built from the latest main branch of this repository with using Nightly images. It is not recommended to use this chart in production. It is only for testing purpose. The procedure to install the Nightly chart is the same as the above, only different on the version, it is 1.0.0-nightly
# List all versions Nightly in the docker-selenium repo
helm search repo docker-selenium --devel
# Install basic grid Nightly version
helm install selenium-grid docker-selenium/selenium-grid --version 1.0.0-nightly
By default, all objects created by the chart will be prefixed with the release name. This is to avoid conflicts with other installations of the chart in the same namespace.
-
If you want to disable this behavior, you can deploy the chart with the release name is
selenium
. -
You can override the component name via
.nameOverride
in a respective component. For examplehub: nameOverride: my-hub-name chromeNode: nameOverride: my-chrome-name
Selenium Grid has the ability to autoscaling browser nodes up/down based on the pending requests in the session queue.
To do this KEDA is used. When enabling
autoscaling using autoscaling.enabling
KEDA is installed automatically. To instead use an existing
installation of KEDA you can enable autoscaling with autoscaling.enableWithExistingKEDA
instead.
KEDA can scale either with
deployments
or jobs and the charts support both types. This
chart support both modes. It is controlled with autoscaling.scalingType
that can be set to either
job (default) or deployment.
There are few settings that are common for both scaling types. These are grouped under autoscaling.scaledOptions
.
In case individual node should be scaled differently, you can override the upstream settings with .scaledOptions
for each node type. For example:
autoscaling:
scaledOptions:
minReplicaCount: 0
maxReplicaCount: 8
pollingInterval: 20
chromeNode:
scaledOptions:
minReplicaCount: 1
maxReplicaCount: 16
pollingInterval: 10
By default, autoscaling.terminationGracePeriodSeconds
is set to 3600 seconds. This is used when scalingType is set to deployment
. You can adjust this value, it will affect to all nodes.
In case individual node which needs to set different period, you can override the upstream settings with .terminationGracePeriodSeconds
for each node type. Note that override value must be greater than upstream setting to take effect. For example:
autoscaling:
terminationGracePeriodSeconds: 3600 #default
chromeNode:
terminationGracePeriodSeconds: 7200 #override
firefoxNode:
terminationGracePeriodSeconds: 1800 #not override
When scaling using deployments the HPA choose pods to terminate randomly. If the chosen pod is currently executing a test rather
than being idle, then there is terminationGracePeriodSeconds
seconds before the test is expected to complete. If your test is
still executing after terminationGracePeriodSeconds
seconds, it would result in failure as the pod will be killed.
During terminationGracePeriodSeconds
period, there is preStop
hook to execute command to wait for the pod can be shut down gracefully which can be defined in .deregisterLifecycle
- There is a
_helpers
template with nameseleniumGrid.node.deregisterLifecycle
render value for podlifecycle.preStop
. By default, hook to execute the script to drain node and wait for current session to complete if any. The script is stored in node ConfigMap, more details can be seen in confignodeConfigMap.
- You can define your custom
preStop
hook which is applied for all nodes viaautoscaling.deregisterLifecycle
- In case individual node which needs different hook, you can override the upstream settings with
.deregisterLifecycle
for each node type. If you want to disable upstream hook in a node, pass the value asfalse
- If an individual node has settings
.lifecycle
itself, it would take the highest precedence to override the above use cases.
autoscaling:
deregisterLifecycle:
preStop:
exec:
command: ["bash", "-c", "echo 'Your custom preStop hook applied for all nodes'"]
chromeNode:
deregisterLifecycle: false #disable upstream hook in chrome node
firefoxNode:
deregisterLifecycle:
preStop:
exec:
command: ["bash", "-c", "echo 'Your custom preStop hook specific for firefox node'"]
edgeNode:
lifecycle:
preStop:
exec:
command: ["bash", "-c", "echo 'preStop hook is defined in edge node lifecycle itself'"]
For other settings that KEDA ScaledObject spec supports, you can set them via autoscaling.scaledObjectOptions
. For example:
autoscaling:
scaledObjectOptions:
cooldownPeriod: 60
Settings that KEDA ScaledJob spec supports can be set via autoscaling.scaledJobOptions
.
Once you have a new chart version, you can update your selenium-grid running:
helm upgrade selenium-grid docker-selenium/selenium-grid
If needed, you can add sidecars for your browser nodes by running:
helm upgrade selenium-grid docker-selenium/selenium-grid --set 'firefoxNode.enabled=true' --set-json 'firefoxNode.sidecars=[{"name":"my-sidecar","image":"my-sidecar:latest","imagePullPolicy":"IfNotPresent","ports":[{"containerPort":8080, "protocol":"TCP"}],"resources":{"limits":{"memory": "128Mi"},"requests":{"cpu": "100m"}}}]'
Note: the parameter used for --set-json is just an example, please refer to Container Spec for an overview of usable parameters.
To uninstall:
helm uninstall selenium-grid
By default, ingress is enabled without annotations set. If NGINX ingress controller is used, you need to set few annotations to override the default timeout values to avoid 504 errors (see #1808). Since in Selenium Grid the default of SE_NODE_SESSION_TIMEOUT
and SE_SESSION_REQUEST_TIMEOUT
is 300
seconds.
To make the user experience better, there are few annotations will be set by default if NGINX ingress controller is used. Mostly relates to timeouts and buffer sizes.
If you are not using NGINX ingress controller, you can disable these default annotations by setting ingress.nginx
to nil
(aka null) via Helm CLI --set ingress.nginx=!
) or via an override-values.yaml as below:
ingress:
nginx: !
Similarly, if you want to disable a sub-config of ingress.nginx
. For example: --set ingress.nginx.proxyBuffer=null
)
You are also able to combine using both default annotations and your own annotations in ingress.annotations
. Duplicated keys will be merged strategy overwrite with your own annotations in ingress.annotations
take precedence.
ingress:
nginx:
proxyTimeout: 3600
annotations:
nginx.ingress.kubernetes.io/proxy-connect-timeout: "7200" # This key will take 7200 instead of 3600
List mapping of chart values and default annotation(s)
# `ingress.nginx.proxyTimeout` pass value to annotation(s)
nginx.ingress.kubernetes.io/proxy-connect-timeout
nginx.ingress.kubernetes.io/proxy-send-timeout
nginx.ingress.kubernetes.io/proxy-read-timeout
nginx.ingress.kubernetes.io/proxy-next-upstream-timeout
nginx.ingress.kubernetes.io/auth-keepalive-timeout
# `ingress.nginx.proxyBuffer` pass value to to annotation(s)
nginx.ingress.kubernetes.io/proxy-request-buffering: "on"
nginx.ingress.kubernetes.io/proxy-buffering: "on"
# `ingress.nginx.proxyBuffer.size` pass value to to annotation(s)
nginx.ingress.kubernetes.io/proxy-buffer-size
nginx.ingress.kubernetes.io/client-body-buffer-size
# `ingress.nginx.proxyBuffer.number` pass value to annotation(s)
nginx.ingress.kubernetes.io/proxy-buffers-number
You can generate a test double self-signed certificate specify for your hostname
, assign it to spec ingress.tls
and NGINX ingress controller default certificate (if it is enabled inline). For example:
tls:
ingress:
generateTLS: true
ingress:
hostname: "your.domain.com"
ingress-nginx:
enabled: true
controller:
extraArgs:
default-ssl-certificate: '$(POD_NAMESPACE)/selenium-tls-secret'
For now, global configuration supported is:
Parameter | Default | Description |
---|---|---|
global.K8S_PUBLIC_IP |
"" |
Public IP of the host running K8s |
global.seleniumGrid.imageRegistry |
selenium |
Distribution registry to pull images |
global.seleniumGrid.imageTag |
4.18.1-20240224 |
Image tag for all selenium components |
global.seleniumGrid.nodesImageTag |
4.18.1-20240224 |
Image tag for browser's nodes |
global.seleniumGrid.videoImageTag |
ffmpeg-6.1-20240224 |
Image tag for browser's video recorder |
global.seleniumGrid.imagePullSecret |
"" |
Pull secret to be used for all images |
global.seleniumGrid.imagePullSecret |
"" |
Pull secret to be used for all images |
global.seleniumGrid.affinity |
{} |
Affinity assigned globally |
global.seleniumGrid.logLevel |
INFO |
Set log level for all components |
global.seleniumGrid.defaultNodeStartupProbe |
exec |
Default startup probe method in Nodes |
global.seleniumGrid.defaultNodeLivenessProbe |
exec |
Default liveness probe method in Nodes |
global.seleniumGrid.stdoutProbeLog |
true |
Enable probe logs output in kubectl logs |
This is the public IP of the host running Kubernetes cluster. Mainly, it is used to construct the URL for the Selenium Grid (Hub or Router) can be accessed from the outside of the cluster for Node register, Grid UI, RemoteWebDriver, etc.
- Ingress is enabled without setting
ingress.hostname
. All the services will be exposed via the public IP is set inK8S_PUBLIC_IP
. - Using NodePort to expose the services. All the services will be exposed via the public IP is set in
K8S_PUBLIC_IP
. - Using LoadBalancer to expose the services. All the services will be exposed via the LB External IP is set in
K8S_PUBLIC_IP
.
For example:
global:
K8S_PUBLIC_IP: "10.10.10.10"
ingress:
enabled: true
hostname: ""
hub:
subPath: "/selenium"
serviceType: NodePort
# Source: selenium-grid/templates/node-configmap.yaml
SE_NODE_GRID_URL: 'http://admin:[email protected]/selenium'
Besides that, from the outside of the cluster, you can access via NodePort http://10.10.10.10:30444/selenium
By default, Node will use port 5555
to listen on container (following this) and expose via Service. You can update this value via .port
in respective node type. This will be used to set SE_NODE_PORT
environment variable to pass to option --port
when starting the node and update in Service accordingly.
By default, if httpGet probes are enabled, it will use .port
value in respective node type unless you override it via e.g. .startupProbe.port
.readinessProbe.port
or .livenessProbe.port
in respective node type.
In a node container, there are other running services can be exposed. For example: VNC, NoVNC, SSH, etc. You can easily expose them on container via .ports
and on Service service.ports
in respective node type.
chromeNode:
port: 6666 # Update `SE_NODE_PORT` to 6666
nodePort: 30666 # Specify a NodePort to expose `SE_NODE_PORT` to outside traffic
ports:
- 5900 # You can give port number alone, default protocol is TCP
- 7900
service:
type: NodePort # Expose entire ports on Service via NodePort
ports:
- name: vnc-port
protocol: TCP
port: 5900
targetPort: 5900
nodePort: 30590 # Specify a NodePort to expose VNC port
- name: novnc-port
protocol: TCP
port: 7900
targetPort: 7900
# NodePort will be assigned randomly if not set
edgeNode:
ports: # You also can give objects following manifest of container ports
- containerPort: 5900
name: vnc
protocol: TCP
- containerPort: 7900
name: novnc
protocol: TCP
By default, startupProbe
is enabled and readinessProbe
and livenessProbe
are disabled. You can enable/disable them via .startupProbe.enabled
.readinessProbe.enabled
.livenessProbe.enabled
in respective node type.
By default, probes are using httpGet
method to check the node state. It will use .port
value in respective node type unless you override it via e.g. .startupProbe.port
.readinessProbe.port
or .livenessProbe.port
in respective node type.
Other settings of probe support to override under .startupProbe
.readinessProbe
.livenessProbe
in respective node type.
schema
path
port
initialDelaySeconds
failureThreshold
timeoutSeconds
periodSeconds
successThreshold
You can configure the probes (as Kubernetes supports) to override the default settings. For example:
edgeNode:
port: 5555
startupProbe:
enabled: true
tcpSocket:
port: 5555
failureThreshold: 10
periodSeconds: 5
This is supported for containers of browser node, video recorder and video uploader. By default, in these containers, there are scripts, config files implemented. In case you want to customize or replace them with your own implementation. Instead of forking the chart, use volume mount. Now, from your external files, you can insert them into ConfigMap via Helm CLI --set-file
or compose them in your own YAML values file and pass to Helm CLI --values
when deploying chart. Any files name that you defined will be picked up into ConfigMap and mounted to the container.
nodeConfigMap:
extraScriptsDirectory: "/opt/selenium"
extraScripts:
nodePreStop.sh: |
#!/bin/bash
echo "Your custom script"
recorderConfigMap:
extraScriptsDirectory: "/opt/bin"
extraScripts:
video.sh: |
#!/bin/bash
echo "Your custom script"
video_graphQLQuery.sh: |
#!/bin/bash
echo "My new script"
uploaderConfigMap:
extraScriptsDirectory: "/opt/bin"
extraScripts:
upload.sh: |
#!/bin/bash
echo "Your custom entry point"
secretFiles:
upload.conf: |
[myremote]
type = s3
Via Helm CLI, you can pass your own files to particular config key. Note that, the file name contains dot .
for file extension, it will impact to the key name convention in Helm CLI. In this case, be careful to escape the dot .
in the file name. For example a command in Unix:
helm upgrade -i test \
--set-file 'nodeConfigMap.extraScripts.nodePreStop\.sh=/path/to/myScript.sh' \
--set-file 'recorderConfigMap.extraScripts.video\.sh=/path/to/myCustom.sh' \
selenium-grid
Files in .extraScripts
will be mounted to the container with the same name within directory is defined in .extraScriptsDirectory
. For example, in the above config, nodePreStop.sh
will be mounted to /opt/selenium/nodePreStop.sh
in the node container.
The video recorder is a sidecar deployed with the browser nodes. It is responsible for recording the video of the browser session. The video recorder is disabled by default. To enable it, you need to set the following values:
videoRecorder:
enabled: true
At chart deployment level, that config will enable video container always. In addition, you can disable video recording process via session capability se:recordVideo
. For example in Python binding:
from selenium.webdriver.chrome.options import Options as ChromeOptions
from selenium import webdriver
options = ChromeOptions()
options.set_capability('se:recordVideo', False)
driver = webdriver.Remote(options=options, command_executor="http://localhost:4444")
In Node will perform query GraphQL in Hub based on Node SessionId and extract the value of se:recordVideo
in capabilities before deciding to start video recording process or not. You can customize by reading on section Configuration extra scripts mount to container.
The uploader is extra utility in the video container. It is responsible for uploading the video to a remote location. The uploader is disabled by default. To enable it, you need to set the following values:
videoRecorder:
uploader:
enabled: true
By default, the uploader uses RCLONE to upload the video to a remote location. RCLONE requires a configuration file to define different remote locations. Refer to RCLONE docs for more details. Config file might contain sensitive information such as access key, secret key, etc. hence it is stored in Secret.
The uploader requires destinationPrefix
to be set. It is used to instruct the uploader where to upload the video. The format of destinationPrefix is remote-name://bucket-name/path
. The remote-name
is configured in RCLONE. The bucket-name
is the name of the bucket in the remote location. The path
is the path to the folder in the bucket.
By default, the config file is empty. You can override the config file via --set-file uploaderConfigMap.secretFiles.upload\.conf=/path/to/your_config.conf
or set via YAML values.
For example, to configure an S3 remote hosted on AWS with named mys3
and the bucket name is mybucket
, you can set the following values:
uploaderConfigMap:
secretFiles:
upload.conf: |
[mys3]
type = s3
provider = AWS
env_auth = true
region = ap-southeast-1
location_constraint = ap-southeast-1
acl = private
access_key_id = xxx
secret_access_key = xxx
videoRecorder:
uploader:
destinationPrefix: "mys3://mybucket/subFolder"
You can prepare a config file with multiple remotes are defined. Ensure that [remoteName]
is unique for each remote.
Instead of using config file, another way that RCLONE also supports to pass the information via environment variables. ENV variable with format: RCLONE_CONFIG_ + name of remote + _ + name of config file option
(make it all uppercase). In this case the remote name it can only contain letters, digits, or the _ (underscore) character. All those ENV variables can be set via videoRecorder.uploader.secrets
, it will be stored in Secret.
For example, the same above config can be set via ENV vars as below:
videoRecorder:
uploader:
destinationPrefix: "mys3://mybucket"
secrets:
RCLONE_CONFIG_MYS3_TYPE: "s3"
RCLONE_CONFIG_MYS3_PROVIDER: "GCS"
RCLONE_CONFIG_MYS3_ENV_AUTH: "true"
RCLONE_CONFIG_MYS3_REGION: "asia-southeast1"
RCLONE_CONFIG_MYS3_LOCATION_CONSTRAINT: "asia-southeast1"
RCLONE_CONFIG_MYS3_ACL: "private"
RCLONE_CONFIG_MYS3_ACCESS_KEY_ID: "xxx"
RCLONE_CONFIG_MYS3_SECRET_ACCESS_KEY: "xxx"
RCLONE_CONFIG_MYS3_ENDPOINT: "https://storage.googleapis.com"
RCLONE_CONFIG_MYS3_NO_CHECK_BUCKET: "true"
Those two ways are equivalent. You can choose one of them or combine them. When both config file and ENV vars are set, value in upload.conf
will take precedence.
Besides the configuration, the script for entry point of uploader container also needed. You can override the script via --set-file uploaderConfigMap.extraScripts.upload\.sh=/path/to/your_script.sh
or set via YAML values. For example:
uploaderConfigMap:
extraScripts:
upload.sh: |
#!/bin/bash
echo "Your custom entry point"
In case you want to configure another sidecar container for uploader, you can set a name for videoRecorder.uploader.name
and create a config key with the same name under videoRecorder
with all the settings for your container. Set name of videoRecorder.uploader.entryPointFileName
if your container start by a different entry point. For example:
uploaderConfigMap:
extraScripts:
upload.sh: |
#!/bin/bash
echo "Script control the uploader process"
videoRecorder:
enabled: true
uploader:
enabled: true
name: "s3"
entryPointFileName: "upload.sh"
destinationPrefix: "s3://mybucket"
secrets:
AWS_REGION: "ap-southeast-1"
AWS_ACCESS_KEY_ID: "xxxx"
AWS_SECRET_ACCESS_KEY: "xxxx"
s3:
imageRegistry: public.ecr.aws
imageName: bitnami/aws-cli
imageTag: latest
Selenium Grid supports secure communication between components. Refer to the instructions and options are able to configure the secure communication. Below is the details on how to enable secure communication in Selenium Grid chart.
In the chart, there is directory certs contains the default certificate, private key (as PKCS8 format), and Java Keystore (JKS) to teach Java about secure connection (since we are using a non-standard CA) for your trial, local testing purpose. You can generate your own self-signed certificate put them in that default directory by using script cert.sh with adjust needed information. The certificate, private key, truststore are mounted to the components via Secret
.
There are multiple ways to configure your certificate, private key, truststore to the components. You can choose one of them or combine them.
-
Use the default directory certs. Rename your own files to be same as the default files and replace them. Give
--set tls.enabled=true
to enable secure communication. -
Use the default directory certs. Copy your own files to there and adjust the file name under config
tls.defaultFile
, those will be picked up when installing chart. For example:tls: enabled: true trustStorePassword: "your_truststore_password" defaultFile: certificate: "certs/your_cert.pem" privateKey: "certs/your_private_key.pkcs8" trustStore: "certs/your_truststore.jks"
For some security reasons, you may not able to put private key in your source code or your customization chart package. You can provide files with contents are encoded in Base64 format, just append
.base64
to the file name for chart able to know and decode them. For example:tls: enabled: true trustStorePassword: "your_truststore_password" defaultFile: certificate: "certs/your_cert.pem.base64" privateKey: "certs/your_private_key.pkcs8.base64" trustStore: "certs/your_truststore.jks.base64"
-
Using Helm CLI
--set-file
to pass your own file to particular config key. For example:helm upgrade -i test selenium-grid \ --set tls.enabled=true \ --set-file tls.certificate=/path/to/your_cert\.pem \ --set-file tls.privateKey=/path/to/your_private_key\.pkcs8 \ --set-file tls.trustStore=/path/to/your_truststore\.jks \ --set-string tls.trustStorePassword=your_truststore_password
If you start NGINX ingress controller inline with Selenium Grid chart, you can configure the default certificate of NGINX ingress controller to use the same certificate as Selenium Grid. For example:
tls:
enabled: true
ingress-nginx:
enabled: true
controller:
extraArgs:
default-ssl-certificate: '$(POD_NAMESPACE)/selenium-tls-secret'
To enable secure in the node registration to make sure that the node is one you control and not a rouge node, you can enable and provide a registration secret string to Distributor, Router and
Node servers in config tls.registrationSecret
. For example:
tls:
enabled: true
registrationSecret:
enabled: true
value: "matchThisSecret"
The chart supports tracing observability via Jaeger. To enable it, you need to set the following values:
tracing:
enabled: true
With this configuration, by default, Jaeger (all-in-one) will be deployed in the same namespace as Selenium Grid.
The Jaeger UI can be accessed via same ingress with prefix /jaeger
, for example: http://your.host.name/jaeger
.
The traces will be collected from all the components of Selenium Grid and can be viewed in the Jaeger UI.
In case you want to use your own existing Jaeger instance, you can set the following values:
tracing:
enabledWithExistingEndpoint: true
exporter: otlp
exporterEndpoint: 'http://jaeger.domain.com:4317'
This table contains the configuration parameters of the chart and their default values:
Parameter | Default | Description |
---|---|---|
basicAuth.enabled |
true |
Enable or disable basic auth for Selenium Grid |
basicAuth.username |
admin |
Username of basic auth for Selenium Grid |
basicAuth.password |
admin |
Password of basic auth for Selenium Grid |
isolateComponents |
false |
Deploy Router, Distributor, EventBus, SessionMap and Nodes separately |
serviceAccount.create |
true |
Enable or disable creation of service account (if false , serviceAccount.nameOverride MUST be specified |
serviceAccount.nameOverride |
"" |
Name of another service account to be made or existing service account to use for all deployments and jobs |
serviceAccount.annotations |
{} |
Custom annotations for service account |
busConfigMap.nameOverride |
`` | Override another configmap that contains SE_EVENT_BUS_HOST, SE_EVENT_BUS_PUBLISH_PORT and SE_EVENT_BUS_SUBSCRIBE_PORT vars |
busConfigMap.annotations |
{} |
Custom annotations for configmap |
nodeConfigMap.nameOverride |
`` | Name of the configmap that contains common environment variables for browser nodes |
nodeConfigMap.annotations |
{} |
Custom annotations for configmap |
ingress.enabled |
true |
Enable or disable ingress resource |
ingress.className |
"" |
Name of ingress class to select which controller will implement ingress resource |
ingress.annotations |
{} |
Custom annotations for ingress resource |
ingress.nginx.proxyTimeout |
3600 |
Value is used to set for NGINX ingress annotations related to proxy timeout |
ingress.nginx.proxyBuffer.size |
512M |
Value is used to set for NGINX ingress annotations on size of the buffer proxy_buffer_size used for reading |
ingress.nginx.proxyBuffer.number |
4 |
Value is used to set for NGINX ingress annotations on number of the buffers in proxy_buffers used for reading |
ingress.ports.http |
80 |
Port to expose for HTTP |
ingress.ports.https |
443 |
Port to expose for HTTPS |
ingress.hostname |
`` | Default host for the ingress resource |
ingress.path |
/ |
Default host path for the ingress resource |
ingress.pathType |
Prefix |
Default path type for the ingress resource |
ingress.paths |
[] |
List of paths config for the ingress resource. This will override the default path |
ingress.tls |
[] |
TLS backend configuration for ingress resource |
autoscaling.enableWithExistingKEDA |
false |
Enable autoscaling of browser nodes. |
autoscaling.enabled |
false |
Same as above plus installation of KEDA |
autoscaling.scalingType |
job |
Which typ of KEDA scaling to use: job or deployment |
autoscaling.scaledOptions |
See values.yaml |
Common options for KEDA scaled resources (both ScaledJobs and ScaledObjects) |
autoscaling.scaledOptions.minReplicaCount |
0 |
Min number of replicas that each browser nodes has when autoscaling |
autoscaling.scaledOptions.maxReplicaCount |
8 |
Max number of replicas that each browser nodes can auto scale up to |
autoscaling.scaledOptions.pollingInterval |
10 |
The interval to check each trigger on |
autoscaling.scaledJobOptions |
See values.yaml |
Options for KEDA ScaledJobs (when scalingType is set to job ) |
autoscaling.scaledObjectOptions |
See values.yaml |
Options for KEDA ScaledObjects (when scalingType is set to deployment ) |
autoscaling.deregisterLifecycle |
See values.yaml |
Lifecycle applied to pods of deployments controlled by KEDA. Makes the node deregister from selenium hub |
chromeNode.enabled |
true |
Enable chrome nodes |
chromeNode.deploymentEnabled |
true |
Enable creation of Deployment for chrome nodes |
chromeNode.replicas |
1 |
Number of chrome nodes. Disabled if autoscaling is enabled. |
chromeNode.imageRegistry |
nil |
Distribution registry to pull the image (this overwrites .global.seleniumGrid.imageRegistry value) |
chromeNode.imageName |
node-chrome |
Image of chrome nodes |
chromeNode.imageTag |
4.18.1-20240224 |
Image of chrome nodes |
chromeNode.imagePullPolicy |
IfNotPresent |
Image pull policy (see https://kubernetes.io/docs/concepts/containers/images/#updating-images) |
chromeNode.imagePullSecret |
"" |
Image pull secret (see https://kubernetes.io/docs/tasks/configure-pod-container/pull-image-private-registry) |
chromeNode.ports |
[] |
Extra ports list to enable on container (e.g VNC, NoVNC, SSH if any) |
chromeNode.port |
5555 |
Port is used to set SE_NODE_PORT |
chromeNode.nodePort |
nil |
NodePort where chrome-node exposed |
chromeNode.annotations |
{} |
Annotations for chrome-node pods |
chromeNode.labels |
{} |
Labels for chrome-node pods |
chromeNode.resources |
See values.yaml |
Resources for chrome-node pods |
chromeNode.securityContext |
See values.yaml |
Security context for chrome-node pods |
chromeNode.tolerations |
[] |
Tolerations for chrome-node pods |
chromeNode.nodeSelector |
{} |
Node Selector for chrome-node pods |
chromeNode.affinity |
{} |
Affinity for chrome-node pods |
chromeNode.hostAliases |
nil |
Custom host aliases for chrome nodes |
chromeNode.priorityClassName |
"" |
Priority class name for chrome-node pods |
chromeNode.extraEnvironmentVariables |
nil |
Custom environment variables for chrome nodes |
chromeNode.extraEnvFrom |
nil |
Custom environment taken from configMap or secret variables for chrome nodes |
chromeNode.service.enabled |
true |
Create a service for node |
chromeNode.service.type |
ClusterIP |
Service type |
chromeNode.service.loadBalancerIP |
`` | Set specific loadBalancerIP when serviceType is LoadBalancer |
chromeNode.service.ports |
[] |
Extra ports exposed in node service |
chromeNode.service.annotations |
{} |
Custom annotations for service |
chromeNode.dshmVolumeSizeLimit |
1Gi |
Size limit for DSH volume mounted in container (if not set, default is "1Gi") |
chromeNode.startupProbe.enabled |
true |
Enable Probe to check pod is started successfully (the following configs see values.yaml ) |
chromeNode.readinessProbe.enabled |
false |
Enable Readiness probe settings (the following configs see values.yaml ) |
chromeNode.livenessProbe.enabled |
false |
Enable Liveness probe settings (the following configs see values.yaml ) |
chromeNode.terminationGracePeriodSeconds |
30 |
Time to graceful terminate container (default: 30s) |
chromeNode.lifecycle |
{} |
hooks to make pod correctly shutdown or started |
chromeNode.extraVolumeMounts |
[] |
Extra mounts of declared ExtraVolumes into pod |
chromeNode.extraVolumes |
[] |
Extra Volumes declarations to be used in the pod (can be any supported volume type: ConfigMap, Secret, PVC, NFS, etc.) |
chromeNode.hpa.url |
{{ include "seleniumGrid.graphqlURL" . }} |
Graphql Url of the hub or the router |
chromeNode.hpa.browserName |
chrome |
BrowserName from the capability |
chromeNode.hpa.browserVersion |
`` | BrowserVersion from the capability |
chromeNode.sidecars |
[] |
Add a sidecars proxy in the same pod of the browser node |
chromeNode.initContainers |
[] |
Add initContainers in the same pod of the browser node |
chromeNode.scaledOptions |
See values.yaml |
Override the global autoscaling.scaledOptions with specific scaled options for chrome nodes |
chromeNode.scaledJobOptions |
See values.yaml |
Override the global autoscaling.scaledJobOptions with specific scaled options for chrome nodes |
chromeNode.scaledObjectOptions |
See values.yaml |
Override the global autoscaling.scaledObjectOptions with specific scaled options for chrome nodes |
firefoxNode.enabled |
true |
Enable firefox nodes |
firefoxNode.deploymentEnabled |
true |
Enable creation of Deployment for firefox nodes |
firefoxNode.replicas |
1 |
Number of firefox nodes. Disabled if autoscaling is enabled. |
firefoxNode.imageRegistry |
nil |
Distribution registry to pull the image (this overwrites .global.seleniumGrid.imageRegistry value) |
firefoxNode.imageName |
node-firefox |
Image of firefox nodes |
firefoxNode.imageTag |
4.18.1-20240224 |
Image of firefox nodes |
firefoxNode.imagePullPolicy |
IfNotPresent |
Image pull policy (see https://kubernetes.io/docs/concepts/containers/images/#updating-images) |
firefoxNode.imagePullSecret |
"" |
Image pull secret (see https://kubernetes.io/docs/tasks/configure-pod-container/pull-image-private-registry) |
firefoxNode.ports |
[] |
Extra ports list to enable on container (e.g VNC, NoVNC, SSH if any) |
firefoxNode.port |
5555 |
Port is used to set SE_NODE_PORT |
firefoxNode.nodePort |
nil |
NodePort where firefox-node exposed |
firefoxNode.annotations |
{} |
Annotations for firefox-node pods |
firefoxNode.labels |
{} |
Labels for firefox-node pods |
firefoxNode.resources |
See values.yaml |
Resources for firefox-node pods |
firefoxNode.securityContext |
See values.yaml |
Security context for firefox-node pods |
firefoxNode.tolerations |
[] |
Tolerations for firefox-node pods |
firefoxNode.nodeSelector |
{} |
Node Selector for firefox-node pods |
firefoxNode.affinity |
{} |
Affinity for firefox-node pods |
firefoxNode.hostAliases |
nil |
Custom host aliases for firefox nodes |
firefoxNode.priorityClassName |
"" |
Priority class name for firefox-node pods |
firefoxNode.extraEnvironmentVariables |
nil |
Custom environment variables for firefox nodes |
firefoxNode.extraEnvFrom |
nil |
Custom environment variables taken from configMap or secret for firefox nodes |
firefoxNode.service.enabled |
true |
Create a service for node |
firefoxNode.service.type |
ClusterIP |
Service type |
firefoxNode.service.loadBalancerIP |
`` | Set specific loadBalancerIP when serviceType is LoadBalancer |
firefoxNode.service.ports |
[] |
Extra ports exposed in node service |
firefoxNode.service.annotations |
{} |
Custom annotations for service |
firefoxNode.dshmVolumeSizeLimit |
1Gi |
Size limit for DSH volume mounted in container (if not set, default is "1Gi") |
firefoxNode.startupProbe.enabled |
true |
Enable Probe to check pod is started successfully (the following configs see values.yaml ) |
firefoxNode.readinessProbe.enabled |
false |
Enable Readiness probe settings (the following configs see values.yaml ) |
firefoxNode.livenessProbe.enabled |
false |
Enable Liveness probe settings (the following configs see values.yaml ) |
firefoxNode.terminationGracePeriodSeconds |
30 |
Time to graceful terminate container (default: 30s) |
firefoxNode.lifecycle |
{} |
hooks to make pod correctly shutdown or started |
firefoxNode.extraVolumeMounts |
[] |
Extra mounts of declared ExtraVolumes into pod |
firefoxNode.extraVolumes |
[] |
Extra Volumes declarations to be used in the pod (can be any supported volume type: ConfigMap, Secret, PVC, NFS, etc.) |
firefoxNode.hpa.url |
{{ include "seleniumGrid.graphqlURL" . }} |
Graphql Url of the hub or the router |
firefoxNode.hpa.browserName |
firefox |
BrowserName from the capability |
firefoxNode.hpa.browserVersion |
`` | BrowserVersion from the capability |
firefoxNode.sidecars |
[] |
Add a sidecars proxy in the same pod of the browser node |
firefoxNode.initContainers |
[] |
Add initContainers in the same pod of the browser node |
firefoxNode.scaledOptions |
See values.yaml |
Override the global autoscaling.scaledOptions with specific scaled options for firefox nodes |
firefoxNode.scaledJobOptions |
See values.yaml |
Override the global autoscaling.scaledJobOptions with specific scaled options for firefox nodes |
firefoxNode.scaledObjectOptions |
See values.yaml |
Override the global autoscaling.scaledObjectOptions with specific scaled options for firefox nodes |
edgeNode.enabled |
true |
Enable edge nodes |
edgeNode.deploymentEnabled |
true |
Enable creation of Deployment for edge nodes |
edgeNode.replicas |
1 |
Number of edge nodes. Disabled if autoscaling is enabled. |
edgeNode.imageRegistry |
nil |
Distribution registry to pull the image (this overwrites .global.seleniumGrid.imageRegistry value) |
edgeNode.imageName |
node-edge |
Image of edge nodes |
edgeNode.imageTag |
4.18.1-20240224 |
Image of edge nodes |
edgeNode.imagePullPolicy |
IfNotPresent |
Image pull policy (see https://kubernetes.io/docs/concepts/containers/images/#updating-images) |
edgeNode.imagePullSecret |
"" |
Image pull secret (see https://kubernetes.io/docs/tasks/configure-pod-container/pull-image-private-registry) |
edgeNode.ports |
[] |
Extra ports list to enable on container (e.g VNC, NoVNC, SSH if any) |
edgeNode.port |
5555 |
Port is used to set SE_NODE_PORT |
edgeNode.nodePort |
nil |
NodePort where edge-node exposed |
edgeNode.annotations |
{} |
Annotations for edge-node pods |
edgeNode.labels |
{} |
Labels for edge-node pods |
edgeNode.resources |
See values.yaml |
Resources for edge-node pods |
edgeNode.securityContext |
See values.yaml |
Security context for edge-node pods |
edgeNode.tolerations |
[] |
Tolerations for edge-node pods |
edgeNode.nodeSelector |
{} |
Node Selector for edge-node pods |
edgeNode.affinity |
{} |
Affinity for edge-node pods |
edgeNode.hostAliases |
nil |
Custom host aliases for edge nodes |
edgeNode.priorityClassName |
"" |
Priority class name for edge-node pods |
edgeNode.extraEnvironmentVariables |
nil |
Custom environment variables for firefox nodes |
edgeNode.extraEnvFrom |
nil |
Custom environment taken from configMap or secret variables for firefox nodes |
edgeNode.service.enabled |
true |
Create a service for node |
edgeNode.service.type |
ClusterIP |
Service type |
edgeNode.service.loadBalancerIP |
`` | Set specific loadBalancerIP when serviceType is LoadBalancer |
edgeNode.service.ports |
[] |
Extra ports exposed in node service |
edgeNode.service.annotations |
{} |
Custom annotations for service |
edgeNode.dshmVolumeSizeLimit |
1Gi |
Size limit for DSH volume mounted in container (if not set, default is "1Gi") |
edgeNode.startupProbe.enabled |
true |
Enable Probe to check pod is started successfully (the following configs see values.yaml ) |
edgeNode.readinessProbe.enabled |
false |
Enable Readiness probe settings (the following configs see values.yaml ) |
edgeNode.livenessProbe.enabled |
false |
Enable Liveness probe settings (the following configs see values.yaml ) |
edgeNode.terminationGracePeriodSeconds |
30 |
Time to graceful terminate container (default: 30s) |
edgeNode.lifecycle |
{} |
hooks to make pod correctly shutdown or started |
edgeNode.extraVolumeMounts |
[] |
Extra mounts of declared ExtraVolumes into pod |
edgeNode.extraVolumes |
[] |
Extra Volumes declarations to be used in the pod (can be any supported volume type: ConfigMap, Secret, PVC, NFS, etc.) |
edgeNode.hpa.url |
{{ include "seleniumGrid.graphqlURL" . }} |
Graphql Url of the hub or the router |
edgeNode.hpa.browserName |
edge |
BrowserName from the capability |
edgeNode.hpa.browserVersion |
`` | BrowserVersion from the capability |
edgeNode.sidecars |
[] |
Add a sidecars proxy in the same pod of the browser node |
edgeNode.initContainers |
[] |
Add initContainers in the same pod of the browser node |
edgeNode.scaledOptions |
See values.yaml |
Override the global autoscaling.scaledOptions with specific scaled options for edge nodes |
edgeNode.scaledJobOptions |
See values.yaml |
Override the global autoscaling.scaledJobOptions with specific scaled options for edge nodes |
edgeNode.scaledObjectOptions |
See values.yaml |
Override the global autoscaling.scaledObjectOptions with specific scaled options for edge nodes |
videoRecorder.enabled |
false |
Enable video recorder for node |
videoRecorder.imageRegistry |
nil |
Distribution registry to pull the image (this overwrites .global.seleniumGrid.imageRegistry value) |
videoRecorder.imageName |
video |
Selenium video recorder image name |
videoRecorder.imageTag |
ffmpeg-6.1-20240224 |
Image tag of video recorder |
videoRecorder.imagePullPolicy |
IfNotPresent |
Image pull policy (see https://kubernetes.io/docs/concepts/containers/images/#updating-images) |
videoRecorder.uploader.enabled |
false |
Enable the uploader for videos |
videoRecorder.uploader.destinationPrefix |
`` | Destination for uploading video file. It is following rclone config |
videoRecorder.uploader.name |
`` | Name of the pluggable uploader container to add and configure |
videoRecorder.uploader.configFileName |
config.conf |
Config file name for rclone in uploader container |
videoRecorder.uploader.entryPointFileName |
entry_point.sh |
Script file name for uploader container entry point |
videoRecorder.uploader.config |
`` | Set value to uploader config file via YAML or --set-file |
videoRecorder.uploader.entryPoint |
`` | Set value to uploader entry point via YAML or --set-file |
videoRecorder.uploader.secrets |
`` | Environment variables to configure the uploader which store in Secret |
videoRecorder.ports |
[9000] |
Port list to enable on video recorder container |
videoRecorder.resources |
See values.yaml |
Resources for video recorder |
videoRecorder.extraEnvironmentVariables |
nil |
Custom environment variables for video recorder |
videoRecorder.extraEnvFrom |
nil |
Custom environment taken from configMap or secret variables for video recorder |
videoRecorder.terminationGracePeriodSeconds |
30 |
Time to graceful terminate container (default: 30s) |
videoRecorder.startupProbe |
{} |
Probe to check pod is started successfully |
videoRecorder.livenessProbe |
{} |
Liveness probe settings |
videoRecorder.lifecycle |
{} |
Define lifecycle events for video recorder |
videoRecorder.extraVolumeMounts |
[] |
Extra mounts of declared ExtraVolumes into pod |
videoRecorder.extraVolumes |
[] |
Extra Volumes declarations to be used in the pod (can be any supported volume type: ConfigMap, Secret, PVC, NFS, etc.) |
customLabels |
{} |
Custom labels for k8s resources |
ingress-nginx.enabled |
false |
Enable the dependency chart Ingress controller for Kubernetes (https://github.com/kubernetes/ingress-nginx) |
If you are setting autoscaling.enabled
to true
, chart KEDA is installed and can be configured with
values with the prefix keda
. So you can for example set keda.prometheus.metricServer.enabled
to
true
to enable the metrics server for KEDA. See
https://github.com/kedacore/charts/blob/main/keda/README.md for more details.
If you are setting ingress-nginx.enabled
to true
, chart Ingress NGINX Controller is installed and can be configured with
values with the prefix ingress-nginx
. See https://github.com/kubernetes/ingress-nginx for more details.
If you are setting tracing.enabled
to true
, chart Jaeger is installed and can be configured with
values with the prefix jaeger
. See https://github.com/jaegertracing/helm-charts for more details.
You can configure the Selenium Hub with these values:
Parameter | Default | Description |
---|---|---|
hub.imageRegistry |
nil |
Distribution registry to pull the image (this overwrites .global.seleniumGrid.imageRegistry value) |
hub.imageName |
hub |
Selenium Hub image name |
hub.imageTag |
nil |
Selenium Hub image tag (this overwrites .global.seleniumGrid.imageTag value) |
hub.imagePullPolicy |
IfNotPresent |
Image pull policy (see https://kubernetes.io/docs/concepts/containers/images/#updating-images) |
hub.imagePullSecret |
"" |
Image pull secret (see https://kubernetes.io/docs/tasks/configure-pod-container/pull-image-private-registry) |
hub.annotations |
{} |
Custom annotations for Selenium Hub pod |
hub.labels |
{} |
Custom labels for Selenium Hub pod |
hub.disableUI |
false |
Disable the Grid UI |
hub.publishPort |
4442 |
Port where events are published |
hub.publishNodePort |
31442 |
NodePort where events are published |
hub.subscribePort |
4443 |
Port where to subscribe for events |
hub.subscribeNodePort |
31443 |
NodePort where to subscribe for events |
hub.port |
4444 |
Selenium Hub port |
hub.nodePort |
31444 |
Selenium Hub NodePort |
hub.livenessProbe |
See values.yaml |
Liveness probe settings |
hub.readinessProbe |
See values.yaml |
Readiness probe settings |
hub.tolerations |
[] |
Tolerations for selenium-hub pods |
hub.nodeSelector |
{} |
Node Selector for selenium-hub pods |
hub.affinity |
{} |
Affinity for selenium-hub pods |
hub.priorityClassName |
"" |
Priority class name for selenium-hub pods |
hub.subPath |
/ |
Custom sub path for the hub deployment |
hub.extraEnvironmentVariables |
nil |
Custom environment variables for selenium-hub |
hub.extraEnvFrom |
nil |
Custom environment variables for selenium taken from configMap or secret -hub |
hub.extraVolumeMounts |
[] |
Extra mounts of declared ExtraVolumes into pod |
hub.extraVolumes |
[] |
Extra Volumes declarations to be used in the pod (can be any supported volume type: ConfigMap, Secret, PVC, NFS, etc.) |
hub.resources |
{} |
Resources for selenium-hub container |
hub.securityContext |
See values.yaml |
Security context for selenium-hub container |
hub.serviceType |
ClusterIP |
Kubernetes service type (see https://kubernetes.io/docs/concepts/services-networking/service/#publishing-services-service-types) |
hub.loadBalancerIP |
nil |
Set specific loadBalancerIP when serviceType is LoadBalancer (see https://kubernetes.io/docs/concepts/services-networking/service/#loadbalancer) |
hub.serviceAnnotations |
{} |
Custom annotations for Selenium Hub service |
If you implement selenium-grid with separate components (isolateComponents: true
), you can configure all components via the following values:
Parameter | Default | Description |
---|---|---|
components.router.imageRegistry |
nil |
Distribution registry to pull the image (this overwrites .global.seleniumGrid.imageRegistry value) |
components.router.imageName |
router |
Router image name |
components.router.imageTag |
nil |
Router image tag (this overwrites .global.seleniumGrid.imageTag value) |
components.router.imagePullPolicy |
IfNotPresent |
Image pull policy (see https://kubernetes.io/docs/concepts/containers/images/#updating-images) |
components.router.imagePullSecret |
"" |
Image pull secret (see https://kubernetes.io/docs/tasks/configure-pod-container/pull-image-private-registry) |
components.router.annotations |
{} |
Custom annotations for router pod |
components.router.port |
4444 |
Router port |
components.router.nodePort |
30444 |
Router NodePort |
components.router.livenessProbe |
See values.yaml |
Liveness probe settings |
components.router.readinessProbe |
See values.yaml |
Readiness probe settings |
components.router.resources |
{} |
Resources for router pod |
components.router.securityContext |
See values.yaml |
Security context for router pod |
components.router.serviceType |
ClusterIP |
Kubernetes service type (see https://kubernetes.io/docs/concepts/services-networking/service/#publishing-services-service-types) |
components.router.loadBalancerIP |
nil |
Set specific loadBalancerIP when serviceType is LoadBalancer (see https://kubernetes.io/docs/concepts/services-networking/service/#loadbalancer) |
components.router.serviceAnnotations |
{} |
Custom annotations for router service |
components.router.tolerations |
[] |
Tolerations for router pods |
components.router.nodeSelector |
{} |
Node Selector for router pods |
components.router.affinity |
{} |
Affinity for router pods |
components.router.priorityClassName |
"" |
Priority class name for router pods |
components.distributor.imageRegistry |
nil |
Distribution registry to pull the image (this overwrites .global.seleniumGrid.imageRegistry value) |
components.distributor.imageName |
distributor |
Distributor image name |
components.distributor.imageTag |
nil |
Distributor image tag (this overwrites .global.seleniumGrid.imageTag value) |
components.distributor.imagePullPolicy |
IfNotPresent |
Image pull policy (see https://kubernetes.io/docs/concepts/containers/images/#updating-images) |
components.distributor.imagePullSecret |
"" |
Image pull secret (see https://kubernetes.io/docs/tasks/configure-pod-container/pull-image-private-registry) |
components.distributor.annotations |
{} |
Custom annotations for Distributor pod |
components.distributor.port |
5553 |
Distributor port |
components.distributor.nodePort |
30553 |
Distributor NodePort |
components.distributor.resources |
{} |
Resources for Distributor pod |
components.distributor.securityContext |
See values.yaml |
Security context for Distributor pod |
components.distributor.serviceType |
ClusterIP |
Kubernetes service type (see https://kubernetes.io/docs/concepts/services-networking/service/#publishing-services-service-types) |
components.distributor.serviceAnnotations |
{} |
Custom annotations for Distributor service |
components.distributor.tolerations |
[] |
Tolerations for Distributor pods |
components.distributor.nodeSelector |
{} |
Node Selector for Distributor pods |
components.distributor.affinity |
{} |
Affinity for Distributor pods |
components.distributor.priorityClassName |
"" |
Priority class name for Distributor pods |
components.eventBus.imageRegistry |
nil |
Distribution registry to pull the image (this overwrites .global.seleniumGrid.imageRegistry value) |
components.eventBus.imageName |
event-bus |
Event Bus image name |
components.eventBus.imageTag |
nil |
Event Bus image tag (this overwrites .global.seleniumGrid.imageTag value) |
components.eventBus.imagePullPolicy |
IfNotPresent |
Image pull policy (see https://kubernetes.io/docs/concepts/containers/images/#updating-images) |
components.eventBus.imagePullSecret |
"" |
Image pull secret (see https://kubernetes.io/docs/tasks/configure-pod-container/pull-image-private-registry) |
components.eventBus.annotations |
{} |
Custom annotations for Event Bus pod |
components.eventBus.port |
5557 |
Event Bus port |
components.eventBus.nodePort |
30557 |
Event Bus NodePort |
components.eventBus.publishPort |
4442 |
Port where events are published |
components.eventBus.publishNodePort |
30442 |
NodePort where events are published |
components.eventBus.subscribePort |
4443 |
Port where to subscribe for events |
components.eventBus.subscribeNodePort |
30443 |
NodePort where to subscribe for events |
components.eventBus.resources |
{} |
Resources for event-bus pod |
components.eventBus.securityContext |
See values.yaml |
Security context for event-bus pod |
components.eventBus.serviceType |
ClusterIP |
Kubernetes service type (see https://kubernetes.io/docs/concepts/services-networking/service/#publishing-services-service-types) |
components.eventBus.serviceAnnotations |
{} |
Custom annotations for Event Bus service |
components.eventBus.tolerations |
[] |
Tolerations for Event Bus pods |
components.eventBus.nodeSelector |
{} |
Node Selector for Event Bus pods |
components.eventBus.affinity |
{} |
Affinity for Event Bus pods |
components.eventBus.priorityClassName |
"" |
Priority class name for Event Bus pods |
components.sessionMap.imageRegistry |
nil |
Distribution registry to pull the image (this overwrites .global.seleniumGrid.imageRegistry value) |
components.sessionMap.imageName |
sessions |
Session Map image name |
components.sessionMap.imageTag |
nil |
Session Map image tag (this overwrites .global.seleniumGrid.imageTag value) |
components.sessionMap.imagePullPolicy |
IfNotPresent |
Image pull policy (see https://kubernetes.io/docs/concepts/containers/images/#updating-images) |
components.sessionMap.imagePullSecret |
"" |
Image pull secret (see https://kubernetes.io/docs/tasks/configure-pod-container/pull-image-private-registry) |
components.sessionMap.annotations |
{} |
Custom annotations for Session Map pod |
components.sessionMap.resources |
{} |
Resources for Session Map pod |
components.sessionMap.securityContext |
See values.yaml |
Security context for Session Map pod |
components.sessionMap.serviceType |
ClusterIP |
Kubernetes service type (see https://kubernetes.io/docs/concepts/services-networking/service/#publishing-services-service-types) |
components.sessionMap.serviceAnnotations |
{} |
Custom annotations for Session Map service |
components.sessionMap.tolerations |
[] |
Tolerations for Session Map pods |
components.sessionMap.nodeSelector |
{} |
Node Selector for Session Map pods |
components.sessionMap.affinity |
{} |
Affinity for Session Map pods |
components.sessionMap.priorityClassName |
"" |
Priority class name for Session Map pods |
components.sessionQueue.imageRegistry |
nil |
Distribution registry to pull the image (this overwrites .global.seleniumGrid.imageRegistry value) |
components.sessionQueue.imageName |
session-queue |
Session Queue image name |
components.sessionQueue.imageTag |
nil |
Session Queue image tag (this overwrites .global.seleniumGrid.imageTag value) |
components.sessionQueue.imagePullPolicy |
IfNotPresent |
Image pull policy (see https://kubernetes.io/docs/concepts/containers/images/#updating-images) |
components.sessionQueue.imagePullSecret |
"" |
Image pull secret (see https://kubernetes.io/docs/tasks/configure-pod-container/pull-image-private-registry) |
components.sessionQueue.annotations |
{} |
Custom annotations for Session Queue pod |
components.sessionQueue.port |
5559 |
Session Queue Port |
components.sessionQueue.nodePort |
30559 |
Session Queue NodePort |
components.sessionQueue.resources |
{} |
Resources for Session Queue pod |
components.sessionQueue.securityContext |
See values.yaml |
Security context for Session Queue pod |
components.sessionQueue.serviceType |
ClusterIP |
Kubernetes service type (see https://kubernetes.io/docs/concepts/services-networking/service/#publishing-services-service-types) |
components.sessionQueue.serviceAnnotations |
{} |
Custom annotations for Session Queue service |
components.sessionQueue.tolerations |
[] |
Tolerations for Session Queue pods |
components.sessionQueue.nodeSelector |
{} |
Node Selector for Session Queue pods |
components.sessionQueue.affinity |
{} |
Affinity for Session Queue pods |
components.sessionQueue.priorityClassName |
"" |
Priority class name for Session Queue pods |
components.subPath |
/ |
Custom sub path for all components |
components.disableUI |
false |
Disable the Grid UI |
components.extraEnvironmentVariables |
nil |
Custom environment variables for all components |
components.extraEnvFrom |
nil |
Custom environment variables taken from configMap or secret for all components |
See how to customize a helm chart installation in the Helm Docs for more information.