Node.js is a runtime environment built on V8 JavaScript engine. Its event-driven, non-blocking I/O model enables the development of fast, scalable, and data-intensive server applications.
Trademarks: This software listing is packaged by Bitnami. The respective trademarks mentioned in the offering are owned by the respective companies, and use of them does not imply any affiliation or endorsement.
$ helm repo add my-repo https://charts.bitnami.com/bitnami
$ helm install my-release my-repo/nodeThis chart bootstraps a Node deployment on a Kubernetes cluster using the Helm package manager.
It clones and deploys a Node.js application from a Git repository. Optionally, you can set up an Ingress resource to access your application and provision an external database using the Kubernetes service catalog and the Open Service Broker for Azure.
Bitnami charts can be used with Kubeapps for deployment and management of Helm Charts in clusters.
- Kubernetes 1.19+
- Helm 3.2.0+
- PV provisioner support in the underlying infrastructure
- ReadWriteMany volumes for deployment scaling
To install the chart with the release name my-release:
$ helm repo add my-repo https://charts.bitnami.com/bitnami
$ helm install my-release my-repo/nodeThese commands deploy Node.js on the Kubernetes cluster in the default configuration. The Parameters section lists the parameters that can be configured during installation. Also includes support for MariaDB chart out of the box.
Due that the Helm Chart clones the application on the /app volume while the container is initializing, a persistent volume is not required.
Tip: List all releases using
helm list
To uninstall/delete the my-release deployment:
$ helm delete my-releaseThe command removes all the Kubernetes components associated with the chart and deletes the release.
| Name | Description | Value |
|---|---|---|
global.imageRegistry |
Global Docker image registry | "" |
global.imagePullSecrets |
Global Docker registry secret names as an array | [] |
global.storageClass |
Global StorageClass for Persistent Volume(s) | "" |
| Name | Description | Value |
|---|---|---|
kubeVersion |
Force target Kubernetes version (using Helm capabilities if not set) | "" |
nameOverride |
String to partially override node.fullname template (will maintain the release name) | "" |
fullnameOverride |
String to fully override node.fullname template | "" |
namespaceOverride |
Override namespace for resources | "" |
commonLabels |
Add labels to all the deployed resources | {} |
commonAnnotations |
Add annotations to all the deployed resources | {} |
| Name | Description | Value |
|---|---|---|
installCommand |
Override default container install command (useful when using custom images or repositories) | ["/bin/bash","-ec","npm install"] |
command |
Override default container command (useful when using custom images) | ["/bin/bash","-ec","npm start"] |
args |
Override default container args (useful when using custom images) | [] |
hostAliases |
Deployment pod host aliases | [] |
extraEnvVars |
Extra environment variables to be set on Node container | [] |
extraEnvVarsCM |
Name of existing ConfigMap containing extra environment variables | "" |
extraEnvVarsSecret |
Name of existing Secret containing extra environment variables | "" |
mongodb.enabled |
Whether to install or not the MongoDB® chart | true |
mongodb.auth.enabled |
Whether to enable auth or not for the MongoDB® chart | true |
mongodb.auth.rootPassword |
MongoDB® admin password | "" |
mongodb.auth.username |
MongoDB® custom user | user |
mongodb.auth.database |
MongoDB® custom database | test_db |
mongodb.auth.password |
MongoDB® custom password | secret_password |
externaldb.enabled |
Enables or disables external database (ignored if mongodb.enabled=true) |
false |
externaldb.ssl |
Set to true if your external database has ssl enabled | false |
externaldb.secretName |
Secret containing existing database credentials | "" |
externaldb.type |
Only if using Kubernetes Service Catalog you can specify the kind of broker used. Available options are osba | gce |
externaldb.broker.serviceInstanceName |
If you provide the serviceInstanceName, the chart will create a ServiceBinding for that ServiceInstance | "" |
| Name | Description | Value |
|---|---|---|
image.registry |
NodeJS image registry | docker.io |
image.repository |
NodeJS image repository | bitnami/node |
image.tag |
NodeJS image tag (immutable tags are recommended) | 16.17.1-debian-11-r0 |
image.digest |
NodeJS image digest in the way sha256:aa.... Please note this parameter, if set, will override the tag | "" |
image.pullPolicy |
NodeJS image pull policy | IfNotPresent |
image.pullSecrets |
Specify docker-registry secret names as an array | [] |
image.debug |
Set to true if you would like to see extra information on logs | false |
replicaCount |
Specify the number of replicas for the application | 1 |
updateStrategy.type |
Strategy to use to replace existing pods. | RollingUpdate |
containerPorts.http |
Specify the port where your application will be running | 3000 |
podAffinityPreset |
Pod affinity preset. Ignored if affinity is set. Allowed values: soft or hard |
"" |
podAntiAffinityPreset |
Pod anti-affinity preset. Ignored if affinity is set. Allowed values: soft or hard |
soft |
nodeAffinityPreset.type |
Node affinity preset type. Ignored if affinity is set. Allowed values: soft or hard |
"" |
nodeAffinityPreset.key |
Node label key to match Ignored if affinity is set. |
"" |
nodeAffinityPreset.values |
Node label values to match. Ignored if affinity is set. |
[] |
affinity |
Affinity for pod assignment. Evaluated as a template. | {} |
nodeSelector |
Node labels for pod assignment. Evaluated as a template. | {} |
tolerations |
Tolerations for pod assignment. Evaluated as a template. | [] |
podAnnotations |
Additional pod annotations | {} |
podLabels |
Additional labels for Node pods | {} |
extraDeploy |
Array of extra objects to deploy with the release (evaluated as a template) | [] |
diagnosticMode.enabled |
Enable diagnostic mode (all probes will be disabled and the command will be overridden) | false |
diagnosticMode.command |
Command to override all containers in the the deployment(s)/statefulset(s) | ["sleep"] |
diagnosticMode.args |
Args to override all containers in the the deployment(s)/statefulset(s) | ["infinity"] |
livenessProbe.enabled |
Enable livenessProbe | true |
livenessProbe.path |
Request path for livenessProbe | / |
livenessProbe.initialDelaySeconds |
Initial delay seconds for livenessProbe | 60 |
livenessProbe.periodSeconds |
Period seconds for livenessProbe | 10 |
livenessProbe.timeoutSeconds |
Timeout seconds for livenessProbe | 5 |
livenessProbe.failureThreshold |
Failure threshold for livenessProbe | 6 |
livenessProbe.successThreshold |
Success threshold for livenessProbe | 1 |
readinessProbe.enabled |
Enable readinessProbe | true |
readinessProbe.path |
Request path for readinessProbe | / |
readinessProbe.initialDelaySeconds |
Initial delay seconds for readinessProbe | 10 |
readinessProbe.periodSeconds |
Period seconds for readinessProbe | 5 |
readinessProbe.timeoutSeconds |
Timeout seconds for readinessProbe | 3 |
readinessProbe.failureThreshold |
Failure threshold for readinessProbe | 3 |
readinessProbe.successThreshold |
Success threshold for readinessProbe | 1 |
startupProbe.enabled |
Enable startupProbe | false |
startupProbe.path |
Request path for startupProbe | / |
startupProbe.initialDelaySeconds |
Initial delay seconds for startupProbe | 5 |
startupProbe.periodSeconds |
Period seconds for startupProbe | 3 |
startupProbe.timeoutSeconds |
Timeout seconds for startupProbe | 1 |
startupProbe.failureThreshold |
Failure threshold for startupProbe | 15 |
startupProbe.successThreshold |
Success threshold for startupProbe | 1 |
customLivenessProbe |
Override default liveness probe | {} |
customReadinessProbe |
Override default readiness probe | {} |
customStartupProbe |
Override default startup probe | {} |
topologySpreadConstraints |
Topology Spread Constraints for pod assignment spread across your cluster among failure-domains. Evaluated as a template | [] |
priorityClassName |
Node priorityClassName | "" |
schedulerName |
Use an alternate scheduler, e.g. "stork". | "" |
terminationGracePeriodSeconds |
Seconds Airflow web pod needs to terminate gracefully | "" |
lifecycleHooks |
lifecycleHooks for the Node container to automate configuration before or after startup. | {} |
sidecars |
Add sidecars to the Node pods | [] |
initContainers |
Add init containers to the Node pods | [] |
extraVolumes |
Extra volumes to add to the deployment | [] |
extraVolumeMounts |
Extra volume mounts to add to the container | [] |
serviceAccount.create |
Enable creation of ServiceAccount for node pod | false |
serviceAccount.name |
The name of the ServiceAccount to use. | "" |
serviceAccount.annotations |
Annotations for service account. Evaluated as a template. | {} |
serviceAccount.automountServiceAccountToken |
Whether to auto mount the service account token | false |
containerSecurityContext.enabled |
Node Container securityContext | true |
containerSecurityContext.runAsUser |
User ID for the Node container | 1001 |
containerSecurityContext.runAsNonRoot |
Set container's Security Context runAsNonRoot | true |
podSecurityContext.enabled |
Enable security context for Node pods | true |
podSecurityContext.fsGroup |
Group ID for the volumes of the pod | 1001 |
resources.limits |
The resources limits for the Node container | {} |
resources.requests |
The requested resources for the Node container | {} |
| Name | Description | Value |
|---|---|---|
git.image.registry |
Git image registry | docker.io |
git.image.repository |
Git image repository | bitnami/git |
git.image.tag |
Git image tag (immutable tags are recommended) | 2.37.3-debian-11-r8 |
git.image.digest |
Git image digest in the way sha256:aa.... Please note this parameter, if set, will override the tag | "" |
git.image.pullPolicy |
Git image pull policy | IfNotPresent |
git.image.pullSecrets |
Specify docker-registry secret names as an array | [] |
git.image.debug |
Set to true if you would like to see extra information on logs | false |
git.extraVolumeMounts |
Add extra volume mounts for the Git container | [] |
getAppFromExternalRepository |
Enable to download app from external git repository | true |
repository |
Git repository http/https url | https://github.com/bitnami/sample-mean.git |
revision |
Git repository revision to checkout | master |
| Name | Description | Value |
|---|---|---|
volumePermissions.enabled |
Enable init container that changes volume permissions in the data directory | false |
volumePermissions.image.registry |
Init container volume-permissions image registry | docker.io |
volumePermissions.image.repository |
Init container volume-permissions image repository | bitnami/bitnami-shell |
volumePermissions.image.tag |
Init container volume-permissions image tag (immutable tags are recommended) | 11-debian-11-r37 |
volumePermissions.image.digest |
Init container volume-permissions image digest in the way sha256:aa.... Please note this parameter, if set, will override the tag | "" |
volumePermissions.image.pullPolicy |
Init container volume-permissions image pull policy | IfNotPresent |
volumePermissions.image.pullSecrets |
Specify docker-registry secret names as an array | [] |
volumePermissions.resources.limits |
The resources limits for the container | {} |
volumePermissions.resources.requests |
The requested resources for the container | {} |
| Name | Description | Value |
|---|---|---|
persistence.enabled |
Enable persistence using PVC | false |
persistence.mountPath |
Path where the volume will be mount at. | /app/data |
persistence.subPath |
Subdirectory of the volume to mount | "" |
persistence.existingClaim |
Name of an existing PVC to use | "" |
persistence.resourcePolicy |
Setting it to "keep" to avoid removing PVCs during a helm delete operation. Leaving it empty will delete PVCs after the chart deleted | "" |
persistence.storageClass |
Persistent Volume Storage Class | "" |
persistence.accessModes |
PVC Access Modes | ["ReadWriteOnce"] |
persistence.size |
PVC Storage Request | 1Gi |
persistence.annotations |
Annotations for the PVC | {} |
| Name | Description | Value |
|---|---|---|
service.type |
Kubernetes Service type | ClusterIP |
service.ports.http |
Kubernetes Service port | 80 |
service.clusterIP |
Service Cluster IP | "" |
service.sessionAffinity |
Control where client requests go, to the same pod or round-robin | None |
service.sessionAffinityConfig |
Additional settings for the sessionAffinity | {} |
service.nodePorts.http |
NodePort if Service type is LoadBalancer or NodePort |
"" |
service.extraPorts |
Extra ports to expose (normally used with the sidecar value) |
[] |
service.loadBalancerIP |
LoadBalancer IP if Service type is LoadBalancer |
"" |
service.loadBalancerSourceRanges |
In order to limit which client IP's can access the Network Load Balancer, specify loadBalancerSourceRanges | [] |
service.externalTrafficPolicy |
Enable client source IP preservation | Cluster |
service.annotations |
Annotations for the Service | {} |
ingress.enabled |
Set to true to enable ingress record generation | false |
ingress.ingressClassName |
IngressClass that will be be used to implement the Ingress (Kubernetes 1.18+) | "" |
ingress.pathType |
Ingress path type | ImplementationSpecific |
ingress.apiVersion |
Override API Version (automatically detected if not set) | "" |
ingress.hostname |
When the ingress is enabled, a host pointing to this will be created | node.local |
ingress.path |
The Path to Node.js. You may need to set this to '/*' in order to use this with ALB ingress controllers. | / |
ingress.annotations |
Additional annotations for the Ingress resource. To enable certificate autogeneration, place here your cert-manager annotations. | {} |
ingress.tls |
Enable TLS configuration for the hostname defined at ingress.hostname parameter | false |
ingress.selfSigned |
Create a TLS secret for this ingress record using self-signed certificates generated by Helm | false |
ingress.extraHosts |
The list of additional hostnames to be covered with this ingress record. | [] |
ingress.extraPaths |
Any additional arbitrary paths that may need to be added to the ingress under the main host. | [] |
ingress.extraTls |
The tls configuration for additional hostnames to be covered with this ingress record. | [] |
ingress.secrets |
If you're providing your own certificates, please use this to add the certificates as secrets | [] |
ingress.extraRules |
Additional rules to be covered with this ingress record | [] |
The above parameters map to the env variables defined in bitnami/node. For more information please refer to the bitnami/node image documentation.
Specify each parameter using the --set key=value[,key=value] argument to helm install. For example,
$ helm install my-release \
--set repository=https://github.com/jbianquetti-nami/simple-node-app.git,replicaCount=2 \
my-repo/nodeThe above command clones the remote git repository to the /app/ directory of the container. Additionally it sets the number of replicaCount to 2.
Alternatively, a YAML file that specifies the values for the above parameters can be provided while installing the chart. For example,
$ helm install my-release -f values.yaml my-repo/nodeTip: You can use the default values.yaml
It is strongly recommended to use immutable tags in a production environment. This ensures your deployment does not change automatically if the same tag is updated with a different image.
Bitnami will release a new chart updating its containers if a new version of the main container, significant changes, or critical vulnerabilities exist.
To modify the application version used in this chart, specify a different version of the image using the image.tag parameter and/or a different repository using the image.repository parameter. Refer to the chart documentation for more information on these parameters and how to use them with images from a private registry.
First install the nginx-ingress controller and then deploy the node helm chart with the following parameters:
ingress.enabled=true
ingress.host=example.com
service.type=ClusterIPYou must manually create a secret containing the certificate and key for your domain. Then ensure you deploy the Helm chart with the following ingress configuration:
ingress:
enabled: false
path: /
host: example.com
annotations:
kubernetes.io/ingress.class: nginx
tls:
hosts:
- example.com- Create a secret containing your database credentials (named
my-database-secretas example), you can use the following options (set with--from-literal) to create the secret:
host=YOUR_DATABASE_HOST
port=YOUR_DATABASE_PORT
username=YOUR_DATABASE_USER
password=YOUR_DATABASE_PASSWORD
database=YOUR_DATABASE_NAMEYOUR_DATABASE_HOST, YOUR_DATABASE_PORT, YOUR_DATABASE_USER, YOUR_DATABASE_PASSWORD, and YOUR_DATABASE_NAME are placeholders that must be replaced with correct values.
- Deploy the node chart specifying the secret name
mongodb.enabled=false
externaldb.enabled=true
externaldb.secretName=my-database-secret- Install Service Catalog in your Kubernetes cluster following this instructions
- Install the Open Service Broker for Azure in your Kubernetes cluster following this instructions
TIP: you may want to install the osba chart setting the
modules.minStability=EXPERIMENTALto see all the available services.azure.subscriptionId=$AZURE_SUBSCRIPTION_ID azure.tenantId=$AZURE_TENANT_ID azure.clientId=$AZURE_CLIENT_ID azure.clientSecret=$AZURE_CLIENT_SECRET modules.minStability=EXPERIMENTAL
- Create and deploy a ServiceInstance to provision a database server in Azure cloud.
apiVersion: servicecatalog.k8s.io/v1beta1
kind: ServiceInstance
metadata:
name: azure-mongodb-instance
labels:
app: mongodb
spec:
clusterServiceClassExternalName: azure-cosmosdb-mongo-account
clusterServicePlanExternalName: account
parameters:
location: YOUR_AZURE_LOCATION
resourceGroup: mongodb-k8s-service-catalog
ipFilters:
allowedIPRanges:
- "0.0.0.0/0"Please update the YOUR_AZURE_LOCATION placeholder in the above example.
-
Deploy the helm chart:
mongodb.enabled=false externaldb.enabled=true externaldb.broker.serviceInstanceName=azure-mongodb-instance externaldb.ssl=true
Once the instance has been provisioned in Azure, a new secret should have been automatically created with the connection parameters for your application.
Deploying the helm chart enabling the Azure external database makes the following assumptions:
- You would want an Azure CosmosDB MongoDB® database
- Your application uses DATABASE_HOST, DATABASE_PORT, DATABASE_USER, DATABASE_PASSWORD, and DATABASE_NAME environment variables to connect to the database.
The Bitnami Node image stores the Node application and configurations at the /app path of the container.
Persistent Volume Claims are used to keep the data across deployments. This is known to work in GCE, AWS, and minikube. See the Parameters section to configure the PVC or to disable persistence.
As the image run as non-root by default, it is necessary to adjust the ownership of the persistent volume so that the container can write data into it.
By default, the chart is configured to use Kubernetes Security Context to automatically change the ownership of the volume. However, this feature does not work in all Kubernetes distributions. As an alternative, this chart supports using an initContainer to change the ownership of the volume before mounting it in the final destination.
You can enable this initContainer by setting volumePermissions.enabled to true.
Find more information about how to deal with common errors related to Bitnami's Helm charts in this troubleshooting guide.
This major updates the MongoDB® subchart to its newest major, 13.0.0. No major issues are expected during the upgrade.
This major release renames several values in this chart and adds missing features, in order to be inline with the rest of assets in the Bitnami charts repository.
Affected values:
applicationPortis replaced bycontainerPorts.httpservice.portis renamed toservice.ports.httpservice.nodePortis renamed toservice.nodePorts.httpaccessModeis replaced byaccessModes(a list instead of a simple string)persistence.pathis renamed topersistence.mountPath(a list instead of a simple string)
Also MongoDB® subchart container images were updated to 5.0.x and it can affect compatibility with older versions of MongoDB®.
In this version, the mongodb-exporter bundled as part of the bitnami/mongodb dependency was updated to a new version which, even it is not a major change, can contain breaking changes (from 0.11.X to 0.30.X).
Please visit the release notes from the upstream project at https://github.com/percona/mongodb_exporter/releases
This version standardizes the way of defining Ingress rules. When configuring a single hostname for the Ingress rule, set the ingress.hostname value. When defining more than one, set the ingress.extraHosts array. Apart from this case, no issues are expected to appear when upgrading.
On November 13, 2020, Helm v2 support was formally finished, this major version is the result of the required changes applied to the Helm Chart to be able to incorporate the different features added in Helm v3 and to be consistent with the Helm project itself regarding the Helm v2 EOL.
What changes were introduced in this major version?
- Previous versions of this Helm Chart use
apiVersion: v1(installable by both Helm 2 and 3), this Helm Chart was updated toapiVersion: v2(installable by Helm 3 only). Here you can find more information about theapiVersionfield. - Move dependency information from the requirements.yaml to the Chart.yaml
- After running
helm dependency update, a Chart.lock file is generated containing the same structure used in the previous requirements.lock - The different fields present in the Chart.yaml file has been ordered alphabetically in a homogeneous way for all the Bitnami Helm Charts
- MongoDB® dependency version was bumped to a new major version
10.X.X. Check MongoDB® Upgrading Notes for more information. - Inclusion of the
bitnami/commonlibrary chart and standardization to include common features found on other charts. securityContext.*is deprecated in favor ofpodSecurityContextandcontainerSecurityContext.replicasis deprecated in favor ofreplicaCount.
Considerations when upgrading to this version
- If you want to upgrade to this version from a previous one installed with Helm v3, you shouldn't face any issues
- If you want to upgrade to this version using Helm v2, this scenario is not supported as this version doesn't support Helm v2 anymore
- If you installed the previous version with Helm v2 and wants to upgrade to this version with Helm v3, please refer to the official Helm documentation about migrating from Helm v2 to v3
Useful links
- https://docs.bitnami.com/tutorials/resolve-helm2-helm3-post-migration-issues/
- https://helm.sh/docs/topics/v2_v3_migration/
- https://helm.sh/blog/migrate-from-helm-v2-to-helm-v3/
MongoDB® subchart container images were updated to 4.4.x and it can affect compatibility with older versions of MongoDB®.
Backwards compatibility is not guaranteed since breaking changes were included in MongoDB® subchart. More information in the link below:
This release includes security contexts, so the containers in the chart are run as non-root. More information in this link.
Backwards compatibility is not guaranteed unless you modify the labels used on the chart's deployments. Use the workaround below to upgrade from versions previous to 6.0.0. The following example assumes that the release name is node:
$ kubectl patch deployment node --type=json -p='[{"op": "remove", "path": "/spec/selector/matchLabels/chart"}]'
$ kubectl patch deployment node-mongodb --type=json -p='[{"op": "remove", "path": "/spec/selector/matchLabels/chart"}]'Please, note this Helm chart is a community-supported solution. This means that the Bitnami team is not actively working on new features/improvements nor providing support through GitHub Issues for this Helm chart. Any new issue will stay open for 20 days to allow the community to contribute, after 15 days without activity the issue will be marked as stale being closed after 5 days.
The Bitnami team will review any PR that is created, feel free to create a PR if you find any issue or want to implement a new feature.
New versions are not going to be affected. Once a new version is released in the upstream project, the Bitnami container image will be updated to use the latest version.
Copyright © 2022 Bitnami
Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License.