@helm-charts/bitnami-node v8.1.3-0.1.0
@helm-charts/bitnami-node
Event-driven I/O server-side JavaScript environment based on V8
Field | Value |
---|---|
Repository Name | bitnami |
Chart Name | node |
Chart Version | 8.1.3 |
NPM Package Version | 0.1.0 |
## Global Docker image parameters
## Please, note that this will override the image parameters, including dependencies, configured to use the global value
## Current available global Docker image parameters: imageRegistry and imagePullSecrets
##
# global:
# imageRegistry: myRegistryName
# imagePullSecrets:
# - myRegistryKeySecretName
## Bitnami node image version
## ref: https://hub.docker.com/r/bitnami/node/tags/
##
image:
registry: docker.io
repository: bitnami/node
tag: 8.16.0
## Specify a imagePullPolicy
## Defaults to 'Always' if image tag is 'latest', else set to 'IfNotPresent'
## ref: http://kubernetes.io/docs/user-guide/images/#pre-pulling-images
##
pullPolicy: IfNotPresent
## Optionally specify an array of imagePullSecrets.
## Secrets must be manually created in the namespace.
## ref: https://kubernetes.io/docs/tasks/configure-pod-container/pull-image-private-registry/
##
# pullSecrets:
# - myRegistryKeySecretName
## Bitnami git image version
## ref: https://hub.docker.com/r/bitnami/git/tags/
##
git:
registry: docker.io
repository: bitnami/git
tag: latest
pullPolicy: IfNotPresent
## Optionally specify an array of imagePullSecrets.
## Secrets must be manually created in the namespace.
## ref: https://kubernetes.io/docs/tasks/configure-pod-container/pull-image-private-registry/
##
# pullSecrets:
# - myRegistryKeySecretName
## Git repository http/https
##
repository: https://github.com/bitnami/sample-mean.git
## Git repository revision to checkout
##
revision: master
## Specify the number of replicas for the application
##
replicas: 1
## Specify the port where your application will be running
##
applicationPort: 3000
# Define custom environment variables to pass to the image here
extraEnv: {}
## Kubernetes Service Configuration
service:
## For minikube, set this to NodePort, elsewhere use LoadBalancer
##
type: ClusterIP
port: 80
## Specify the nodePort value for the LoadBalancer and NodePort service types.
## ref: https://kubernetes.io/docs/concepts/services-networking/service/#type-nodeport
##
# nodePort:
## Provide any additional annotations which may be required. This can be used to
## set the LoadBalancer service type to internal only.
## ref: https://kubernetes.io/docs/concepts/services-networking/service/#internal-load-balancer
##
annotations: {}
# loadBalancerIP:
## Enable persistence using Persistent Volume Claims
## ref: http://kubernetes.io/docs/user-guide/persistent-volumes/
##
persistence:
enabled: false
path: /app/data
## If defined, volume.beta.kubernetes.io/storage-class: <storageClass>
## Default: volume.alpha.kubernetes.io/storage-class: default
##
# storageClass:
accessMode: ReadWriteOnce
size: 1Gi
## Configure resource requests and limits
## ref: http://kubernetes.io/docs/user-guide/compute-resources/
##
resources:
{}
# limits:
# cpu: 500m
# memory: 512Mi
# requests:
# cpu: 500m
# memory: 512Mi
##
## MongoDB chart configuration
##
## https://github.com/helm/charts/blob/master/stable/mongodb/values.yaml
##
mongodb:
## Whether to deploy a mongodb server to satisfy the applications database requirements.
## To use an external database set this to false and configure the externaldb parameters
install: true # Check mongodb chart for configuration values
## MongoDB custom user and database
## ref: https://github.com/bitnami/bitnami-docker-mongodb/blob/master/README.md#creating-a-user-and-database-on-first-run
##
mongodbUsername: user
mongodbDatabase: test_db
mongodbPassword: secret_password
## Provision an external database (Only if mongodb.install is false)
## You can:
## 1) Pass an already existing Secret with your database credentials
## 2) Pass an already existing ServiceInstance name and specify the service catalog broker to automatically create a ServiceBinding for your application.
externaldb:
# Set to true if your external database has ssl enabled
ssl: false
# You can use an existing secret containing your database credentials
# Please refer to the respective section in the README to know the details about this secret.
secretName:
# Only if using Kubernetes Service Catalog you can specify the kind of broker used. Available options are osba|gce|aws
type: osba
# If you provide the serviceInstanceName, the chart will create a ServiceBinding for that ServiceInstance
broker:
serviceInstanceName:
## Pod Security Context
## ref: https://kubernetes.io/docs/tasks/configure-pod-container/security-context/
##
securityContext:
enabled: true
fsGroup: 1001
runAsUser: 1001
## Configure ingress resource that allow you to access the application.
## ref: http://kubernetes.io/docs/user-guide/ingress/
##
ingress:
## Set to true to enable ingress record generation
enabled: false
## The list of hostnames to be covered with this ingress record.
## Most likely this will be just one host, but in the event more hosts are needed, this is an array
hosts:
- name: node.local
## Set this to true in order to enable TLS on the ingress record
tls: false
## Set this to true in order to add the corresponding annotations for cert-manager
certManager: false
## If TLS is set to true, you must declare what secret will store the key/certificate for TLS
tlsSecret: node.local-tls
## Ingress annotations done as key:value pairs
## For a full list of possible ingress annotations, please see
## ref: https://github.com/kubernetes/ingress-nginx/blob/master/docs/annotations.md
##
## If certManager is set to true, annotation kubernetes.io/tls-acme: "true" will automatically be set
annotations:
# kubernetes.io/ingress.class: nginx
secrets:
## If you're providing your own certificates, please use this to add the certificates as secrets
## key and certificate should start with -----BEGIN CERTIFICATE----- or
## -----BEGIN RSA PRIVATE KEY-----
##
## name should line up with a tlsSecret set further up
## If you're using cert-manager, this is unneeded, as it will create the secret for you if it is not set
##
## It is also possible to create and manage the certificates outside of this helm chart
## Please see README.md for more information
# - name: node.local-tls
# key:
# certificate:
Node
Node Event-driven I/O server-side JavaScript environment based on V8.
TL;DR
$ helm repo add bitnami https://charts.bitnami.com/bitnami
$ helm install bitnami/node
Introduction
This chart bootstraps a Node deployment on a Kubernetes cluster using the Helm package manager.
It clones and deploys a Node.js application from a Git repository. Optionally, you can set up an Ingress resource to access your application and provision an external database using the Kubernetes service catalog and the Open Service Broker for Azure.
Bitnami charts can be used with Kubeapps for deployment and management of Helm Charts in clusters. This Helm chart has been tested on top of Bitnami Kubernetes Production Runtime (BKPR). Deploy BKPR to get automated TLS certificates, logging and monitoring for your applications.
Prerequisites
- Kubernetes 1.4+ with Beta APIs enabled
- PV provisioner support in the underlying infrastructure
Installing the Chart
To install the chart with the release name my-release
:
$ helm install --name my-release bitnami/node
The command deploys Node.js on the Kubernetes cluster in the default configuration. The configuration section lists the parameters that can be configured during installation. Also includes support for MariaDB chart out of the box.
Due that the Helm Chart clones the application on the /app volume while the container is initializing, a persistent volume is not required.
Tip: List all releases using
helm list
Uninstalling the Chart
To uninstall/delete the my-release
deployment:
$ helm delete my-release
The command removes all the Kubernetes components associated with the chart and deletes the release.
Configuration
The following table lists the configurable parameters of the Node chart and their default values.
Parameter | Description | Default |
---|---|---|
global.imageRegistry | Global Docker image registry | nil |
global.imagePullSecrets | Global Docker registry secret names as an array | [] (does not add image pull secrets to deployed pods) |
image.registry | NodeJS image registry | docker.io |
image.repository | NodeJS image name | bitnami/node |
image.tag | NodeJS image tag | {VERSION} |
image.pullPolicy | NodeJS image pull policy | IfNotPresent |
image.pullSecrets | Specify docker-registry secret names as an array | [] (does not add image pull secrets to deployed pods) |
git.registry | Git image registry | docker.io |
git.repository | Git image name | bitnami/git |
git.tag | Git image tag | latest |
git.pullPolicy | Git image pull policy | Always if imageTag is latest , else IfNotPresent |
repository | Repo of the application | https://github.com/bitnami/sample-mean.git |
revision | Revision to checkout | master |
replicas | Number of replicas for the application | 1 |
applicationPort | Port where the application will be running | 3000 |
extraEnv | Any extra environment variables to be pass to the pods | {} |
securityContext.enabled | Enable security context | true |
securityContext.fsGroup | Group ID for the container | 1001 |
securityContext.runAsUser | User ID for the container | 1001 |
service.type | Kubernetes Service type | ClusterIP |
service.port | Kubernetes Service port | 80 |
service.annotations | Annotations for the Service | {} |
service.loadBalancerIP | LoadBalancer IP if Service type is LoadBalancer | nil |
service.nodePort | NodePort if Service type is LoadBalancer or NodePort | nil |
persistence.enabled | Enable persistence using PVC | false |
persistence.path | Path to persisted directory | /app/data |
persistence.accessMode | PVC Access Mode | ReadWriteOnce |
persistence.size | PVC Storage Request | 1Gi |
mongodb.install | Wheter to install or not the MongoDB chart | true |
externaldb.secretName | Secret containing existing database credentials | nil |
externaldb.type | Type of database that defines the database secret mapping | osba |
externaldb.broker.serviceInstanceName | The existing ServiceInstance to be used | nil |
ingress.enabled | Enable ingress controller resource | false |
ingress.hosts[0].name | Hostname to your Node installation | node.local |
ingress.hosts[0].path | Path within the url structure | / |
ingress.hosts[0].tls | Utilize TLS backend in ingress | false |
ingress.hosts[0].certManager | Add annotations for cert-manager | false |
ingress.hosts[0].tlsSecret | TLS Secret (certificates) | node.local-tls-secret |
ingress.hosts[0].annotations | Annotations for this host's ingress record | [] |
ingress.secrets[0].name | TLS Secret Name | nil |
ingress.secrets[0].certificate | TLS Secret Certificate | nil |
ingress.secrets[0].key | TLS Secret Key | nil |
The above parameters map to the env variables defined in bitnami/node. For more information please refer to the bitnami/node image documentation.
Specify each parameter using the --set key=value[,key=value]
argument to helm install
. For example,
$ helm install --name my-release \
--set repository=https://github.com/jbianquetti-nami/simple-node-app.git,replicas=2 \
bitnami/node
The above command clones the remote git repository to the /app/
directory of the container. Additionally it sets the number of replicas
to 2
.
Alternatively, a YAML file that specifies the values for the above parameters can be provided while installing the chart. For example,
$ helm install --name my-release -f values.yaml bitnami/node
Tip: You can use the default values.yaml
Persistence
The Bitnami Node image stores the Node application and configurations at the /app
path of the container.
Persistent Volume Claims are used to keep the data across deployments. This is known to work in GCE, AWS, and minikube. See the Configuration section to configure the PVC or to disable persistence.
Set up an Ingress controller
First install the nginx-ingress controller via helm:
$ helm install stable/nginx-ingress
Now deploy the node helm chart:
$ helm install --name my-release bitnami/node --set ingress.enabled=true,ingress.host=example.com,service.type=ClusterIP
Configure TLS termination for your ingress controller
You must manually create a secret containing the certificate and key for your domain. You can do it with this command:
$ kubectl create secret tls my-tls-secret --cert=path/to/file.cert --key=path/to/file.key
Then ensure you deploy the Helm chart with the following ingress configuration:
ingress:
enabled: false
path: /
host: example.com
annotations:
kubernetes.io/ingress.class: nginx
tls:
hosts:
- example.com
Connect your application to an already existing database
- Create a secret containing your database credentials:
$ kubectl create secret generic my-database-secret --from-literal=host=YOUR_DATABASE_HOST --from-literal=port=YOUR_DATABASE_PORT --from-literal=username=YOUR_DATABASE_USER --from-literal=password=YOUR_DATABASE_PASSWORD --from-literal=database=YOUR_DATABASE_NAME
YOUR_DATABASE_HOST
, YOUR_DATABASE_PORT
, YOUR_DATABASE_USER
, YOUR_DATABASE_PASSWORD
, and YOUR_DATABASE_NAME
are placeholders that must be replaced with correct values.
- Deploy the node chart specifying the secret name
$ helm install --name node-app --set mongodb.install=false,externaldb.secretName=my-database-secret bitnami/node
Provision a database using the Open Service Broker for Azure
- Install Service Catalog in your Kubernetes cluster following this instructions
- Install the Open Service Broker for Azure in your Kubernetes cluster following this instructions
TIP: you may want to install the osba chart setting the
modules.minStability=EXPERIMENTAL
to see all the available services.$ helm install azure/open-service-broker-azure --name osba --namespace osba \ --set azure.subscriptionId=$AZURE_SUBSCRIPTION_ID \ --set azure.tenantId=$AZURE_TENANT_ID \ --set azure.clientId=$AZURE_CLIENT_ID \ --set azure.clientSecret=$AZURE_CLIENT_SECRET \ --set modules.minStability=EXPERIMENTAL
- Create and deploy a ServiceInstance to provision a database server in Azure cloud.
apiVersion: servicecatalog.k8s.io/v1beta1
kind: ServiceInstance
metadata:
name: azure-mongodb-instance
labels:
app: mongodb
spec:
clusterServiceClassExternalName: azure-cosmosdb-mongo-account
clusterServicePlanExternalName: account
parameters:
location: YOUR_AZURE_LOCATION
resourceGroup: mongodb-k8s-service-catalog
ipFilters:
allowedIPRanges:
- '0.0.0.0/0'
Please update the YOUR_AZURE_LOCATION
placeholder in the above example.
$ kubectl create -f mongodb-service-instance.yml
Deploy the helm chart:
$ helm install --name node-app --set mongodb.install=false,externaldb.broker.serviceInstanceName=azure-mongodb-instance,externaldb.ssl=true bitnami/node
Once the instance has been provisioned in Azure, a new secret should have been automatically created with the connection parameters for your application.
Deploying the helm chart enabling the Azure external database makes the following assumptions:
- You would want an Azure CosmosDB MongoDB database
- Your application uses DATABASE_HOST, DATABASE_PORT, DATABASE_USER, DATABASE_PASSWORD, and DATABASE_NAME environment variables to connect to the database.
You can read more about the kubernetes service catalog at https://github.com/kubernetes-bitnami/service-catalog
Notable changes
7.0.0
This release includes security contexts, so the containers in the chart are run as non-root. More information in this link.
Upgrading
To 6.0.0
Backwards compatibility is not guaranteed unless you modify the labels used on the chart's deployments. Use the workaround below to upgrade from versions previous to 6.0.0. The following example assumes that the release name is node:
$ kubectl patch deployment node --type=json -p='[{"op": "remove", "path": "/spec/selector/matchLabels/chart"}]'
$ kubectl patch deployment node-mongodb --type=json -p='[{"op": "remove", "path": "/spec/selector/matchLabels/chart"}]'
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago
5 years ago