Using Google Kubernetes Engine (GKE)
Before creating a new cluster, set up your computer as follows:
-
You need a GKE account if you do not already have one.
Creating or Configuring Your Cluster
-
To create your GKE cluster with the
gcloud
command line-tool, run the following command line:gcloud container clusters create <cluster name> \ (1) --cluster-version <version> \ (2) --zone <zone> \ (3) --num-nodes 2 \ --machine-type n1-standard-4 \ --enable-dataplane-v2 \ --enable-ip-alias \ --node-labels io.saagie/type=common,io.saagie/installationId=<installationId> (4) gcloud container node-pools create platform-<platformId> \ (5) --cluster <cluster name> \ --machine-type n1-standard-4 \ --node-labels io.saagie/type=platform,io.saagie/installationId=<installationId>,io.saagie/platform-assignable=<platformId> (4) (5)
Where:
1 <version>
must be replaced with a Kubernetes version that is compatible with Saagie. Current compatible Kubernetes versions are1.20.x
,1.21.x
,1.23.x
, and1.24.x
.2 <cluster name>
and<zone>
must be replaced with the name of your cluster and your zone.3 n1-standard-4
is the minimum machine type that supports machine node and storage prerequisites.4 <installationId>
must be replaced with your installation ID, which must match the prefix you have determined for your DNS entry.5 platformId
must be replaced with the ID of the platform, which is determined during the configuration of your platform. Its value is defined according to the number of platforms and their order, starting from one. You can therefore predict it.The order in which the platforms are declared during configuration must match the order of the platform IDs you entered here in the node pool. So remember it for later.
You can also create your GKE cluster using the Google Cloud Console, but we recommend using the gcloud command line-tool.
|
-
If you are using an existing GKE cluster, create your configuration file by running the following command line:
gcloud container clusters get-credentials <cluster name> --zone <zone> (1)
Where:
1 <cluster name>
and<zone>
must be replaced with the name of your cluster and your zone.
Verifying Your Kubernetes Cluster
-
Open your Google Console and select Connect for your cluster.
-
Run the following command line to verify that you have access to your Kubernetes cluster:
kubectl get nodes
The output of the command should look like the following:
NAME STATUS ROLES AGE VERSION gke-tests1811-pool-d3d39e4c-ct5j Ready <none> 141m v1.14.10-gke.37 gke-tests1811-pool-d3d39e4c-jhhq Ready <none> 141m v1.14.10-gke.37 gke-tests1811-pool-d3d39e4c-pzxf Ready <none> 141m v1.14.10-gke.37
All nodes must have the status ready
.
Creating Storage Classes for Your Saagie Platform
-
Create the
storage.yml
file for your GKE Kubernetes cluster.The following sample storage.yml
file for GKE can be customized according to your needs.--- apiVersion: storage.k8s.io/v1 kind: StorageClass metadata: name: common-storageclass parameters: type: pd-standard provisioner: kubernetes.io/gce-pd --- apiVersion: storage.k8s.io/v1 kind: StorageClass metadata: name: <installationId>-storageclass (1) parameters: type: pd-standard provisioner: kubernetes.io/gce-pd
Where:
1 <installationId>
must be replaced with your installation ID, which must match the prefix you have determined for your DNS entry. -
To store app data and job data on different provisioners, include the following lines in the same
storage.yml
file:--- apiVersion: storage.k8s.io/v1 kind: StorageClass metadata: name: <installationId>-app-storageclass (1) parameters: (2) provisioner: (3)
Where:
1 <installationId>
must be replaced with your installation ID, which must match the prefix you have determined for your DNS entry.2 The parameters
value must contain the parameters for app data.3 The provisioner
value must indicate your second provisioner used to store app data. -
Apply the
storage.yml
file by running the following command line:kubectl apply -f storage.yml
-
Confirm that the storage classes are available by running the following command line:
kubectl get sc
Creating the requirements.yml
File
-
Create your
requirements.yml
file with the code as follows:--- apiVersion: v1 kind: Namespace metadata: name: <installationId> --- apiVersion: v1 kind: ServiceAccount metadata: name: sa-saagie-deploy namespace: <installationId> automountServiceAccountToken: true imagePullSecrets: - name: saagie-docker-config --- kind: ClusterRoleBinding apiVersion: rbac.authorization.k8s.io/v1 metadata: name: sa-saagie-deploy-crbinding namespace: <installationId> roleRef: kind: ClusterRole name: cluster-admin apiGroup: rbac.authorization.k8s.io subjects: - kind: ServiceAccount name: sa-saagie-deploy namespace: <installationId> --- apiVersion: v1 kind: ServiceAccount metadata: name: traefik-ingress-controller namespace: <installationId> imagePullSecrets: - name: saagie-docker-config --- kind: ClusterRoleBinding apiVersion: rbac.authorization.k8s.io/v1 metadata: name: traefik-ingress-cluster-binding subjects: - kind: ServiceAccount name: traefik-ingress-controller namespace: <installationId> roleRef: kind: ClusterRole name: traefik-ingress-cluster apiGroup: rbac.authorization.k8s.io --- kind: ClusterRole apiVersion: rbac.authorization.k8s.io/v1 metadata: name: traefik-ingress-cluster rules: - apiGroups: - "" resources: - services - endpoints - secrets verbs: - get - list - watch - apiGroups: - networking.k8s.io resources: - ingresses - ingressclasses verbs: - get - list - watch - apiGroups: - networking.k8s.io resources: - ingresses/status verbs: - update - apiGroups: - traefik.containo.us resources: - middlewares - middlewaretcps - ingressroutes - traefikservices - ingressroutetcps - ingressrouteudps - tlsoptions - tlsstores - serverstransports verbs: - get - list - watch - apiGroups: - apiextensions.k8s.io resources: - customresourcedefinitions verbs: - create - apiGroups: - apiextensions.k8s.io resourceNames: - middlewares.traefik.containo.us - middlewaretcps.traefik.containo.us - ingressroutes.traefik.containo.us - traefikservices.traefik.containo.us - ingressroutetcps.traefik.containo.us - ingressrouteudps.traefik.containo.us - tlsoptions.traefik.containo.us - tlsstores.traefik.containo.us - serverstransports.traefik.containo.us resources: - customresourcedefinitions verbs: - get --- apiVersion: policy/v1beta1 kind: PodSecurityPolicy metadata: labels: addonmanager.kubernetes.io/mode: Reconcile kubernetes.io/cluster-service: "true" name: 00-saagie-common-psp spec: allowPrivilegeEscalation: false allowedHostPaths: - pathPrefix: /etc/machine-id readOnly: true - pathPrefix: /etc/fluent-bit readOnly: false - pathPrefix: /var/log readOnly: true - pathPrefix: /var/lib/docker/containers readOnly: true - pathPrefix: /data/docker/containers readOnly: true fsGroup: rule: RunAsAny runAsUser: rule: RunAsAny seLinux: rule: RunAsAny supplementalGroups: rule: RunAsAny volumes: - configMap - emptyDir - secret - persistentVolumeClaim - hostPath - projected - downwardAPI --- apiVersion: policy/v1beta1 kind: PodSecurityPolicy metadata: labels: addonmanager.kubernetes.io/mode: Reconcile kubernetes.io/cluster-service: "true" name: 00-saagie-project-psp spec: allowPrivilegeEscalation: true fsGroup: rule: RunAsAny runAsUser: rule: RunAsAny seLinux: rule: RunAsAny supplementalGroups: rule: RunAsAny volumes: - configMap - emptyDir - secret - persistentVolumeClaim - projected - downwardAPI --- apiVersion: rbac.authorization.k8s.io/v1 kind: ClusterRole metadata: labels: addonmanager.kubernetes.io/mode: Reconcile kubernetes.io/cluster-service: "true" name: psp:saagie-common:saagie-common-cluster-psp rules: - apiGroups: - policy resourceNames: - 00-saagie-common-psp resources: - podsecuritypolicies verbs: - use --- apiVersion: rbac.authorization.k8s.io/v1 kind: ClusterRole metadata: labels: addonmanager.kubernetes.io/mode: Reconcile kubernetes.io/cluster-service: "true" name: psp:saagie-common:saagie-project-cluster-psp rules: - apiGroups: - policy resourceNames: - 00-saagie-common-psp resources: - podsecuritypolicies verbs: - use --- apiVersion: rbac.authorization.k8s.io/v1 kind: RoleBinding metadata: name: psp:saagie-common:saagie-deploy-psp-crbinding namespace: <installationId> roleRef: apiGroup: rbac.authorization.k8s.io kind: ClusterRole name: psp:saagie-common:saagie-common-cluster-psp subjects: - kind: Group name: system:serviceaccounts:saagie-common
Where:
-
<installationId>
must be replaced with your installation ID, which must match the prefix you have determined for your DNS entry.
-
-
Apply your
requirements.yml
file by running the following command line:kubectl apply -f requirements.yml
The output of the command should look like the following:
namespace/<installationId> created serviceaccount/sa-saagie-deploy created ... rolebinding.rbac.authorization.k8s.io/psp:saagie-commin:saagie-deploy-psp-cribinding created
Where:
-
<installationId>
must be replaced with your installation ID, which must match the prefix you have determined for your DNS entry.
-
Applying or Installing Secret saagie-docker-config
-
Apply or install the secret:
-
Apply: If you receive the credentials in a Kubernetes secret file, apply the secret to your cluster by running the following
kubectl
command line:kubectl apply -n <installationId> -f saagie-docker-config.yaml (1)
Where:
1 <installationId>
must be replaced with your installation ID, which must match the prefix you have determined for your DNS entry. -
Install: If you receive a username and password, install the secret on your cluster by running the following
kubectl
command line:kubectl create secret docker-registry -n <installationId> saagie-docker-config \ (1) --docker-server=<registry server \ (2) --docker-username=<username> \ (3) --docker-password=<password> (4)
Where:
1 <installationId>
must be replaced with your installation ID, which must match the prefix you have determined for your DNS entry.2 <registry server>
must be replaced with the Docker repository hosting Saagie images.3 <username>
must be replaced with the username provided to you.4 <password>
must be replaced with the password provided to you.
-
-
Edit the default service account to reference the
saagie-docker-config
secret by running the followingkubectl
command line:kubectl patch serviceaccount -n <installationId> default -p '{"imagePullSecrets":[{"name" : "saagie-docker-config"}]}'
Where:
-
<installationId>
must be replaced with your installation ID, which must match the prefix you have determined for your DNS entry.
-
-
Confirm that the secret is properly installed by running the following command line:
kubectl get secret -n <installationId>
Where:
-
<installationId>
must be replaced with your installation ID, which must match the prefix you have determined for your DNS entry.
The output of the command should look like the following:
NAME TYPE DATA AGE saagie-docker-config kubernetes.io/dockerconfigjson 1 2m43s
-
Installing Saagie in Offline Mode
Uploading Docker Images
To upload the Docker images to your registry, make sure you meet all the following prerequisites:
-
A machine with access to your Docker registry.
-
The
tar
archives provided by Saagie, which include the Saagie product and technologies. -
The Skopeo command line tool installed on your machine. For more information, you can refer to the Git repository dedicated to Skopeo.
-
The credentials to push the images into the registry (if any).
-
Run the following command line to decompress the archive:
untar xvf <product-tar-archive> (1)
Where:
1 tar archive
is the file name of the Saagie product provided by Saagie itself. -
OPTIONAL: If you need to require authentication, configure the user and password to connect to your registry using
skopeo login
. For more information, you can refer to the Git repository dedicated to Skopeo). -
Run the following command line in the decompressed archive to start the image upload:
./pushall.sh <registry> (1)
Where:
1 <registry>
is the hostname of your Docker registry.
The process is the same as for uploading Saagie product archives. |
-
Run the following command line to decompress the archive:
untar xvf <technologies-tar-archive> (1)
Where:
1 tar archive
is the file name of the Saagie technologies provided by Saagie itself. -
OPTIONAL: If you need to require authentication, configure the user and password to connect to your registry using
skopeo login
. For more information, you can refer to the Git repository dedicated to Skopeo).If you configured authentication when you uploaded the first tar archive
file, you will not need to configure it again. -
Run the following command line in the decompressed archive to start the image upload:
./pushall.sh <registry> (1)
Where:
1 <registry>
is the hostname of your Docker registry.
Installing Technology Repository
For more information on adding technologies, see our SDK documentation. |
-
Copy the path to the
technologies.zip
file that contains your technologies. -
Run the following
saagiectl
command line to install the repository in your cluster:./bin/saagiectl upload technologies --file <technologies-file> (1)
Where:
1 <technologies-file>
must be replaced with the path to yourtechnologies.zip
file.
Setting Up SMTP (Simple Mail Transfer Protocol) Requirements
An SMTP server is mandatory to send, receive, and relay outgoing mail between your Saagie platform and users' email address. Saagie must therefore have access to your SMTP server and be compatible with the following configurations:
-
SMTP authentication can be anonymous or require authentication.
-
SMTP transport can be SMTP or SMTPS.
-
You must have a valid SSL certificate.
Once configured, you will be able to use your user email address to receive status alerts or change/reset the password associated with your Saagie account.
Deploying and Updating Your SSL Certificate
Make sure that your SSL certificate is valid by checking the following constraints:
-
The certificate’s validity date must be correct.
-
The certificate must include at least the Saagie product URL.
-
The
KeyUsage
attribute must include thedigitalSignature
andkeyEncipherment
elements.
-
Open your preferred terminal command.
-
To deploy (or update) your SSL certificate, run the following command line:
kubectl create secret tls saagie-common-tls --cert=cert.pem --key=cert.key -n <installationId> --dry-run=client -o yaml | kubectl apply -f -
Where:
-
<installationId>
must be replaced with your installation ID, which must match the prefix you have determined for your DNS entry.
-