mirror of
https://github.com/edgelesssys/constellation.git
synced 2024-10-01 01:36:09 -04:00
add documentation for Rook/CephFS
This commit is contained in:
parent
c378cb19e1
commit
41a312f945
206
.github/docs/nfs.md
vendored
Normal file
206
.github/docs/nfs.md
vendored
Normal file
@ -0,0 +1,206 @@
|
||||
# Deploying NFS in Constellation using Rook
|
||||
|
||||
This document describes how to deploy NFS in Constellation using Rook.
|
||||
|
||||
## Prerequisites
|
||||
Rook/Ceph currently only works with containterd version 1.5.9 or below. Therefore, you need to use a custom VM image. Ask @3u13r for details.
|
||||
|
||||
|
||||
## Create a Cluster
|
||||
|
||||
The cluster needs at least 3 worker nodes, default machines are powerful enough.
|
||||
|
||||
```bash
|
||||
$ constellation create --name nfs -c 1 -w 3
|
||||
```
|
||||
|
||||
## Deploy CSI Driver
|
||||
We need block storage form somewhere. We will use the official Azure CSI for that. We need to create the azure config secret again with the expected fields. Replace "XXX" with the corresponding value from the secret `azureconfig`.
|
||||
|
||||
```bash
|
||||
kubectl create secret generic -n kube-system --from-literal=cloud-config='{"cloud":"AzurePublicCloud","useInstanceMetadata":true,"vmType":"vmss","tenantId":"XXX","subscriptionId":"XXX","resourceGroup":"XXX","location":"XXX", "aadClientId":"XXX","aadClientSecret":"XXX"}' azure-config
|
||||
|
||||
helm repo add azuredisk-csi-driver https://raw.githubusercontent.com/kubernetes-sigs/azuredisk-csi-driver/master/charts
|
||||
helm repo update azuredisk-csi-driver
|
||||
helm install azuredisk-csi-driver azuredisk-csi-driver/azuredisk-csi-driver --namespace kube-system --set linux.distro=fedora --set controller.cloudConfigSecretName=azure-config --set node.cloudConfigSecretName=azure-config
|
||||
```
|
||||
## Deploy the StorageClass:
|
||||
```yaml
|
||||
apiVersion: storage.k8s.io/v1
|
||||
kind: StorageClass
|
||||
metadata:
|
||||
name: managed-premium
|
||||
provisioner: disk.csi.azure.com
|
||||
parameters:
|
||||
skuName: Premium_LRS
|
||||
cachingmode: ReadOnly
|
||||
kind: Managed
|
||||
volumeBindingMode: WaitForFirstConsumer
|
||||
```
|
||||
|
||||
## Deploy Rook:
|
||||
```bash
|
||||
git clone https://github.com/rook/rook.git
|
||||
cd rook/deploy/examples
|
||||
kubectl apply -f common.yaml -f crds.yaml -f operator.yaml
|
||||
kubectl rollout status -n rook-ceph deployment/rook-ceph-operator
|
||||
```
|
||||
|
||||
Apply the following changes to `cluster-on-pvc.yaml`:
|
||||
```diff
|
||||
euler@work:~/projects/rook/deploy/examples$ git diff cluster-on-pvc.yaml
|
||||
diff --git a/deploy/examples/cluster-on-pvc.yaml b/deploy/examples/cluster-on-pvc.yaml
|
||||
index ee4976be2..b5cf294cb 100644
|
||||
--- a/deploy/examples/cluster-on-pvc.yaml
|
||||
+++ b/deploy/examples/cluster-on-pvc.yaml
|
||||
@@ -28,7 +28,7 @@ spec:
|
||||
# size appropriate for monitor data will be used.
|
||||
volumeClaimTemplate:
|
||||
spec:
|
||||
- storageClassName: gp2
|
||||
+ storageClassName: managed-premium
|
||||
resources:
|
||||
requests:
|
||||
storage: 10Gi
|
||||
@@ -59,11 +59,11 @@ spec:
|
||||
# Certain storage class in the Cloud are slow
|
||||
# Rook can configure the OSD running on PVC to accommodate that by tuning some of the Ceph internal
|
||||
# Currently, "gp2" has been identified as such
|
||||
- tuneDeviceClass: true
|
||||
+ tuneDeviceClass: false
|
||||
# Certain storage class in the Cloud are fast
|
||||
# Rook can configure the OSD running on PVC to accommodate that by tuning some of the Ceph internal
|
||||
# Currently, "managed-premium" has been identified as such
|
||||
- tuneFastDeviceClass: false
|
||||
+ tuneFastDeviceClass: true
|
||||
# whether to encrypt the deviceSet or not
|
||||
encrypted: false
|
||||
# Since the OSDs could end up on any node, an effort needs to be made to spread the OSDs
|
||||
@@ -100,7 +100,7 @@ spec:
|
||||
topologySpreadConstraints:
|
||||
- maxSkew: 1
|
||||
# IMPORTANT: If you don't have zone labels, change this to another key such as kubernetes.io/hostname
|
||||
- topologyKey: topology.kubernetes.io/zone
|
||||
+ topologyKey: kubernetes.io/hostname
|
||||
whenUnsatisfiable: DoNotSchedule
|
||||
labelSelector:
|
||||
matchExpressions:
|
||||
@@ -127,7 +127,7 @@ spec:
|
||||
requests:
|
||||
storage: 10Gi
|
||||
# IMPORTANT: Change the storage class depending on your environment
|
||||
- storageClassName: gp2
|
||||
+ storageClassName: managed-premium
|
||||
volumeMode: Block
|
||||
accessModes:
|
||||
- ReadWriteOnce
|
||||
```
|
||||
|
||||
Now apply the yaml:
|
||||
```bash
|
||||
kubectl apply -f cluster-on-pvc.yaml
|
||||
```
|
||||
|
||||
Verify the health of the ceph cluster:
|
||||
```bash
|
||||
$ kubectl apply -f toolbox.yaml
|
||||
$ kubectl -n rook-ceph exec -it deploy/rook-ceph-tools -- ceph status
|
||||
cluster:
|
||||
id: 7c220b31-29f7-4f17-a291-3ef39a9553b3
|
||||
health: HEALTH_OK
|
||||
|
||||
services:
|
||||
mon: 3 daemons, quorum a,b,c (age 2m)
|
||||
mgr: a(active, since 72s)
|
||||
osd: 3 osds: 3 up (since 61s), 3 in (since 81s)
|
||||
|
||||
data:
|
||||
pools: 1 pools, 1 pgs
|
||||
objects: 2 objects, 449 KiB
|
||||
usage: 62 MiB used, 30 GiB / 30 GiB avail
|
||||
pgs: 1 active+clean
|
||||
```
|
||||
|
||||
Deploy the filesystem:
|
||||
```bash
|
||||
$ kubectl apply -f filesystem.yaml
|
||||
$ kubectl -n rook-ceph exec -it deploy/rook-ceph-tools -- ceph status
|
||||
cluster:
|
||||
id: 7c220b31-29f7-4f17-a291-3ef39a9553b3
|
||||
health: HEALTH_OK
|
||||
|
||||
services:
|
||||
mon: 3 daemons, quorum a,b,c (age 3m)
|
||||
mgr: a(active, since 2m)
|
||||
mds: 1/1 daemons up, 1 hot standby
|
||||
osd: 3 osds: 3 up (since 2m), 3 in (since 2m)
|
||||
|
||||
data:
|
||||
volumes: 1/1 healthy
|
||||
pools: 3 pools, 34 pgs
|
||||
objects: 24 objects, 451 KiB
|
||||
usage: 63 MiB used, 30 GiB / 30 GiB avail
|
||||
pgs: 34 active+clean
|
||||
|
||||
io:
|
||||
client: 853 B/s rd, 1 op/s rd, 0 op/s wr
|
||||
|
||||
progress:
|
||||
```
|
||||
|
||||
Deploy the StorageClass:
|
||||
```bash
|
||||
$ kubectl apply -f csi/cephfs/storageclass.yaml
|
||||
```
|
||||
|
||||
## Use the NFS
|
||||
The following deployment will create a PVC based on NFS and mount it into 3 pods.
|
||||
```yaml
|
||||
apiVersion: v1
|
||||
kind: PersistentVolumeClaim
|
||||
metadata:
|
||||
name: nfs
|
||||
spec:
|
||||
accessModes:
|
||||
- ReadWriteMany
|
||||
resources:
|
||||
requests:
|
||||
storage: 5Gi
|
||||
storageClassName: rook-cephfs
|
||||
---
|
||||
# from https://github.com/Azure/kubernetes-volume-drivers/tree/master/nfs
|
||||
apiVersion: apps/v1
|
||||
kind: StatefulSet
|
||||
metadata:
|
||||
name: statefulset-nfs
|
||||
labels:
|
||||
app: nginx
|
||||
spec:
|
||||
serviceName: statefulset-nfs
|
||||
replicas: 3
|
||||
template:
|
||||
metadata:
|
||||
labels:
|
||||
app: nginx
|
||||
spec:
|
||||
containers:
|
||||
- name: statefulset-nfs
|
||||
image: nginx
|
||||
command:
|
||||
- "/bin/sh"
|
||||
- "-c"
|
||||
- "sleep 9999999"
|
||||
volumeMounts:
|
||||
- name: persistent-storage
|
||||
mountPath: /mnt/nfs
|
||||
volumes:
|
||||
- name: persistent-storage
|
||||
persistentVolumeClaim:
|
||||
claimName: nfs
|
||||
readOnly: false
|
||||
updateStrategy:
|
||||
type: RollingUpdate
|
||||
selector:
|
||||
matchLabels:
|
||||
app: nginx
|
||||
```
|
Loading…
Reference in New Issue
Block a user