I am running a Dask
cluster and a Jupyter
notebook server on cloud resources using Kubernetes
and Helm
I am using a yaml
file for the Dask
cluster and Jupyter
, initially taken from https://docs.dask.org/en/latest/setup/kubernetes-helm.html:
apiVersion: v1
kind: Pod
worker:
replicas: 2 #number of workers
resources:
limits:
cpu: 2
memory: 2G
requests:
cpu: 2
memory: 2G
env:
- name: EXTRA_PIP_PACKAGES
value: s3fs --upgrade
# We want to keep the same packages on the workers and jupyter environments
jupyter:
enabled: true
env:
- name: EXTRA_PIP_PACKAGES
value: s3fs --upgrade
resources:
limits:
cpu: 1
memory: 2G
requests:
cpu: 1
memory: 2G
an I am using another yaml
file to create the storage locally.
#CREATE A PERSISTENT VOLUME CLAIM // attached to our pod config
apiVersion: 1
kind: PersistentVolumeClaim
metadata:
name: dask-cluster-persistent-volume-claim
spec:
accessModes:
- ReadWriteOne #can be used by a single node -ReadOnlyMany : for multiple nodes -ReadWriteMany: read/written to/by many nodes
ressources:
requests:
storage: 2Gi # storage capacity
I would like to add a persistent volume claim to the first yaml
file, I couldn't figure out where the add volumes
and volumeMounts
.
if you have an idea, please share it, thank you
I started by creating a pvc claim with the
YAML
file:with lunching in bash:
I checked the creation of persitent volume:
I made major changes to the
Dask
clusterYAML
: I added thevolumes
andvolumeMounts
where I read/write from a directory/data
from the persistent volume created previously, I specifiedServiceType
toLoadBalancer
withport
:Then, I install my
Dask
configuration usinghelm
:Finally, I accessed my
jupyter
interactively:to examine the existence of the
/data
folder