Ceph

PLEASE NOTE: This document applies to v0.5 version and not to the latest stable release v1.9

Shared File System Quickstart

A shared file system can be mounted read-write from multiple pods. This may be useful for applications which can be clustered using a shared filesystem.

This example runs a shared file system for the kube-registry.

Prerequisites

This guide assumes you have created a Rook cluster and pool as explained in the main Kubernetes guide

Rook Client

Setting up the Rook file system requires running rookctl commands with the Rook toolbox. This will be simplified in the future with a CRD for the file system.

Create the File System

Create the file system with the default pools.

rookctl filesystem create --name registryFS

If you are consuming the filesystem from a namespace other than rook you will need to copy the key to the desired namespace. In this example we are copying to the kube-system namespace.

kubectl get secret rook-admin -n rook -o json | jq '.metadata.namespace = "kube-system"' | kubectl apply -f -

Optional: Adjust pool parameters

By default the pools do not have any redundancy. To create another copy of the data, let’s set the replication to 2.

First you will need to launch the Rook toolbox in order to run ceph commands.

Now from the toolbox pod we can modify the pool size:

ceph osd pool set registryFS-data size 2
ceph osd pool set registryFS-metadata size 2

Deploy the Application

Save the following spec as kube-registry.yaml:

apiVersion: v1
kind: ReplicationController
metadata:
  name: kube-registry-v0
  namespace: kube-system
  labels:
    k8s-app: kube-registry
    version: v0
    kubernetes.io/cluster-service: "true"
spec:
  replicas: 3
  selector:
    k8s-app: kube-registry
    version: v0
  template:
    metadata:
      labels:
        k8s-app: kube-registry
        version: v0
        kubernetes.io/cluster-service: "true"
    spec:
      containers:
      - name: registry
        image: registry:2
        resources:
          limits:
            cpu: 100m
            memory: 100Mi
        env:
        - name: REGISTRY_HTTP_ADDR
          value: :5000
        - name: REGISTRY_STORAGE_FILESYSTEM_ROOTDIRECTORY
          value: /var/lib/registry
        volumeMounts:
        - name: image-store
          mountPath: /var/lib/registry
        ports:
        - containerPort: 5000
          name: registry
          protocol: TCP
      volumes:
      - name: image-store
        cephfs:
          monitors:
          - INSERT_MONS_HERE
          user: admin
          secretRef:
            name: rook-admin

We will need to update the yaml with the monitor IP addresses with the following commands. In the future this step will be improved with a Rook volume plugin.

cd cluster/examples/kubernetes
export MONS=$(kubectl -n rook get service -l app=rook-ceph-mon -o json|jq ".items[].spec.clusterIP"|tr -d "\""|sed -e 's/$/:6790/'|paste -s -d, -)
sed "s/INSERT_MONS_HERE/$MONS/g" kube-registry.yaml | kubectl create -f -

You now have a docker registry which is HA with persistent storage.

Test the storage

Once you have pushed an image to the registry (see the instructions to expose and use the kube-registry), verify that kube-registry is using the filesystem that was configured above by mounting the shared file system in the toolbox pod.

Start and connect to the Rook toolbox.

# Mount the same filesystem that the kube-registry is using
mkdir /tmp/registry
rookctl filesystem mount --name registryFS --path /tmp/registry

# If you have pushed images to the registry you will see a directory called docker
ls /tmp/registry

# Cleanup the filesystem mount
rookctl filesystem unmount --path /tmp/registry
rmdir /tmp/registry

Teardown

To clean up all the artifacts created by the file system demo:

kubectl -n kube-system delete secret rook-admin
kubectl delete -f kube-registry.yaml