Storage
GlusterFS storage can be expanded for provisioned upstream k8s clusters. GlusterFS storage expansion can be done
- By Adding a new storage device to the existing storage node or
- By Adding a brand new node with a Storage Role.
Add New Storage Device To Node¶
- Add the additional raw/unformatted volume/disk to your VMs or instances
- Sign into the Web Console
- Select the cluster and click on nodes
- Click "Add Storage Device" on the node and wait for the newly added storage devices to be discovered by the controller
- Select the new storage device to expand the GlusterFS storage to.
-
Click on "Save" to start expanding the storage. It will take couple minutes for the additional storage to be added.
-
Once it is completed, the new storage device will be shown under the node
Add New Node With Storage Role¶
- Create required VMs or instances with the raw/unformatted volume attached to it.
- Sign into the Web Console
- Select the cluster and click on nodes
- Click "Add Node" and Follow the Node Installation Instructions to install the bootstrap agent on the VM
- Approve the node.
Users can optionally enable "auto approval" for new nodes to join the cluster. Enable the auto approval "toggle" in the cluster configuration page as shown in the screenshot below.
- Click on "Configure" to configure the node with the Storage Role and select the storage device and Save the settings.
- Click on "Provision" and confirm to start adding this node as a new node with storage role to the existing cluster
- It will take couple minutes for the additional node to be provisioned.
- Once the node is provisioned, it will join the cluster with the Storage role
Add rook/ceph Storage¶
Optionally, users can also add rook-ceph storage to the new nodes or the existing node by introducing a new device on Day 2, post upstream cluster provision. The rook-ceph version 1.8.1 is supported.
- Select the required upstream cluster and click Nodes tab
- Select the required node and click the down arrow to expand the Nodes section
- Click Edit Labels button
- Click Create Key-Value Label to add two labels for rook-ceph storage as shown in the below image
- Key role and Value storage
- Key storage and Value rook-ceph
- Click Save
- The newly added storage labels appears as shown below
Once these labels are added, user must update the blueprint to default-upstream to deploy the rook-ceph storage to the cluster. Refer Update Blueprint to know more about update blueprint process
Important
Storage expansion or update may get failed when adding a new device or a new storage node. To overcome this issue, restart the rook-operator pod.