croit demo setup with Vagrant
This is the easiest way to try out croit by using our virtual demo based on Vagrant. If you just want to use our software in a production cluster, follow our production guide.
With the help of our instructions below, you can quickly, easily and conveniently perform your own installation of the croit Ceph Storage Management Software. Our software will lead you to a functional Ceph Cluster with a few steps after the first start.
After the initial installation of your cluster, our software is available to you for 30 days free of charge under the Enterprise license. After the first 30 days, you can purchase the license from us, or continue to use the functionality of the free Community Edition.
The following requirements must be met to run this demo:
- Vagrant with VirtualBox installed
- At least 8 GB RAM
- At least 10 GB disk space, 40 GB if you want to fill up your cluster completely
- A reasonably fast Internet connection, the initial setup will download ~3 GB of data
- We recommend using a recent version of Chrome or Firefox
Note that PXE booting is an unusual scenario for both Vagrant and VirtualBox and may cause unexpected issues. We have last tested these procedures with VirtualBox 5.0.32 and Vagrant 1.9.3 on macOS and Linux.
Caution: The VirtualBox PXE stack sometimes hangs when a server reboots. Shut down the VM completely and restart it instead of resetting it if you encounter this problem.
Installing the croit demo is as simple as running our Vagrant file:
git clone https://github.com/croit/vagrant-demo croit-vagrant cd croit-vagrant vagrant up
You can select a specific croit version by setting the environment variable
CROIT_TAG (a list of available tags can be found here https://hub.docker.com/r/croit/croit/tags) before
Note that this will take several minutes. Don’t worry if the output looks unresponsive while the Docker image (1.5 GB) is being downloaded.
croit is made available by the Vagrant environment at http://localhost:8080 or via https at https://localhost:8443 with a self-signed certificate. Hint: Please not that currently the console Terminal is only working on port 8080 and not over ssl due to Virtualbox NAT restrictions
VMs in this setup
Our Vagrant file defines three VMs:
croit is the main VM running the croit management engine, the three
ceph VMs are identically configured, each featuring three disks to be used as mon, osd, and journal.
The other VMs are started later once croit is configured.
croit VM also contains a docker container with our management node. This is where Ceph is installed, so if you want to access Ceph:
vagrant ssh croit sudo docker exec -it croit bash
Setting up croit
croit will guide you through the initial setup of your Ceph cluster.
admin with password
Now you have to read and accept the unfortunately necessary EULA.
Setup management interface
croit uses one network interface as management interface for all management communication from PXE boot to monitoring running servers.
This demo setup preconfigures an interface with the IP address
192.168.0.2/24 in the croit VM, we will use this as our management interface.
The interface should be named
eth1, but the name might change.
Setup PXE network
This demo setup only features a single network that is used for management, PXE, and Ceph traffic alike.
Create a new setup and configure it as
192.168.0.0/24 with a DHCP pool for the initial boot of new servers.
We recommend using multiple networks for a real setup, but a single network for all traffic is sufficient for this demo. Click save to continue to the next step.
Boot the first server
We now have everything we need to boot our first server.
Run the following Vagrant command to boot the first VM.
vagrant up ceph1
Caution: The live image is downloaded asynchronously from croit.io after the first start, the menu will instruct you to wait and retry if the download is not yet finished.
The server will automatically boot our live image and it will show up in our frontend as soon as you see the boot menu.
Rename the server to something reasonable like
ceph-server1 using the Edit button.
Disks and NICs will show up a few seconds after the server finishes booting.
Configure one of the 1 GB disks as Ceph mon.
mon disks are used to store the Ceph monitor database, each server running a Ceph monitor service needs one disk configured as mon disk.
This formats and mounts the disk on the server, the ‘Set to MON’ button leading to the next step becomes active once this action completes successfully.
Caution: Vagrant doesn’t know about our network and login configuration, it will hence get stuck with the following message:
Warning: Authentication failure. Retrying...
It is safe to cancel Vagrant with
ctrl-c as soon as the VM is up.
The VM will continue to run even though Vagrant complains.
Create the cluster
The last step is creating the Ceph cluster. Each server that was configured in the previous step can be selected as a mon.
Caution: A Ceph monitor has exactly one IP address (even if the server has multiple IPs). This IP needs to be reachable by all Ceph services, clients, and the croit management daemon.
This demo setup only has a single network, making this an easy choice. You will have to create your designated public network during setup if you are using segregated storage and management networks in a real setup.
Our cluster starts out in a bad state as we haven’t created any OSDs yet. Let’s get started by booting the remaining two VMs.
vagrant up ceph2 vagrant up ceph3
The servers will show up in the ‘Hardware’ tab where they can be renamed with the ‘Edit’ action.
Create OSDs on all three servers by selecting ‘Show Disks’ followed by ‘Set to OSD’ on the largest disk. You can configure one of the smaller disks as journal disk to test external journal partitions.
You can also click on ‘Disks’ below the table while not selecting any server to get an overview of all disks on all servers in a single view.
Add hosts to the crush map
croit does not automatically add new hosts to the crush map to avoid data movement before a server is fully configured. Select the CRUSH map tab and add the servers to the default root via drag and drop by dragging them at the drag indicator icon (=). Click ‘Save changes’ to update the crush map.
You can also add CRUSH buckets here to create complex hierarchies.
(Optional) Adding mons
Configure a disk as mon, then select ‘Show services’ to create an additional Ceph monitor service.
Keep in mind that there should be an odd number of mons for optimal fault tolerance.
Test the cluster with RBD
Your cluster is now ready to be used. The health should show up as “Ok” in the status view. Let’s test it by mounting and using a disk image.
Creating an RBD pool and image
You can create an RBD image from the ‘Pools’ view. Click ‘Create Pool’ and select application ‘RBD’, then click ‘Show RBDs’ on the newly created pool. Create an RBD image with KRBD compatibility enabled, this allows us to mount and use the image directly from Linux.
Configure a Ceph client
We will use the croit mangement VM as test client for simplicity here.
The ‘Keys’ view (Settings -> Keys) allows downloading Ceph keyrings and a slimmed-down
ceph.conf for Ceph clients. Use the
All downloads offered by croit can either be downloaded directly or as a copy-pastable link that is valid for 30 minutes.
vagrant ssh # connects to the croit management VM sudo curl -k -o /etc/ceph/ceph.client.admin.keyring http://localhost:8080/api/download/XXXXX/ceph.client.admin.keyring sudo curl -k -o /etc/ceph/ceph.conf http://localhost:8080/api/download/XXXXX/ceph.conf
Caution: The link contains a unique token, copy it from the croit frontend and not from the textbox above.
Caution: Vagrant translates ports, this might break the generated link. The correct ports to use inside the VM are 8080 for HTTP and the default port (443) for HTTPS.
Test the configuration by running
Mount and use the rbd
Run the following inside the VM to mount and use the RBD.
sudo rbd map rbd/test-image # change image name if necessary sudo mkfs.xfs /dev/rbd0 mkdir rbd-test sudo mount /dev/rbd0 rbd-test cd rbd-test sudo dd if=/dev/zero of=./testfile bs=4M count=250
You can use the statistics tab to observe IO and drill down to detailed performance statistics of individual servers.
All of this data is stored in graphite. This allows you feed this data into your existing monitoring systems such as icinga2 or Grafana.
Test the cluster with S3
Description coming soon.