ceph training

4 Days Ceph In-Depth Remote/Online Training

What awaits you:

Type:

Virtual training (Instructor-led)

Language:

English (German on request)

Duration:

28 hours in total. Total duration can be split over 4x full, 5x 3/4, or 8x 1/2 days.

Prerequisites:

No prior knowledge of Ceph is required, but basic knowledge of Linux and computer networking is necessary.

Certificate:

Participants will get an official certificate of completion at the end of the training.

Cost:

€2,760.00 (Rest of World) $3,277.00 (North America only) Price per person (excl. VAT)

CEPH – OPEN SOURCE SCALE-OUT STORAGE

In the following we would like to give you an insight into our training agenda. Please do not hesitate to contact us if you have any questions. We are also happy to include other topics or tailor the course to your individual needs. We can also offer in-house training.

Ceph is a high performance open source storage solution. Thanks to its massive and simple scalability, Ceph is suitable for almost all application scenarios. These include virtual servers, cloud, backup, and much more.

TRAINER

We pay special attention to direct practical relevance in all courses. Therefore all our trainers have extensive knowledge in dealing with Ceph, Linux, networks and software development. As developers of the innovative Ceph Storage Management Software, we can fall back on knowledge from a multitude of projects and thus offer the best possible range of services. We strongly believe that this is unique in the market.

MATERIALS PROVIDED

  • Digital documentation
  • Virtual training center

TARGET GROUP:

Our course is aimed at IT administrators who need to ensure 24/7 operation of the storage environment. Typically, they are in regular contact with developers, users or run their own storage applications. As previous knowledge we expect basic knowledge in Linux.

AGENDA

INTRODUCTION
  • Introduction of the trainer and croit
  • Demonstration of the training process
  • Setting up access to the test environment
GENERAL CEPH BASICS
  • History of Ceph
  • Horizontal vs. vertical scalability
  • Ceph vs. X
  • Typical Ceph use cases
  • Overview of Ceph management systems
  • Introduction Low-Level Object Store (RADOS)
  • Introduction: RADOS Block Store (RBD)
  • Introduction: POSIX Filesystem (CephFS)
  • Introduction High-Level Object Store (S3/Swift)
COMPONENTS OF A CEPH CLUSTER
  • MONs (monitor daemons)
  • MGRs (Manager Daemons)
  • OSDs (Object Storage Daemons
  • CRUSH Map
  • MDSs (Meta Data Server)
  • RGWs (RADOS Gateways)
  • Replication
  • Erasure Code (EC)
  • Example
INITIAL CEPH CLUSTER SETUP
  • Obtainng Ceph
  • Deployment automation
  • Setting up the first MON service
  • Adding MON services
  • MGR services
  • Summary
DEPLOYING OSDS
  • BlueStore device setup variants
  • Using ceph-deploy
  • ceph-volume
  • Summary
DISTRIBUTING DATA INTO POOLS WITH CRUSH
  • Ceph pools and Placement Groups
  • CRUSH rules
  • Adding CRUSH rules
  • Erasure coded pools
  • Summary
RBD DETAILS
  • Mounting RBD devices
  • Using RBD devices
  • Details of erasure coded pools
  • Best practices
CEPHFS DETAILS
  • Creating and mounting CephFS
  • CephFS attributes
OTHER CONNECTORS
  • Setting up and use Samba (SMB / CIFS)
  • Setting up and use NFS
RADOSGW DETAILS
  • Deploying RadosGW
  • Access via S3 API

CLUSTER PLANNING
  • Cluster planning for 1y,3y,5y
  • Clarify requirements
  • How much performance can you expect from hard drives?
  • Hardware sizing done right
  • Failure domains
  • Exercise
NETWORK PLANNING
  • Simple network design
  • Complex network design
  • Typical network related problems
  • Example possible network setups
  • Dos and Don’ts of cluster networks
AUTHORIZATION MANAGEMENT
  • Ceph keys
  • Permissions for RBD
  • Permissions for CephFS
  • Firewall configuration
OPERATIONS: ERROR HANDLING, UPGRADES AND ALL THAT
  • Scenario: a disk died, what now?
  • PG states
  • Debugging crashes with log files
  • Using the objectstore tool
  • Cluster is running full
  • Controlling recovery speed
  • Upgrading Ceph
MANAGING A CLUSTER WITH CEPHADM
  • Convert an existing cluster to cephadm
  • Configure Ceph through cephadm & dashboard
  • Deploy & troubleshoot ceph services
  • Create custom service configuration
  • Update ceph services
MANAGING A CLUSTER WITH CROIT (OPTIONAL)
  • Import existing cluster to croit
  • Configure Ceph with croit
  • Setting up NFS & SMB with croit
  • Run an RGW HA group
  • Update croit (cluster & container)
CASE STUDIES
  • RGW for Big Data
  • RGW for video streaming
  • RBD for virtualization
  • CephFS with NFS and SMB for backups
  • CephFS for a large number of small files
MONITORING AND PERFORMANCE
  • Alerting when something goes wrong
  • Tuning performance
  • Monitoring Performance

NEXT STEPS

  • Register via the form; you’ll receive a confirmation email.
  • We'll send a quote (in EUR or USD, depending on your region) to the billing contact/company details provided.
  • Enrollment is confirmed on a rolling basis and secured upon receipt of a paid invoice.
  • Class sizes are limited, please complete registration and payment promptly to guarantee your spot.
  • Training access details and the final schedule are sent approximately one week before the course start date.
If none of the listed dates fit your schedule, please contact us at info@croit.io and we’ll be happy to discuss available options.