4 Days Ceph In-Depth Remote/Online Training
What awaits you:
Type:
Virtual training (Instructor-led)
Language:
English
Duration:
28 hours in total. Total duration can be split over 4 days, or 8x 1/2 day.
Class size:
up to 8 participants
Prerequisites:
No prior knowledge of Ceph is required, but basic knowledge of Linux and computer networking is necessary.
Certificate:
Participants will get an official certificate of completion at the end of the training.
CEPH – OPEN SOURCE SCALE-OUT STORAGE
In the following we would like to give you an insight into our training agenda. Please do not hesitate to contact us if you have any questions. We are also happy to include other topics or tailor the course to your individual needs. We can also offer in-house training.
Ceph is a high performance open source storage solution. Thanks to its massive and simple scalability, Ceph is suitable for almost all application scenarios. These include virtual servers, cloud, backup, and much more.
TRAINER
We pay special attention to direct practical relevance in all courses. Therefore all our trainers have extensive knowledge in dealing with Ceph, Linux, networks and software development. As developers of the innovative Ceph Storage Management Software, we can fall back on knowledge from a multitude of projects and thus offer the best possible range of services. We strongly believe that this is unique in the market.
MATERIALS PROVIDED
- Digital documentation
- Virtual training center
TARGET GROUP:
Our course is aimed at IT administrators who need to ensure 24/7 operation of the storage environment. Typically, they are in regular contact with developers, users or run their own storage applications. As previous knowledge we expect basic knowledge in Linux.
AGENDA
INTRODUCTION:
- Introduction of the trainer and croit
- Demonstration of the training process
- Setting up access to the test environment
GENERAL CEPH BASICS:
- History of Ceph
- Horizontal vs. vertical scalability
- Ceph vs. X
- Typical Ceph use cases
- Overview of Ceph management systems
- Introduction Low-Level Object Store (RADOS)
- Introduction: RADOS Block Store (RBD)
- Introduction: POSIX Filesystem (CephFS)
- Introduction High-Level Object Store (S3/Swift)
COMPONENTS OF A CEPH CLUSTER:
- MONs (monitor daemons)
- MGRs (Manager Daemons)
- OSDs (Object Storage Daemons
- CRUSH Map
- MDSs (Meta Data Server)
- RGWs (RADOS Gateways)
- Replication
- Erasure Code (EC)
- Example
INITIAL CEPH CLUSTER SETUP:
- Obtainng Ceph
- Deployment automation
- Setting up the first MON service
- Adding MON services
- MGR services
- Summary
DEPLOYING OSDS:
- BlueStore device setup variants
- Using ceph-deploy
- ceph-volume
- Summary
DISTRIBUTING DATA INTO POOLS WITH CRUSH:
- Ceph pools and Placement Groups
- CRUSH rules
- Adding CRUSH rules
- Erasure coded pools
- Summary
RBD DETAILS:
- Mounting RBD devices
- Using RBD devices
- Details of erasure coded pools
- Best practices
CEPHFS DETAILS:
- Creating and mounting CephFS
- CephFS attributes
OTHER CONNECTORS:
- Setting up and use Samba (SMB / CIFS)
- Setting up and use NFS
RADOSGW DETAILS:
- Deploying RadosGW
- Access via S3 API
CLUSTER PLANNING:
- Cluster planning for 1y,3y,5y
- Clarify requirements
- How much performance can you expect from hard drives?
- Hardware sizing done right
- Failure domains
- Exercise
NETWORK PLANNING:
- Simple network design
- Complex network design
- Typical network related problems
- Example possible network setups
- Dos and Don’ts of cluster networks
AUTHORIZATION MANAGEMENT:
- Ceph keys
- Permissions for RBD
- Permissions for CephFS
- Firewall configuration
OPERATIONS: ERROR HANDLING, UPGRADES AND ALL THAT:
- Scenario: a disk died, what now?
- PG states
- Debugging crashes with log files
- Using the objectstore tool
- Cluster is running full
- Controlling recovery speed
- Upgrading Ceph
MANAGING A CLUSTER WITH CEPHADM:
- Convert an existing cluster to cephadm
- Configure Ceph through cephadm & dashboard
- Deploy & troubleshoot ceph services
- Create custom service configuration
- Update ceph services
MANAGING A CLUSTER WITH CROIT (OPTIONAL):
- Import existing cluster to croit
- Configure Ceph with croit
- Setting up NFS & SMB with croit
- Run an RGW HA group
- Update croit (cluster & container)
CASE STUDIES:
- RGW for Big Data
- RGW for video streaming
- RBD for virtualization
- CephFS with NFS and SMB for backups
- CephFS for a large number of small files
MONITORING AND PERFORMANCE:
- Alerting when something goes wrong
- Tuning performance
- Monitoring Performance
Ceph online training [EAST]
Start: Mar 23, 06:00 AM GMT
End: Apr 3, 09:00 AM GMT
€2,760.00
Remote/Online
Ceph online training [WEST]
Start: Mar 23, 03:00 PM GMT
End: Apr 3, 06:00 PM GMT
€2,760.00
Remote/Online
Ceph online training [EAST]
Start: Sep 29, 05:00 AM GMT
End: Oct 9, 09:00 AM GMT
€2,760.00
Remote/Online
Ceph online training [WEST]
Start: Sep 29, 02:00 PM GMT
End: Oct 9, 06:00 PM GMT
€2,760.00
Remote/Online