wip
This commit is contained in:
parent
0b5e0016ec
commit
4f93665533
2 changed files with 63 additions and 1 deletions
61
technology/linux/filesystems/Ceph.md
Normal file
61
technology/linux/filesystems/Ceph.md
Normal file
|
@ -0,0 +1,61 @@
|
|||
---
|
||||
obj: filesystem
|
||||
---
|
||||
|
||||
# Ceph
|
||||
#wip
|
||||
|
||||
Ceph is a distributed storage system providing Object, Block and Filesystem Storage.
|
||||
|
||||
## Concepts
|
||||
- Monitors: A Ceph Monitor (`ceph-mon`) maintains maps of the cluster state, including the monitor map, manager map, the OSD map, the MDS map, and the CRUSH map. These maps are critical cluster state required for Ceph daemons to coordinate with each other. Monitors are also responsible for managing authentication between daemons and clients. At least three monitors are normally required for redundancy and high availability.
|
||||
- Managers: A Ceph Manager daemon (`ceph-mgr`) is responsible for keeping track of runtime metrics and the current state of the Ceph cluster, including storage utilization, current performance metrics, and system load. The Ceph Manager daemons also host python-based modules to manage and expose Ceph cluster information, including a web-based Ceph Dashboard and REST API. At least two managers are normally required for high availability.
|
||||
- Ceph OSDs: An Object Storage Daemon (Ceph OSD, `ceph-osd`) stores data, handles data replication, recovery, rebalancing, and provides some monitoring information to Ceph Monitors and Managers by checking other Ceph OSD Daemons for a heartbeat. At least three Ceph OSDs are normally required for redundancy and high availability.
|
||||
- MDSs: A Ceph Metadata Server (MDS, `ceph-mds`) stores metadata for the Ceph File System. Ceph Metadata Servers allow CephFS users to run basic commands (like ls, find, etc.) without placing a burden on the Ceph Storage Cluster.
|
||||
|
||||
Ceph stores data as objects within logical storage pools. Using the CRUSH algorithm, Ceph calculates which placement group (PG) should contain the object, and which OSD should store the placement group. The CRUSH algorithm enables the Ceph Storage Cluster to scale, rebalance, and recover dynamically.
|
||||
|
||||
## Setup
|
||||
Cephadm creates a new Ceph cluster by bootstrapping a single host, expanding the cluster to encompass any additional hosts, and then deploying the needed services.
|
||||
|
||||
Run the ceph bootstrap command with the IP of the first cluster host:
|
||||
```
|
||||
cephadm bootstrap --mon-ip <mon-ip>
|
||||
```
|
||||
|
||||
This command will:
|
||||
- Create a Monitor and a Manager daemon for the new cluster on the local host.
|
||||
- Generate a new SSH key for the Ceph cluster and add it to the root user’s `/root/.ssh/authorized_keys` file.
|
||||
- Write a copy of the public key to `/etc/ceph/ceph.pub`.
|
||||
- Write a minimal configuration file to `/etc/ceph/ceph.conf`. This file is needed to communicate with Ceph daemons.
|
||||
- Write a copy of the `client.admin` administrative (privileged!) secret key to `/etc/ceph/ceph.client.admin.keyring`.
|
||||
- Add the `_admin` label to the bootstrap host. By default, any host with this label will (also) get a copy of `/etc/ceph/ceph.conf` and `/etc/ceph/ceph.client.admin.keyring`.
|
||||
|
||||
### Ceph CLI
|
||||
The `cephadm shell` command launches a bash shell in a container with all of the Ceph packages installed. By default, if configuration and keyring files are found in `/etc/ceph` on the host, they are passed into the container environment so that the shell is fully functional. Note that when executed on a MON host, cephadm shell will infer the config from the MON container instead of using the default configuration. If `--mount <path>` is given, then the host `<path>` (file or directory) will appear under `/mnt` inside the container:
|
||||
|
||||
```shell
|
||||
cephadm shell
|
||||
```
|
||||
|
||||
To execute ceph commands, you can also run commands like this:
|
||||
```shell
|
||||
cephadm shell -- ceph -s
|
||||
```
|
||||
|
||||
You can install the ceph-common package, which contains all of the ceph commands, including ceph, rbd, mount.ceph (for mounting CephFS file systems), etc.:
|
||||
```shell
|
||||
cephadm add-repo --release reef
|
||||
cephadm install ceph-common
|
||||
```
|
||||
|
||||
Confirm that the ceph command is accessible with:
|
||||
```shell
|
||||
ceph -v
|
||||
ceph status
|
||||
```
|
||||
|
||||
## Host Management
|
||||
#todo -> https://docs.ceph.com/en/latest/cephadm/host-management/
|
||||
|
||||
|
|
@ -12,4 +12,5 @@ obj: meta/collection
|
|||
- [MergerFS](MergerFS.md)
|
||||
- [exFAT](exFAT.md)
|
||||
- [SSHFS](SSHFS.md)
|
||||
- [XFS](XFS.md)
|
||||
- [XFS](XFS.md)
|
||||
- [Ceph](Ceph.md)
|
||||
|
|
Loading…
Reference in a new issue