Difference between revisions of "Kubernetes/SAN-Storage"
Jump to navigation
Jump to search
(→Ceph) |
(→Ceph) |
||
Line 1: | Line 1: | ||
= Ceph = | = Ceph = | ||
Configure access to ceph cluster | |||
<source lang=bash> | <source lang=bash> | ||
# Ubuntu 20.04 | # Ubuntu 20.04 | ||
Line 45: | Line 46: | ||
io: | io: | ||
client: 27 MiB/s rd, 5.6 MiB/s wr, 3.88k op/s rd, 191 op/s wr | client: 27 MiB/s rd, 5.6 MiB/s wr, 3.88k op/s rd, 191 op/s wr | ||
# Aliases | # Aliases | ||
alias ceph="ceph -c ~/.ceph/ceph.conf" | alias ceph="ceph -c ~/.ceph/ceph.conf" | ||
alias rbd="rbd -c ~/.ceph/ceph.conf" | alias rbd="rbd -c ~/.ceph/ceph.conf" | ||
</source> | |||
[https://access.redhat.com/documentation/en-us/red_hat_ceph_storage/1.2.3/html/ceph_block_device/get-a-list-of-images Operations] | |||
<source lang=bash> | |||
# List block device images | |||
rbd ls {poolname} -c ~/.ceph/ceph.conf | |||
kubernetes-dynamic-pvc-aaa9e0ff-14d9-479e-a425-aaaaaaaaaaaa | |||
kubernetes-dynamic-pvc-aaa194fb-cdc3-4cb4-85e9-aaaaaaaaaaaa | |||
myapp-postgresql-pv-0 | |||
myapp-postgresql-pv-1 | |||
otheapp-pg-data-0-volume | |||
# Create a block device image | |||
rbd create {image-name} --size {megabytes} --pool {pool-name} | |||
rbd create dev-prometheus-data-0-vol --size 3072 --pool DEV_block # 3Gb | |||
rbd info dev-prometheus-data-0-vol --pool DEV_block -c ~/.ceph/ceph.conf | |||
rbd image 'dev-prometheus-data-0-vol': | |||
size 3 GiB in 768 objects | |||
order 22 (4 MiB objects) | |||
snapshot_count: 0 | |||
id: 0f582d6be65aaa | |||
block_name_prefix: rbd_data.0f582d6be65aaa | |||
format: 2 | |||
features: layering, exclusive-lock, object-map, fast-diff, deep-flatten | |||
op_features: | |||
flags: | |||
create_timestamp: Tue Nov 11 13:30:53 2021 | |||
access_timestamp: Tue Nov 11 13:30:53 2021 | |||
modify_timestamp: Tue Nov 11 13:30:53 2021 | |||
</source> | </source> |
Revision as of 14:45, 16 November 2021
Ceph
Configure access to ceph cluster
# Ubuntu 20.04 sudo apt install ceph-common # Config file cat > ~/.ceph/ceph.conf <<EOF [global] mon_host = XXXXXXX keyring = /home/myuser/.ceph/ceph.client.admin.keyring # requires full absolute path auth_cluster_required = cephx auth_service_required = cephx auth_client_required = cephx EOF cat > ~/.ceph/ceph.client.admin.keyring <<EOF [client.admin] key = XXXXXXXXXXXX== EOF # Test ceph -c ~/.ceph/ceph.conf status cluster: id: aaaaaaaa-bbbb-cccc-dddd-eeeeeeeeeeeee health: HEALTH_OK services: mon: 3 daemons, quorum dc1ceph11-2222,dc2ceph21-3333,dc3ceph31-4444 (age 4h) mgr: dc1ceph11-2222.dddddd(active, since 51m), standbys: dc2ceph21-3333.eeeeee mds: devcephfs:1 {0=devcephfs.dc3ceph31-4444.nmngty=up:active} 2 up:standby osd: 20 osds: 19 up (since 4d), 19 in (since 4d) rgw: 1 daemon active (admin) task status: scrub status: mds.devcephfs.dc3ceph31-4444.nmngty: idle data: pools: 22 pools, 449 pgs objects: 10.77M objects, 25 TiB usage: 54 TiB used, 85 TiB / 139 TiB avail pgs: 447 active+clean 2 active+clean+scrubbing+deep io: client: 27 MiB/s rd, 5.6 MiB/s wr, 3.88k op/s rd, 191 op/s wr # Aliases alias ceph="ceph -c ~/.ceph/ceph.conf" alias rbd="rbd -c ~/.ceph/ceph.conf"
# List block device images rbd ls {poolname} -c ~/.ceph/ceph.conf kubernetes-dynamic-pvc-aaa9e0ff-14d9-479e-a425-aaaaaaaaaaaa kubernetes-dynamic-pvc-aaa194fb-cdc3-4cb4-85e9-aaaaaaaaaaaa myapp-postgresql-pv-0 myapp-postgresql-pv-1 otheapp-pg-data-0-volume # Create a block device image rbd create {image-name} --size {megabytes} --pool {pool-name} rbd create dev-prometheus-data-0-vol --size 3072 --pool DEV_block # 3Gb rbd info dev-prometheus-data-0-vol --pool DEV_block -c ~/.ceph/ceph.conf rbd image 'dev-prometheus-data-0-vol': size 3 GiB in 768 objects order 22 (4 MiB objects) snapshot_count: 0 id: 0f582d6be65aaa block_name_prefix: rbd_data.0f582d6be65aaa format: 2 features: layering, exclusive-lock, object-map, fast-diff, deep-flatten op_features: flags: create_timestamp: Tue Nov 11 13:30:53 2021 access_timestamp: Tue Nov 11 13:30:53 2021 modify_timestamp: Tue Nov 11 13:30:53 2021