文章目录
- 1.查看集群的状态信息
- 2.动态的查看集群的状态信息
- 3.查看集群的利用率
- 4.查看OSD的资源利用率
- 5.查看OSD的列表
- 6.查看各组件的状态
- 7.查看集群的仲裁信息
- 8.查看/修改集群组件sock的配置参数
 
1.查看集群的状态信息
通过集群状态信息可以看到集群的健康状态、各个组件的运行状态以及数据使用情况。
查看集群状态有两种方式:
- 使用ceph命令加参数的方式-s或者status
[root@ceph-node-1 ~]# ceph -scluster:id:     a5ec192a-8d13-4624-b253-5b350a616041health: HEALTH_OKservices:mon: 3 daemons, quorum ceph-node-1,ceph-node-2,ceph-node-3 (age 2d)mgr: ceph-node-1(active, since 2d), standbys: ceph-node-2, ceph-node-3mds: cephfs-storage:1 {0=ceph-node-2=up:active} 2 up:standbyosd: 7 osds: 7 up (since 95m), 7 in (since 2d); 1 remapped pgsrgw: 1 daemon active (ceph-node-1)task status:data:pools:   9 pools, 240 pgsobjects: 275 objects, 588 KiBusage:   7.6 GiB used, 62 GiB / 70 GiB availpgs:     239 active+clean1   active+clean+remapped
- 使用ceph交互式界面查看状态
[root@ceph-node-1 ~]# ceph
ceph> statuscluster:id:     a5ec192a-8d13-4624-b253-5b350a616041health: HEALTH_OKservices:mon: 3 daemons, quorum ceph-node-1,ceph-node-2,ceph-node-3 (age 2d)mgr: ceph-node-1(active, since 2d), standbys: ceph-node-2, ceph-node-3mds: cephfs-storage:1 {0=ceph-node-2=up:active} 2 up:standbyosd: 7 osds: 7 up (since 95m), 7 in (since 2d); 1 remapped pgsrgw: 1 daemon active (ceph-node-1)task status:data:pools:   9 pools, 240 pgsobjects: 275 objects, 588 KiBusage:   7.6 GiB used, 62 GiB / 70 GiB availpgs:     239 active+clean1   active+clean+remapped
2.动态的查看集群的状态信息
类似于tail命令,持续变化刷新集群的状态信息。
[root@ceph-node-1 ~]# ceph -wcluster:id:     a5ec192a-8d13-4624-b253-5b350a616041health: HEALTH_OKservices:mon: 3 daemons, quorum ceph-node-1,ceph-node-2,ceph-node-3 (age 2d)mgr: ceph-node-1(active, since 2d), standbys: ceph-node-2, ceph-node-3mds: cephfs-storage:1 {0=ceph-node-2=up:active} 2 up:standbyosd: 7 osds: 7 up (since 99m), 7 in (since 2d); 1 remapped pgsrgw: 1 daemon active (ceph-node-1)task status:data:pools:   9 pools, 240 pgsobjects: 275 objects, 588 KiBusage:   7.6 GiB used, 62 GiB / 70 GiB availpgs:     239 active+clean1   active+clean+remapped
3.查看集群的利用率
通过ceph df命令可以看到集群的资源利用率,会显示出Ceph集群的总空间容量、剩余容量、使用的容量,还会显示出集群中Pool资源池的利用率。
[root@ceph-node-1 ~]# ceph df
RAW STORAGE:CLASS     SIZE       AVAIL      USED        RAW USED     %RAW USED hdd       70 GiB     62 GiB     644 MiB      7.6 GiB         10.90 TOTAL     70 GiB     62 GiB     644 MiB      7.6 GiB         10.90 POOLS:POOL                          ID     PGS     STORED      OBJECTS     USED        %USED     MAX AVAIL ceph-rbd-data                  1      16        19 B           3     192 KiB         0        20 GiB .rgw.root                      2      32     1.2 KiB           4     768 KiB         0        20 GiB default.rgw.control            3      32         0 B           8         0 B         0        20 GiB default.rgw.meta               4      32     3.4 KiB          19       3 MiB         0        20 GiB default.rgw.log                5      32         0 B         207         0 B         0        20 GiB default.rgw.buckets.index      6      32         0 B           3         0 B         0        20 GiB default.rgw.buckets.data       7      32     157 KiB           8     1.9 MiB         0        20 GiB cephfs_metadata                8      16     426 KiB          23     2.8 MiB         0        20 GiB cephfs_data                    9      16         0 B           0         0 B         0        20 GiB
4.查看OSD的资源利用率
通过ceph osd df命令可以查看集群中每个OSD的资源利用率,包括使用率和可用空间的大小,以及OSD的状态,通过此命令可以观察OSD的空间剩余,如果OSD的空间不足,也会导致集群出现问题,当OSD空间不足时,可以扩容OSD或者删除上层文件来解决。
[root@ceph-node-1 ~]# ceph osd df
ID CLASS WEIGHT  REWEIGHT SIZE   RAW USE DATA    OMAP   META     AVAIL   %USE  VAR  PGS STATUS 0   hdd 0.00980  1.00000 10 GiB 1.1 GiB  92 MiB 23 KiB 1024 MiB 8.9 GiB 10.90 1.00  79     up 3   hdd 0.00980  1.00000 10 GiB 1.1 GiB  92 MiB 20 KiB 1024 MiB 8.9 GiB 10.91 1.00  93     up 6   hdd 0.00980  1.00000 10 GiB 1.1 GiB  91 MiB 20 KiB 1024 MiB 8.9 GiB 10.89 1.00  68     up 1   hdd 0.00980  1.00000 10 GiB 1.1 GiB  92 MiB    0 B    1 GiB 8.9 GiB 10.90 1.00  74     up 4   hdd 0.00980  1.00000 10 GiB 1.1 GiB  92 MiB    0 B    1 GiB 8.9 GiB 10.90 1.00  74     up 7   hdd 0.00980  1.00000 10 GiB 1.1 GiB  91 MiB    0 B    1 GiB 8.9 GiB 10.90 1.00  92     up 2   hdd 0.00980  1.00000 10 GiB 1.1 GiB  94 MiB 31 KiB 1024 MiB 8.9 GiB 10.92 1.00 240     up TOTAL 70 GiB 7.6 GiB 644 MiB 95 KiB  7.0 GiB  62 GiB 10.90     
5.查看OSD的列表
[root@ceph-node-1 ~]# ceph osd tree
ID CLASS WEIGHT  TYPE NAME            STATUS REWEIGHT PRI-AFF 
-1       0.06857 root default                                 
-3       0.02939     host ceph-node-1                         0   hdd 0.00980         osd.0            up  1.00000 1.00000 3   hdd 0.00980         osd.3            up  1.00000 1.00000 6   hdd 0.00980         osd.6            up  1.00000 1.00000 
-5       0.02939     host ceph-node-2                         1   hdd 0.00980         osd.1            up  1.00000 1.00000 4   hdd 0.00980         osd.4            up  1.00000 1.00000 7   hdd 0.00980         osd.7            up  1.00000 1.00000 
-7       0.00980     host ceph-node-3                         2   hdd 0.00980         osd.2            up  1.00000 1.00000 
6.查看各组件的状态
查看组件的状态:
 命令格式ceph {组件} stat
查看组件的详细信息:
 命令格式ceph {组件} dump
[root@ceph-node-1 ~]# ceph mon stat
e1: 3 mons at {ceph-node-1=[v2:192.168.20.20:3300/0,v1:192.168.20.20:6789/0],ceph-node-2=[v2:192.168.20.21:3300/0,v1:192.168.20.21:6789/0],ceph-node-3=[v2:192.168.20.22:3300/0,v1:192.168.20.22:6789/0]}, election epoch 44, leader 0 ceph-node-1, quorum 0,1,2 ceph-node-1,ceph-node-2,ceph-node-3[root@ceph-node-1 ~]# ceph mon dump
epoch 1
fsid a5ec192a-8d13-4624-b253-5b350a616041
last_changed 2022-04-02 22:09:57.238072
created 2022-04-02 22:09:57.238072
min_mon_release 14 (nautilus)
0: [v2:192.168.20.20:3300/0,v1:192.168.20.20:6789/0] mon.ceph-node-1
1: [v2:192.168.20.21:3300/0,v1:192.168.20.21:6789/0] mon.ceph-node-2
2: [v2:192.168.20.22:3300/0,v1:192.168.20.22:6789/0] mon.ceph-node-3
7.查看集群的仲裁信息
[root@ceph-node-1 ~]# ceph quorum_status
{"election_epoch":44,"quorum":[0,1,2],"quorum_names":["ceph-node-1","ceph-node-2","ceph-node-3"],"quorum_leader_name":"ceph-node-1","quorum_age":193280,"monmap":{"epoch":1,"fsid":"a5ec192a-8d13-4624-b253-5b350a616041","modified":"2022-04-02 22:09:57.238072","created":"2022-04-02 22:09:57.238072","min_mon_release":14,"min_mon_release_name":"nautilus","features":{"persistent":["kraken","luminous","mimic","osdmap-prune","nautilus"],"optional":[]},"mons":[{"rank":0,"name":"ceph-node-1","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.20.20:3300","nonce":0},{"type":"v1","addr":"192.168.20.20:6789","nonce":0}]},"addr":"192.168.20.20:6789/0","public_addr":"192.168.20.20:6789/0"},{"rank":1,"name":"ceph-node-2","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.20.21:3300","nonce":0},{"type":"v1","addr":"192.168.20.21:6789","nonce":0}]},"addr":"192.168.20.21:6789/0","public_addr":"192.168.20.21:6789/0"},{"rank":2,"name":"ceph-node-3","public_addrs":{"addrvec":[{"type":"v2","addr":"192.168.20.22:3300","nonce":0},{"type":"v1","addr":"192.168.20.22:6789","nonce":0}]},"addr":"192.168.20.22:6789/0","public_addr":"192.168.20.22:6789/0"}]}}
8.查看/修改集群组件sock的配置参数
查看组件的配置信息:
 命令格式:ceph --admin-daemon {sock文件路径} config show
设置组件sock的配置参数:
 命令格式:ceph --admin-daemon {sock文件路径} config set 参数 值
查看组件sock的配置参数:
 命令格式:ceph --admin-daemon {sock文件路径} config get参数
集群各组件的sock文件都位于/var/run/ceph路径。

1)查看集群组件的配置信息
[root@ceph-node-1 ~]# ceph --admin-daemon /var/run/ceph/ceph-mon.ceph-node-1.asok config show

2)查看mon组件的仲裁选举信息
[root@ceph-node-1 ~]# ceph --admin-daemon /var/run/ceph/ceph-mon.ceph-node-1.asok quorum_status

3)调整mon组件的配置参数
[root@ceph-node-1 ~]# ceph --admin-daemon /var/run/ceph/ceph-mon.ceph-node-1.asok config set mon_clock_drift_allowed 1.0
{"success": "mon_clock_drift_allowed = '1.000000' (not observed, change may require restart) "
}