分布式存储系统之Ceph集群状态获取及ceph配置文件说明

今天我们来聊一聊获取ceph集群状态和ceph配置文件说明相关话题;

  Ceph集群状态获取常用命令

  1、ceph -s :该命令用于输出ceph集群系统状态信息

  提示:ceph -s主要输出有三类信息,一类是集群相关信息,比如集群id,健康状态;第二类是服务类相关信息,比如集群运行了几个mon节点,几个mgr节点,几个mds,osd和rgw;这些服务都处于什么样的状态等等;我们把这些信息称为集群运行状况,它可以让我们一目了然的了解到集群现有运行状况;第三类信息是数据存储类的信息;比如有多少个存储池,和pg数量;usage用来展示集群使用容量和剩余容量以及总容量;这里需要注意一点,集群显示的总磁盘大小,它不等于可以存储这么多对象数据;因为每一个对象数据都多个副本,所以真正能够存储对象数据的量应该根据副本的数量来计算;默认情况下,我们创建的存储都是副本型存储池,副本数量是3个(其中一个主,两个从),即每一个对象数据都会存储三份,所以真正能够存储对象数据的空间只有总空间的三分之一。  

  获取集群的即时状态信息

  2、获取pg的状态



1

2

3

[cephadm@ceph-admin ceph-cluster]$ ceph pg stat

304 pgs: 304 active+clean; 3.8 KiB data, 10 GiB used, 890 GiB / 900 GiB avail

[cephadm@ceph-admin ceph-cluster]$



  3、获取存储池的状态



1

2

3

4

5

6

7

8

9

10

11

12

13

14

15

16

17

18

19

20

21

22

23

24

25

26

[cephadm@ceph-admin ceph-cluster]$ ceph osd pool stats

pool testpool id 1

nothing is going on

pool rbdpool id 2

nothing is going on

pool .rgw.root id 3

nothing is going on

pool default.rgw.control id 4

nothing is going on

pool default.rgw.meta id 5

nothing is going on

pool default.rgw.log id 6

nothing is going on

pool cephfs-metadatpool id 7

nothing is going on

pool cephfs-datapool id 8

nothing is going on

[cephadm@ceph-admin ceph-cluster]$



  提示:如果后面没有跟指定的存储表示获取所有存储的状态;

  4、获取存储池大小和空间使用情况



1

2

3

4

5

6

7

8

9

10

11

12

13

14

15

[cephadm@ceph-admin ceph-cluster]$ ceph df

GLOBAL:

SIZE AVAIL RAW USED %RAW USED

900 GiB 890 GiB 10 GiB 1.13

POOLS:

NAME ID USED %USED MAX AVAIL OBJECTS

testpool 1 0 B 0 281 GiB 0

rbdpool 2 389 B 0 281 GiB 5

.rgw.root 3 1.1 KiB 0 281 GiB 4

default.rgw.control 4 0 B 0 281 GiB 8

default.rgw.meta 5 0 B 0 281 GiB 0

default.rgw.log 6 0 B 0 281 GiB 175

cephfs-metadatpool 7 2.2 KiB 0 281 GiB 22

cephfs-datapool 8 0 B 0 281 GiB 0

[cephadm@ceph-admin ceph-cluster]$



  提示:ceph df输出的内容主要分两大段,第一段是global,全局存储空间用量情况;size表示总空间大小,avail表示剩余空间大小;RAW USED表示已用到原始存储空间;%RAW USED表示已用原始空间占比重空间的比例;第二段是相关存储空间使用情况;其中MAX AVAIL表示对应存储池能够使用的最大容量;OBJECTS表示该存储池中对象的个数;

  获取存储空间用量详细情况



1

2

3

4

5

6

7

8

9

10

11

12

13

14

15

[cephadm@ceph-admin ceph-cluster]$ ceph df detail

GLOBAL:

SIZE AVAIL RAW USED %RAW USED OBJECTS

900 GiB 890 GiB 10 GiB 1.13 214

POOLS:

NAME ID QUOTA OBJECTS QUOTA BYTES USED %USED MAX AVAIL OBJECTS DIRTY READ WRITE RAW USED

testpool 1 N/A N/A 0 B 0 281 GiB 0 0 2 B 2 B 0 B

rbdpool 2 N/A N/A 389 B 0 281 GiB 5 5 75 B 19 B 1.1 KiB

.rgw.root 3 N/A N/A 1.1 KiB 0 281 GiB 4 4 66 B 4 B 3.4 KiB

default.rgw.control 4 N/A N/A 0 B 0 281 GiB 8 8 0 B 0 B 0 B

default.rgw.meta 5 N/A N/A 0 B 0 281 GiB 0 0 0 B 0 B 0 B

default.rgw.log 6 N/A N/A 0 B 0 281 GiB 175 175 7.2 KiB 4.8 KiB 0 B

cephfs-metadatpool 7 N/A N/A 2.2 KiB 0 281 GiB 22 22 0 B 45 B 6.7 KiB

cephfs-datapool 8 N/A N/A 0 B 0 281 GiB 0 0 0 B 0 B 0 B

[cephadm@ceph-admin ceph-cluster]$



  5、检查OSD和MON的状态



1

2

3

4

5

6

7

8

9

10

11

12

13

14

15

16

17

18

19

20

21

22

23

24

25

26

27

28

29

30

31

32

33

34

35

36

[cephadm@ceph-admin ceph-cluster]$ ceph osd stat

10 osds: 10 up, 10 in; epoch: e99

[cephadm@ceph-admin ceph-cluster]$ ceph osd dump

epoch 99

fsid 7fd4a619-9767-4b46-9cee-78b9dfe88f34

created 2022-09-24 00:36:13.639715

modified 2022-09-25 12:33:15.111283

flags sortbitwise,recovery_deletes,purged_snapdirs

crush_version 25

full_ratio 0.95

backfillfull_ratio 0.9

nearfull_ratio 0.85

require_min_compat_client jewel

min_compat_client jewel

require_osd_release mimic

pool 1 'testpool' replicated size 3 min_size 2 crush_rule 0 object_hash rjenkins pg_num 16 pgp_num 16 last_change 42 flags hashpspool stripe_width 0

pool 2 'rbdpool' replicated size 3 min_size 2 crush_rule 0 object_hash rjenkins pg_num 64 pgp_num 64 last_change 81 flags hashpspool,selfmanaged_snaps stripe_width 0 application rbd

removed_snaps [1~3]

pool 3 '.rgw.root' replicated size 3 min_size 2 crush_rule 0 object_hash rjenkins pg_num 8 pgp_num 8 last_change 84 owner 18446744073709551615 flags hashpspool stripe_width 0 application rgw

pool 4 'default.rgw.control' replicated size 3 min_size 2 crush_rule 0 object_hash rjenkins pg_num 8 pgp_num 8 last_change 87 owner 18446744073709551615 flags hashpspool stripe_width 0 application rgw

pool 5 'default.rgw.meta' replicated size 3 min_size 2 crush_rule 0 object_hash rjenkins pg_num 8 pgp_num 8 last_change 89 owner 18446744073709551615 flags hashpspool stripe_width 0 application rgw

pool 6 'default.rgw.log' replicated size 3 min_size 2 crush_rule 0 object_hash rjenkins pg_num 8 pgp_num 8 last_change 91 owner 18446744073709551615 flags hashpspool stripe_width 0 application rgw

pool 7 'cephfs-metadatpool' replicated size 3 min_size 2 crush_rule 0 object_hash rjenkins pg_num 64 pgp_num 64 last_change 99 flags hashpspool stripe_width 0 application cephfs

pool 8 'cephfs-datapool' replicated size 3 min_size 2 crush_rule 0 object_hash rjenkins pg_num 128 pgp_num 128 last_change 99 flags hashpspool stripe_width 0 application cephfs

max_osd 10

osd.0 up in weight 1 up_from 67 up_thru 96 down_at 66 last_clean_interval [64,65) 192.168.0.71:6802/1361 172.16.30.71:6802/1361 172.16.30.71:6803/1361 192.168.0.71:6803/1361 exists,up bf3649af-e3f4-41a2-a5ce-8f1a316d344e

osd.1 up in weight 1 up_from 68 up_thru 96 down_at 66 last_clean_interval [64,65) 192.168.0.71:6800/1346 172.16.30.71:6800/1346 172.16.30.71:6801/1346 192.168.0.71:6801/1346 exists,up 7293a12a-7b4e-4c86-82dc-0acc15c3349e

osd.2 up in weight 1 up_from 67 up_thru 96 down_at 66 last_clean_interval [60,65) 192.168.0.72:6800/1389 172.16.30.72:6800/1389 172.16.30.72:6801/1389 192.168.0.72:6801/1389 exists,up 96c437c5-8e82-4486-910f-9e98d195e4f9

osd.3 up in weight 1 up_from 67 up_thru 96 down_at 66 last_clean_interval [60,65) 192.168.0.72:6802/1406 172.16.30.72:6802/1406 172.16.30.72:6803/1406 192.168.0.72:6803/1406 exists,up 4659d2a9-09c7-49d5-bce0-4d2e65f5198c

osd.4 up in weight 1 up_from 71 up_thru 96 down_at 68 last_clean_interval [59,66) 192.168.0.73:6802/1332 172.16.30.73:6802/1332 172.16.30.73:6803/1332 192.168.0.73:6803/1332 exists,up de019aa8-3d2a-4079-a99e-ec2da2d4edb9

osd.5 up in weight 1 up_from 71 up_thru 96 down_at 68 last_clean_interval [58,66) 192.168.0.73:6800/1333 172.16.30.73:6800/1333 172.16.30.73:6801/1333 192.168.0.73:6801/1333 exists,up 119c8748-af3b-4ac4-ac74-6171c90c82cc

osd.6 up in weight 1 up_from 69 up_thru 96 down_at 68 last_clean_interval [59,66) 192.168.0.74:6800/1306 172.16.30.74:6800/1306 172.16.30.74:6801/1306 192.168.0.74:6801/1306 exists,up 08d8dd8b-cdfe-4338-83c0-b1e2b5c2a799

osd.7 up in weight 1 up_from 69 up_thru 96 down_at 68 last_clean_interval [60,65) 192.168.0.74:6802/1301 172.16.30.74:6802/1301 172.16.30.74:6803/1301 192.168.0.74:6803/1301 exists,up 9de6cbd0-bb1b-49e9-835c-3e714a867393

osd.8 up in weight 1 up_from 73 up_thru 96 down_at 66 last_clean_interval [59,65) 192.168.0.75:6800/1565 172.16.30.75:6800/1565 172.16.30.75:6801/1565 192.168.0.75:6801/1565 exists,up 63aaa0b8-4e52-4d74-82a8-fbbe7b48c837

osd.9 up in weight 1 up_from 73 up_thru 96 down_at 66 last_clean_interval [59,65) 192.168.0.75:6802/1558 172.16.30.75:6802/1558 172.16.30.75:6803/1558 192.168.0.75:6803/1558 exists,up 6bf3204a-b64c-4808-a782-434a93ac578c

[cephadm@ceph-admin ceph-cluster]$



  除了上述命令来检查osd状态,我们还可以根据OSD在CRUSH MPA中的位置查看osd



1

2

3

4

5

6

7

8

9

10

11

12

13

14

15

16

17

18

19

[cephadm@ceph-admin ceph-cluster]$ ceph osd tree

ID CLASS WEIGHT TYPE NAME STATUS REWEIGHT PRI-AFF

-1 0.87891 root default

-9 0.17578 host ceph-mgr01

6 hdd 0.07809 osd.6 up 1.00000 1.00000

7 hdd 0.09769 osd.7 up 1.00000 1.00000

-3 0.17578 host ceph-mon01

0 hdd 0.07809 osd.0 up 1.00000 1.00000

1 hdd 0.09769 osd.1 up 1.00000 1.00000

-5 0.17578 host ceph-mon02

2 hdd 0.07809 osd.2 up 1.00000 1.00000

3 hdd 0.09769 osd.3 up 1.00000 1.00000

-7 0.17578 host ceph-mon03

4 hdd 0.07809 osd.4 up 1.00000 1.00000

5 hdd 0.09769 osd.5 up 1.00000 1.00000

-11 0.17578 host node01

8 hdd 0.07809 osd.8 up 1.00000 1.00000

9 hdd 0.09769 osd.9 up 1.00000 1.00000

[cephadm@ceph-admin ceph-cluster]$



  提示:从上面的输出信息我们可以看到每台主机上osd编号情况,以及每个OSD的权重;

  检查mon节点状态



1

2

3

4

5

6

7

8

9

10

11

12

[cephadm@ceph-admin ceph-cluster]$ ceph mon stat

e3: 3 mons at {ceph-mon01=192.168.0.71:6789/0,ceph-mon02=192.168.0.72:6789/0,ceph-mon03=192.168.0.73:6789/0}, election epoch 18, leader 0 ceph-mon01, quorum 0,1,2 ceph-mon01,ceph-mon02,ceph-mon03

[cephadm@ceph-admin ceph-cluster]$ ceph mon dump

dumped monmap epoch 3

epoch 3

fsid 7fd4a619-9767-4b46-9cee-78b9dfe88f34

last_changed 2022-09-24 01:56:24.196075

created 2022-09-24 00:36:13.210155

0: 192.168.0.71:6789/0 mon.ceph-mon01

1: 192.168.0.72:6789/0 mon.ceph-mon02

2: 192.168.0.73:6789/0 mon.ceph-mon03

[cephadm@ceph-admin ceph-cluster]$



  提示:上述两条命令都能显示出集群有多少个mon节点,以及对应节点的ip地址和监听端口,以及mon节点编号等信息;ceph mon stat除了能显示有多少mon节点和mon的详细信息外,它还显示领导节点的编号,以及选举次数;

  查看仲裁状态



1

2

3

[cephadm@ceph-admin ceph-cluster]$ ceph quorum_status

{"election_epoch":18,"quorum":[0,1,2],"quorum_names":["ceph-mon01","ceph-mon02","ceph-mon03"],"quorum_leader_name":"ceph-mon01","monmap":{"epoch":3,"fsid":"7fd4a619-9767-4b46-9cee-78b9dfe88f34","modified":"2022-09-24 01:56:24.196075","created":"2022-09-24 00:36:13.210155","features":{"persistent":["kraken","luminous","mimic","osdmap-prune"],"optional":[]},"mons":[{"rank":0,"name":"ceph-mon01","addr":"192.168.0.71:6789/0","public_addr":"192.168.0.71:6789/0"},{"rank":1,"name":"ceph-mon02","addr":"192.168.0.72:6789/0","public_addr":"192.168.0.72:6789/0"},{"rank":2,"name":"ceph-mon03","addr":"192.168.0.73:6789/0","public_addr":"192.168.0.73:6789/0"}]}}

[cephadm@ceph-admin ceph-cluster]$



  使用管理套接字查询集群状态

  Ceph的管理套接字接口常用于查询守护进程,套接字默认保存 于/var/run/ceph目录,此接口的使用不能以远程方式进程,只能在对应节点上使用;

  命令的使用格式:ceph --admin-daemon /var/run/ceph/socket-name 命令;比如获取帮助信息 ceph --admin-daemon /var/run/ceph/socket-name help



1

2

3

4

5

6

7

8

9

10

11

12

13

14

15

16

17

18

19

20

21

22

23

24

25

26

27

28

29

30

31

32

33

34

35

36

37

38

39

40

41

42

43

44

45

46

47

48

49

50

51

52

53

54

55

56

57

58

59

60

61

62

[root@ceph-mon01 ~]# ceph --admin-daemon /var/run/ceph/ceph-osd.0.asok help

{

"calc_objectstore_db_histogram": "Generate key value histogram of kvdb(rocksdb) which used by bluestore",

"compact": "Commpact object store's omap. WARNING: Compaction probably slows your requests",

"config diff": "dump diff of current config and default config",

"config diff get": "dump diff get : dump diff of current and default config setting ",

"config get": "config get : get the config value",

"config help": "get config setting schema and descriptions",

"config set": "config set [ ...]: set a config variable",

"config show": "dump current config settings",

"config unset": "config unset : unset a config variable",

"dump_blacklist": "dump blacklisted clients and times",

"dump_blocked_ops": "show the blocked ops currently in flight",

"dump_historic_ops": "show recent ops",

"dump_historic_ops_by_duration": "show slowest recent ops, sorted by duration",

"dump_historic_slow_ops": "show slowest recent ops",

"dump_mempools": "get mempool stats",

"dump_objectstore_kv_stats": "print statistics of kvdb which used by bluestore",

"dump_op_pq_state": "dump op priority queue state",

"dump_ops_in_flight": "show the ops currently in flight",

"dump_osd_network": "Dump osd heartbeat network ping times",

"dump_pgstate_history": "show recent state history",

"dump_reservations": "show recovery reservations",

"dump_scrubs": "print scheduled scrubs",

"dump_watchers": "show clients which have active watches, and on which objects",

"flush_journal": "flush the journal to permanent store",

"flush_store_cache": "Flush bluestore internal cache",

"get_command_descriptions": "list available commands",

"get_heap_property": "get malloc extension heap property",

"get_latest_osdmap": "force osd to update the latest map from the mon",

"get_mapped_pools": "dump pools whose PG(s) are mapped to this OSD.",

"getomap": "output entire object map",

"git_version": "get git sha1",

"heap": "show heap usage info (available only if compiled with tcmalloc)",

"help": "list available commands",

"injectdataerr": "inject data error to an object",

"injectfull": "Inject a full disk (optional count times)",

"injectmdataerr": "inject metadata error to an object",

"list_devices": "list OSD devices.",

"log dump": "dump recent log entries to log file",

"log flush": "flush log entries to log file",

"log reopen": "reopen log file",

"objecter_requests": "show in-progress osd requests",

"ops": "show the ops currently in flight",

"perf dump": "dump perfcounters value",

"perf histogram dump": "dump perf histogram values",

"perf histogram schema": "dump perf histogram schema",

"perf reset": "perf reset : perf reset all or one perfcounter name",

"perf schema": "dump perfcounters schema",

"rmomapkey": "remove omap key",

"set_heap_property": "update malloc extension heap property",

"set_recovery_delay": "Delay osd recovery by specified seconds",

"setomapheader": "set omap header",

"setomapval": "set omap key",

"smart": "probe OSD devices for SMART data.",

"status": "high-level status of OSD",

"trigger_deep_scrub": "Trigger a scheduled deep scrub ",

"trigger_scrub": "Trigger a scheduled scrub ",

"truncobj": "truncate object to length",

"version": "get ceph version"

}

[root@ceph-mon01 ~]#



  比如获取mon01的版本信息



1

2

3

[root@ceph-mon01 ~]# ceph --admin-daemon /var/run/ceph/ceph-mon.ceph-mon01.asok version

{"version":"13.2.10","release":"mimic","release_type":"stable"}

[root@ceph-mon01 ~]#



  获取osd的状态信息



1

2

3

4

5

6

7

8

9

10

11

[root@ceph-mon01 ~]# ceph --admin-daemon /var/run/ceph/ceph-osd.0.asok status

{

"cluster_fsid": "7fd4a619-9767-4b46-9cee-78b9dfe88f34",

"osd_fsid": "bf3649af-e3f4-41a2-a5ce-8f1a316d344e",

"whoami": 0,

"state": "active",

"oldest_map": 1,

"newest_map": 114,

"num_pgs": 83

}

[root@ceph-mon01 ~]#



  进程的运行时配置

  我们可以使用ceph daemon命令来动态的配置ceph进程,即不停服务动态配置进程;

  比如,获取osd.0的公网地址



1

2

3

4

5

[root@ceph-mon01 ~]# ceph daemon osd.0 config get public_addr

{

"public_addr": "192.168.0.71:0/0"

}

[root@ceph-mon01 ~]#



  获取帮助信息:命令格式:ceph daemon {daemon-type}.{id} help



1

2

3

4

5

6

7

8

9

10

11

12

13

14

15

16

17

18

19

20

21

22

23

24

25

26

27

28

29

30

31

32

33

34

35

36

37

38

39

40

41

42

43

44

45

46

47

48

49

50

51

52

53

54

55

56

57

58

59

60

61

62

[root@ceph-mon01 ~]# ceph daemon osd.1 help

{

"calc_objectstore_db_histogram": "Generate key value histogram of kvdb(rocksdb) which used by bluestore",

"compact": "Commpact object store's omap. WARNING: Compaction probably slows your requests",

"config diff": "dump diff of current config and default config",

"config diff get": "dump diff get : dump diff of current and default config setting ",

"config get": "config get : get the config value",

"config help": "get config setting schema and descriptions",

"config set": "config set [ ...]: set a config variable",

"config show": "dump current config settings",

"config unset": "config unset : unset a config variable",

"dump_blacklist": "dump blacklisted clients and times",

"dump_blocked_ops": "show the blocked ops currently in flight",

"dump_historic_ops": "show recent ops",

"dump_historic_ops_by_duration": "show slowest recent ops, sorted by duration",

"dump_historic_slow_ops": "show slowest recent ops",

"dump_mempools": "get mempool stats",

"dump_objectstore_kv_stats": "print statistics of kvdb which used by bluestore",

"dump_op_pq_state": "dump op priority queue state",

"dump_ops_in_flight": "show the ops currently in flight",

"dump_osd_network": "Dump osd heartbeat network ping times",

"dump_pgstate_history": "show recent state history",

"dump_reservations": "show recovery reservations",

"dump_scrubs": "print scheduled scrubs",

"dump_watchers": "show clients which have active watches, and on which objects",

"flush_journal": "flush the journal to permanent store",

"flush_store_cache": "Flush bluestore internal cache",

"get_command_descriptions": "list available commands",

"get_heap_property": "get malloc extension heap property",

"get_latest_osdmap": "force osd to update the latest map from the mon",

"get_mapped_pools": "dump pools whose PG(s) are mapped to this OSD.",

"getomap": "output entire object map",

"git_version": "get git sha1",

"heap": "show heap usage info (available only if compiled with tcmalloc)",

"help": "list available commands",

"injectdataerr": "inject data error to an object",

"injectfull": "Inject a full disk (optional count times)",

"injectmdataerr": "inject metadata error to an object",

"list_devices": "list OSD devices.",

"log dump": "dump recent log entries to log file",

"log flush": "flush log entries to log file",

"log reopen": "reopen log file",

"objecter_requests": "show in-progress osd requests",

"ops": "show the ops currently in flight",

"perf dump": "dump perfcounters value",

"perf histogram dump": "dump perf histogram values",

"perf histogram schema": "dump perf histogram schema",

"perf reset": "perf reset : perf reset all or one perfcounter name",

"perf schema": "dump perfcounters schema",

"rmomapkey": "remove omap key",

"set_heap_property": "update malloc extension heap property",

"set_recovery_delay": "Delay osd recovery by specified seconds",

"setomapheader": "set omap header",

"setomapval": "set omap key",

"smart": "probe OSD devices for SMART data.",

"status": "high-level status of OSD",

"trigger_deep_scrub": "Trigger a scheduled deep scrub ",

"trigger_scrub": "Trigger a scheduled scrub ",

"truncobj": "truncate object to length",

"version": "get ceph version"

}

[root@ceph-mon01 ~]#



  提示:ceph daemon获取某个进程的信息时,需要在对应主机上用root执行命令;

  动态设置进程参数有两种方式,一种是通过mon向对应进程发送配置,一种是通过admin socket发送配置给进程

  通过mon向对应进程发送配置命令格式: ceph tell {daemon-type}.{daemon id or *} injectargs --{name} {value} [--{name} {value}]



1

2

[cephadm@ceph-admin ceph-cluster]$ ceph tell osd.1 injectargs '--debug-osd 0/5'

[cephadm@ceph-admin ceph-cluster]$



  提示:这种方式可以在集群任意主机上执行;

  通过admin socket的方式发送配置命令格式:ceph daemon {daemon-type}.{id} set {name} {value}



1

2

3

4

5

[root@ceph-mon01 ~]# ceph daemon osd.0 config set debug_osd 0/5

{

"success": ""

}

[root@ceph-mon01 ~]#



  提示:这种方式只能在进程所在主机上执行;

  停止或重启Ceph集群步骤

  停止ceph集群步骤

  1、告知Ceph集群不要将OSD标记为out,命令:ceph osd set noout



1

2

3

[cephadm@ceph-admin ceph-cluster]$ ceph osd set noout

noout is set

[cephadm@ceph-admin ceph-cluster]$



  2、按如下顺序停止守护进程和节点:停止存储客户端--->网关,如rgw--->元数据服务器,MDS---> Ceph OSD---> Ceph Manager--->Ceph Monitor;然后关闭对应主机;

  启动ceph集群步骤

  1、以与停止过程相关的顺序启动节点:Ceph Monitor--->Ceph Manager--->Ceph OSD--->元数据服务器,MDS--->网关,如rgw---->存储客户端;

  2、删除noout标志,命令ceph osd unset noout



1

2

3

[cephadm@ceph-admin ceph-cluster]$ ceph osd unset noout

noout is unset

[cephadm@ceph-admin ceph-cluster]$



  提示:集群重新启动起来,需要将noout标记取消,以免但真正有osd故障时,能够将osd及时下线,避免将对应存取数据的操作调度到对应osd上进行操作而引发的故障;

  ceph是一个对象存储集群,在生产环境中,如有不慎可能导致不可预估的后果,所以停止和启动顺序都非常重要;上述过程主要是尽量减少丢失数据的几率,但不保证一定不丢数据;

  Ceph配置文件ceph.conf说明



1

2

3

4

5

6

7

8

9

10

11

12

[cephadm@ceph-admin ceph-cluster]$ cat /etc/ceph/ceph.conf

[global]

fsid = 7fd4a619-9767-4b46-9cee-78b9dfe88f34

mon_initial_members = ceph-mon01

mon_host = 192.168.0.71

public_network = 192.168.0.0/24

cluster_network = 172.16.30.0/24

auth_cluster_required = cephx

auth_service_required = cephx

auth_client_required = cephx

[cephadm@ceph-admin ceph-cluster]$



  提示:ceph.conf配置文件严格遵守ini配置文件风格的语法和格式;其中井号‘#’和冒号‘;’用于注释;ceph.conf主要有[global]、[osd]、[mon]、[client]这4个配置段组成;其中global配置段适用于全局配置,即各组件的公共配置;【osd】配置段作用范围是集群所有osd都生效的配置;【mon】作用范围是集群所以mon都生效的配置;【client】作用范围是所有客户端,比如rbd、rgw;

  mon和osd的独有配置段

  上面的【osd】和【mon】配置段都是针对所有osd和mon生效,如果我们只想配置单独某一个osd或mon该怎么配置呢?ceph.conf中我们使用[type.ID]来表示某一个osd或mon的配置;比如,我们只想配置osd.0,则我们可以在配置段里写[osd.0]来表示该段配置只针对osd.0生效;同样mon也是同样的逻辑,不同于osd的是,mon的ID不是数字;我们可以使用ceph mon dump来查看mon的ID;

  获取osd的编号

  提示:osd的编号都是数字,从0开始;

  ceph.conf配置段生效优先级

  如果以上公共配置段里的配置和专用配置段的配置重复,则专用配置段覆盖公共配置段里的配置,即专用配置段里的配置生效;配置生效优先级顺序为:【global】小于【osd】、【mon】、【client】;【osd】小于【osd.ID】,【mon】小于【mon.a】;总之配置段作用范围越小,越优先生效;

  ceph配置文件生效优先级

  ceph启动时会按如下顺序查找配置文件

  1、$CEPH_CONF:该环境变量所指定的配置文件;

  2、-c path/path :命令行使用-c选项指定的配置文件路径;

  3、/etc/ceph/ceph.conf:默认配置文件路径

  4、 /.ceph/config:当前用户家目录下.ceph/config文件

  5、./ceph.conf :当前用户所在目录下的ceph.conf文件

  配置文件生效顺序是$CEPH_CONF---->-c path/path ---->/etc/ceph/ceph.conf----> /.ceph/config---->./ceph.conf ;

  Ceph配置文件常用的元参数

  ceph配置文件支持用元参数来替换对应配置信息,比如$cluster就表示当前Ceph集群的名称;$type表示当前服务的类型名称;比如osd、mon;$id表示守护进程的标识符,比如以osd.0来说,它的标识符就是0;$host表示守护进程所在主机的主机名;$name表示当前服务的类型名称和进程标识符的组合;即$name=$type.$id;

原文链接:https://www.cnblogs.com/qiuhom-1874/p/16727820.html

展开阅读全文

页面更新:2024-04-22

标签:集群   状态   分布式   节点   顺序   进程   命令   对象   提示   数据   信息

1 2 3 4 5

上滑加载更多 ↓
推荐阅读:
友情链接:
更多:

本站资料均由网友自行发布提供,仅用于学习交流。如有版权问题,请与我联系,QQ:4156828  

© CopyRight 2008-2024 All Rights Reserved. Powered By bs178.com 闽ICP备11008920号-3
闽公网安备35020302034844号

Top