blob: a9cce29539aa1f7d8c7a2cdc16e5cbca266f07b9 (
plain)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
|
[global]
chdir = ""
pid file = /var/run/ceph/$cluster-$name.pid
auth supported = cephx
filestore xattr use omap = true
mon clock drift allowed = 1.000
osd crush chooseleaf type = 0
auth debug = true
ms die on old message = true
ms die on bug = true
mon max pg per osd = 10000 # >= luminous
mon pg warn max object skew = 0
# disable pg_autoscaler by default for new pools
osd_pool_default_pg_autoscale_mode = off
osd pool default size = 2
mon osd allow primary affinity = true
mon osd allow pg remap = true
mon warn on legacy crush tunables = false
mon warn on crush straw calc version zero = false
mon warn on no sortbitwise = false
mon warn on osd down out interval zero = false
mon warn on too few osds = false
mon_warn_on_pool_pg_num_not_power_of_two = false
mon_warn_on_pool_no_redundancy = false
mon_allow_pool_size_one = true
osd pool default erasure code profile = "plugin=jerasure technique=reed_sol_van k=2 m=1 crush-failure-domain=osd"
osd default data pool replay window = 5
mon allow pool delete = true
mon cluster log file level = debug
debug asserts on shutdown = true
mon health detail to clog = false
[osd]
osd journal size = 100
osd scrub load threshold = 5.0
osd scrub max interval = 600
osd mclock profile = high_recovery_ops
osd recover clone overlap = true
osd recovery max chunk = 1048576
osd debug shutdown = true
osd debug op order = true
osd debug verify stray on activate = true
osd open classes on start = true
osd debug pg log writeout = true
osd deep scrub update digest min age = 30
osd map max advance = 10
journal zero on create = true
filestore ondisk finisher threads = 3
filestore apply finisher threads = 3
bdev debug aio = true
osd debug misdirected ops = true
[mgr]
debug ms = 1
debug mgr = 20
debug mon = 20
debug auth = 20
mon reweight min pgs per osd = 4
mon reweight min bytes per osd = 10
mgr/telemetry/nag = false
[mon]
debug ms = 1
debug mon = 20
debug paxos = 20
debug auth = 20
mon data avail warn = 5
mon mgr mkfs grace = 240
mon reweight min pgs per osd = 4
mon osd reporter subtree level = osd
mon osd prime pg temp = true
mon reweight min bytes per osd = 10
# rotate auth tickets quickly to exercise renewal paths
auth mon ticket ttl = 660 # 11m
auth service ticket ttl = 240 # 4m
# don't complain about insecure global_id in the test suite
mon_warn_on_insecure_global_id_reclaim = false
mon_warn_on_insecure_global_id_reclaim_allowed = false
# 1m isn't quite enough
mon_down_mkfs_grace = 2m
mon_warn_on_filestore_osds = false
[client]
rgw cache enabled = true
rgw enable ops log = true
rgw enable usage log = true
log file = /var/log/ceph/$cluster-$name.$pid.log
admin socket = /var/run/ceph/$cluster-$name.$pid.asok
|