summaryrefslogtreecommitdiffstats
path: root/qa/tasks/ceph.conf.template
blob: 609a124b82d891753e5f7a40f4969668931e6359 (plain)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
[global]
	chdir = ""
	pid file = /var/run/ceph/$cluster-$name.pid
        auth supported = cephx

	filestore xattr use omap = true

	mon clock drift allowed = 1.000

	osd crush chooseleaf type = 0
        auth debug = true

	ms die on old message = true
	ms die on bug = true

	mon max pg per osd = 10000        # >= luminous
	mon pg warn max object skew = 0

	# disable pg_autoscaler by default for new pools
        osd_pool_default_pg_autoscale_mode = off

	osd pool default size = 2

	mon osd allow primary affinity = true
	mon osd allow pg remap = true
	mon warn on legacy crush tunables = false
	mon warn on crush straw calc version zero = false
	mon warn on no sortbitwise = false
	mon warn on osd down out interval zero = false
	mon warn on too few osds = false
	mon_warn_on_pool_pg_num_not_power_of_two = false
        mon_warn_on_pool_no_redundancy = false
	mon_allow_pool_size_one = true

        osd pool default erasure code profile = "plugin=jerasure technique=reed_sol_van k=2 m=1 ruleset-failure-domain=osd crush-failure-domain=osd"

	osd default data pool replay window = 5

	mon allow pool delete = true

	mon cluster log file level = debug
	debug asserts on shutdown = true
	mon health detail to clog = false

[osd]
        osd journal size = 100

        osd scrub load threshold = 5.0
	osd scrub max interval = 600

	osd recover clone overlap = true
	osd recovery max chunk = 1048576

	osd debug shutdown = true
        osd debug op order = true
        osd debug verify stray on activate = true

	osd open classes on start = true
        osd debug pg log writeout = true

	osd deep scrub update digest min age = 30

	osd map max advance = 10

        journal zero on create = true

	filestore ondisk finisher threads = 3
	filestore apply finisher threads = 3

	bdev debug aio = true
	osd debug misdirected ops = true

[mgr]
	debug ms = 1
	debug mgr = 20
	debug mon = 20
	debug auth = 20
	mon reweight min pgs per osd = 4
	mon reweight min bytes per osd = 10
	mgr/telemetry/nag = false

[mon]
	debug ms = 1
	debug mon = 20
	debug paxos = 20
	debug auth = 20
	mon data avail warn = 5
	mon mgr mkfs grace = 240
	mon reweight min pgs per osd = 4
	mon osd reporter subtree level = osd
	mon osd prime pg temp = true
	mon reweight min bytes per osd = 10

	# rotate auth tickets quickly to exercise renewal paths
	auth mon ticket ttl = 660      # 11m
	auth service ticket ttl = 240  # 4m

	# don't complain about insecure global_id in the test suite
	mon_warn_on_insecure_global_id_reclaim = false
	mon_warn_on_insecure_global_id_reclaim_allowed = false

[client]
	rgw cache enabled = true
	rgw enable ops log = true
	rgw enable usage log = true
	log file = /var/log/ceph/$cluster-$name.$pid.log
	admin socket = /var/run/ceph/$cluster-$name.$pid.asok