diff options
Diffstat (limited to 'src/common/options/mgr.yaml.in')
-rw-r--r-- | src/common/options/mgr.yaml.in | 362 |
1 files changed, 362 insertions, 0 deletions
diff --git a/src/common/options/mgr.yaml.in b/src/common/options/mgr.yaml.in new file mode 100644 index 000000000..7d7b68035 --- /dev/null +++ b/src/common/options/mgr.yaml.in @@ -0,0 +1,362 @@ +# -*- mode: YAML -*- +--- + +options: +- name: mgr_data + type: str + level: advanced + desc: Filesystem path to the ceph-mgr data directory, used to contain keyring. + fmt_desc: Path to load daemon data (such as keyring) + default: /var/lib/ceph/mgr/$cluster-$id + services: + - mgr + flags: + - no_mon_update +- name: mgr_pool + type: bool + level: dev + desc: Allow use/creation of .mgr pool. + default: true + services: + - mgr + flags: + - startup +- name: mgr_stats_period + type: int + level: basic + desc: Period in seconds of OSD/MDS stats reports to manager + long_desc: Use this setting to control the granularity of time series data collection + from daemons. Adjust upwards if the manager CPU load is too high, or if you simply + do not require the most up to date performance counter data. + default: 5 + services: + - mgr + - common +- name: mgr_client_bytes + type: size + level: dev + default: 128_M + services: + - mgr +- name: mgr_client_messages + type: uint + level: dev + default: 512 + services: + - mgr +- name: mgr_osd_bytes + type: size + level: dev + default: 512_M + services: + - mgr +- name: mgr_osd_messages + type: uint + level: dev + default: 8_K + services: + - mgr +- name: mgr_mds_bytes + type: size + level: dev + default: 128_M + services: + - mgr +- name: mgr_mds_messages + type: uint + level: dev + default: 128 + services: + - mgr +- name: mgr_mon_bytes + type: size + level: dev + default: 128_M + services: + - mgr +- name: mgr_mon_messages + type: uint + level: dev + default: 128 + services: + - mgr +- name: mgr_service_beacon_grace + type: float + level: advanced + desc: Period in seconds from last beacon to manager dropping state about a monitored + service (RGW, rbd-mirror etc) + default: 1_min + services: + - mgr +- name: mgr_debug_aggressive_pg_num_changes + type: bool + level: dev + desc: Bypass most throttling and safety checks in pg[p]_num controller + default: false + services: + - mgr +- name: mgr_max_pg_num_change + type: int + level: advanced + desc: maximum change in pg_num + default: 128 + services: + - mgr + with_legacy: true +- name: mgr_module_path + type: str + level: advanced + desc: Filesystem path to manager modules. + fmt_desc: Path to load modules from + default: @CEPH_INSTALL_DATADIR@/mgr + services: + - mgr +- name: mgr_standby_modules + type: bool + default: true + level: advanced + desc: Start modules in standby (redirect) mode when mgr is standby + long_desc: By default, the standby modules will answer incoming requests with a + HTTP redirect to the active manager, allowing users to point their browser at any + mgr node and find their way to an active mgr. However, this mode is problematic + when using a load balancer because (1) the redirect locations are usually private + IPs and (2) the load balancer can't identify which mgr is the right one to send + traffic to. If a load balancer is being used, set this to false. +- name: mgr_disabled_modules + type: str + level: advanced + desc: List of manager modules never get loaded + long_desc: A comma delimited list of module names. This list is read by manager + when it starts. By default, manager loads all modules found in specified 'mgr_module_path', + and it starts the enabled ones as instructed. The modules in this list will not + be loaded at all. + default: @mgr_disabled_modules@ + services: + - mgr + see_also: + - mgr_module_path + flags: + - startup +- name: mgr_initial_modules + type: str + level: basic + desc: List of manager modules to enable when the cluster is first started + long_desc: This list of module names is read by the monitor when the cluster is + first started after installation, to populate the list of enabled manager modules. Subsequent + updates are done using the 'mgr module [enable|disable]' commands. List may be + comma or space separated. + default: restful iostat nfs + services: + - mon + - common + flags: + - no_mon_update + - cluster_create +- name: cephadm_path + type: str + level: advanced + desc: Path to cephadm utility + default: /usr/sbin/cephadm + services: + - mgr +- name: mon_delta_reset_interval + type: float + level: advanced + desc: window duration for rate calculations in 'ceph status' + fmt_desc: Seconds of inactivity before we reset the PG delta to 0. We keep + track of the delta of the used space of each pool, so, for + example, it would be easier for us to understand the progress of + recovery or the performance of cache tier. But if there's no + activity reported for a certain pool, we just reset the history of + deltas of that pool. + default: 10 + services: + - mgr + with_legacy: true +- name: mon_stat_smooth_intervals + type: uint + level: advanced + desc: number of PGMaps stats over which we calc the average read/write throughput + of the whole cluster + fmt_desc: Ceph will smooth statistics over the last ``N`` PG maps. + default: 6 + services: + - mgr + min: 1 +- name: mon_pool_quota_warn_threshold + type: int + level: advanced + desc: percent of quota at which to issue warnings + default: 0 + services: + - mgr +- name: mon_pool_quota_crit_threshold + type: int + level: advanced + desc: percent of quota at which to issue errors + default: 0 + services: + - mgr +- name: mon_cache_target_full_warn_ratio + type: float + level: advanced + desc: issue CACHE_POOL_NEAR_FULL health warning when cache pool utilization exceeds + this ratio of usable space + fmt_desc: Position between pool's ``cache_target_full`` and ``target_max_object`` + where we start warning + default: 0.66 + services: + - mgr + flags: + - no_mon_update + - cluster_create + with_legacy: true +- name: mon_pg_check_down_all_threshold + type: float + level: advanced + desc: threshold of down osds after which we check all pgs + fmt_desc: Percentage threshold of ``down`` OSDs above which we check all PGs + for stale ones. + default: 0.5 + services: + - mgr + with_legacy: true +- name: mon_pg_stuck_threshold + type: int + level: advanced + desc: number of seconds after which pgs can be considered stuck inactive, unclean, + etc + long_desc: see doc/control.rst under dump_stuck for more info + fmt_desc: Number of seconds after which PGs can be considered as + being stuck. + default: 1_min + services: + - mgr +- name: mon_pg_warn_min_per_osd + type: uint + level: advanced + desc: minimal number PGs per (in) osd before we warn the admin + fmt_desc: Raise ``HEALTH_WARN`` if the average number + of PGs per ``in`` OSD is under this number. A non-positive number + disables this. + default: 0 + services: + - mgr +- name: mon_pg_warn_max_object_skew + type: float + level: advanced + desc: max skew few average in objects per pg + fmt_desc: Raise ``HEALTH_WARN`` if the average RADOS object count per PG + of any pool is greater than ``mon_pg_warn_max_object_skew`` times + the average RADOS object count per PG of all pools. Zero or a non-positive + number disables this. Note that this option applies to ``ceph-mgr`` daemons. + default: 10 + services: + - mgr +- name: mon_pg_warn_min_objects + type: int + level: advanced + desc: 'do not warn below this object #' + fmt_desc: Do not warn if the total number of RADOS objects in cluster is below + this number + default: 10000 + services: + - mgr +- name: mon_pg_warn_min_pool_objects + type: int + level: advanced + desc: 'do not warn on pools below this object #' + fmt_desc: Do not warn on pools whose RADOS object count is below this number + default: 1000 + services: + - mgr +- name: mon_warn_on_misplaced + type: bool + level: advanced + desc: Issue a health warning if there are misplaced objects + default: false + services: + - mgr + with_legacy: true +- name: mon_warn_on_pool_no_app + type: bool + level: dev + desc: issue POOL_APP_NOT_ENABLED health warning if pool has not application enabled + default: true + services: + - mgr +- name: mon_warn_on_too_few_osds + type: bool + level: advanced + desc: Issue a health warning if there are fewer OSDs than osd_pool_default_size + default: true + services: + - mgr +- name: mon_target_pg_per_osd + type: uint + level: advanced + desc: Automated PG management creates this many PGs per OSD + long_desc: When creating pools, the automated PG management logic will attempt to + reach this target. In some circumstances, it may exceed this target, up to the + ``mon_max_pg_per_osd`` limit. Conversely, a lower number of PGs per OSD may be + created if the cluster is not yet fully utilised + default: 100 + min: 1 +# min pgs per osd for reweight-by-pg command +- name: mon_reweight_min_pgs_per_osd + type: uint + level: advanced + default: 10 + services: + - mgr + with_legacy: true +# min bytes per osd for reweight-by-utilization command +- name: mon_reweight_min_bytes_per_osd + type: size + level: advanced + default: 100_M + services: + - mgr + with_legacy: true +# max osds to change per reweight-by-* command +- name: mon_reweight_max_osds + type: int + level: advanced + default: 4 + services: + - mgr + with_legacy: true +- name: mon_reweight_max_change + type: float + level: advanced + default: 0.05 + services: + - mgr + with_legacy: true +- name: mgr_stats_threshold + type: int + level: advanced + desc: Lowest perfcounter priority collected by mgr + long_desc: Daemons only set perf counter data to the manager daemon if the counter + has a priority higher than this. + default: 5 + min: 0 + max: 11 +- name: mgr_tick_period + type: secs + level: advanced + desc: Period in seconds of beacon messages to monitor + fmt_desc: How many seconds between mgr beacons to monitors, and other + periodic checks. + default: 2 + services: + - mgr + - mon +- name: mon_osd_err_op_age_ratio + type: float + level: advanced + desc: issue REQUEST_STUCK health error if OSD ops are slower than is age (seconds) + default: 128 + services: + - mgr + with_legacy: true |