From 19fcec84d8d7d21e796c7624e521b60d28ee21ed Mon Sep 17 00:00:00 2001 From: Daniel Baumann Date: Sun, 7 Apr 2024 20:45:59 +0200 Subject: Adding upstream version 16.2.11+ds. Signed-off-by: Daniel Baumann --- qa/suites/rados/.qa | 1 + qa/suites/rados/basic/% | 0 qa/suites/rados/basic/.qa | 1 + qa/suites/rados/basic/ceph.yaml | 18 +++ qa/suites/rados/basic/clusters/+ | 0 qa/suites/rados/basic/clusters/.qa | 1 + qa/suites/rados/basic/clusters/fixed-2.yaml | 1 + qa/suites/rados/basic/clusters/openstack.yaml | 4 + qa/suites/rados/basic/mon_election | 1 + qa/suites/rados/basic/msgr | 1 + qa/suites/rados/basic/msgr-failures/.qa | 1 + qa/suites/rados/basic/msgr-failures/few.yaml | 8 ++ qa/suites/rados/basic/msgr-failures/many.yaml | 8 ++ qa/suites/rados/basic/objectstore | 1 + qa/suites/rados/basic/rados.yaml | 1 + qa/suites/rados/basic/supported-random-distro$ | 1 + qa/suites/rados/basic/tasks/.qa | 1 + qa/suites/rados/basic/tasks/libcephsqlite.yaml | 24 ++++ qa/suites/rados/basic/tasks/rados_api_tests.yaml | 25 ++++ qa/suites/rados/basic/tasks/rados_cls_all.yaml | 13 ++ qa/suites/rados/basic/tasks/rados_python.yaml | 15 ++ .../rados/basic/tasks/rados_stress_watch.yaml | 11 ++ qa/suites/rados/basic/tasks/rados_striper.yaml | 7 + .../basic/tasks/rados_workunit_loadgen_big.yaml | 11 ++ .../basic/tasks/rados_workunit_loadgen_mix.yaml | 11 ++ .../tasks/rados_workunit_loadgen_mostlyread.yaml | 11 ++ qa/suites/rados/basic/tasks/readwrite.yaml | 17 +++ qa/suites/rados/basic/tasks/repair_test.yaml | 31 +++++ qa/suites/rados/basic/tasks/scrub_test.yaml | 30 ++++ qa/suites/rados/cephadm | 1 + qa/suites/rados/dashboard/% | 0 qa/suites/rados/dashboard/.qa | 1 + .../dashboard/centos_8.stream_container_tools.yaml | 1 + qa/suites/rados/dashboard/clusters/+ | 0 qa/suites/rados/dashboard/clusters/.qa | 1 + qa/suites/rados/dashboard/clusters/2-node-mgr.yaml | 1 + qa/suites/rados/dashboard/debug/.qa | 1 + qa/suites/rados/dashboard/debug/mgr.yaml | 1 + qa/suites/rados/dashboard/mon_election | 1 + qa/suites/rados/dashboard/random-objectstore$ | 1 + qa/suites/rados/dashboard/tasks/.qa | 1 + qa/suites/rados/dashboard/tasks/dashboard.yaml | 61 ++++++++ qa/suites/rados/mgr/% | 0 qa/suites/rados/mgr/.qa | 1 + qa/suites/rados/mgr/clusters/+ | 0 qa/suites/rados/mgr/clusters/.qa | 1 + qa/suites/rados/mgr/clusters/2-node-mgr.yaml | 1 + qa/suites/rados/mgr/debug/.qa | 1 + qa/suites/rados/mgr/debug/mgr.yaml | 1 + qa/suites/rados/mgr/mgr_ttl_cache/.qa | 1 + qa/suites/rados/mgr/mgr_ttl_cache/disable.yaml | 1 + qa/suites/rados/mgr/mgr_ttl_cache/enable.yaml | 1 + qa/suites/rados/mgr/mon_election | 1 + qa/suites/rados/mgr/objectstore | 1 + qa/suites/rados/mgr/supported-random-distro$ | 1 + qa/suites/rados/mgr/tasks/.qa | 1 + qa/suites/rados/mgr/tasks/crash.yaml | 17 +++ qa/suites/rados/mgr/tasks/failover.yaml | 16 +++ qa/suites/rados/mgr/tasks/insights.yaml | 19 +++ qa/suites/rados/mgr/tasks/module_selftest.yaml | 26 ++++ qa/suites/rados/mgr/tasks/progress.yaml | 25 ++++ qa/suites/rados/mgr/tasks/prometheus.yaml | 16 +++ qa/suites/rados/mgr/tasks/workunits.yaml | 16 +++ qa/suites/rados/monthrash/% | 0 qa/suites/rados/monthrash/.qa | 1 + qa/suites/rados/monthrash/ceph.yaml | 25 ++++ qa/suites/rados/monthrash/clusters/.qa | 1 + qa/suites/rados/monthrash/clusters/3-mons.yaml | 7 + qa/suites/rados/monthrash/clusters/9-mons.yaml | 7 + qa/suites/rados/monthrash/mon_election | 1 + qa/suites/rados/monthrash/msgr | 1 + qa/suites/rados/monthrash/msgr-failures/.qa | 1 + qa/suites/rados/monthrash/msgr-failures/few.yaml | 8 ++ .../rados/monthrash/msgr-failures/mon-delay.yaml | 14 ++ qa/suites/rados/monthrash/objectstore | 1 + qa/suites/rados/monthrash/rados.yaml | 1 + qa/suites/rados/monthrash/supported-random-distro$ | 1 + qa/suites/rados/monthrash/thrashers/.qa | 1 + .../rados/monthrash/thrashers/force-sync-many.yaml | 12 ++ qa/suites/rados/monthrash/thrashers/many.yaml | 16 +++ qa/suites/rados/monthrash/thrashers/one.yaml | 9 ++ qa/suites/rados/monthrash/thrashers/sync-many.yaml | 14 ++ qa/suites/rados/monthrash/thrashers/sync.yaml | 13 ++ qa/suites/rados/monthrash/workloads/.qa | 1 + .../monthrash/workloads/pool-create-delete.yaml | 58 ++++++++ .../rados/monthrash/workloads/rados_5925.yaml | 9 ++ .../rados/monthrash/workloads/rados_api_tests.yaml | 26 ++++ .../workloads/rados_mon_osdmap_prune.yaml | 22 +++ .../monthrash/workloads/rados_mon_workunits.yaml | 17 +++ .../monthrash/workloads/snaps-few-objects.yaml | 13 ++ qa/suites/rados/multimon/% | 0 qa/suites/rados/multimon/.qa | 1 + qa/suites/rados/multimon/clusters/.qa | 1 + qa/suites/rados/multimon/clusters/21.yaml | 8 ++ qa/suites/rados/multimon/clusters/3.yaml | 7 + qa/suites/rados/multimon/clusters/6.yaml | 7 + qa/suites/rados/multimon/clusters/9.yaml | 8 ++ qa/suites/rados/multimon/mon_election | 1 + qa/suites/rados/multimon/msgr | 1 + qa/suites/rados/multimon/msgr-failures/.qa | 1 + qa/suites/rados/multimon/msgr-failures/few.yaml | 8 ++ qa/suites/rados/multimon/msgr-failures/many.yaml | 9 ++ qa/suites/rados/multimon/no_pools.yaml | 5 + qa/suites/rados/multimon/objectstore | 1 + qa/suites/rados/multimon/rados.yaml | 1 + qa/suites/rados/multimon/supported-random-distro$ | 1 + qa/suites/rados/multimon/tasks/.qa | 1 + .../rados/multimon/tasks/mon_clock_no_skews.yaml | 11 ++ .../rados/multimon/tasks/mon_clock_with_skews.yaml | 24 ++++ qa/suites/rados/multimon/tasks/mon_recovery.yaml | 10 ++ qa/suites/rados/objectstore/% | 0 qa/suites/rados/objectstore/.qa | 1 + qa/suites/rados/objectstore/backends/.qa | 1 + .../rados/objectstore/backends/alloc-hint.yaml | 22 +++ .../backends/ceph_objectstore_tool.yaml | 25 ++++ .../rados/objectstore/backends/filejournal.yaml | 13 ++ .../backends/filestore-idempotent-aio-journal.yaml | 14 ++ .../objectstore/backends/filestore-idempotent.yaml | 11 ++ .../rados/objectstore/backends/fusestore.yaml | 9 ++ .../rados/objectstore/backends/keyvaluedb.yaml | 8 ++ .../objectstore/backends/objectcacher-stress.yaml | 14 ++ .../rados/objectstore/backends/objectstore.yaml | 12 ++ .../rados/objectstore/supported-random-distro$ | 1 + qa/suites/rados/perf/% | 0 qa/suites/rados/perf/.qa | 1 + qa/suites/rados/perf/ceph.yaml | 13 ++ qa/suites/rados/perf/mon_election | 1 + qa/suites/rados/perf/objectstore/.qa | 1 + .../bluestore-basic-min-osd-mem-target.yaml | 25 ++++ .../rados/perf/objectstore/bluestore-bitmap.yaml | 43 ++++++ .../rados/perf/objectstore/bluestore-comp.yaml | 23 +++ .../objectstore/bluestore-low-osd-mem-target.yaml | 25 ++++ .../rados/perf/objectstore/bluestore-stupid.yaml | 43 ++++++ qa/suites/rados/perf/openstack.yaml | 4 + .../perf/scheduler/dmclock_1Shard_16Threads.yaml | 7 + .../perf/scheduler/dmclock_default_shards.yaml | 5 + .../rados/perf/scheduler/wpq_default_shards.yaml | 5 + qa/suites/rados/perf/settings/.qa | 1 + qa/suites/rados/perf/settings/optimized.yaml | 74 ++++++++++ qa/suites/rados/perf/ubuntu_latest.yaml | 1 + qa/suites/rados/perf/workloads/.qa | 1 + .../rados/perf/workloads/fio_4K_rand_read.yaml | 24 ++++ qa/suites/rados/perf/workloads/fio_4K_rand_rw.yaml | 24 ++++ .../rados/perf/workloads/fio_4M_rand_read.yaml | 24 ++++ qa/suites/rados/perf/workloads/fio_4M_rand_rw.yaml | 24 ++++ .../rados/perf/workloads/fio_4M_rand_write.yaml | 24 ++++ .../perf/workloads/radosbench_4K_rand_read.yaml | 24 ++++ .../perf/workloads/radosbench_4K_seq_read.yaml | 23 +++ .../perf/workloads/radosbench_4M_rand_read.yaml | 24 ++++ .../perf/workloads/radosbench_4M_seq_read.yaml | 23 +++ .../rados/perf/workloads/radosbench_4M_write.yaml | 23 +++ .../perf/workloads/radosbench_omap_write.yaml | 7 + qa/suites/rados/perf/workloads/sample_fio.yaml | 24 ++++ .../rados/perf/workloads/sample_radosbench.yaml | 23 +++ qa/suites/rados/rest/% | 0 qa/suites/rados/rest/.qa | 1 + qa/suites/rados/rest/mgr-restful.yaml | 30 ++++ qa/suites/rados/rest/supported-random-distro$ | 1 + qa/suites/rados/rook | 1 + qa/suites/rados/singleton-bluestore/% | 0 qa/suites/rados/singleton-bluestore/.qa | 1 + qa/suites/rados/singleton-bluestore/all/.qa | 1 + .../rados/singleton-bluestore/all/cephtool.yaml | 48 +++++++ qa/suites/rados/singleton-bluestore/mon_election | 1 + qa/suites/rados/singleton-bluestore/msgr | 1 + qa/suites/rados/singleton-bluestore/msgr-failures | 1 + .../rados/singleton-bluestore/objectstore/.qa | 1 + .../objectstore/bluestore-bitmap.yaml | 1 + .../objectstore/bluestore-comp-lz4.yaml | 1 + .../objectstore/bluestore-comp-snappy.yaml | 1 + qa/suites/rados/singleton-bluestore/rados.yaml | 1 + .../singleton-bluestore/supported-random-distro$ | 1 + qa/suites/rados/singleton-nomsgr/% | 0 qa/suites/rados/singleton-nomsgr/.qa | 1 + qa/suites/rados/singleton-nomsgr/all/.qa | 1 + .../singleton-nomsgr/all/admin_socket_output.yaml | 27 ++++ qa/suites/rados/singleton-nomsgr/all/balancer.yaml | 14 ++ .../rados/singleton-nomsgr/all/cache-fs-trunc.yaml | 54 ++++++++ .../singleton-nomsgr/all/ceph-kvstore-tool.yaml | 23 +++ .../rados/singleton-nomsgr/all/ceph-post-file.yaml | 12 ++ .../singleton-nomsgr/all/export-after-evict.yaml | 40 ++++++ .../rados/singleton-nomsgr/all/full-tiering.yaml | 40 ++++++ .../singleton-nomsgr/all/health-warnings.yaml | 22 +++ .../all/large-omap-object-warnings.yaml | 30 ++++ .../all/lazy_omap_stats_output.yaml | 18 +++ .../singleton-nomsgr/all/librados_hello_world.yaml | 24 ++++ qa/suites/rados/singleton-nomsgr/all/msgr.yaml | 23 +++ .../all/multi-backfill-reject.yaml | 50 +++++++ .../singleton-nomsgr/all/osd_stale_reads.yaml | 29 ++++ .../rados/singleton-nomsgr/all/pool-access.yaml | 15 ++ .../all/recovery-unfound-found.yaml | 60 ++++++++ .../all/version-number-sanity.yaml | 15 ++ qa/suites/rados/singleton-nomsgr/mon_election | 1 + qa/suites/rados/singleton-nomsgr/rados.yaml | 1 + .../singleton-nomsgr/supported-random-distro$ | 1 + qa/suites/rados/singleton/% | 0 qa/suites/rados/singleton/.qa | 1 + qa/suites/rados/singleton/all/.qa | 1 + qa/suites/rados/singleton/all/admin-socket.yaml | 26 ++++ .../rados/singleton/all/backfill-toofull.yaml | 37 +++++ qa/suites/rados/singleton/all/deduptool.yaml | 28 ++++ .../rados/singleton/all/divergent_priors.yaml | 28 ++++ .../rados/singleton/all/divergent_priors2.yaml | 28 ++++ qa/suites/rados/singleton/all/dump-stuck.yaml | 21 +++ .../rados/singleton/all/ec-inconsistent-hinfo.yaml | 36 +++++ qa/suites/rados/singleton/all/ec-lost-unfound.yaml | 29 ++++ .../singleton/all/erasure-code-nonregression.yaml | 17 +++ .../rados/singleton/all/lost-unfound-delete.yaml | 27 ++++ qa/suites/rados/singleton/all/lost-unfound.yaml | 27 ++++ .../singleton/all/max-pg-per-osd.from-mon.yaml | 29 ++++ .../singleton/all/max-pg-per-osd.from-primary.yaml | 34 +++++ .../singleton/all/max-pg-per-osd.from-replica.yaml | 34 +++++ qa/suites/rados/singleton/all/mon-auth-caps.yaml | 19 +++ .../rados/singleton/all/mon-config-key-caps.yaml | 19 +++ qa/suites/rados/singleton/all/mon-config-keys.yaml | 22 +++ qa/suites/rados/singleton/all/mon-config.yaml | 22 +++ .../all/mon-memory-target-compliance.yaml.disabled | 154 +++++++++++++++++++++ qa/suites/rados/singleton/all/osd-backfill.yaml | 28 ++++ .../singleton/all/osd-recovery-incomplete.yaml | 30 ++++ qa/suites/rados/singleton/all/osd-recovery.yaml | 32 +++++ qa/suites/rados/singleton/all/peer.yaml | 27 ++++ .../singleton/all/pg-autoscaler-progress-off.yaml | 44 ++++++ qa/suites/rados/singleton/all/pg-autoscaler.yaml | 36 +++++ .../singleton/all/pg-removal-interruption.yaml | 36 +++++ qa/suites/rados/singleton/all/radostool.yaml | 28 ++++ qa/suites/rados/singleton/all/random-eio.yaml | 46 ++++++ qa/suites/rados/singleton/all/rebuild-mondb.yaml | 37 +++++ .../rados/singleton/all/recovery-preemption.yaml | 60 ++++++++ .../rados/singleton/all/resolve_stuck_peering.yaml | 19 +++ qa/suites/rados/singleton/all/test-crash.yaml | 20 +++ .../rados/singleton/all/test-noautoscale-flag.yaml | 39 ++++++ .../all/test_envlibrados_for_rocksdb.yaml | 22 +++ .../rados/singleton/all/thrash-backfill-full.yaml | 52 +++++++ qa/suites/rados/singleton/all/thrash-eio.yaml | 49 +++++++ qa/suites/rados/singleton/all/thrash-rados/+ | 0 qa/suites/rados/singleton/all/thrash-rados/.qa | 1 + .../singleton/all/thrash-rados/thrash-rados.yaml | 27 ++++ .../all/thrash-rados/thrashosds-health.yaml | 1 + .../all/thrash_cache_writeback_proxy_none.yaml | 70 ++++++++++ .../singleton/all/watch-notify-same-primary.yaml | 34 +++++ qa/suites/rados/singleton/mon_election | 1 + qa/suites/rados/singleton/msgr | 1 + qa/suites/rados/singleton/msgr-failures/.qa | 1 + qa/suites/rados/singleton/msgr-failures/few.yaml | 9 ++ qa/suites/rados/singleton/msgr-failures/many.yaml | 13 ++ qa/suites/rados/singleton/msgr-failures/none.yaml | 0 qa/suites/rados/singleton/objectstore | 1 + qa/suites/rados/singleton/rados.yaml | 1 + qa/suites/rados/singleton/supported-random-distro$ | 1 + qa/suites/rados/standalone/% | 0 qa/suites/rados/standalone/.qa | 1 + .../rados/standalone/supported-random-distro$ | 1 + qa/suites/rados/standalone/workloads/.qa | 1 + qa/suites/rados/standalone/workloads/crush.yaml | 18 +++ .../rados/standalone/workloads/erasure-code.yaml | 18 +++ qa/suites/rados/standalone/workloads/mgr.yaml | 18 +++ qa/suites/rados/standalone/workloads/misc.yaml | 18 +++ qa/suites/rados/standalone/workloads/mon.yaml | 18 +++ qa/suites/rados/standalone/workloads/osd.yaml | 18 +++ qa/suites/rados/standalone/workloads/scrub.yaml | 18 +++ qa/suites/rados/thrash-erasure-code-big/% | 0 qa/suites/rados/thrash-erasure-code-big/.qa | 1 + qa/suites/rados/thrash-erasure-code-big/ceph.yaml | 1 + qa/suites/rados/thrash-erasure-code-big/cluster/+ | 0 .../rados/thrash-erasure-code-big/cluster/.qa | 1 + .../thrash-erasure-code-big/cluster/12-osds.yaml | 4 + .../thrash-erasure-code-big/cluster/openstack.yaml | 4 + .../rados/thrash-erasure-code-big/mon_election | 1 + .../rados/thrash-erasure-code-big/msgr-failures | 1 + .../rados/thrash-erasure-code-big/objectstore | 1 + qa/suites/rados/thrash-erasure-code-big/rados.yaml | 1 + .../thrash-erasure-code-big/recovery-overrides | 1 + .../supported-random-distro$ | 1 + .../rados/thrash-erasure-code-big/thrashers/.qa | 1 + .../thrash-erasure-code-big/thrashers/careful.yaml | 20 +++ .../thrash-erasure-code-big/thrashers/default.yaml | 21 +++ .../thrashers/fastread.yaml | 20 +++ .../thrash-erasure-code-big/thrashers/mapgap.yaml | 22 +++ .../thrashers/morepggrow.yaml | 16 +++ .../thrash-erasure-code-big/thrashers/pggrow.yaml | 16 +++ .../thrash-erasure-code-big/thrashosds-health.yaml | 1 + .../rados/thrash-erasure-code-big/workloads/.qa | 1 + .../ec-rados-plugin=jerasure-k=4-m=2.yaml | 1 + .../workloads/ec-rados-plugin=lrc-k=4-m=2-l=3.yaml | 1 + qa/suites/rados/thrash-erasure-code-isa/% | 0 qa/suites/rados/thrash-erasure-code-isa/.qa | 1 + qa/suites/rados/thrash-erasure-code-isa/arch/.qa | 1 + .../rados/thrash-erasure-code-isa/arch/x86_64.yaml | 1 + qa/suites/rados/thrash-erasure-code-isa/ceph.yaml | 1 + qa/suites/rados/thrash-erasure-code-isa/clusters | 1 + .../rados/thrash-erasure-code-isa/mon_election | 1 + .../rados/thrash-erasure-code-isa/msgr-failures | 1 + .../rados/thrash-erasure-code-isa/objectstore | 1 + qa/suites/rados/thrash-erasure-code-isa/rados.yaml | 1 + .../thrash-erasure-code-isa/recovery-overrides | 1 + .../supported-random-distro$ | 1 + qa/suites/rados/thrash-erasure-code-isa/thrashers | 1 + .../thrash-erasure-code-isa/thrashosds-health.yaml | 1 + .../rados/thrash-erasure-code-isa/workloads/.qa | 1 + .../workloads/ec-rados-plugin=isa-k=2-m=1.yaml | 1 + qa/suites/rados/thrash-erasure-code-overwrites/% | 0 qa/suites/rados/thrash-erasure-code-overwrites/.qa | 1 + .../bluestore-bitmap.yaml | 1 + .../rados/thrash-erasure-code-overwrites/ceph.yaml | 1 + .../rados/thrash-erasure-code-overwrites/clusters | 1 + .../rados/thrash-erasure-code-overwrites/fast | 1 + .../thrash-erasure-code-overwrites/mon_election | 1 + .../thrash-erasure-code-overwrites/msgr-failures | 1 + .../thrash-erasure-code-overwrites/rados.yaml | 1 + .../recovery-overrides | 1 + .../supported-random-distro$ | 1 + .../rados/thrash-erasure-code-overwrites/thrashers | 1 + .../thrashosds-health.yaml | 1 + .../thrash-erasure-code-overwrites/workloads/.qa | 1 + .../ec-pool-snaps-few-objects-overwrites.yaml | 23 +++ .../ec-small-objects-fast-read-overwrites.yaml | 29 ++++ .../workloads/ec-small-objects-overwrites.yaml | 28 ++++ .../workloads/ec-snaps-few-objects-overwrites.yaml | 22 +++ qa/suites/rados/thrash-erasure-code-shec/% | 0 qa/suites/rados/thrash-erasure-code-shec/.qa | 1 + qa/suites/rados/thrash-erasure-code-shec/ceph.yaml | 1 + .../rados/thrash-erasure-code-shec/clusters/+ | 0 .../rados/thrash-erasure-code-shec/clusters/.qa | 1 + .../thrash-erasure-code-shec/clusters/fixed-4.yaml | 1 + .../clusters/openstack.yaml | 4 + .../rados/thrash-erasure-code-shec/mon_election | 1 + .../rados/thrash-erasure-code-shec/msgr-failures | 1 + .../rados/thrash-erasure-code-shec/objectstore | 1 + .../rados/thrash-erasure-code-shec/rados.yaml | 1 + .../thrash-erasure-code-shec/recovery-overrides | 1 + .../supported-random-distro$ | 1 + .../rados/thrash-erasure-code-shec/thrashers/.qa | 1 + .../thrashers/careful.yaml | 20 +++ .../thrashers/default.yaml | 21 +++ .../thrashosds-health.yaml | 1 + .../rados/thrash-erasure-code-shec/workloads/.qa | 1 + .../ec-rados-plugin=shec-k=4-m=3-c=2.yaml | 1 + qa/suites/rados/thrash-erasure-code/% | 0 qa/suites/rados/thrash-erasure-code/.qa | 1 + qa/suites/rados/thrash-erasure-code/ceph.yaml | 3 + qa/suites/rados/thrash-erasure-code/clusters | 1 + qa/suites/rados/thrash-erasure-code/fast/.qa | 1 + qa/suites/rados/thrash-erasure-code/fast/fast.yaml | 5 + .../rados/thrash-erasure-code/fast/normal.yaml | 0 qa/suites/rados/thrash-erasure-code/mon_election | 1 + qa/suites/rados/thrash-erasure-code/msgr-failures | 1 + qa/suites/rados/thrash-erasure-code/objectstore | 1 + qa/suites/rados/thrash-erasure-code/rados.yaml | 1 + .../rados/thrash-erasure-code/recovery-overrides | 1 + .../thrash-erasure-code/supported-random-distro$ | 1 + qa/suites/rados/thrash-erasure-code/thrashers/.qa | 1 + .../thrash-erasure-code/thrashers/careful.yaml | 19 +++ .../thrash-erasure-code/thrashers/default.yaml | 20 +++ .../thrash-erasure-code/thrashers/fastread.yaml | 20 +++ .../thrashers/minsize_recovery.yaml | 18 +++ .../thrash-erasure-code/thrashers/morepggrow.yaml | 16 +++ .../thrash-erasure-code/thrashers/pggrow.yaml | 16 +++ .../thrash-erasure-code/thrashosds-health.yaml | 1 + qa/suites/rados/thrash-erasure-code/workloads/.qa | 1 + .../workloads/ec-rados-plugin=clay-k=4-m=2.yaml | 1 + .../ec-rados-plugin=jerasure-k=2-m=1.yaml | 1 + .../ec-rados-plugin=jerasure-k=3-m=1.yaml | 1 + .../workloads/ec-radosbench.yaml | 27 ++++ .../workloads/ec-small-objects-balanced.yaml | 21 +++ .../workloads/ec-small-objects-fast-read.yaml | 21 +++ .../workloads/ec-small-objects-many-deletes.yaml | 14 ++ .../workloads/ec-small-objects.yaml | 20 +++ qa/suites/rados/thrash/% | 0 qa/suites/rados/thrash/.qa | 1 + .../rados/thrash/0-size-min-size-overrides/.qa | 1 + .../2-size-2-min-size.yaml | 1 + .../3-size-2-min-size.yaml | 1 + qa/suites/rados/thrash/1-pg-log-overrides/.qa | 1 + .../thrash/1-pg-log-overrides/normal_pg_log.yaml | 0 .../thrash/1-pg-log-overrides/short_pg_log.yaml | 1 + qa/suites/rados/thrash/2-recovery-overrides/$ | 0 qa/suites/rados/thrash/2-recovery-overrides/.qa | 1 + .../rados/thrash/2-recovery-overrides/default.yaml | 0 .../2-recovery-overrides/more-active-recovery.yaml | 1 + .../more-async-partial-recovery.yaml | 6 + .../2-recovery-overrides/more-async-recovery.yaml | 5 + .../more-partial-recovery.yaml | 5 + qa/suites/rados/thrash/3-scrub-overrides/$ | 0 qa/suites/rados/thrash/3-scrub-overrides/.qa | 1 + .../rados/thrash/3-scrub-overrides/default.yaml | 0 .../max-simultaneous-scrubs-2.yaml | 5 + .../max-simultaneous-scrubs-3.yaml | 5 + qa/suites/rados/thrash/backoff/.qa | 1 + qa/suites/rados/thrash/backoff/normal.yaml | 0 qa/suites/rados/thrash/backoff/peering.yaml | 5 + .../rados/thrash/backoff/peering_and_degraded.yaml | 6 + qa/suites/rados/thrash/ceph.yaml | 6 + qa/suites/rados/thrash/clusters/+ | 0 qa/suites/rados/thrash/clusters/.qa | 1 + qa/suites/rados/thrash/clusters/fixed-2.yaml | 1 + qa/suites/rados/thrash/clusters/openstack.yaml | 4 + .../thrash/crc-failures/bad_map_crc_failure.yaml | 7 + qa/suites/rados/thrash/crc-failures/default.yaml | 0 qa/suites/rados/thrash/d-balancer/.qa | 1 + .../rados/thrash/d-balancer/crush-compat.yaml | 6 + qa/suites/rados/thrash/d-balancer/on.yaml | 0 qa/suites/rados/thrash/mon_election | 1 + qa/suites/rados/thrash/msgr | 1 + qa/suites/rados/thrash/msgr-failures/.qa | 1 + .../rados/thrash/msgr-failures/fastclose.yaml | 9 ++ qa/suites/rados/thrash/msgr-failures/few.yaml | 10 ++ .../rados/thrash/msgr-failures/osd-delay.yaml | 12 ++ .../thrash/msgr-failures/osd-dispatch-delay.yaml | 7 + qa/suites/rados/thrash/objectstore | 1 + qa/suites/rados/thrash/rados.yaml | 1 + qa/suites/rados/thrash/supported-random-distro$ | 1 + qa/suites/rados/thrash/thrashers/.qa | 1 + qa/suites/rados/thrash/thrashers/careful.yaml | 26 ++++ qa/suites/rados/thrash/thrashers/default.yaml | 28 ++++ qa/suites/rados/thrash/thrashers/mapgap.yaml | 27 ++++ qa/suites/rados/thrash/thrashers/morepggrow.yaml | 22 +++ qa/suites/rados/thrash/thrashers/none.yaml | 0 qa/suites/rados/thrash/thrashers/pggrow.yaml | 24 ++++ qa/suites/rados/thrash/thrashosds-health.yaml | 1 + qa/suites/rados/thrash/workloads/.qa | 1 + .../workloads/admin_socket_objecter_requests.yaml | 13 ++ .../rados/thrash/workloads/cache-agent-big.yaml | 36 +++++ .../rados/thrash/workloads/cache-agent-small.yaml | 34 +++++ .../workloads/cache-pool-snaps-readproxy.yaml | 39 ++++++ .../rados/thrash/workloads/cache-pool-snaps.yaml | 44 ++++++ .../thrash/workloads/cache-snaps-balanced.yaml | 40 ++++++ qa/suites/rados/thrash/workloads/cache-snaps.yaml | 39 ++++++ qa/suites/rados/thrash/workloads/cache.yaml | 36 +++++ .../thrash/workloads/pool-snaps-few-objects.yaml | 18 +++ .../rados/thrash/workloads/rados_api_tests.yaml | 20 +++ .../workloads/radosbench-high-concurrency.yaml | 49 +++++++ qa/suites/rados/thrash/workloads/radosbench.yaml | 24 ++++ qa/suites/rados/thrash/workloads/redirect.yaml | 15 ++ .../thrash/workloads/redirect_promote_tests.yaml | 14 ++ .../thrash/workloads/redirect_set_object.yaml | 13 ++ .../rados/thrash/workloads/set-chunks-read.yaml | 13 ++ .../thrash/workloads/small-objects-balanced.yaml | 22 +++ .../thrash/workloads/small-objects-localized.yaml | 22 +++ .../rados/thrash/workloads/small-objects.yaml | 21 +++ .../workloads/snaps-few-objects-balanced.yaml | 14 ++ .../workloads/snaps-few-objects-localized.yaml | 14 ++ .../rados/thrash/workloads/snaps-few-objects.yaml | 13 ++ .../thrash/workloads/write_fadvise_dontneed.yaml | 8 ++ qa/suites/rados/upgrade/.qa | 1 + qa/suites/rados/upgrade/nautilus-x-singleton | 1 + qa/suites/rados/valgrind-leaks/% | 0 qa/suites/rados/valgrind-leaks/.qa | 1 + qa/suites/rados/valgrind-leaks/1-start.yaml | 30 ++++ .../rados/valgrind-leaks/2-inject-leak/mon.yaml | 7 + .../rados/valgrind-leaks/2-inject-leak/none.yaml | 0 .../rados/valgrind-leaks/2-inject-leak/osd.yaml | 7 + qa/suites/rados/valgrind-leaks/centos_latest.yaml | 1 + qa/suites/rados/verify/% | 0 qa/suites/rados/verify/.qa | 1 + qa/suites/rados/verify/centos_latest.yaml | 1 + qa/suites/rados/verify/ceph.yaml | 15 ++ qa/suites/rados/verify/clusters/+ | 0 qa/suites/rados/verify/clusters/.qa | 1 + qa/suites/rados/verify/clusters/fixed-2.yaml | 1 + qa/suites/rados/verify/clusters/openstack.yaml | 4 + qa/suites/rados/verify/d-thrash/.qa | 1 + qa/suites/rados/verify/d-thrash/default/+ | 0 qa/suites/rados/verify/d-thrash/default/.qa | 1 + .../rados/verify/d-thrash/default/default.yaml | 11 ++ .../verify/d-thrash/default/thrashosds-health.yaml | 1 + qa/suites/rados/verify/d-thrash/none.yaml | 0 qa/suites/rados/verify/mon_election | 1 + qa/suites/rados/verify/msgr | 1 + qa/suites/rados/verify/msgr-failures/.qa | 1 + qa/suites/rados/verify/msgr-failures/few.yaml | 8 ++ qa/suites/rados/verify/objectstore | 1 + qa/suites/rados/verify/rados.yaml | 1 + qa/suites/rados/verify/tasks/.qa | 1 + qa/suites/rados/verify/tasks/mon_recovery.yaml | 12 ++ qa/suites/rados/verify/tasks/rados_api_tests.yaml | 30 ++++ qa/suites/rados/verify/tasks/rados_cls_all.yaml | 11 ++ qa/suites/rados/verify/validater/.qa | 1 + qa/suites/rados/verify/validater/lockdep.yaml | 5 + qa/suites/rados/verify/validater/valgrind.yaml | 31 +++++ 479 files changed, 5239 insertions(+) create mode 120000 qa/suites/rados/.qa create mode 100644 qa/suites/rados/basic/% create mode 120000 qa/suites/rados/basic/.qa create mode 100644 qa/suites/rados/basic/ceph.yaml create mode 100644 qa/suites/rados/basic/clusters/+ create mode 120000 qa/suites/rados/basic/clusters/.qa create mode 120000 qa/suites/rados/basic/clusters/fixed-2.yaml create mode 100644 qa/suites/rados/basic/clusters/openstack.yaml create mode 120000 qa/suites/rados/basic/mon_election create mode 120000 qa/suites/rados/basic/msgr create mode 120000 qa/suites/rados/basic/msgr-failures/.qa create mode 100644 qa/suites/rados/basic/msgr-failures/few.yaml create mode 100644 qa/suites/rados/basic/msgr-failures/many.yaml create mode 120000 qa/suites/rados/basic/objectstore create mode 120000 qa/suites/rados/basic/rados.yaml create mode 120000 qa/suites/rados/basic/supported-random-distro$ create mode 120000 qa/suites/rados/basic/tasks/.qa create mode 100644 qa/suites/rados/basic/tasks/libcephsqlite.yaml create mode 100644 qa/suites/rados/basic/tasks/rados_api_tests.yaml create mode 100644 qa/suites/rados/basic/tasks/rados_cls_all.yaml create mode 100644 qa/suites/rados/basic/tasks/rados_python.yaml create mode 100644 qa/suites/rados/basic/tasks/rados_stress_watch.yaml create mode 100644 qa/suites/rados/basic/tasks/rados_striper.yaml create mode 100644 qa/suites/rados/basic/tasks/rados_workunit_loadgen_big.yaml create mode 100644 qa/suites/rados/basic/tasks/rados_workunit_loadgen_mix.yaml create mode 100644 qa/suites/rados/basic/tasks/rados_workunit_loadgen_mostlyread.yaml create mode 100644 qa/suites/rados/basic/tasks/readwrite.yaml create mode 100644 qa/suites/rados/basic/tasks/repair_test.yaml create mode 100644 qa/suites/rados/basic/tasks/scrub_test.yaml create mode 120000 qa/suites/rados/cephadm create mode 100644 qa/suites/rados/dashboard/% create mode 120000 qa/suites/rados/dashboard/.qa create mode 120000 qa/suites/rados/dashboard/centos_8.stream_container_tools.yaml create mode 100644 qa/suites/rados/dashboard/clusters/+ create mode 120000 qa/suites/rados/dashboard/clusters/.qa create mode 120000 qa/suites/rados/dashboard/clusters/2-node-mgr.yaml create mode 120000 qa/suites/rados/dashboard/debug/.qa create mode 120000 qa/suites/rados/dashboard/debug/mgr.yaml create mode 120000 qa/suites/rados/dashboard/mon_election create mode 120000 qa/suites/rados/dashboard/random-objectstore$ create mode 120000 qa/suites/rados/dashboard/tasks/.qa create mode 100644 qa/suites/rados/dashboard/tasks/dashboard.yaml create mode 100644 qa/suites/rados/mgr/% create mode 120000 qa/suites/rados/mgr/.qa create mode 100644 qa/suites/rados/mgr/clusters/+ create mode 120000 qa/suites/rados/mgr/clusters/.qa create mode 120000 qa/suites/rados/mgr/clusters/2-node-mgr.yaml create mode 120000 qa/suites/rados/mgr/debug/.qa create mode 120000 qa/suites/rados/mgr/debug/mgr.yaml create mode 120000 qa/suites/rados/mgr/mgr_ttl_cache/.qa create mode 120000 qa/suites/rados/mgr/mgr_ttl_cache/disable.yaml create mode 120000 qa/suites/rados/mgr/mgr_ttl_cache/enable.yaml create mode 120000 qa/suites/rados/mgr/mon_election create mode 120000 qa/suites/rados/mgr/objectstore create mode 120000 qa/suites/rados/mgr/supported-random-distro$ create mode 120000 qa/suites/rados/mgr/tasks/.qa create mode 100644 qa/suites/rados/mgr/tasks/crash.yaml create mode 100644 qa/suites/rados/mgr/tasks/failover.yaml create mode 100644 qa/suites/rados/mgr/tasks/insights.yaml create mode 100644 qa/suites/rados/mgr/tasks/module_selftest.yaml create mode 100644 qa/suites/rados/mgr/tasks/progress.yaml create mode 100644 qa/suites/rados/mgr/tasks/prometheus.yaml create mode 100644 qa/suites/rados/mgr/tasks/workunits.yaml create mode 100644 qa/suites/rados/monthrash/% create mode 120000 qa/suites/rados/monthrash/.qa create mode 100644 qa/suites/rados/monthrash/ceph.yaml create mode 120000 qa/suites/rados/monthrash/clusters/.qa create mode 100644 qa/suites/rados/monthrash/clusters/3-mons.yaml create mode 100644 qa/suites/rados/monthrash/clusters/9-mons.yaml create mode 120000 qa/suites/rados/monthrash/mon_election create mode 120000 qa/suites/rados/monthrash/msgr create mode 120000 qa/suites/rados/monthrash/msgr-failures/.qa create mode 100644 qa/suites/rados/monthrash/msgr-failures/few.yaml create mode 100644 qa/suites/rados/monthrash/msgr-failures/mon-delay.yaml create mode 120000 qa/suites/rados/monthrash/objectstore create mode 120000 qa/suites/rados/monthrash/rados.yaml create mode 120000 qa/suites/rados/monthrash/supported-random-distro$ create mode 120000 qa/suites/rados/monthrash/thrashers/.qa create mode 100644 qa/suites/rados/monthrash/thrashers/force-sync-many.yaml create mode 100644 qa/suites/rados/monthrash/thrashers/many.yaml create mode 100644 qa/suites/rados/monthrash/thrashers/one.yaml create mode 100644 qa/suites/rados/monthrash/thrashers/sync-many.yaml create mode 100644 qa/suites/rados/monthrash/thrashers/sync.yaml create mode 120000 qa/suites/rados/monthrash/workloads/.qa create mode 100644 qa/suites/rados/monthrash/workloads/pool-create-delete.yaml create mode 100644 qa/suites/rados/monthrash/workloads/rados_5925.yaml create mode 100644 qa/suites/rados/monthrash/workloads/rados_api_tests.yaml create mode 100644 qa/suites/rados/monthrash/workloads/rados_mon_osdmap_prune.yaml create mode 100644 qa/suites/rados/monthrash/workloads/rados_mon_workunits.yaml create mode 100644 qa/suites/rados/monthrash/workloads/snaps-few-objects.yaml create mode 100644 qa/suites/rados/multimon/% create mode 120000 qa/suites/rados/multimon/.qa create mode 120000 qa/suites/rados/multimon/clusters/.qa create mode 100644 qa/suites/rados/multimon/clusters/21.yaml create mode 100644 qa/suites/rados/multimon/clusters/3.yaml create mode 100644 qa/suites/rados/multimon/clusters/6.yaml create mode 100644 qa/suites/rados/multimon/clusters/9.yaml create mode 120000 qa/suites/rados/multimon/mon_election create mode 120000 qa/suites/rados/multimon/msgr create mode 120000 qa/suites/rados/multimon/msgr-failures/.qa create mode 100644 qa/suites/rados/multimon/msgr-failures/few.yaml create mode 100644 qa/suites/rados/multimon/msgr-failures/many.yaml create mode 100644 qa/suites/rados/multimon/no_pools.yaml create mode 120000 qa/suites/rados/multimon/objectstore create mode 120000 qa/suites/rados/multimon/rados.yaml create mode 120000 qa/suites/rados/multimon/supported-random-distro$ create mode 120000 qa/suites/rados/multimon/tasks/.qa create mode 100644 qa/suites/rados/multimon/tasks/mon_clock_no_skews.yaml create mode 100644 qa/suites/rados/multimon/tasks/mon_clock_with_skews.yaml create mode 100644 qa/suites/rados/multimon/tasks/mon_recovery.yaml create mode 100644 qa/suites/rados/objectstore/% create mode 120000 qa/suites/rados/objectstore/.qa create mode 120000 qa/suites/rados/objectstore/backends/.qa create mode 100644 qa/suites/rados/objectstore/backends/alloc-hint.yaml create mode 100644 qa/suites/rados/objectstore/backends/ceph_objectstore_tool.yaml create mode 100644 qa/suites/rados/objectstore/backends/filejournal.yaml create mode 100644 qa/suites/rados/objectstore/backends/filestore-idempotent-aio-journal.yaml create mode 100644 qa/suites/rados/objectstore/backends/filestore-idempotent.yaml create mode 100644 qa/suites/rados/objectstore/backends/fusestore.yaml create mode 100644 qa/suites/rados/objectstore/backends/keyvaluedb.yaml create mode 100644 qa/suites/rados/objectstore/backends/objectcacher-stress.yaml create mode 100644 qa/suites/rados/objectstore/backends/objectstore.yaml create mode 120000 qa/suites/rados/objectstore/supported-random-distro$ create mode 100644 qa/suites/rados/perf/% create mode 120000 qa/suites/rados/perf/.qa create mode 100644 qa/suites/rados/perf/ceph.yaml create mode 120000 qa/suites/rados/perf/mon_election create mode 120000 qa/suites/rados/perf/objectstore/.qa create mode 100644 qa/suites/rados/perf/objectstore/bluestore-basic-min-osd-mem-target.yaml create mode 100644 qa/suites/rados/perf/objectstore/bluestore-bitmap.yaml create mode 100644 qa/suites/rados/perf/objectstore/bluestore-comp.yaml create mode 100644 qa/suites/rados/perf/objectstore/bluestore-low-osd-mem-target.yaml create mode 100644 qa/suites/rados/perf/objectstore/bluestore-stupid.yaml create mode 100644 qa/suites/rados/perf/openstack.yaml create mode 100644 qa/suites/rados/perf/scheduler/dmclock_1Shard_16Threads.yaml create mode 100644 qa/suites/rados/perf/scheduler/dmclock_default_shards.yaml create mode 100644 qa/suites/rados/perf/scheduler/wpq_default_shards.yaml create mode 120000 qa/suites/rados/perf/settings/.qa create mode 100644 qa/suites/rados/perf/settings/optimized.yaml create mode 120000 qa/suites/rados/perf/ubuntu_latest.yaml create mode 120000 qa/suites/rados/perf/workloads/.qa create mode 100644 qa/suites/rados/perf/workloads/fio_4K_rand_read.yaml create mode 100644 qa/suites/rados/perf/workloads/fio_4K_rand_rw.yaml create mode 100644 qa/suites/rados/perf/workloads/fio_4M_rand_read.yaml create mode 100644 qa/suites/rados/perf/workloads/fio_4M_rand_rw.yaml create mode 100644 qa/suites/rados/perf/workloads/fio_4M_rand_write.yaml create mode 100644 qa/suites/rados/perf/workloads/radosbench_4K_rand_read.yaml create mode 100644 qa/suites/rados/perf/workloads/radosbench_4K_seq_read.yaml create mode 100644 qa/suites/rados/perf/workloads/radosbench_4M_rand_read.yaml create mode 100644 qa/suites/rados/perf/workloads/radosbench_4M_seq_read.yaml create mode 100644 qa/suites/rados/perf/workloads/radosbench_4M_write.yaml create mode 100644 qa/suites/rados/perf/workloads/radosbench_omap_write.yaml create mode 100644 qa/suites/rados/perf/workloads/sample_fio.yaml create mode 100644 qa/suites/rados/perf/workloads/sample_radosbench.yaml create mode 100644 qa/suites/rados/rest/% create mode 120000 qa/suites/rados/rest/.qa create mode 100644 qa/suites/rados/rest/mgr-restful.yaml create mode 120000 qa/suites/rados/rest/supported-random-distro$ create mode 120000 qa/suites/rados/rook create mode 100644 qa/suites/rados/singleton-bluestore/% create mode 120000 qa/suites/rados/singleton-bluestore/.qa create mode 120000 qa/suites/rados/singleton-bluestore/all/.qa create mode 100644 qa/suites/rados/singleton-bluestore/all/cephtool.yaml create mode 120000 qa/suites/rados/singleton-bluestore/mon_election create mode 120000 qa/suites/rados/singleton-bluestore/msgr create mode 120000 qa/suites/rados/singleton-bluestore/msgr-failures create mode 120000 qa/suites/rados/singleton-bluestore/objectstore/.qa create mode 120000 qa/suites/rados/singleton-bluestore/objectstore/bluestore-bitmap.yaml create mode 120000 qa/suites/rados/singleton-bluestore/objectstore/bluestore-comp-lz4.yaml create mode 120000 qa/suites/rados/singleton-bluestore/objectstore/bluestore-comp-snappy.yaml create mode 120000 qa/suites/rados/singleton-bluestore/rados.yaml create mode 120000 qa/suites/rados/singleton-bluestore/supported-random-distro$ create mode 100644 qa/suites/rados/singleton-nomsgr/% create mode 120000 qa/suites/rados/singleton-nomsgr/.qa create mode 120000 qa/suites/rados/singleton-nomsgr/all/.qa create mode 100644 qa/suites/rados/singleton-nomsgr/all/admin_socket_output.yaml create mode 100644 qa/suites/rados/singleton-nomsgr/all/balancer.yaml create mode 100644 qa/suites/rados/singleton-nomsgr/all/cache-fs-trunc.yaml create mode 100644 qa/suites/rados/singleton-nomsgr/all/ceph-kvstore-tool.yaml create mode 100644 qa/suites/rados/singleton-nomsgr/all/ceph-post-file.yaml create mode 100644 qa/suites/rados/singleton-nomsgr/all/export-after-evict.yaml create mode 100644 qa/suites/rados/singleton-nomsgr/all/full-tiering.yaml create mode 100644 qa/suites/rados/singleton-nomsgr/all/health-warnings.yaml create mode 100644 qa/suites/rados/singleton-nomsgr/all/large-omap-object-warnings.yaml create mode 100644 qa/suites/rados/singleton-nomsgr/all/lazy_omap_stats_output.yaml create mode 100644 qa/suites/rados/singleton-nomsgr/all/librados_hello_world.yaml create mode 100644 qa/suites/rados/singleton-nomsgr/all/msgr.yaml create mode 100644 qa/suites/rados/singleton-nomsgr/all/multi-backfill-reject.yaml create mode 100644 qa/suites/rados/singleton-nomsgr/all/osd_stale_reads.yaml create mode 100644 qa/suites/rados/singleton-nomsgr/all/pool-access.yaml create mode 100644 qa/suites/rados/singleton-nomsgr/all/recovery-unfound-found.yaml create mode 100644 qa/suites/rados/singleton-nomsgr/all/version-number-sanity.yaml create mode 120000 qa/suites/rados/singleton-nomsgr/mon_election create mode 120000 qa/suites/rados/singleton-nomsgr/rados.yaml create mode 120000 qa/suites/rados/singleton-nomsgr/supported-random-distro$ create mode 100644 qa/suites/rados/singleton/% create mode 120000 qa/suites/rados/singleton/.qa create mode 120000 qa/suites/rados/singleton/all/.qa create mode 100644 qa/suites/rados/singleton/all/admin-socket.yaml create mode 100644 qa/suites/rados/singleton/all/backfill-toofull.yaml create mode 100644 qa/suites/rados/singleton/all/deduptool.yaml create mode 100644 qa/suites/rados/singleton/all/divergent_priors.yaml create mode 100644 qa/suites/rados/singleton/all/divergent_priors2.yaml create mode 100644 qa/suites/rados/singleton/all/dump-stuck.yaml create mode 100644 qa/suites/rados/singleton/all/ec-inconsistent-hinfo.yaml create mode 100644 qa/suites/rados/singleton/all/ec-lost-unfound.yaml create mode 100644 qa/suites/rados/singleton/all/erasure-code-nonregression.yaml create mode 100644 qa/suites/rados/singleton/all/lost-unfound-delete.yaml create mode 100644 qa/suites/rados/singleton/all/lost-unfound.yaml create mode 100644 qa/suites/rados/singleton/all/max-pg-per-osd.from-mon.yaml create mode 100644 qa/suites/rados/singleton/all/max-pg-per-osd.from-primary.yaml create mode 100644 qa/suites/rados/singleton/all/max-pg-per-osd.from-replica.yaml create mode 100644 qa/suites/rados/singleton/all/mon-auth-caps.yaml create mode 100644 qa/suites/rados/singleton/all/mon-config-key-caps.yaml create mode 100644 qa/suites/rados/singleton/all/mon-config-keys.yaml create mode 100644 qa/suites/rados/singleton/all/mon-config.yaml create mode 100644 qa/suites/rados/singleton/all/mon-memory-target-compliance.yaml.disabled create mode 100644 qa/suites/rados/singleton/all/osd-backfill.yaml create mode 100644 qa/suites/rados/singleton/all/osd-recovery-incomplete.yaml create mode 100644 qa/suites/rados/singleton/all/osd-recovery.yaml create mode 100644 qa/suites/rados/singleton/all/peer.yaml create mode 100644 qa/suites/rados/singleton/all/pg-autoscaler-progress-off.yaml create mode 100644 qa/suites/rados/singleton/all/pg-autoscaler.yaml create mode 100644 qa/suites/rados/singleton/all/pg-removal-interruption.yaml create mode 100644 qa/suites/rados/singleton/all/radostool.yaml create mode 100644 qa/suites/rados/singleton/all/random-eio.yaml create mode 100644 qa/suites/rados/singleton/all/rebuild-mondb.yaml create mode 100644 qa/suites/rados/singleton/all/recovery-preemption.yaml create mode 100644 qa/suites/rados/singleton/all/resolve_stuck_peering.yaml create mode 100644 qa/suites/rados/singleton/all/test-crash.yaml create mode 100644 qa/suites/rados/singleton/all/test-noautoscale-flag.yaml create mode 100644 qa/suites/rados/singleton/all/test_envlibrados_for_rocksdb.yaml create mode 100644 qa/suites/rados/singleton/all/thrash-backfill-full.yaml create mode 100644 qa/suites/rados/singleton/all/thrash-eio.yaml create mode 100644 qa/suites/rados/singleton/all/thrash-rados/+ create mode 120000 qa/suites/rados/singleton/all/thrash-rados/.qa create mode 100644 qa/suites/rados/singleton/all/thrash-rados/thrash-rados.yaml create mode 120000 qa/suites/rados/singleton/all/thrash-rados/thrashosds-health.yaml create mode 100644 qa/suites/rados/singleton/all/thrash_cache_writeback_proxy_none.yaml create mode 100644 qa/suites/rados/singleton/all/watch-notify-same-primary.yaml create mode 120000 qa/suites/rados/singleton/mon_election create mode 120000 qa/suites/rados/singleton/msgr create mode 120000 qa/suites/rados/singleton/msgr-failures/.qa create mode 100644 qa/suites/rados/singleton/msgr-failures/few.yaml create mode 100644 qa/suites/rados/singleton/msgr-failures/many.yaml create mode 100644 qa/suites/rados/singleton/msgr-failures/none.yaml create mode 120000 qa/suites/rados/singleton/objectstore create mode 120000 qa/suites/rados/singleton/rados.yaml create mode 120000 qa/suites/rados/singleton/supported-random-distro$ create mode 100644 qa/suites/rados/standalone/% create mode 120000 qa/suites/rados/standalone/.qa create mode 120000 qa/suites/rados/standalone/supported-random-distro$ create mode 120000 qa/suites/rados/standalone/workloads/.qa create mode 100644 qa/suites/rados/standalone/workloads/crush.yaml create mode 100644 qa/suites/rados/standalone/workloads/erasure-code.yaml create mode 100644 qa/suites/rados/standalone/workloads/mgr.yaml create mode 100644 qa/suites/rados/standalone/workloads/misc.yaml create mode 100644 qa/suites/rados/standalone/workloads/mon.yaml create mode 100644 qa/suites/rados/standalone/workloads/osd.yaml create mode 100644 qa/suites/rados/standalone/workloads/scrub.yaml create mode 100644 qa/suites/rados/thrash-erasure-code-big/% create mode 120000 qa/suites/rados/thrash-erasure-code-big/.qa create mode 120000 qa/suites/rados/thrash-erasure-code-big/ceph.yaml create mode 100644 qa/suites/rados/thrash-erasure-code-big/cluster/+ create mode 120000 qa/suites/rados/thrash-erasure-code-big/cluster/.qa create mode 100644 qa/suites/rados/thrash-erasure-code-big/cluster/12-osds.yaml create mode 100644 qa/suites/rados/thrash-erasure-code-big/cluster/openstack.yaml create mode 120000 qa/suites/rados/thrash-erasure-code-big/mon_election create mode 120000 qa/suites/rados/thrash-erasure-code-big/msgr-failures create mode 120000 qa/suites/rados/thrash-erasure-code-big/objectstore create mode 120000 qa/suites/rados/thrash-erasure-code-big/rados.yaml create mode 120000 qa/suites/rados/thrash-erasure-code-big/recovery-overrides create mode 120000 qa/suites/rados/thrash-erasure-code-big/supported-random-distro$ create mode 120000 qa/suites/rados/thrash-erasure-code-big/thrashers/.qa create mode 100644 qa/suites/rados/thrash-erasure-code-big/thrashers/careful.yaml create mode 100644 qa/suites/rados/thrash-erasure-code-big/thrashers/default.yaml create mode 100644 qa/suites/rados/thrash-erasure-code-big/thrashers/fastread.yaml create mode 100644 qa/suites/rados/thrash-erasure-code-big/thrashers/mapgap.yaml create mode 100644 qa/suites/rados/thrash-erasure-code-big/thrashers/morepggrow.yaml create mode 100644 qa/suites/rados/thrash-erasure-code-big/thrashers/pggrow.yaml create mode 120000 qa/suites/rados/thrash-erasure-code-big/thrashosds-health.yaml create mode 120000 qa/suites/rados/thrash-erasure-code-big/workloads/.qa create mode 120000 qa/suites/rados/thrash-erasure-code-big/workloads/ec-rados-plugin=jerasure-k=4-m=2.yaml create mode 120000 qa/suites/rados/thrash-erasure-code-big/workloads/ec-rados-plugin=lrc-k=4-m=2-l=3.yaml create mode 100644 qa/suites/rados/thrash-erasure-code-isa/% create mode 120000 qa/suites/rados/thrash-erasure-code-isa/.qa create mode 120000 qa/suites/rados/thrash-erasure-code-isa/arch/.qa create mode 100644 qa/suites/rados/thrash-erasure-code-isa/arch/x86_64.yaml create mode 120000 qa/suites/rados/thrash-erasure-code-isa/ceph.yaml create mode 120000 qa/suites/rados/thrash-erasure-code-isa/clusters create mode 120000 qa/suites/rados/thrash-erasure-code-isa/mon_election create mode 120000 qa/suites/rados/thrash-erasure-code-isa/msgr-failures create mode 120000 qa/suites/rados/thrash-erasure-code-isa/objectstore create mode 120000 qa/suites/rados/thrash-erasure-code-isa/rados.yaml create mode 120000 qa/suites/rados/thrash-erasure-code-isa/recovery-overrides create mode 120000 qa/suites/rados/thrash-erasure-code-isa/supported-random-distro$ create mode 120000 qa/suites/rados/thrash-erasure-code-isa/thrashers create mode 120000 qa/suites/rados/thrash-erasure-code-isa/thrashosds-health.yaml create mode 120000 qa/suites/rados/thrash-erasure-code-isa/workloads/.qa create mode 120000 qa/suites/rados/thrash-erasure-code-isa/workloads/ec-rados-plugin=isa-k=2-m=1.yaml create mode 100644 qa/suites/rados/thrash-erasure-code-overwrites/% create mode 120000 qa/suites/rados/thrash-erasure-code-overwrites/.qa create mode 120000 qa/suites/rados/thrash-erasure-code-overwrites/bluestore-bitmap.yaml create mode 120000 qa/suites/rados/thrash-erasure-code-overwrites/ceph.yaml create mode 120000 qa/suites/rados/thrash-erasure-code-overwrites/clusters create mode 120000 qa/suites/rados/thrash-erasure-code-overwrites/fast create mode 120000 qa/suites/rados/thrash-erasure-code-overwrites/mon_election create mode 120000 qa/suites/rados/thrash-erasure-code-overwrites/msgr-failures create mode 120000 qa/suites/rados/thrash-erasure-code-overwrites/rados.yaml create mode 120000 qa/suites/rados/thrash-erasure-code-overwrites/recovery-overrides create mode 120000 qa/suites/rados/thrash-erasure-code-overwrites/supported-random-distro$ create mode 120000 qa/suites/rados/thrash-erasure-code-overwrites/thrashers create mode 120000 qa/suites/rados/thrash-erasure-code-overwrites/thrashosds-health.yaml create mode 120000 qa/suites/rados/thrash-erasure-code-overwrites/workloads/.qa create mode 100644 qa/suites/rados/thrash-erasure-code-overwrites/workloads/ec-pool-snaps-few-objects-overwrites.yaml create mode 100644 qa/suites/rados/thrash-erasure-code-overwrites/workloads/ec-small-objects-fast-read-overwrites.yaml create mode 100644 qa/suites/rados/thrash-erasure-code-overwrites/workloads/ec-small-objects-overwrites.yaml create mode 100644 qa/suites/rados/thrash-erasure-code-overwrites/workloads/ec-snaps-few-objects-overwrites.yaml create mode 100644 qa/suites/rados/thrash-erasure-code-shec/% create mode 120000 qa/suites/rados/thrash-erasure-code-shec/.qa create mode 120000 qa/suites/rados/thrash-erasure-code-shec/ceph.yaml create mode 100644 qa/suites/rados/thrash-erasure-code-shec/clusters/+ create mode 120000 qa/suites/rados/thrash-erasure-code-shec/clusters/.qa create mode 120000 qa/suites/rados/thrash-erasure-code-shec/clusters/fixed-4.yaml create mode 100644 qa/suites/rados/thrash-erasure-code-shec/clusters/openstack.yaml create mode 120000 qa/suites/rados/thrash-erasure-code-shec/mon_election create mode 120000 qa/suites/rados/thrash-erasure-code-shec/msgr-failures create mode 120000 qa/suites/rados/thrash-erasure-code-shec/objectstore create mode 120000 qa/suites/rados/thrash-erasure-code-shec/rados.yaml create mode 120000 qa/suites/rados/thrash-erasure-code-shec/recovery-overrides create mode 120000 qa/suites/rados/thrash-erasure-code-shec/supported-random-distro$ create mode 120000 qa/suites/rados/thrash-erasure-code-shec/thrashers/.qa create mode 100644 qa/suites/rados/thrash-erasure-code-shec/thrashers/careful.yaml create mode 100644 qa/suites/rados/thrash-erasure-code-shec/thrashers/default.yaml create mode 120000 qa/suites/rados/thrash-erasure-code-shec/thrashosds-health.yaml create mode 120000 qa/suites/rados/thrash-erasure-code-shec/workloads/.qa create mode 120000 qa/suites/rados/thrash-erasure-code-shec/workloads/ec-rados-plugin=shec-k=4-m=3-c=2.yaml create mode 100644 qa/suites/rados/thrash-erasure-code/% create mode 120000 qa/suites/rados/thrash-erasure-code/.qa create mode 100644 qa/suites/rados/thrash-erasure-code/ceph.yaml create mode 120000 qa/suites/rados/thrash-erasure-code/clusters create mode 120000 qa/suites/rados/thrash-erasure-code/fast/.qa create mode 100644 qa/suites/rados/thrash-erasure-code/fast/fast.yaml create mode 100644 qa/suites/rados/thrash-erasure-code/fast/normal.yaml create mode 120000 qa/suites/rados/thrash-erasure-code/mon_election create mode 120000 qa/suites/rados/thrash-erasure-code/msgr-failures create mode 120000 qa/suites/rados/thrash-erasure-code/objectstore create mode 120000 qa/suites/rados/thrash-erasure-code/rados.yaml create mode 120000 qa/suites/rados/thrash-erasure-code/recovery-overrides create mode 120000 qa/suites/rados/thrash-erasure-code/supported-random-distro$ create mode 120000 qa/suites/rados/thrash-erasure-code/thrashers/.qa create mode 100644 qa/suites/rados/thrash-erasure-code/thrashers/careful.yaml create mode 100644 qa/suites/rados/thrash-erasure-code/thrashers/default.yaml create mode 100644 qa/suites/rados/thrash-erasure-code/thrashers/fastread.yaml create mode 100644 qa/suites/rados/thrash-erasure-code/thrashers/minsize_recovery.yaml create mode 100644 qa/suites/rados/thrash-erasure-code/thrashers/morepggrow.yaml create mode 100644 qa/suites/rados/thrash-erasure-code/thrashers/pggrow.yaml create mode 120000 qa/suites/rados/thrash-erasure-code/thrashosds-health.yaml create mode 120000 qa/suites/rados/thrash-erasure-code/workloads/.qa create mode 120000 qa/suites/rados/thrash-erasure-code/workloads/ec-rados-plugin=clay-k=4-m=2.yaml create mode 120000 qa/suites/rados/thrash-erasure-code/workloads/ec-rados-plugin=jerasure-k=2-m=1.yaml create mode 120000 qa/suites/rados/thrash-erasure-code/workloads/ec-rados-plugin=jerasure-k=3-m=1.yaml create mode 100644 qa/suites/rados/thrash-erasure-code/workloads/ec-radosbench.yaml create mode 100644 qa/suites/rados/thrash-erasure-code/workloads/ec-small-objects-balanced.yaml create mode 100644 qa/suites/rados/thrash-erasure-code/workloads/ec-small-objects-fast-read.yaml create mode 100644 qa/suites/rados/thrash-erasure-code/workloads/ec-small-objects-many-deletes.yaml create mode 100644 qa/suites/rados/thrash-erasure-code/workloads/ec-small-objects.yaml create mode 100644 qa/suites/rados/thrash/% create mode 120000 qa/suites/rados/thrash/.qa create mode 120000 qa/suites/rados/thrash/0-size-min-size-overrides/.qa create mode 120000 qa/suites/rados/thrash/0-size-min-size-overrides/2-size-2-min-size.yaml create mode 120000 qa/suites/rados/thrash/0-size-min-size-overrides/3-size-2-min-size.yaml create mode 120000 qa/suites/rados/thrash/1-pg-log-overrides/.qa create mode 100644 qa/suites/rados/thrash/1-pg-log-overrides/normal_pg_log.yaml create mode 120000 qa/suites/rados/thrash/1-pg-log-overrides/short_pg_log.yaml create mode 100644 qa/suites/rados/thrash/2-recovery-overrides/$ create mode 120000 qa/suites/rados/thrash/2-recovery-overrides/.qa create mode 100644 qa/suites/rados/thrash/2-recovery-overrides/default.yaml create mode 120000 qa/suites/rados/thrash/2-recovery-overrides/more-active-recovery.yaml create mode 100644 qa/suites/rados/thrash/2-recovery-overrides/more-async-partial-recovery.yaml create mode 100644 qa/suites/rados/thrash/2-recovery-overrides/more-async-recovery.yaml create mode 100644 qa/suites/rados/thrash/2-recovery-overrides/more-partial-recovery.yaml create mode 100644 qa/suites/rados/thrash/3-scrub-overrides/$ create mode 120000 qa/suites/rados/thrash/3-scrub-overrides/.qa create mode 100644 qa/suites/rados/thrash/3-scrub-overrides/default.yaml create mode 100644 qa/suites/rados/thrash/3-scrub-overrides/max-simultaneous-scrubs-2.yaml create mode 100644 qa/suites/rados/thrash/3-scrub-overrides/max-simultaneous-scrubs-3.yaml create mode 120000 qa/suites/rados/thrash/backoff/.qa create mode 100644 qa/suites/rados/thrash/backoff/normal.yaml create mode 100644 qa/suites/rados/thrash/backoff/peering.yaml create mode 100644 qa/suites/rados/thrash/backoff/peering_and_degraded.yaml create mode 100644 qa/suites/rados/thrash/ceph.yaml create mode 100644 qa/suites/rados/thrash/clusters/+ create mode 120000 qa/suites/rados/thrash/clusters/.qa create mode 120000 qa/suites/rados/thrash/clusters/fixed-2.yaml create mode 100644 qa/suites/rados/thrash/clusters/openstack.yaml create mode 100644 qa/suites/rados/thrash/crc-failures/bad_map_crc_failure.yaml create mode 100644 qa/suites/rados/thrash/crc-failures/default.yaml create mode 120000 qa/suites/rados/thrash/d-balancer/.qa create mode 100644 qa/suites/rados/thrash/d-balancer/crush-compat.yaml create mode 100644 qa/suites/rados/thrash/d-balancer/on.yaml create mode 120000 qa/suites/rados/thrash/mon_election create mode 120000 qa/suites/rados/thrash/msgr create mode 120000 qa/suites/rados/thrash/msgr-failures/.qa create mode 100644 qa/suites/rados/thrash/msgr-failures/fastclose.yaml create mode 100644 qa/suites/rados/thrash/msgr-failures/few.yaml create mode 100644 qa/suites/rados/thrash/msgr-failures/osd-delay.yaml create mode 100644 qa/suites/rados/thrash/msgr-failures/osd-dispatch-delay.yaml create mode 120000 qa/suites/rados/thrash/objectstore create mode 120000 qa/suites/rados/thrash/rados.yaml create mode 120000 qa/suites/rados/thrash/supported-random-distro$ create mode 120000 qa/suites/rados/thrash/thrashers/.qa create mode 100644 qa/suites/rados/thrash/thrashers/careful.yaml create mode 100644 qa/suites/rados/thrash/thrashers/default.yaml create mode 100644 qa/suites/rados/thrash/thrashers/mapgap.yaml create mode 100644 qa/suites/rados/thrash/thrashers/morepggrow.yaml create mode 100644 qa/suites/rados/thrash/thrashers/none.yaml create mode 100644 qa/suites/rados/thrash/thrashers/pggrow.yaml create mode 120000 qa/suites/rados/thrash/thrashosds-health.yaml create mode 120000 qa/suites/rados/thrash/workloads/.qa create mode 100644 qa/suites/rados/thrash/workloads/admin_socket_objecter_requests.yaml create mode 100644 qa/suites/rados/thrash/workloads/cache-agent-big.yaml create mode 100644 qa/suites/rados/thrash/workloads/cache-agent-small.yaml create mode 100644 qa/suites/rados/thrash/workloads/cache-pool-snaps-readproxy.yaml create mode 100644 qa/suites/rados/thrash/workloads/cache-pool-snaps.yaml create mode 100644 qa/suites/rados/thrash/workloads/cache-snaps-balanced.yaml create mode 100644 qa/suites/rados/thrash/workloads/cache-snaps.yaml create mode 100644 qa/suites/rados/thrash/workloads/cache.yaml create mode 100644 qa/suites/rados/thrash/workloads/pool-snaps-few-objects.yaml create mode 100644 qa/suites/rados/thrash/workloads/rados_api_tests.yaml create mode 100644 qa/suites/rados/thrash/workloads/radosbench-high-concurrency.yaml create mode 100644 qa/suites/rados/thrash/workloads/radosbench.yaml create mode 100644 qa/suites/rados/thrash/workloads/redirect.yaml create mode 100644 qa/suites/rados/thrash/workloads/redirect_promote_tests.yaml create mode 100644 qa/suites/rados/thrash/workloads/redirect_set_object.yaml create mode 100644 qa/suites/rados/thrash/workloads/set-chunks-read.yaml create mode 100644 qa/suites/rados/thrash/workloads/small-objects-balanced.yaml create mode 100644 qa/suites/rados/thrash/workloads/small-objects-localized.yaml create mode 100644 qa/suites/rados/thrash/workloads/small-objects.yaml create mode 100644 qa/suites/rados/thrash/workloads/snaps-few-objects-balanced.yaml create mode 100644 qa/suites/rados/thrash/workloads/snaps-few-objects-localized.yaml create mode 100644 qa/suites/rados/thrash/workloads/snaps-few-objects.yaml create mode 100644 qa/suites/rados/thrash/workloads/write_fadvise_dontneed.yaml create mode 120000 qa/suites/rados/upgrade/.qa create mode 120000 qa/suites/rados/upgrade/nautilus-x-singleton create mode 100644 qa/suites/rados/valgrind-leaks/% create mode 120000 qa/suites/rados/valgrind-leaks/.qa create mode 100644 qa/suites/rados/valgrind-leaks/1-start.yaml create mode 100644 qa/suites/rados/valgrind-leaks/2-inject-leak/mon.yaml create mode 100644 qa/suites/rados/valgrind-leaks/2-inject-leak/none.yaml create mode 100644 qa/suites/rados/valgrind-leaks/2-inject-leak/osd.yaml create mode 120000 qa/suites/rados/valgrind-leaks/centos_latest.yaml create mode 100644 qa/suites/rados/verify/% create mode 120000 qa/suites/rados/verify/.qa create mode 120000 qa/suites/rados/verify/centos_latest.yaml create mode 100644 qa/suites/rados/verify/ceph.yaml create mode 100644 qa/suites/rados/verify/clusters/+ create mode 120000 qa/suites/rados/verify/clusters/.qa create mode 120000 qa/suites/rados/verify/clusters/fixed-2.yaml create mode 100644 qa/suites/rados/verify/clusters/openstack.yaml create mode 120000 qa/suites/rados/verify/d-thrash/.qa create mode 100644 qa/suites/rados/verify/d-thrash/default/+ create mode 120000 qa/suites/rados/verify/d-thrash/default/.qa create mode 100644 qa/suites/rados/verify/d-thrash/default/default.yaml create mode 120000 qa/suites/rados/verify/d-thrash/default/thrashosds-health.yaml create mode 100644 qa/suites/rados/verify/d-thrash/none.yaml create mode 120000 qa/suites/rados/verify/mon_election create mode 120000 qa/suites/rados/verify/msgr create mode 120000 qa/suites/rados/verify/msgr-failures/.qa create mode 100644 qa/suites/rados/verify/msgr-failures/few.yaml create mode 120000 qa/suites/rados/verify/objectstore create mode 120000 qa/suites/rados/verify/rados.yaml create mode 120000 qa/suites/rados/verify/tasks/.qa create mode 100644 qa/suites/rados/verify/tasks/mon_recovery.yaml create mode 100644 qa/suites/rados/verify/tasks/rados_api_tests.yaml create mode 100644 qa/suites/rados/verify/tasks/rados_cls_all.yaml create mode 120000 qa/suites/rados/verify/validater/.qa create mode 100644 qa/suites/rados/verify/validater/lockdep.yaml create mode 100644 qa/suites/rados/verify/validater/valgrind.yaml (limited to 'qa/suites/rados') diff --git a/qa/suites/rados/.qa b/qa/suites/rados/.qa new file mode 120000 index 000000000..a602a0353 --- /dev/null +++ b/qa/suites/rados/.qa @@ -0,0 +1 @@ +../.qa/ \ No newline at end of file diff --git a/qa/suites/rados/basic/% b/qa/suites/rados/basic/% new file mode 100644 index 000000000..e69de29bb diff --git a/qa/suites/rados/basic/.qa b/qa/suites/rados/basic/.qa new file mode 120000 index 000000000..a602a0353 --- /dev/null +++ b/qa/suites/rados/basic/.qa @@ -0,0 +1 @@ +../.qa/ \ No newline at end of file diff --git a/qa/suites/rados/basic/ceph.yaml b/qa/suites/rados/basic/ceph.yaml new file mode 100644 index 000000000..c12a671f0 --- /dev/null +++ b/qa/suites/rados/basic/ceph.yaml @@ -0,0 +1,18 @@ +overrides: + ceph: + conf: + mon: + mon min osdmap epochs: 50 + paxos service trim min: 10 + # prune full osdmaps regularly + mon osdmap full prune min: 15 + mon osdmap full prune interval: 2 + mon osdmap full prune txsize: 2 +tasks: +- install: + extra_system_packages: + rpm: + - sqlite-devel + deb: + - sqlite3 +- ceph: diff --git a/qa/suites/rados/basic/clusters/+ b/qa/suites/rados/basic/clusters/+ new file mode 100644 index 000000000..e69de29bb diff --git a/qa/suites/rados/basic/clusters/.qa b/qa/suites/rados/basic/clusters/.qa new file mode 120000 index 000000000..a602a0353 --- /dev/null +++ b/qa/suites/rados/basic/clusters/.qa @@ -0,0 +1 @@ +../.qa/ \ No newline at end of file diff --git a/qa/suites/rados/basic/clusters/fixed-2.yaml b/qa/suites/rados/basic/clusters/fixed-2.yaml new file mode 120000 index 000000000..230ff0fda --- /dev/null +++ b/qa/suites/rados/basic/clusters/fixed-2.yaml @@ -0,0 +1 @@ +.qa/clusters/fixed-2.yaml \ No newline at end of file diff --git a/qa/suites/rados/basic/clusters/openstack.yaml b/qa/suites/rados/basic/clusters/openstack.yaml new file mode 100644 index 000000000..e559d9126 --- /dev/null +++ b/qa/suites/rados/basic/clusters/openstack.yaml @@ -0,0 +1,4 @@ +openstack: + - volumes: # attached to each instance + count: 4 + size: 10 # GB diff --git a/qa/suites/rados/basic/mon_election b/qa/suites/rados/basic/mon_election new file mode 120000 index 000000000..3f331e621 --- /dev/null +++ b/qa/suites/rados/basic/mon_election @@ -0,0 +1 @@ +.qa/mon_election \ No newline at end of file diff --git a/qa/suites/rados/basic/msgr b/qa/suites/rados/basic/msgr new file mode 120000 index 000000000..57bee80db --- /dev/null +++ b/qa/suites/rados/basic/msgr @@ -0,0 +1 @@ +.qa/msgr \ No newline at end of file diff --git a/qa/suites/rados/basic/msgr-failures/.qa b/qa/suites/rados/basic/msgr-failures/.qa new file mode 120000 index 000000000..a602a0353 --- /dev/null +++ b/qa/suites/rados/basic/msgr-failures/.qa @@ -0,0 +1 @@ +../.qa/ \ No newline at end of file diff --git a/qa/suites/rados/basic/msgr-failures/few.yaml b/qa/suites/rados/basic/msgr-failures/few.yaml new file mode 100644 index 000000000..519288992 --- /dev/null +++ b/qa/suites/rados/basic/msgr-failures/few.yaml @@ -0,0 +1,8 @@ +overrides: + ceph: + conf: + global: + ms inject socket failures: 5000 + mon client directed command retry: 5 + log-ignorelist: + - \(OSD_SLOW_PING_TIME diff --git a/qa/suites/rados/basic/msgr-failures/many.yaml b/qa/suites/rados/basic/msgr-failures/many.yaml new file mode 100644 index 000000000..075d959a7 --- /dev/null +++ b/qa/suites/rados/basic/msgr-failures/many.yaml @@ -0,0 +1,8 @@ +overrides: + ceph: + conf: + global: + ms inject socket failures: 1500 + mon client directed command retry: 5 + log-ignorelist: + - \(OSD_SLOW_PING_TIME diff --git a/qa/suites/rados/basic/objectstore b/qa/suites/rados/basic/objectstore new file mode 120000 index 000000000..848c65f9e --- /dev/null +++ b/qa/suites/rados/basic/objectstore @@ -0,0 +1 @@ +.qa/objectstore_debug \ No newline at end of file diff --git a/qa/suites/rados/basic/rados.yaml b/qa/suites/rados/basic/rados.yaml new file mode 120000 index 000000000..d256979c0 --- /dev/null +++ b/qa/suites/rados/basic/rados.yaml @@ -0,0 +1 @@ +.qa/config/rados.yaml \ No newline at end of file diff --git a/qa/suites/rados/basic/supported-random-distro$ b/qa/suites/rados/basic/supported-random-distro$ new file mode 120000 index 000000000..0862b4457 --- /dev/null +++ b/qa/suites/rados/basic/supported-random-distro$ @@ -0,0 +1 @@ +.qa/distros/supported-random-distro$ \ No newline at end of file diff --git a/qa/suites/rados/basic/tasks/.qa b/qa/suites/rados/basic/tasks/.qa new file mode 120000 index 000000000..a602a0353 --- /dev/null +++ b/qa/suites/rados/basic/tasks/.qa @@ -0,0 +1 @@ +../.qa/ \ No newline at end of file diff --git a/qa/suites/rados/basic/tasks/libcephsqlite.yaml b/qa/suites/rados/basic/tasks/libcephsqlite.yaml new file mode 100644 index 000000000..12498fb15 --- /dev/null +++ b/qa/suites/rados/basic/tasks/libcephsqlite.yaml @@ -0,0 +1,24 @@ +overrides: + ceph: + conf: + client: + debug ms: 1 + debug client: 20 + debug cephsqlite: 20 + log-ignorelist: + - POOL_APP_NOT_ENABLED + - do not have an application enabled +tasks: +- exec: + client.0: + - ceph osd pool create cephsqlite + - ceph auth get-or-create client.libcephsqlite mon 'profile simple-rados-client-with-blocklist' osd 'allow rwx pool=cephsqlite' >> /etc/ceph/ceph.keyring +- exec: + client.0: + - ceph_test_libcephsqlite --id libcephsqlite --no-log-to-stderr +- workunit: + clients: + client.0: + - rados/test_libcephsqlite.sh cephsqlite + env: + CEPH_ARGS: --id libcephsqlite --no-log-to-stderr diff --git a/qa/suites/rados/basic/tasks/rados_api_tests.yaml b/qa/suites/rados/basic/tasks/rados_api_tests.yaml new file mode 100644 index 000000000..a54e03b1a --- /dev/null +++ b/qa/suites/rados/basic/tasks/rados_api_tests.yaml @@ -0,0 +1,25 @@ +overrides: + ceph: + log-ignorelist: + - reached quota + - but it is still running + - overall HEALTH_ + - \(POOL_FULL\) + - \(SMALLER_PGP_NUM\) + - \(CACHE_POOL_NO_HIT_SET\) + - \(CACHE_POOL_NEAR_FULL\) + - \(POOL_APP_NOT_ENABLED\) + - \(PG_AVAILABILITY\) + - \(PG_DEGRADED\) + conf: + client: + debug ms: 1 + mon: + mon warn on pool no app: false +tasks: +- workunit: + clients: + client.0: + - rados/test.sh + - rados/test_pool_quota.sh + diff --git a/qa/suites/rados/basic/tasks/rados_cls_all.yaml b/qa/suites/rados/basic/tasks/rados_cls_all.yaml new file mode 100644 index 000000000..8f5b79027 --- /dev/null +++ b/qa/suites/rados/basic/tasks/rados_cls_all.yaml @@ -0,0 +1,13 @@ +overrides: + ceph: + log-ignorelist: + - \(PG_AVAILABILITY\) + conf: + osd: + osd_class_load_list: "*" + osd_class_default_list: "*" +tasks: +- workunit: + clients: + client.0: + - cls diff --git a/qa/suites/rados/basic/tasks/rados_python.yaml b/qa/suites/rados/basic/tasks/rados_python.yaml new file mode 100644 index 000000000..965909450 --- /dev/null +++ b/qa/suites/rados/basic/tasks/rados_python.yaml @@ -0,0 +1,15 @@ +overrides: + ceph: + log-ignorelist: + - but it is still running + - overall HEALTH_ + - \(OSDMAP_FLAGS\) + - \(PG_ + - \(OSD_ + - \(OBJECT_ + - \(POOL_APP_NOT_ENABLED\) +tasks: +- workunit: + clients: + client.0: + - rados/test_python.sh diff --git a/qa/suites/rados/basic/tasks/rados_stress_watch.yaml b/qa/suites/rados/basic/tasks/rados_stress_watch.yaml new file mode 100644 index 000000000..7b5c89b9b --- /dev/null +++ b/qa/suites/rados/basic/tasks/rados_stress_watch.yaml @@ -0,0 +1,11 @@ +overrides: + ceph: + log-ignorelist: + - overall HEALTH_ + - \(CACHE_POOL_NO_HIT_SET\) + - \(TOO_FEW_PGS\) +tasks: +- workunit: + clients: + client.0: + - rados/stress_watch.sh diff --git a/qa/suites/rados/basic/tasks/rados_striper.yaml b/qa/suites/rados/basic/tasks/rados_striper.yaml new file mode 100644 index 000000000..c19cc83a9 --- /dev/null +++ b/qa/suites/rados/basic/tasks/rados_striper.yaml @@ -0,0 +1,7 @@ +tasks: +- exec: + client.0: + - ceph_test_rados_striper_api_io + - ceph_test_rados_striper_api_aio + - ceph_test_rados_striper_api_striping + diff --git a/qa/suites/rados/basic/tasks/rados_workunit_loadgen_big.yaml b/qa/suites/rados/basic/tasks/rados_workunit_loadgen_big.yaml new file mode 100644 index 000000000..53effb42d --- /dev/null +++ b/qa/suites/rados/basic/tasks/rados_workunit_loadgen_big.yaml @@ -0,0 +1,11 @@ +overrides: + ceph: + log-ignorelist: + - but it is still running + - overall HEALTH_ + - \(POOL_APP_NOT_ENABLED\) +tasks: +- workunit: + clients: + all: + - rados/load-gen-big.sh diff --git a/qa/suites/rados/basic/tasks/rados_workunit_loadgen_mix.yaml b/qa/suites/rados/basic/tasks/rados_workunit_loadgen_mix.yaml new file mode 100644 index 000000000..847aedb21 --- /dev/null +++ b/qa/suites/rados/basic/tasks/rados_workunit_loadgen_mix.yaml @@ -0,0 +1,11 @@ +overrides: + ceph: + log-ignorelist: + - but it is still running + - overall HEALTH_ + - \(POOL_APP_NOT_ENABLED\) +tasks: +- workunit: + clients: + all: + - rados/load-gen-mix.sh diff --git a/qa/suites/rados/basic/tasks/rados_workunit_loadgen_mostlyread.yaml b/qa/suites/rados/basic/tasks/rados_workunit_loadgen_mostlyread.yaml new file mode 100644 index 000000000..b25392ffa --- /dev/null +++ b/qa/suites/rados/basic/tasks/rados_workunit_loadgen_mostlyread.yaml @@ -0,0 +1,11 @@ +overrides: + ceph: + log-ignorelist: + - but it is still running + - overall HEALTH_ + - \(POOL_APP_NOT_ENABLED\) +tasks: +- workunit: + clients: + all: + - rados/load-gen-mostlyread.sh diff --git a/qa/suites/rados/basic/tasks/readwrite.yaml b/qa/suites/rados/basic/tasks/readwrite.yaml new file mode 100644 index 000000000..f135107c7 --- /dev/null +++ b/qa/suites/rados/basic/tasks/readwrite.yaml @@ -0,0 +1,17 @@ +overrides: + ceph: + crush_tunables: optimal + conf: + mon: + mon osd initial require min compat client: luminous + osd: + osd_discard_disconnected_ops: false +tasks: +- rados: + clients: [client.0] + ops: 4000 + objects: 500 + op_weights: + read: 45 + write: 45 + delete: 10 diff --git a/qa/suites/rados/basic/tasks/repair_test.yaml b/qa/suites/rados/basic/tasks/repair_test.yaml new file mode 100644 index 000000000..0d3749f4d --- /dev/null +++ b/qa/suites/rados/basic/tasks/repair_test.yaml @@ -0,0 +1,31 @@ +overrides: + ceph: + wait-for-scrub: false + log-ignorelist: + - candidate had a stat error + - candidate had a read error + - deep-scrub 0 missing, 1 inconsistent objects + - deep-scrub 0 missing, 4 inconsistent objects + - deep-scrub [0-9]+ errors + - '!= omap_digest' + - '!= data_digest' + - repair 0 missing, 1 inconsistent objects + - repair 0 missing, 4 inconsistent objects + - repair [0-9]+ errors, [0-9]+ fixed + - scrub 0 missing, 1 inconsistent objects + - scrub [0-9]+ errors + - 'size 1 != size' + - attr name mismatch + - Regular scrub request, deep-scrub details will be lost + - candidate size [0-9]+ info size [0-9]+ mismatch + - overall HEALTH_ + - \(OSDMAP_FLAGS\) + - \(OSD_ + - \(PG_ + conf: + osd: + filestore debug inject read err: true + bluestore debug inject read err: true +tasks: +- repair_test: + diff --git a/qa/suites/rados/basic/tasks/scrub_test.yaml b/qa/suites/rados/basic/tasks/scrub_test.yaml new file mode 100644 index 000000000..dde468940 --- /dev/null +++ b/qa/suites/rados/basic/tasks/scrub_test.yaml @@ -0,0 +1,30 @@ +overrides: + ceph: + wait-for-scrub: false + log-ignorelist: + - '!= data_digest' + - '!= omap_digest' + - '!= size' + - 'deep-scrub 0 missing, 1 inconsistent objects' + - 'deep-scrub [0-9]+ errors' + - 'repair 0 missing, 1 inconsistent objects' + - 'repair [0-9]+ errors, [0-9]+ fixed' + - 'shard [0-9]+ .* : missing' + - 'deep-scrub 1 missing, 1 inconsistent objects' + - 'does not match object info size' + - 'attr name mistmatch' + - 'deep-scrub 1 missing, 0 inconsistent objects' + - 'failed to pick suitable auth object' + - 'candidate size [0-9]+ info size [0-9]+ mismatch' + - overall HEALTH_ + - \(OSDMAP_FLAGS\) + - \(OSD_ + - \(PG_ + - \(OSD_SCRUB_ERRORS\) + - \(TOO_FEW_PGS\) + conf: + osd: + osd deep scrub update digest min age: 0 + osd skip data digest: false +tasks: +- scrub_test: diff --git a/qa/suites/rados/cephadm b/qa/suites/rados/cephadm new file mode 120000 index 000000000..0e8b6774d --- /dev/null +++ b/qa/suites/rados/cephadm @@ -0,0 +1 @@ +../orch/cephadm \ No newline at end of file diff --git a/qa/suites/rados/dashboard/% b/qa/suites/rados/dashboard/% new file mode 100644 index 000000000..e69de29bb diff --git a/qa/suites/rados/dashboard/.qa b/qa/suites/rados/dashboard/.qa new file mode 120000 index 000000000..a602a0353 --- /dev/null +++ b/qa/suites/rados/dashboard/.qa @@ -0,0 +1 @@ +../.qa/ \ No newline at end of file diff --git a/qa/suites/rados/dashboard/centos_8.stream_container_tools.yaml b/qa/suites/rados/dashboard/centos_8.stream_container_tools.yaml new file mode 120000 index 000000000..7a86f967f --- /dev/null +++ b/qa/suites/rados/dashboard/centos_8.stream_container_tools.yaml @@ -0,0 +1 @@ +.qa/distros/podman/centos_8.stream_container_tools.yaml \ No newline at end of file diff --git a/qa/suites/rados/dashboard/clusters/+ b/qa/suites/rados/dashboard/clusters/+ new file mode 100644 index 000000000..e69de29bb diff --git a/qa/suites/rados/dashboard/clusters/.qa b/qa/suites/rados/dashboard/clusters/.qa new file mode 120000 index 000000000..a602a0353 --- /dev/null +++ b/qa/suites/rados/dashboard/clusters/.qa @@ -0,0 +1 @@ +../.qa/ \ No newline at end of file diff --git a/qa/suites/rados/dashboard/clusters/2-node-mgr.yaml b/qa/suites/rados/dashboard/clusters/2-node-mgr.yaml new file mode 120000 index 000000000..8a0b9123b --- /dev/null +++ b/qa/suites/rados/dashboard/clusters/2-node-mgr.yaml @@ -0,0 +1 @@ +.qa/clusters/2-node-mgr.yaml \ No newline at end of file diff --git a/qa/suites/rados/dashboard/debug/.qa b/qa/suites/rados/dashboard/debug/.qa new file mode 120000 index 000000000..a602a0353 --- /dev/null +++ b/qa/suites/rados/dashboard/debug/.qa @@ -0,0 +1 @@ +../.qa/ \ No newline at end of file diff --git a/qa/suites/rados/dashboard/debug/mgr.yaml b/qa/suites/rados/dashboard/debug/mgr.yaml new file mode 120000 index 000000000..651e5f8a8 --- /dev/null +++ b/qa/suites/rados/dashboard/debug/mgr.yaml @@ -0,0 +1 @@ +.qa/debug/mgr.yaml \ No newline at end of file diff --git a/qa/suites/rados/dashboard/mon_election b/qa/suites/rados/dashboard/mon_election new file mode 120000 index 000000000..3f331e621 --- /dev/null +++ b/qa/suites/rados/dashboard/mon_election @@ -0,0 +1 @@ +.qa/mon_election \ No newline at end of file diff --git a/qa/suites/rados/dashboard/random-objectstore$ b/qa/suites/rados/dashboard/random-objectstore$ new file mode 120000 index 000000000..c40bd3261 --- /dev/null +++ b/qa/suites/rados/dashboard/random-objectstore$ @@ -0,0 +1 @@ +.qa/objectstore \ No newline at end of file diff --git a/qa/suites/rados/dashboard/tasks/.qa b/qa/suites/rados/dashboard/tasks/.qa new file mode 120000 index 000000000..a602a0353 --- /dev/null +++ b/qa/suites/rados/dashboard/tasks/.qa @@ -0,0 +1 @@ +../.qa/ \ No newline at end of file diff --git a/qa/suites/rados/dashboard/tasks/dashboard.yaml b/qa/suites/rados/dashboard/tasks/dashboard.yaml new file mode 100644 index 000000000..01b3ecd90 --- /dev/null +++ b/qa/suites/rados/dashboard/tasks/dashboard.yaml @@ -0,0 +1,61 @@ + +tasks: + - install: + - ceph: + # tests may leave mgrs broken, so don't try and call into them + # to invoke e.g. pg dump during teardown. + wait-for-scrub: false + log-ignorelist: + - overall HEALTH_ + - \(MGR_DOWN\) + - \(PG_ + - replacing it with standby + - No standby daemons available + - \(FS_DEGRADED\) + - \(MDS_FAILED\) + - \(MDS_DEGRADED\) + - \(FS_WITH_FAILED_MDS\) + - \(MDS_DAMAGE\) + - \(MDS_ALL_DOWN\) + - \(MDS_UP_LESS_THAN_MAX\) + - \(OSD_DOWN\) + - \(OSD_HOST_DOWN\) + - \(POOL_APP_NOT_ENABLED\) + - \(OSDMAP_FLAGS\) + - \(OSD_FLAGS\) + - \(TELEMETRY_CHANGED\) + - pauserd,pausewr flag\(s\) set + - Monitor daemon marked osd\.[[:digit:]]+ down, but it is still running + - evicting unresponsive client .+ + - MON_DOWN + - rgw: [client.0] + - cephfs_test_runner: + fail_on_skip: false + modules: + - tasks.mgr.test_dashboard + - tasks.mgr.dashboard.test_api + - tasks.mgr.dashboard.test_auth + - tasks.mgr.dashboard.test_cephfs + - tasks.mgr.dashboard.test_cluster + - tasks.mgr.dashboard.test_cluster_configuration + - tasks.mgr.dashboard.test_crush_rule + - tasks.mgr.dashboard.test_erasure_code_profile + - tasks.mgr.dashboard.test_health + - tasks.mgr.dashboard.test_host + - tasks.mgr.dashboard.test_logs + - tasks.mgr.dashboard.test_mgr_module + - tasks.mgr.dashboard.test_monitor + - tasks.mgr.dashboard.test_motd + - tasks.mgr.dashboard.test_orchestrator + - tasks.mgr.dashboard.test_osd + - tasks.mgr.dashboard.test_perf_counters + - tasks.mgr.dashboard.test_pool + - tasks.mgr.dashboard.test_rbd + - tasks.mgr.dashboard.test_rbd_mirroring + - tasks.mgr.dashboard.test_requests + - tasks.mgr.dashboard.test_rgw + - tasks.mgr.dashboard.test_role + - tasks.mgr.dashboard.test_settings + - tasks.mgr.dashboard.test_summary + - tasks.mgr.dashboard.test_telemetry + - tasks.mgr.dashboard.test_user diff --git a/qa/suites/rados/mgr/% b/qa/suites/rados/mgr/% new file mode 100644 index 000000000..e69de29bb diff --git a/qa/suites/rados/mgr/.qa b/qa/suites/rados/mgr/.qa new file mode 120000 index 000000000..a602a0353 --- /dev/null +++ b/qa/suites/rados/mgr/.qa @@ -0,0 +1 @@ +../.qa/ \ No newline at end of file diff --git a/qa/suites/rados/mgr/clusters/+ b/qa/suites/rados/mgr/clusters/+ new file mode 100644 index 000000000..e69de29bb diff --git a/qa/suites/rados/mgr/clusters/.qa b/qa/suites/rados/mgr/clusters/.qa new file mode 120000 index 000000000..a602a0353 --- /dev/null +++ b/qa/suites/rados/mgr/clusters/.qa @@ -0,0 +1 @@ +../.qa/ \ No newline at end of file diff --git a/qa/suites/rados/mgr/clusters/2-node-mgr.yaml b/qa/suites/rados/mgr/clusters/2-node-mgr.yaml new file mode 120000 index 000000000..8a0b9123b --- /dev/null +++ b/qa/suites/rados/mgr/clusters/2-node-mgr.yaml @@ -0,0 +1 @@ +.qa/clusters/2-node-mgr.yaml \ No newline at end of file diff --git a/qa/suites/rados/mgr/debug/.qa b/qa/suites/rados/mgr/debug/.qa new file mode 120000 index 000000000..a602a0353 --- /dev/null +++ b/qa/suites/rados/mgr/debug/.qa @@ -0,0 +1 @@ +../.qa/ \ No newline at end of file diff --git a/qa/suites/rados/mgr/debug/mgr.yaml b/qa/suites/rados/mgr/debug/mgr.yaml new file mode 120000 index 000000000..651e5f8a8 --- /dev/null +++ b/qa/suites/rados/mgr/debug/mgr.yaml @@ -0,0 +1 @@ +.qa/debug/mgr.yaml \ No newline at end of file diff --git a/qa/suites/rados/mgr/mgr_ttl_cache/.qa b/qa/suites/rados/mgr/mgr_ttl_cache/.qa new file mode 120000 index 000000000..fea2489fd --- /dev/null +++ b/qa/suites/rados/mgr/mgr_ttl_cache/.qa @@ -0,0 +1 @@ +../.qa \ No newline at end of file diff --git a/qa/suites/rados/mgr/mgr_ttl_cache/disable.yaml b/qa/suites/rados/mgr/mgr_ttl_cache/disable.yaml new file mode 120000 index 000000000..d7db486dd --- /dev/null +++ b/qa/suites/rados/mgr/mgr_ttl_cache/disable.yaml @@ -0,0 +1 @@ +.qa/mgr_ttl_cache/disable.yaml \ No newline at end of file diff --git a/qa/suites/rados/mgr/mgr_ttl_cache/enable.yaml b/qa/suites/rados/mgr/mgr_ttl_cache/enable.yaml new file mode 120000 index 000000000..18286a656 --- /dev/null +++ b/qa/suites/rados/mgr/mgr_ttl_cache/enable.yaml @@ -0,0 +1 @@ +.qa/mgr_ttl_cache/enable.yaml \ No newline at end of file diff --git a/qa/suites/rados/mgr/mon_election b/qa/suites/rados/mgr/mon_election new file mode 120000 index 000000000..3f331e621 --- /dev/null +++ b/qa/suites/rados/mgr/mon_election @@ -0,0 +1 @@ +.qa/mon_election \ No newline at end of file diff --git a/qa/suites/rados/mgr/objectstore b/qa/suites/rados/mgr/objectstore new file mode 120000 index 000000000..848c65f9e --- /dev/null +++ b/qa/suites/rados/mgr/objectstore @@ -0,0 +1 @@ +.qa/objectstore_debug \ No newline at end of file diff --git a/qa/suites/rados/mgr/supported-random-distro$ b/qa/suites/rados/mgr/supported-random-distro$ new file mode 120000 index 000000000..7cef21eef --- /dev/null +++ b/qa/suites/rados/mgr/supported-random-distro$ @@ -0,0 +1 @@ +../basic/supported-random-distro$ \ No newline at end of file diff --git a/qa/suites/rados/mgr/tasks/.qa b/qa/suites/rados/mgr/tasks/.qa new file mode 120000 index 000000000..a602a0353 --- /dev/null +++ b/qa/suites/rados/mgr/tasks/.qa @@ -0,0 +1 @@ +../.qa/ \ No newline at end of file diff --git a/qa/suites/rados/mgr/tasks/crash.yaml b/qa/suites/rados/mgr/tasks/crash.yaml new file mode 100644 index 000000000..af4c40642 --- /dev/null +++ b/qa/suites/rados/mgr/tasks/crash.yaml @@ -0,0 +1,17 @@ + +tasks: + - install: + - ceph: + # tests may leave mgrs broken, so don't try and call into them + # to invoke e.g. pg dump during teardown. + wait-for-scrub: false + log-ignorelist: + - overall HEALTH_ + - \(MGR_DOWN\) + - \(PG_ + - \(RECENT_CRASH\) + - replacing it with standby + - No standby daemons available + - cephfs_test_runner: + modules: + - tasks.mgr.test_crash diff --git a/qa/suites/rados/mgr/tasks/failover.yaml b/qa/suites/rados/mgr/tasks/failover.yaml new file mode 100644 index 000000000..42c2f5c5b --- /dev/null +++ b/qa/suites/rados/mgr/tasks/failover.yaml @@ -0,0 +1,16 @@ + +tasks: + - install: + - ceph: + # tests may leave mgrs broken, so don't try and call into them + # to invoke e.g. pg dump during teardown. + wait-for-scrub: false + log-ignorelist: + - overall HEALTH_ + - \(MGR_DOWN\) + - \(PG_ + - replacing it with standby + - No standby daemons available + - cephfs_test_runner: + modules: + - tasks.mgr.test_failover diff --git a/qa/suites/rados/mgr/tasks/insights.yaml b/qa/suites/rados/mgr/tasks/insights.yaml new file mode 100644 index 000000000..5cb124bf7 --- /dev/null +++ b/qa/suites/rados/mgr/tasks/insights.yaml @@ -0,0 +1,19 @@ + +tasks: + - install: + - ceph: + # tests may leave mgrs broken, so don't try and call into them + # to invoke e.g. pg dump during teardown. + wait-for-scrub: false + log-ignorelist: + - overall HEALTH_ + - \(MGR_DOWN\) + - \(MGR_INSIGHTS_WARNING\) + - \(insights_health_check + - \(PG_ + - \(RECENT_CRASH\) + - replacing it with standby + - No standby daemons available + - cephfs_test_runner: + modules: + - tasks.mgr.test_insights diff --git a/qa/suites/rados/mgr/tasks/module_selftest.yaml b/qa/suites/rados/mgr/tasks/module_selftest.yaml new file mode 100644 index 000000000..905e6f783 --- /dev/null +++ b/qa/suites/rados/mgr/tasks/module_selftest.yaml @@ -0,0 +1,26 @@ + +tasks: + - install: + - ceph: + # tests may leave mgrs broken, so don't try and call into them + # to invoke e.g. pg dump during teardown. + wait-for-scrub: false + log-ignorelist: + - overall HEALTH_ + - \(MGR_DOWN\) + - \(PG_ + - replacing it with standby + - No standby daemons available + - Reduced data availability + - Degraded data redundancy + - objects misplaced + - Synthetic exception in serve + - influxdb python module not found + - \(MGR_ZABBIX_ + - foo bar + - Failed to open Telegraf + - evicting unresponsive client + - cephfs_test_runner: + modules: + - tasks.mgr.test_module_selftest + fail_on_skip: false diff --git a/qa/suites/rados/mgr/tasks/progress.yaml b/qa/suites/rados/mgr/tasks/progress.yaml new file mode 100644 index 000000000..73bbe3c99 --- /dev/null +++ b/qa/suites/rados/mgr/tasks/progress.yaml @@ -0,0 +1,25 @@ + +tasks: + - install: + - ceph: + config: + global: + osd pool default size : 3 + osd pool default min size : 2 + # tests may leave mgrs broken, so don't try and call into them + # to invoke e.g. pg dump during teardown. + wait-for-scrub: false + log-ignorelist: + - overall HEALTH_ + - \(MGR_DOWN\) + - \(MDS_ALL_DOWN\) + - \(MDS_UP_LESS_THAN_MAX\) + - \(FS_WITH_FAILED_MDS\) + - \(FS_DEGRADED\) + - \(PG_ + - \(OSDMAP_FLAGS\) + - replacing it with standby + - No standby daemons available + - cephfs_test_runner: + modules: + - tasks.mgr.test_progress diff --git a/qa/suites/rados/mgr/tasks/prometheus.yaml b/qa/suites/rados/mgr/tasks/prometheus.yaml new file mode 100644 index 000000000..388e19678 --- /dev/null +++ b/qa/suites/rados/mgr/tasks/prometheus.yaml @@ -0,0 +1,16 @@ + +tasks: + - install: + - ceph: + # tests may leave mgrs broken, so don't try and call into them + # to invoke e.g. pg dump during teardown. + wait-for-scrub: false + log-ignorelist: + - overall HEALTH_ + - \(MGR_DOWN\) + - \(PG_ + - replacing it with standby + - No standby daemons available + - cephfs_test_runner: + modules: + - tasks.mgr.test_prometheus diff --git a/qa/suites/rados/mgr/tasks/workunits.yaml b/qa/suites/rados/mgr/tasks/workunits.yaml new file mode 100644 index 000000000..6074de0ed --- /dev/null +++ b/qa/suites/rados/mgr/tasks/workunits.yaml @@ -0,0 +1,16 @@ +tasks: + - install: + - ceph: + # tests may leave mgrs broken, so don't try and call into them + # to invoke e.g. pg dump during teardown. + wait-for-scrub: false + log-ignorelist: + - overall HEALTH_ + - \(MGR_DOWN\) + - \(PG_ + - replacing it with standby + - No standby daemons available + - workunit: + clients: + client.0: + - mgr \ No newline at end of file diff --git a/qa/suites/rados/monthrash/% b/qa/suites/rados/monthrash/% new file mode 100644 index 000000000..e69de29bb diff --git a/qa/suites/rados/monthrash/.qa b/qa/suites/rados/monthrash/.qa new file mode 120000 index 000000000..a602a0353 --- /dev/null +++ b/qa/suites/rados/monthrash/.qa @@ -0,0 +1 @@ +../.qa/ \ No newline at end of file diff --git a/qa/suites/rados/monthrash/ceph.yaml b/qa/suites/rados/monthrash/ceph.yaml new file mode 100644 index 000000000..124ac3850 --- /dev/null +++ b/qa/suites/rados/monthrash/ceph.yaml @@ -0,0 +1,25 @@ +overrides: + ceph: + conf: + mon: + mon min osdmap epochs: 25 + paxos service trim min: 5 + # prune full osdmaps regularly + mon osdmap full prune min: 15 + mon osdmap full prune interval: 2 + mon osdmap full prune txsize: 2 + mon scrub inject crc mismatch: 0.01 + mon scrub inject missing keys: 0.05 +# thrashing monitors may make mgr have trouble w/ its keepalive + log-ignorelist: + - ScrubResult + - scrub mismatch + - overall HEALTH_ + - \(MGR_DOWN\) +# slow mons -> slow peering -> PG_AVAILABILITY + - \(PG_AVAILABILITY\) + - \(SLOW_OPS\) + - slow request +tasks: +- install: +- ceph: diff --git a/qa/suites/rados/monthrash/clusters/.qa b/qa/suites/rados/monthrash/clusters/.qa new file mode 120000 index 000000000..a602a0353 --- /dev/null +++ b/qa/suites/rados/monthrash/clusters/.qa @@ -0,0 +1 @@ +../.qa/ \ No newline at end of file diff --git a/qa/suites/rados/monthrash/clusters/3-mons.yaml b/qa/suites/rados/monthrash/clusters/3-mons.yaml new file mode 100644 index 000000000..4b721ef80 --- /dev/null +++ b/qa/suites/rados/monthrash/clusters/3-mons.yaml @@ -0,0 +1,7 @@ +roles: +- [mon.a, mon.c, osd.0, osd.1, osd.2] +- [mon.b, mgr.x, osd.3, osd.4, osd.5, client.0] +openstack: + - volumes: # attached to each instance + count: 3 + size: 10 # GB diff --git a/qa/suites/rados/monthrash/clusters/9-mons.yaml b/qa/suites/rados/monthrash/clusters/9-mons.yaml new file mode 100644 index 000000000..a2874c1d0 --- /dev/null +++ b/qa/suites/rados/monthrash/clusters/9-mons.yaml @@ -0,0 +1,7 @@ +roles: +- [mon.a, mon.b, mon.c, mon.d, mon.e, osd.0, osd.1, osd.2] +- [mon.f, mon.g, mon.h, mon.i, mgr.x, osd.3, osd.4, osd.5, client.0] +openstack: + - volumes: # attached to each instance + count: 3 + size: 10 # GB diff --git a/qa/suites/rados/monthrash/mon_election b/qa/suites/rados/monthrash/mon_election new file mode 120000 index 000000000..3f331e621 --- /dev/null +++ b/qa/suites/rados/monthrash/mon_election @@ -0,0 +1 @@ +.qa/mon_election \ No newline at end of file diff --git a/qa/suites/rados/monthrash/msgr b/qa/suites/rados/monthrash/msgr new file mode 120000 index 000000000..57bee80db --- /dev/null +++ b/qa/suites/rados/monthrash/msgr @@ -0,0 +1 @@ +.qa/msgr \ No newline at end of file diff --git a/qa/suites/rados/monthrash/msgr-failures/.qa b/qa/suites/rados/monthrash/msgr-failures/.qa new file mode 120000 index 000000000..a602a0353 --- /dev/null +++ b/qa/suites/rados/monthrash/msgr-failures/.qa @@ -0,0 +1 @@ +../.qa/ \ No newline at end of file diff --git a/qa/suites/rados/monthrash/msgr-failures/few.yaml b/qa/suites/rados/monthrash/msgr-failures/few.yaml new file mode 100644 index 000000000..519288992 --- /dev/null +++ b/qa/suites/rados/monthrash/msgr-failures/few.yaml @@ -0,0 +1,8 @@ +overrides: + ceph: + conf: + global: + ms inject socket failures: 5000 + mon client directed command retry: 5 + log-ignorelist: + - \(OSD_SLOW_PING_TIME diff --git a/qa/suites/rados/monthrash/msgr-failures/mon-delay.yaml b/qa/suites/rados/monthrash/msgr-failures/mon-delay.yaml new file mode 100644 index 000000000..83b136518 --- /dev/null +++ b/qa/suites/rados/monthrash/msgr-failures/mon-delay.yaml @@ -0,0 +1,14 @@ +overrides: + ceph: + conf: + global: + ms inject socket failures: 2500 + ms inject delay type: mon + ms inject delay probability: .005 + ms inject delay max: 1 + ms inject internal delays: .002 + mon client directed command retry: 5 + mgr: + debug monc: 10 + log-ignorelist: + - \(OSD_SLOW_PING_TIME diff --git a/qa/suites/rados/monthrash/objectstore b/qa/suites/rados/monthrash/objectstore new file mode 120000 index 000000000..848c65f9e --- /dev/null +++ b/qa/suites/rados/monthrash/objectstore @@ -0,0 +1 @@ +.qa/objectstore_debug \ No newline at end of file diff --git a/qa/suites/rados/monthrash/rados.yaml b/qa/suites/rados/monthrash/rados.yaml new file mode 120000 index 000000000..d256979c0 --- /dev/null +++ b/qa/suites/rados/monthrash/rados.yaml @@ -0,0 +1 @@ +.qa/config/rados.yaml \ No newline at end of file diff --git a/qa/suites/rados/monthrash/supported-random-distro$ b/qa/suites/rados/monthrash/supported-random-distro$ new file mode 120000 index 000000000..7cef21eef --- /dev/null +++ b/qa/suites/rados/monthrash/supported-random-distro$ @@ -0,0 +1 @@ +../basic/supported-random-distro$ \ No newline at end of file diff --git a/qa/suites/rados/monthrash/thrashers/.qa b/qa/suites/rados/monthrash/thrashers/.qa new file mode 120000 index 000000000..a602a0353 --- /dev/null +++ b/qa/suites/rados/monthrash/thrashers/.qa @@ -0,0 +1 @@ +../.qa/ \ No newline at end of file diff --git a/qa/suites/rados/monthrash/thrashers/force-sync-many.yaml b/qa/suites/rados/monthrash/thrashers/force-sync-many.yaml new file mode 100644 index 000000000..c2ec78fd3 --- /dev/null +++ b/qa/suites/rados/monthrash/thrashers/force-sync-many.yaml @@ -0,0 +1,12 @@ +overrides: + ceph: + log-ignorelist: + - overall HEALTH_ + - \(MON_DOWN\) + - \(TOO_FEW_PGS\) +tasks: +- mon_thrash: + revive_delay: 90 + thrash_delay: 1 + store_thrash: true + thrash_many: true diff --git a/qa/suites/rados/monthrash/thrashers/many.yaml b/qa/suites/rados/monthrash/thrashers/many.yaml new file mode 100644 index 000000000..958831232 --- /dev/null +++ b/qa/suites/rados/monthrash/thrashers/many.yaml @@ -0,0 +1,16 @@ +overrides: + ceph: + log-ignorelist: + - overall HEALTH_ + - \(MON_DOWN\) + conf: + osd: + mon client ping interval: 4 + mon client ping timeout: 12 +tasks: +- mon_thrash: + revive_delay: 20 + thrash_delay: 1 + thrash_many: true + freeze_mon_duration: 20 + freeze_mon_probability: 10 diff --git a/qa/suites/rados/monthrash/thrashers/one.yaml b/qa/suites/rados/monthrash/thrashers/one.yaml new file mode 100644 index 000000000..e969a0d8d --- /dev/null +++ b/qa/suites/rados/monthrash/thrashers/one.yaml @@ -0,0 +1,9 @@ +overrides: + ceph: + log-ignorelist: + - overall HEALTH_ + - \(MON_DOWN\) +tasks: +- mon_thrash: + revive_delay: 20 + thrash_delay: 1 diff --git a/qa/suites/rados/monthrash/thrashers/sync-many.yaml b/qa/suites/rados/monthrash/thrashers/sync-many.yaml new file mode 100644 index 000000000..e721b9b38 --- /dev/null +++ b/qa/suites/rados/monthrash/thrashers/sync-many.yaml @@ -0,0 +1,14 @@ +overrides: + ceph: + log-ignorelist: + - overall HEALTH_ + - \(MON_DOWN\) + conf: + mon: + paxos min: 10 + paxos trim min: 10 +tasks: +- mon_thrash: + revive_delay: 90 + thrash_delay: 1 + thrash_many: true diff --git a/qa/suites/rados/monthrash/thrashers/sync.yaml b/qa/suites/rados/monthrash/thrashers/sync.yaml new file mode 100644 index 000000000..8fdd1ad48 --- /dev/null +++ b/qa/suites/rados/monthrash/thrashers/sync.yaml @@ -0,0 +1,13 @@ +overrides: + ceph: + log-ignorelist: + - overall HEALTH_ + - \(MON_DOWN\) + conf: + mon: + paxos min: 10 + paxos trim min: 10 +tasks: +- mon_thrash: + revive_delay: 90 + thrash_delay: 1 diff --git a/qa/suites/rados/monthrash/workloads/.qa b/qa/suites/rados/monthrash/workloads/.qa new file mode 120000 index 000000000..a602a0353 --- /dev/null +++ b/qa/suites/rados/monthrash/workloads/.qa @@ -0,0 +1 @@ +../.qa/ \ No newline at end of file diff --git a/qa/suites/rados/monthrash/workloads/pool-create-delete.yaml b/qa/suites/rados/monthrash/workloads/pool-create-delete.yaml new file mode 100644 index 000000000..522302cd7 --- /dev/null +++ b/qa/suites/rados/monthrash/workloads/pool-create-delete.yaml @@ -0,0 +1,58 @@ +overrides: + ceph: + log-ignorelist: + - slow request + - overall HEALTH_ + - \(POOL_APP_NOT_ENABLED\) +tasks: +- exec: + client.0: + - ceph_test_rados_delete_pools_parallel + - ceph_test_rados_delete_pools_parallel + - ceph_test_rados_delete_pools_parallel + - ceph_test_rados_delete_pools_parallel + - ceph_test_rados_delete_pools_parallel + - ceph_test_rados_delete_pools_parallel + - ceph_test_rados_delete_pools_parallel + - ceph_test_rados_delete_pools_parallel + - ceph_test_rados_delete_pools_parallel + - ceph_test_rados_delete_pools_parallel + - ceph_test_rados_delete_pools_parallel + - ceph_test_rados_delete_pools_parallel + - ceph_test_rados_delete_pools_parallel + - ceph_test_rados_delete_pools_parallel + - ceph_test_rados_delete_pools_parallel + - ceph_test_rados_delete_pools_parallel + - ceph_test_rados_delete_pools_parallel + - ceph_test_rados_delete_pools_parallel + - ceph_test_rados_delete_pools_parallel + - ceph_test_rados_delete_pools_parallel + - ceph_test_rados_delete_pools_parallel + - ceph_test_rados_delete_pools_parallel + - ceph_test_rados_delete_pools_parallel + - ceph_test_rados_delete_pools_parallel + - ceph_test_rados_delete_pools_parallel + - ceph_test_rados_delete_pools_parallel + - ceph_test_rados_delete_pools_parallel + - ceph_test_rados_delete_pools_parallel + - ceph_test_rados_delete_pools_parallel + - ceph_test_rados_delete_pools_parallel + - ceph_test_rados_delete_pools_parallel + - ceph_test_rados_delete_pools_parallel + - ceph_test_rados_delete_pools_parallel + - ceph_test_rados_delete_pools_parallel + - ceph_test_rados_delete_pools_parallel + - ceph_test_rados_delete_pools_parallel + - ceph_test_rados_delete_pools_parallel + - ceph_test_rados_delete_pools_parallel + - ceph_test_rados_delete_pools_parallel + - ceph_test_rados_delete_pools_parallel + - ceph_test_rados_delete_pools_parallel + - ceph_test_rados_delete_pools_parallel + - ceph_test_rados_delete_pools_parallel + - ceph_test_rados_delete_pools_parallel + - ceph_test_rados_delete_pools_parallel + - ceph_test_rados_delete_pools_parallel + - ceph_test_rados_delete_pools_parallel + - ceph_test_rados_delete_pools_parallel + - ceph_test_rados_delete_pools_parallel diff --git a/qa/suites/rados/monthrash/workloads/rados_5925.yaml b/qa/suites/rados/monthrash/workloads/rados_5925.yaml new file mode 100644 index 000000000..ad19bd341 --- /dev/null +++ b/qa/suites/rados/monthrash/workloads/rados_5925.yaml @@ -0,0 +1,9 @@ +overrides: + ceph: + log-ignorelist: + - overall HEALTH_ + - \(POOL_APP_NOT_ENABLED\) +tasks: +- exec: + client.0: + - ceph_test_rados_delete_pools_parallel --debug_objecter 20 --debug_ms 1 --debug_rados 20 --debug_monc 20 diff --git a/qa/suites/rados/monthrash/workloads/rados_api_tests.yaml b/qa/suites/rados/monthrash/workloads/rados_api_tests.yaml new file mode 100644 index 000000000..e5ce0f6fd --- /dev/null +++ b/qa/suites/rados/monthrash/workloads/rados_api_tests.yaml @@ -0,0 +1,26 @@ +overrides: + ceph: + log-ignorelist: + - reached quota + - overall HEALTH_ + - \(CACHE_POOL_NO_HIT_SET\) + - \(CACHE_POOL_NEAR_FULL\) + - \(POOL_FULL\) + - \(SLOW_OPS\) + - \(MON_DOWN\) + - \(PG_ + - \(POOL_APP_NOT_ENABLED\) + - \(SMALLER_PGP_NUM\) + - slow request + conf: + global: + debug objecter: 20 + debug rados: 20 + debug ms: 1 + mon: + mon warn on pool no app: false +tasks: +- workunit: + clients: + client.0: + - rados/test.sh diff --git a/qa/suites/rados/monthrash/workloads/rados_mon_osdmap_prune.yaml b/qa/suites/rados/monthrash/workloads/rados_mon_osdmap_prune.yaml new file mode 100644 index 000000000..372bf2561 --- /dev/null +++ b/qa/suites/rados/monthrash/workloads/rados_mon_osdmap_prune.yaml @@ -0,0 +1,22 @@ +overrides: + ceph: + conf: + mon: + mon debug extra checks: true + mon min osdmap epochs: 100 + mon osdmap full prune enabled: true + mon osdmap full prune min: 200 + mon osdmap full prune interval: 10 + mon osdmap full prune txsize: 100 + osd: + osd beacon report interval: 10 + log-ignorelist: + # setting/unsetting noup will trigger health warns, + # causing tests to fail due to health warns, even if + # the tests themselves are successful. + - \(OSDMAP_FLAGS\) +tasks: +- workunit: + clients: + client.0: + - mon/test_mon_osdmap_prune.sh diff --git a/qa/suites/rados/monthrash/workloads/rados_mon_workunits.yaml b/qa/suites/rados/monthrash/workloads/rados_mon_workunits.yaml new file mode 100644 index 000000000..2f9729c92 --- /dev/null +++ b/qa/suites/rados/monthrash/workloads/rados_mon_workunits.yaml @@ -0,0 +1,17 @@ +overrides: + ceph: + log-ignorelist: + - but it is still running + - overall HEALTH_ + - \(PG_ + - \(MON_DOWN\) + - \(AUTH_BAD_CAPS\) +tasks: +- workunit: + clients: + client.0: + - mon/pool_ops.sh + - mon/crush_ops.sh + - mon/osd.sh + - mon/caps.sh + diff --git a/qa/suites/rados/monthrash/workloads/snaps-few-objects.yaml b/qa/suites/rados/monthrash/workloads/snaps-few-objects.yaml new file mode 100644 index 000000000..aa82d973a --- /dev/null +++ b/qa/suites/rados/monthrash/workloads/snaps-few-objects.yaml @@ -0,0 +1,13 @@ +tasks: +- rados: + clients: [client.0] + ops: 4000 + objects: 50 + op_weights: + read: 100 + write: 100 + delete: 50 + snap_create: 50 + snap_remove: 50 + rollback: 50 + copy_from: 50 diff --git a/qa/suites/rados/multimon/% b/qa/suites/rados/multimon/% new file mode 100644 index 000000000..e69de29bb diff --git a/qa/suites/rados/multimon/.qa b/qa/suites/rados/multimon/.qa new file mode 120000 index 000000000..a602a0353 --- /dev/null +++ b/qa/suites/rados/multimon/.qa @@ -0,0 +1 @@ +../.qa/ \ No newline at end of file diff --git a/qa/suites/rados/multimon/clusters/.qa b/qa/suites/rados/multimon/clusters/.qa new file mode 120000 index 000000000..a602a0353 --- /dev/null +++ b/qa/suites/rados/multimon/clusters/.qa @@ -0,0 +1 @@ +../.qa/ \ No newline at end of file diff --git a/qa/suites/rados/multimon/clusters/21.yaml b/qa/suites/rados/multimon/clusters/21.yaml new file mode 100644 index 000000000..aae968665 --- /dev/null +++ b/qa/suites/rados/multimon/clusters/21.yaml @@ -0,0 +1,8 @@ +roles: +- [mon.a, mon.d, mon.g, mon.j, mon.m, mon.p, mon.s] +- [mon.b, mon.e, mon.h, mon.k, mon.n, mon.q, mon.t, mgr.x] +- [mon.c, mon.f, mon.i, mon.l, mon.o, mon.r, mon.u] +openstack: +- volumes: # attached to each instance + count: 1 + size: 10 # GB diff --git a/qa/suites/rados/multimon/clusters/3.yaml b/qa/suites/rados/multimon/clusters/3.yaml new file mode 100644 index 000000000..11adef16b --- /dev/null +++ b/qa/suites/rados/multimon/clusters/3.yaml @@ -0,0 +1,7 @@ +roles: +- [mon.a, mon.c] +- [mon.b, mgr.x] +openstack: +- volumes: # attached to each instance + count: 2 + size: 10 # GB diff --git a/qa/suites/rados/multimon/clusters/6.yaml b/qa/suites/rados/multimon/clusters/6.yaml new file mode 100644 index 000000000..29c74dc7f --- /dev/null +++ b/qa/suites/rados/multimon/clusters/6.yaml @@ -0,0 +1,7 @@ +roles: +- [mon.a, mon.c, mon.e, mgr.x] +- [mon.b, mon.d, mon.f, mgr.y] +openstack: +- volumes: # attached to each instance + count: 1 + size: 10 # GB diff --git a/qa/suites/rados/multimon/clusters/9.yaml b/qa/suites/rados/multimon/clusters/9.yaml new file mode 100644 index 000000000..d51168556 --- /dev/null +++ b/qa/suites/rados/multimon/clusters/9.yaml @@ -0,0 +1,8 @@ +roles: +- [mon.a, mon.d, mon.g] +- [mon.b, mon.e, mon.h, mgr.x] +- [mon.c, mon.f, mon.i] +openstack: +- volumes: # attached to each instance + count: 1 + size: 10 # GB diff --git a/qa/suites/rados/multimon/mon_election b/qa/suites/rados/multimon/mon_election new file mode 120000 index 000000000..3f331e621 --- /dev/null +++ b/qa/suites/rados/multimon/mon_election @@ -0,0 +1 @@ +.qa/mon_election \ No newline at end of file diff --git a/qa/suites/rados/multimon/msgr b/qa/suites/rados/multimon/msgr new file mode 120000 index 000000000..57bee80db --- /dev/null +++ b/qa/suites/rados/multimon/msgr @@ -0,0 +1 @@ +.qa/msgr \ No newline at end of file diff --git a/qa/suites/rados/multimon/msgr-failures/.qa b/qa/suites/rados/multimon/msgr-failures/.qa new file mode 120000 index 000000000..a602a0353 --- /dev/null +++ b/qa/suites/rados/multimon/msgr-failures/.qa @@ -0,0 +1 @@ +../.qa/ \ No newline at end of file diff --git a/qa/suites/rados/multimon/msgr-failures/few.yaml b/qa/suites/rados/multimon/msgr-failures/few.yaml new file mode 100644 index 000000000..519288992 --- /dev/null +++ b/qa/suites/rados/multimon/msgr-failures/few.yaml @@ -0,0 +1,8 @@ +overrides: + ceph: + conf: + global: + ms inject socket failures: 5000 + mon client directed command retry: 5 + log-ignorelist: + - \(OSD_SLOW_PING_TIME diff --git a/qa/suites/rados/multimon/msgr-failures/many.yaml b/qa/suites/rados/multimon/msgr-failures/many.yaml new file mode 100644 index 000000000..d47b466b9 --- /dev/null +++ b/qa/suites/rados/multimon/msgr-failures/many.yaml @@ -0,0 +1,9 @@ +overrides: + ceph: + conf: + global: + ms inject socket failures: 1000 + mon client directed command retry: 5 + mon mgr beacon grace: 90 + log-ignorelist: + - \(OSD_SLOW_PING_TIME diff --git a/qa/suites/rados/multimon/no_pools.yaml b/qa/suites/rados/multimon/no_pools.yaml new file mode 100644 index 000000000..dc6c769ca --- /dev/null +++ b/qa/suites/rados/multimon/no_pools.yaml @@ -0,0 +1,5 @@ +overrides: + ceph: + create_rbd_pool: false + pre-mgr-commands: + - sudo ceph config set mgr mgr/devicehealth/enable_monitoring false --force diff --git a/qa/suites/rados/multimon/objectstore b/qa/suites/rados/multimon/objectstore new file mode 120000 index 000000000..848c65f9e --- /dev/null +++ b/qa/suites/rados/multimon/objectstore @@ -0,0 +1 @@ +.qa/objectstore_debug \ No newline at end of file diff --git a/qa/suites/rados/multimon/rados.yaml b/qa/suites/rados/multimon/rados.yaml new file mode 120000 index 000000000..d256979c0 --- /dev/null +++ b/qa/suites/rados/multimon/rados.yaml @@ -0,0 +1 @@ +.qa/config/rados.yaml \ No newline at end of file diff --git a/qa/suites/rados/multimon/supported-random-distro$ b/qa/suites/rados/multimon/supported-random-distro$ new file mode 120000 index 000000000..7cef21eef --- /dev/null +++ b/qa/suites/rados/multimon/supported-random-distro$ @@ -0,0 +1 @@ +../basic/supported-random-distro$ \ No newline at end of file diff --git a/qa/suites/rados/multimon/tasks/.qa b/qa/suites/rados/multimon/tasks/.qa new file mode 120000 index 000000000..a602a0353 --- /dev/null +++ b/qa/suites/rados/multimon/tasks/.qa @@ -0,0 +1 @@ +../.qa/ \ No newline at end of file diff --git a/qa/suites/rados/multimon/tasks/mon_clock_no_skews.yaml b/qa/suites/rados/multimon/tasks/mon_clock_no_skews.yaml new file mode 100644 index 000000000..2a4bf2bac --- /dev/null +++ b/qa/suites/rados/multimon/tasks/mon_clock_no_skews.yaml @@ -0,0 +1,11 @@ +tasks: +- install: +- ceph: + log-ignorelist: + - slow request + - .*clock.*skew.* + - clocks not synchronized + - overall HEALTH_ + - \(MON_CLOCK_SKEW\) +- mon_clock_skew_check: + expect-skew: false diff --git a/qa/suites/rados/multimon/tasks/mon_clock_with_skews.yaml b/qa/suites/rados/multimon/tasks/mon_clock_with_skews.yaml new file mode 100644 index 000000000..41749349e --- /dev/null +++ b/qa/suites/rados/multimon/tasks/mon_clock_with_skews.yaml @@ -0,0 +1,24 @@ +tasks: +- install: +- exec: + mon.b: + - sudo systemctl stop chronyd.service || true + - sudo systemctl stop systemd-timesync.service || true + - sudo systemctl stop ntpd.service || true + - sudo systemctl stop ntp.service || true + - date -u -s @$(expr $(date -u +%s) + 2) +- ceph: + wait-for-healthy: false + log-ignorelist: + - .*clock.*skew.* + - clocks not synchronized + - overall HEALTH_ + - \(MON_CLOCK_SKEW\) + - \(MGR_DOWN\) + - \(MON_DOWN\) + - \(PG_ + - \(SLOW_OPS\) + - No standby daemons available + - slow request +- mon_clock_skew_check: + expect-skew: true diff --git a/qa/suites/rados/multimon/tasks/mon_recovery.yaml b/qa/suites/rados/multimon/tasks/mon_recovery.yaml new file mode 100644 index 000000000..6373663c5 --- /dev/null +++ b/qa/suites/rados/multimon/tasks/mon_recovery.yaml @@ -0,0 +1,10 @@ +tasks: +- install: +- ceph: + log-ignorelist: + - overall HEALTH_ + - \(MON_DOWN\) + - \(PG_AVAILABILITY\) + - \(SLOW_OPS\) + - slow request +- mon_recovery: diff --git a/qa/suites/rados/objectstore/% b/qa/suites/rados/objectstore/% new file mode 100644 index 000000000..e69de29bb diff --git a/qa/suites/rados/objectstore/.qa b/qa/suites/rados/objectstore/.qa new file mode 120000 index 000000000..a602a0353 --- /dev/null +++ b/qa/suites/rados/objectstore/.qa @@ -0,0 +1 @@ +../.qa/ \ No newline at end of file diff --git a/qa/suites/rados/objectstore/backends/.qa b/qa/suites/rados/objectstore/backends/.qa new file mode 120000 index 000000000..a602a0353 --- /dev/null +++ b/qa/suites/rados/objectstore/backends/.qa @@ -0,0 +1 @@ +../.qa/ \ No newline at end of file diff --git a/qa/suites/rados/objectstore/backends/alloc-hint.yaml b/qa/suites/rados/objectstore/backends/alloc-hint.yaml new file mode 100644 index 000000000..047b02fa0 --- /dev/null +++ b/qa/suites/rados/objectstore/backends/alloc-hint.yaml @@ -0,0 +1,22 @@ +roles: +- [mon.a, mgr.x, osd.0, osd.1, osd.2, client.0] +openstack: + - volumes: # attached to each instance + count: 3 + size: 10 # GB + +overrides: + ceph: + fs: xfs + conf: + osd: + filestore xfs extsize: true + osd objectstore: filestore + +tasks: +- install: +- ceph: +- workunit: + clients: + all: + - rados/test_alloc_hint.sh diff --git a/qa/suites/rados/objectstore/backends/ceph_objectstore_tool.yaml b/qa/suites/rados/objectstore/backends/ceph_objectstore_tool.yaml new file mode 100644 index 000000000..72b143ad9 --- /dev/null +++ b/qa/suites/rados/objectstore/backends/ceph_objectstore_tool.yaml @@ -0,0 +1,25 @@ +roles: +- [mon.a, mgr.x, osd.0, osd.1, osd.2, osd.3, osd.4, osd.5, client.0] +openstack: +- volumes: # attached to each instance + count: 6 + size: 10 # GB +tasks: +- install: +- ceph: + fs: xfs + conf: + global: + osd max object name len: 460 + osd max object namespace len: 64 + osd: + osd objectstore: filestore + log-ignorelist: + - overall HEALTH_ + - \(OSDMAP_FLAGS\) + - \(OSD_ + - \(PG_ + - \(TOO_FEW_PGS\) + - \(POOL_APP_NOT_ENABLED\) +- ceph_objectstore_tool: + objects: 20 diff --git a/qa/suites/rados/objectstore/backends/filejournal.yaml b/qa/suites/rados/objectstore/backends/filejournal.yaml new file mode 100644 index 000000000..b0af80084 --- /dev/null +++ b/qa/suites/rados/objectstore/backends/filejournal.yaml @@ -0,0 +1,13 @@ +roles: +- [mon.a, mgr.x, osd.0, osd.1, client.0] +openstack: +- volumes: # attached to each instance + count: 2 + size: 10 # GB +tasks: +- install: +- ceph: + fs: xfs +- exec: + client.0: + - ceph_test_filejournal diff --git a/qa/suites/rados/objectstore/backends/filestore-idempotent-aio-journal.yaml b/qa/suites/rados/objectstore/backends/filestore-idempotent-aio-journal.yaml new file mode 100644 index 000000000..58b5197d6 --- /dev/null +++ b/qa/suites/rados/objectstore/backends/filestore-idempotent-aio-journal.yaml @@ -0,0 +1,14 @@ +roles: +- [mon.a, mgr.x, osd.0, osd.1, client.0] +openstack: +- volumes: # attached to each instance + count: 2 + size: 10 # GB +tasks: +- install: +- ceph: + fs: xfs + conf: + global: + journal aio: true +- filestore_idempotent: diff --git a/qa/suites/rados/objectstore/backends/filestore-idempotent.yaml b/qa/suites/rados/objectstore/backends/filestore-idempotent.yaml new file mode 100644 index 000000000..2d3f3c692 --- /dev/null +++ b/qa/suites/rados/objectstore/backends/filestore-idempotent.yaml @@ -0,0 +1,11 @@ +roles: +- [mon.a, mgr.x, osd.0, osd.1, client.0] +openstack: +- volumes: # attached to each instance + count: 2 + size: 10 # GB +tasks: +- install: +- ceph: + fs: xfs +- filestore_idempotent: diff --git a/qa/suites/rados/objectstore/backends/fusestore.yaml b/qa/suites/rados/objectstore/backends/fusestore.yaml new file mode 100644 index 000000000..1c34fcaea --- /dev/null +++ b/qa/suites/rados/objectstore/backends/fusestore.yaml @@ -0,0 +1,9 @@ +roles: +- [mon.a, mgr.x, osd.0, osd.1, client.0] +tasks: +- install: +- workunit: + clients: + all: + - objectstore/test_fuse.sh + diff --git a/qa/suites/rados/objectstore/backends/keyvaluedb.yaml b/qa/suites/rados/objectstore/backends/keyvaluedb.yaml new file mode 100644 index 000000000..efff8d379 --- /dev/null +++ b/qa/suites/rados/objectstore/backends/keyvaluedb.yaml @@ -0,0 +1,8 @@ +roles: +- [mon.a, mgr.x, osd.0, osd.1, client.0] +tasks: +- install: +- exec: + client.0: + - mkdir $TESTDIR/kvtest && cd $TESTDIR/kvtest && ceph_test_keyvaluedb + - rm -rf $TESTDIR/kvtest diff --git a/qa/suites/rados/objectstore/backends/objectcacher-stress.yaml b/qa/suites/rados/objectstore/backends/objectcacher-stress.yaml new file mode 100644 index 000000000..e407a3918 --- /dev/null +++ b/qa/suites/rados/objectstore/backends/objectcacher-stress.yaml @@ -0,0 +1,14 @@ +roles: +- [mon.a, mgr.x, osd.0, osd.1, client.0] +openstack: +- volumes: # attached to each instance + count: 2 + size: 10 # GB +tasks: +- install: +- ceph: + fs: xfs +- workunit: + clients: + all: + - osdc/stress_objectcacher.sh diff --git a/qa/suites/rados/objectstore/backends/objectstore.yaml b/qa/suites/rados/objectstore/backends/objectstore.yaml new file mode 100644 index 000000000..d68270a8e --- /dev/null +++ b/qa/suites/rados/objectstore/backends/objectstore.yaml @@ -0,0 +1,12 @@ +roles: +- [mon.a, mgr.x, osd.0, osd.1, client.0] +openstack: +- volumes: # attached to each instance + count: 2 + size: 10 # GB +tasks: +- install: +- exec: + client.0: + - mkdir $TESTDIR/archive/ostest && cd $TESTDIR/archive/ostest && ulimit -Sn 16384 && CEPH_ARGS="--no-log-to-stderr --log-file $TESTDIR/archive/ceph_test_objectstore.log --debug-filestore 20 --debug-bluestore 20" ceph_test_objectstore --gtest_filter=-*/3 --gtest_catch_exceptions=0 + - rm -rf $TESTDIR/archive/ostest diff --git a/qa/suites/rados/objectstore/supported-random-distro$ b/qa/suites/rados/objectstore/supported-random-distro$ new file mode 120000 index 000000000..7cef21eef --- /dev/null +++ b/qa/suites/rados/objectstore/supported-random-distro$ @@ -0,0 +1 @@ +../basic/supported-random-distro$ \ No newline at end of file diff --git a/qa/suites/rados/perf/% b/qa/suites/rados/perf/% new file mode 100644 index 000000000..e69de29bb diff --git a/qa/suites/rados/perf/.qa b/qa/suites/rados/perf/.qa new file mode 120000 index 000000000..a602a0353 --- /dev/null +++ b/qa/suites/rados/perf/.qa @@ -0,0 +1 @@ +../.qa/ \ No newline at end of file diff --git a/qa/suites/rados/perf/ceph.yaml b/qa/suites/rados/perf/ceph.yaml new file mode 100644 index 000000000..f7f1c8556 --- /dev/null +++ b/qa/suites/rados/perf/ceph.yaml @@ -0,0 +1,13 @@ +roles: +- [mon.a, mgr.x, osd.0, osd.1, osd.2, client.0] +tasks: +- install: +- ceph: + fs: xfs + wait-for-scrub: false + log-ignorelist: + - \(PG_ + - \(OSD_ + - \(OBJECT_ + - overall HEALTH +- ssh_keys: diff --git a/qa/suites/rados/perf/mon_election b/qa/suites/rados/perf/mon_election new file mode 120000 index 000000000..3f331e621 --- /dev/null +++ b/qa/suites/rados/perf/mon_election @@ -0,0 +1 @@ +.qa/mon_election \ No newline at end of file diff --git a/qa/suites/rados/perf/objectstore/.qa b/qa/suites/rados/perf/objectstore/.qa new file mode 120000 index 000000000..a602a0353 --- /dev/null +++ b/qa/suites/rados/perf/objectstore/.qa @@ -0,0 +1 @@ +../.qa/ \ No newline at end of file diff --git a/qa/suites/rados/perf/objectstore/bluestore-basic-min-osd-mem-target.yaml b/qa/suites/rados/perf/objectstore/bluestore-basic-min-osd-mem-target.yaml new file mode 100644 index 000000000..32f596da1 --- /dev/null +++ b/qa/suites/rados/perf/objectstore/bluestore-basic-min-osd-mem-target.yaml @@ -0,0 +1,25 @@ +overrides: + thrashosds: + bdev_inject_crash: 2 + bdev_inject_crash_probability: .5 + ceph: + fs: xfs + conf: + osd: + osd objectstore: bluestore + osd memory target: 2147483648 # min recommended is 2_G + bluestore block size: 96636764160 + debug bluestore: 20 + debug bluefs: 20 + debug rocksdb: 10 + bluestore fsck on mount: true + # lower the full ratios since we can fill up a 100gb osd so quickly + mon osd full ratio: .9 + mon osd backfillfull_ratio: .85 + mon osd nearfull ratio: .8 + osd failsafe full ratio: .95 +# this doesn't work with failures bc the log writes are not atomic across the two backends +# bluestore bluefs env mirror: true + bdev enable discard: true + bdev async discard: true + diff --git a/qa/suites/rados/perf/objectstore/bluestore-bitmap.yaml b/qa/suites/rados/perf/objectstore/bluestore-bitmap.yaml new file mode 100644 index 000000000..b18e04bee --- /dev/null +++ b/qa/suites/rados/perf/objectstore/bluestore-bitmap.yaml @@ -0,0 +1,43 @@ +overrides: + thrashosds: + bdev_inject_crash: 2 + bdev_inject_crash_probability: .5 + ceph: + fs: xfs + conf: + osd: + osd objectstore: bluestore + bluestore block size: 96636764160 + debug bluestore: 20 + debug bluefs: 20 + debug rocksdb: 10 + bluestore fsck on mount: true + bluestore allocator: bitmap + # lower the full ratios since we can fill up a 100gb osd so quickly + mon osd full ratio: .9 + mon osd backfillfull_ratio: .85 + mon osd nearfull ratio: .8 + osd failsafe full ratio: .95 +# this doesn't work with failures bc the log writes are not atomic across the two backends +# bluestore bluefs env mirror: true + bdev enable discard: true + bdev async discard: true + ceph-deploy: + fs: xfs + bluestore: yes + conf: + osd: + osd objectstore: bluestore + bluestore block size: 96636764160 + debug bluestore: 20 + debug bluefs: 20 + debug rocksdb: 10 + bluestore fsck on mount: true + # lower the full ratios since we can fill up a 100gb osd so quickly + mon osd full ratio: .9 + mon osd backfillfull_ratio: .85 + mon osd nearfull ratio: .8 + osd failsafe full ratio: .95 + bdev enable discard: true + bdev async discard: true + diff --git a/qa/suites/rados/perf/objectstore/bluestore-comp.yaml b/qa/suites/rados/perf/objectstore/bluestore-comp.yaml new file mode 100644 index 000000000..b408032fd --- /dev/null +++ b/qa/suites/rados/perf/objectstore/bluestore-comp.yaml @@ -0,0 +1,23 @@ +overrides: + thrashosds: + bdev_inject_crash: 2 + bdev_inject_crash_probability: .5 + ceph: + fs: xfs + conf: + osd: + osd objectstore: bluestore + bluestore block size: 96636764160 + debug bluestore: 20 + debug bluefs: 20 + debug rocksdb: 10 + bluestore compression mode: aggressive + bluestore fsck on mount: true + # lower the full ratios since we can fill up a 100gb osd so quickly + mon osd full ratio: .9 + mon osd backfillfull_ratio: .85 + mon osd nearfull ratio: .8 + osd failsafe full ratio: .95 + +# this doesn't work with failures bc the log writes are not atomic across the two backends +# bluestore bluefs env mirror: true diff --git a/qa/suites/rados/perf/objectstore/bluestore-low-osd-mem-target.yaml b/qa/suites/rados/perf/objectstore/bluestore-low-osd-mem-target.yaml new file mode 100644 index 000000000..b2a49790b --- /dev/null +++ b/qa/suites/rados/perf/objectstore/bluestore-low-osd-mem-target.yaml @@ -0,0 +1,25 @@ +overrides: + thrashosds: + bdev_inject_crash: 2 + bdev_inject_crash_probability: .5 + ceph: + fs: xfs + conf: + osd: + osd objectstore: bluestore + osd memory target: 1610612736 # reduced to 1.5_G + bluestore block size: 96636764160 + debug bluestore: 20 + debug bluefs: 20 + debug rocksdb: 10 + bluestore fsck on mount: true + # lower the full ratios since we can fill up a 100gb osd so quickly + mon osd full ratio: .9 + mon osd backfillfull_ratio: .85 + mon osd nearfull ratio: .8 + osd failsafe full ratio: .95 +# this doesn't work with failures bc the log writes are not atomic across the two backends +# bluestore bluefs env mirror: true + bdev enable discard: true + bdev async discard: true + diff --git a/qa/suites/rados/perf/objectstore/bluestore-stupid.yaml b/qa/suites/rados/perf/objectstore/bluestore-stupid.yaml new file mode 100644 index 000000000..ca811f131 --- /dev/null +++ b/qa/suites/rados/perf/objectstore/bluestore-stupid.yaml @@ -0,0 +1,43 @@ +overrides: + thrashosds: + bdev_inject_crash: 2 + bdev_inject_crash_probability: .5 + ceph: + fs: xfs + conf: + osd: + osd objectstore: bluestore + bluestore block size: 96636764160 + debug bluestore: 20 + debug bluefs: 20 + debug rocksdb: 10 + bluestore fsck on mount: true + bluestore allocator: stupid + # lower the full ratios since we can fill up a 100gb osd so quickly + mon osd full ratio: .9 + mon osd backfillfull_ratio: .85 + mon osd nearfull ratio: .8 + osd failsafe full ratio: .95 +# this doesn't work with failures bc the log writes are not atomic across the two backends +# bluestore bluefs env mirror: true + bdev enable discard: true + bdev async discard: true + ceph-deploy: + fs: xfs + bluestore: yes + conf: + osd: + osd objectstore: bluestore + bluestore block size: 96636764160 + debug bluestore: 20 + debug bluefs: 20 + debug rocksdb: 10 + bluestore fsck on mount: true + # lower the full ratios since we can fill up a 100gb osd so quickly + mon osd full ratio: .9 + mon osd backfillfull_ratio: .85 + mon osd nearfull ratio: .8 + osd failsafe full ratio: .95 + bdev enable discard: true + bdev async discard: true + diff --git a/qa/suites/rados/perf/openstack.yaml b/qa/suites/rados/perf/openstack.yaml new file mode 100644 index 000000000..f4d1349b4 --- /dev/null +++ b/qa/suites/rados/perf/openstack.yaml @@ -0,0 +1,4 @@ +openstack: + - volumes: # attached to each instance + count: 3 + size: 30 # GB diff --git a/qa/suites/rados/perf/scheduler/dmclock_1Shard_16Threads.yaml b/qa/suites/rados/perf/scheduler/dmclock_1Shard_16Threads.yaml new file mode 100644 index 000000000..10388ad72 --- /dev/null +++ b/qa/suites/rados/perf/scheduler/dmclock_1Shard_16Threads.yaml @@ -0,0 +1,7 @@ +overrides: + ceph: + conf: + osd: + osd op num shards: 1 + osd op num threads per shard: 16 + osd op queue: mclock_scheduler diff --git a/qa/suites/rados/perf/scheduler/dmclock_default_shards.yaml b/qa/suites/rados/perf/scheduler/dmclock_default_shards.yaml new file mode 100644 index 000000000..57a0ed912 --- /dev/null +++ b/qa/suites/rados/perf/scheduler/dmclock_default_shards.yaml @@ -0,0 +1,5 @@ +overrides: + ceph: + conf: + osd: + osd op queue: mclock_scheduler diff --git a/qa/suites/rados/perf/scheduler/wpq_default_shards.yaml b/qa/suites/rados/perf/scheduler/wpq_default_shards.yaml new file mode 100644 index 000000000..25d358f27 --- /dev/null +++ b/qa/suites/rados/perf/scheduler/wpq_default_shards.yaml @@ -0,0 +1,5 @@ +overrides: + ceph: + conf: + osd: + osd op queue: wpq diff --git a/qa/suites/rados/perf/settings/.qa b/qa/suites/rados/perf/settings/.qa new file mode 120000 index 000000000..a602a0353 --- /dev/null +++ b/qa/suites/rados/perf/settings/.qa @@ -0,0 +1 @@ +../.qa/ \ No newline at end of file diff --git a/qa/suites/rados/perf/settings/optimized.yaml b/qa/suites/rados/perf/settings/optimized.yaml new file mode 100644 index 000000000..dc4dcbb96 --- /dev/null +++ b/qa/suites/rados/perf/settings/optimized.yaml @@ -0,0 +1,74 @@ +overrides: + ceph: + conf: + mon: + debug mon: "0/0" + debug ms: "0/0" + debug paxos: "0/0" + osd: + debug filestore: "0/0" + debug journal: "0/0" + debug ms: "0/0" + debug osd: "0/0" + global: + auth client required: none + auth cluster required: none + auth service required: none + auth supported: none + + debug lockdep: "0/0" + debug context: "0/0" + debug crush: "0/0" + debug mds: "0/0" + debug mds balancer: "0/0" + debug mds locker: "0/0" + debug mds log: "0/0" + debug mds log expire: "0/0" + debug mds migrator: "0/0" + debug buffer: "0/0" + debug timer: "0/0" + debug filer: "0/0" + debug striper: "0/0" + debug objecter: "0/0" + debug rados: "0/0" + debug rbd: "0/0" + debug rbd mirror: "0/0" + debug rbd replay: "0/0" + debug journaler: "0/0" + debug objectcacher: "0/0" + debug client: "0/0" + debug osd: "0/0" + debug optracker: "0/0" + debug objclass: "0/0" + debug filestore: "0/0" + debug journal: "0/0" + debug ms: "0/0" + debug mon: "0/0" + debug monc: "0/0" + debug paxos: "0/0" + debug tp: "0/0" + debug auth: "0/0" + debug crypto: "0/0" + debug finisher: "0/0" + debug heartbeatmap: "0/0" + debug perfcounter: "0/0" + debug rgw: "0/0" + debug rgw sync: "0/0" + debug civetweb: "0/0" + debug javaclient: "0/0" + debug asok: "0/0" + debug throttle: "0/0" + debug refs: "0/0" + debug compressor: "0/0" + debug bluestore: "0/0" + debug bluefs: "0/0" + debug bdev: "0/0" + debug kstore: "0/0" + debug rocksdb: "0/0" + debug leveldb: "0/0" + debug memdb: "0/0" + debug fuse: "0/0" + debug mgr: "0/0" + debug mgrc: "0/0" + debug dpdk: "0/0" + debug eventtrace: "0/0" diff --git a/qa/suites/rados/perf/ubuntu_latest.yaml b/qa/suites/rados/perf/ubuntu_latest.yaml new file mode 120000 index 000000000..3a09f9abb --- /dev/null +++ b/qa/suites/rados/perf/ubuntu_latest.yaml @@ -0,0 +1 @@ +.qa/distros/supported/ubuntu_latest.yaml \ No newline at end of file diff --git a/qa/suites/rados/perf/workloads/.qa b/qa/suites/rados/perf/workloads/.qa new file mode 120000 index 000000000..a602a0353 --- /dev/null +++ b/qa/suites/rados/perf/workloads/.qa @@ -0,0 +1 @@ +../.qa/ \ No newline at end of file diff --git a/qa/suites/rados/perf/workloads/fio_4K_rand_read.yaml b/qa/suites/rados/perf/workloads/fio_4K_rand_read.yaml new file mode 100644 index 000000000..d5ef33ab3 --- /dev/null +++ b/qa/suites/rados/perf/workloads/fio_4K_rand_read.yaml @@ -0,0 +1,24 @@ +tasks: +- cbt: + benchmarks: + librbdfio: + op_size: [4096] + time: 60 + mode: ['randread'] + norandommap: True + vol_size: 4096 + procs_per_volume: [1] + volumes_per_client: [2] + iodepth: [32] + osd_ra: [4096] + pool_profile: 'rbd' + log_avg_msec: 100 + cluster: + user: 'ubuntu' + osds_per_node: 3 + iterations: 1 + pool_profiles: + rbd: + pg_size: 128 + pgp_size: 128 + replication: 3 diff --git a/qa/suites/rados/perf/workloads/fio_4K_rand_rw.yaml b/qa/suites/rados/perf/workloads/fio_4K_rand_rw.yaml new file mode 100644 index 000000000..14d43f143 --- /dev/null +++ b/qa/suites/rados/perf/workloads/fio_4K_rand_rw.yaml @@ -0,0 +1,24 @@ +tasks: +- cbt: + benchmarks: + librbdfio: + op_size: [4096] + time: 60 + mode: ['randrw'] + norandommap: True + vol_size: 4096 + procs_per_volume: [1] + volumes_per_client: [2] + iodepth: [32] + osd_ra: [4096] + pool_profile: 'rbd' + log_avg_msec: 100 + cluster: + user: 'ubuntu' + osds_per_node: 3 + iterations: 1 + pool_profiles: + rbd: + pg_size: 128 + pgp_size: 128 + replication: 3 diff --git a/qa/suites/rados/perf/workloads/fio_4M_rand_read.yaml b/qa/suites/rados/perf/workloads/fio_4M_rand_read.yaml new file mode 100644 index 000000000..b07432243 --- /dev/null +++ b/qa/suites/rados/perf/workloads/fio_4M_rand_read.yaml @@ -0,0 +1,24 @@ +tasks: +- cbt: + benchmarks: + librbdfio: + op_size: [4194304] + time: 60 + mode: ['randread'] + norandommap: True + vol_size: 4096 + procs_per_volume: [1] + volumes_per_client: [2] + iodepth: [32] + osd_ra: [4096] + pool_profile: 'rbd' + log_avg_msec: 100 + cluster: + user: 'ubuntu' + osds_per_node: 3 + iterations: 1 + pool_profiles: + rbd: + pg_size: 128 + pgp_size: 128 + replication: 3 diff --git a/qa/suites/rados/perf/workloads/fio_4M_rand_rw.yaml b/qa/suites/rados/perf/workloads/fio_4M_rand_rw.yaml new file mode 100644 index 000000000..5fd6e2877 --- /dev/null +++ b/qa/suites/rados/perf/workloads/fio_4M_rand_rw.yaml @@ -0,0 +1,24 @@ +tasks: +- cbt: + benchmarks: + librbdfio: + op_size: [4194304] + time: 60 + mode: ['randrw'] + norandommap: True + vol_size: 4096 + procs_per_volume: [1] + volumes_per_client: [2] + iodepth: [32] + osd_ra: [4096] + pool_profile: 'rbd' + log_avg_msec: 100 + cluster: + user: 'ubuntu' + osds_per_node: 3 + iterations: 1 + pool_profiles: + rbd: + pg_size: 128 + pgp_size: 128 + replication: 3 diff --git a/qa/suites/rados/perf/workloads/fio_4M_rand_write.yaml b/qa/suites/rados/perf/workloads/fio_4M_rand_write.yaml new file mode 100644 index 000000000..2d9d83611 --- /dev/null +++ b/qa/suites/rados/perf/workloads/fio_4M_rand_write.yaml @@ -0,0 +1,24 @@ +tasks: +- cbt: + benchmarks: + librbdfio: + op_size: [4194304] + time: 60 + mode: ['randwrite'] + norandommap: True + vol_size: 4096 + procs_per_volume: [1] + volumes_per_client: [2] + iodepth: [32] + osd_ra: [4096] + pool_profile: 'rbd' + log_avg_msec: 100 + cluster: + user: 'ubuntu' + osds_per_node: 3 + iterations: 1 + pool_profiles: + rbd: + pg_size: 128 + pgp_size: 128 + replication: 3 diff --git a/qa/suites/rados/perf/workloads/radosbench_4K_rand_read.yaml b/qa/suites/rados/perf/workloads/radosbench_4K_rand_read.yaml new file mode 100644 index 000000000..f1de9b41b --- /dev/null +++ b/qa/suites/rados/perf/workloads/radosbench_4K_rand_read.yaml @@ -0,0 +1,24 @@ +tasks: +- cbt: + benchmarks: + radosbench: + concurrent_ops: 4 + concurrent_procs: 2 + op_size: [4096] + pool_monitoring_list: + - collectl + pool_profile: 'replicated' + run_monitoring_list: + - collectl + time: 60 + write_only: false + readmode: 'rand' + cluster: + user: 'ubuntu' + osds_per_node: 3 + iterations: 1 + pool_profiles: + replicated: + pg_size: 256 + pgp_size: 256 + replication: 'replicated' diff --git a/qa/suites/rados/perf/workloads/radosbench_4K_seq_read.yaml b/qa/suites/rados/perf/workloads/radosbench_4K_seq_read.yaml new file mode 100644 index 000000000..8fb204a2f --- /dev/null +++ b/qa/suites/rados/perf/workloads/radosbench_4K_seq_read.yaml @@ -0,0 +1,23 @@ +tasks: +- cbt: + benchmarks: + radosbench: + concurrent_ops: 4 + concurrent_procs: 2 + op_size: [4096] + pool_monitoring_list: + - collectl + pool_profile: 'replicated' + run_monitoring_list: + - collectl + time: 60 + write_only: false + cluster: + user: 'ubuntu' + osds_per_node: 3 + iterations: 1 + pool_profiles: + replicated: + pg_size: 256 + pgp_size: 256 + replication: 'replicated' diff --git a/qa/suites/rados/perf/workloads/radosbench_4M_rand_read.yaml b/qa/suites/rados/perf/workloads/radosbench_4M_rand_read.yaml new file mode 100644 index 000000000..cc1c74489 --- /dev/null +++ b/qa/suites/rados/perf/workloads/radosbench_4M_rand_read.yaml @@ -0,0 +1,24 @@ +tasks: +- cbt: + benchmarks: + radosbench: + concurrent_ops: 4 + concurrent_procs: 2 + op_size: [4194304] + pool_monitoring_list: + - collectl + pool_profile: 'replicated' + run_monitoring_list: + - collectl + time: 60 + write_only: false + readmode: 'rand' + cluster: + user: 'ubuntu' + osds_per_node: 3 + iterations: 1 + pool_profiles: + replicated: + pg_size: 256 + pgp_size: 256 + replication: 'replicated' diff --git a/qa/suites/rados/perf/workloads/radosbench_4M_seq_read.yaml b/qa/suites/rados/perf/workloads/radosbench_4M_seq_read.yaml new file mode 100644 index 000000000..3ab55cf51 --- /dev/null +++ b/qa/suites/rados/perf/workloads/radosbench_4M_seq_read.yaml @@ -0,0 +1,23 @@ +tasks: +- cbt: + benchmarks: + radosbench: + concurrent_ops: 4 + concurrent_procs: 2 + op_size: [4194304] + pool_monitoring_list: + - collectl + pool_profile: 'replicated' + run_monitoring_list: + - collectl + time: 60 + write_only: false + cluster: + user: 'ubuntu' + osds_per_node: 3 + iterations: 1 + pool_profiles: + replicated: + pg_size: 256 + pgp_size: 256 + replication: 'replicated' diff --git a/qa/suites/rados/perf/workloads/radosbench_4M_write.yaml b/qa/suites/rados/perf/workloads/radosbench_4M_write.yaml new file mode 100644 index 000000000..f6a5d715c --- /dev/null +++ b/qa/suites/rados/perf/workloads/radosbench_4M_write.yaml @@ -0,0 +1,23 @@ +tasks: +- cbt: + benchmarks: + radosbench: + concurrent_ops: 4 + concurrent_procs: 2 + op_size: [4194304] + pool_monitoring_list: + - collectl + pool_profile: 'replicated' + run_monitoring_list: + - collectl + time: 60 + write_only: true + cluster: + user: 'ubuntu' + osds_per_node: 3 + iterations: 1 + pool_profiles: + replicated: + pg_size: 256 + pgp_size: 256 + replication: 'replicated' diff --git a/qa/suites/rados/perf/workloads/radosbench_omap_write.yaml b/qa/suites/rados/perf/workloads/radosbench_omap_write.yaml new file mode 100644 index 000000000..5df4674d9 --- /dev/null +++ b/qa/suites/rados/perf/workloads/radosbench_omap_write.yaml @@ -0,0 +1,7 @@ +tasks: +- radosbench: + clients: [client.0] + write-omap: True + objectsize: 4096 + size: 4096 + time: 300 diff --git a/qa/suites/rados/perf/workloads/sample_fio.yaml b/qa/suites/rados/perf/workloads/sample_fio.yaml new file mode 100644 index 000000000..98411392d --- /dev/null +++ b/qa/suites/rados/perf/workloads/sample_fio.yaml @@ -0,0 +1,24 @@ +tasks: +- cbt: + benchmarks: + librbdfio: + op_size: [4096] + time: 60 + mode: ['randwrite'] + norandommap: True + vol_size: 4096 + procs_per_volume: [1] + volumes_per_client: [2] + iodepth: [32] + osd_ra: [4096] + pool_profile: 'rbd' + log_avg_msec: 100 + cluster: + user: 'ubuntu' + osds_per_node: 3 + iterations: 1 + pool_profiles: + rbd: + pg_size: 128 + pgp_size: 128 + replication: 3 diff --git a/qa/suites/rados/perf/workloads/sample_radosbench.yaml b/qa/suites/rados/perf/workloads/sample_radosbench.yaml new file mode 100644 index 000000000..e3dc47ae6 --- /dev/null +++ b/qa/suites/rados/perf/workloads/sample_radosbench.yaml @@ -0,0 +1,23 @@ +tasks: +- cbt: + benchmarks: + radosbench: + concurrent_ops: 4 + concurrent_procs: 2 + op_size: [4096] + pool_monitoring_list: + - collectl + pool_profile: 'replicated' + run_monitoring_list: + - collectl + time: 60 + write_only: true + cluster: + user: 'ubuntu' + osds_per_node: 3 + iterations: 1 + pool_profiles: + replicated: + pg_size: 256 + pgp_size: 256 + replication: 'replicated' diff --git a/qa/suites/rados/rest/% b/qa/suites/rados/rest/% new file mode 100644 index 000000000..e69de29bb diff --git a/qa/suites/rados/rest/.qa b/qa/suites/rados/rest/.qa new file mode 120000 index 000000000..a602a0353 --- /dev/null +++ b/qa/suites/rados/rest/.qa @@ -0,0 +1 @@ +../.qa/ \ No newline at end of file diff --git a/qa/suites/rados/rest/mgr-restful.yaml b/qa/suites/rados/rest/mgr-restful.yaml new file mode 100644 index 000000000..c863463de --- /dev/null +++ b/qa/suites/rados/rest/mgr-restful.yaml @@ -0,0 +1,30 @@ +openstack: +- volumes: # attached to each instance + count: 3 + size: 10 # GB +roles: +- [mon.a, mgr.x, osd.0, osd.1, osd.2, mds.a, client.a] +tasks: +- install: +- ceph: + log-ignorelist: + - overall HEALTH_ + - \(MGR_DOWN\) + - \(PG_ + - \(OSD_ + - \(OBJECT_ + - \(OSDMAP_FLAGS\) +- exec: + mon.a: + - ceph restful create-key admin + - ceph restful create-self-signed-cert + - ceph restful restart +- workunit: + clients: + client.a: + - rest/test-restful.sh +- exec: + mon.a: + - ceph restful delete-key admin + - ceph restful list-keys | jq ".admin" | grep null + diff --git a/qa/suites/rados/rest/supported-random-distro$ b/qa/suites/rados/rest/supported-random-distro$ new file mode 120000 index 000000000..7cef21eef --- /dev/null +++ b/qa/suites/rados/rest/supported-random-distro$ @@ -0,0 +1 @@ +../basic/supported-random-distro$ \ No newline at end of file diff --git a/qa/suites/rados/rook b/qa/suites/rados/rook new file mode 120000 index 000000000..1249cc92d --- /dev/null +++ b/qa/suites/rados/rook @@ -0,0 +1 @@ +../orch/rook \ No newline at end of file diff --git a/qa/suites/rados/singleton-bluestore/% b/qa/suites/rados/singleton-bluestore/% new file mode 100644 index 000000000..e69de29bb diff --git a/qa/suites/rados/singleton-bluestore/.qa b/qa/suites/rados/singleton-bluestore/.qa new file mode 120000 index 000000000..a602a0353 --- /dev/null +++ b/qa/suites/rados/singleton-bluestore/.qa @@ -0,0 +1 @@ +../.qa/ \ No newline at end of file diff --git a/qa/suites/rados/singleton-bluestore/all/.qa b/qa/suites/rados/singleton-bluestore/all/.qa new file mode 120000 index 000000000..a602a0353 --- /dev/null +++ b/qa/suites/rados/singleton-bluestore/all/.qa @@ -0,0 +1 @@ +../.qa/ \ No newline at end of file diff --git a/qa/suites/rados/singleton-bluestore/all/cephtool.yaml b/qa/suites/rados/singleton-bluestore/all/cephtool.yaml new file mode 100644 index 000000000..f86be3459 --- /dev/null +++ b/qa/suites/rados/singleton-bluestore/all/cephtool.yaml @@ -0,0 +1,48 @@ +roles: +- - mon.a + - mon.b + - mon.c + - mgr.x + - osd.0 + - osd.1 + - osd.2 + - client.0 +openstack: + - volumes: # attached to each instance + count: 3 + size: 10 # GB +tasks: +- install: +- ceph: + log-ignorelist: + - but it is still running + - had wrong client addr + - had wrong cluster addr + - must scrub before tier agent can activate + - failsafe engaged, dropping updates + - failsafe disengaged, no longer dropping updates + - overall HEALTH_ + - \(OSDMAP_FLAGS\) + - \(OSD_ + - \(PG_ + - \(SMALLER_PG_NUM\) + - \(SMALLER_PGP_NUM\) + - \(CACHE_POOL_NO_HIT_SET\) + - \(CACHE_POOL_NEAR_FULL\) + - \(FS_WITH_FAILED_MDS\) + - \(FS_DEGRADED\) + - \(POOL_BACKFILLFULL\) + - \(POOL_FULL\) + - \(SMALLER_PGP_NUM\) + - \(POOL_NEARFULL\) + - \(POOL_APP_NOT_ENABLED\) + - \(AUTH_BAD_CAPS\) + - \(FS_INLINE_DATA_DEPRECATED\) + - \(MON_DOWN\) + - \(SLOW_OPS\) + - slow request +- workunit: + clients: + all: + - cephtool + - mon/pool_ops.sh diff --git a/qa/suites/rados/singleton-bluestore/mon_election b/qa/suites/rados/singleton-bluestore/mon_election new file mode 120000 index 000000000..3f331e621 --- /dev/null +++ b/qa/suites/rados/singleton-bluestore/mon_election @@ -0,0 +1 @@ +.qa/mon_election \ No newline at end of file diff --git a/qa/suites/rados/singleton-bluestore/msgr b/qa/suites/rados/singleton-bluestore/msgr new file mode 120000 index 000000000..57bee80db --- /dev/null +++ b/qa/suites/rados/singleton-bluestore/msgr @@ -0,0 +1 @@ +.qa/msgr \ No newline at end of file diff --git a/qa/suites/rados/singleton-bluestore/msgr-failures b/qa/suites/rados/singleton-bluestore/msgr-failures new file mode 120000 index 000000000..3ded97b94 --- /dev/null +++ b/qa/suites/rados/singleton-bluestore/msgr-failures @@ -0,0 +1 @@ +../singleton/msgr-failures \ No newline at end of file diff --git a/qa/suites/rados/singleton-bluestore/objectstore/.qa b/qa/suites/rados/singleton-bluestore/objectstore/.qa new file mode 120000 index 000000000..a602a0353 --- /dev/null +++ b/qa/suites/rados/singleton-bluestore/objectstore/.qa @@ -0,0 +1 @@ +../.qa/ \ No newline at end of file diff --git a/qa/suites/rados/singleton-bluestore/objectstore/bluestore-bitmap.yaml b/qa/suites/rados/singleton-bluestore/objectstore/bluestore-bitmap.yaml new file mode 120000 index 000000000..66cf2bc75 --- /dev/null +++ b/qa/suites/rados/singleton-bluestore/objectstore/bluestore-bitmap.yaml @@ -0,0 +1 @@ +.qa/objectstore_debug/bluestore-bitmap.yaml \ No newline at end of file diff --git a/qa/suites/rados/singleton-bluestore/objectstore/bluestore-comp-lz4.yaml b/qa/suites/rados/singleton-bluestore/objectstore/bluestore-comp-lz4.yaml new file mode 120000 index 000000000..da2e2598c --- /dev/null +++ b/qa/suites/rados/singleton-bluestore/objectstore/bluestore-comp-lz4.yaml @@ -0,0 +1 @@ +.qa/objectstore_debug/bluestore-comp-lz4.yaml \ No newline at end of file diff --git a/qa/suites/rados/singleton-bluestore/objectstore/bluestore-comp-snappy.yaml b/qa/suites/rados/singleton-bluestore/objectstore/bluestore-comp-snappy.yaml new file mode 120000 index 000000000..f75b0e1b4 --- /dev/null +++ b/qa/suites/rados/singleton-bluestore/objectstore/bluestore-comp-snappy.yaml @@ -0,0 +1 @@ +.qa/objectstore_debug/bluestore-comp-snappy.yaml \ No newline at end of file diff --git a/qa/suites/rados/singleton-bluestore/rados.yaml b/qa/suites/rados/singleton-bluestore/rados.yaml new file mode 120000 index 000000000..d256979c0 --- /dev/null +++ b/qa/suites/rados/singleton-bluestore/rados.yaml @@ -0,0 +1 @@ +.qa/config/rados.yaml \ No newline at end of file diff --git a/qa/suites/rados/singleton-bluestore/supported-random-distro$ b/qa/suites/rados/singleton-bluestore/supported-random-distro$ new file mode 120000 index 000000000..7cef21eef --- /dev/null +++ b/qa/suites/rados/singleton-bluestore/supported-random-distro$ @@ -0,0 +1 @@ +../basic/supported-random-distro$ \ No newline at end of file diff --git a/qa/suites/rados/singleton-nomsgr/% b/qa/suites/rados/singleton-nomsgr/% new file mode 100644 index 000000000..e69de29bb diff --git a/qa/suites/rados/singleton-nomsgr/.qa b/qa/suites/rados/singleton-nomsgr/.qa new file mode 120000 index 000000000..a602a0353 --- /dev/null +++ b/qa/suites/rados/singleton-nomsgr/.qa @@ -0,0 +1 @@ +../.qa/ \ No newline at end of file diff --git a/qa/suites/rados/singleton-nomsgr/all/.qa b/qa/suites/rados/singleton-nomsgr/all/.qa new file mode 120000 index 000000000..a602a0353 --- /dev/null +++ b/qa/suites/rados/singleton-nomsgr/all/.qa @@ -0,0 +1 @@ +../.qa/ \ No newline at end of file diff --git a/qa/suites/rados/singleton-nomsgr/all/admin_socket_output.yaml b/qa/suites/rados/singleton-nomsgr/all/admin_socket_output.yaml new file mode 100644 index 000000000..04c40197a --- /dev/null +++ b/qa/suites/rados/singleton-nomsgr/all/admin_socket_output.yaml @@ -0,0 +1,27 @@ +openstack: + - volumes: # attached to each instance + count: 2 + size: 10 # GB +roles: +- [mon.a, mds.a, mgr.x, osd.0, osd.1, client.0] +overrides: + ceph: + log-ignorelist: + - MDS in read-only mode + - force file system read-only + - overall HEALTH_ + - \(FS_DEGRADED\) + - \(OSDMAP_FLAGS\) + - \(OSD_FULL\) + - \(MDS_READ_ONLY\) + - \(POOL_FULL\) +tasks: +- install: +- ceph: + pre-mgr-commands: + - sudo ceph config set mgr mgr/devicehealth/enable_monitoring false --force +- rgw: + - client.0 +- exec: + client.0: + - ceph_test_admin_socket_output --all diff --git a/qa/suites/rados/singleton-nomsgr/all/balancer.yaml b/qa/suites/rados/singleton-nomsgr/all/balancer.yaml new file mode 100644 index 000000000..d4c6e3ca5 --- /dev/null +++ b/qa/suites/rados/singleton-nomsgr/all/balancer.yaml @@ -0,0 +1,14 @@ +roles: +- [mon.a, mgr.x, osd.0, osd.1, osd.2, client.0] +tasks: +- install: +- ceph: + pre-mgr-commands: + - sudo ceph config set mgr mgr/devicehealth/enable_monitoring false --force + fs: xfs + log-ignorelist: + - \(PG_AVAILABILITY\) +- cram: + clients: + client.0: + - src/test/cli-integration/balancer/misplaced.t diff --git a/qa/suites/rados/singleton-nomsgr/all/cache-fs-trunc.yaml b/qa/suites/rados/singleton-nomsgr/all/cache-fs-trunc.yaml new file mode 100644 index 000000000..f998c51c9 --- /dev/null +++ b/qa/suites/rados/singleton-nomsgr/all/cache-fs-trunc.yaml @@ -0,0 +1,54 @@ +openstack: + - volumes: # attached to each instance + count: 3 + size: 10 # GB +roles: +- [mon.a, mgr.x, mds.a, osd.0, osd.1, osd.2, client.0, client.1] +tasks: +- install: +- ceph: + pre-mgr-commands: + - sudo ceph config set mgr mgr/devicehealth/enable_monitoring false --force + log-ignorelist: + - overall HEALTH_ + - \(CACHE_POOL_NO_HIT_SET\) + conf: + global: + osd max object name len: 460 + osd max object namespace len: 64 + debug client: 20 + debug mds: 20 + debug ms: 1 +- exec: + client.0: + - ceph osd pool create data_cache 4 + - ceph osd tier add cephfs_data data_cache + - ceph osd tier cache-mode data_cache writeback + - ceph osd tier set-overlay cephfs_data data_cache + - ceph osd pool set data_cache hit_set_type bloom + - ceph osd pool set data_cache hit_set_count 8 + - ceph osd pool set data_cache hit_set_period 3600 + - ceph osd pool set data_cache min_read_recency_for_promote 0 +- ceph-fuse: +- exec: + client.0: + - sudo chmod 777 $TESTDIR/mnt.0/ + - dd if=/dev/urandom of=$TESTDIR/mnt.0/foo bs=1M count=5 + - ls -al $TESTDIR/mnt.0/foo + - truncate --size 0 $TESTDIR/mnt.0/foo + - ls -al $TESTDIR/mnt.0/foo + - dd if=/dev/urandom of=$TESTDIR/mnt.0/foo bs=1M count=5 + - ls -al $TESTDIR/mnt.0/foo + - cp $TESTDIR/mnt.0/foo /tmp/foo + - sync + - rados -p data_cache ls - + - sleep 10 + - rados -p data_cache ls - + - rados -p data_cache cache-flush-evict-all + - rados -p data_cache ls - + - sleep 1 +- exec: + client.1: + - hexdump -C /tmp/foo | head + - hexdump -C $TESTDIR/mnt.1/foo | head + - cmp $TESTDIR/mnt.1/foo /tmp/foo diff --git a/qa/suites/rados/singleton-nomsgr/all/ceph-kvstore-tool.yaml b/qa/suites/rados/singleton-nomsgr/all/ceph-kvstore-tool.yaml new file mode 100644 index 000000000..e116b5ae0 --- /dev/null +++ b/qa/suites/rados/singleton-nomsgr/all/ceph-kvstore-tool.yaml @@ -0,0 +1,23 @@ +openstack: + - volumes: # attached to each instance + count: 3 + size: 10 # GB +roles: +- [mon.a, mgr.x, osd.0, osd.1, osd.2, client.0] + +overrides: + ceph: + pre-mgr-commands: + - sudo ceph config set mgr mgr/devicehealth/enable_monitoring false --force + log-ignorelist: + - but it is still running + - overall HEALTH_ + - \(POOL_APP_NOT_ENABLED\) + +tasks: +- install: +- ceph: +- workunit: + clients: + all: + - cephtool/test_kvstore_tool.sh diff --git a/qa/suites/rados/singleton-nomsgr/all/ceph-post-file.yaml b/qa/suites/rados/singleton-nomsgr/all/ceph-post-file.yaml new file mode 100644 index 000000000..530dc42a7 --- /dev/null +++ b/qa/suites/rados/singleton-nomsgr/all/ceph-post-file.yaml @@ -0,0 +1,12 @@ +openstack: + - volumes: # attached to each instance + count: 3 + size: 10 # GB +roles: +- [mon.a, mgr.x, osd.0, osd.1, osd.2, client.0] +tasks: +- install: +- workunit: + clients: + all: + - post-file.sh diff --git a/qa/suites/rados/singleton-nomsgr/all/export-after-evict.yaml b/qa/suites/rados/singleton-nomsgr/all/export-after-evict.yaml new file mode 100644 index 000000000..ee800e5a7 --- /dev/null +++ b/qa/suites/rados/singleton-nomsgr/all/export-after-evict.yaml @@ -0,0 +1,40 @@ +openstack: + - volumes: # attached to each instance + count: 3 + size: 10 # GB +roles: +- - mon.a + - mgr.x + - osd.0 + - osd.1 + - osd.2 + - client.0 +tasks: +- install: +- ceph: + pre-mgr-commands: + - sudo ceph config set mgr mgr/devicehealth/enable_monitoring false --force + log-ignorelist: + - overall HEALTH_ + - \(CACHE_POOL_NO_HIT_SET\) + conf: + global: + osd max object name len: 460 + osd max object namespace len: 64 +- exec: + client.0: + - ceph osd pool create base-pool 4 + - ceph osd pool application enable base-pool rados + - ceph osd pool create cache-pool 4 + - ceph osd tier add base-pool cache-pool + - ceph osd tier cache-mode cache-pool writeback + - ceph osd tier set-overlay base-pool cache-pool + - dd if=/dev/urandom of=$TESTDIR/foo bs=1M count=1 + - rbd import --image-format 2 $TESTDIR/foo base-pool/bar + - rbd snap create base-pool/bar@snap + - rados -p base-pool cache-flush-evict-all + - rbd export base-pool/bar $TESTDIR/bar + - rbd export base-pool/bar@snap $TESTDIR/snap + - cmp $TESTDIR/foo $TESTDIR/bar + - cmp $TESTDIR/foo $TESTDIR/snap + - rm $TESTDIR/foo $TESTDIR/bar $TESTDIR/snap diff --git a/qa/suites/rados/singleton-nomsgr/all/full-tiering.yaml b/qa/suites/rados/singleton-nomsgr/all/full-tiering.yaml new file mode 100644 index 000000000..8d26cd323 --- /dev/null +++ b/qa/suites/rados/singleton-nomsgr/all/full-tiering.yaml @@ -0,0 +1,40 @@ +# verify #13098 fix +openstack: + - volumes: # attached to each instance + count: 3 + size: 10 # GB +roles: +- [mon.a, mgr.x, osd.0, osd.1, osd.2, client.0] +overrides: + ceph: + log-ignorelist: + - is full + - overall HEALTH_ + - \(POOL_FULL\) + - \(POOL_NEAR_FULL\) + - \(CACHE_POOL_NO_HIT_SET\) + - \(CACHE_POOL_NEAR_FULL\) +tasks: +- install: +- ceph: + pre-mgr-commands: + - sudo ceph config set mgr mgr/devicehealth/enable_monitoring false --force + conf: + global: + osd max object name len: 460 + osd max object namespace len: 64 +- exec: + client.0: + - ceph osd pool create ec-ca 1 1 + - ceph osd pool create ec 1 1 erasure default + - ceph osd pool application enable ec rados + - ceph osd tier add ec ec-ca + - ceph osd tier cache-mode ec-ca readproxy + - ceph osd tier set-overlay ec ec-ca + - ceph osd pool set ec-ca hit_set_type bloom + - ceph osd pool set-quota ec-ca max_bytes 20480000 + - ceph osd pool set-quota ec max_bytes 20480000 + - ceph osd pool set ec-ca target_max_bytes 20480000 + - timeout 30 rados -p ec-ca bench 30 write || true + - ceph osd pool set-quota ec-ca max_bytes 0 + - ceph osd pool set-quota ec max_bytes 0 diff --git a/qa/suites/rados/singleton-nomsgr/all/health-warnings.yaml b/qa/suites/rados/singleton-nomsgr/all/health-warnings.yaml new file mode 100644 index 000000000..bc57a9cd9 --- /dev/null +++ b/qa/suites/rados/singleton-nomsgr/all/health-warnings.yaml @@ -0,0 +1,22 @@ +roles: +- [mon.a, mgr.x, osd.0, osd.1, osd.2, osd.3, osd.4, osd.5, osd.6, osd.7, osd.8, osd.9, client.0] +tasks: +- install: +- ceph: + pre-mgr-commands: + - sudo ceph config set mgr mgr/devicehealth/enable_monitoring false --force + conf: + osd: +# we may land on ext4 + osd max object name len: 400 + osd max object namespace len: 64 + log-ignorelist: + - but it is still running + - overall HEALTH_ + - \(OSDMAP_FLAGS\) + - \(OSD_ + - \(PG_ +- workunit: + clients: + all: + - rados/test_health_warnings.sh diff --git a/qa/suites/rados/singleton-nomsgr/all/large-omap-object-warnings.yaml b/qa/suites/rados/singleton-nomsgr/all/large-omap-object-warnings.yaml new file mode 100644 index 000000000..b08ab343f --- /dev/null +++ b/qa/suites/rados/singleton-nomsgr/all/large-omap-object-warnings.yaml @@ -0,0 +1,30 @@ +openstack: + - volumes: # attached to each instance + count: 2 + size: 10 # GB +roles: +- [mon.a, mgr.x, osd.0, osd.1, client.0] +overrides: + ceph: + pre-mgr-commands: + - sudo ceph config set mgr mgr/devicehealth/enable_monitoring false --force + log-ignorelist: + - \(OSDMAP_FLAGS\) + - \(OSD_FULL\) + - \(MDS_READ_ONLY\) + - \(POOL_APP_NOT_ENABLED\) + - large omap objects + - Large omap object found + - application not enabled + conf: + osd: + osd scrub backoff ratio: 0 + osd deep scrub large omap object value sum threshold: 8800000 + osd deep scrub large omap object key threshold: 20000 +tasks: +- install: +- ceph: +- workunit: + clients: + all: + - rados/test_large_omap_detection.py diff --git a/qa/suites/rados/singleton-nomsgr/all/lazy_omap_stats_output.yaml b/qa/suites/rados/singleton-nomsgr/all/lazy_omap_stats_output.yaml new file mode 100644 index 000000000..7228522be --- /dev/null +++ b/qa/suites/rados/singleton-nomsgr/all/lazy_omap_stats_output.yaml @@ -0,0 +1,18 @@ +openstack: + - volumes: # attached to each instance + count: 2 + size: 10 # GB +roles: +- [mon.a, mgr.x, osd.0, osd.1, osd.2, client.0] +overrides: + ceph: + log-ignorelist: + - \(POOL_APP_NOT_ENABLED\) +tasks: +- install: +- ceph: + pre-mgr-commands: + - sudo ceph config set mgr mgr/devicehealth/enable_monitoring false --force +- exec: + client.0: + - ceph_test_lazy_omap_stats diff --git a/qa/suites/rados/singleton-nomsgr/all/librados_hello_world.yaml b/qa/suites/rados/singleton-nomsgr/all/librados_hello_world.yaml new file mode 100644 index 000000000..f670a0849 --- /dev/null +++ b/qa/suites/rados/singleton-nomsgr/all/librados_hello_world.yaml @@ -0,0 +1,24 @@ +roles: +- [mon.a, mds.a, mgr.x, osd.0, osd.1, client.0] +overrides: + ceph: + log-ignorelist: + - \(POOL_APP_NOT_ENABLED\) +tasks: +- install: + extra_packages: + deb: + - libradosstriper-dev + - librados-dev + - libradospp-dev + rpm: + - libradosstriper-devel + - librados-devel + - libradospp-devel +- ceph: + pre-mgr-commands: + - sudo ceph config set mgr mgr/devicehealth/enable_monitoring false --force +- workunit: + clients: + all: + - rados/test_librados_build.sh diff --git a/qa/suites/rados/singleton-nomsgr/all/msgr.yaml b/qa/suites/rados/singleton-nomsgr/all/msgr.yaml new file mode 100644 index 000000000..d1852ae2b --- /dev/null +++ b/qa/suites/rados/singleton-nomsgr/all/msgr.yaml @@ -0,0 +1,23 @@ +roles: +- [mon.a, mgr.x, osd.0, osd.1, client.0] +tasks: +- install: +- exec: + client.0: + - ceph_test_async_driver + - ceph_test_msgr +openstack: + - machine: + disk: 40 # GB + ram: 15000 # MB + cpus: 1 + volumes: # attached to each instance + count: 0 + size: 1 # GB +overrides: + ceph: + pre-mgr-commands: + - sudo ceph config set mgr mgr/devicehealth/enable_monitoring false --force + conf: + client: + debug ms: 20 diff --git a/qa/suites/rados/singleton-nomsgr/all/multi-backfill-reject.yaml b/qa/suites/rados/singleton-nomsgr/all/multi-backfill-reject.yaml new file mode 100644 index 000000000..a3ce46e6a --- /dev/null +++ b/qa/suites/rados/singleton-nomsgr/all/multi-backfill-reject.yaml @@ -0,0 +1,50 @@ +openstack: + - volumes: # attached to each instance + count: 3 + size: 10 # GB +roles: +- - mon.a + - mgr.x + - osd.0 + - osd.1 + - osd.2 + - client.0 +- - osd.3 + - osd.4 + - osd.5 +tasks: +- install: +- ceph: + pre-mgr-commands: + - sudo ceph config set mgr mgr/devicehealth/enable_monitoring false --force + log-ignorelist: + - overall HEALTH_ + - \(PG_ + - \(OSD_ + - \(OBJECT_ + conf: + osd: + osd debug reject backfill probability: .3 + osd min pg log entries: 25 + osd max pg log entries: 100 + osd max object name len: 460 + osd max object namespace len: 64 +- exec: + client.0: + - sudo ceph osd pool create foo 64 + - sudo ceph osd pool application enable foo rados + - rados -p foo bench 60 write -b 1024 --no-cleanup + - sudo ceph osd pool set foo size 3 + - sudo ceph osd out 0 1 +- sleep: + duration: 60 +- exec: + client.0: + - sudo ceph osd in 0 1 +- sleep: + duration: 60 +- exec: + client.0: + - sudo ceph osd pool set foo size 2 +- sleep: + duration: 300 diff --git a/qa/suites/rados/singleton-nomsgr/all/osd_stale_reads.yaml b/qa/suites/rados/singleton-nomsgr/all/osd_stale_reads.yaml new file mode 100644 index 000000000..408268a09 --- /dev/null +++ b/qa/suites/rados/singleton-nomsgr/all/osd_stale_reads.yaml @@ -0,0 +1,29 @@ +openstack: + - volumes: # attached to each instance + count: 3 + size: 10 # GB +roles: +- [mon.a, mgr.x, osd.0, osd.1, osd.2, client.0] +overrides: + ceph: + log-ignorelist: + - \(OSD_DOWN\) + - \(POOL_APP_NOT_ENABLED\) + - \(SLOW_OPS\) + - \(PG_AVAILABILITY\) + - \(PG_DEGRADED\) + - application not enabled + - slow request + conf: + osd: + osd scrub backoff ratio: 0 + osd deep scrub large omap object value sum threshold: 8800000 + osd deep scrub large omap object key threshold: 20000 +tasks: +- install: +- ceph: + pre-mgr-commands: + - sudo ceph config set mgr mgr/devicehealth/enable_monitoring false --force +- exec: + client.0: + - ceph_test_osd_stale_read diff --git a/qa/suites/rados/singleton-nomsgr/all/pool-access.yaml b/qa/suites/rados/singleton-nomsgr/all/pool-access.yaml new file mode 100644 index 000000000..6485a6871 --- /dev/null +++ b/qa/suites/rados/singleton-nomsgr/all/pool-access.yaml @@ -0,0 +1,15 @@ +openstack: + - volumes: # attached to each instance + count: 2 + size: 10 # GB +roles: +- [mon.a, mgr.x, osd.0, osd.1, client.0] +tasks: +- install: +- ceph: + pre-mgr-commands: + - sudo ceph config set mgr mgr/devicehealth/enable_monitoring false --force +- workunit: + clients: + all: + - rados/test_pool_access.sh diff --git a/qa/suites/rados/singleton-nomsgr/all/recovery-unfound-found.yaml b/qa/suites/rados/singleton-nomsgr/all/recovery-unfound-found.yaml new file mode 100644 index 000000000..9cf4eec89 --- /dev/null +++ b/qa/suites/rados/singleton-nomsgr/all/recovery-unfound-found.yaml @@ -0,0 +1,60 @@ +roles: +- - mon.a + - mon.b + - mon.c + - mgr.x + - osd.0 + - osd.1 +openstack: + - volumes: # attached to each instance + count: 2 + size: 20 # GB +tasks: +- install: +- ceph: + pre-mgr-commands: + - sudo ceph config set mgr mgr/devicehealth/enable_monitoring false --force + fs: xfs + conf: + osd: + osd recovery sleep: .1 + osd objectstore: filestore + log-ignorelist: + - \(POOL_APP_NOT_ENABLED\) + - \(OSDMAP_FLAGS\) + - \(OSD_ + - \(OBJECT_ + - \(PG_ + - overall HEALTH +- exec: + osd.0: + - ceph osd pool create foo 32 + - ceph osd pool application enable foo foo + - rados -p foo bench 30 write -b 4096 --no-cleanup + - ceph osd set noup +- ceph.restart: + daemons: [osd.0] + wait-for-up: false + wait-for-healthy: false +- exec: + osd.0: + - sleep 5 + - rados -p foo bench 3 write -b 4096 --no-cleanup + - ceph osd unset noup + - sleep 10 + - ceph osd set noup +- ceph.restart: + daemons: [osd.1] + wait-for-up: false + wait-for-healthy: false +- exec: + osd.0: + - ceph osd out 0 + - sleep 10 + - ceph osd unset noup +- ceph.healthy: + wait-for-healthy: false # only wait for osds up and pgs clean, ignore misplaced +- exec: + osd.0: + - ceph osd in 0 +- ceph.healthy: diff --git a/qa/suites/rados/singleton-nomsgr/all/version-number-sanity.yaml b/qa/suites/rados/singleton-nomsgr/all/version-number-sanity.yaml new file mode 100644 index 000000000..6d48796f0 --- /dev/null +++ b/qa/suites/rados/singleton-nomsgr/all/version-number-sanity.yaml @@ -0,0 +1,15 @@ +roles: +- [mon.a, mds.a, mgr.x, osd.0, osd.1, client.0] +overrides: + ceph: + log-ignorelist: + - \(POOL_APP_NOT_ENABLED\) +tasks: +- install: +- ceph: + pre-mgr-commands: + - sudo ceph config set mgr mgr/devicehealth/enable_monitoring false --force +- workunit: + clients: + all: + - rados/version_number_sanity.sh diff --git a/qa/suites/rados/singleton-nomsgr/mon_election b/qa/suites/rados/singleton-nomsgr/mon_election new file mode 120000 index 000000000..3f331e621 --- /dev/null +++ b/qa/suites/rados/singleton-nomsgr/mon_election @@ -0,0 +1 @@ +.qa/mon_election \ No newline at end of file diff --git a/qa/suites/rados/singleton-nomsgr/rados.yaml b/qa/suites/rados/singleton-nomsgr/rados.yaml new file mode 120000 index 000000000..d256979c0 --- /dev/null +++ b/qa/suites/rados/singleton-nomsgr/rados.yaml @@ -0,0 +1 @@ +.qa/config/rados.yaml \ No newline at end of file diff --git a/qa/suites/rados/singleton-nomsgr/supported-random-distro$ b/qa/suites/rados/singleton-nomsgr/supported-random-distro$ new file mode 120000 index 000000000..7cef21eef --- /dev/null +++ b/qa/suites/rados/singleton-nomsgr/supported-random-distro$ @@ -0,0 +1 @@ +../basic/supported-random-distro$ \ No newline at end of file diff --git a/qa/suites/rados/singleton/% b/qa/suites/rados/singleton/% new file mode 100644 index 000000000..e69de29bb diff --git a/qa/suites/rados/singleton/.qa b/qa/suites/rados/singleton/.qa new file mode 120000 index 000000000..a602a0353 --- /dev/null +++ b/qa/suites/rados/singleton/.qa @@ -0,0 +1 @@ +../.qa/ \ No newline at end of file diff --git a/qa/suites/rados/singleton/all/.qa b/qa/suites/rados/singleton/all/.qa new file mode 120000 index 000000000..a602a0353 --- /dev/null +++ b/qa/suites/rados/singleton/all/.qa @@ -0,0 +1 @@ +../.qa/ \ No newline at end of file diff --git a/qa/suites/rados/singleton/all/admin-socket.yaml b/qa/suites/rados/singleton/all/admin-socket.yaml new file mode 100644 index 000000000..76bee411c --- /dev/null +++ b/qa/suites/rados/singleton/all/admin-socket.yaml @@ -0,0 +1,26 @@ +roles: +- - mon.a + - mgr.x + - osd.0 + - osd.1 + - client.a +openstack: + - volumes: # attached to each instance + count: 2 + size: 10 # GB +tasks: +- install: +- ceph: +- admin_socket: + osd.0: + version: + git_version: + help: + config show: + config help: + config set bluestore_csum_type xxhash64: + perf dump: + perf schema: + get_heap_property tcmalloc.max_total_thread_cache_byte: + set_heap_property tcmalloc.max_total_thread_cache_bytes 67108864: + set_heap_property tcmalloc.max_total_thread_cache_bytes 33554432: diff --git a/qa/suites/rados/singleton/all/backfill-toofull.yaml b/qa/suites/rados/singleton/all/backfill-toofull.yaml new file mode 100644 index 000000000..fcc3d0e29 --- /dev/null +++ b/qa/suites/rados/singleton/all/backfill-toofull.yaml @@ -0,0 +1,37 @@ +roles: +- - mon.a + - mon.b + - mon.c + - mgr.x + - osd.0 + - osd.1 + - osd.2 + - osd.3 +openstack: + - volumes: # attached to each instance + count: 4 + size: 10 # GB +tasks: +- install: +- ceph: + create_rbd_pool: false + pre-mgr-commands: + - sudo ceph config set mgr mgr/devicehealth/enable_monitoring false --force + log-ignorelist: + - Error + - overall HEALTH_ + - \(OBJECT_ + - \(OSDMAP_FLAGS\) + - \(OSD_ + - \(PG_ + - \(POOL_BACKFILLFULL\) + - \(POOL_NEARFULL\) + - \(SLOW_OPS\) + - \(TOO_FEW_PGS\) + - Monitor daemon marked osd\.[[:digit:]]+ down, but it is still running + - slow request + conf: + osd: + osd min pg log entries: 5 + osd max pg log entries: 5 +- backfill_toofull: diff --git a/qa/suites/rados/singleton/all/deduptool.yaml b/qa/suites/rados/singleton/all/deduptool.yaml new file mode 100644 index 000000000..616a0b33c --- /dev/null +++ b/qa/suites/rados/singleton/all/deduptool.yaml @@ -0,0 +1,28 @@ +roles: +- - mon.a + - mgr.x + - osd.0 + - osd.1 + - osd.2 + - client.0 +openstack: + - volumes: # attached to each instance + count: 2 + size: 10 # GB +tasks: +- install: +- ceph: + pre-mgr-commands: + - sudo ceph config set mgr mgr/devicehealth/enable_monitoring false --force + log-ignorelist: + - but it is still running + - had wrong client addr + - had wrong cluster addr + - reached quota + - overall HEALTH_ + - \(POOL_FULL\) + - \(POOL_APP_NOT_ENABLED\) +- workunit: + clients: + all: + - rados/test_dedup_tool.sh diff --git a/qa/suites/rados/singleton/all/divergent_priors.yaml b/qa/suites/rados/singleton/all/divergent_priors.yaml new file mode 100644 index 000000000..24b42557f --- /dev/null +++ b/qa/suites/rados/singleton/all/divergent_priors.yaml @@ -0,0 +1,28 @@ +roles: +- - mon.a + - mgr.x + - osd.0 + - osd.1 + - osd.2 + - client.0 +openstack: + - volumes: # attached to each instance + count: 3 + size: 10 # GB + +overrides: + ceph: + log-ignorelist: + - overall HEALTH_ + - \(OSDMAP_FLAGS\) + - \(OSD_ + - \(PG_ + - \(OBJECT_ + - \(POOL_APP_NOT_ENABLED\) + +tasks: +- install: +- ceph: + pre-mgr-commands: + - sudo ceph config set mgr mgr/devicehealth/enable_monitoring false --force +- divergent_priors: diff --git a/qa/suites/rados/singleton/all/divergent_priors2.yaml b/qa/suites/rados/singleton/all/divergent_priors2.yaml new file mode 100644 index 000000000..6bef63958 --- /dev/null +++ b/qa/suites/rados/singleton/all/divergent_priors2.yaml @@ -0,0 +1,28 @@ +roles: +- - mon.a + - mgr.x + - osd.0 + - osd.1 + - osd.2 + - client.0 +openstack: + - volumes: # attached to each instance + count: 3 + size: 10 # GB + +overrides: + ceph: + log-ignorelist: + - overall HEALTH_ + - \(OSDMAP_FLAGS\) + - \(OSD_ + - \(PG_ + - \(OBJECT_ + - \(POOL_APP_NOT_ENABLED\) + +tasks: +- install: +- ceph: + pre-mgr-commands: + - sudo ceph config set mgr mgr/devicehealth/enable_monitoring false --force +- divergent_priors2: diff --git a/qa/suites/rados/singleton/all/dump-stuck.yaml b/qa/suites/rados/singleton/all/dump-stuck.yaml new file mode 100644 index 000000000..c1d28ee8e --- /dev/null +++ b/qa/suites/rados/singleton/all/dump-stuck.yaml @@ -0,0 +1,21 @@ +roles: +- - mon.a + - mgr.x + - osd.0 + - osd.1 +openstack: + - volumes: # attached to each instance + count: 2 + size: 10 # GB +tasks: +- install: +- ceph: + pre-mgr-commands: + - sudo ceph config set mgr mgr/devicehealth/enable_monitoring false --force + log-ignorelist: + - but it is still running + - overall HEALTH_ + - \(OSDMAP_FLAGS\) + - \(OSD_ + - \(PG_ +- dump_stuck: diff --git a/qa/suites/rados/singleton/all/ec-inconsistent-hinfo.yaml b/qa/suites/rados/singleton/all/ec-inconsistent-hinfo.yaml new file mode 100644 index 000000000..d71eab149 --- /dev/null +++ b/qa/suites/rados/singleton/all/ec-inconsistent-hinfo.yaml @@ -0,0 +1,36 @@ +roles: +- - mon.a + - mon.b + - mon.c + - mgr.x + - osd.0 + - osd.1 + - osd.2 + - osd.3 +openstack: + - volumes: # attached to each instance + count: 4 + size: 10 # GB +tasks: +- install: +- ceph: + create_rbd_pool: false + pre-mgr-commands: + - sudo ceph config set mgr mgr/devicehealth/enable_monitoring false --force + log-ignorelist: + - \(OBJECT_ + - \(OSDMAP_FLAGS\) + - \(OSD_ + - \(PG_ + - \(SLOW_OPS\) + - deep-scrub + - missing + - overall HEALTH_ + - repair + - slow request + - unfound + conf: + osd: + osd min pg log entries: 5 + osd max pg log entries: 5 +- ec_inconsistent_hinfo: diff --git a/qa/suites/rados/singleton/all/ec-lost-unfound.yaml b/qa/suites/rados/singleton/all/ec-lost-unfound.yaml new file mode 100644 index 000000000..9c423c8d8 --- /dev/null +++ b/qa/suites/rados/singleton/all/ec-lost-unfound.yaml @@ -0,0 +1,29 @@ +roles: +- - mon.a + - mon.b + - mon.c + - mgr.x + - osd.0 + - osd.1 + - osd.2 + - osd.3 +openstack: + - volumes: # attached to each instance + count: 4 + size: 10 # GB +tasks: +- install: +- ceph: + create_rbd_pool: false + pre-mgr-commands: + - sudo ceph config set mgr mgr/devicehealth/enable_monitoring false --force + log-ignorelist: + - objects unfound and apparently lost + - overall HEALTH_ + - \(OSDMAP_FLAGS\) + - \(OSD_ + - \(PG_ + - \(OBJECT_ + - \(SLOW_OPS\) + - slow request +- ec_lost_unfound: diff --git a/qa/suites/rados/singleton/all/erasure-code-nonregression.yaml b/qa/suites/rados/singleton/all/erasure-code-nonregression.yaml new file mode 100644 index 000000000..e8201ee0b --- /dev/null +++ b/qa/suites/rados/singleton/all/erasure-code-nonregression.yaml @@ -0,0 +1,17 @@ +roles: +- - mon.a + - mgr.x + - osd.0 + - osd.1 + - osd.2 + - client.0 +openstack: + - volumes: # attached to each instance + count: 3 + size: 10 # GB +tasks: +- install: +- workunit: + clients: + all: + - erasure-code/encode-decode-non-regression.sh diff --git a/qa/suites/rados/singleton/all/lost-unfound-delete.yaml b/qa/suites/rados/singleton/all/lost-unfound-delete.yaml new file mode 100644 index 000000000..bb170b506 --- /dev/null +++ b/qa/suites/rados/singleton/all/lost-unfound-delete.yaml @@ -0,0 +1,27 @@ +roles: +- - mon.a + - mon.b + - mon.c + - mgr.x + - osd.0 + - osd.1 + - osd.2 +openstack: + - volumes: # attached to each instance + count: 3 + size: 10 # GB +tasks: +- install: +- ceph: + pre-mgr-commands: + - sudo ceph config set mgr mgr/devicehealth/enable_monitoring false --force + log-ignorelist: + - objects unfound and apparently lost + - overall HEALTH_ + - \(OSDMAP_FLAGS\) + - \(OSD_ + - \(PG_ + - \(OBJECT_ + - \(SLOW_OPS\) + - slow request +- rep_lost_unfound_delete: diff --git a/qa/suites/rados/singleton/all/lost-unfound.yaml b/qa/suites/rados/singleton/all/lost-unfound.yaml new file mode 100644 index 000000000..fceee20c0 --- /dev/null +++ b/qa/suites/rados/singleton/all/lost-unfound.yaml @@ -0,0 +1,27 @@ +roles: +- - mon.a + - mon.b + - mon.c + - mgr.x + - osd.0 + - osd.1 + - osd.2 +openstack: + - volumes: # attached to each instance + count: 3 + size: 10 # GB +tasks: +- install: +- ceph: + pre-mgr-commands: + - sudo ceph config set mgr mgr/devicehealth/enable_monitoring false --force + log-ignorelist: + - objects unfound and apparently lost + - overall HEALTH_ + - \(OSDMAP_FLAGS\) + - \(OSD_ + - \(PG_ + - \(OBJECT_ + - \(SLOW_OPS\) + - slow request +- lost_unfound: diff --git a/qa/suites/rados/singleton/all/max-pg-per-osd.from-mon.yaml b/qa/suites/rados/singleton/all/max-pg-per-osd.from-mon.yaml new file mode 100644 index 000000000..e5999bc9b --- /dev/null +++ b/qa/suites/rados/singleton/all/max-pg-per-osd.from-mon.yaml @@ -0,0 +1,29 @@ +roles: +- - mon.a + - mgr.x + - osd.0 + - osd.1 +openstack: + - volumes: # attached to each instance + count: 2 + size: 10 # GB +overrides: + ceph: + create_rbd_pool: False + pre-mgr-commands: + - sudo ceph config set mgr mgr/devicehealth/enable_monitoring false --force + conf: + mon: + osd pool default size: 2 + osd: + mon max pg per osd : 2 + osd max pg per osd hard ratio : 1 + log-ignorelist: + - \(TOO_FEW_PGS\) + - \(PENDING_CREATING_PGS\) +tasks: +- install: +- ceph: +- osd_max_pg_per_osd: + test_create_from_mon: True + pg_num: 2 diff --git a/qa/suites/rados/singleton/all/max-pg-per-osd.from-primary.yaml b/qa/suites/rados/singleton/all/max-pg-per-osd.from-primary.yaml new file mode 100644 index 000000000..075d6be1f --- /dev/null +++ b/qa/suites/rados/singleton/all/max-pg-per-osd.from-primary.yaml @@ -0,0 +1,34 @@ +roles: +- - mon.a + - mgr.x + - osd.0 + - osd.1 + - osd.2 + - osd.3 +openstack: + - volumes: # attached to each instance + count: 4 + size: 10 # GB +overrides: + ceph: + create_rbd_pool: False + pre-mgr-commands: + - sudo ceph config set mgr mgr/devicehealth/enable_monitoring false --force + conf: + mon: + osd pool default size: 2 + osd: + mon max pg per osd : 1 + osd max pg per osd hard ratio : 1 + log-ignorelist: + - \(TOO_FEW_PGS\) + - \(PG_ + - \(PENDING_CREATING_PGS\) +tasks: +- install: +- ceph: +- osd_max_pg_per_osd: + test_create_from_mon: False + pg_num: 1 + pool_size: 2 + from_primary: True diff --git a/qa/suites/rados/singleton/all/max-pg-per-osd.from-replica.yaml b/qa/suites/rados/singleton/all/max-pg-per-osd.from-replica.yaml new file mode 100644 index 000000000..db2856484 --- /dev/null +++ b/qa/suites/rados/singleton/all/max-pg-per-osd.from-replica.yaml @@ -0,0 +1,34 @@ +roles: +- - mon.a + - mgr.x + - osd.0 + - osd.1 + - osd.2 + - osd.3 +openstack: + - volumes: # attached to each instance + count: 4 + size: 10 # GB +overrides: + ceph: + create_rbd_pool: False + pre-mgr-commands: + - sudo ceph config set mgr mgr/devicehealth/enable_monitoring false --force + conf: + mon: + osd pool default size: 2 + osd: + mon max pg per osd : 1 + osd max pg per osd hard ratio : 1 + log-ignorelist: + - \(TOO_FEW_PGS\) + - \(PG_ + - \(PENDING_CREATING_PGS\) +tasks: +- install: +- ceph: +- osd_max_pg_per_osd: + test_create_from_mon: False + pg_num: 1 + pool_size: 2 + from_primary: False diff --git a/qa/suites/rados/singleton/all/mon-auth-caps.yaml b/qa/suites/rados/singleton/all/mon-auth-caps.yaml new file mode 100644 index 000000000..8c23c0bc9 --- /dev/null +++ b/qa/suites/rados/singleton/all/mon-auth-caps.yaml @@ -0,0 +1,19 @@ +roles: +- - mon.a + - mgr.x + - osd.0 + - osd.1 + - osd.2 + - client.0 +tasks: +- install: +- ceph: + pre-mgr-commands: + - sudo ceph config set mgr mgr/devicehealth/enable_monitoring false --force + log-ignorelist: + - overall HEALTH_ + - \(AUTH_BAD_CAPS\) +- workunit: + clients: + all: + - mon/auth_caps.sh diff --git a/qa/suites/rados/singleton/all/mon-config-key-caps.yaml b/qa/suites/rados/singleton/all/mon-config-key-caps.yaml new file mode 100644 index 000000000..f987f3c98 --- /dev/null +++ b/qa/suites/rados/singleton/all/mon-config-key-caps.yaml @@ -0,0 +1,19 @@ +roles: +- - mon.a + - mgr.x + - osd.0 + - osd.1 + - osd.2 + - client.0 +tasks: +- install: +- ceph: + pre-mgr-commands: + - sudo ceph config set mgr mgr/devicehealth/enable_monitoring false --force + log-ignorelist: + - overall HEALTH_ + - \(AUTH_BAD_CAPS\) +- workunit: + clients: + all: + - mon/test_config_key_caps.sh diff --git a/qa/suites/rados/singleton/all/mon-config-keys.yaml b/qa/suites/rados/singleton/all/mon-config-keys.yaml new file mode 100644 index 000000000..7d8b920cb --- /dev/null +++ b/qa/suites/rados/singleton/all/mon-config-keys.yaml @@ -0,0 +1,22 @@ +roles: +- - mon.a + - mon.b + - mon.c + - mgr.x + - osd.0 + - osd.1 + - osd.2 + - client.0 +openstack: + - volumes: # attached to each instance + count: 3 + size: 10 # GB +tasks: +- install: +- ceph: + pre-mgr-commands: + - sudo ceph config set mgr mgr/devicehealth/enable_monitoring false --force +- workunit: + clients: + all: + - mon/test_mon_config_key.py diff --git a/qa/suites/rados/singleton/all/mon-config.yaml b/qa/suites/rados/singleton/all/mon-config.yaml new file mode 100644 index 000000000..3627e17df --- /dev/null +++ b/qa/suites/rados/singleton/all/mon-config.yaml @@ -0,0 +1,22 @@ +roles: +- - mon.a + - mon.b + - mon.c + - mgr.x + - osd.0 + - osd.1 + - osd.2 + - client.0 +openstack: + - volumes: # attached to each instance + count: 3 + size: 10 # GB +tasks: +- install: +- ceph: + pre-mgr-commands: + - sudo ceph config set mgr mgr/devicehealth/enable_monitoring false --force +- workunit: + clients: + all: + - mon/config.sh diff --git a/qa/suites/rados/singleton/all/mon-memory-target-compliance.yaml.disabled b/qa/suites/rados/singleton/all/mon-memory-target-compliance.yaml.disabled new file mode 100644 index 000000000..120e073a7 --- /dev/null +++ b/qa/suites/rados/singleton/all/mon-memory-target-compliance.yaml.disabled @@ -0,0 +1,154 @@ +roles: +- - mon.a + - mgr.x + - osd.0 + - osd.1 + - osd.2 + - osd.3 + - osd.4 + - osd.5 + - osd.6 + - osd.7 + - osd.8 + - osd.9 + - osd.10 + - osd.11 + - osd.12 + - osd.13 + - osd.14 + - client.0 +openstack: + - volumes: # attached to each instance + count: 4 + size: 1 # GB +overrides: + ceph: + conf: + mon: + mon memory target: 134217728 # reduced to 128_M + rocksdb cache size: 67108864 # reduced to 64_M + mon osd cache size: 100000 + mon osd cache size min: 134217728 + osd: + osd memory target: 1610612736 # reduced to 1.5_G + osd objectstore: bluestore + debug bluestore: 20 + osd scrub min interval: 60 + osd scrub max interval: 120 + osd max backfills: 9 + +tasks: +- install: + branch: wip-sseshasa2-testing-2019-07-30-1825 # change as appropriate +- ceph: + create_rbd_pool: false + pre-mgr-commands: + - sudo ceph config set mgr mgr/devicehealth/enable_monitoring false --force + log-ignorelist: + - overall HEALTH_ + - \(OSDMAP_FLAGS\) + - \(OSD_ + - \(PG_ + - \(POOL_ + - \(CACHE_POOL_ + - \(OBJECT_ + - \(SLOW_OPS\) + - \(REQUEST_SLOW\) + - \(TOO_FEW_PGS\) + - slow request +- interactive: +- parallel: + - log-mon-rss + - stress-tasks + - benchload +- exec: + client.0: + - "ceph_test_mon_memory_target 134217728" # mon memory target + - "ceph_test_mon_rss_usage 134217728" +log-mon-rss: +- background_exec: + client.0: + - while true + - do /usr/bin/ceph_test_log_rss_usage ceph-mon >> /var/log/ceph/ceph-mon-rss-usage.log + - sleep 300 # log rss usage every 5 mins. May be modified accordingly + - done +- exec: + client.0: + - sleep 37860 # sum total of the radosbench test times below plus 60 secs +benchload: # The total radosbench test below translates to 10.5 hrs +- full_sequential: + - radosbench: + clients: [client.0] + time: 1800 + - radosbench: + clients: [client.0] + time: 1800 + - radosbench: + clients: [client.0] + time: 1800 + - radosbench: + clients: [client.0] + time: 1800 + - radosbench: + clients: [client.0] + time: 1800 + - radosbench: + clients: [client.0] + time: 1800 + - radosbench: + clients: [client.0] + time: 1800 + - radosbench: + clients: [client.0] + time: 1800 + - radosbench: + clients: [client.0] + time: 1800 + - radosbench: + clients: [client.0] + time: 1800 + - radosbench: + clients: [client.0] + time: 1800 + - radosbench: + clients: [client.0] + time: 1800 + - radosbench: + clients: [client.0] + time: 1800 + - radosbench: + clients: [client.0] + time: 1800 + - radosbench: + clients: [client.0] + time: 1800 + - radosbench: + clients: [client.0] + time: 1800 + - radosbench: + clients: [client.0] + time: 1800 + - radosbench: + clients: [client.0] + time: 1800 + - radosbench: + clients: [client.0] + time: 1800 + - radosbench: + clients: [client.0] + time: 1800 + - radosbench: + clients: [client.0] + time: 1800 +stress-tasks: +- thrashosds: + op_delay: 1 + bdev_inject_crash: 1 + bdev_inject_crash_probability: .8 + chance_down: 80 + chance_pgnum_grow: 3 + chance_pgpnum_fix: 1 + chance_thrash_cluster_full: 0 + chance_thrash_pg_upmap: 3 + chance_thrash_pg_upmap_items: 3 + min_in: 2 diff --git a/qa/suites/rados/singleton/all/osd-backfill.yaml b/qa/suites/rados/singleton/all/osd-backfill.yaml new file mode 100644 index 000000000..bbbd9b4b3 --- /dev/null +++ b/qa/suites/rados/singleton/all/osd-backfill.yaml @@ -0,0 +1,28 @@ +roles: +- - mon.a + - mon.b + - mon.c + - mgr.x + - osd.0 + - osd.1 + - osd.2 +openstack: + - volumes: # attached to each instance + count: 3 + size: 10 # GB +tasks: +- install: +- ceph: + pre-mgr-commands: + - sudo ceph config set mgr mgr/devicehealth/enable_monitoring false --force + log-ignorelist: + - but it is still running + - overall HEALTH_ + - \(OSDMAP_FLAGS\) + - \(OSD_ + - \(PG_ + - \(OBJECT_ + conf: + osd: + osd min pg log entries: 5 +- osd_backfill: diff --git a/qa/suites/rados/singleton/all/osd-recovery-incomplete.yaml b/qa/suites/rados/singleton/all/osd-recovery-incomplete.yaml new file mode 100644 index 000000000..15a0ea342 --- /dev/null +++ b/qa/suites/rados/singleton/all/osd-recovery-incomplete.yaml @@ -0,0 +1,30 @@ +roles: +- - mon.a + - mon.b + - mon.c + - mgr.x + - osd.0 + - osd.1 + - osd.2 + - osd.3 +openstack: + - volumes: # attached to each instance + count: 4 + size: 10 # GB +tasks: +- install: +- ceph: + pre-mgr-commands: + - sudo ceph config set mgr mgr/devicehealth/enable_monitoring false --force + log-ignorelist: + - but it is still running + - overall HEALTH_ + - \(OSDMAP_FLAGS\) + - \(OSD_ + - \(PG_ + - \(OBJECT_ + conf: + osd: + osd min pg log entries: 5 + osd_fast_fail_on_connection_refused: false +- osd_recovery.test_incomplete_pgs: diff --git a/qa/suites/rados/singleton/all/osd-recovery.yaml b/qa/suites/rados/singleton/all/osd-recovery.yaml new file mode 100644 index 000000000..75cea6a94 --- /dev/null +++ b/qa/suites/rados/singleton/all/osd-recovery.yaml @@ -0,0 +1,32 @@ +roles: +- - mon.a + - mon.b + - mon.c + - mgr.x + - osd.0 + - osd.1 + - osd.2 +openstack: + - volumes: # attached to each instance + count: 3 + size: 10 # GB +tasks: +- install: +- ceph: + pre-mgr-commands: + - sudo ceph config set mgr mgr/devicehealth/enable_monitoring false --force + log-ignorelist: + - but it is still running + - overall HEALTH_ + - \(OSDMAP_FLAGS\) + - \(OSD_ + - \(PG_ + - \(OBJECT_DEGRADED\) + - \(SLOW_OPS\) + - slow request + conf: + osd: + osd min pg log entries: 5 + osd pg log trim min: 0 + osd_fast_fail_on_connection_refused: false +- osd_recovery: diff --git a/qa/suites/rados/singleton/all/peer.yaml b/qa/suites/rados/singleton/all/peer.yaml new file mode 100644 index 000000000..24fd74b82 --- /dev/null +++ b/qa/suites/rados/singleton/all/peer.yaml @@ -0,0 +1,27 @@ +roles: +- - mon.a + - mon.b + - mon.c + - mgr.x + - osd.0 + - osd.1 + - osd.2 +openstack: + - volumes: # attached to each instance + count: 3 + size: 10 # GB +tasks: +- install: +- ceph: + pre-mgr-commands: + - sudo ceph config set mgr mgr/devicehealth/enable_monitoring false --force + config: + global: + osd pool default min size : 1 + log-ignorelist: + - objects unfound and apparently lost + - overall HEALTH_ + - \(OSDMAP_FLAGS\) + - \(OSD_ + - \(PG_ +- peer: diff --git a/qa/suites/rados/singleton/all/pg-autoscaler-progress-off.yaml b/qa/suites/rados/singleton/all/pg-autoscaler-progress-off.yaml new file mode 100644 index 000000000..042c3d78e --- /dev/null +++ b/qa/suites/rados/singleton/all/pg-autoscaler-progress-off.yaml @@ -0,0 +1,44 @@ +roles: +- - mon.a + - mgr.x + - osd.0 + - osd.1 + - osd.2 + - osd.3 + - client.0 +- - mon.b + - mon.c + - osd.4 + - osd.5 + - osd.6 + - osd.7 +openstack: + - volumes: # attached to each instance + count: 4 + size: 10 # GB +tasks: +- install: +- ceph: + create_rbd_pool: false + pre-mgr-commands: + - sudo ceph config set mgr mgr/devicehealth/enable_monitoring false --force + log-ignorelist: + - overall HEALTH_ + - \(OSDMAP_FLAGS\) + - \(OSD_ + - \(PG_ + - \(POOL_ + - \(CACHE_POOL_ + - \(OBJECT_ + - \(SLOW_OPS\) + - \(REQUEST_SLOW\) + - \(TOO_FEW_PGS\) + - slow request +- exec: + client.0: + - ceph progress off + +- workunit: + clients: + all: + - mon/pg_autoscaler.sh diff --git a/qa/suites/rados/singleton/all/pg-autoscaler.yaml b/qa/suites/rados/singleton/all/pg-autoscaler.yaml new file mode 100644 index 000000000..c2ab618ca --- /dev/null +++ b/qa/suites/rados/singleton/all/pg-autoscaler.yaml @@ -0,0 +1,36 @@ +roles: +- - mon.a + - mgr.x + - osd.0 + - osd.1 + - osd.2 + - osd.3 + - osd.4 + - osd.5 + - client.0 +openstack: + - volumes: # attached to each instance + count: 4 + size: 10 # GB +tasks: +- install: +- ceph: + create_rbd_pool: false + pre-mgr-commands: + - sudo ceph config set mgr mgr/devicehealth/enable_monitoring false --force + log-ignorelist: + - overall HEALTH_ + - \(OSDMAP_FLAGS\) + - \(OSD_ + - \(PG_ + - \(POOL_ + - \(CACHE_POOL_ + - \(OBJECT_ + - \(SLOW_OPS\) + - \(REQUEST_SLOW\) + - \(TOO_FEW_PGS\) + - slow request +- workunit: + clients: + all: + - mon/pg_autoscaler.sh diff --git a/qa/suites/rados/singleton/all/pg-removal-interruption.yaml b/qa/suites/rados/singleton/all/pg-removal-interruption.yaml new file mode 100644 index 000000000..b3f11264f --- /dev/null +++ b/qa/suites/rados/singleton/all/pg-removal-interruption.yaml @@ -0,0 +1,36 @@ +roles: +- - mon.a + - mgr.x + - osd.0 + - osd.1 + - osd.2 + - client.0 +openstack: + - volumes: # attached to each instance + count: 3 + size: 10 # GB +tasks: +- install: +- ceph: + pre-mgr-commands: + - sudo ceph config set mgr mgr/devicehealth/enable_monitoring false --force + log-ignorelist: + - but it is still running + - slow request + - overall HEALTH_ + - \(OSDMAP_FLAGS\) + - \(OSD_ + - \(PG_ +- exec: + client.0: + - sudo ceph osd pool create foo 128 128 + - sudo ceph osd pool application enable foo rados + - sleep 5 + - sudo ceph tell osd.0 injectargs -- --osd-inject-failure-on-pg-removal + - sudo ceph osd pool delete foo foo --yes-i-really-really-mean-it +- ceph.wait_for_failure: [osd.0] +- exec: + client.0: + - sudo ceph osd down 0 +- ceph.restart: [osd.0] +- ceph.healthy: diff --git a/qa/suites/rados/singleton/all/radostool.yaml b/qa/suites/rados/singleton/all/radostool.yaml new file mode 100644 index 000000000..fa3a1b0f7 --- /dev/null +++ b/qa/suites/rados/singleton/all/radostool.yaml @@ -0,0 +1,28 @@ +roles: +- - mon.a + - mgr.x + - osd.0 + - osd.1 + - osd.2 + - client.0 +openstack: + - volumes: # attached to each instance + count: 2 + size: 10 # GB +tasks: +- install: +- ceph: + pre-mgr-commands: + - sudo ceph config set mgr mgr/devicehealth/enable_monitoring false --force + log-ignorelist: + - but it is still running + - had wrong client addr + - had wrong cluster addr + - reached quota + - overall HEALTH_ + - \(POOL_FULL\) + - \(POOL_APP_NOT_ENABLED\) +- workunit: + clients: + all: + - rados/test_rados_tool.sh diff --git a/qa/suites/rados/singleton/all/random-eio.yaml b/qa/suites/rados/singleton/all/random-eio.yaml new file mode 100644 index 000000000..782b906d6 --- /dev/null +++ b/qa/suites/rados/singleton/all/random-eio.yaml @@ -0,0 +1,46 @@ +roles: +- - mon.a + - mgr.x + - osd.0 + - osd.1 + - osd.2 +- - osd.3 + - osd.4 + - osd.5 + - client.0 +openstack: + - volumes: # attached to each instance + count: 3 + size: 10 # GB +tasks: +- install: +- ceph: + pre-mgr-commands: + - sudo ceph config set mgr mgr/devicehealth/enable_monitoring false --force + log-ignorelist: + - missing primary copy of + - objects unfound and apparently lost + - had a read error + - overall HEALTH_ + - \(POOL_APP_NOT_ENABLED\) + - \(PG_DEGRADED\) + - \(OSD_TOO_MANY_REPAIRS\) +- full_sequential: + - exec: + client.0: + - sudo ceph tell osd.1 injectargs -- --filestore_debug_random_read_err=0.33 + - sudo ceph tell osd.1 injectargs -- --bluestore_debug_random_read_err=0.33 + - sudo ceph osd pool create test 16 16 + - sudo ceph osd pool set test size 3 + - sudo ceph pg dump pgs --format=json-pretty + - radosbench: + clients: [client.0] + time: 360 + type: rand + objectsize: 1048576 + pool: test + create_pool: false + - exec: + client.0: + - sudo ceph tell osd.1 injectargs -- --filestore_debug_random_read_err=0.0 + - sudo ceph tell osd.1 injectargs -- --bluestore_debug_random_read_err=0.0 diff --git a/qa/suites/rados/singleton/all/rebuild-mondb.yaml b/qa/suites/rados/singleton/all/rebuild-mondb.yaml new file mode 100644 index 000000000..3f1c74831 --- /dev/null +++ b/qa/suites/rados/singleton/all/rebuild-mondb.yaml @@ -0,0 +1,37 @@ +roles: +- - mon.a + - mon.b + - mon.c + - mgr.x + - osd.0 + - osd.1 + - osd.2 + - client.0 +openstack: + - volumes: # attached to each instance + count: 3 + size: 10 # GB +tasks: +- install: +- ceph: + pre-mgr-commands: + - sudo ceph config set mgr mgr/devicehealth/enable_monitoring false --force + log-ignorelist: + - no reply from + - overall HEALTH_ + - \(MON_DOWN\) + - \(MGR_DOWN\) + - \(OSDMAP_FLAGS\) + - \(OSD_ + - \(PG_ + conf: + mon: + debug auth: 30 +- full_sequential: + - radosbench: + clients: [client.0] + time: 30 + - rebuild_mondb: + - radosbench: + clients: [client.0] + time: 30 diff --git a/qa/suites/rados/singleton/all/recovery-preemption.yaml b/qa/suites/rados/singleton/all/recovery-preemption.yaml new file mode 100644 index 000000000..7438f9e77 --- /dev/null +++ b/qa/suites/rados/singleton/all/recovery-preemption.yaml @@ -0,0 +1,60 @@ +roles: +- - mon.a + - mon.b + - mon.c + - mgr.x + - osd.0 + - osd.1 + - osd.2 + - osd.3 +openstack: + - volumes: # attached to each instance + count: 3 + size: 20 # GB +tasks: +- install: +- ceph: + pre-mgr-commands: + - sudo ceph config set mgr mgr/devicehealth/enable_monitoring false --force + conf: + osd: + osd recovery sleep: .1 + osd min pg log entries: 10 + osd max pg log entries: 1000 + osd_target_pg_log_entries_per_osd: 0 + osd pg log trim min: 10 + log-ignorelist: + - \(POOL_APP_NOT_ENABLED\) + - \(OSDMAP_FLAGS\) + - \(OSD_ + - \(OBJECT_ + - \(PG_ + - \(SLOW_OPS\) + - overall HEALTH + - slow request +- exec: + osd.0: + - ceph osd pool create foo 128 + - ceph osd pool application enable foo foo + - sleep 5 +- ceph.healthy: +- exec: + osd.0: + - rados -p foo bench 30 write -b 4096 --no-cleanup + - ceph osd out 0 + - sleep 5 + - ceph osd set noup +- ceph.restart: + daemons: [osd.1] + wait-for-up: false + wait-for-healthy: false +- exec: + osd.0: + - rados -p foo bench 3 write -b 4096 --no-cleanup + - ceph osd unset noup + - sleep 10 + - for f in 0 1 2 3 ; do sudo ceph daemon osd.$f config set osd_recovery_sleep 0 ; sudo ceph daemon osd.$f config set osd_recovery_max_active 20 ; done +- ceph.healthy: +- exec: + osd.0: + - egrep '(defer backfill|defer recovery)' /var/log/ceph/ceph-osd.*.log diff --git a/qa/suites/rados/singleton/all/resolve_stuck_peering.yaml b/qa/suites/rados/singleton/all/resolve_stuck_peering.yaml new file mode 100644 index 000000000..2756ebe82 --- /dev/null +++ b/qa/suites/rados/singleton/all/resolve_stuck_peering.yaml @@ -0,0 +1,19 @@ +roles: +- [mon.a, mgr.x] +- [osd.0, osd.1, osd.2, client.0] + +tasks: +- install: +- ceph: + pre-mgr-commands: + - sudo ceph config set mgr mgr/devicehealth/enable_monitoring false --force + fs: xfs + log-ignorelist: + - overall HEALTH_ + - \(OSDMAP_FLAGS\) + - \(OSD_ + - \(PG_ + - \(OBJECT_DEGRADED\) + - \(POOL_APP_NOT_ENABLED\) +- resolve_stuck_peering: + diff --git a/qa/suites/rados/singleton/all/test-crash.yaml b/qa/suites/rados/singleton/all/test-crash.yaml new file mode 100644 index 000000000..beb83f0bb --- /dev/null +++ b/qa/suites/rados/singleton/all/test-crash.yaml @@ -0,0 +1,20 @@ +roles: + - [client.0, mon.a, mgr.x, osd.0, osd.1, osd.2] + +tasks: + - install: + - ceph: + pre-mgr-commands: + - sudo ceph config set mgr mgr/devicehealth/enable_monitoring false --force + log-ignorelist: + - Reduced data availability + - OSD_.*DOWN + - \(RECENT_CRASH\) + - workunit: + clients: + client.0: + - rados/test_crash.sh + - ceph.restart: [osd.*] + - exec: + mon.a: + - find $TESTDIR/archive/coredump -type f -exec rm -f {} \; diff --git a/qa/suites/rados/singleton/all/test-noautoscale-flag.yaml b/qa/suites/rados/singleton/all/test-noautoscale-flag.yaml new file mode 100644 index 000000000..23caa745d --- /dev/null +++ b/qa/suites/rados/singleton/all/test-noautoscale-flag.yaml @@ -0,0 +1,39 @@ +roles: +- - mon.a + - mgr.x + - osd.0 + - osd.1 + - osd.2 + - osd.3 + - client.0 +openstack: + - volumes: # attached to each instance + count: 4 + size: 10 # GB +overrides: + ceph: + create_rbd_pool: false + pre-mgr-commands: + - sudo ceph config set mgr mgr/devicehealth/enable_monitoring false --force + conf: + mon: + osd pool default pg autoscale mode: on + log-ignorelist: + - overall HEALTH_ + - \(OSDMAP_FLAGS\) + - \(OSD_ + - \(PG_ + - \(POOL_ + - \(CACHE_POOL_ + - \(OBJECT_ + - \(SLOW_OPS\) + - \(REQUEST_SLOW\) + - \(TOO_FEW_PGS\) + - slow request +tasks: +- install: +- ceph: +- workunit: + clients: + all: + - mon/test_noautoscale_flag.sh diff --git a/qa/suites/rados/singleton/all/test_envlibrados_for_rocksdb.yaml b/qa/suites/rados/singleton/all/test_envlibrados_for_rocksdb.yaml new file mode 100644 index 000000000..a76f6a8f0 --- /dev/null +++ b/qa/suites/rados/singleton/all/test_envlibrados_for_rocksdb.yaml @@ -0,0 +1,22 @@ +overrides: + ceph: + fs: ext4 + conf: + global: + osd max object name len: 460 + osd max object namespace len: 64 + osd client message cap: 5000 +roles: +- [mon.a, mgr.x, osd.0, osd.1, osd.2, client.0] +tasks: +- install: +- ceph: + pre-mgr-commands: + - sudo ceph config set mgr mgr/devicehealth/enable_monitoring false --force + log-ignorelist: + - overall HEALTH_ + - \(POOL_APP_NOT_ENABLED\) +- workunit: + clients: + all: + - rados/test_envlibrados_for_rocksdb.sh diff --git a/qa/suites/rados/singleton/all/thrash-backfill-full.yaml b/qa/suites/rados/singleton/all/thrash-backfill-full.yaml new file mode 100644 index 000000000..0f2924db3 --- /dev/null +++ b/qa/suites/rados/singleton/all/thrash-backfill-full.yaml @@ -0,0 +1,52 @@ +roles: +- - mon.a + - mgr.x + - osd.0 + - osd.1 + - osd.2 +- - osd.3 + - osd.4 + - osd.5 + - client.0 +openstack: + - volumes: # attached to each instance + count: 3 + size: 10 # GB +override: + ceph: + conf: + mon: + osd default pool size: 3 + osd min pg log entries: 5 + osd max pg log entries: 10 +tasks: +- install: +- ceph: + pre-mgr-commands: + - sudo ceph config set mgr mgr/devicehealth/enable_monitoring false --force + log-ignorelist: + - but it is still running + - missing primary copy of + - objects unfound and apparently lost + - overall HEALTH_ + - \(OSDMAP_FLAGS\) + - \(SLOW_OPS\) + - \(PG_ + - \(OBJECT_MISPLACED\) + - \(OSD_ + - \(OBJECT_ + - \(TOO_FEW_PGS\) + - \(POOL_BACKFILLFULL\) + - slow request +- thrashosds: + op_delay: 30 + clean_interval: 120 + chance_down: .75 + min_live: 5 + min_in: 5 + chance_test_backfill_full: .5 +- radosbench: + clients: [client.0] + time: 1800 + type: rand + objectsize: 1048576 diff --git a/qa/suites/rados/singleton/all/thrash-eio.yaml b/qa/suites/rados/singleton/all/thrash-eio.yaml new file mode 100644 index 000000000..5d9770061 --- /dev/null +++ b/qa/suites/rados/singleton/all/thrash-eio.yaml @@ -0,0 +1,49 @@ +roles: +- - mon.a + - mgr.x + - osd.0 + - osd.1 + - osd.2 +- - osd.3 + - osd.4 + - osd.5 + - client.0 +openstack: + - volumes: # attached to each instance + count: 3 + size: 10 # GB +override: + ceph: + conf: + mon: + osd default pool size: 3 +tasks: +- install: +- ceph: + pre-mgr-commands: + - sudo ceph config set mgr mgr/devicehealth/enable_monitoring false --force + log-ignorelist: + - but it is still running + - missing primary copy of + - objects unfound and apparently lost + - overall HEALTH_ + - \(OSDMAP_FLAGS\) + - \(SLOW_OPS\) + - \(PG_ + - \(OBJECT_MISPLACED\) + - \(OSD_ + - \(OBJECT_ + - \(TOO_FEW_PGS\) + - slow request +- thrashosds: + op_delay: 30 + clean_interval: 120 + chance_down: .5 + random_eio: .33 + min_live: 5 + min_in: 5 +- radosbench: + clients: [client.0] + time: 720 + type: rand + objectsize: 1048576 diff --git a/qa/suites/rados/singleton/all/thrash-rados/+ b/qa/suites/rados/singleton/all/thrash-rados/+ new file mode 100644 index 000000000..e69de29bb diff --git a/qa/suites/rados/singleton/all/thrash-rados/.qa b/qa/suites/rados/singleton/all/thrash-rados/.qa new file mode 120000 index 000000000..a602a0353 --- /dev/null +++ b/qa/suites/rados/singleton/all/thrash-rados/.qa @@ -0,0 +1 @@ +../.qa/ \ No newline at end of file diff --git a/qa/suites/rados/singleton/all/thrash-rados/thrash-rados.yaml b/qa/suites/rados/singleton/all/thrash-rados/thrash-rados.yaml new file mode 100644 index 000000000..1caef6db5 --- /dev/null +++ b/qa/suites/rados/singleton/all/thrash-rados/thrash-rados.yaml @@ -0,0 +1,27 @@ +roles: +- - mon.a + - mgr.x + - osd.0 + - osd.1 + - osd.2 +- - osd.3 + - osd.4 + - osd.5 + - client.0 +openstack: + - volumes: # attached to each instance + count: 3 + size: 10 # GB +tasks: +- install: +- ceph: + log-ignorelist: + - but it is still running +- thrashosds: + op_delay: 30 + clean_interval: 120 + chance_down: .5 +- workunit: + clients: + all: + - rados/load-gen-mix-small.sh diff --git a/qa/suites/rados/singleton/all/thrash-rados/thrashosds-health.yaml b/qa/suites/rados/singleton/all/thrash-rados/thrashosds-health.yaml new file mode 120000 index 000000000..9124eb1aa --- /dev/null +++ b/qa/suites/rados/singleton/all/thrash-rados/thrashosds-health.yaml @@ -0,0 +1 @@ +.qa/tasks/thrashosds-health.yaml \ No newline at end of file diff --git a/qa/suites/rados/singleton/all/thrash_cache_writeback_proxy_none.yaml b/qa/suites/rados/singleton/all/thrash_cache_writeback_proxy_none.yaml new file mode 100644 index 000000000..ab210abd7 --- /dev/null +++ b/qa/suites/rados/singleton/all/thrash_cache_writeback_proxy_none.yaml @@ -0,0 +1,70 @@ +roles: +- - mon.a + - mgr.x + - osd.0 + - osd.1 + - osd.2 +- - osd.3 + - osd.4 + - osd.5 + - client.0 +openstack: + - volumes: # attached to each instance + count: 3 + size: 30 # GB +tasks: +- install: +- ceph: + pre-mgr-commands: + - sudo ceph config set mgr mgr/devicehealth/enable_monitoring false --force + log-ignorelist: + - but it is still running + - slow request + - overall HEALTH_ + - \(CACHE_POOL_ +- exec: + client.0: + - sudo ceph osd pool create base 4 + - sudo ceph osd pool application enable base rados + - sudo ceph osd pool create cache 4 + - sudo ceph osd tier add base cache + - sudo ceph osd tier cache-mode cache writeback + - sudo ceph osd tier set-overlay base cache + - sudo ceph osd pool set cache hit_set_type bloom + - sudo ceph osd pool set cache hit_set_count 8 + - sudo ceph osd pool set cache hit_set_period 60 + - sudo ceph osd pool set cache target_max_objects 500 +- background_exec: + mon.a: + - while true + - do sleep 30 + - sudo ceph osd pool set cache cache_target_full_ratio .001 + - echo cache-try-flush-evict-all + - rados -p cache cache-try-flush-evict-all + - sleep 5 + - echo cache-flush-evict-all + - rados -p cache cache-flush-evict-all + - sleep 5 + - echo remove overlay + - sudo ceph osd tier remove-overlay base + - sleep 20 + # Disabled due to https://tracker.ceph.com/issues/46323 + #- echo add writeback overlay + #- sudo ceph osd tier cache-mode cache writeback + #- sudo ceph osd pool set cache cache_target_full_ratio .8 + #- sudo ceph osd tier set-overlay base cache + #- sleep 30 + #- sudo ceph osd tier cache-mode cache readproxy + - done +- rados: + clients: [client.0] + pools: [base] + max_seconds: 600 + ops: 400000 + objects: 10000 + size: 1024 + op_weights: + read: 100 + write: 100 + delete: 50 + copy_from: 50 diff --git a/qa/suites/rados/singleton/all/watch-notify-same-primary.yaml b/qa/suites/rados/singleton/all/watch-notify-same-primary.yaml new file mode 100644 index 000000000..eeb585c88 --- /dev/null +++ b/qa/suites/rados/singleton/all/watch-notify-same-primary.yaml @@ -0,0 +1,34 @@ +roles: +- - mon.a + - mon.b + - mon.c + - mgr.x + - osd.0 + - osd.1 + - osd.2 + - client.0 +openstack: + - volumes: # attached to each instance + count: 3 + size: 10 # GB +tasks: +- install: +- ceph: + pre-mgr-commands: + - sudo ceph config set mgr mgr/devicehealth/enable_monitoring false --force + config: + global: + osd pool default min size : 1 + client: + debug ms: 1 + debug objecter: 20 + debug rados: 20 + log-ignorelist: + - objects unfound and apparently lost + - overall HEALTH_ + - \(OSDMAP_FLAGS\) + - \(OSD_ + - \(PG_ + - \(OBJECT_DEGRADED\) +- watch_notify_same_primary: + clients: [client.0] diff --git a/qa/suites/rados/singleton/mon_election b/qa/suites/rados/singleton/mon_election new file mode 120000 index 000000000..3f331e621 --- /dev/null +++ b/qa/suites/rados/singleton/mon_election @@ -0,0 +1 @@ +.qa/mon_election \ No newline at end of file diff --git a/qa/suites/rados/singleton/msgr b/qa/suites/rados/singleton/msgr new file mode 120000 index 000000000..57bee80db --- /dev/null +++ b/qa/suites/rados/singleton/msgr @@ -0,0 +1 @@ +.qa/msgr \ No newline at end of file diff --git a/qa/suites/rados/singleton/msgr-failures/.qa b/qa/suites/rados/singleton/msgr-failures/.qa new file mode 120000 index 000000000..a602a0353 --- /dev/null +++ b/qa/suites/rados/singleton/msgr-failures/.qa @@ -0,0 +1 @@ +../.qa/ \ No newline at end of file diff --git a/qa/suites/rados/singleton/msgr-failures/few.yaml b/qa/suites/rados/singleton/msgr-failures/few.yaml new file mode 100644 index 000000000..8fd638744 --- /dev/null +++ b/qa/suites/rados/singleton/msgr-failures/few.yaml @@ -0,0 +1,9 @@ +overrides: + ceph: + conf: + global: + ms inject socket failures: 5000 + mon client directed command retry: 5 + log-ignorelist: + - \(OSD_SLOW_PING_TIME + - \(MON_DOWN\) diff --git a/qa/suites/rados/singleton/msgr-failures/many.yaml b/qa/suites/rados/singleton/msgr-failures/many.yaml new file mode 100644 index 000000000..206da3ec1 --- /dev/null +++ b/qa/suites/rados/singleton/msgr-failures/many.yaml @@ -0,0 +1,13 @@ +overrides: + ceph: + conf: + global: + ms inject socket failures: 1000 + mon mgr beacon grace: 90 + mon client hunt interval max multiple: 2 + mon client directed command retry: 5 + mgr: + debug monc: 10 + log-ignorelist: + - \(OSD_SLOW_PING_TIME + - \(MON_DOWN\) diff --git a/qa/suites/rados/singleton/msgr-failures/none.yaml b/qa/suites/rados/singleton/msgr-failures/none.yaml new file mode 100644 index 000000000..e69de29bb diff --git a/qa/suites/rados/singleton/objectstore b/qa/suites/rados/singleton/objectstore new file mode 120000 index 000000000..848c65f9e --- /dev/null +++ b/qa/suites/rados/singleton/objectstore @@ -0,0 +1 @@ +.qa/objectstore_debug \ No newline at end of file diff --git a/qa/suites/rados/singleton/rados.yaml b/qa/suites/rados/singleton/rados.yaml new file mode 120000 index 000000000..d256979c0 --- /dev/null +++ b/qa/suites/rados/singleton/rados.yaml @@ -0,0 +1 @@ +.qa/config/rados.yaml \ No newline at end of file diff --git a/qa/suites/rados/singleton/supported-random-distro$ b/qa/suites/rados/singleton/supported-random-distro$ new file mode 120000 index 000000000..7cef21eef --- /dev/null +++ b/qa/suites/rados/singleton/supported-random-distro$ @@ -0,0 +1 @@ +../basic/supported-random-distro$ \ No newline at end of file diff --git a/qa/suites/rados/standalone/% b/qa/suites/rados/standalone/% new file mode 100644 index 000000000..e69de29bb diff --git a/qa/suites/rados/standalone/.qa b/qa/suites/rados/standalone/.qa new file mode 120000 index 000000000..a602a0353 --- /dev/null +++ b/qa/suites/rados/standalone/.qa @@ -0,0 +1 @@ +../.qa/ \ No newline at end of file diff --git a/qa/suites/rados/standalone/supported-random-distro$ b/qa/suites/rados/standalone/supported-random-distro$ new file mode 120000 index 000000000..7cef21eef --- /dev/null +++ b/qa/suites/rados/standalone/supported-random-distro$ @@ -0,0 +1 @@ +../basic/supported-random-distro$ \ No newline at end of file diff --git a/qa/suites/rados/standalone/workloads/.qa b/qa/suites/rados/standalone/workloads/.qa new file mode 120000 index 000000000..a602a0353 --- /dev/null +++ b/qa/suites/rados/standalone/workloads/.qa @@ -0,0 +1 @@ +../.qa/ \ No newline at end of file diff --git a/qa/suites/rados/standalone/workloads/crush.yaml b/qa/suites/rados/standalone/workloads/crush.yaml new file mode 100644 index 000000000..a62a0dd81 --- /dev/null +++ b/qa/suites/rados/standalone/workloads/crush.yaml @@ -0,0 +1,18 @@ +roles: +- - mon.a + - mgr.x + - osd.0 + - osd.1 + - osd.2 + - client.0 +openstack: + - volumes: # attached to each instance + count: 3 + size: 10 # GB +tasks: +- install: +- workunit: + basedir: qa/standalone + clients: + all: + - crush diff --git a/qa/suites/rados/standalone/workloads/erasure-code.yaml b/qa/suites/rados/standalone/workloads/erasure-code.yaml new file mode 100644 index 000000000..7d79753ce --- /dev/null +++ b/qa/suites/rados/standalone/workloads/erasure-code.yaml @@ -0,0 +1,18 @@ +roles: +- - mon.a + - mgr.x + - osd.0 + - osd.1 + - osd.2 + - client.0 +openstack: + - volumes: # attached to each instance + count: 3 + size: 10 # GB +tasks: +- install: +- workunit: + basedir: qa/standalone + clients: + all: + - erasure-code diff --git a/qa/suites/rados/standalone/workloads/mgr.yaml b/qa/suites/rados/standalone/workloads/mgr.yaml new file mode 100644 index 000000000..997fae865 --- /dev/null +++ b/qa/suites/rados/standalone/workloads/mgr.yaml @@ -0,0 +1,18 @@ +roles: +- - mon.a + - mgr.x + - osd.0 + - osd.1 + - osd.2 + - client.0 +openstack: + - volumes: # attached to each instance + count: 3 + size: 10 # GB +tasks: +- install: +- workunit: + basedir: qa/standalone + clients: + all: + - mgr diff --git a/qa/suites/rados/standalone/workloads/misc.yaml b/qa/suites/rados/standalone/workloads/misc.yaml new file mode 100644 index 000000000..4aa9ee27e --- /dev/null +++ b/qa/suites/rados/standalone/workloads/misc.yaml @@ -0,0 +1,18 @@ +roles: +- - mon.a + - mgr.x + - osd.0 + - osd.1 + - osd.2 + - client.0 +openstack: + - volumes: # attached to each instance + count: 3 + size: 10 # GB +tasks: +- install: +- workunit: + basedir: qa/standalone + clients: + all: + - misc diff --git a/qa/suites/rados/standalone/workloads/mon.yaml b/qa/suites/rados/standalone/workloads/mon.yaml new file mode 100644 index 000000000..c19606f42 --- /dev/null +++ b/qa/suites/rados/standalone/workloads/mon.yaml @@ -0,0 +1,18 @@ +roles: +- - mon.a + - mgr.x + - osd.0 + - osd.1 + - osd.2 + - client.0 +openstack: + - volumes: # attached to each instance + count: 3 + size: 10 # GB +tasks: +- install: +- workunit: + basedir: qa/standalone + clients: + all: + - mon diff --git a/qa/suites/rados/standalone/workloads/osd.yaml b/qa/suites/rados/standalone/workloads/osd.yaml new file mode 100644 index 000000000..e28b52210 --- /dev/null +++ b/qa/suites/rados/standalone/workloads/osd.yaml @@ -0,0 +1,18 @@ +roles: +- - mon.a + - mgr.x + - osd.0 + - osd.1 + - osd.2 + - client.0 +openstack: + - volumes: # attached to each instance + count: 3 + size: 10 # GB +tasks: +- install: +- workunit: + basedir: qa/standalone + clients: + all: + - osd diff --git a/qa/suites/rados/standalone/workloads/scrub.yaml b/qa/suites/rados/standalone/workloads/scrub.yaml new file mode 100644 index 000000000..7f6fad406 --- /dev/null +++ b/qa/suites/rados/standalone/workloads/scrub.yaml @@ -0,0 +1,18 @@ +roles: +- - mon.a + - mgr.x + - osd.0 + - osd.1 + - osd.2 + - client.0 +openstack: + - volumes: # attached to each instance + count: 3 + size: 10 # GB +tasks: +- install: +- workunit: + basedir: qa/standalone + clients: + all: + - scrub diff --git a/qa/suites/rados/thrash-erasure-code-big/% b/qa/suites/rados/thrash-erasure-code-big/% new file mode 100644 index 000000000..e69de29bb diff --git a/qa/suites/rados/thrash-erasure-code-big/.qa b/qa/suites/rados/thrash-erasure-code-big/.qa new file mode 120000 index 000000000..a602a0353 --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code-big/.qa @@ -0,0 +1 @@ +../.qa/ \ No newline at end of file diff --git a/qa/suites/rados/thrash-erasure-code-big/ceph.yaml b/qa/suites/rados/thrash-erasure-code-big/ceph.yaml new file mode 120000 index 000000000..a2fd139cb --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code-big/ceph.yaml @@ -0,0 +1 @@ +../thrash/ceph.yaml \ No newline at end of file diff --git a/qa/suites/rados/thrash-erasure-code-big/cluster/+ b/qa/suites/rados/thrash-erasure-code-big/cluster/+ new file mode 100644 index 000000000..e69de29bb diff --git a/qa/suites/rados/thrash-erasure-code-big/cluster/.qa b/qa/suites/rados/thrash-erasure-code-big/cluster/.qa new file mode 120000 index 000000000..a602a0353 --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code-big/cluster/.qa @@ -0,0 +1 @@ +../.qa/ \ No newline at end of file diff --git a/qa/suites/rados/thrash-erasure-code-big/cluster/12-osds.yaml b/qa/suites/rados/thrash-erasure-code-big/cluster/12-osds.yaml new file mode 100644 index 000000000..1c45ee352 --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code-big/cluster/12-osds.yaml @@ -0,0 +1,4 @@ +roles: +- [osd.0, osd.1, osd.2, osd.3, client.0, mon.a] +- [osd.4, osd.5, osd.6, osd.7, mon.b, mgr.x] +- [osd.8, osd.9, osd.10, osd.11, mon.c] diff --git a/qa/suites/rados/thrash-erasure-code-big/cluster/openstack.yaml b/qa/suites/rados/thrash-erasure-code-big/cluster/openstack.yaml new file mode 100644 index 000000000..e559d9126 --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code-big/cluster/openstack.yaml @@ -0,0 +1,4 @@ +openstack: + - volumes: # attached to each instance + count: 4 + size: 10 # GB diff --git a/qa/suites/rados/thrash-erasure-code-big/mon_election b/qa/suites/rados/thrash-erasure-code-big/mon_election new file mode 120000 index 000000000..3f331e621 --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code-big/mon_election @@ -0,0 +1 @@ +.qa/mon_election \ No newline at end of file diff --git a/qa/suites/rados/thrash-erasure-code-big/msgr-failures b/qa/suites/rados/thrash-erasure-code-big/msgr-failures new file mode 120000 index 000000000..03689aa44 --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code-big/msgr-failures @@ -0,0 +1 @@ +../thrash/msgr-failures \ No newline at end of file diff --git a/qa/suites/rados/thrash-erasure-code-big/objectstore b/qa/suites/rados/thrash-erasure-code-big/objectstore new file mode 120000 index 000000000..848c65f9e --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code-big/objectstore @@ -0,0 +1 @@ +.qa/objectstore_debug \ No newline at end of file diff --git a/qa/suites/rados/thrash-erasure-code-big/rados.yaml b/qa/suites/rados/thrash-erasure-code-big/rados.yaml new file mode 120000 index 000000000..d256979c0 --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code-big/rados.yaml @@ -0,0 +1 @@ +.qa/config/rados.yaml \ No newline at end of file diff --git a/qa/suites/rados/thrash-erasure-code-big/recovery-overrides b/qa/suites/rados/thrash-erasure-code-big/recovery-overrides new file mode 120000 index 000000000..1957f2c42 --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code-big/recovery-overrides @@ -0,0 +1 @@ +../thrash/2-recovery-overrides \ No newline at end of file diff --git a/qa/suites/rados/thrash-erasure-code-big/supported-random-distro$ b/qa/suites/rados/thrash-erasure-code-big/supported-random-distro$ new file mode 120000 index 000000000..7cef21eef --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code-big/supported-random-distro$ @@ -0,0 +1 @@ +../basic/supported-random-distro$ \ No newline at end of file diff --git a/qa/suites/rados/thrash-erasure-code-big/thrashers/.qa b/qa/suites/rados/thrash-erasure-code-big/thrashers/.qa new file mode 120000 index 000000000..a602a0353 --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code-big/thrashers/.qa @@ -0,0 +1 @@ +../.qa/ \ No newline at end of file diff --git a/qa/suites/rados/thrash-erasure-code-big/thrashers/careful.yaml b/qa/suites/rados/thrash-erasure-code-big/thrashers/careful.yaml new file mode 100644 index 000000000..f032d88e7 --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code-big/thrashers/careful.yaml @@ -0,0 +1,20 @@ +overrides: + ceph: + log-ignorelist: + - but it is still running + - objects unfound and apparently lost + - slow request + conf: + osd: + osd debug reject backfill probability: .3 + osd scrub min interval: 60 + osd scrub max interval: 120 + osd max backfills: 6 +tasks: +- thrashosds: + timeout: 1200 + chance_pgnum_grow: 1 + chance_pgnum_shrink: 1 + chance_pgpnum_fix: 1 + min_in: 8 + aggressive_pg_num_changes: false diff --git a/qa/suites/rados/thrash-erasure-code-big/thrashers/default.yaml b/qa/suites/rados/thrash-erasure-code-big/thrashers/default.yaml new file mode 100644 index 000000000..d2c7b85b5 --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code-big/thrashers/default.yaml @@ -0,0 +1,21 @@ +overrides: + ceph: + log-ignorelist: + - but it is still running + - objects unfound and apparently lost + - slow request + conf: + osd: + osd debug reject backfill probability: .1 + osd scrub min interval: 60 + osd scrub max interval: 120 + osd max backfills: 6 +tasks: +- thrashosds: + timeout: 1200 + chance_pgnum_grow: 1 + chance_pgnum_shrink: 1 + chance_pgpnum_fix: 1 + min_in: 8 + chance_bluestore_reshard: 1 + bluestore_new_sharding: random diff --git a/qa/suites/rados/thrash-erasure-code-big/thrashers/fastread.yaml b/qa/suites/rados/thrash-erasure-code-big/thrashers/fastread.yaml new file mode 100644 index 000000000..371ed570b --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code-big/thrashers/fastread.yaml @@ -0,0 +1,20 @@ +overrides: + ceph: + log-ignorelist: + - but it is still running + - objects unfound and apparently lost + conf: + mon: + osd pool default ec fast read: true + osd: + osd debug reject backfill probability: .1 + osd scrub min interval: 60 + osd scrub max interval: 120 + osd max backfills: 2 +tasks: +- thrashosds: + timeout: 1200 + chance_pgnum_grow: 1 + chance_pgnum_shrink: 1 + chance_pgpnum_fix: 1 + min_in: 4 diff --git a/qa/suites/rados/thrash-erasure-code-big/thrashers/mapgap.yaml b/qa/suites/rados/thrash-erasure-code-big/thrashers/mapgap.yaml new file mode 100644 index 000000000..18843d872 --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code-big/thrashers/mapgap.yaml @@ -0,0 +1,22 @@ +overrides: + ceph: + log-ignorelist: + - but it is still running + - objects unfound and apparently lost + - osd_map_cache_size + conf: + mon: + mon min osdmap epochs: 2 + osd: + osd map cache size: 1 + osd scrub min interval: 60 + osd scrub max interval: 120 + osd max backfills: 6 +tasks: +- thrashosds: + timeout: 1800 + chance_pgnum_grow: 1 + chance_pgnum_shrink: 1 + chance_pgpnum_fix: 1 + chance_test_map_discontinuity: 0.5 + min_in: 8 diff --git a/qa/suites/rados/thrash-erasure-code-big/thrashers/morepggrow.yaml b/qa/suites/rados/thrash-erasure-code-big/thrashers/morepggrow.yaml new file mode 100644 index 000000000..23c9f7d84 --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code-big/thrashers/morepggrow.yaml @@ -0,0 +1,16 @@ +overrides: + ceph: + conf: + osd: + osd scrub min interval: 60 + osd scrub max interval: 120 + osd max backfills: 9 + log-ignorelist: + - but it is still running + - objects unfound and apparently lost +tasks: +- thrashosds: + timeout: 1200 + chance_pgnum_grow: 3 + chance_pgpnum_fix: 1 + min_in: 8 diff --git a/qa/suites/rados/thrash-erasure-code-big/thrashers/pggrow.yaml b/qa/suites/rados/thrash-erasure-code-big/thrashers/pggrow.yaml new file mode 100644 index 000000000..9cbb80dba --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code-big/thrashers/pggrow.yaml @@ -0,0 +1,16 @@ +overrides: + ceph: + log-ignorelist: + - but it is still running + - objects unfound and apparently lost + conf: + osd: + osd scrub min interval: 60 + osd scrub max interval: 120 + osd max backfills: 6 +tasks: +- thrashosds: + timeout: 1200 + chance_pgnum_grow: 2 + chance_pgpnum_fix: 1 + min_in: 8 diff --git a/qa/suites/rados/thrash-erasure-code-big/thrashosds-health.yaml b/qa/suites/rados/thrash-erasure-code-big/thrashosds-health.yaml new file mode 120000 index 000000000..9124eb1aa --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code-big/thrashosds-health.yaml @@ -0,0 +1 @@ +.qa/tasks/thrashosds-health.yaml \ No newline at end of file diff --git a/qa/suites/rados/thrash-erasure-code-big/workloads/.qa b/qa/suites/rados/thrash-erasure-code-big/workloads/.qa new file mode 120000 index 000000000..a602a0353 --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code-big/workloads/.qa @@ -0,0 +1 @@ +../.qa/ \ No newline at end of file diff --git a/qa/suites/rados/thrash-erasure-code-big/workloads/ec-rados-plugin=jerasure-k=4-m=2.yaml b/qa/suites/rados/thrash-erasure-code-big/workloads/ec-rados-plugin=jerasure-k=4-m=2.yaml new file mode 120000 index 000000000..c18bec161 --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code-big/workloads/ec-rados-plugin=jerasure-k=4-m=2.yaml @@ -0,0 +1 @@ +.qa/erasure-code/ec-rados-plugin=jerasure-k=4-m=2.yaml \ No newline at end of file diff --git a/qa/suites/rados/thrash-erasure-code-big/workloads/ec-rados-plugin=lrc-k=4-m=2-l=3.yaml b/qa/suites/rados/thrash-erasure-code-big/workloads/ec-rados-plugin=lrc-k=4-m=2-l=3.yaml new file mode 120000 index 000000000..d66fd7960 --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code-big/workloads/ec-rados-plugin=lrc-k=4-m=2-l=3.yaml @@ -0,0 +1 @@ +.qa/erasure-code/ec-rados-plugin=lrc-k=4-m=2-l=3.yaml \ No newline at end of file diff --git a/qa/suites/rados/thrash-erasure-code-isa/% b/qa/suites/rados/thrash-erasure-code-isa/% new file mode 100644 index 000000000..e69de29bb diff --git a/qa/suites/rados/thrash-erasure-code-isa/.qa b/qa/suites/rados/thrash-erasure-code-isa/.qa new file mode 120000 index 000000000..a602a0353 --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code-isa/.qa @@ -0,0 +1 @@ +../.qa/ \ No newline at end of file diff --git a/qa/suites/rados/thrash-erasure-code-isa/arch/.qa b/qa/suites/rados/thrash-erasure-code-isa/arch/.qa new file mode 120000 index 000000000..a602a0353 --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code-isa/arch/.qa @@ -0,0 +1 @@ +../.qa/ \ No newline at end of file diff --git a/qa/suites/rados/thrash-erasure-code-isa/arch/x86_64.yaml b/qa/suites/rados/thrash-erasure-code-isa/arch/x86_64.yaml new file mode 100644 index 000000000..c2409f5d0 --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code-isa/arch/x86_64.yaml @@ -0,0 +1 @@ +arch: x86_64 diff --git a/qa/suites/rados/thrash-erasure-code-isa/ceph.yaml b/qa/suites/rados/thrash-erasure-code-isa/ceph.yaml new file mode 120000 index 000000000..a2fd139cb --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code-isa/ceph.yaml @@ -0,0 +1 @@ +../thrash/ceph.yaml \ No newline at end of file diff --git a/qa/suites/rados/thrash-erasure-code-isa/clusters b/qa/suites/rados/thrash-erasure-code-isa/clusters new file mode 120000 index 000000000..7aac47be3 --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code-isa/clusters @@ -0,0 +1 @@ +../thrash/clusters \ No newline at end of file diff --git a/qa/suites/rados/thrash-erasure-code-isa/mon_election b/qa/suites/rados/thrash-erasure-code-isa/mon_election new file mode 120000 index 000000000..3f331e621 --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code-isa/mon_election @@ -0,0 +1 @@ +.qa/mon_election \ No newline at end of file diff --git a/qa/suites/rados/thrash-erasure-code-isa/msgr-failures b/qa/suites/rados/thrash-erasure-code-isa/msgr-failures new file mode 120000 index 000000000..03689aa44 --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code-isa/msgr-failures @@ -0,0 +1 @@ +../thrash/msgr-failures \ No newline at end of file diff --git a/qa/suites/rados/thrash-erasure-code-isa/objectstore b/qa/suites/rados/thrash-erasure-code-isa/objectstore new file mode 120000 index 000000000..848c65f9e --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code-isa/objectstore @@ -0,0 +1 @@ +.qa/objectstore_debug \ No newline at end of file diff --git a/qa/suites/rados/thrash-erasure-code-isa/rados.yaml b/qa/suites/rados/thrash-erasure-code-isa/rados.yaml new file mode 120000 index 000000000..d256979c0 --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code-isa/rados.yaml @@ -0,0 +1 @@ +.qa/config/rados.yaml \ No newline at end of file diff --git a/qa/suites/rados/thrash-erasure-code-isa/recovery-overrides b/qa/suites/rados/thrash-erasure-code-isa/recovery-overrides new file mode 120000 index 000000000..1957f2c42 --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code-isa/recovery-overrides @@ -0,0 +1 @@ +../thrash/2-recovery-overrides \ No newline at end of file diff --git a/qa/suites/rados/thrash-erasure-code-isa/supported-random-distro$ b/qa/suites/rados/thrash-erasure-code-isa/supported-random-distro$ new file mode 120000 index 000000000..7cef21eef --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code-isa/supported-random-distro$ @@ -0,0 +1 @@ +../basic/supported-random-distro$ \ No newline at end of file diff --git a/qa/suites/rados/thrash-erasure-code-isa/thrashers b/qa/suites/rados/thrash-erasure-code-isa/thrashers new file mode 120000 index 000000000..f461dadc3 --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code-isa/thrashers @@ -0,0 +1 @@ +../thrash/thrashers \ No newline at end of file diff --git a/qa/suites/rados/thrash-erasure-code-isa/thrashosds-health.yaml b/qa/suites/rados/thrash-erasure-code-isa/thrashosds-health.yaml new file mode 120000 index 000000000..9124eb1aa --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code-isa/thrashosds-health.yaml @@ -0,0 +1 @@ +.qa/tasks/thrashosds-health.yaml \ No newline at end of file diff --git a/qa/suites/rados/thrash-erasure-code-isa/workloads/.qa b/qa/suites/rados/thrash-erasure-code-isa/workloads/.qa new file mode 120000 index 000000000..a602a0353 --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code-isa/workloads/.qa @@ -0,0 +1 @@ +../.qa/ \ No newline at end of file diff --git a/qa/suites/rados/thrash-erasure-code-isa/workloads/ec-rados-plugin=isa-k=2-m=1.yaml b/qa/suites/rados/thrash-erasure-code-isa/workloads/ec-rados-plugin=isa-k=2-m=1.yaml new file mode 120000 index 000000000..19342b9d8 --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code-isa/workloads/ec-rados-plugin=isa-k=2-m=1.yaml @@ -0,0 +1 @@ +.qa/erasure-code/ec-rados-plugin=isa-k=2-m=1.yaml \ No newline at end of file diff --git a/qa/suites/rados/thrash-erasure-code-overwrites/% b/qa/suites/rados/thrash-erasure-code-overwrites/% new file mode 100644 index 000000000..e69de29bb diff --git a/qa/suites/rados/thrash-erasure-code-overwrites/.qa b/qa/suites/rados/thrash-erasure-code-overwrites/.qa new file mode 120000 index 000000000..a602a0353 --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code-overwrites/.qa @@ -0,0 +1 @@ +../.qa/ \ No newline at end of file diff --git a/qa/suites/rados/thrash-erasure-code-overwrites/bluestore-bitmap.yaml b/qa/suites/rados/thrash-erasure-code-overwrites/bluestore-bitmap.yaml new file mode 120000 index 000000000..635085f7f --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code-overwrites/bluestore-bitmap.yaml @@ -0,0 +1 @@ +../thrash-erasure-code/objectstore/bluestore-bitmap.yaml \ No newline at end of file diff --git a/qa/suites/rados/thrash-erasure-code-overwrites/ceph.yaml b/qa/suites/rados/thrash-erasure-code-overwrites/ceph.yaml new file mode 120000 index 000000000..a2fd139cb --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code-overwrites/ceph.yaml @@ -0,0 +1 @@ +../thrash/ceph.yaml \ No newline at end of file diff --git a/qa/suites/rados/thrash-erasure-code-overwrites/clusters b/qa/suites/rados/thrash-erasure-code-overwrites/clusters new file mode 120000 index 000000000..646ea04cd --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code-overwrites/clusters @@ -0,0 +1 @@ +../thrash-erasure-code/clusters \ No newline at end of file diff --git a/qa/suites/rados/thrash-erasure-code-overwrites/fast b/qa/suites/rados/thrash-erasure-code-overwrites/fast new file mode 120000 index 000000000..6170b30e0 --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code-overwrites/fast @@ -0,0 +1 @@ +../thrash-erasure-code/fast \ No newline at end of file diff --git a/qa/suites/rados/thrash-erasure-code-overwrites/mon_election b/qa/suites/rados/thrash-erasure-code-overwrites/mon_election new file mode 120000 index 000000000..3f331e621 --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code-overwrites/mon_election @@ -0,0 +1 @@ +.qa/mon_election \ No newline at end of file diff --git a/qa/suites/rados/thrash-erasure-code-overwrites/msgr-failures b/qa/suites/rados/thrash-erasure-code-overwrites/msgr-failures new file mode 120000 index 000000000..70c9ca130 --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code-overwrites/msgr-failures @@ -0,0 +1 @@ +../thrash-erasure-code/msgr-failures \ No newline at end of file diff --git a/qa/suites/rados/thrash-erasure-code-overwrites/rados.yaml b/qa/suites/rados/thrash-erasure-code-overwrites/rados.yaml new file mode 120000 index 000000000..017df6f60 --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code-overwrites/rados.yaml @@ -0,0 +1 @@ +../thrash-erasure-code/rados.yaml \ No newline at end of file diff --git a/qa/suites/rados/thrash-erasure-code-overwrites/recovery-overrides b/qa/suites/rados/thrash-erasure-code-overwrites/recovery-overrides new file mode 120000 index 000000000..1957f2c42 --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code-overwrites/recovery-overrides @@ -0,0 +1 @@ +../thrash/2-recovery-overrides \ No newline at end of file diff --git a/qa/suites/rados/thrash-erasure-code-overwrites/supported-random-distro$ b/qa/suites/rados/thrash-erasure-code-overwrites/supported-random-distro$ new file mode 120000 index 000000000..7cef21eef --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code-overwrites/supported-random-distro$ @@ -0,0 +1 @@ +../basic/supported-random-distro$ \ No newline at end of file diff --git a/qa/suites/rados/thrash-erasure-code-overwrites/thrashers b/qa/suites/rados/thrash-erasure-code-overwrites/thrashers new file mode 120000 index 000000000..40ff82cf7 --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code-overwrites/thrashers @@ -0,0 +1 @@ +../thrash-erasure-code/thrashers \ No newline at end of file diff --git a/qa/suites/rados/thrash-erasure-code-overwrites/thrashosds-health.yaml b/qa/suites/rados/thrash-erasure-code-overwrites/thrashosds-health.yaml new file mode 120000 index 000000000..9124eb1aa --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code-overwrites/thrashosds-health.yaml @@ -0,0 +1 @@ +.qa/tasks/thrashosds-health.yaml \ No newline at end of file diff --git a/qa/suites/rados/thrash-erasure-code-overwrites/workloads/.qa b/qa/suites/rados/thrash-erasure-code-overwrites/workloads/.qa new file mode 120000 index 000000000..a602a0353 --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code-overwrites/workloads/.qa @@ -0,0 +1 @@ +../.qa/ \ No newline at end of file diff --git a/qa/suites/rados/thrash-erasure-code-overwrites/workloads/ec-pool-snaps-few-objects-overwrites.yaml b/qa/suites/rados/thrash-erasure-code-overwrites/workloads/ec-pool-snaps-few-objects-overwrites.yaml new file mode 100644 index 000000000..d2ad70a57 --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code-overwrites/workloads/ec-pool-snaps-few-objects-overwrites.yaml @@ -0,0 +1,23 @@ +overrides: + ceph: + conf: + global: + enable experimental unrecoverable data corrupting features: '*' + thrashosds: + disable_objectstore_tool_tests: true +tasks: +- rados: + clients: [client.0] + ops: 4000 + objects: 50 + pool_snaps: true + ec_pool: true + erasure_code_use_overwrites: true + op_weights: + read: 100 + write: 100 + delete: 50 + snap_create: 50 + snap_remove: 50 + rollback: 50 + copy_from: 50 diff --git a/qa/suites/rados/thrash-erasure-code-overwrites/workloads/ec-small-objects-fast-read-overwrites.yaml b/qa/suites/rados/thrash-erasure-code-overwrites/workloads/ec-small-objects-fast-read-overwrites.yaml new file mode 100644 index 000000000..b3f831b77 --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code-overwrites/workloads/ec-small-objects-fast-read-overwrites.yaml @@ -0,0 +1,29 @@ +overrides: + ceph: + conf: + global: + enable experimental unrecoverable data corrupting features: '*' + thrashosds: + disable_objectstore_tool_tests: true +tasks: +- rados: + clients: [client.0] + ops: 400000 + max_seconds: 600 + max_in_flight: 64 + objects: 1024 + size: 16384 + ec_pool: true + erasure_code_use_overwrites: true + fast_read: true + op_weights: + read: 100 + write: 100 + append: 100 + delete: 50 + snap_create: 50 + snap_remove: 50 + rollback: 50 + copy_from: 50 + setattr: 25 + rmattr: 25 diff --git a/qa/suites/rados/thrash-erasure-code-overwrites/workloads/ec-small-objects-overwrites.yaml b/qa/suites/rados/thrash-erasure-code-overwrites/workloads/ec-small-objects-overwrites.yaml new file mode 100644 index 000000000..9baacef48 --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code-overwrites/workloads/ec-small-objects-overwrites.yaml @@ -0,0 +1,28 @@ +overrides: + ceph: + conf: + global: + enable experimental unrecoverable data corrupting features: '*' + thrashosds: + disable_objectstore_tool_tests: true +tasks: +- rados: + clients: [client.0] + ops: 400000 + max_seconds: 600 + max_in_flight: 64 + objects: 1024 + size: 16384 + ec_pool: true + erasure_code_use_overwrites: true + op_weights: + read: 100 + write: 100 + append: 100 + delete: 50 + snap_create: 50 + snap_remove: 50 + rollback: 50 + copy_from: 50 + setattr: 25 + rmattr: 25 diff --git a/qa/suites/rados/thrash-erasure-code-overwrites/workloads/ec-snaps-few-objects-overwrites.yaml b/qa/suites/rados/thrash-erasure-code-overwrites/workloads/ec-snaps-few-objects-overwrites.yaml new file mode 100644 index 000000000..b7c538199 --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code-overwrites/workloads/ec-snaps-few-objects-overwrites.yaml @@ -0,0 +1,22 @@ +overrides: + ceph: + conf: + global: + enable experimental unrecoverable data corrupting features: '*' + thrashosds: + disable_objectstore_tool_tests: true +tasks: +- rados: + clients: [client.0] + ops: 4000 + objects: 50 + ec_pool: true + erasure_code_use_overwrites: true + op_weights: + read: 100 + write: 100 + delete: 50 + snap_create: 50 + snap_remove: 50 + rollback: 50 + copy_from: 50 diff --git a/qa/suites/rados/thrash-erasure-code-shec/% b/qa/suites/rados/thrash-erasure-code-shec/% new file mode 100644 index 000000000..e69de29bb diff --git a/qa/suites/rados/thrash-erasure-code-shec/.qa b/qa/suites/rados/thrash-erasure-code-shec/.qa new file mode 120000 index 000000000..a602a0353 --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code-shec/.qa @@ -0,0 +1 @@ +../.qa/ \ No newline at end of file diff --git a/qa/suites/rados/thrash-erasure-code-shec/ceph.yaml b/qa/suites/rados/thrash-erasure-code-shec/ceph.yaml new file mode 120000 index 000000000..a2fd139cb --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code-shec/ceph.yaml @@ -0,0 +1 @@ +../thrash/ceph.yaml \ No newline at end of file diff --git a/qa/suites/rados/thrash-erasure-code-shec/clusters/+ b/qa/suites/rados/thrash-erasure-code-shec/clusters/+ new file mode 100644 index 000000000..e69de29bb diff --git a/qa/suites/rados/thrash-erasure-code-shec/clusters/.qa b/qa/suites/rados/thrash-erasure-code-shec/clusters/.qa new file mode 120000 index 000000000..a602a0353 --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code-shec/clusters/.qa @@ -0,0 +1 @@ +../.qa/ \ No newline at end of file diff --git a/qa/suites/rados/thrash-erasure-code-shec/clusters/fixed-4.yaml b/qa/suites/rados/thrash-erasure-code-shec/clusters/fixed-4.yaml new file mode 120000 index 000000000..aa8830071 --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code-shec/clusters/fixed-4.yaml @@ -0,0 +1 @@ +.qa/clusters/fixed-4.yaml \ No newline at end of file diff --git a/qa/suites/rados/thrash-erasure-code-shec/clusters/openstack.yaml b/qa/suites/rados/thrash-erasure-code-shec/clusters/openstack.yaml new file mode 100644 index 000000000..e559d9126 --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code-shec/clusters/openstack.yaml @@ -0,0 +1,4 @@ +openstack: + - volumes: # attached to each instance + count: 4 + size: 10 # GB diff --git a/qa/suites/rados/thrash-erasure-code-shec/mon_election b/qa/suites/rados/thrash-erasure-code-shec/mon_election new file mode 120000 index 000000000..3f331e621 --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code-shec/mon_election @@ -0,0 +1 @@ +.qa/mon_election \ No newline at end of file diff --git a/qa/suites/rados/thrash-erasure-code-shec/msgr-failures b/qa/suites/rados/thrash-erasure-code-shec/msgr-failures new file mode 120000 index 000000000..03689aa44 --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code-shec/msgr-failures @@ -0,0 +1 @@ +../thrash/msgr-failures \ No newline at end of file diff --git a/qa/suites/rados/thrash-erasure-code-shec/objectstore b/qa/suites/rados/thrash-erasure-code-shec/objectstore new file mode 120000 index 000000000..848c65f9e --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code-shec/objectstore @@ -0,0 +1 @@ +.qa/objectstore_debug \ No newline at end of file diff --git a/qa/suites/rados/thrash-erasure-code-shec/rados.yaml b/qa/suites/rados/thrash-erasure-code-shec/rados.yaml new file mode 120000 index 000000000..d256979c0 --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code-shec/rados.yaml @@ -0,0 +1 @@ +.qa/config/rados.yaml \ No newline at end of file diff --git a/qa/suites/rados/thrash-erasure-code-shec/recovery-overrides b/qa/suites/rados/thrash-erasure-code-shec/recovery-overrides new file mode 120000 index 000000000..1957f2c42 --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code-shec/recovery-overrides @@ -0,0 +1 @@ +../thrash/2-recovery-overrides \ No newline at end of file diff --git a/qa/suites/rados/thrash-erasure-code-shec/supported-random-distro$ b/qa/suites/rados/thrash-erasure-code-shec/supported-random-distro$ new file mode 120000 index 000000000..7cef21eef --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code-shec/supported-random-distro$ @@ -0,0 +1 @@ +../basic/supported-random-distro$ \ No newline at end of file diff --git a/qa/suites/rados/thrash-erasure-code-shec/thrashers/.qa b/qa/suites/rados/thrash-erasure-code-shec/thrashers/.qa new file mode 120000 index 000000000..a602a0353 --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code-shec/thrashers/.qa @@ -0,0 +1 @@ +../.qa/ \ No newline at end of file diff --git a/qa/suites/rados/thrash-erasure-code-shec/thrashers/careful.yaml b/qa/suites/rados/thrash-erasure-code-shec/thrashers/careful.yaml new file mode 100644 index 000000000..94b5d07f8 --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code-shec/thrashers/careful.yaml @@ -0,0 +1,20 @@ +overrides: + ceph: + log-ignorelist: + - but it is still running + - objects unfound and apparently lost + - slow request + conf: + osd: + osd debug reject backfill probability: .3 + osd scrub min interval: 60 + osd scrub max interval: 120 + osd max backfills: 3 +tasks: +- thrashosds: + timeout: 1200 + chance_pgnum_grow: 1 + chance_pgnum_shrink: 1 + chance_pgpnum_fix: 1 + min_in: 8 + aggressive_pg_num_changes: false diff --git a/qa/suites/rados/thrash-erasure-code-shec/thrashers/default.yaml b/qa/suites/rados/thrash-erasure-code-shec/thrashers/default.yaml new file mode 100644 index 000000000..a869369fa --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code-shec/thrashers/default.yaml @@ -0,0 +1,21 @@ +overrides: + ceph: + log-ignorelist: + - but it is still running + - objects unfound and apparently lost + - slow request + conf: + osd: + osd debug reject backfill probability: .1 + osd scrub min interval: 60 + osd scrub max interval: 120 + osd max backfills: 3 +tasks: +- thrashosds: + timeout: 1200 + chance_pgnum_grow: 1 + chance_pgnum_shrink: 1 + chance_pgpnum_fix: 1 + min_in: 8 + chance_bluestore_reshard: 1 + bluestore_new_sharding: random diff --git a/qa/suites/rados/thrash-erasure-code-shec/thrashosds-health.yaml b/qa/suites/rados/thrash-erasure-code-shec/thrashosds-health.yaml new file mode 120000 index 000000000..9124eb1aa --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code-shec/thrashosds-health.yaml @@ -0,0 +1 @@ +.qa/tasks/thrashosds-health.yaml \ No newline at end of file diff --git a/qa/suites/rados/thrash-erasure-code-shec/workloads/.qa b/qa/suites/rados/thrash-erasure-code-shec/workloads/.qa new file mode 120000 index 000000000..a602a0353 --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code-shec/workloads/.qa @@ -0,0 +1 @@ +../.qa/ \ No newline at end of file diff --git a/qa/suites/rados/thrash-erasure-code-shec/workloads/ec-rados-plugin=shec-k=4-m=3-c=2.yaml b/qa/suites/rados/thrash-erasure-code-shec/workloads/ec-rados-plugin=shec-k=4-m=3-c=2.yaml new file mode 120000 index 000000000..8f318cc33 --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code-shec/workloads/ec-rados-plugin=shec-k=4-m=3-c=2.yaml @@ -0,0 +1 @@ +.qa/erasure-code/ec-rados-plugin=shec-k=4-m=3-c=2.yaml \ No newline at end of file diff --git a/qa/suites/rados/thrash-erasure-code/% b/qa/suites/rados/thrash-erasure-code/% new file mode 100644 index 000000000..e69de29bb diff --git a/qa/suites/rados/thrash-erasure-code/.qa b/qa/suites/rados/thrash-erasure-code/.qa new file mode 120000 index 000000000..a602a0353 --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code/.qa @@ -0,0 +1 @@ +../.qa/ \ No newline at end of file diff --git a/qa/suites/rados/thrash-erasure-code/ceph.yaml b/qa/suites/rados/thrash-erasure-code/ceph.yaml new file mode 100644 index 000000000..2030acb90 --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code/ceph.yaml @@ -0,0 +1,3 @@ +tasks: +- install: +- ceph: diff --git a/qa/suites/rados/thrash-erasure-code/clusters b/qa/suites/rados/thrash-erasure-code/clusters new file mode 120000 index 000000000..7aac47be3 --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code/clusters @@ -0,0 +1 @@ +../thrash/clusters \ No newline at end of file diff --git a/qa/suites/rados/thrash-erasure-code/fast/.qa b/qa/suites/rados/thrash-erasure-code/fast/.qa new file mode 120000 index 000000000..a602a0353 --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code/fast/.qa @@ -0,0 +1 @@ +../.qa/ \ No newline at end of file diff --git a/qa/suites/rados/thrash-erasure-code/fast/fast.yaml b/qa/suites/rados/thrash-erasure-code/fast/fast.yaml new file mode 100644 index 000000000..8ebfee0a9 --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code/fast/fast.yaml @@ -0,0 +1,5 @@ +overrides: + ceph: + conf: + global: + osd pool default ec fast read: true diff --git a/qa/suites/rados/thrash-erasure-code/fast/normal.yaml b/qa/suites/rados/thrash-erasure-code/fast/normal.yaml new file mode 100644 index 000000000..e69de29bb diff --git a/qa/suites/rados/thrash-erasure-code/mon_election b/qa/suites/rados/thrash-erasure-code/mon_election new file mode 120000 index 000000000..3f331e621 --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code/mon_election @@ -0,0 +1 @@ +.qa/mon_election \ No newline at end of file diff --git a/qa/suites/rados/thrash-erasure-code/msgr-failures b/qa/suites/rados/thrash-erasure-code/msgr-failures new file mode 120000 index 000000000..03689aa44 --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code/msgr-failures @@ -0,0 +1 @@ +../thrash/msgr-failures \ No newline at end of file diff --git a/qa/suites/rados/thrash-erasure-code/objectstore b/qa/suites/rados/thrash-erasure-code/objectstore new file mode 120000 index 000000000..848c65f9e --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code/objectstore @@ -0,0 +1 @@ +.qa/objectstore_debug \ No newline at end of file diff --git a/qa/suites/rados/thrash-erasure-code/rados.yaml b/qa/suites/rados/thrash-erasure-code/rados.yaml new file mode 120000 index 000000000..d256979c0 --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code/rados.yaml @@ -0,0 +1 @@ +.qa/config/rados.yaml \ No newline at end of file diff --git a/qa/suites/rados/thrash-erasure-code/recovery-overrides b/qa/suites/rados/thrash-erasure-code/recovery-overrides new file mode 120000 index 000000000..1957f2c42 --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code/recovery-overrides @@ -0,0 +1 @@ +../thrash/2-recovery-overrides \ No newline at end of file diff --git a/qa/suites/rados/thrash-erasure-code/supported-random-distro$ b/qa/suites/rados/thrash-erasure-code/supported-random-distro$ new file mode 120000 index 000000000..7cef21eef --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code/supported-random-distro$ @@ -0,0 +1 @@ +../basic/supported-random-distro$ \ No newline at end of file diff --git a/qa/suites/rados/thrash-erasure-code/thrashers/.qa b/qa/suites/rados/thrash-erasure-code/thrashers/.qa new file mode 120000 index 000000000..a602a0353 --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code/thrashers/.qa @@ -0,0 +1 @@ +../.qa/ \ No newline at end of file diff --git a/qa/suites/rados/thrash-erasure-code/thrashers/careful.yaml b/qa/suites/rados/thrash-erasure-code/thrashers/careful.yaml new file mode 100644 index 000000000..9a708db31 --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code/thrashers/careful.yaml @@ -0,0 +1,19 @@ +overrides: + ceph: + log-ignorelist: + - but it is still running + - objects unfound and apparently lost + conf: + osd: + osd debug reject backfill probability: .3 + osd scrub min interval: 60 + osd scrub max interval: 120 + osd max backfills: 2 +tasks: +- thrashosds: + timeout: 1200 + chance_pgnum_grow: 1 + chance_pgnum_shrink: 1 + chance_pgpnum_fix: 1 + min_in: 4 + aggressive_pg_num_changes: false diff --git a/qa/suites/rados/thrash-erasure-code/thrashers/default.yaml b/qa/suites/rados/thrash-erasure-code/thrashers/default.yaml new file mode 100644 index 000000000..3728bd8e7 --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code/thrashers/default.yaml @@ -0,0 +1,20 @@ +overrides: + ceph: + log-ignorelist: + - but it is still running + - objects unfound and apparently lost + conf: + osd: + osd debug reject backfill probability: .1 + osd scrub min interval: 60 + osd scrub max interval: 120 + osd max backfills: 2 +tasks: +- thrashosds: + timeout: 1200 + chance_pgnum_grow: 1 + chance_pgnum_shrink: 1 + chance_pgpnum_fix: 1 + min_in: 4 + chance_bluestore_reshard: 1 + bluestore_new_sharding: random diff --git a/qa/suites/rados/thrash-erasure-code/thrashers/fastread.yaml b/qa/suites/rados/thrash-erasure-code/thrashers/fastread.yaml new file mode 100644 index 000000000..4ba4eb2a2 --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code/thrashers/fastread.yaml @@ -0,0 +1,20 @@ +overrides: + ceph: + log-ignorelist: + - but it is still running + - objects unfound and apparently lost + conf: + mon: + osd pool default ec fast read: true + osd: + osd debug reject backfill probability: .1 + osd scrub min interval: 60 + osd scrub max interval: 120 + osd max backfills: 3 +tasks: +- thrashosds: + timeout: 1200 + chance_pgnum_grow: 1 + chance_pgnum_shrink: 1 + chance_pgpnum_fix: 1 + min_in: 4 diff --git a/qa/suites/rados/thrash-erasure-code/thrashers/minsize_recovery.yaml b/qa/suites/rados/thrash-erasure-code/thrashers/minsize_recovery.yaml new file mode 100644 index 000000000..8362b6b1d --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code/thrashers/minsize_recovery.yaml @@ -0,0 +1,18 @@ +overrides: + ceph: + log-ignorelist: + - but it is still running + - objects unfound and apparently lost + create_rbd_pool: False + pre-mgr-commands: + - sudo ceph config set mgr mgr/devicehealth/enable_monitoring false --force + conf: + osd: + osd debug reject backfill probability: .3 + osd scrub min interval: 60 + osd scrub max interval: 120 + osd max backfills: 2 +tasks: +- thrashosds: + timeout: 1200 + chance_test_min_size: 3 diff --git a/qa/suites/rados/thrash-erasure-code/thrashers/morepggrow.yaml b/qa/suites/rados/thrash-erasure-code/thrashers/morepggrow.yaml new file mode 100644 index 000000000..a3e66e693 --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code/thrashers/morepggrow.yaml @@ -0,0 +1,16 @@ +overrides: + ceph: + conf: + osd: + osd scrub min interval: 60 + osd scrub max interval: 120 + osd max backfills: 9 + log-ignorelist: + - but it is still running + - objects unfound and apparently lost +tasks: +- thrashosds: + timeout: 1200 + chance_pgnum_grow: 3 + chance_pgpnum_fix: 1 + min_in: 4 diff --git a/qa/suites/rados/thrash-erasure-code/thrashers/pggrow.yaml b/qa/suites/rados/thrash-erasure-code/thrashers/pggrow.yaml new file mode 100644 index 000000000..98f87d6df --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code/thrashers/pggrow.yaml @@ -0,0 +1,16 @@ +overrides: + ceph: + log-ignorelist: + - but it is still running + - objects unfound and apparently lost + conf: + osd: + osd scrub min interval: 60 + osd scrub max interval: 120 + osd max backfills: 4 +tasks: +- thrashosds: + timeout: 1200 + chance_pgnum_grow: 2 + chance_pgpnum_fix: 1 + min_in: 4 diff --git a/qa/suites/rados/thrash-erasure-code/thrashosds-health.yaml b/qa/suites/rados/thrash-erasure-code/thrashosds-health.yaml new file mode 120000 index 000000000..9124eb1aa --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code/thrashosds-health.yaml @@ -0,0 +1 @@ +.qa/tasks/thrashosds-health.yaml \ No newline at end of file diff --git a/qa/suites/rados/thrash-erasure-code/workloads/.qa b/qa/suites/rados/thrash-erasure-code/workloads/.qa new file mode 120000 index 000000000..a602a0353 --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code/workloads/.qa @@ -0,0 +1 @@ +../.qa/ \ No newline at end of file diff --git a/qa/suites/rados/thrash-erasure-code/workloads/ec-rados-plugin=clay-k=4-m=2.yaml b/qa/suites/rados/thrash-erasure-code/workloads/ec-rados-plugin=clay-k=4-m=2.yaml new file mode 120000 index 000000000..08155ed6e --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code/workloads/ec-rados-plugin=clay-k=4-m=2.yaml @@ -0,0 +1 @@ +.qa/erasure-code/ec-rados-plugin=clay-k=4-m=2.yaml \ No newline at end of file diff --git a/qa/suites/rados/thrash-erasure-code/workloads/ec-rados-plugin=jerasure-k=2-m=1.yaml b/qa/suites/rados/thrash-erasure-code/workloads/ec-rados-plugin=jerasure-k=2-m=1.yaml new file mode 120000 index 000000000..af6d8042f --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code/workloads/ec-rados-plugin=jerasure-k=2-m=1.yaml @@ -0,0 +1 @@ +.qa/erasure-code/ec-rados-plugin=jerasure-k=2-m=1.yaml \ No newline at end of file diff --git a/qa/suites/rados/thrash-erasure-code/workloads/ec-rados-plugin=jerasure-k=3-m=1.yaml b/qa/suites/rados/thrash-erasure-code/workloads/ec-rados-plugin=jerasure-k=3-m=1.yaml new file mode 120000 index 000000000..cdf551995 --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code/workloads/ec-rados-plugin=jerasure-k=3-m=1.yaml @@ -0,0 +1 @@ +.qa/erasure-code/ec-rados-plugin=jerasure-k=3-m=1.yaml \ No newline at end of file diff --git a/qa/suites/rados/thrash-erasure-code/workloads/ec-radosbench.yaml b/qa/suites/rados/thrash-erasure-code/workloads/ec-radosbench.yaml new file mode 100644 index 000000000..3c2ff7af0 --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code/workloads/ec-radosbench.yaml @@ -0,0 +1,27 @@ +tasks: +- full_sequential: + - radosbench: + clients: [client.0] + time: 150 + unique_pool: true + ec_pool: true + - radosbench: + clients: [client.0] + time: 150 + unique_pool: true + ec_pool: true + - radosbench: + clients: [client.0] + time: 150 + unique_pool: true + ec_pool: true + - radosbench: + clients: [client.0] + time: 150 + unique_pool: true + ec_pool: true + - radosbench: + clients: [client.0] + time: 150 + unique_pool: true + ec_pool: true diff --git a/qa/suites/rados/thrash-erasure-code/workloads/ec-small-objects-balanced.yaml b/qa/suites/rados/thrash-erasure-code/workloads/ec-small-objects-balanced.yaml new file mode 100644 index 000000000..af0ac3931 --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code/workloads/ec-small-objects-balanced.yaml @@ -0,0 +1,21 @@ +tasks: +- rados: + clients: [client.0] + ops: 400000 + max_seconds: 600 + max_in_flight: 64 + objects: 1024 + size: 16384 + ec_pool: true + balanced_reads: true + op_weights: + read: 100 + write: 0 + append: 100 + delete: 50 + snap_create: 50 + snap_remove: 50 + rollback: 50 + copy_from: 50 + setattr: 25 + rmattr: 25 diff --git a/qa/suites/rados/thrash-erasure-code/workloads/ec-small-objects-fast-read.yaml b/qa/suites/rados/thrash-erasure-code/workloads/ec-small-objects-fast-read.yaml new file mode 100644 index 000000000..e732ec6fa --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code/workloads/ec-small-objects-fast-read.yaml @@ -0,0 +1,21 @@ +tasks: +- rados: + clients: [client.0] + ops: 400000 + max_seconds: 600 + max_in_flight: 64 + objects: 1024 + size: 16384 + ec_pool: true + fast_read: true + op_weights: + read: 100 + write: 0 + append: 100 + delete: 50 + snap_create: 50 + snap_remove: 50 + rollback: 50 + copy_from: 50 + setattr: 25 + rmattr: 25 diff --git a/qa/suites/rados/thrash-erasure-code/workloads/ec-small-objects-many-deletes.yaml b/qa/suites/rados/thrash-erasure-code/workloads/ec-small-objects-many-deletes.yaml new file mode 100644 index 000000000..25b38e14f --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code/workloads/ec-small-objects-many-deletes.yaml @@ -0,0 +1,14 @@ +tasks: +- rados: + clients: [client.0] + ops: 400000 + max_seconds: 600 + max_in_flight: 8 + objects: 20 + size: 16384 + ec_pool: true + op_weights: + write: 0 + read: 0 + append: 10 + delete: 20 diff --git a/qa/suites/rados/thrash-erasure-code/workloads/ec-small-objects.yaml b/qa/suites/rados/thrash-erasure-code/workloads/ec-small-objects.yaml new file mode 100644 index 000000000..a8ac39716 --- /dev/null +++ b/qa/suites/rados/thrash-erasure-code/workloads/ec-small-objects.yaml @@ -0,0 +1,20 @@ +tasks: +- rados: + clients: [client.0] + ops: 400000 + max_seconds: 600 + max_in_flight: 64 + objects: 1024 + size: 16384 + ec_pool: true + op_weights: + read: 100 + write: 0 + append: 100 + delete: 50 + snap_create: 50 + snap_remove: 50 + rollback: 50 + copy_from: 50 + setattr: 25 + rmattr: 25 diff --git a/qa/suites/rados/thrash/% b/qa/suites/rados/thrash/% new file mode 100644 index 000000000..e69de29bb diff --git a/qa/suites/rados/thrash/.qa b/qa/suites/rados/thrash/.qa new file mode 120000 index 000000000..a602a0353 --- /dev/null +++ b/qa/suites/rados/thrash/.qa @@ -0,0 +1 @@ +../.qa/ \ No newline at end of file diff --git a/qa/suites/rados/thrash/0-size-min-size-overrides/.qa b/qa/suites/rados/thrash/0-size-min-size-overrides/.qa new file mode 120000 index 000000000..a602a0353 --- /dev/null +++ b/qa/suites/rados/thrash/0-size-min-size-overrides/.qa @@ -0,0 +1 @@ +../.qa/ \ No newline at end of file diff --git a/qa/suites/rados/thrash/0-size-min-size-overrides/2-size-2-min-size.yaml b/qa/suites/rados/thrash/0-size-min-size-overrides/2-size-2-min-size.yaml new file mode 120000 index 000000000..5393a7554 --- /dev/null +++ b/qa/suites/rados/thrash/0-size-min-size-overrides/2-size-2-min-size.yaml @@ -0,0 +1 @@ +.qa/overrides/2-size-2-min-size.yaml \ No newline at end of file diff --git a/qa/suites/rados/thrash/0-size-min-size-overrides/3-size-2-min-size.yaml b/qa/suites/rados/thrash/0-size-min-size-overrides/3-size-2-min-size.yaml new file mode 120000 index 000000000..5ff70eadf --- /dev/null +++ b/qa/suites/rados/thrash/0-size-min-size-overrides/3-size-2-min-size.yaml @@ -0,0 +1 @@ +.qa/overrides/3-size-2-min-size.yaml \ No newline at end of file diff --git a/qa/suites/rados/thrash/1-pg-log-overrides/.qa b/qa/suites/rados/thrash/1-pg-log-overrides/.qa new file mode 120000 index 000000000..a602a0353 --- /dev/null +++ b/qa/suites/rados/thrash/1-pg-log-overrides/.qa @@ -0,0 +1 @@ +../.qa/ \ No newline at end of file diff --git a/qa/suites/rados/thrash/1-pg-log-overrides/normal_pg_log.yaml b/qa/suites/rados/thrash/1-pg-log-overrides/normal_pg_log.yaml new file mode 100644 index 000000000..e69de29bb diff --git a/qa/suites/rados/thrash/1-pg-log-overrides/short_pg_log.yaml b/qa/suites/rados/thrash/1-pg-log-overrides/short_pg_log.yaml new file mode 120000 index 000000000..abd86d7d9 --- /dev/null +++ b/qa/suites/rados/thrash/1-pg-log-overrides/short_pg_log.yaml @@ -0,0 +1 @@ +.qa/overrides/short_pg_log.yaml \ No newline at end of file diff --git a/qa/suites/rados/thrash/2-recovery-overrides/$ b/qa/suites/rados/thrash/2-recovery-overrides/$ new file mode 100644 index 000000000..e69de29bb diff --git a/qa/suites/rados/thrash/2-recovery-overrides/.qa b/qa/suites/rados/thrash/2-recovery-overrides/.qa new file mode 120000 index 000000000..a602a0353 --- /dev/null +++ b/qa/suites/rados/thrash/2-recovery-overrides/.qa @@ -0,0 +1 @@ +../.qa/ \ No newline at end of file diff --git a/qa/suites/rados/thrash/2-recovery-overrides/default.yaml b/qa/suites/rados/thrash/2-recovery-overrides/default.yaml new file mode 100644 index 000000000..e69de29bb diff --git a/qa/suites/rados/thrash/2-recovery-overrides/more-active-recovery.yaml b/qa/suites/rados/thrash/2-recovery-overrides/more-active-recovery.yaml new file mode 120000 index 000000000..47afd7020 --- /dev/null +++ b/qa/suites/rados/thrash/2-recovery-overrides/more-active-recovery.yaml @@ -0,0 +1 @@ +.qa/overrides/more-active-recovery.yaml \ No newline at end of file diff --git a/qa/suites/rados/thrash/2-recovery-overrides/more-async-partial-recovery.yaml b/qa/suites/rados/thrash/2-recovery-overrides/more-async-partial-recovery.yaml new file mode 100644 index 000000000..0bbc72db7 --- /dev/null +++ b/qa/suites/rados/thrash/2-recovery-overrides/more-async-partial-recovery.yaml @@ -0,0 +1,6 @@ +overrides: + ceph: + conf: + global: + osd_async_recovery_min_cost: 1 + osd_object_clean_region_max_num_intervals: 1000 diff --git a/qa/suites/rados/thrash/2-recovery-overrides/more-async-recovery.yaml b/qa/suites/rados/thrash/2-recovery-overrides/more-async-recovery.yaml new file mode 100644 index 000000000..4aed086bc --- /dev/null +++ b/qa/suites/rados/thrash/2-recovery-overrides/more-async-recovery.yaml @@ -0,0 +1,5 @@ +overrides: + ceph: + conf: + global: + osd_async_recovery_min_cost: 1 diff --git a/qa/suites/rados/thrash/2-recovery-overrides/more-partial-recovery.yaml b/qa/suites/rados/thrash/2-recovery-overrides/more-partial-recovery.yaml new file mode 100644 index 000000000..88f15f2f6 --- /dev/null +++ b/qa/suites/rados/thrash/2-recovery-overrides/more-partial-recovery.yaml @@ -0,0 +1,5 @@ +overrides: + ceph: + conf: + global: + osd_object_clean_region_max_num_intervals: 1000 diff --git a/qa/suites/rados/thrash/3-scrub-overrides/$ b/qa/suites/rados/thrash/3-scrub-overrides/$ new file mode 100644 index 000000000..e69de29bb diff --git a/qa/suites/rados/thrash/3-scrub-overrides/.qa b/qa/suites/rados/thrash/3-scrub-overrides/.qa new file mode 120000 index 000000000..fea2489fd --- /dev/null +++ b/qa/suites/rados/thrash/3-scrub-overrides/.qa @@ -0,0 +1 @@ +../.qa \ No newline at end of file diff --git a/qa/suites/rados/thrash/3-scrub-overrides/default.yaml b/qa/suites/rados/thrash/3-scrub-overrides/default.yaml new file mode 100644 index 000000000..e69de29bb diff --git a/qa/suites/rados/thrash/3-scrub-overrides/max-simultaneous-scrubs-2.yaml b/qa/suites/rados/thrash/3-scrub-overrides/max-simultaneous-scrubs-2.yaml new file mode 100644 index 000000000..abf852e98 --- /dev/null +++ b/qa/suites/rados/thrash/3-scrub-overrides/max-simultaneous-scrubs-2.yaml @@ -0,0 +1,5 @@ +overrides: + ceph: + conf: + osd: + osd max scrubs: 2 diff --git a/qa/suites/rados/thrash/3-scrub-overrides/max-simultaneous-scrubs-3.yaml b/qa/suites/rados/thrash/3-scrub-overrides/max-simultaneous-scrubs-3.yaml new file mode 100644 index 000000000..3b3dfd61f --- /dev/null +++ b/qa/suites/rados/thrash/3-scrub-overrides/max-simultaneous-scrubs-3.yaml @@ -0,0 +1,5 @@ +overrides: + ceph: + conf: + osd: + osd max scrubs: 3 diff --git a/qa/suites/rados/thrash/backoff/.qa b/qa/suites/rados/thrash/backoff/.qa new file mode 120000 index 000000000..a602a0353 --- /dev/null +++ b/qa/suites/rados/thrash/backoff/.qa @@ -0,0 +1 @@ +../.qa/ \ No newline at end of file diff --git a/qa/suites/rados/thrash/backoff/normal.yaml b/qa/suites/rados/thrash/backoff/normal.yaml new file mode 100644 index 000000000..e69de29bb diff --git a/qa/suites/rados/thrash/backoff/peering.yaml b/qa/suites/rados/thrash/backoff/peering.yaml new file mode 100644 index 000000000..66d06117e --- /dev/null +++ b/qa/suites/rados/thrash/backoff/peering.yaml @@ -0,0 +1,5 @@ +overrides: + ceph: + conf: + osd: + osd backoff on peering: true diff --git a/qa/suites/rados/thrash/backoff/peering_and_degraded.yaml b/qa/suites/rados/thrash/backoff/peering_and_degraded.yaml new file mode 100644 index 000000000..e61099065 --- /dev/null +++ b/qa/suites/rados/thrash/backoff/peering_and_degraded.yaml @@ -0,0 +1,6 @@ +overrides: + ceph: + conf: + osd: + osd backoff on peering: true + osd backoff on degraded: true diff --git a/qa/suites/rados/thrash/ceph.yaml b/qa/suites/rados/thrash/ceph.yaml new file mode 100644 index 000000000..67393c564 --- /dev/null +++ b/qa/suites/rados/thrash/ceph.yaml @@ -0,0 +1,6 @@ +tasks: +- install: +- ceph: + conf: + osd: + debug monc: 20 diff --git a/qa/suites/rados/thrash/clusters/+ b/qa/suites/rados/thrash/clusters/+ new file mode 100644 index 000000000..e69de29bb diff --git a/qa/suites/rados/thrash/clusters/.qa b/qa/suites/rados/thrash/clusters/.qa new file mode 120000 index 000000000..a602a0353 --- /dev/null +++ b/qa/suites/rados/thrash/clusters/.qa @@ -0,0 +1 @@ +../.qa/ \ No newline at end of file diff --git a/qa/suites/rados/thrash/clusters/fixed-2.yaml b/qa/suites/rados/thrash/clusters/fixed-2.yaml new file mode 120000 index 000000000..230ff0fda --- /dev/null +++ b/qa/suites/rados/thrash/clusters/fixed-2.yaml @@ -0,0 +1 @@ +.qa/clusters/fixed-2.yaml \ No newline at end of file diff --git a/qa/suites/rados/thrash/clusters/openstack.yaml b/qa/suites/rados/thrash/clusters/openstack.yaml new file mode 100644 index 000000000..b0f3b9b4d --- /dev/null +++ b/qa/suites/rados/thrash/clusters/openstack.yaml @@ -0,0 +1,4 @@ +openstack: + - volumes: # attached to each instance + count: 4 + size: 30 # GB diff --git a/qa/suites/rados/thrash/crc-failures/bad_map_crc_failure.yaml b/qa/suites/rados/thrash/crc-failures/bad_map_crc_failure.yaml new file mode 100644 index 000000000..5bbb4385e --- /dev/null +++ b/qa/suites/rados/thrash/crc-failures/bad_map_crc_failure.yaml @@ -0,0 +1,7 @@ +overrides: + ceph: + conf: + osd: + osd inject bad map crc probability: 0.1 + log-ignorelist: + - failed to encode map diff --git a/qa/suites/rados/thrash/crc-failures/default.yaml b/qa/suites/rados/thrash/crc-failures/default.yaml new file mode 100644 index 000000000..e69de29bb diff --git a/qa/suites/rados/thrash/d-balancer/.qa b/qa/suites/rados/thrash/d-balancer/.qa new file mode 120000 index 000000000..a602a0353 --- /dev/null +++ b/qa/suites/rados/thrash/d-balancer/.qa @@ -0,0 +1 @@ +../.qa/ \ No newline at end of file diff --git a/qa/suites/rados/thrash/d-balancer/crush-compat.yaml b/qa/suites/rados/thrash/d-balancer/crush-compat.yaml new file mode 100644 index 000000000..aa867660d --- /dev/null +++ b/qa/suites/rados/thrash/d-balancer/crush-compat.yaml @@ -0,0 +1,6 @@ +tasks: +- exec: + mon.a: + - while ! ceph balancer status ; do sleep 1 ; done + - ceph balancer mode crush-compat + - ceph balancer on diff --git a/qa/suites/rados/thrash/d-balancer/on.yaml b/qa/suites/rados/thrash/d-balancer/on.yaml new file mode 100644 index 000000000..e69de29bb diff --git a/qa/suites/rados/thrash/mon_election b/qa/suites/rados/thrash/mon_election new file mode 120000 index 000000000..3f331e621 --- /dev/null +++ b/qa/suites/rados/thrash/mon_election @@ -0,0 +1 @@ +.qa/mon_election \ No newline at end of file diff --git a/qa/suites/rados/thrash/msgr b/qa/suites/rados/thrash/msgr new file mode 120000 index 000000000..57bee80db --- /dev/null +++ b/qa/suites/rados/thrash/msgr @@ -0,0 +1 @@ +.qa/msgr \ No newline at end of file diff --git a/qa/suites/rados/thrash/msgr-failures/.qa b/qa/suites/rados/thrash/msgr-failures/.qa new file mode 120000 index 000000000..a602a0353 --- /dev/null +++ b/qa/suites/rados/thrash/msgr-failures/.qa @@ -0,0 +1 @@ +../.qa/ \ No newline at end of file diff --git a/qa/suites/rados/thrash/msgr-failures/fastclose.yaml b/qa/suites/rados/thrash/msgr-failures/fastclose.yaml new file mode 100644 index 000000000..ec45f8882 --- /dev/null +++ b/qa/suites/rados/thrash/msgr-failures/fastclose.yaml @@ -0,0 +1,9 @@ +overrides: + ceph: + conf: + global: + ms inject socket failures: 2500 + ms tcp read timeout: 5 + mon client directed command retry: 5 + log-ignorelist: + - \(OSD_SLOW_PING_TIME diff --git a/qa/suites/rados/thrash/msgr-failures/few.yaml b/qa/suites/rados/thrash/msgr-failures/few.yaml new file mode 100644 index 000000000..cc9a3ae69 --- /dev/null +++ b/qa/suites/rados/thrash/msgr-failures/few.yaml @@ -0,0 +1,10 @@ +overrides: + ceph: + conf: + global: + ms inject socket failures: 5000 + mon client directed command retry: 5 + osd: + osd heartbeat use min delay socket: true + log-ignorelist: + - \(OSD_SLOW_PING_TIME diff --git a/qa/suites/rados/thrash/msgr-failures/osd-delay.yaml b/qa/suites/rados/thrash/msgr-failures/osd-delay.yaml new file mode 100644 index 000000000..d7cec6f36 --- /dev/null +++ b/qa/suites/rados/thrash/msgr-failures/osd-delay.yaml @@ -0,0 +1,12 @@ +overrides: + ceph: + conf: + global: + ms inject socket failures: 2500 + ms inject delay type: osd + ms inject delay probability: .005 + ms inject delay max: 1 + ms inject internal delays: .002 + mon client directed command retry: 5 + log-ignorelist: + - \(OSD_SLOW_PING_TIME diff --git a/qa/suites/rados/thrash/msgr-failures/osd-dispatch-delay.yaml b/qa/suites/rados/thrash/msgr-failures/osd-dispatch-delay.yaml new file mode 100644 index 000000000..aff059fb8 --- /dev/null +++ b/qa/suites/rados/thrash/msgr-failures/osd-dispatch-delay.yaml @@ -0,0 +1,7 @@ +overrides: + ceph: + conf: + global: + osd debug inject dispatch delay duration: 0.1 + osd debug inject dispatch delay probability: 0.1 + diff --git a/qa/suites/rados/thrash/objectstore b/qa/suites/rados/thrash/objectstore new file mode 120000 index 000000000..848c65f9e --- /dev/null +++ b/qa/suites/rados/thrash/objectstore @@ -0,0 +1 @@ +.qa/objectstore_debug \ No newline at end of file diff --git a/qa/suites/rados/thrash/rados.yaml b/qa/suites/rados/thrash/rados.yaml new file mode 120000 index 000000000..d256979c0 --- /dev/null +++ b/qa/suites/rados/thrash/rados.yaml @@ -0,0 +1 @@ +.qa/config/rados.yaml \ No newline at end of file diff --git a/qa/suites/rados/thrash/supported-random-distro$ b/qa/suites/rados/thrash/supported-random-distro$ new file mode 120000 index 000000000..7cef21eef --- /dev/null +++ b/qa/suites/rados/thrash/supported-random-distro$ @@ -0,0 +1 @@ +../basic/supported-random-distro$ \ No newline at end of file diff --git a/qa/suites/rados/thrash/thrashers/.qa b/qa/suites/rados/thrash/thrashers/.qa new file mode 120000 index 000000000..a602a0353 --- /dev/null +++ b/qa/suites/rados/thrash/thrashers/.qa @@ -0,0 +1 @@ +../.qa/ \ No newline at end of file diff --git a/qa/suites/rados/thrash/thrashers/careful.yaml b/qa/suites/rados/thrash/thrashers/careful.yaml new file mode 100644 index 000000000..8190657f3 --- /dev/null +++ b/qa/suites/rados/thrash/thrashers/careful.yaml @@ -0,0 +1,26 @@ +overrides: + ceph: + log-ignorelist: + - but it is still running + - objects unfound and apparently lost + conf: + osd: + osd debug reject backfill probability: .3 + osd scrub min interval: 60 + osd scrub max interval: 120 + osd max backfills: 3 + osd snap trim sleep: 2 + mon: + mon min osdmap epochs: 50 + paxos service trim min: 10 + # prune full osdmaps regularly + mon osdmap full prune min: 15 + mon osdmap full prune interval: 2 + mon osdmap full prune txsize: 2 +tasks: +- thrashosds: + timeout: 1200 + chance_pgnum_grow: 1 + chance_pgnum_shrink: 1 + chance_pgpnum_fix: 1 + aggressive_pg_num_changes: false diff --git a/qa/suites/rados/thrash/thrashers/default.yaml b/qa/suites/rados/thrash/thrashers/default.yaml new file mode 100644 index 000000000..5a300a9ff --- /dev/null +++ b/qa/suites/rados/thrash/thrashers/default.yaml @@ -0,0 +1,28 @@ +overrides: + ceph: + log-ignorelist: + - but it is still running + - objects unfound and apparently lost + conf: + osd: + osd debug reject backfill probability: .3 + osd scrub min interval: 60 + osd scrub max interval: 120 + osd max backfills: 3 + osd snap trim sleep: 2 + osd delete sleep: 1 + mon: + mon min osdmap epochs: 50 + paxos service trim min: 10 + # prune full osdmaps regularly + mon osdmap full prune min: 15 + mon osdmap full prune interval: 2 + mon osdmap full prune txsize: 2 +tasks: +- thrashosds: + timeout: 1200 + chance_pgnum_grow: 1 + chance_pgnum_shrink: 1 + chance_pgpnum_fix: 1 + chance_bluestore_reshard: 1 + bluestore_new_sharding: random diff --git a/qa/suites/rados/thrash/thrashers/mapgap.yaml b/qa/suites/rados/thrash/thrashers/mapgap.yaml new file mode 100644 index 000000000..3b34f5b6b --- /dev/null +++ b/qa/suites/rados/thrash/thrashers/mapgap.yaml @@ -0,0 +1,27 @@ +overrides: + ceph: + log-ignorelist: + - but it is still running + - objects unfound and apparently lost + - osd_map_cache_size + conf: + mon: + mon min osdmap epochs: 50 + paxos service trim min: 10 + # prune full osdmaps regularly + mon osdmap full prune min: 15 + mon osdmap full prune interval: 2 + mon osdmap full prune txsize: 2 + osd: + osd map cache size: 1 + osd scrub min interval: 60 + osd scrub max interval: 120 + osd scrub during recovery: false + osd max backfills: 6 +tasks: +- thrashosds: + timeout: 1800 + chance_pgnum_grow: 0.25 + chance_pgnum_shrink: 0.25 + chance_pgpnum_fix: 0.25 + chance_test_map_discontinuity: 2 diff --git a/qa/suites/rados/thrash/thrashers/morepggrow.yaml b/qa/suites/rados/thrash/thrashers/morepggrow.yaml new file mode 100644 index 000000000..f18a88711 --- /dev/null +++ b/qa/suites/rados/thrash/thrashers/morepggrow.yaml @@ -0,0 +1,22 @@ +overrides: + ceph: + conf: + osd: + osd scrub min interval: 60 + osd scrub max interval: 120 + journal throttle high multiple: 2 + journal throttle max multiple: 10 + filestore queue throttle high multiple: 2 + filestore queue throttle max multiple: 10 + osd max backfills: 9 + log-ignorelist: + - but it is still running + - objects unfound and apparently lost +tasks: +- thrashosds: + timeout: 1200 + chance_pgnum_grow: 3 + chance_pgpnum_fix: 1 +openstack: +- volumes: + size: 50 diff --git a/qa/suites/rados/thrash/thrashers/none.yaml b/qa/suites/rados/thrash/thrashers/none.yaml new file mode 100644 index 000000000..e69de29bb diff --git a/qa/suites/rados/thrash/thrashers/pggrow.yaml b/qa/suites/rados/thrash/thrashers/pggrow.yaml new file mode 100644 index 000000000..54498d0cf --- /dev/null +++ b/qa/suites/rados/thrash/thrashers/pggrow.yaml @@ -0,0 +1,24 @@ +overrides: + ceph: + log-ignorelist: + - but it is still running + - objects unfound and apparently lost + conf: + osd: + osd scrub min interval: 60 + osd scrub max interval: 120 + filestore odsync write: true + osd max backfills: 2 + osd snap trim sleep: .5 + mon: + mon min osdmap epochs: 50 + paxos service trim min: 10 + # prune full osdmaps regularly + mon osdmap full prune min: 15 + mon osdmap full prune interval: 2 + mon osdmap full prune txsize: 2 +tasks: +- thrashosds: + timeout: 1200 + chance_pgnum_grow: 2 + chance_pgpnum_fix: 1 diff --git a/qa/suites/rados/thrash/thrashosds-health.yaml b/qa/suites/rados/thrash/thrashosds-health.yaml new file mode 120000 index 000000000..9124eb1aa --- /dev/null +++ b/qa/suites/rados/thrash/thrashosds-health.yaml @@ -0,0 +1 @@ +.qa/tasks/thrashosds-health.yaml \ No newline at end of file diff --git a/qa/suites/rados/thrash/workloads/.qa b/qa/suites/rados/thrash/workloads/.qa new file mode 120000 index 000000000..a602a0353 --- /dev/null +++ b/qa/suites/rados/thrash/workloads/.qa @@ -0,0 +1 @@ +../.qa/ \ No newline at end of file diff --git a/qa/suites/rados/thrash/workloads/admin_socket_objecter_requests.yaml b/qa/suites/rados/thrash/workloads/admin_socket_objecter_requests.yaml new file mode 100644 index 000000000..8c9764ade --- /dev/null +++ b/qa/suites/rados/thrash/workloads/admin_socket_objecter_requests.yaml @@ -0,0 +1,13 @@ +overrides: + ceph: + conf: + client.0: + admin socket: /var/run/ceph/ceph-$name.asok +tasks: +- radosbench: + clients: [client.0] + time: 150 +- admin_socket: + client.0: + objecter_requests: + test: "http://git.ceph.com/?p={repo};a=blob_plain;f=src/test/admin_socket/objecter_requests;hb={branch}" diff --git a/qa/suites/rados/thrash/workloads/cache-agent-big.yaml b/qa/suites/rados/thrash/workloads/cache-agent-big.yaml new file mode 100644 index 000000000..3f377858a --- /dev/null +++ b/qa/suites/rados/thrash/workloads/cache-agent-big.yaml @@ -0,0 +1,36 @@ +overrides: + ceph: + log-ignorelist: + - must scrub before tier agent can activate + conf: + osd: + # override short_pg_log_entries.yaml (which sets these under [global]) + osd_min_pg_log_entries: 3000 + osd_max_pg_log_entries: 3000 +tasks: +- exec: + client.0: + - sudo ceph osd erasure-code-profile set myprofile crush-failure-domain=osd m=2 k=2 + - sudo ceph osd pool create base 4 4 erasure myprofile + - sudo ceph osd pool application enable base rados + - sudo ceph osd pool set base min_size 2 + - sudo ceph osd pool create cache 4 + - sudo ceph osd tier add base cache + - sudo ceph osd tier cache-mode cache writeback + - sudo ceph osd tier set-overlay base cache + - sudo ceph osd pool set cache hit_set_type bloom + - sudo ceph osd pool set cache hit_set_count 8 + - sudo ceph osd pool set cache hit_set_period 60 + - sudo ceph osd pool set cache target_max_objects 5000 +- rados: + clients: [client.0] + pools: [base] + ops: 10000 + objects: 6600 + max_seconds: 1200 + size: 1024 + op_weights: + read: 100 + write: 100 + delete: 50 + copy_from: 50 diff --git a/qa/suites/rados/thrash/workloads/cache-agent-small.yaml b/qa/suites/rados/thrash/workloads/cache-agent-small.yaml new file mode 100644 index 000000000..29219a7e8 --- /dev/null +++ b/qa/suites/rados/thrash/workloads/cache-agent-small.yaml @@ -0,0 +1,34 @@ +overrides: + ceph: + log-ignorelist: + - must scrub before tier agent can activate + conf: + osd: + # override short_pg_log_entries.yaml (which sets these under [global]) + osd_min_pg_log_entries: 3000 + osd_max_pg_log_entries: 3000 +tasks: +- exec: + client.0: + - sudo ceph osd pool create base 4 + - sudo ceph osd pool application enable base rados + - sudo ceph osd pool create cache 4 + - sudo ceph osd tier add base cache + - sudo ceph osd tier cache-mode cache writeback + - sudo ceph osd tier set-overlay base cache + - sudo ceph osd pool set cache hit_set_type bloom + - sudo ceph osd pool set cache hit_set_count 8 + - sudo ceph osd pool set cache hit_set_period 60 + - sudo ceph osd pool set cache target_max_objects 250 + - sudo ceph osd pool set cache min_read_recency_for_promote 2 + - sudo ceph osd pool set cache min_write_recency_for_promote 2 +- rados: + clients: [client.0] + pools: [base] + ops: 4000 + objects: 500 + op_weights: + read: 100 + write: 100 + delete: 50 + copy_from: 50 diff --git a/qa/suites/rados/thrash/workloads/cache-pool-snaps-readproxy.yaml b/qa/suites/rados/thrash/workloads/cache-pool-snaps-readproxy.yaml new file mode 100644 index 000000000..808968d6b --- /dev/null +++ b/qa/suites/rados/thrash/workloads/cache-pool-snaps-readproxy.yaml @@ -0,0 +1,39 @@ +overrides: + ceph: + log-ignorelist: + - must scrub before tier agent can activate + conf: + osd: + # override short_pg_log_entries.yaml (which sets these under [global]) + osd_min_pg_log_entries: 3000 + osd_max_pg_log_entries: 3000 +tasks: +- exec: + client.0: + - sudo ceph osd pool create base 4 + - sudo ceph osd pool application enable base rados + - sudo ceph osd pool create cache 4 + - sudo ceph osd tier add base cache + - sudo ceph osd tier cache-mode cache readproxy + - sudo ceph osd tier set-overlay base cache + - sudo ceph osd pool set cache hit_set_type bloom + - sudo ceph osd pool set cache hit_set_count 8 + - sudo ceph osd pool set cache hit_set_period 3600 + - sudo ceph osd pool set cache target_max_objects 250 +- rados: + clients: [client.0] + pools: [base] + ops: 4000 + objects: 500 + pool_snaps: true + op_weights: + read: 100 + write: 100 + delete: 50 + copy_from: 50 + cache_flush: 50 + cache_try_flush: 50 + cache_evict: 50 + snap_create: 50 + snap_remove: 50 + rollback: 50 diff --git a/qa/suites/rados/thrash/workloads/cache-pool-snaps.yaml b/qa/suites/rados/thrash/workloads/cache-pool-snaps.yaml new file mode 100644 index 000000000..4aec8611c --- /dev/null +++ b/qa/suites/rados/thrash/workloads/cache-pool-snaps.yaml @@ -0,0 +1,44 @@ +overrides: + ceph: + log-ignorelist: + - must scrub before tier agent can activate + conf: + osd: + # override short_pg_log_entries.yaml (which sets these under [global]) + osd_min_pg_log_entries: 3000 + osd_max_pg_log_entries: 3000 +tasks: +- exec: + client.0: + - sudo ceph osd pool create base 4 + - sudo ceph osd pool application enable base rados + - sudo ceph osd pool create cache 4 + - sudo ceph osd tier add base cache + - sudo ceph osd tier cache-mode cache writeback + - sudo ceph osd tier set-overlay base cache + - sudo ceph osd pool set cache hit_set_type bloom + - sudo ceph osd pool set cache hit_set_count 8 + - sudo ceph osd pool set cache hit_set_period 3600 + - sudo ceph osd pool set cache target_max_objects 250 + - sudo ceph osd pool set cache min_read_recency_for_promote 0 + - sudo ceph osd pool set cache min_write_recency_for_promote 0 +- rados: + clients: [client.0] + pools: [base] + ops: 4000 + objects: 500 + pool_snaps: true + op_weights: + read: 100 + write: 100 + delete: 50 + copy_from: 50 + cache_flush: 50 + cache_try_flush: 50 + cache_evict: 50 + snap_create: 50 + snap_remove: 50 + rollback: 50 +openstack: + - machine: + ram: 15000 # MB diff --git a/qa/suites/rados/thrash/workloads/cache-snaps-balanced.yaml b/qa/suites/rados/thrash/workloads/cache-snaps-balanced.yaml new file mode 100644 index 000000000..1e55f573f --- /dev/null +++ b/qa/suites/rados/thrash/workloads/cache-snaps-balanced.yaml @@ -0,0 +1,40 @@ +overrides: + ceph: + log-ignorelist: + - must scrub before tier agent can activate + conf: + osd: + # override short_pg_log_entries.yaml (which sets these under [global]) + osd_min_pg_log_entries: 3000 + osd_max_pg_log_entries: 3000 +tasks: +- exec: + client.0: + - sudo ceph osd pool create base 4 + - sudo ceph osd pool application enable base rados + - sudo ceph osd pool create cache 4 + - sudo ceph osd tier add base cache + - sudo ceph osd tier cache-mode cache writeback + - sudo ceph osd tier set-overlay base cache + - sudo ceph osd pool set cache hit_set_type bloom + - sudo ceph osd pool set cache hit_set_count 8 + - sudo ceph osd pool set cache hit_set_period 3600 + - sudo ceph osd pool set cache target_max_objects 250 + - sudo ceph osd pool set cache min_read_recency_for_promote 2 +- rados: + clients: [client.0] + pools: [base] + ops: 4000 + objects: 500 + balance_reads: true + op_weights: + read: 100 + write: 100 + delete: 50 + copy_from: 50 + cache_flush: 50 + cache_try_flush: 50 + cache_evict: 50 + snap_create: 50 + snap_remove: 50 + rollback: 50 diff --git a/qa/suites/rados/thrash/workloads/cache-snaps.yaml b/qa/suites/rados/thrash/workloads/cache-snaps.yaml new file mode 100644 index 000000000..11401d7e2 --- /dev/null +++ b/qa/suites/rados/thrash/workloads/cache-snaps.yaml @@ -0,0 +1,39 @@ +overrides: + ceph: + log-ignorelist: + - must scrub before tier agent can activate + conf: + osd: + # override short_pg_log_entries.yaml (which sets these under [global]) + osd_min_pg_log_entries: 3000 + osd_max_pg_log_entries: 3000 +tasks: +- exec: + client.0: + - sudo ceph osd pool create base 4 + - sudo ceph osd pool application enable base rados + - sudo ceph osd pool create cache 4 + - sudo ceph osd tier add base cache + - sudo ceph osd tier cache-mode cache writeback + - sudo ceph osd tier set-overlay base cache + - sudo ceph osd pool set cache hit_set_type bloom + - sudo ceph osd pool set cache hit_set_count 8 + - sudo ceph osd pool set cache hit_set_period 3600 + - sudo ceph osd pool set cache target_max_objects 250 + - sudo ceph osd pool set cache min_read_recency_for_promote 2 +- rados: + clients: [client.0] + pools: [base] + ops: 4000 + objects: 500 + op_weights: + read: 100 + write: 100 + delete: 50 + copy_from: 50 + cache_flush: 50 + cache_try_flush: 50 + cache_evict: 50 + snap_create: 50 + snap_remove: 50 + rollback: 50 diff --git a/qa/suites/rados/thrash/workloads/cache.yaml b/qa/suites/rados/thrash/workloads/cache.yaml new file mode 100644 index 000000000..c557a6751 --- /dev/null +++ b/qa/suites/rados/thrash/workloads/cache.yaml @@ -0,0 +1,36 @@ +overrides: + ceph: + log-ignorelist: + - must scrub before tier agent can activate + conf: + osd: + # override short_pg_log_entries.yaml (which sets these under [global]) + osd_min_pg_log_entries: 3000 + osd_max_pg_log_entries: 3000 +tasks: +- exec: + client.0: + - sudo ceph osd pool create base 4 + - sudo ceph osd pool application enable base rados + - sudo ceph osd pool create cache 4 + - sudo ceph osd tier add base cache + - sudo ceph osd tier cache-mode cache writeback + - sudo ceph osd tier set-overlay base cache + - sudo ceph osd pool set cache hit_set_type bloom + - sudo ceph osd pool set cache hit_set_count 8 + - sudo ceph osd pool set cache hit_set_period 3600 + - sudo ceph osd pool set cache min_read_recency_for_promote 0 + - sudo ceph osd pool set cache min_write_recency_for_promote 0 +- rados: + clients: [client.0] + pools: [base] + ops: 4000 + objects: 500 + op_weights: + read: 100 + write: 100 + delete: 50 + copy_from: 50 + cache_flush: 50 + cache_try_flush: 50 + cache_evict: 50 diff --git a/qa/suites/rados/thrash/workloads/pool-snaps-few-objects.yaml b/qa/suites/rados/thrash/workloads/pool-snaps-few-objects.yaml new file mode 100644 index 000000000..1f0759d96 --- /dev/null +++ b/qa/suites/rados/thrash/workloads/pool-snaps-few-objects.yaml @@ -0,0 +1,18 @@ +override: + conf: + osd: + osd deep scrub update digest min age: 0 +tasks: +- rados: + clients: [client.0] + ops: 4000 + objects: 50 + pool_snaps: true + op_weights: + read: 100 + write: 100 + delete: 50 + snap_create: 50 + snap_remove: 50 + rollback: 50 + copy_from: 50 diff --git a/qa/suites/rados/thrash/workloads/rados_api_tests.yaml b/qa/suites/rados/thrash/workloads/rados_api_tests.yaml new file mode 100644 index 000000000..eb4c9b1ee --- /dev/null +++ b/qa/suites/rados/thrash/workloads/rados_api_tests.yaml @@ -0,0 +1,20 @@ +overrides: + ceph: + log-ignorelist: + - reached quota + - \(POOL_APP_NOT_ENABLED\) + - \(PG_AVAILABILITY\) + crush_tunables: jewel + conf: + client: + debug ms: 1 + debug objecter: 20 + debug rados: 20 + mon: + mon warn on pool no app: false + debug mgrc: 20 +tasks: +- workunit: + clients: + client.0: + - rados/test.sh diff --git a/qa/suites/rados/thrash/workloads/radosbench-high-concurrency.yaml b/qa/suites/rados/thrash/workloads/radosbench-high-concurrency.yaml new file mode 100644 index 000000000..902c4b56a --- /dev/null +++ b/qa/suites/rados/thrash/workloads/radosbench-high-concurrency.yaml @@ -0,0 +1,49 @@ +overrides: + ceph: + conf: + client.0: + debug ms: 1 + debug objecter: 20 + debug rados: 20 +tasks: +- full_sequential: + - radosbench: + clients: [client.0] + concurrency: 128 + size: 8192 + time: 90 + - radosbench: + clients: [client.0] + concurrency: 128 + size: 8192 + time: 90 + - radosbench: + clients: [client.0] + concurrency: 128 + size: 8192 + time: 90 + - radosbench: + clients: [client.0] + concurrency: 128 + size: 8192 + time: 90 + - radosbench: + clients: [client.0] + concurrency: 128 + size: 8192 + time: 90 + - radosbench: + clients: [client.0] + concurrency: 128 + size: 8192 + time: 90 + - radosbench: + clients: [client.0] + concurrency: 128 + size: 8192 + time: 90 + - radosbench: + clients: [client.0] + concurrency: 128 + size: 8192 + time: 90 diff --git a/qa/suites/rados/thrash/workloads/radosbench.yaml b/qa/suites/rados/thrash/workloads/radosbench.yaml new file mode 100644 index 000000000..071f55e39 --- /dev/null +++ b/qa/suites/rados/thrash/workloads/radosbench.yaml @@ -0,0 +1,24 @@ +overrides: + ceph: + conf: + client.0: + debug ms: 1 + debug objecter: 20 + debug rados: 20 +tasks: +- full_sequential: + - radosbench: + clients: [client.0] + time: 90 + - radosbench: + clients: [client.0] + time: 90 + - radosbench: + clients: [client.0] + time: 90 + - radosbench: + clients: [client.0] + time: 90 + - radosbench: + clients: [client.0] + time: 90 diff --git a/qa/suites/rados/thrash/workloads/redirect.yaml b/qa/suites/rados/thrash/workloads/redirect.yaml new file mode 100644 index 000000000..bebce845c --- /dev/null +++ b/qa/suites/rados/thrash/workloads/redirect.yaml @@ -0,0 +1,15 @@ +tasks: +- exec: + client.0: + - sudo ceph osd pool create low_tier 4 +- rados: + clients: [client.0] + low_tier_pool: 'low_tier' + ops: 4000 + objects: 500 + set_redirect: true + op_weights: + read: 100 + write: 100 + delete: 50 + copy_from: 50 diff --git a/qa/suites/rados/thrash/workloads/redirect_promote_tests.yaml b/qa/suites/rados/thrash/workloads/redirect_promote_tests.yaml new file mode 100644 index 000000000..c2787c43b --- /dev/null +++ b/qa/suites/rados/thrash/workloads/redirect_promote_tests.yaml @@ -0,0 +1,14 @@ +tasks: +- exec: + client.0: + - sudo ceph osd pool create low_tier 4 +- rados: + clients: [client.0] + low_tier_pool: 'low_tier' + ops: 4000 + objects: 500 + set_redirect: true + op_weights: + set_redirect: 100 + read: 50 + tier_promote: 30 diff --git a/qa/suites/rados/thrash/workloads/redirect_set_object.yaml b/qa/suites/rados/thrash/workloads/redirect_set_object.yaml new file mode 100644 index 000000000..06ba60c7c --- /dev/null +++ b/qa/suites/rados/thrash/workloads/redirect_set_object.yaml @@ -0,0 +1,13 @@ +tasks: +- exec: + client.0: + - sudo ceph osd pool create low_tier 4 +- rados: + clients: [client.0] + low_tier_pool: 'low_tier' + ops: 4000 + objects: 500 + set_redirect: true + op_weights: + set_redirect: 100 + copy_from: 100 diff --git a/qa/suites/rados/thrash/workloads/set-chunks-read.yaml b/qa/suites/rados/thrash/workloads/set-chunks-read.yaml new file mode 100644 index 000000000..1abbdd757 --- /dev/null +++ b/qa/suites/rados/thrash/workloads/set-chunks-read.yaml @@ -0,0 +1,13 @@ +tasks: +- exec: + client.0: + - sudo ceph osd pool create low_tier 4 +- rados: + clients: [client.0] + low_tier_pool: 'low_tier' + ops: 4000 + objects: 300 + set_chunk: true + op_weights: + chunk_read: 100 + tier_promote: 10 diff --git a/qa/suites/rados/thrash/workloads/small-objects-balanced.yaml b/qa/suites/rados/thrash/workloads/small-objects-balanced.yaml new file mode 100644 index 000000000..56aaaf543 --- /dev/null +++ b/qa/suites/rados/thrash/workloads/small-objects-balanced.yaml @@ -0,0 +1,22 @@ +overrides: + ceph: + crush_tunables: jewel +tasks: +- rados: + clients: [client.0] + ops: 400000 + max_seconds: 600 + max_in_flight: 64 + objects: 1024 + size: 16384 + balance_reads: true + op_weights: + read: 100 + write: 100 + delete: 50 + snap_create: 50 + snap_remove: 50 + rollback: 50 + copy_from: 50 + setattr: 25 + rmattr: 25 diff --git a/qa/suites/rados/thrash/workloads/small-objects-localized.yaml b/qa/suites/rados/thrash/workloads/small-objects-localized.yaml new file mode 100644 index 000000000..e06c717f4 --- /dev/null +++ b/qa/suites/rados/thrash/workloads/small-objects-localized.yaml @@ -0,0 +1,22 @@ +overrides: + ceph: + crush_tunables: jewel +tasks: +- rados: + clients: [client.0] + ops: 400000 + max_seconds: 600 + max_in_flight: 64 + objects: 1024 + size: 16384 + localize_reads: true + op_weights: + read: 100 + write: 100 + delete: 50 + snap_create: 50 + snap_remove: 50 + rollback: 50 + copy_from: 50 + setattr: 25 + rmattr: 25 diff --git a/qa/suites/rados/thrash/workloads/small-objects.yaml b/qa/suites/rados/thrash/workloads/small-objects.yaml new file mode 100644 index 000000000..ee20bc3c5 --- /dev/null +++ b/qa/suites/rados/thrash/workloads/small-objects.yaml @@ -0,0 +1,21 @@ +overrides: + ceph: + crush_tunables: jewel +tasks: +- rados: + clients: [client.0] + ops: 400000 + max_seconds: 600 + max_in_flight: 64 + objects: 1024 + size: 16384 + op_weights: + read: 100 + write: 100 + delete: 50 + snap_create: 50 + snap_remove: 50 + rollback: 50 + copy_from: 50 + setattr: 25 + rmattr: 25 diff --git a/qa/suites/rados/thrash/workloads/snaps-few-objects-balanced.yaml b/qa/suites/rados/thrash/workloads/snaps-few-objects-balanced.yaml new file mode 100644 index 000000000..5752633c8 --- /dev/null +++ b/qa/suites/rados/thrash/workloads/snaps-few-objects-balanced.yaml @@ -0,0 +1,14 @@ +tasks: +- rados: + clients: [client.0] + ops: 4000 + objects: 50 + balance_reads: true + op_weights: + read: 100 + write: 100 + delete: 50 + snap_create: 50 + snap_remove: 50 + rollback: 50 + copy_from: 50 diff --git a/qa/suites/rados/thrash/workloads/snaps-few-objects-localized.yaml b/qa/suites/rados/thrash/workloads/snaps-few-objects-localized.yaml new file mode 100644 index 000000000..d02a5784e --- /dev/null +++ b/qa/suites/rados/thrash/workloads/snaps-few-objects-localized.yaml @@ -0,0 +1,14 @@ +tasks: +- rados: + clients: [client.0] + ops: 4000 + objects: 50 + localize_reads: true + op_weights: + read: 100 + write: 100 + delete: 50 + snap_create: 50 + snap_remove: 50 + rollback: 50 + copy_from: 50 diff --git a/qa/suites/rados/thrash/workloads/snaps-few-objects.yaml b/qa/suites/rados/thrash/workloads/snaps-few-objects.yaml new file mode 100644 index 000000000..aa82d973a --- /dev/null +++ b/qa/suites/rados/thrash/workloads/snaps-few-objects.yaml @@ -0,0 +1,13 @@ +tasks: +- rados: + clients: [client.0] + ops: 4000 + objects: 50 + op_weights: + read: 100 + write: 100 + delete: 50 + snap_create: 50 + snap_remove: 50 + rollback: 50 + copy_from: 50 diff --git a/qa/suites/rados/thrash/workloads/write_fadvise_dontneed.yaml b/qa/suites/rados/thrash/workloads/write_fadvise_dontneed.yaml new file mode 100644 index 000000000..606dcae69 --- /dev/null +++ b/qa/suites/rados/thrash/workloads/write_fadvise_dontneed.yaml @@ -0,0 +1,8 @@ +tasks: +- rados: + clients: [client.0] + ops: 4000 + objects: 500 + write_fadvise_dontneed: true + op_weights: + write: 100 diff --git a/qa/suites/rados/upgrade/.qa b/qa/suites/rados/upgrade/.qa new file mode 120000 index 000000000..a602a0353 --- /dev/null +++ b/qa/suites/rados/upgrade/.qa @@ -0,0 +1 @@ +../.qa/ \ No newline at end of file diff --git a/qa/suites/rados/upgrade/nautilus-x-singleton b/qa/suites/rados/upgrade/nautilus-x-singleton new file mode 120000 index 000000000..d0e26e498 --- /dev/null +++ b/qa/suites/rados/upgrade/nautilus-x-singleton @@ -0,0 +1 @@ +../../upgrade/nautilus-x-singleton \ No newline at end of file diff --git a/qa/suites/rados/valgrind-leaks/% b/qa/suites/rados/valgrind-leaks/% new file mode 100644 index 000000000..e69de29bb diff --git a/qa/suites/rados/valgrind-leaks/.qa b/qa/suites/rados/valgrind-leaks/.qa new file mode 120000 index 000000000..fea2489fd --- /dev/null +++ b/qa/suites/rados/valgrind-leaks/.qa @@ -0,0 +1 @@ +../.qa \ No newline at end of file diff --git a/qa/suites/rados/valgrind-leaks/1-start.yaml b/qa/suites/rados/valgrind-leaks/1-start.yaml new file mode 100644 index 000000000..9263f2a83 --- /dev/null +++ b/qa/suites/rados/valgrind-leaks/1-start.yaml @@ -0,0 +1,30 @@ +openstack: + - volumes: # attached to each instance + count: 2 + size: 10 # GB + +overrides: + install: + ceph: + debuginfo: true + ceph: + log-ignorelist: + - overall HEALTH_ + - \(PG_ + conf: + global: + osd heartbeat grace: 40 + osd max object name len: 460 + osd max object namespace len: 64 + mon: + mon osd crush smoke test: false + osd: + osd fast shutdown: false + valgrind: + mon: [--tool=memcheck, --leak-check=full, --show-reachable=yes] + osd: [--tool=memcheck] +roles: +- [mon.a, mon.b, mon.c, mgr.x, mgr.y, osd.0, osd.1, osd.2, client.0] +tasks: +- install: +- ceph: diff --git a/qa/suites/rados/valgrind-leaks/2-inject-leak/mon.yaml b/qa/suites/rados/valgrind-leaks/2-inject-leak/mon.yaml new file mode 100644 index 000000000..695a9f8af --- /dev/null +++ b/qa/suites/rados/valgrind-leaks/2-inject-leak/mon.yaml @@ -0,0 +1,7 @@ +overrides: + ceph: + expect_valgrind_errors: true +tasks: +- exec: + mon.a: + - ceph tell mon.a leak_some_memory diff --git a/qa/suites/rados/valgrind-leaks/2-inject-leak/none.yaml b/qa/suites/rados/valgrind-leaks/2-inject-leak/none.yaml new file mode 100644 index 000000000..e69de29bb diff --git a/qa/suites/rados/valgrind-leaks/2-inject-leak/osd.yaml b/qa/suites/rados/valgrind-leaks/2-inject-leak/osd.yaml new file mode 100644 index 000000000..f249f16e1 --- /dev/null +++ b/qa/suites/rados/valgrind-leaks/2-inject-leak/osd.yaml @@ -0,0 +1,7 @@ +overrides: + ceph: + expect_valgrind_errors: true +tasks: +- exec: + mon.a: + - ceph tell osd.0 leak_some_memory diff --git a/qa/suites/rados/valgrind-leaks/centos_latest.yaml b/qa/suites/rados/valgrind-leaks/centos_latest.yaml new file mode 120000 index 000000000..bd9854e70 --- /dev/null +++ b/qa/suites/rados/valgrind-leaks/centos_latest.yaml @@ -0,0 +1 @@ +.qa/distros/supported/centos_latest.yaml \ No newline at end of file diff --git a/qa/suites/rados/verify/% b/qa/suites/rados/verify/% new file mode 100644 index 000000000..e69de29bb diff --git a/qa/suites/rados/verify/.qa b/qa/suites/rados/verify/.qa new file mode 120000 index 000000000..a602a0353 --- /dev/null +++ b/qa/suites/rados/verify/.qa @@ -0,0 +1 @@ +../.qa/ \ No newline at end of file diff --git a/qa/suites/rados/verify/centos_latest.yaml b/qa/suites/rados/verify/centos_latest.yaml new file mode 120000 index 000000000..bd9854e70 --- /dev/null +++ b/qa/suites/rados/verify/centos_latest.yaml @@ -0,0 +1 @@ +.qa/distros/supported/centos_latest.yaml \ No newline at end of file diff --git a/qa/suites/rados/verify/ceph.yaml b/qa/suites/rados/verify/ceph.yaml new file mode 100644 index 000000000..fc5ce350a --- /dev/null +++ b/qa/suites/rados/verify/ceph.yaml @@ -0,0 +1,15 @@ +overrides: + ceph: + conf: + mon: + mon min osdmap epochs: 50 + paxos service trim min: 10 + # prune full osdmaps regularly + mon osdmap full prune min: 15 + mon osdmap full prune interval: 2 + mon osdmap full prune txsize: 2 + osd: + debug monc: 20 +tasks: +- install: +- ceph: diff --git a/qa/suites/rados/verify/clusters/+ b/qa/suites/rados/verify/clusters/+ new file mode 100644 index 000000000..e69de29bb diff --git a/qa/suites/rados/verify/clusters/.qa b/qa/suites/rados/verify/clusters/.qa new file mode 120000 index 000000000..a602a0353 --- /dev/null +++ b/qa/suites/rados/verify/clusters/.qa @@ -0,0 +1 @@ +../.qa/ \ No newline at end of file diff --git a/qa/suites/rados/verify/clusters/fixed-2.yaml b/qa/suites/rados/verify/clusters/fixed-2.yaml new file mode 120000 index 000000000..230ff0fda --- /dev/null +++ b/qa/suites/rados/verify/clusters/fixed-2.yaml @@ -0,0 +1 @@ +.qa/clusters/fixed-2.yaml \ No newline at end of file diff --git a/qa/suites/rados/verify/clusters/openstack.yaml b/qa/suites/rados/verify/clusters/openstack.yaml new file mode 100644 index 000000000..e559d9126 --- /dev/null +++ b/qa/suites/rados/verify/clusters/openstack.yaml @@ -0,0 +1,4 @@ +openstack: + - volumes: # attached to each instance + count: 4 + size: 10 # GB diff --git a/qa/suites/rados/verify/d-thrash/.qa b/qa/suites/rados/verify/d-thrash/.qa new file mode 120000 index 000000000..a602a0353 --- /dev/null +++ b/qa/suites/rados/verify/d-thrash/.qa @@ -0,0 +1 @@ +../.qa/ \ No newline at end of file diff --git a/qa/suites/rados/verify/d-thrash/default/+ b/qa/suites/rados/verify/d-thrash/default/+ new file mode 100644 index 000000000..e69de29bb diff --git a/qa/suites/rados/verify/d-thrash/default/.qa b/qa/suites/rados/verify/d-thrash/default/.qa new file mode 120000 index 000000000..a602a0353 --- /dev/null +++ b/qa/suites/rados/verify/d-thrash/default/.qa @@ -0,0 +1 @@ +../.qa/ \ No newline at end of file diff --git a/qa/suites/rados/verify/d-thrash/default/default.yaml b/qa/suites/rados/verify/d-thrash/default/default.yaml new file mode 100644 index 000000000..41b35926f --- /dev/null +++ b/qa/suites/rados/verify/d-thrash/default/default.yaml @@ -0,0 +1,11 @@ +overrides: + ceph: + log-ignorelist: + - but it is still running + - objects unfound and apparently lost +tasks: +- thrashosds: + timeout: 1200 + chance_pgnum_grow: 1 + chance_pgnum_shrink: 1 + chance_pgpnum_fix: 1 diff --git a/qa/suites/rados/verify/d-thrash/default/thrashosds-health.yaml b/qa/suites/rados/verify/d-thrash/default/thrashosds-health.yaml new file mode 120000 index 000000000..9124eb1aa --- /dev/null +++ b/qa/suites/rados/verify/d-thrash/default/thrashosds-health.yaml @@ -0,0 +1 @@ +.qa/tasks/thrashosds-health.yaml \ No newline at end of file diff --git a/qa/suites/rados/verify/d-thrash/none.yaml b/qa/suites/rados/verify/d-thrash/none.yaml new file mode 100644 index 000000000..e69de29bb diff --git a/qa/suites/rados/verify/mon_election b/qa/suites/rados/verify/mon_election new file mode 120000 index 000000000..3f331e621 --- /dev/null +++ b/qa/suites/rados/verify/mon_election @@ -0,0 +1 @@ +.qa/mon_election \ No newline at end of file diff --git a/qa/suites/rados/verify/msgr b/qa/suites/rados/verify/msgr new file mode 120000 index 000000000..57bee80db --- /dev/null +++ b/qa/suites/rados/verify/msgr @@ -0,0 +1 @@ +.qa/msgr \ No newline at end of file diff --git a/qa/suites/rados/verify/msgr-failures/.qa b/qa/suites/rados/verify/msgr-failures/.qa new file mode 120000 index 000000000..a602a0353 --- /dev/null +++ b/qa/suites/rados/verify/msgr-failures/.qa @@ -0,0 +1 @@ +../.qa/ \ No newline at end of file diff --git a/qa/suites/rados/verify/msgr-failures/few.yaml b/qa/suites/rados/verify/msgr-failures/few.yaml new file mode 100644 index 000000000..519288992 --- /dev/null +++ b/qa/suites/rados/verify/msgr-failures/few.yaml @@ -0,0 +1,8 @@ +overrides: + ceph: + conf: + global: + ms inject socket failures: 5000 + mon client directed command retry: 5 + log-ignorelist: + - \(OSD_SLOW_PING_TIME diff --git a/qa/suites/rados/verify/objectstore b/qa/suites/rados/verify/objectstore new file mode 120000 index 000000000..848c65f9e --- /dev/null +++ b/qa/suites/rados/verify/objectstore @@ -0,0 +1 @@ +.qa/objectstore_debug \ No newline at end of file diff --git a/qa/suites/rados/verify/rados.yaml b/qa/suites/rados/verify/rados.yaml new file mode 120000 index 000000000..d256979c0 --- /dev/null +++ b/qa/suites/rados/verify/rados.yaml @@ -0,0 +1 @@ +.qa/config/rados.yaml \ No newline at end of file diff --git a/qa/suites/rados/verify/tasks/.qa b/qa/suites/rados/verify/tasks/.qa new file mode 120000 index 000000000..a602a0353 --- /dev/null +++ b/qa/suites/rados/verify/tasks/.qa @@ -0,0 +1 @@ +../.qa/ \ No newline at end of file diff --git a/qa/suites/rados/verify/tasks/mon_recovery.yaml b/qa/suites/rados/verify/tasks/mon_recovery.yaml new file mode 100644 index 000000000..06d9602e6 --- /dev/null +++ b/qa/suites/rados/verify/tasks/mon_recovery.yaml @@ -0,0 +1,12 @@ +overrides: + ceph: + log-ignorelist: + - overall HEALTH_ + - \(MON_DOWN\) + - \(OSDMAP_FLAGS\) + - \(SMALLER_PGP_NUM\) + - \(POOL_APP_NOT_ENABLED\) + - \(SLOW OPS\) + - slow request +tasks: +- mon_recovery: diff --git a/qa/suites/rados/verify/tasks/rados_api_tests.yaml b/qa/suites/rados/verify/tasks/rados_api_tests.yaml new file mode 100644 index 000000000..f89109aa1 --- /dev/null +++ b/qa/suites/rados/verify/tasks/rados_api_tests.yaml @@ -0,0 +1,30 @@ +overrides: + ceph: + log-ignorelist: + - reached quota + - overall HEALTH_ + - \(CACHE_POOL_NO_HIT_SET\) + - \(POOL_FULL\) + - \(SMALLER_PGP_NUM\) + - \(SLOW_OPS\) + - \(CACHE_POOL_NEAR_FULL\) + - \(POOL_APP_NOT_ENABLED\) + - \(PG_AVAILABILITY\) + - \(OBJECT_MISPLACED\) + - slow request + conf: + client: + debug ms: 1 + debug objecter: 20 + debug rados: 20 + debug monc: 20 + mon: + mon warn on pool no app: false +tasks: +- workunit: + timeout: 6h + env: + ALLOW_TIMEOUTS: "1" + clients: + client.0: + - rados/test.sh diff --git a/qa/suites/rados/verify/tasks/rados_cls_all.yaml b/qa/suites/rados/verify/tasks/rados_cls_all.yaml new file mode 100644 index 000000000..163bc2c08 --- /dev/null +++ b/qa/suites/rados/verify/tasks/rados_cls_all.yaml @@ -0,0 +1,11 @@ +overrides: + ceph: + conf: + osd: + osd_class_load_list: "*" + osd_class_default_list: "*" +tasks: +- workunit: + clients: + client.0: + - cls diff --git a/qa/suites/rados/verify/validater/.qa b/qa/suites/rados/verify/validater/.qa new file mode 120000 index 000000000..a602a0353 --- /dev/null +++ b/qa/suites/rados/verify/validater/.qa @@ -0,0 +1 @@ +../.qa/ \ No newline at end of file diff --git a/qa/suites/rados/verify/validater/lockdep.yaml b/qa/suites/rados/verify/validater/lockdep.yaml new file mode 100644 index 000000000..25f84355c --- /dev/null +++ b/qa/suites/rados/verify/validater/lockdep.yaml @@ -0,0 +1,5 @@ +overrides: + ceph: + conf: + global: + lockdep: true diff --git a/qa/suites/rados/verify/validater/valgrind.yaml b/qa/suites/rados/verify/validater/valgrind.yaml new file mode 100644 index 000000000..03accceaf --- /dev/null +++ b/qa/suites/rados/verify/validater/valgrind.yaml @@ -0,0 +1,31 @@ +# see http://tracker.ceph.com/issues/20360 and http://tracker.ceph.com/issues/18126 +os_type: centos + +overrides: + install: + ceph: + debuginfo: true + ceph: + conf: + global: + osd heartbeat grace: 80 + mon: + mon osd crush smoke test: false + osd: + osd fast shutdown: false + debug bluestore: 1 + debug bluefs: 1 + log-ignorelist: + - overall HEALTH_ +# valgrind is slow.. we might get PGs stuck peering etc + - \(PG_ +# mons sometimes are left off of initial quorum due to valgrind slowness. ok to ignore here because we'll still catch an actual crash due to the core + - \(MON_DOWN\) + - \(SLOW_OPS\) + - slow request + valgrind: + mon: [--tool=memcheck, --leak-check=full, --show-reachable=yes] + osd: [--tool=memcheck] + mds: [--tool=memcheck] +# https://tracker.ceph.com/issues/38621 +# mgr: [--tool=memcheck] -- cgit v1.2.3