From 19fcec84d8d7d21e796c7624e521b60d28ee21ed Mon Sep 17 00:00:00 2001 From: Daniel Baumann Date: Sun, 7 Apr 2024 20:45:59 +0200 Subject: Adding upstream version 16.2.11+ds. Signed-off-by: Daniel Baumann --- qa/suites/rbd/librbd/pool/.qa | 1 + qa/suites/rbd/librbd/pool/ec-data-pool.yaml | 24 ++++++++++++++++++++++ qa/suites/rbd/librbd/pool/none.yaml | 0 .../rbd/librbd/pool/replicated-data-pool.yaml | 11 ++++++++++ qa/suites/rbd/librbd/pool/small-cache-pool.yaml | 17 +++++++++++++++ 5 files changed, 53 insertions(+) create mode 120000 qa/suites/rbd/librbd/pool/.qa create mode 100644 qa/suites/rbd/librbd/pool/ec-data-pool.yaml create mode 100644 qa/suites/rbd/librbd/pool/none.yaml create mode 100644 qa/suites/rbd/librbd/pool/replicated-data-pool.yaml create mode 100644 qa/suites/rbd/librbd/pool/small-cache-pool.yaml (limited to 'qa/suites/rbd/librbd/pool') diff --git a/qa/suites/rbd/librbd/pool/.qa b/qa/suites/rbd/librbd/pool/.qa new file mode 120000 index 000000000..a602a0353 --- /dev/null +++ b/qa/suites/rbd/librbd/pool/.qa @@ -0,0 +1 @@ +../.qa/ \ No newline at end of file diff --git a/qa/suites/rbd/librbd/pool/ec-data-pool.yaml b/qa/suites/rbd/librbd/pool/ec-data-pool.yaml new file mode 100644 index 000000000..f39a5bb4c --- /dev/null +++ b/qa/suites/rbd/librbd/pool/ec-data-pool.yaml @@ -0,0 +1,24 @@ +tasks: +- exec: + client.0: + - sudo ceph osd erasure-code-profile set teuthologyprofile crush-failure-domain=osd m=1 k=2 + - sudo ceph osd pool create datapool 4 4 erasure teuthologyprofile + - sudo ceph osd pool set datapool allow_ec_overwrites true + - rbd pool init datapool + +overrides: + thrashosds: + bdev_inject_crash: 2 + bdev_inject_crash_probability: .5 + ceph: + fs: xfs + conf: + client: + rbd default data pool: datapool + osd: # force bluestore since it's required for ec overwrites + osd objectstore: bluestore + bluestore block size: 96636764160 + enable experimental unrecoverable data corrupting features: "*" + osd debug randomize hobject sort order: false +# this doesn't work with failures bc the log writes are not atomic across the two backends +# bluestore bluefs env mirror: true diff --git a/qa/suites/rbd/librbd/pool/none.yaml b/qa/suites/rbd/librbd/pool/none.yaml new file mode 100644 index 000000000..e69de29bb diff --git a/qa/suites/rbd/librbd/pool/replicated-data-pool.yaml b/qa/suites/rbd/librbd/pool/replicated-data-pool.yaml new file mode 100644 index 000000000..c5647dba1 --- /dev/null +++ b/qa/suites/rbd/librbd/pool/replicated-data-pool.yaml @@ -0,0 +1,11 @@ +tasks: +- exec: + client.0: + - sudo ceph osd pool create datapool 4 + - rbd pool init datapool + +overrides: + ceph: + conf: + client: + rbd default data pool: datapool diff --git a/qa/suites/rbd/librbd/pool/small-cache-pool.yaml b/qa/suites/rbd/librbd/pool/small-cache-pool.yaml new file mode 100644 index 000000000..bad95eadd --- /dev/null +++ b/qa/suites/rbd/librbd/pool/small-cache-pool.yaml @@ -0,0 +1,17 @@ +overrides: + ceph: + log-ignorelist: + - overall HEALTH_ + - \(CACHE_POOL_NEAR_FULL\) + - \(CACHE_POOL_NO_HIT_SET\) +tasks: +- exec: + client.0: + - sudo ceph osd pool create cache 4 + - sudo ceph osd tier add rbd cache + - sudo ceph osd tier cache-mode cache writeback + - sudo ceph osd tier set-overlay rbd cache + - sudo ceph osd pool set cache hit_set_type bloom + - sudo ceph osd pool set cache hit_set_count 8 + - sudo ceph osd pool set cache hit_set_period 60 + - sudo ceph osd pool set cache target_max_objects 250 -- cgit v1.2.3