meta: - desc: | Run ceph on two nodes, using one of them as a client, with a separate client-only node. Use xfs beneath the osds. install ceph/nautilus v14.2.2 point version run workload and upgrade-sequence in parallel (every point reslease should be tested) run workload and upgrade-sequence in parallel install ceph/nautilus latest version run workload and upgrade-sequence in parallel overrides: ceph: log-whitelist: - evicting unresponsive client - reached quota - scrub - osd_map_max_advance - wrongly marked - FS_DEGRADED - POOL_APP_NOT_ENABLED - CACHE_POOL_NO_HIT_SET - POOL_FULL - SMALLER_PG - pool\(s\) full - OSD_DOWN - missing hit_sets - CACHE_POOL_NEAR_FULL - PG_AVAILABILITY - PG_DEGRADED - application not enabled - cache pools at or near target size - filesystem is degraded - OBJECT_MISPLACED ### ref: https://tracker.ceph.com/issues/40251 #removed see ^ - failed to encode map fs: xfs conf: global: mon_warn_on_pool_no_app: false mon pg warn min per osd: 0 mon: mon debug unsafe allow tier with nonempty snaps: true osd: osd map max advance: 1000 osd_class_default_list: "*" osd_class_load_list: "*" client: rgw_crypt_require_ssl: false rgw crypt s3 kms encryption keys: testkey-1=YmluCmJvb3N0CmJvb3N0LWJ1aWxkCmNlcGguY29uZgo= testkey-2=aWIKTWFrZWZpbGUKbWFuCm91dApzcmMKVGVzdGluZwo= roles: - - mon.a - mds.a - osd.0 - osd.1 - osd.2 - mgr.x - - mon.b - mon.c - osd.3 - osd.4 - osd.5 - client.0 - - client.1 openstack: - volumes: # attached to each instance count: 3 size: 30 # GB tasks: # v14.2.0 removed per http://tracker.ceph.com/issues/40251 - print: "**** done nautilus v14.2.2 about to install" - install: tag: v14.2.2 # line below can be removed its from jewel test #exclude_packages: ['ceph-mgr','libcephfs2','libcephfs-devel','libcephfs-dev', 'librgw2'] - print: "**** done v14.2.2 install" - ceph: fs: xfs add_osds_to_crush: true - print: "**** done ceph xfs" - sequential: - workload - print: "**** done workload v14.2.2" # v14.2.1 removed per http://tracker.ceph.com/issues/40251 # v14.2.2 ####### upgrade to v14.2.?? PLACEHOLDER #- install.upgrade: # #exclude_packages: ['ceph-mgr','libcephfs2','libcephfs-devel','libcephfs-dev'] # mon.a: # tag: v14.2.?? # mon.b: # tag: v14.2.?? # # Note that client.a IS NOT upgraded at this point #- parallel: # - workload_nautilus # - upgrade-sequence_nautilus #- print: "**** done parallel nautilus v14.2.??" #### upgrade to latest nautilus - install.upgrade: mon.a: mon.b: - parallel: - workload_nautilus - upgrade-sequence_nautilus - print: "**** done parallel nautilus branch" ####################### workload: sequential: - workunit: clients: client.0: - suites/blogbench.sh workload_nautilus: full_sequential: - workunit: branch: nautilus #tag: v14.2.0 clients: client.1: - rados/test.sh - cls env: CLS_RBD_GTEST_FILTER: '*:-TestClsRbd.snapshots_namespaces' - print: "**** done rados/test.sh & cls workload_nautilus" - sequential: - rgw: [client.0] - print: "**** done rgw workload_nautilus" - s3tests: client.0: force-branch: ceph-nautilus rgw_server: client.0 scan_for_encryption_keys: false - print: "**** done s3tests workload_nautilus" - rbd_fsx: clients: [client.0] size: 134217728 - print: "**** done rbd_fsx workload_nautilus" upgrade-sequence_nautilus: sequential: - print: "**** done branch: nautilus install.upgrade" - ceph.restart: [mds.a] - sleep: duration: 60 - ceph.restart: [osd.0] - sleep: duration: 30 - ceph.restart: [osd.1] - sleep: duration: 30 - ceph.restart: [osd.2] - sleep: duration: 30 - ceph.restart: [osd.3] - sleep: duration: 30 - ceph.restart: [osd.4] - sleep: duration: 30 - ceph.restart: [osd.5] - sleep: duration: 60 - ceph.restart: [mgr.x] - sleep: duration: 60 - ceph.restart: [mon.a] - sleep: duration: 60 - ceph.restart: [mon.b] - sleep: duration: 60 - ceph.restart: [mon.c] - sleep: duration: 60 - print: "**** done ceph.restart all nautilus branch mds/osd/mon"