blob: 362e0373416ea1478aaed9aa8c8fb1b47e273393 (
plain)
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
|
tasks:
- parallel:
- upgrade-tasks
- workload-tasks
upgrade-tasks:
sequential:
- cephadm.shell:
env: [sha1]
host.a:
- ceph config set mon mon_warn_on_insecure_global_id_reclaim false --force
- ceph config set mon mon_warn_on_insecure_global_id_reclaim_allowed false --force
- ceph config set global log_to_journald false --force
- ceph mgr module enable nfs --force
- ceph orch upgrade start --image quay.ceph.io/ceph-ci/ceph:$sha1
- cephadm.shell:
env: [sha1]
host.a:
- while ceph orch upgrade status | jq '.in_progress' | grep true && ! ceph orch upgrade status | jq '.message' | grep Error ; do ceph orch ps ; ceph versions ; ceph orch upgrade status ; sleep 30 ; done
- ceph orch ps
- ceph versions
- echo "wait for servicemap items w/ changing names to refresh"
- sleep 60
- ceph orch ps
- ceph versions
- ceph versions | jq -e '.overall | length == 1'
- ceph versions | jq -e '.overall | keys' | grep $sha1
# this should be a no-op, but confirms nfs.ganesha-foo was remapped to nfs.foo
- cephadm.wait_for_service:
service: nfs.foo
workload-tasks:
sequential:
- exec:
host.a:
- cd /mnt/foo && dbench 5 -t 600 || true # might fail with ESTALE
# make sure mount works
- umount /mnt/foo
- while ! mount -t nfs $(hostname):/fake /mnt/foo ; do sleep 5 ; done
- cd /mnt/foo && dbench 5 -t 5
|