3 Run ceph on two nodes, using one of them as a client,
4 with a separate client-only node.
5 Use xfs beneath the osds.
6 install ceph/jewel v10.2.0 point version
7 run workload and upgrade-sequence in parallel
8 install ceph/jewel latest version
9 run workload and upgrade-sequence in parallel
10 install ceph/-x version (jewel or kraken)
11 run workload and upgrade-sequence in parallel
22 mon debug unsafe allow tier with nonempty snaps: true
24 osd map max advance: 1000
40 - volumes: # attached to each instance
44 - print: "**** v10.2.0 about to install"
47 exclude_packages: ['ceph-mgr','libcephfs2','libcephfs-devel','libcephfs-dev', 'librgw2']
48 - print: "**** done v10.2.0 install"
51 skip_mgr_daemons: true
52 - print: "**** done ceph xfs"
55 - print: "**** done workload v10.2.0"
57 exclude_packages: ['ceph-mgr','libcephfs2','libcephfs-devel','libcephfs-dev']
62 # Note that client.a IS NOT upgraded at this point
67 - upgrade-sequence_jewel
68 - print: "**** done parallel jewel branch"
70 exclude_packages: ['ceph-mgr','libcephfs2','libcephfs-devel','libcephfs-dev']
73 - print: "**** done branch: jewel install.upgrade on client.1"
77 - print: "**** done branch: -x install.upgrade on mon.a and mon.b"
81 - print: "**** done parallel -x branch"
82 # Run librados tests on the -x upgraded cluster
89 - rados/test-upgrade-v11.0.0.sh
91 - print: "**** done final test on -x cluster"
92 #######################
108 CLS_RBD_GTEST_FILTER: '*:-TestClsRbd.mirror_image'
109 - print: "**** done rados/test.sh & cls workload_jewel"
112 - print: "**** done rgw workload_jewel"
115 force-branch: ceph-jewel
117 scan_for_encryption_keys: false
118 - print: "**** done s3tests workload_jewel"
119 upgrade-sequence_jewel:
121 - print: "**** done branch: jewel install.upgrade"
122 - ceph.restart: [mds.a]
125 - ceph.restart: [osd.0]
128 - ceph.restart: [osd.1]
131 - ceph.restart: [osd.2]
134 - ceph.restart: [osd.3]
137 - ceph.restart: [osd.4]
140 - ceph.restart: [osd.5]
143 - ceph.restart: [mon.a]
146 - ceph.restart: [mon.b]
149 - ceph.restart: [mon.c]
152 - print: "**** done ceph.restart all jewel branch mds/osd/mon"
159 - rados/test-upgrade-v11.0.0.sh
162 CLS_RBD_GTEST_FILTER: '*:-TestClsRbd.mirror_image'
163 - print: "**** done rados/test-upgrade-v11.0.0.sh & cls workload_x NOT upgraded client"
168 - rados/test-upgrade-v11.0.0.sh
170 - print: "**** done rados/test-upgrade-v11.0.0.sh & cls workload_x upgraded client"
172 - print: "**** done rgw workload_x"
175 force-branch: ceph-jewel
177 scan_for_encryption_keys: false
178 - print: "**** done s3tests workload_x"
181 - ceph.restart: [mds.a]
184 - ceph.restart: [mon.a]
187 - ceph.restart: [mon.b]
190 - ceph.restart: [mon.c]
193 - ceph.restart: [osd.0]
196 - ceph.restart: [osd.1]
199 - ceph.restart: [osd.2]
202 - ceph.restart: [osd.3]
205 - ceph.restart: [osd.4]
210 wait-for-healthy: false
211 wait-for-up-osds: true
214 - mkdir -p /var/lib/ceph/mgr/ceph-x
215 - ceph auth get-or-create-key mgr.x mon 'allow profile mgr'
216 - ceph auth export mgr.x > /var/lib/ceph/mgr/ceph-x/keyring
219 wait-for-healthy: false
222 - ceph osd require-osd-release luminous
223 - ceph osd set-require-min-compat-client luminous
225 - print: "**** done ceph.restart all -x branch mds/osd/mon"