3 # Copyright (C) 2017 Red Hat <contact@redhat.com>
6 # Author: Kefu Chai <kchai@redhat.com>
7 # Author: David Zafman <dzafman@redhat.com>
9 # This program is free software; you can redistribute it and/or modify
10 # it under the terms of the GNU Library Public License as published by
11 # the Free Software Foundation; either version 2, or (at your option)
14 # This program is distributed in the hope that it will be useful,
15 # but WITHOUT ANY WARRANTY; without even the implied warranty of
16 # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
17 # GNU Library Public License for more details.
20 source $CEPH_ROOT/qa
/standalone
/ceph-helpers.sh
28 export CEPH_MON
="127.0.0.1:7140" # git grep '\<7140\>' : there must be only one
30 CEPH_ARGS
+="--fsid=$(uuidgen) --auth-supported=none "
31 CEPH_ARGS
+="--mon-host=$CEPH_MON "
32 CEPH_ARGS
+="--osd-mclock-profile=high_recovery_ops "
35 local funcs
=${@:-$(set | sed -n -e 's/^\(TEST_[0-9a-z_]*\) .*/\1/p')}
36 for func
in $funcs ; do
37 setup
$dir ||
return 1
38 # set warning amount in case default changes
39 run_mon
$dir a
--mon_osd_warn_num_repaired=$warnings ||
return 1
40 run_mgr
$dir x ||
return 1
41 ceph osd pool create foo
8 ||
return 1
43 $func $dir ||
return 1
44 teardown
$dir ||
return 1
48 function setup_osds
() {
53 for id
in $
(seq 0 $
(expr $count - 1)) ; do
54 run_osd
${type} $dir $id ||
return 1
56 wait_for_clean ||
return 1
59 function get_state
() {
62 ceph
--format json pg dump pgs
2>/dev
/null | \
63 jq
-r ".pg_stats | .[] | select(.pgid==\"$pgid\") | .$sname"
66 function rados_put
() {
69 local objname
=${3:-SOMETHING}
71 for marker
in AAA BBB CCCC DDDD
; do
72 printf "%*s" 1024 $marker
75 # get and put an object, compare they are equal
77 rados
--pool $poolname put
$objname $dir/ORIGINAL ||
return 1
80 function rados_get
() {
83 local objname
=${3:-SOMETHING}
87 # Expect a failure to get object
89 if [ $expect = "fail" ];
91 ! rados
--pool $poolname get
$objname $dir/COPY
95 # Expect hang trying to get object
97 if [ $expect = "hang" ];
99 timeout
5 rados
--pool $poolname get
$objname $dir/COPY
104 # get an object, compare with $dir/ORIGINAL
106 rados
--pool $poolname get
$objname $dir/COPY ||
return 1
107 diff $dir/ORIGINAL
$dir/COPY ||
return 1
111 function rados_get_data
() {
116 local poolname
=pool-rep
117 local objname
=obj-
$inject-$$
118 local pgid
=$
(get_pg
$poolname $objname)
120 rados_put
$dir $poolname $objname ||
return 1
121 inject_
$inject rep data
$poolname $objname $dir 0 ||
return 1
122 rados_get
$dir $poolname $objname ||
return 1
125 COUNT
=$
(ceph pg
$pgid query | jq
'.info.stats.stat_sum.num_objects_repaired')
126 test "$COUNT" = "1" ||
return 1
128 COUNT
=$
(ceph pg dump
--format=json-pretty | jq
".pg_map.osd_stats_sum.num_shards_repaired")
129 test "$COUNT" = "1" ||
return 1
131 local object_osds
=($
(get_osds
$poolname $objname))
132 local primary
=${object_osds[0]}
133 local bad_peer
=${object_osds[1]}
134 inject_
$inject rep data
$poolname $objname $dir 0 ||
return 1
135 inject_
$inject rep data
$poolname $objname $dir 1 ||
return 1
136 # Force primary to pull from the bad peer, so we can repair it too!
137 set_config osd
$primary osd_debug_feed_pullee
$bad_peer ||
return 1
138 rados_get
$dir $poolname $objname ||
return 1
140 # Wait until automatic repair of bad peer is done
141 wait_for_clean ||
return 1
143 inject_
$inject rep data
$poolname $objname $dir 0 ||
return 1
144 inject_
$inject rep data
$poolname $objname $dir 2 ||
return 1
145 rados_get
$dir $poolname $objname ||
return 1
148 COUNT
=$
(ceph pg
$pgid query | jq
'.info.stats.stat_sum.num_objects_repaired')
149 test "$COUNT" = "3" ||
return 1
151 COUNT
=$
(ceph pg dump
--format=json-pretty | jq
".pg_map.osd_stats_sum.num_shards_repaired")
152 test "$COUNT" = "4" ||
return 1
154 inject_
$inject rep data
$poolname $objname $dir 0 ||
return 1
155 inject_
$inject rep data
$poolname $objname $dir 1 ||
return 1
156 inject_
$inject rep data
$poolname $objname $dir 2 ||
return 1
157 rados_get
$dir $poolname $objname hang ||
return 1
160 # After hang another repair couldn't happen, so count stays the same
161 COUNT
=$
(ceph pg
$pgid query | jq
'.info.stats.stat_sum.num_objects_repaired')
162 test "$COUNT" = "3" ||
return 1
164 COUNT
=$
(ceph pg dump
--format=json-pretty | jq
".pg_map.osd_stats_sum.num_shards_repaired")
165 test "$COUNT" = "4" ||
return 1
168 function TEST_rados_get_with_eio
() {
171 setup_osds
4 ||
return 1
173 local poolname
=pool-rep
174 create_pool
$poolname 1 1 ||
return 1
175 wait_for_clean ||
return 1
176 rados_get_data eio
$dir ||
return 1
178 delete_pool
$poolname
181 function TEST_rados_repair_warning
() {
183 local OBJS
=$
(expr $warnings + 1)
185 setup_osds
4 ||
return 1
187 local poolname
=pool-rep
188 create_pool
$poolname 1 1 ||
return 1
189 wait_for_clean ||
return 1
191 local poolname
=pool-rep
192 local objbase
=obj-warn
195 for i
in $
(seq 1 $OBJS)
197 rados_put
$dir $poolname ${objbase}-$i ||
return 1
198 inject_
$inject rep data
$poolname ${objbase}-$i $dir 0 ||
return 1
199 rados_get
$dir $poolname ${objbase}-$i ||
return 1
201 local pgid
=$
(get_pg
$poolname ${objbase}-1)
203 local object_osds
=($
(get_osds
$poolname ${objbase}-1))
204 local primary
=${object_osds[0]}
205 local bad_peer
=${object_osds[1]}
208 COUNT
=$
(ceph pg
$pgid query | jq
'.info.stats.stat_sum.num_objects_repaired')
209 test "$COUNT" = "$OBJS" ||
return 1
211 COUNT
=$
(ceph pg dump
--format=json-pretty | jq
".pg_map.osd_stats_sum.num_shards_repaired")
212 test "$COUNT" = "$OBJS" ||
return 1
214 ceph health |
grep -q "Too many repaired reads on 1 OSDs" ||
return 1
215 ceph health detail |
grep -q "osd.$primary had $OBJS reads repaired" ||
return 1
217 ceph health mute OSD_TOO_MANY_REPAIRS
220 ceph health | $
(! grep -q "Too many repaired reads on 1 OSDs") ||
return 1
223 for i
in $
(seq 1 $OBJS)
225 inject_
$inject rep data
$poolname ${objbase}-$i $dir 0 ||
return 1
226 inject_
$inject rep data
$poolname ${objbase}-$i $dir 1 ||
return 1
227 # Force primary to pull from the bad peer, so we can repair it too!
228 set_config osd
$primary osd_debug_feed_pullee
$bad_peer ||
return 1
229 rados_get
$dir $poolname ${objbase}-$i ||
return 1
233 COUNT
=$
(ceph pg
$pgid query | jq
'.info.stats.stat_sum.num_objects_repaired')
234 test "$COUNT" = "$(expr $OBJS \* 2)" ||
return 1
236 COUNT
=$
(ceph pg dump
--format=json-pretty | jq
".pg_map.osd_stats_sum.num_shards_repaired")
237 test "$COUNT" = "$(expr $OBJS \* 3)" ||
return 1
239 # Give mon a chance to notice additional OSD and unmute
240 # The default tick time is 5 seconds
246 if ceph health |
grep -q "Too many repaired reads on 2 OSDs"
250 LOOPS
=$
(expr $LOOPS + 1)
251 if test "$LOOPS" = "$CHECKTIME"
253 echo "Too many repaired reads not seen after $CHECKTIME seconds"
257 ceph health detail |
grep -q "osd.$primary had $(expr $OBJS \* 2) reads repaired" ||
return 1
258 ceph health detail |
grep -q "osd.$bad_peer had $OBJS reads repaired" ||
return 1
260 delete_pool
$poolname
263 # Test backfill with unfound object
264 function TEST_rep_backfill_unfound
() {
266 local objname
=myobject
268 # Must be between 1 and $lastobj
272 CEPH_ARGS
+=' --osd_min_pg_log_entries=5 --osd_max_pg_log_entries=10'
273 setup_osds
3 ||
return 1
275 local poolname
=test-pool
276 create_pool
$poolname 1 1 ||
return 1
277 wait_for_clean ||
return 1
281 rados_put
$dir $poolname $objname ||
return 1
283 local -a initial_osds
=($
(get_osds
$poolname $objname))
284 local last_osd
=${initial_osds[-1]}
285 kill_daemons
$dir TERM osd.
${last_osd} 2>&2 < /dev
/null ||
return 1
286 ceph osd down
${last_osd} ||
return 1
287 ceph osd out
${last_osd} ||
return 1
291 dd if=/dev
/urandom of
=${dir}/ORIGINAL bs
=1024 count
=4
292 for i
in $
(seq 1 $lastobj)
294 rados
--pool $poolname put obj
${i} $dir/ORIGINAL ||
return 1
297 inject_eio rep data
$poolname $testobj $dir 0 ||
return 1
298 inject_eio rep data
$poolname $testobj $dir 1 ||
return 1
300 activate_osd
$dir ${last_osd} ||
return 1
301 ceph osd
in ${last_osd} ||
return 1
305 for tmp
in $
(seq 1 360); do
306 state
=$
(get_state
2.0)
307 echo $state |
grep backfill_unfound
308 if [ "$?" = "0" ]; then
316 ceph pg
2.0 list_unfound |
grep -q $testobj ||
return 1
318 # Command should hang because object is unfound
319 timeout
5 rados
-p $poolname get
$testobj $dir/CHECK
320 test $?
= "124" ||
return 1
322 ceph pg
2.0 mark_unfound_lost delete
324 wait_for_clean ||
return 1
326 for i
in $
(seq 1 $lastobj)
328 if [ obj
${i} = "$testobj" ]; then
329 # Doesn't exist anymore
330 ! rados
-p $poolname get
$testobj $dir/CHECK ||
return 1
332 rados
--pool $poolname get obj
${i} $dir/CHECK ||
return 1
333 diff -q $dir/ORIGINAL
$dir/CHECK ||
return 1
337 rm -f ${dir}/ORIGINAL
${dir}/CHECK
339 delete_pool
$poolname
342 # Test recovery with unfound object
343 function TEST_rep_recovery_unfound
() {
345 local objname
=myobject
347 # Must be between 1 and $lastobj
350 setup_osds
3 ||
return 1
352 local poolname
=test-pool
353 create_pool
$poolname 1 1 ||
return 1
354 wait_for_clean ||
return 1
358 rados_put
$dir $poolname $objname ||
return 1
360 local -a initial_osds
=($
(get_osds
$poolname $objname))
361 local last_osd
=${initial_osds[-1]}
362 kill_daemons
$dir TERM osd.
${last_osd} 2>&2 < /dev
/null ||
return 1
363 ceph osd down
${last_osd} ||
return 1
364 ceph osd out
${last_osd} ||
return 1
368 dd if=/dev
/urandom of
=${dir}/ORIGINAL bs
=1024 count
=4
369 for i
in $
(seq 1 $lastobj)
371 rados
--pool $poolname put obj
${i} $dir/ORIGINAL ||
return 1
374 inject_eio rep data
$poolname $testobj $dir 0 ||
return 1
375 inject_eio rep data
$poolname $testobj $dir 1 ||
return 1
377 activate_osd
$dir ${last_osd} ||
return 1
378 ceph osd
in ${last_osd} ||
return 1
382 for tmp
in $
(seq 1 100); do
383 state
=$
(get_state
2.0)
384 echo $state |
grep -v recovering
385 if [ "$?" = "0" ]; then
393 ceph pg
2.0 list_unfound |
grep -q $testobj ||
return 1
395 # Command should hang because object is unfound
396 timeout
5 rados
-p $poolname get
$testobj $dir/CHECK
397 test $?
= "124" ||
return 1
399 ceph pg
2.0 mark_unfound_lost delete
401 wait_for_clean ||
return 1
403 for i
in $
(seq 1 $lastobj)
405 if [ obj
${i} = "$testobj" ]; then
406 # Doesn't exist anymore
407 ! rados
-p $poolname get
$testobj $dir/CHECK ||
return 1
409 rados
--pool $poolname get obj
${i} $dir/CHECK ||
return 1
410 diff -q $dir/ORIGINAL
$dir/CHECK ||
return 1
414 rm -f ${dir}/ORIGINAL
${dir}/CHECK
416 delete_pool
$poolname
419 # This is a filestore only test because it requires data digest in object info
420 function TEST_rep_read_unfound
() {
422 local objname
=myobject
424 setup_osds
3 _filestore ||
return 1
426 ceph osd pool delete foo foo
--yes-i-really-really-mean-it ||
return 1
427 local poolname
=test-pool
428 create_pool
$poolname 1 1 ||
return 1
429 ceph osd pool
set $poolname size
2
430 wait_for_clean ||
return 1
434 dd if=/dev
/urandom bs
=8k count
=1 of
=$dir/ORIGINAL
435 rados
-p $poolname put
$objname $dir/ORIGINAL
437 local primary
=$
(get_primary
$poolname $objname)
438 local other
=$
(get_not_primary
$poolname $objname)
440 dd if=/dev
/urandom bs
=8k count
=1 of
=$dir/CORRUPT
441 objectstore_tool
$dir $primary $objname set-bytes
$dir/CORRUPT ||
return 1
442 objectstore_tool
$dir $other $objname set-bytes
$dir/CORRUPT ||
return 1
444 timeout
30 rados
-p $poolname get
$objname $dir/tmp
&
449 ceph
--format=json pg dump pgs | jq
'.'
451 if ! ceph
--format=json pg dump pgs | jq
'.pg_stats | .[0].state' |
grep -q recovery_unfound
453 echo "Failure to get to recovery_unfound state"
457 objectstore_tool
$dir $other $objname set-bytes
$dir/ORIGINAL ||
return 1
461 if ! cmp $dir/ORIGINAL
$dir/tmp
463 echo "Bad data after primary repair"
468 main osd-rep-recov-eio.sh
"$@"
471 # compile-command: "cd ../../../build ; make -j4 && ../qa/run-standalone.sh osd-rep-recov-eio.sh"