11 - volumes: # attached to each instance
18 - sudo ceph config set mgr mgr/devicehealth/enable_monitoring false --force
21 osd recovery sleep: .1
22 osd min pg log entries: 10
23 osd max pg log entries: 1000
24 osd_target_pg_log_entries_per_osd: 0
25 osd pg log trim min: 10
27 - \(POOL_APP_NOT_ENABLED\)
37 - ceph osd pool create foo 128
38 - ceph osd pool application enable foo foo
43 - rados -p foo bench 30 write -b 4096 --no-cleanup
50 wait-for-healthy: false
53 - rados -p foo bench 3 write -b 4096 --no-cleanup
56 - for f in 0 1 2 3 ; do sudo ceph daemon osd.$f config set osd_recovery_sleep 0 ; sudo ceph daemon osd.$f config set osd_recovery_max_active 20 ; done
60 - egrep '(defer backfill|defer recovery)' /var/log/ceph/ceph-osd.*.log