]>
Commit | Line | Data |
---|---|---|
7c673cae | 1 | # verify #13098 fix |
11fdf7f2 TL |
2 | openstack: |
3 | - volumes: # attached to each instance | |
4 | count: 3 | |
5 | size: 10 # GB | |
7c673cae FG |
6 | roles: |
7 | - [mon.a, mgr.x, osd.0, osd.1, osd.2, client.0] | |
8 | overrides: | |
9 | ceph: | |
cd265ab1 | 10 | log-ignorelist: |
7c673cae | 11 | - is full |
224ce89b | 12 | - overall HEALTH_ |
d2e6a577 FG |
13 | - \(POOL_FULL\) |
14 | - \(POOL_NEAR_FULL\) | |
15 | - \(CACHE_POOL_NO_HIT_SET\) | |
16 | - \(CACHE_POOL_NEAR_FULL\) | |
7c673cae FG |
17 | tasks: |
18 | - install: | |
19 | - ceph: | |
9f95a23c TL |
20 | pre-mgr-commands: |
21 | - sudo ceph config set mgr mgr/devicehealth/enable_monitoring false --force | |
7c673cae FG |
22 | conf: |
23 | global: | |
24 | osd max object name len: 460 | |
25 | osd max object namespace len: 64 | |
26 | - exec: | |
27 | client.0: | |
28 | - ceph osd pool create ec-ca 1 1 | |
29 | - ceph osd pool create ec 1 1 erasure default | |
c07f9fc5 | 30 | - ceph osd pool application enable ec rados |
7c673cae FG |
31 | - ceph osd tier add ec ec-ca |
32 | - ceph osd tier cache-mode ec-ca readproxy | |
33 | - ceph osd tier set-overlay ec ec-ca | |
34 | - ceph osd pool set ec-ca hit_set_type bloom | |
35 | - ceph osd pool set-quota ec-ca max_bytes 20480000 | |
36 | - ceph osd pool set-quota ec max_bytes 20480000 | |
37 | - ceph osd pool set ec-ca target_max_bytes 20480000 | |
38 | - timeout 30 rados -p ec-ca bench 30 write || true | |
31f18b77 FG |
39 | - ceph osd pool set-quota ec-ca max_bytes 0 |
40 | - ceph osd pool set-quota ec max_bytes 0 |