From: Chunmei Liu Date: Thu, 27 Nov 2025 07:47:37 +0000 (+0000) Subject: qa/suites/crimson-rados: add fio test case for osd shard number changes upon restart X-Git-Url: http://git-server-git.apps.pok.os.sepia.ceph.com/?a=commitdiff_plain;h=cf7be3c2877aa6a8d12e98d40191a920a6d07321;p=ceph-ci.git qa/suites/crimson-rados: add fio test case for osd shard number changes upon restart Signed-off-by: Chunmei Liu --- diff --git a/qa/suites/crimson-rados/osd_shards/.qa b/qa/suites/crimson-rados/osd_shards/.qa new file mode 120000 index 00000000000..a602a0353e7 --- /dev/null +++ b/qa/suites/crimson-rados/osd_shards/.qa @@ -0,0 +1 @@ +../.qa/ \ No newline at end of file diff --git a/qa/suites/crimson-rados/osd_shards/0-crimson_install.yaml b/qa/suites/crimson-rados/osd_shards/0-crimson_install.yaml new file mode 120000 index 00000000000..e1c3e1bb67f --- /dev/null +++ b/qa/suites/crimson-rados/osd_shards/0-crimson_install.yaml @@ -0,0 +1 @@ +.qa/config/crimson_install.yaml \ No newline at end of file diff --git a/qa/suites/crimson-rados/osd_shards/clusters/crimson-fixed.yaml b/qa/suites/crimson-rados/osd_shards/clusters/crimson-fixed.yaml new file mode 100644 index 00000000000..7e6a0db8644 --- /dev/null +++ b/qa/suites/crimson-rados/osd_shards/clusters/crimson-fixed.yaml @@ -0,0 +1,16 @@ +roles: +- [mon.a, mgr.x, osd.0, osd.1, osd.2, client.0] +overrides: + ceph: + conf: + osd: + osd shutdown pgref assert: true + crimson cpu num: 3 + global: + ms cluster mode: crc + ms service mode: crc + ms client mode: crc + ms mon service mode: crc + ms mon cluster mode: crc + ms mon client mode: crc + diff --git a/qa/suites/crimson-rados/osd_shards/crimson-supported-all-distro b/qa/suites/crimson-rados/osd_shards/crimson-supported-all-distro new file mode 120000 index 00000000000..a5b729b9efa --- /dev/null +++ b/qa/suites/crimson-rados/osd_shards/crimson-supported-all-distro @@ -0,0 +1 @@ +.qa/distros/crimson-supported-all-distro/ \ No newline at end of file diff --git a/qa/suites/crimson-rados/osd_shards/crimson_qa_overrides.yaml b/qa/suites/crimson-rados/osd_shards/crimson_qa_overrides.yaml new file mode 120000 index 00000000000..2bf67af1b18 --- /dev/null +++ b/qa/suites/crimson-rados/osd_shards/crimson_qa_overrides.yaml @@ -0,0 +1 @@ +.qa/config/crimson_qa_overrides.yaml \ No newline at end of file diff --git a/qa/suites/crimson-rados/osd_shards/deploy/.qa b/qa/suites/crimson-rados/osd_shards/deploy/.qa new file mode 120000 index 00000000000..a602a0353e7 --- /dev/null +++ b/qa/suites/crimson-rados/osd_shards/deploy/.qa @@ -0,0 +1 @@ +../.qa/ \ No newline at end of file diff --git a/qa/suites/crimson-rados/osd_shards/deploy/ceph.yaml b/qa/suites/crimson-rados/osd_shards/deploy/ceph.yaml new file mode 100644 index 00000000000..1af2f4b1bd2 --- /dev/null +++ b/qa/suites/crimson-rados/osd_shards/deploy/ceph.yaml @@ -0,0 +1,13 @@ +tasks: +- install: +- ceph: + conf: + osd: + debug monc: 20 + mon: + mon min osdmap epochs: 50 + paxos service trim min: 10 + # prune full osdmaps regularly + mon osdmap full prune min: 15 + mon osdmap full prune interval: 2 + mon osdmap full prune txsize: 2 diff --git a/qa/suites/crimson-rados/osd_shards/deploy/cephadm.yaml.disabled b/qa/suites/crimson-rados/osd_shards/deploy/cephadm.yaml.disabled new file mode 100644 index 00000000000..0c2062240ee --- /dev/null +++ b/qa/suites/crimson-rados/osd_shards/deploy/cephadm.yaml.disabled @@ -0,0 +1,16 @@ +# no need to verify os + flavor + sha1 +verify_ceph_hash: false +tasks: +- cephadm: + conf: + mgr: + debug ms: 1 + debug mgr: 20 + debug osd: 10 +- cephadm.shell: + mon.a: + - ceph orch status + - ceph orch ps + - ceph orch ls + - ceph orch host ls + - ceph orch device ls diff --git a/qa/suites/crimson-rados/osd_shards/objectstore b/qa/suites/crimson-rados/osd_shards/objectstore new file mode 120000 index 00000000000..921836a8a5e --- /dev/null +++ b/qa/suites/crimson-rados/osd_shards/objectstore @@ -0,0 +1 @@ +.qa/objectstore/crimson \ No newline at end of file diff --git a/qa/suites/crimson-rados/osd_shards/settings b/qa/suites/crimson-rados/osd_shards/settings new file mode 120000 index 00000000000..e9ff073ab4b --- /dev/null +++ b/qa/suites/crimson-rados/osd_shards/settings @@ -0,0 +1 @@ +../perf/settings \ No newline at end of file diff --git a/qa/suites/crimson-rados/osd_shards/tasks/crimson_fio_restart.yaml b/qa/suites/crimson-rados/osd_shards/tasks/crimson_fio_restart.yaml new file mode 100644 index 00000000000..fae124c200d --- /dev/null +++ b/qa/suites/crimson-rados/osd_shards/tasks/crimson_fio_restart.yaml @@ -0,0 +1,124 @@ +tasks: +- exec: + mon.a: + - ceph -s +- cbt: + benchmarks: + librbdfio: + op_size: [4096] + time: 60 + mode: ['randwrite'] + norandommap: True + vol_size: 4096 + procs_per_volume: [1] + volumes_per_client: [2] + iodepth: [32] + osd_ra: [4096] + pool_profile: 'rbd' + log_avg_msec: 100 + cluster: + user: 'ubuntu' + osds_per_node: 3 + iterations: 1 + pool_profiles: + rbd: + pg_size: 128 + pgp_size: 128 + replication: 3 + monitoring_profiles: + perf: + nodes: + - osds + perf_cmd: 'perf' + args: 'stat -p {pid} -o {perf_dir}/perf_stat.{pid}' + pid_glob: crimson-osd.*.pid + +- exec: + mon.a: + - cephadm shell -- ceph orch stop osd --all +- exec: + mon.a: + - | + ceph config set osd.* crimson_cpu_num 2 + for osd in $(cephadm shell -- ceph osd ls | awk '{print $1}'); do + cephadm shell -- ceph orch daemon restart osd.$osd + done + +- exec: + mon.a: + - ceph -s +- cbt: + benchmarks: + librbdfio: + op_size: [4096] + time: 60 + mode: ['randwrite'] + norandommap: True + vol_size: 4096 + procs_per_volume: [1] + volumes_per_client: [2] + iodepth: [32] + osd_ra: [4096] + pool_profile: 'rbd' + log_avg_msec: 100 + cluster: + user: 'ubuntu' + osds_per_node: 3 + iterations: 1 + pool_profiles: + rbd: + pg_size: 128 + pgp_size: 128 + replication: 3 + monitoring_profiles: + perf: + nodes: + - osds + perf_cmd: 'perf' + args: 'stat -p {pid} -o {perf_dir}/perf_stat.{pid}' + pid_glob: crimson-osd.*.pid + +- exec: + mon.a: + - cephadm shell -- ceph orch stop osd --all +- exec: + mon.a: + - | + ceph config set osd.* crimson_cpu_num 5 + for osd in $(cephadm shell -- ceph osd ls | awk '{print $1}'); do + cephadm shell -- ceph orch daemon restart osd.$osd + done + +- exec: + mon.a: + - ceph -s +- cbt: + benchmarks: + librbdfio: + op_size: [4096] + time: 60 + mode: ['randwrite'] + norandommap: True + vol_size: 4096 + procs_per_volume: [1] + volumes_per_client: [2] + iodepth: [32] + osd_ra: [4096] + pool_profile: 'rbd' + log_avg_msec: 100 + cluster: + user: 'ubuntu' + osds_per_node: 3 + iterations: 1 + pool_profiles: + rbd: + pg_size: 128 + pgp_size: 128 + replication: 3 + monitoring_profiles: + perf: + nodes: + - osds + perf_cmd: 'perf' + args: 'stat -p {pid} -o {perf_dir}/perf_stat.{pid}' + pid_glob: crimson-osd.*.pid