pg_size: 128
pgp_size: 128
replication: 3
+ monitoring_profiles:
+ perf:
+ nodes:
+ - osds
+ perf_cmd: 'perf'
+ args: 'stat -p {pid} -o {perf_dir}/perf_stat.{pid}'
+ pid_glob: ceph-osd.*.pid
osd_ra: [4096]
pool_profile: 'rbd'
log_avg_msec: 100
+ monitoring_profiles:
+ perf:
+ nodes:
+ - osds
+ perf_cmd: 'perf'
+ args: 'stat -p {pid} -o {perf_dir}/perf_stat.{pid}'
+ pid_glob: ceph-osd.*.pid
cluster:
user: 'ubuntu'
osds_per_node: 3
pg_size: 128
pgp_size: 128
replication: 3
+ pid_dir: /var/run/ceph
pg_size: 128
pgp_size: 128
replication: 3
+ monitoring_profiles:
+ perf:
+ nodes:
+ - osds
+ perf_cmd: 'perf'
+ args: 'stat -p {pid} -o {perf_dir}/perf_stat.{pid}'
+ pid_glob: ceph-osd.*.pid
pg_size: 128
pgp_size: 128
replication: 3
+ monitoring_profiles:
+ perf:
+ nodes:
+ - osds
+ perf_cmd: 'perf'
+ args: 'stat -p {pid} -o {perf_dir}/perf_stat.{pid}'
+ pid_glob: ceph-osd.*.pid
pg_size: 128
pgp_size: 128
replication: 3
+ monitoring_profiles:
+ perf:
+ nodes:
+ - osds
+ perf_cmd: 'perf'
+ args: 'stat -p {pid} -o {perf_dir}/perf_stat.{pid}'
+ pid_glob: ceph-osd.*.pid
concurrent_ops: 4
concurrent_procs: 2
op_size: [4096]
- pool_monitoring_list:
- - collectl
pool_profile: 'replicated'
- run_monitoring_list:
- - collectl
time: 60
write_only: false
readmode: 'rand'
pg_size: 256
pgp_size: 256
replication: 'replicated'
+ monitoring_profiles:
+ perf:
+ nodes:
+ - osds
+ perf_cmd: 'perf'
+ args: 'stat -p {pid} -o {perf_dir}/perf_stat.{pid}'
+ pid_glob: ceph-osd.*.pid
pg_size: 256
pgp_size: 256
replication: 'replicated'
+ monitoring_profiles:
+ perf:
+ nodes:
+ - osds
+ perf_cmd: 'perf'
+ args: 'stat -p {pid} -o {perf_dir}/perf_stat.{pid}'
+ pid_glob: ceph-osd.*.pid
pg_size: 256
pgp_size: 256
replication: 'replicated'
+ monitoring_profiles:
+ perf:
+ nodes:
+ - osds
+ perf_cmd: 'perf'
+ args: 'stat -p {pid} -o {perf_dir}/perf_stat.{pid}'
+ pid_glob: ceph-osd.*.pid
pg_size: 256
pgp_size: 256
replication: 'replicated'
+ monitoring_profiles:
+ perf:
+ nodes:
+ - osds
+ perf_cmd: 'perf'
+ args: 'stat -p {pid} -o {perf_dir}/perf_stat.{pid}'
+ pid_glob: ceph-osd.*.pid
pg_size: 256
pgp_size: 256
replication: 'replicated'
+ monitoring_profiles:
+ perf:
+ nodes:
+ - osds
+ perf_cmd: 'perf'
+ args: 'stat -p {pid} -o {perf_dir}/perf_stat.{pid}'
+ pid_glob: ceph-osd.*.pid
pg_size: 128
pgp_size: 128
replication: 3
+ monitoring_profiles:
+ perf:
+ nodes:
+ - osds
+ perf_cmd: 'perf'
+ args: 'stat -p {pid} -o {perf_dir}/perf_stat.{pid}'
+ pid_glob: ceph-osd.*.pid
pg_size: 256
pgp_size: 256
replication: 'replicated'
+ monitoring_profiles:
+ perf:
+ nodes:
+ - osds
+ perf_cmd: 'perf'
+ args: 'stat -p {pid} -o {perf_dir}/perf_stat.{pid}'
+ pid_glob: ceph-osd.*.pid
perf:
nodes:
- osds
+ perf_cmd: 'perf'
args: 'stat -p {pid} -o {perf_dir}/perf_stat.{pid}'
cluster:
osds_per_node: 3