_merge_results()
{
- echo Results for $bench benchmark >>$OUT
+ echo Results for $bench benchmark
headers=`$here/run.$bench -h`
- echo "[$headers]" >>$OUT
- echo results.$bench.* | sort -nu | xargs cat >>$OUT
- echo >>$OUT
+ echo "[$headers]"
+ echo results.$bench.* | sort -nu | xargs cat
+ echo
}
# real QA test starts here
trap "_cleanup; exit \$status" 0 1 2 3 15
_require_scratch
+rm -f bench.*
-echo "MKFS_OPTIONS=$MKFS_OPTIONS" >>$OUT
-echo "MOUNT_OPTIONS=$MOUNT_OPTIONS" >>$OUT
-echo "" >>$OUT
+# $OUT is the report which will ultimately be sent, keep it tidy.
+cat >$OUT <<EOF
+bench: MKFS_OPTIONS=$MKFS_OPTIONS
+bench: MOUNT_OPTIONS=$MOUNT_OPTIONS
+
+EOF
-rm -f bench.*
for bench in $benches
do
echo "" >$FULL
umount $SCRATCH_DEV 2>&1 | _fix_malloc >>$FULL
_run_benchmark | _fix_malloc
- _merge_results
+ _merge_results >>$OUT
_log "*** done $bench"
done
SOAK_PROC=3 # -p option to fsstress
SOAK_STRESS=10000 # -n option to fsstress
SOAK_PASSES=-1 # count of repetitions of fsstress (while soaking)
-BENCH_PASSES=5 # count of repetitions of benchmarks (want averages)
EMAIL=root@localhost # where auto-qa will send its status messages
export MKFS_OPTIONS=${MKFS_OPTIONS:=-bsize=4096}
export MOUNT_OPTIONS=${MOUNT_OPTIONS:=-ologbufs=2}
export MALLOCLIB=${MALLOCLIB:=/usr/lib/libefence.a}
+export BENCH_PASSES=${BENCH_PASSES:=5}
case "$HOST"
in