test: add timing calls to run_test

And clean out the corresponding calls to timing_enter and timing_exit
from the various test scripts.

Change-Id: I0759417b5a529e4c3649ce04cca1799c089da278
Signed-off-by: Seth Howell <seth.howell@intel.com>
Reviewed-on: https://review.gerrithub.io/c/spdk/spdk/+/476804
Tested-by: SPDK CI Jenkins <sys_sgci@intel.com>
Reviewed-by: Ben Walker <benjamin.walker@intel.com>
Reviewed-by: Jim Harris <james.r.harris@intel.com>
Community-CI: Broadcom SPDK FC-NVMe CI <spdk-ci.pdl@broadcom.com>
This commit is contained in:
Seth Howell 2019-12-04 13:47:50 -07:00 committed by Tomasz Zawadzki
parent fd17f1aabc
commit 1fafd71bc3
76 changed files with 4 additions and 248 deletions

View File

@ -144,10 +144,8 @@ fi
#####################
if [ $SPDK_TEST_UNITTEST -eq 1 ]; then
timing_enter unittest
run_test suite "unittest" ./test/unit/unittest.sh
report_test_completion "unittest"
timing_exit unittest
fi
@ -223,12 +221,10 @@ if [ $SPDK_RUN_FUNCTIONAL_TEST -eq 1 ]; then
fi
if [ $SPDK_TEST_LVOL -eq 1 ]; then
timing_enter lvol
run_test suite "lvol" ./test/lvol/lvol.sh --test-cases=all
run_test suite "lvol2" ./test/lvol/lvol2.sh
run_test suite "blob_io_wait" ./test/blobstore/blob_io_wait/blob_io_wait.sh
report_test_completion "lvol"
timing_exit lvol
fi
if [ $SPDK_TEST_VHOST_INIT -eq 1 ]; then

View File

@ -112,11 +112,9 @@ function raid_function_test() {
return 0
}
timing_enter bdev_raid
trap 'on_error_exit;' ERR
raid_function_test
rm -f $tmp_file
report_test_completion "bdev_raid"
timing_exit bdev_raid

View File

@ -143,8 +143,6 @@ function qos_function_test() {
fi
}
timing_enter bdev
# Create a file to be used as an AIO backend
dd if=/dev/zero of=/tmp/aiofile bs=2048 count=5000
@ -301,4 +299,3 @@ rm -f /tmp/spdk-pmem-pool
rm -f $testdir/bdev.conf
rbd_cleanup
report_test_completion "bdev"
timing_exit bdev

View File

@ -122,8 +122,6 @@ function blobfs_fuse_test() {
killprocess $blobfs_pid
}
timing_enter blobfs
trap 'on_error_exit;' ERR
# Create one temp file as test bdev
@ -146,5 +144,3 @@ blobfs_fuse_test
rm -rf $mount_dir
rm -f $tmp_file
report_test_completion "blobfs"
timing_exit blobfs

View File

@ -39,8 +39,6 @@ if [ ! -e $DB_BENCH_DIR ]; then
exit 0
fi
timing_enter rocksdb
timing_enter db_bench_build
pushd $DB_BENCH_DIR
@ -154,4 +152,3 @@ run_bsdump
rm -f $ROCKSDB_CONF
report_test_completion "blobfs"
timing_exit rocksdb

View File

@ -12,8 +12,6 @@ source $rootdir/test/common/autotest_common.sh
source $rootdir/test/nvmf/common.sh
rpc_py="$rootdir/scripts/rpc.py"
timing_enter blob_bdev_io_wait
truncate -s 64M $testdir/aio.bdev
$rootdir/test/app/bdev_svc/bdev_svc &
@ -63,4 +61,3 @@ rm -rf $testdir/aio.bdev
trap - SIGINT SIGTERM EXIT
report_test_completion "blob_io_wait"
timing_exit bdev_io_wait

View File

@ -10,8 +10,6 @@ testdir=$(readlink -f $(dirname $0))
rootdir=$(readlink -f $testdir/../..)
source $rootdir/test/common/autotest_common.sh
timing_enter blobstore
# Nvme0 target configuration
$rootdir/scripts/gen_nvme.sh > $testdir/blobcli.conf
@ -33,5 +31,3 @@ rm -rf $testdir/btest.out
rm -rf $testdir/blobcli.conf
rm -rf $testdir/*.blob
rm -rf $testdir/test.pattern
timing_exit blobstore

View File

@ -575,6 +575,8 @@ function run_test() {
shift
local test_name="$1"
shift
timing_enter $test_name
echo "************************************"
echo "START TEST $test_type $test_name"
echo "************************************"
@ -584,6 +586,8 @@ function run_test() {
echo "************************************"
echo "END TEST $test_type $test_name"
echo "************************************"
timing_exit $test_name
xtrace_restore
}

View File

@ -73,7 +73,6 @@ function run_bdevperf() {
killprocess $bdevperf_pid
}
timing_enter compress_test
mkdir -p /tmp/pmem
# per patch bdevperf uses slightly different params than nightly
@ -110,4 +109,3 @@ if [ $RUN_NIGHTLY -eq 1 ]; then
fi
rm -rf /tmp/pmem
timing_exit compress_test

3
test/env/env.sh vendored
View File

@ -4,8 +4,6 @@ testdir=$(readlink -f $(dirname $0))
rootdir=$(readlink -f $testdir/../..)
source $rootdir/test/common/autotest_common.sh
timing_enter env
timing_enter memory
$testdir/memory/memory_ut
timing_exit memory
@ -41,4 +39,3 @@ if [ $(uname) = Linux ]; then
fi
report_test_completion "env"
timing_exit env

View File

@ -4,9 +4,7 @@ testdir=$(readlink -f $(dirname $0))
rootdir=$(readlink -f $testdir/../..)
source $rootdir/test/common/autotest_common.sh
timing_enter event
$testdir/event_perf/event_perf -m 0xF -t 1
$testdir/reactor/reactor -t 1
$testdir/reactor_perf/reactor_perf -t 1
report_test_completion "event"
timing_exit event

View File

@ -45,34 +45,21 @@ if [ -z "$nv_cache" ]; then
echo "Couldn't find NVMe device to be used as non-volatile cache"
fi
timing_enter ftl
timing_enter bdevperf
run_test suite "ftl_bdevperf" $testdir/bdevperf.sh $device
timing_exit bdevperf
timing_enter restore
run_test suite "ftl_restore" $testdir/restore.sh $device
if [ -n "$nv_cache" ]; then
run_test suite "ftl_restore_nv_cache" $testdir/restore.sh -c $nv_cache $device
fi
timing_exit restore
if [ -n "$nv_cache" ]; then
timing_enter dirty_shutdown
run_test suite "ftl_dirty_shutdown" $testdir/dirty_shutdown.sh -c $nv_cache $device
timing_exit dirty_shutdown
fi
timing_enter json
run_test suite "ftl_json" $testdir/json.sh $device
timing_exit json
if [ $SPDK_TEST_FTL_EXTENDED -eq 1 ]; then
timing_enter fio_basic
run_test suite "ftl_fio_basic" $testdir/fio.sh $device basic
timing_exit fio_basic
$rootdir/app/spdk_tgt/spdk_tgt &
svc_pid=$!
@ -85,9 +72,5 @@ if [ $SPDK_TEST_FTL_EXTENDED -eq 1 ]; then
trap - SIGINT SIGTERM EXIT
timing_enter fio_extended
run_test suite "ftl_fio_extended" $testdir/fio.sh $device extended $uuid
timing_exit fio_extended
fi
timing_exit ftl

View File

@ -4,8 +4,6 @@ testdir=$(readlink -f $(dirname $0))
rootdir=$(readlink -f $testdir/../..)
source $rootdir/test/common/autotest_common.sh
timing_enter ioat
timing_enter perf
$rootdir/examples/ioat/perf/ioat_perf -t 1
timing_exit perf
@ -15,4 +13,3 @@ $rootdir/examples/ioat/verify/verify -t 1
timing_exit verify
report_test_completion "ioat"
timing_exit ioat

View File

@ -7,8 +7,6 @@ source $rootdir/test/iscsi_tgt/common.sh
iscsitestinit $1 $2
timing_enter bdev_io_wait
MALLOC_BDEV_SIZE=64
MALLOC_BLOCK_SIZE=512
@ -55,4 +53,3 @@ killprocess $pid
iscsitestfini $1 $2
report_test_completion "bdev_io_wait"
timing_exit bdev_io_wait

View File

@ -13,8 +13,6 @@ if [ ! -d /usr/local/calsoft ]; then
exit 0
fi
timing_enter calsoft
MALLOC_BDEV_SIZE=64
MALLOC_BLOCK_SIZE=512
@ -67,5 +65,4 @@ trap - SIGINT SIGTERM EXIT
killprocess $pid
delete_tmp_conf_files
timing_exit calsoft
exit $failed

View File

@ -57,8 +57,6 @@ function iscsi_header_data_digest_test() {
timing_exit both_enabled
}
timing_enter digests
MALLOC_BDEV_SIZE=64
MALLOC_BLOCK_SIZE=512
@ -106,4 +104,3 @@ trap - SIGINT SIGTERM EXIT
iscsicleanup
killprocess $pid
iscsitestfini $1 $2
timing_exit digests

View File

@ -9,8 +9,6 @@ source $rootdir/test/iscsi_tgt/common.sh
# $2 = test type posix or vpp. defaults to posix.
iscsitestinit $1 $2
timing_enter ext4test
rpc_py="$rootdir/scripts/rpc.py"
node_base="iqn.2013-06.com.intel.ch.spdk"
@ -132,4 +130,3 @@ fi
killprocess $pid
iscsitestfini $1 $2
report_test_completion "nightly_iscsi_ext4test"
timing_exit ext4test

View File

@ -10,8 +10,6 @@ source $rootdir/scripts/common.sh
# $2 = test type posix or vpp. defaults to posix.
iscsitestinit $1 $2
timing_enter filesystem
rpc_py="$rootdir/scripts/rpc.py"
# Remove lvol bdevs and stores.
function remove_backends() {
@ -144,4 +142,3 @@ iscsicleanup
remove_backends
killprocess $pid
iscsitestfini $1 $2
timing_exit filesystem

View File

@ -54,8 +54,6 @@ if [ -z "$INITIATOR_IP" ]; then
exit 1
fi
timing_enter fio
MALLOC_BDEV_SIZE=64
MALLOC_BLOCK_SIZE=4096
@ -151,5 +149,3 @@ trap - SIGINT SIGTERM EXIT
killprocess $pid
iscsitestfini $1 $2
timing_exit fio

View File

@ -9,8 +9,6 @@ source $rootdir/test/iscsi_tgt/common.sh
# $2 = test type posix or vpp. defaults to posix.
iscsitestinit $1 $2
timing_enter initiator
MALLOC_BDEV_SIZE=64
MALLOC_BLOCK_SIZE=512
@ -57,4 +55,3 @@ killprocess $pid
iscsitestfini $1 $2
report_test_completion "iscsi_initiator"
timing_exit initiator

View File

@ -75,8 +75,6 @@ function rpc_add_target_node() {
$rpc_py -s $1 net_interface_delete_ip_address 1 $MIGRATION_ADDRESS
}
timing_enter ip_migration
echo "Running ip migration tests"
for ((i = 0; i < 2; i++)); do
timing_enter start_iscsi_tgt_$i
@ -130,4 +128,3 @@ iscsicleanup
$rpc_py -s $rpc_second_addr spdk_kill_instance SIGTERM
iscsitestfini $1 $2
report_test_completion "iscsi_ip_migration"
timing_exit ip_migration

View File

@ -9,8 +9,6 @@ fi
source $rootdir/test/iscsi_tgt/common.sh
timing_enter iscsi_tgt
# $1 = test type (posix/vpp)
if [ "$1" == "posix" ] || [ "$1" == "vpp" ]; then
TEST_TYPE=$1
@ -82,4 +80,3 @@ fi
cleanup_veth_interfaces $TEST_TYPE
trap - SIGINT SIGTERM EXIT
timing_exit iscsi_tgt

View File

@ -9,8 +9,6 @@ source $rootdir/test/iscsi_tgt/common.sh
# $2 = test type posix or vpp. defaults to posix.
iscsitestinit $1 $2
timing_enter iscsi_lvol
MALLOC_BDEV_SIZE=128
MALLOC_BLOCK_SIZE=512
if [ $RUN_NIGHTLY -eq 1 ]; then
@ -85,4 +83,3 @@ rm -f ./local-job*
iscsicleanup
killprocess $pid
iscsitestfini $1 $2
timing_exit iscsi_lvol

View File

@ -34,8 +34,6 @@ function remove_backends() {
return 0
}
timing_enter multiconnection
timing_enter start_iscsi_tgt
$ISCSI_APP --wait-for-rpc &
iscsipid=$!
@ -84,4 +82,3 @@ iscsicleanup
remove_backends
killprocess $iscsipid
iscsitestfini $1 $2
timing_exit multiconnection

View File

@ -53,8 +53,6 @@ function run_nvme_remote() {
iscsiadm -m node --login -p $TARGET_IP:$ISCSI_PORT
}
timing_enter nvme_remote
# Start the NVMf target
NVMF_APP="$rootdir/app/nvmf_tgt/nvmf_tgt"
$NVMF_APP -m 0x2 -p 1 -s 512 --wait-for-rpc &
@ -102,4 +100,3 @@ $rpc_py nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1
report_test_completion "iscsi_nvme_remote"
iscsitestfini $1 $2
nvmftestfini
timing_exit nvme_remote

View File

@ -15,8 +15,6 @@ PMEM_PER_TGT=1
rpc_py="$rootdir/scripts/rpc.py"
fio_py="$rootdir/scripts/fio.py"
timing_enter iscsi_pmem
timing_enter start_iscsi_target
$ISCSI_APP -m $ISCSI_TEST_CORE_MASK --wait-for-rpc &
pid=$!
@ -75,4 +73,3 @@ rm -f ./local-job*
rm -f /tmp/pool_file*
killprocess $pid
report_test_completion "nightly_iscsi_pmem"
timing_exit iscsi_pmem

View File

@ -50,8 +50,6 @@ if [ -z "$INITIATOR_IP" ]; then
exit 1
fi
timing_enter qos
MALLOC_BDEV_SIZE=64
MALLOC_BLOCK_SIZE=512
IOPS_RESULT=
@ -145,4 +143,3 @@ trap - SIGINT SIGTERM EXIT
killprocess $pid
iscsitestfini $1 $2
timing_exit qos

View File

@ -19,8 +19,6 @@ rbd_setup $TARGET_IP $TARGET_NAMESPACE
trap 'rbd_cleanup; exit 1' SIGINT SIGTERM EXIT
timing_exit rbd_setup
timing_enter rbd
rpc_py="$rootdir/scripts/rpc.py"
fio_py="$rootdir/scripts/fio.py"
@ -68,4 +66,3 @@ rbd_cleanup
iscsitestfini $1 $2
report_test_completion "iscsi_rbd"
timing_exit rbd

View File

@ -9,8 +9,6 @@ source $rootdir/test/iscsi_tgt/common.sh
# $2 = test type posix or vpp. defaults to posix.
iscsitestinit $1 $2
timing_enter reset
MALLOC_BDEV_SIZE=64
MALLOC_BLOCK_SIZE=512
@ -77,4 +75,3 @@ trap - SIGINT SIGTERM EXIT
iscsicleanup
killprocess $pid
iscsitestfini $1 $2
timing_exit reset

View File

@ -9,8 +9,6 @@ source $rootdir/test/iscsi_tgt/common.sh
# $2 = "iso" - triggers isolation mode (setting up required environment).
iscsitestinit $2 $1
timing_enter rpc_config
if [ "$1" == "posix" ] || [ "$1" == "vpp" ]; then
TEST_TYPE=$1
else
@ -63,5 +61,3 @@ iscsicleanup
killprocess $pid
iscsitestfini $2 $1
timing_exit rpc_config

View File

@ -28,8 +28,6 @@ if [ -z "$INITIATOR_IP" ]; then
exit 1
fi
timing_enter trace_record
NUM_TRACE_ENTRIES=4096
MALLOC_BDEV_SIZE=64
MALLOC_BLOCK_SIZE=4096
@ -135,4 +133,3 @@ done
trap - SIGINT SIGTERM EXIT
iscsitestfini $1 $2
timing_exit trace_record

View File

@ -89,8 +89,6 @@ if [ ! -e "$base_img" ]; then
exit 0
fi
timing_enter hotplug
timing_enter start_qemu
qemu-img create -b "$base_img" -f qcow2 "$test_img"
@ -146,5 +144,3 @@ rm "$test_img"
report_test_completion "nvme_hotplug"
timing_exit hotplug_test
timing_exit hotplug

View File

@ -5,8 +5,6 @@ rootdir=$(readlink -f $testdir/../..)
source $rootdir/scripts/common.sh
source $rootdir/test/common/autotest_common.sh
timing_enter nvme
if [ $(uname) = Linux ]; then
# check that our setup.sh script does not bind NVMe devices to uio/vfio if they
# have an active mountpoint
@ -160,5 +158,3 @@ if [ -d /usr/src/fio ]; then
done
timing_exit fio_plugin
fi
timing_exit nvme

View File

@ -17,8 +17,6 @@ if [ ! -d $spdk_nvme_cli ]; then
exit 0
fi
timing_enter nvme_cli
if [ $(uname) = Linux ]; then
start_stub "-s 2048 -i 0 -m 0xF"
trap "kill_stub; exit 1" SIGINT SIGTERM EXIT
@ -52,4 +50,3 @@ if [ $(uname) = Linux ]; then
fi
report_test_completion spdk_nvme_cli
timing_exit nvme_cli

View File

@ -5,8 +5,6 @@ rootdir=$(readlink -f $testdir/../..)
source $rootdir/scripts/common.sh
source $rootdir/test/common/autotest_common.sh
timing_enter nvme_cli_cuse
NVME_CMD=/usr/local/src/nvme-cli/nvme
rpc_py=$rootdir/scripts/rpc.py
@ -74,4 +72,3 @@ trap - SIGINT SIGTERM EXIT
killprocess $spdk_tgt_pid
report_test_completion spdk_nvme_cli_cuse
timing_exit nvme_cli_cuse

View File

@ -7,7 +7,6 @@ source $rootdir/test/nvmf/common.sh
rpc_py="$rootdir/scripts/rpc.py"
timing_enter aer
nvmftestinit
nvmfappstart "-m 0xF"
@ -49,4 +48,3 @@ $rpc_py nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1
trap - SIGINT SIGTERM EXIT
nvmftestfini
timing_exit aer

View File

@ -23,8 +23,6 @@ function tgt_init()
nvmftestinit
timing_enter bdevperf
tgt_init
echo "[Nvme]" > $testdir/bdevperf.conf
@ -48,4 +46,3 @@ $rpc_py nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1
trap - SIGINT SIGTERM EXIT
nvmftestfini
timing_exit bdevperf

View File

@ -15,7 +15,6 @@ if [ ! -d /usr/src/fio ]; then
exit 1
fi
timing_enter fio
timing_enter start_nvmf_tgt
$NVMF_APP -m 0xF &
@ -81,4 +80,3 @@ trap - SIGINT SIGTERM EXIT
rm -f ./local-test-0-verify.state
nvmftestfini
timing_exit fio

View File

@ -12,7 +12,6 @@ rpc_py="$rootdir/scripts/rpc.py"
nvmftestinit
timing_enter identify
timing_enter start_nvmf_tgt
$NVMF_APP -m 0xF &
@ -53,4 +52,3 @@ $rpc_py nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1
trap - SIGINT SIGTERM EXIT
nvmftestfini
timing_exit identify

View File

@ -7,8 +7,6 @@ source $rootdir/test/nvmf/common.sh
nvmftestinit
timing_enter identify_kernel_nvmf_tgt
subsystemname=nqn.2016-06.io.spdk:testnqn
modprobe null_blk nr_devices=1
@ -71,4 +69,3 @@ rmmod null_blk
rmmod nvmet
nvmftestfini
timing_exit identify_kernel_nvmf_tgt

View File

@ -10,8 +10,6 @@ MALLOC_BLOCK_SIZE=512
rpc_py="$rootdir/scripts/rpc.py"
timing_enter perf
nvmftestinit
nvmfappstart "-m 0xF"
@ -92,4 +90,3 @@ fi
trap - SIGINT SIGTERM EXIT
nvmftestfini
timing_exit perf

View File

@ -34,9 +34,6 @@ if [ $TEST_TRANSPORT == "rdma" ] && check_ip_is_soft_roce $NVMF_FIRST_TARGET_IP;
exit 0
fi
timing_enter target_disconnect
# Test to make sure we don't segfault or access null pointers when we try to connect to
# a discovery controller that doesn't exist yet.
set +e
@ -88,4 +85,3 @@ fi
trap - SIGINT SIGTERM EXIT
rm -f $PLUGIN_DIR/example_config_extended.fio || true
nvmftestfini
timing_exit target_disconnect

View File

@ -9,8 +9,6 @@ fi
source $rootdir/test/nvmf/common.sh
timing_enter nvmf_tgt
trap "exit 1" SIGINT SIGTERM EXIT
TEST_ARGS=( "$@" )
@ -59,4 +57,3 @@ trap - SIGINT SIGTERM EXIT
revert_soft_roce
report_test_completion "nvmf"
timing_exit nvmf_tgt

View File

@ -10,7 +10,6 @@ MALLOC_BLOCK_SIZE=512
rpc_py="$rootdir/scripts/rpc.py"
timing_enter bdev_io_wait
nvmftestinit
nvmfappstart "-m 0xF --wait-for-rpc"
@ -47,4 +46,3 @@ $rpc_py nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1
trap - SIGINT SIGTERM EXIT
nvmftestfini
timing_exit bdev_io_wait

View File

@ -10,7 +10,6 @@ MALLOC_BLOCK_SIZE=512
rpc_py="$rootdir/scripts/rpc.py"
timing_enter bdevio
nvmftestinit
# Don't use cores 0 - 2 to avoid overlap with bdevio.
nvmfappstart "-m 0x78"
@ -32,4 +31,3 @@ $rpc_py nvmf_delete_subsystem nqn.2016-06.io.spdk:cnode1
trap - SIGINT SIGTERM EXIT
nvmftestfini
timing_exit bdev_io_wait

View File

@ -11,8 +11,6 @@ MALLOC_BLOCK_SIZE=512
rpc_py="$rootdir/scripts/rpc.py"
# connect disconnect is geared towards ensuring that we are properly freeing resources after disconnecting qpairs.
timing_enter connect_disconnect
nvmftestinit
nvmfappstart "-m 0xF"
@ -43,4 +41,3 @@ set -x
trap - SIGINT SIGTERM EXIT
nvmftestfini
timing_exit connect_disconnect

View File

@ -15,7 +15,6 @@ if ! hash nvme; then
exit 0
fi
timing_enter cr_trprt
nvmftestinit
nvmfappstart "-m 0xF"
@ -51,4 +50,3 @@ fi
trap - SIGINT SIGTERM EXIT
nvmftestfini
timing_exit crt_trprt

View File

@ -15,7 +15,6 @@ if ! hash nvme; then
exit 0
fi
timing_enter discovery
nvmftestinit
nvmfappstart "-m 0xF"
@ -48,4 +47,3 @@ fi
trap - SIGINT SIGTERM EXIT
nvmftestfini
timing_exit discovery

View File

@ -10,8 +10,6 @@ MALLOC_BLOCK_SIZE=512
rpc_py="$rootdir/scripts/rpc.py"
timing_enter fs_test
nvmftestinit
for incapsule in 0 4096; do
@ -72,4 +70,3 @@ for incapsule in 0 4096; do
done
nvmftestfini
timing_exit fs_test

View File

@ -10,7 +10,6 @@ MALLOC_BLOCK_SIZE=512
rpc_py="$rootdir/scripts/rpc.py"
timing_enter fio
nvmftestinit
nvmfappstart "-m 0xF"
@ -78,4 +77,3 @@ rm -f ./local-job2-2-verify.state
trap - SIGINT SIGTERM EXIT
nvmftestfini
timing_exit fio

View File

@ -9,8 +9,6 @@ rpc_py="$rootdir/scripts/rpc.py"
nvmftestinit
timing_enter fuzz_test
$NVMF_APP -m 0xF >$output_dir/nvmf_fuzz_tgt_output.txt 2>&1 &
nvmfpid=$!
@ -42,4 +40,3 @@ nvmfcleanup
killprocess $nvmfpid
nvmftestfini
timing_exit fuzz_test

View File

@ -10,7 +10,6 @@ MALLOC_BLOCK_SIZE=512
rpc_py="$rootdir/scripts/rpc.py"
timing_enter initiator_timeout
nvmftestinit
nvmfappstart "-m 0xF"
@ -70,4 +69,3 @@ rm -f ./local-job0-0-verify.state
trap - SIGINT SIGTERM EXIT
nvmftestfini
timing_exit initiator_timeout

View File

@ -12,7 +12,6 @@ NVMF_SUBSYS=11
rpc_py="$rootdir/scripts/rpc.py"
timing_enter multiconnection
nvmftestinit
nvmfappstart "-m 0xF"
@ -55,4 +54,3 @@ rm -f ./local-job0-0-verify.state
trap - SIGINT SIGTERM EXIT
nvmftestfini
timing_exit multiconnection

View File

@ -8,7 +8,6 @@ source $rootdir/test/nvmf/common.sh
# multitarget RPCs as the in-tree applications don't support multi-target functionality.
rpc_py="$rootdir/test/nvmf/target/multitarget_rpc.py"
timing_enter multitarget
nvmftestinit
nvmfappstart "-m 0xF"
@ -36,4 +35,3 @@ fi
trap - SIGINT SIGTERM EXIT
nvmftestfini
timing_exit multitarget

View File

@ -10,7 +10,6 @@ MALLOC_BLOCK_SIZE=512
rpc_py="$rootdir/scripts/rpc.py"
timing_enter nmic
nvmftestinit
nvmfappstart "-m 0xF"
@ -55,4 +54,3 @@ nvme disconnect -n "nqn.2016-06.io.spdk:cnode1" || true
trap - SIGINT SIGTERM EXIT
nvmftestfini
timing_exit nmic

View File

@ -17,7 +17,6 @@ MALLOC_BLOCK_SIZE=512
rpc_py="$rootdir/scripts/rpc.py"
timing_enter nvme_cli
nvmftestinit
nvmfappstart "-m 0xF"
@ -75,4 +74,3 @@ trap - SIGINT SIGTERM EXIT
nvmftestfini
report_test_completion "nvmf_spdk_nvme_cli"
timing_exit nvme_cli

View File

@ -12,7 +12,6 @@ LVOL_BDEV_FINAL_SIZE=30
rpc_py="$rootdir/scripts/rpc.py"
timing_enter lvol_integrity
nvmftestinit
nvmfappstart "-m 0x7"
@ -67,4 +66,3 @@ rm -f ./local-job*
trap - SIGINT SIGTERM EXIT
nvmftestfini
timing_exit lvol_integrity

View File

@ -21,7 +21,6 @@ VHOST_SOCK="/tmp/vhost_rpc.sock"
VHOST_APP="$rootdir/app/vhost/vhost -p 0 -r $VHOST_SOCK -u"
VHOST_RPC="$rootdir/scripts/rpc.py -s $VHOST_SOCK"
timing_enter nvmf_vhost
nvmftestinit
# Start Apps
@ -68,4 +67,3 @@ trap - SIGINT SIGTERM EXIT
killprocess $vhostpid
nvmftestfini
timing_exit nvmf_vhost

View File

@ -19,7 +19,6 @@ function jsum()
jq "$filter" | awk '{s+=$1}END{print s}'
}
timing_enter rpc
nvmftestinit
nvmfappstart "-m 0xF"
@ -127,4 +126,3 @@ fi
trap - SIGINT SIGTERM EXIT
nvmftestfini
timing_exit rpc

View File

@ -34,7 +34,6 @@ function waitforio() {
return $ret
}
timing_enter shutdown
nvmftestinit
nvmfappstart "-m 0x1E"
@ -140,4 +139,3 @@ trap - SIGINT SIGTERM EXIT
timing_enter testfini
nvmftestfini
timing_exit testfini
timing_exit shutdown

View File

@ -10,7 +10,6 @@ MALLOC_BLOCK_SIZE=512
rpc_py="$rootdir/scripts/rpc.py"
timing_enter srq_overwhelm
nvmftestinit
if check_ip_is_soft_roce $NVMF_FIRST_TARGET_IP; then
@ -49,4 +48,3 @@ done
trap - SIGINT SIGTERM EXIT
nvmftestfini
timing_exit srq_overwhelm

View File

@ -7,13 +7,9 @@ source $rootdir/test/common/autotest_common.sh
function suite()
{
timing_enter $(basename "$@")
run_test suite "ocf_$(basename "$@")" "$@"
timing_exit $(basename "$@")
}
timing_enter ocf
suite "$testdir/integrity/fio-modes.sh"
suite "$testdir/integrity/bdevperf-iotypes.sh"
suite "$testdir/integrity/stats.sh"
@ -22,5 +18,4 @@ suite "$testdir/management/multicore.sh"
suite "$testdir/management/persistent-metadata.sh"
suite "$testdir/management/remove.sh"
timing_exit ocf
report_test_completion "ocf"

View File

@ -645,7 +645,6 @@ function delete_bdev_tc2()
return 0
}
timing_enter pmem
vhost_start
if ! $enable_script_debug; then
set +x
@ -694,4 +693,3 @@ fi
pmem_clean_pool_file
report_test_completion "pmem"
vhost_kill 0
timing_exit pmem

View File

@ -4,8 +4,6 @@ testdir=$(readlink -f $(dirname $0))
rootdir=$(readlink -f $testdir/../..)
source $rootdir/test/common/autotest_common.sh
timing_enter rpc_client
$rootdir/test/rpc_client/rpc_client_test
timing_exit rpc_client
trap - SIGINT SIGTERM EXIT

View File

@ -9,7 +9,6 @@ source $rootdir/test/iscsi_tgt/common.sh
MATCH_FILE="spdkcli_iscsi.test"
SPDKCLI_BRANCH="/iscsi"
timing_enter spdkcli_iscsi
trap 'on_error_exit;' ERR
timing_enter run_iscsi_tgt
@ -71,5 +70,4 @@ $spdkcli_job "'/iscsi/auth_groups delete_secret 1 test2' 'user=test2'
timing_exit spdkcli_clear_iscsi_config
killprocess $iscsi_tgt_pid
timing_exit spdkcli_iscsi
report_test_completion spdk_cli

View File

@ -9,7 +9,6 @@ source $rootdir/test/nvmf/common.sh
MATCH_FILE="spdkcli_nvmf.test"
SPDKCLI_BRANCH="/nvmf"
timing_enter spdkcli_nvmf
trap 'on_error_exit; revert_soft_roce' ERR
rdma_device_init
@ -84,5 +83,4 @@ timing_exit spdkcli_clear_nvmf_config
killprocess $nvmf_tgt_pid
#revert_soft_roce
timing_exit spdkcli_nvmf
report_test_completion spdk_cli_nvmf

View File

@ -8,7 +8,6 @@ source $rootdir/test/spdkcli/common.sh
MATCH_FILE="spdkcli_pmem.test"
SPDKCLI_BRANCH="/bdevs/pmemblk"
timing_enter spdkcli_pmem
trap 'rm -f $testdir/match_files/spdkcli_pmem_info.test; on_error_exit;' ERR
timing_enter run_spdk_tgt
@ -46,5 +45,4 @@ rm -f $testdir/match_files/spdkcli_pmem_info.test
timing_exit spdkcli_clear_pmem_config
killprocess $spdk_tgt_pid
timing_exit spdkcli_pmem
report_test_completion spdk_cli

View File

@ -11,7 +11,6 @@ SPDKCLI_BRANCH="/bdevs"
testdir=$(readlink -f $(dirname $0))
. $testdir/common.sh
timing_enter spdkcli_raid
trap 'on_error_exit;' ERR
timing_enter run_spdk_tgt
@ -45,5 +44,4 @@ $spdkcli_job "'/bdevs/malloc delete Malloc1' '' True
timing_exit spdkcli_delete_malloc
killprocess $spdk_tgt_pid
timing_exit spdkcli_raid
report_test_completion spdk_cli

View File

@ -8,7 +8,6 @@ source $rootdir/test/spdkcli/common.sh
MATCH_FILE="spdkcli_rbd.test"
SPDKCLI_BRANCH="/bdevs/rbd"
timing_enter spdk_cli_rbd
trap 'on_error_exit' ERR
timing_enter run_spdk_tgt
run_spdk_tgt
@ -32,5 +31,4 @@ timing_exit spdkcli_clear_rbd_config
killprocess $spdk_tgt_pid
timing_exit spdk_cli_rbd
report_test_completion spdk_cli_rbd

View File

@ -46,7 +46,6 @@ function waitfortcplisten() {
IP_ADDRESS="127.0.0.1"
PORT="9998"
timing_enter spdkcli_tcp
trap 'on_error_exit;' ERR
timing_enter run_spdk_tgt_tcp
@ -60,6 +59,4 @@ timing_exit run_spdk_tgt_tcp
killprocess $spdk_tgt_pid
timing_exit spdkcli_tcp
report_test_completion spdkcli_tcp

View File

@ -8,7 +8,6 @@ source $rootdir/test/spdkcli/common.sh
MATCH_FILE="spdkcli_vhost.test"
SPDKCLI_BRANCH="/"
timing_enter spdk_cli_vhost
trap 'on_error_exit' ERR
timing_enter run_vhost_tgt
run_vhost_tgt
@ -144,5 +143,4 @@ timing_exit spdkcli_load_config
killprocess $vhost_tgt_pid
timing_exit spdk_cli_vhost
report_test_completion spdk_cli_vhost

View File

@ -6,7 +6,6 @@ source $rootdir/test/common/autotest_common.sh
source $rootdir/test/spdkcli/common.sh
trap 'killprocess $virtio_pid; on_error_exit' ERR
timing_enter spdk_cli_vhost_init
timing_enter run_spdk_tgt
run_spdk_tgt
@ -77,5 +76,3 @@ timing_exit spdkcli_clear_virtio_config
killprocess $virtio_pid
killprocess $spdk_tgt_pid
timing_exit spdk_cli_vhost_init

View File

@ -13,8 +13,6 @@ vhost_rpc_py="$rootdir/scripts/rpc.py"
fuzz_generic_rpc_py="$rootdir/scripts/rpc.py -s $FUZZ_RPC_SOCK"
fuzz_specific_rpc_py="$rootdir/test/app/fuzz/common/fuzz_rpc.py -s $FUZZ_RPC_SOCK"
timing_enter fuzz_test
$VHOST_APP >$output_dir/vhost_fuzz_tgt_output.txt 2>&1 &
vhostpid=$!
waitforlisten $vhostpid
@ -66,5 +64,3 @@ wait $fuzzpid
trap - SIGINT SIGTERM exit
killprocess $vhostpid
timing_exit fuzz_test

View File

@ -29,49 +29,35 @@ DISKS_NUMBER=$(lspci -mm -n | grep 0108 | tr -d '"' | awk -F " " '{print "0000:"
WORKDIR=$(readlink -f $(dirname $0))
timing_enter vhost
timing_enter negative
run_test case "vhost_negative" $WORKDIR/other/negative.sh
report_test_completion "vhost_negative"
timing_exit negative
timing_enter vhost_boot
run_test suite "vhost_boot" $WORKDIR/vhost_boot/vhost_boot.sh --vm_image=$VM_IMAGE
report_test_completion "vhost_boot"
timing_exit vhost_boot
if [ $RUN_NIGHTLY -eq 1 ]; then
timing_enter integrity_blk
echo 'Running blk integrity suite...'
run_test case "vhost_blk_integrity" $WORKDIR/fiotest/fio.sh -x --fio-bin=$FIO_BIN \
--vm=0,$VM_IMAGE,Nvme0n1p0:RaidBdev0:RaidBdev1:RaidBdev2 \
--test-type=spdk_vhost_blk \
--fio-job=$WORKDIR/common/fio_jobs/default_integrity.job
report_test_completion "nightly_vhost_integrity_blk"
timing_exit integrity_blk
timing_enter integrity
echo 'Running SCSI integrity suite...'
run_test case "vhost_scsi_integrity" $WORKDIR/fiotest/fio.sh -x --fio-bin=$FIO_BIN \
--vm=0,$VM_IMAGE,Nvme0n1p0:RaidBdev0:RaidBdev1:RaidBdev2 \
--test-type=spdk_vhost_scsi \
--fio-job=$WORKDIR/common/fio_jobs/default_integrity.job
report_test_completion "nightly_vhost_integrity"
timing_exit integrity
timing_enter fs_integrity_scsi
echo 'Running filesystem integrity suite with SCSI...'
run_test case "vhost_scsi_fs_integrity" $WORKDIR/integrity/integrity_start.sh --ctrl-type=spdk_vhost_scsi --fs="xfs ntfs btrfs ext4"
report_test_completion "vhost_fs_integrity_scsi"
timing_exit fs_integrity_scsi
timing_enter fs_integrity_blk
echo 'Running filesystem integrity suite with BLK...'
run_test case "vhost_blk_fs_integrity" $WORKDIR/integrity/integrity_start.sh --ctrl-type=spdk_vhost_blk --fs="xfs ntfs btrfs ext4"
report_test_completion "vhost_fs_integrity_blk"
timing_exit fs_integrity_blk
timing_enter integrity_lvol_scsi_nightly
if [[ $DISKS_NUMBER -ge 2 ]]; then
echo 'Running lvol integrity nightly suite with two cores and two controllers'
run_test case "vhost_scsi_2core_2ctrl" $WORKDIR/lvol/lvol_test.sh --fio-bin=$FIO_BIN \
@ -89,9 +75,7 @@ if [ $RUN_NIGHTLY -eq 1 ]; then
echo 'Running lvol integrity nightly suite with one core and one controller'
run_test case "vhost_scsi_1core_1ctrl" $WORKDIR/lvol/lvol_test.sh --fio-bin=$FIO_BIN \
--ctrl-type=spdk_vhost_scsi --max-disks=1
timing_exit integrity_lvol_scsi_nightly
timing_enter integrity_lvol_blk_nightly
if [[ $DISKS_NUMBER -ge 2 ]]; then
echo 'Running lvol integrity nightly suite with two cores and two controllers'
run_test case "vhost_blk_2core_2ctrl" $WORKDIR/lvol/lvol_test.sh --fio-bin=$FIO_BIN \
@ -109,37 +93,24 @@ if [ $RUN_NIGHTLY -eq 1 ]; then
echo 'Running lvol integrity nightly suite with one core and one controller'
run_test case "vhost_lvol_integrity_1core_1ctrl" $WORKDIR/lvol/lvol_test.sh --fio-bin=$FIO_BIN \
--ctrl-type=spdk_vhost_blk --max-disks=1
timing_exit integrity_lvol_blk_nightly
timing_enter readonly
echo 'Running readonly tests suite...'
run_test case "vhost_readonly" $WORKDIR/readonly/readonly.sh --vm_image=$VM_IMAGE --disk=Nvme0n1 -x
report_test_completion "vhost_readonly"
timing_exit readonly
timing_enter vhost_migration
echo 'Running migration suite...'
run_test case "vhost_migration" $WORKDIR/migration/migration.sh -x \
--fio-bin=$FIO_BIN --os=$VM_IMAGE --test-cases=1,2
timing_exit vhost_migration
fi
timing_enter integrity_lvol_scsi
echo 'Running lvol integrity suite...'
run_test case "vhost_scsi_lvol_integrity" $WORKDIR/lvol/lvol_test.sh -x --fio-bin=$FIO_BIN \
--ctrl-type=spdk_vhost_scsi --thin-provisioning
report_test_completion "vhost_integrity_lvol_scsi"
timing_exit integrity_lvol_scsi
timing_enter integrity_lvol_blk
echo 'Running lvol integrity suite...'
run_test case "vhost_blk_lvol_integrity" $WORKDIR/lvol/lvol_test.sh -x --fio-bin=$FIO_BIN \
--ctrl-type=spdk_vhost_blk
report_test_completion "vhost_integrity_lvol_blk"
timing_exit integrity_lvol_blk
timing_enter spdk_cli
run_test suite "spdkcli_vhost" ./test/spdkcli/vhost.sh
timing_exit spdk_cli
timing_exit vhost

View File

@ -58,7 +58,6 @@ fi
vhosttestinit
timing_enter vhost_boot
trap 'err_clean "${FUNCNAME}" "${LINENO}"' ERR
timing_enter start_vhost
vhost_run 0
@ -122,6 +121,4 @@ $rpc_py bdev_lvol_delete_lvstore -u $lvs_u
vhost_kill 0
timing_exit clean_vhost
timing_exit vhost_boot
vhosttestfini

View File

@ -14,8 +14,6 @@ if [ -z "$pci_devs" ]; then
exit 1
fi
timing_enter vmd
timing_enter identify
for bdf in $pci_devs; do
$rootdir/examples/nvme/identify/identify -i 0 -V -r "trtype:PCIe traddr:$bdf"
@ -56,5 +54,3 @@ done
trap - SIGINT SIGTERM EXIT
killprocess $svcpid
timing_exit vmd