diff --git a/test/common/skipped_tests.txt b/test/common/skipped_tests.txt index f978e4147..c5f00a00c 100644 --- a/test/common/skipped_tests.txt +++ b/test/common/skipped_tests.txt @@ -50,8 +50,7 @@ iscsi_tgt_fuzz nvmf_fuzz nvmf_multiconnection nvmf_initiator_timeout -vhost_blk_2core_2ctrl -vhost_blk_1core_2ctrl +vhost_blk_cores_2ctrl vhost_blk_fs_integrity vhost_blk_integrity vhost_blk_nightly @@ -63,6 +62,4 @@ vhost_readonly vhost_scsi_fs_integrity vhost_scsi_integrity vhost_scsi_nightly -vhost_scsi_2core_2ctrl -vhost_scsi_1core_2ctrl -vhost_scsi_1core_1ctrl +vhost_scsi_cores_2ctrl diff --git a/test/vhost/lvol/lvol_test.sh b/test/vhost/lvol/lvol_test.sh index 5ff8ffc33..0bcf93b0d 100755 --- a/test/vhost/lvol/lvol_test.sh +++ b/test/vhost/lvol/lvol_test.sh @@ -9,7 +9,6 @@ source $rootdir/scripts/common.sh rpc_py="$rootdir/scripts/rpc.py -s $(get_vhost_dir 0)/rpc.sock" vm_count=1 -max_disks="" ctrl_type="spdk_vhost_scsi" use_fs=false nested_lvol=false @@ -28,8 +27,6 @@ function usage() { echo " --vm-count=INT Virtual machines to use in test;" echo " Each VM will get one lvol bdev on each NVMe." echo " Default: 1" - echo " --max-disks=INT Maximum number of NVMe drives to use in test." - echo " Default: will use all available NVMes." echo " --ctrl-type=TYPE Controller type to use for test:" echo " spdk_vhost_scsi - use spdk vhost scsi" echo " spdk_vhost_blk - use spdk vhost block" @@ -67,10 +64,8 @@ function clean_lvol_cfg() { done notice "Removing lvol stores" - for lvol_store in "${lvol_stores[@]}"; do - $rpc_py bdev_lvol_delete_lvstore -u $lvol_store - notice "lvol store $lvol_store removed" - done + $rpc_py bdev_lvol_delete_lvstore -u "$ls_guid" + notice "lvol store $ls_guid removed" } while getopts 'xh-:' optchar; do @@ -80,7 +75,6 @@ while getopts 'xh-:' optchar; do help) usage $0 ;; fio-bin=*) fio_bin="--fio-bin=${OPTARG#*=}" ;; vm-count=*) vm_count="${OPTARG#*=}" ;; - max-disks=*) max_disks="${OPTARG#*=}" ;; ctrl-type=*) ctrl_type="${OPTARG#*=}" ;; nested-lvol) nested_lvol=true ;; distribute-cores) distribute_cores=true ;; @@ -99,17 +93,6 @@ done vhosttestinit -notice "Get NVMe disks:" -nvmes=($(get_nvme_bdfs)) - -if [[ -z $max_disks ]]; then - max_disks=${#nvmes[@]} -fi - -if ((${#nvmes[@]} < max_disks)); then - fail "Number of NVMe drives (${#nvmes[@]}) is lower than number of requested disks for test ($max_disks)" -fi - if $distribute_cores; then # FIXME: this need to be handled entirely in common.sh source $testdir/autotest.config @@ -125,48 +108,42 @@ notice "..." trap 'clean_lvol_cfg; error_exit "${FUNCNAME}" "${LINENO}"' SIGTERM SIGABRT ERR -lvol_stores=() lvol_bdevs=() nest_lvol_stores=() nest_lvol_bdevs=() used_vms="" -# On each NVMe create one lvol store -for ((i = 0; i < max_disks; i++)); do +id=0 +# Create base lvol store on NVMe +notice "Creating lvol store on device Nvme${id}n1" +ls_guid=$($rpc_py bdev_lvol_create_lvstore Nvme0n1 lvs_$id -c 4194304) +if $nested_lvol; then + free_mb=$(get_lvs_free_mb "$ls_guid") + size=$((free_mb / (vm_count + 1))) - # Create base lvol store on NVMe - notice "Creating lvol store on device Nvme${i}n1" - ls_guid=$($rpc_py bdev_lvol_create_lvstore Nvme${i}n1 lvs_$i -c 4194304) - lvol_stores+=("$ls_guid") + notice "Creating lvol bdev on lvol store: $ls_guid" + lb_name=$($rpc_py bdev_lvol_create -u $ls_guid lbd_nest $size $thin) - if $nested_lvol; then - free_mb=$(get_lvs_free_mb "$ls_guid") - size=$((free_mb / (vm_count + 1))) + notice "Creating nested lvol store on lvol bdev: $lb_name" + nest_ls_guid=$($rpc_py bdev_lvol_create_lvstore $lb_name lvs_n_$id -c 4194304) + nest_lvol_stores+=("$nest_ls_guid") - notice "Creating lvol bdev on lvol store: $ls_guid" - lb_name=$($rpc_py bdev_lvol_create -u $ls_guid lbd_nest $size $thin) - - notice "Creating nested lvol store on lvol bdev: $lb_name" - nest_ls_guid=$($rpc_py bdev_lvol_create_lvstore $lb_name lvs_n_$i -c 4194304) - nest_lvol_stores+=("$nest_ls_guid") - - for ((j = 0; j < vm_count; j++)); do - notice "Creating nested lvol bdev for VM $i on lvol store $nest_ls_guid" - free_mb=$(get_lvs_free_mb "$nest_ls_guid") - nest_size=$((free_mb / (vm_count - j))) - lb_name=$($rpc_py bdev_lvol_create -u $nest_ls_guid lbd_vm_$j $nest_size $thin) - nest_lvol_bdevs+=("$lb_name") - done - fi - - # Create base lvol bdevs for ((j = 0; j < vm_count; j++)); do - notice "Creating lvol bdev for VM $i on lvol store $ls_guid" - free_mb=$(get_lvs_free_mb "$ls_guid") - size=$((free_mb / (vm_count - j))) - lb_name=$($rpc_py bdev_lvol_create -u $ls_guid lbd_vm_$j $size $thin) - lvol_bdevs+=("$lb_name") + notice "Creating nested lvol bdev for VM $id on lvol store $nest_ls_guid" + free_mb=$(get_lvs_free_mb "$nest_ls_guid") + nest_size=$((free_mb / (vm_count - j))) + lb_name=$($rpc_py bdev_lvol_create -u $nest_ls_guid lbd_vm_$j $nest_size $thin) + nest_lvol_bdevs+=("$lb_name") done +fi + +# Create base lvol bdevs +for ((j = 0; j < vm_count; j++)); do + notice "Creating lvol bdev for VM $id on lvol store $ls_guid" + free_mb=$(get_lvs_free_mb "$ls_guid") + size=$((free_mb / (vm_count - j))) + lb_name=$($rpc_py bdev_lvol_create -u $ls_guid lbd_vm_$j $size $thin) + lvol_bdevs+=("$lb_name") done bdev_info=$($rpc_py bdev_get_bdevs) diff --git a/test/vhost/manual.sh b/test/vhost/manual.sh index 2b204ff6e..7f6792509 100755 --- a/test/vhost/manual.sh +++ b/test/vhost/manual.sh @@ -44,8 +44,6 @@ if [[ ! -r "${VM_IMAGE}" ]]; then exit 1 fi -DISKS_NUMBER=$(lspci -mm -n | grep 0108 | tr -d '"' | awk -F " " '{print "0000:"$1}' | wc -l) - WORKDIR=$(readlink -f $(dirname $0)) case $1 in diff --git a/test/vhost/vhost.sh b/test/vhost/vhost.sh index 25d815aeb..e53bb9887 100755 --- a/test/vhost/vhost.sh +++ b/test/vhost/vhost.sh @@ -23,8 +23,6 @@ if [[ ! -r "${VM_IMAGE}" ]]; then exit 1 fi -DISKS_NUMBER=$(lspci -mm -n | grep 0108 | tr -d '"' | awk -F " " '{print "0000:"$1}' | wc -l) - WORKDIR=$(readlink -f $(dirname $0)) run_test "vhost_negative" $WORKDIR/other/negative.sh @@ -50,31 +48,13 @@ if [ $RUN_NIGHTLY -eq 1 ]; then echo 'Running filesystem integrity suite with BLK...' run_test "vhost_blk_fs_integrity" $WORKDIR/integrity/integrity_start.sh --ctrl-type=spdk_vhost_blk --fs="xfs ntfs btrfs ext4" - if [[ $DISKS_NUMBER -ge 2 ]]; then - echo 'Running lvol integrity nightly suite with two cores and two controllers' - run_test "vhost_scsi_2core_2ctrl" $WORKDIR/lvol/lvol_test.sh --fio-bin=$FIO_BIN \ - --ctrl-type=spdk_vhost_scsi --max-disks=2 --distribute-cores --vm-count=2 + echo 'Running lvol integrity nightly suite with multiple cores and two vhost controllers (vhost_scsi)' + run_test "vhost_scsi_cores_2ctrl" $WORKDIR/lvol/lvol_test.sh --fio-bin=$FIO_BIN \ + --ctrl-type=spdk_vhost_scsi --distribute-cores --vm-count=2 - echo 'Running lvol integrity nightly suite with one core and two controllers' - run_test "vhost_scsi_1core_2ctrl" $WORKDIR/lvol/lvol_test.sh --fio-bin=$FIO_BIN \ - --ctrl-type=spdk_vhost_scsi --max-disks=2 --vm-count=2 - fi - echo 'Running lvol integrity nightly suite with one core and one controller' - run_test "vhost_scsi_1core_1ctrl" $WORKDIR/lvol/lvol_test.sh --fio-bin=$FIO_BIN \ - --ctrl-type=spdk_vhost_scsi --max-disks=1 - - if [[ $DISKS_NUMBER -ge 2 ]]; then - echo 'Running lvol integrity nightly suite with two cores and two controllers' - run_test "vhost_blk_2core_2ctrl" $WORKDIR/lvol/lvol_test.sh --fio-bin=$FIO_BIN \ - --ctrl-type=spdk_vhost_blk --max-disks=2 --distribute-cores --vm-count=2 - - echo 'Running lvol integrity nightly suite with one core and two controllers' - run_test "vhost_blk_1core_2ctrl" $WORKDIR/lvol/lvol_test.sh --fio-bin=$FIO_BIN \ - --ctrl-type=spdk_vhost_blk --max-disks=2 --vm-count=2 - fi - echo 'Running lvol integrity nightly suite with one core and one controller' - run_test "vhost_lvol_integrity_1core_1ctrl" $WORKDIR/lvol/lvol_test.sh --fio-bin=$FIO_BIN \ - --ctrl-type=spdk_vhost_blk --max-disks=1 + echo 'Running lvol integrity nightly suite with multiple cores and two vhost controllers (vhost_blk)' + run_test "vhost_blk_cores_2ctrl" $WORKDIR/lvol/lvol_test.sh --fio-bin=$FIO_BIN \ + --ctrl-type=spdk_vhost_blk --distribute-cores --vm-count=2 echo 'Running readonly tests suite...' run_test "vhost_readonly" $WORKDIR/readonly/readonly.sh --vm_image=$VM_IMAGE --disk=Nvme0n1 -x