From 9a4a87b5738969377ccc61af0d88b46c82f240f9 Mon Sep 17 00:00:00 2001 From: Maciej Wawryk Date: Tue, 5 Nov 2019 10:28:35 +0100 Subject: [PATCH] test: Shellcheck - correct rule: $/${} is unnecessary on arithmetic variables. Correct shellcheck rule SC2004: $/${} is unnecessary on arithmetic variables. Signed-off-by: Maciej Wawryk Change-Id: Ibf2879360bc50cc058b0f4434a5777c53c0eeffb Reviewed-on: https://review.gerrithub.io/c/spdk/spdk/+/473265 Tested-by: SPDK CI Jenkins Reviewed-by: Ben Walker Reviewed-by: Jim Harris Reviewed-by: Shuhei Matsumoto --- scripts/check_format.sh | 2 +- scripts/prep_benchmarks.sh | 2 +- scripts/setup.sh | 4 +-- test/ftl/common.sh | 4 +-- test/ftl/dirty_shutdown.sh | 4 +-- test/ftl/restore.sh | 2 +- test/iscsi_tgt/common.sh | 6 ++-- test/iscsi_tgt/lvol/iscsi_lvol.sh | 2 +- .../multiconnection/multiconnection.sh | 2 +- test/iscsi_tgt/perf/iscsi_target.sh | 2 +- test/iscsi_tgt/qos/qos.sh | 20 +++++------ test/iscsi_tgt/trace_record/trace_record.sh | 2 +- test/json_config/json_config.sh | 2 +- test/nvme/hw_hotplug.sh | 2 +- test/nvme/perf/common.sh | 36 +++++++++---------- test/nvme/perf/run_perf.sh | 24 ++++++------- test/nvmf/common.sh | 4 +-- test/vhost/common.sh | 2 +- test/vhost/lvol/lvol_test.sh | 12 +++---- test/vhost/perf_bench/vhost_perf.sh | 14 ++++---- 20 files changed, 74 insertions(+), 74 deletions(-) diff --git a/scripts/check_format.sh b/scripts/check_format.sh index 95bb6f9e3..55de05f36 100755 --- a/scripts/check_format.sh +++ b/scripts/check_format.sh @@ -240,7 +240,7 @@ if hash shellcheck 2>/dev/null; then # go to: https://trello.com/c/29Z90j1W # Error descriptions can also be found at: https://github.com/koalaman/shellcheck/wiki # This SHCK_EXCLUDE list is out "to do" and we work to fix all of this errors. - SHCK_EXCLUDE="SC1083,SC2002,SC2004,\ + SHCK_EXCLUDE="SC1083,SC2002,\ SC2010,SC2012,SC2016,SC2034,SC2045,SC2046,SC2068,SC2086,SC2089,SC2090,\ SC2097,SC2098,SC2119,SC2120,SC2121,SC2124,SC2126,SC2128,\ SC2129,SC2140,SC2142,SC2143,SC2154,SC2155,SC2162" diff --git a/scripts/prep_benchmarks.sh b/scripts/prep_benchmarks.sh index 1bb91385e..2b0642b07 100755 --- a/scripts/prep_benchmarks.sh +++ b/scripts/prep_benchmarks.sh @@ -20,7 +20,7 @@ function configure_performance() { echo -n "Moving all interrupts off of core 0..." count=$(($(nproc) / 4)) cpumask="e" - for ((i=1; i<$count; i++)); do + for ((i=1; i /dev/null && break - counter=$(( $counter - 1 )) + counter=$(( counter - 1 )) sleep 0.5 done xtrace_restore @@ -178,8 +178,8 @@ function start_vpp() { sleep 3 # SC1010: ping -M do - in this case do is an option not bash special word # shellcheck disable=SC1010 - ping -c 1 $TARGET_IP -s $(( $MTU - 28 )) -M do - vppctl ping $INITIATOR_IP repeat 1 size $(( $MTU - (28 + 8) )) verbose + ping -c 1 $TARGET_IP -s $(( MTU - 28 )) -M do + vppctl ping $INITIATOR_IP repeat 1 size $(( MTU - (28 + 8) )) verbose } function kill_vpp() { diff --git a/test/iscsi_tgt/lvol/iscsi_lvol.sh b/test/iscsi_tgt/lvol/iscsi_lvol.sh index fcf36c1ff..a54c40755 100755 --- a/test/iscsi_tgt/lvol/iscsi_lvol.sh +++ b/test/iscsi_tgt/lvol/iscsi_lvol.sh @@ -71,7 +71,7 @@ sleep 1 timing_enter discovery iscsiadm -m discovery -t sendtargets -p $TARGET_IP:$ISCSI_PORT iscsiadm -m node --login -p $TARGET_IP:$ISCSI_PORT -waitforiscsidevices $(( $NUM_LVS * $NUM_LVOL )) +waitforiscsidevices $(( NUM_LVS * NUM_LVOL )) timing_exit discovery timing_enter fio diff --git a/test/iscsi_tgt/multiconnection/multiconnection.sh b/test/iscsi_tgt/multiconnection/multiconnection.sh index 1fccbd0ca..0eb09b651 100755 --- a/test/iscsi_tgt/multiconnection/multiconnection.sh +++ b/test/iscsi_tgt/multiconnection/multiconnection.sh @@ -56,7 +56,7 @@ ls_guid=$($rpc_py bdev_lvol_create_lvstore "Nvme0n1" "lvs0" -c 1048576) # Assign even size for each lvol_bdev. get_lvs_free_mb $ls_guid -lvol_bdev_size=$(($free_mb / $CONNECTION_NUMBER)) +lvol_bdev_size=$((free_mb / CONNECTION_NUMBER)) for i in $(seq 1 $CONNECTION_NUMBER); do $rpc_py bdev_lvol_create -u $ls_guid lbd_$i $lvol_bdev_size done diff --git a/test/iscsi_tgt/perf/iscsi_target.sh b/test/iscsi_tgt/perf/iscsi_target.sh index 69158d3ac..4c122f96d 100755 --- a/test/iscsi_tgt/perf/iscsi_target.sh +++ b/test/iscsi_tgt/perf/iscsi_target.sh @@ -100,7 +100,7 @@ fi $rpc_py iscsi_create_portal_group $PORTAL_TAG $TARGET_IP:$ISCSI_PORT $rpc_py iscsi_create_initiator_group $INITIATOR_TAG $INITIATOR_NAME $NETMASK -for (( i=0; i < $DISKNO; i++ )) +for (( i=0; i < DISKNO; i++ )) do $rpc_py iscsi_create_target_node Target${i} Target${i}_alias "${bdevs[i]}:0" "$PORTAL_TAG:$INITIATOR_TAG" 64 -d done diff --git a/test/iscsi_tgt/qos/qos.sh b/test/iscsi_tgt/qos/qos.sh index 6da04cb9e..fd5a24bb0 100755 --- a/test/iscsi_tgt/qos/qos.sh +++ b/test/iscsi_tgt/qos/qos.sh @@ -23,8 +23,8 @@ function run_fio() { local end_io_count=$(jq -r '.bdevs[0].num_read_ops' <<< "$iostats") local end_bytes_read=$(jq -r '.bdevs[0].bytes_read' <<< "$iostats") - IOPS_RESULT=$(((end_io_count-start_io_count)/$run_time)) - BANDWIDTH_RESULT=$(((end_bytes_read-start_bytes_read)/$run_time)) + IOPS_RESULT=$(((end_io_count-start_io_count)/run_time)) + BANDWIDTH_RESULT=$(((end_bytes_read-start_bytes_read)/run_time)) } function verify_qos_limits() { @@ -84,19 +84,19 @@ trap 'iscsicleanup; killprocess $pid; iscsitestfini $1 $2; exit 1' SIGINT SIGTER run_fio Malloc0 # Set IOPS/bandwidth limit to 50% of the actual unrestrained performance. -IOPS_LIMIT=$(($IOPS_RESULT/2)) -BANDWIDTH_LIMIT=$(($BANDWIDTH_RESULT/2)) +IOPS_LIMIT=$((IOPS_RESULT/2)) +BANDWIDTH_LIMIT=$((BANDWIDTH_RESULT/2)) # Set READ bandwidth limit to 50% of the RW bandwidth limit to be able # to differentiate those two. -READ_BANDWIDTH_LIMIT=$(($BANDWIDTH_LIMIT/2)) +READ_BANDWIDTH_LIMIT=$((BANDWIDTH_LIMIT/2)) # Also round them down to nearest multiple of either 1000 IOPS or 1MB BW # which are the minimal QoS granularities -IOPS_LIMIT=$(($IOPS_LIMIT/1000*1000)) -BANDWIDTH_LIMIT_MB=$(($BANDWIDTH_LIMIT/1024/1024)) -BANDWIDTH_LIMIT=$(($BANDWIDTH_LIMIT_MB*1024*1024)) -READ_BANDWIDTH_LIMIT_MB=$(($READ_BANDWIDTH_LIMIT/1024/1024)) -READ_BANDWIDTH_LIMIT=$(($READ_BANDWIDTH_LIMIT_MB*1024*1024)) +IOPS_LIMIT=$((IOPS_LIMIT/1000*1000)) +BANDWIDTH_LIMIT_MB=$((BANDWIDTH_LIMIT/1024/1024)) +BANDWIDTH_LIMIT=$((BANDWIDTH_LIMIT_MB*1024*1024)) +READ_BANDWIDTH_LIMIT_MB=$((READ_BANDWIDTH_LIMIT/1024/1024)) +READ_BANDWIDTH_LIMIT=$((READ_BANDWIDTH_LIMIT_MB*1024*1024)) # Limit the I/O rate by RPC, then confirm the observed rate matches. $rpc_py bdev_set_qos_limit Malloc0 --rw_ios_per_sec $IOPS_LIMIT diff --git a/test/iscsi_tgt/trace_record/trace_record.sh b/test/iscsi_tgt/trace_record/trace_record.sh index 5e6f231db..75dd19d97 100755 --- a/test/iscsi_tgt/trace_record/trace_record.sh +++ b/test/iscsi_tgt/trace_record/trace_record.sh @@ -73,7 +73,7 @@ sleep 1 iscsiadm -m discovery -t sendtargets -p $TARGET_IP:$ISCSI_PORT iscsiadm -m node --login -p $TARGET_IP:$ISCSI_PORT -waitforiscsidevices $(( $CONNECTION_NUMBER + 1 )) +waitforiscsidevices $(( CONNECTION_NUMBER + 1 )) trap 'iscsicleanup; killprocess $iscsi_pid; killprocess $record_pid; delete_tmp_files; iscsitestfini $1 $2; exit 1' SIGINT SIGTERM EXIT diff --git a/test/json_config/json_config.sh b/test/json_config/json_config.sh index 6c12cd57c..29881adb1 100755 --- a/test/json_config/json_config.sh +++ b/test/json_config/json_config.sh @@ -415,7 +415,7 @@ function json_config_clear() { count=100 while [ $count -gt 0 ] ; do $rootdir/scripts/rpc.py -s "${app_socket[$1]}" save_config | $config_filter -method delete_global_parameters | $config_filter -method check_empty && break - count=$(( $count -1 )) + count=$(( count -1 )) sleep 0.1 done diff --git a/test/nvme/hw_hotplug.sh b/test/nvme/hw_hotplug.sh index 6b99fa2e3..276e61ad4 100755 --- a/test/nvme/hw_hotplug.sh +++ b/test/nvme/hw_hotplug.sh @@ -50,7 +50,7 @@ trap 'killprocess $example_pid; exit 1' SIGINT SIGTERM EXIT i=0 while ! grep "Starting I/O" log.txt; do [ $i -lt 20 ] || break - i=$(($i+1)) + i=$((i+1)) sleep 1 done diff --git a/test/nvme/perf/common.sh b/test/nvme/perf/common.sh index 7dd102557..bf72b289f 100755 --- a/test/nvme/perf/common.sh +++ b/test/nvme/perf/common.sh @@ -107,7 +107,7 @@ function get_disks_on_numa(){ for (( i=0; i<${#devs[@]}; i++ )) do if [ ${numas[$i]} = $numa_no ]; then - disks_on_numa=$(($disks_on_numa+1)) + disks_on_numa=$((disks_on_numa+1)) fi done echo $disks_on_numa @@ -124,18 +124,18 @@ function create_fio_config(){ local filename="" local cores_numa=($(get_cores_numa_node "$5")) - local disks_per_core=$(($disk_no/$no_cores)) - local disks_per_core_mod=$(($disk_no%$no_cores)) + local disks_per_core=$((disk_no/no_cores)) + local disks_per_core_mod=$((disk_no%no_cores)) # For kernel dirver, each disk will be alligned with all cpus on the same NUMA node if [ "$plugin" != "nvme" ] && [ "$plugin" != "bdev" ]; then - for (( i=0; i<$disk_no; i++ )) + for (( i=0; i> $BASE_DIR/config.fio done else - for (( i=0; i<$no_cores; i++ )) + for (( i=0; i> $result_file echo "num_of_disks,iops,avg_lat[usec],p99[usec],p99.99[usec],stdev[usec],avg_slat[usec],avg_clat[usec],bw[Kib/s]" >> $result_file #Run each workolad $REPEAT_NO times -for (( j=0; j < $REPEAT_NO; j++ )) +for (( j=0; j < REPEAT_NO; j++ )) do #Start with $DISKNO disks and remove 2 disks for each run to avoid preconditioning before each run. - for (( k=$DISKNO; k >= 1; k-=2 )) + for (( k=DISKNO; k >= 1; k-=2 )) do cp $BASE_DIR/config.fio.tmp $BASE_DIR/config.fio echo "" >> $BASE_DIR/config.fio #The SPDK fio plugin supports submitting/completing I/Os to multiple SSDs from a single thread. #Therefore, the per thread queue depth is set to the desired IODEPTH/device X the number of devices per thread. if [ "$PLUGIN" = "nvme" ] || [ "$PLUGIN" = "bdev" ] && [ "$NOIOSCALING" = false ]; then - qd=$(( $IODEPTH * $k )) + qd=$(( IODEPTH * k )) else qd=$IODEPTH fi @@ -161,17 +161,17 @@ do done done #Write results to csv file -for (( k=$DISKNO; k >= 1; k-=2 )) +for (( k=DISKNO; k >= 1; k-=2 )) do - iops_disks[$k]=$((${iops_disks[$k]} / $REPEAT_NO)) + iops_disks[$k]=$((${iops_disks[$k]} / REPEAT_NO)) if [ $PLUGIN != "bdevperf" ]; then - mean_lat_disks_usec[$k]=$((${mean_lat_disks_usec[$k]} / $REPEAT_NO)) - p99_lat_disks_usec[$k]=$((${p99_lat_disks_usec[$k]} / $REPEAT_NO)) - p99_99_lat_disks_usec[$k]=$((${p99_99_lat_disks_usec[$k]} / $REPEAT_NO)) - stdev_disks_usec[$k]=$((${stdev_disks_usec[$k]} / $REPEAT_NO)) - mean_slat_disks_usec[$k]=$((${mean_slat_disks_usec[$k]} / $REPEAT_NO)) - mean_clat_disks_usec[$k]=$((${mean_clat_disks_usec[$k]} / $REPEAT_NO)) + mean_lat_disks_usec[$k]=$((${mean_lat_disks_usec[$k]} / REPEAT_NO)) + p99_lat_disks_usec[$k]=$((${p99_lat_disks_usec[$k]} / REPEAT_NO)) + p99_99_lat_disks_usec[$k]=$((${p99_99_lat_disks_usec[$k]} / REPEAT_NO)) + stdev_disks_usec[$k]=$((${stdev_disks_usec[$k]} / REPEAT_NO)) + mean_slat_disks_usec[$k]=$((${mean_slat_disks_usec[$k]} / REPEAT_NO)) + mean_clat_disks_usec[$k]=$((${mean_clat_disks_usec[$k]} / REPEAT_NO)) else mean_lat_disks_usec[$k]=0 p99_lat_disks_usec[$k]=0 @@ -181,7 +181,7 @@ do mean_clat_disks_usec[$k]=0 fi - bw[$k]=$((${bw[$k]} / $REPEAT_NO)) + bw[$k]=$((${bw[$k]} / REPEAT_NO)) printf "%s,%s,%s,%s,%s,%s,%s,%s,%s\n" ${k} ${iops_disks[$k]} ${mean_lat_disks_usec[$k]} ${p99_lat_disks_usec[$k]}\ ${p99_99_lat_disks_usec[$k]} ${stdev_disks_usec[$k]} ${mean_slat_disks_usec[$k]} ${mean_clat_disks_usec[$k]} ${bw[$k]} >> $result_file diff --git a/test/nvmf/common.sh b/test/nvmf/common.sh index 19079cd2b..49a09d180 100644 --- a/test/nvmf/common.sh +++ b/test/nvmf/common.sh @@ -109,13 +109,13 @@ function detect_rdma_nics() function allocate_nic_ips() { - (( count=$NVMF_IP_LEAST_ADDR )) + (( count=NVMF_IP_LEAST_ADDR )) for nic_name in $(get_rdma_if_list); do ip="$(get_ip_address $nic_name)" if [ -z $ip ]; then ip addr add $NVMF_IP_PREFIX.$count/24 dev $nic_name ip link set $nic_name up - (( count=$count+1 )) + (( count=count+1 )) fi # dump configuration for debug log ip addr show $nic_name diff --git a/test/vhost/common.sh b/test/vhost/common.sh index aa5051ce2..2ba1c1890 100644 --- a/test/vhost/common.sh +++ b/test/vhost/common.sh @@ -655,7 +655,7 @@ function vm_setup() for c in $cpu_list; do # if range is detected - count how many cpus if [[ $c =~ [0-9]+-[0-9]+ ]]; then - val=$(($c-1)) + val=$((c-1)) val=${val#-} else val=1 diff --git a/test/vhost/lvol/lvol_test.sh b/test/vhost/lvol/lvol_test.sh index e72fca016..6e69a2044 100755 --- a/test/vhost/lvol/lvol_test.sh +++ b/test/vhost/lvol/lvol_test.sh @@ -132,7 +132,7 @@ nest_lvol_bdevs=() used_vms="" # On each NVMe create one lvol store -for (( i=0; i<$max_disks; i++ ));do +for (( i=0; i