Using new_id attribute is global in scope, meaning that depending on the kernel's setup seen prior running setup.sh, single write to it may re-bind ALL matching devices. This doesn't play well with our PCI_{ALLOWED,BLOCKED} options as they can't be enforced in such a case. Consider the following example: > modprobe -r nvme # all nvme ctrls are detached from the kernel > echo 0xdead 0xbeef >/sys/bus/pci/drivers/uio_pci_generic/new_id # setup.sh-wise > modprobe -r nvme > PCI_BLOCKED=some:dead:beef.bdf setup.sh # PCI_BLOCKED device still ends up bound to userspace driver. After this single write, ALL matching devices will end up bound to uio_pci_generic. To avoid this, we should override preferred driver on per-bdf basis. Signed-off-by: Michal Berger <michal.berger@intel.com> Change-Id: Ic4613e33321303b92b47ce3f4d7e1f29ecca3036 Reviewed-on: https://review.spdk.io/gerrit/c/spdk/spdk/+/13813 Tested-by: SPDK CI Jenkins <sys_sgci@intel.com> Reviewed-by: Jim Harris <james.r.harris@intel.com> Reviewed-by: Kamil Godzwon <kamilx.godzwon@intel.com> Reviewed-by: Konrad Sztyber <konrad.sztyber@intel.com>
148 lines
3.6 KiB
Bash
Executable File
148 lines
3.6 KiB
Bash
Executable File
#!/usr/bin/env bash
|
|
# SPDX-License-Identifier: BSD-3-Clause
|
|
# Copyright (C) 2022 Intel Corporation
|
|
# All rights reserved.
|
|
#
|
|
testdir=$(readlink -f $(dirname $0))
|
|
rootdir=$(readlink -f $testdir/../..)
|
|
source $rootdir/scripts/common.sh
|
|
source $rootdir/test/common/autotest_common.sh
|
|
|
|
# Pci bus hotplug
|
|
# Helper function to remove/attach cotrollers
|
|
remove_attach_helper() {
|
|
local hotplug_events=$1
|
|
local hotplug_wait=$2
|
|
local use_bdev=$3
|
|
local dev
|
|
|
|
# We need to make sure we wait long enough for hotplug to initialize the devices
|
|
# and start IO - if we start removing devices before that happens we will end up
|
|
# stepping on hotplug's toes forcing it to fail to report proper count of given
|
|
# events.
|
|
sleep "$hotplug_wait"
|
|
|
|
while ((hotplug_events--)); do
|
|
for dev in "${nvmes[@]}"; do
|
|
echo 1 > "/sys/bus/pci/devices/$dev/remove"
|
|
if $use_bdev; then
|
|
sleep "$hotplug_wait"
|
|
rpc_cmd bdev_get_bdevs \
|
|
| jq '.[] | .driver_specific | .nvme | .[] | .pci_address' \
|
|
| NOT grep $dev
|
|
fi
|
|
done
|
|
|
|
echo 1 > "/sys/bus/pci/rescan"
|
|
|
|
if $use_bdev; then
|
|
local retries_left=10
|
|
|
|
"$rootdir/scripts/setup.sh"
|
|
|
|
while [ "$(rpc_cmd bdev_get_bdevs \
|
|
| jq '.[] | .driver_specific | .nvme | .[] | .pci_address' \
|
|
| wc -l)" -lt "$nvme_count" ]; do
|
|
[ $retries_left -gt 0 ] || break
|
|
retries_left=$((retries_left - 1))
|
|
sleep 1
|
|
done
|
|
|
|
# Bdev name has already changed, search for the bdev's PCI addresses
|
|
for dev in "${nvmes[@]}"; do
|
|
rpc_cmd bdev_get_bdevs \
|
|
| jq '.[] | .driver_specific | .nvme | .[] | .pci_address' \
|
|
| grep $dev
|
|
done
|
|
fi
|
|
|
|
# Avoid setup.sh as it does some extra work which is not relevant for this test.
|
|
for dev in "${nvmes[@]}"; do
|
|
echo "${pci_bus_driver["$dev"]}" > "/sys/bus/pci/devices/$dev/driver_override"
|
|
echo "$dev" > "/sys/bus/pci/devices/$dev/driver/unbind"
|
|
echo "$dev" > "/sys/bus/pci/drivers_probe"
|
|
echo "" > "/sys/bus/pci/devices/$dev/driver_override"
|
|
done
|
|
|
|
# Wait now for hotplug to reattach to the devices
|
|
sleep "$hotplug_wait"
|
|
done
|
|
}
|
|
|
|
run_hotplug() {
|
|
trap 'killprocess $hotplug_pid; exit 1' SIGINT SIGTERM EXIT
|
|
|
|
"$SPDK_EXAMPLE_DIR/hotplug" \
|
|
-i 0 \
|
|
-t $((hotplug_events * hotplug_wait + hotplug_wait * 3)) \
|
|
-n $((hotplug_events * nvme_count)) \
|
|
-r $((hotplug_events * nvme_count)) \
|
|
-l warning &
|
|
hotplug_pid=$!
|
|
|
|
remove_attach_helper "$hotplug_events" "$hotplug_wait" false
|
|
|
|
# Wait in case hotplug app is lagging behind
|
|
# and kill it, if it hung.
|
|
sleep $hotplug_wait
|
|
|
|
if ! kill -0 "$hotplug_pid"; then
|
|
# hotplug already finished, check for the error code.
|
|
wait "$hotplug_pid"
|
|
else
|
|
echo "Killing hotplug application"
|
|
killprocess $hotplug_pid
|
|
return 1
|
|
fi
|
|
|
|
trap - SIGINT SIGTERM EXIT
|
|
}
|
|
|
|
# SPDK target hotplug
|
|
tgt_run_hotplug() {
|
|
local i=0
|
|
local dev
|
|
|
|
$SPDK_BIN_DIR/spdk_tgt &
|
|
spdk_tgt_pid=$!
|
|
|
|
trap 'killprocess ${spdk_tgt_pid}; echo 1 > /sys/bus/pci/rescan; exit 1' SIGINT SIGTERM EXIT
|
|
waitforlisten $spdk_tgt_pid
|
|
|
|
for dev in "${nvmes[@]}"; do
|
|
rpc_cmd bdev_nvme_attach_controller -b "Nvme0$i" -t PCIe -a $dev
|
|
waitforbdev "Nvme0${i}n1" "$hotplug_wait"
|
|
((i = i + 1))
|
|
done
|
|
|
|
rpc_cmd bdev_nvme_set_hotplug -e
|
|
|
|
remove_attach_helper "$hotplug_events" "$hotplug_wait" true
|
|
# Verify reregistering hotplug poller
|
|
rpc_cmd bdev_nvme_set_hotplug -d
|
|
rpc_cmd bdev_nvme_set_hotplug -e
|
|
|
|
remove_attach_helper "$hotplug_events" "$hotplug_wait" true
|
|
|
|
trap - SIGINT SIGTERM EXIT
|
|
killprocess $spdk_tgt_pid
|
|
}
|
|
|
|
# Preparation
|
|
"$rootdir/scripts/setup.sh"
|
|
|
|
hotplug_wait=6
|
|
hotplug_events=3
|
|
nvmes=($(nvme_in_userspace))
|
|
nvme_count=${#nvmes[@]}
|
|
|
|
xtrace_disable
|
|
cache_pci_bus_sysfs
|
|
xtrace_restore
|
|
|
|
# Run pci bus hotplug test
|
|
run_hotplug
|
|
|
|
# Run SPDK target based hotplug
|
|
tgt_run_hotplug
|