Commit e0335bb1 authored by Michal Berger's avatar Michal Berger Committed by Tomasz Zawadzki
Browse files

test/vhost: Drop support for multiple nvme controllers



This test path is currently not being run through CI, hence no need
to support it.

Signed-off-by: default avatarMichal Berger <michalx.berger@intel.com>
Change-Id: I9a7cc7005c4e20f25724d2a0a417656deb09abb1
Reviewed-on: https://review.spdk.io/gerrit/c/spdk/spdk/+/6543


Reviewed-by: default avatarTomasz Zawadzki <tomasz.zawadzki@intel.com>
Reviewed-by: default avatarKarol Latecki <karol.latecki@intel.com>
Reviewed-by: default avatarJim Harris <james.r.harris@intel.com>
Community-CI: Mellanox Build Bot
Tested-by: default avatarSPDK CI Jenkins <sys_sgci@intel.com>
parent 3369ef45
Loading
Loading
Loading
Loading
+2 −5
Original line number Diff line number Diff line
@@ -50,8 +50,7 @@ iscsi_tgt_fuzz
nvmf_fuzz
nvmf_multiconnection
nvmf_initiator_timeout
vhost_blk_2core_2ctrl
vhost_blk_1core_2ctrl
vhost_blk_cores_2ctrl
vhost_blk_fs_integrity
vhost_blk_integrity
vhost_blk_nightly
@@ -63,6 +62,4 @@ vhost_readonly
vhost_scsi_fs_integrity
vhost_scsi_integrity
vhost_scsi_nightly
vhost_scsi_2core_2ctrl
vhost_scsi_1core_2ctrl
vhost_scsi_1core_1ctrl
vhost_scsi_cores_2ctrl
+28 −51
Original line number Diff line number Diff line
@@ -9,7 +9,6 @@ source $rootdir/scripts/common.sh
rpc_py="$rootdir/scripts/rpc.py -s $(get_vhost_dir 0)/rpc.sock"

vm_count=1
max_disks=""
ctrl_type="spdk_vhost_scsi"
use_fs=false
nested_lvol=false
@@ -28,8 +27,6 @@ function usage() {
	echo "    --vm-count=INT        Virtual machines to use in test;"
	echo "                          Each VM will get one lvol bdev on each NVMe."
	echo "                          Default: 1"
	echo "    --max-disks=INT       Maximum number of NVMe drives to use in test."
	echo "                          Default: will use all available NVMes."
	echo "    --ctrl-type=TYPE      Controller type to use for test:"
	echo "                          spdk_vhost_scsi - use spdk vhost scsi"
	echo "                          spdk_vhost_blk - use spdk vhost block"
@@ -67,10 +64,8 @@ function clean_lvol_cfg() {
	done

	notice "Removing lvol stores"
	for lvol_store in "${lvol_stores[@]}"; do
		$rpc_py bdev_lvol_delete_lvstore -u $lvol_store
		notice "lvol store $lvol_store removed"
	done
	$rpc_py bdev_lvol_delete_lvstore -u "$ls_guid"
	notice "lvol store $ls_guid removed"
}

while getopts 'xh-:' optchar; do
@@ -80,7 +75,6 @@ while getopts 'xh-:' optchar; do
				help) usage $0 ;;
				fio-bin=*) fio_bin="--fio-bin=${OPTARG#*=}" ;;
				vm-count=*) vm_count="${OPTARG#*=}" ;;
				max-disks=*) max_disks="${OPTARG#*=}" ;;
				ctrl-type=*) ctrl_type="${OPTARG#*=}" ;;
				nested-lvol) nested_lvol=true ;;
				distribute-cores) distribute_cores=true ;;
@@ -99,17 +93,6 @@ done

vhosttestinit

notice "Get NVMe disks:"
nvmes=($(get_nvme_bdfs))

if [[ -z $max_disks ]]; then
	max_disks=${#nvmes[@]}
fi

if ((${#nvmes[@]} < max_disks)); then
	fail "Number of NVMe drives (${#nvmes[@]}) is lower than number of requested disks for test ($max_disks)"
fi

if $distribute_cores; then
	# FIXME: this need to be handled entirely in common.sh
	source $testdir/autotest.config
@@ -125,20 +108,15 @@ notice "..."

trap 'clean_lvol_cfg; error_exit "${FUNCNAME}" "${LINENO}"' SIGTERM SIGABRT ERR

lvol_stores=()
lvol_bdevs=()
nest_lvol_stores=()
nest_lvol_bdevs=()
used_vms=""

# On each NVMe create one lvol store
for ((i = 0; i < max_disks; i++)); do

id=0
# Create base lvol store on NVMe
	notice "Creating lvol store on device Nvme${i}n1"
	ls_guid=$($rpc_py bdev_lvol_create_lvstore Nvme${i}n1 lvs_$i -c 4194304)
	lvol_stores+=("$ls_guid")

notice "Creating lvol store on device Nvme${id}n1"
ls_guid=$($rpc_py bdev_lvol_create_lvstore Nvme0n1 lvs_$id -c 4194304)
if $nested_lvol; then
	free_mb=$(get_lvs_free_mb "$ls_guid")
	size=$((free_mb / (vm_count + 1)))
@@ -147,11 +125,11 @@ for ((i = 0; i < max_disks; i++)); do
	lb_name=$($rpc_py bdev_lvol_create -u $ls_guid lbd_nest $size $thin)

	notice "Creating nested lvol store on lvol bdev: $lb_name"
		nest_ls_guid=$($rpc_py bdev_lvol_create_lvstore $lb_name lvs_n_$i -c 4194304)
	nest_ls_guid=$($rpc_py bdev_lvol_create_lvstore $lb_name lvs_n_$id -c 4194304)
	nest_lvol_stores+=("$nest_ls_guid")

	for ((j = 0; j < vm_count; j++)); do
			notice "Creating nested lvol bdev for VM $i on lvol store $nest_ls_guid"
		notice "Creating nested lvol bdev for VM $id on lvol store $nest_ls_guid"
		free_mb=$(get_lvs_free_mb "$nest_ls_guid")
		nest_size=$((free_mb / (vm_count - j)))
		lb_name=$($rpc_py bdev_lvol_create -u $nest_ls_guid lbd_vm_$j $nest_size $thin)
@@ -161,13 +139,12 @@ for ((i = 0; i < max_disks; i++)); do

# Create base lvol bdevs
for ((j = 0; j < vm_count; j++)); do
		notice "Creating lvol bdev for VM $i on lvol store $ls_guid"
	notice "Creating lvol bdev for VM $id on lvol store $ls_guid"
	free_mb=$(get_lvs_free_mb "$ls_guid")
	size=$((free_mb / (vm_count - j)))
	lb_name=$($rpc_py bdev_lvol_create -u $ls_guid lbd_vm_$j $size $thin)
	lvol_bdevs+=("$lb_name")
done
done

bdev_info=$($rpc_py bdev_get_bdevs)
notice "Configuration after initial set-up:"
+0 −2
Original line number Diff line number Diff line
@@ -44,8 +44,6 @@ if [[ ! -r "${VM_IMAGE}" ]]; then
	exit 1
fi

DISKS_NUMBER=$(lspci -mm -n | grep 0108 | tr -d '"' | awk -F " " '{print "0000:"$1}' | wc -l)

WORKDIR=$(readlink -f $(dirname $0))

case $1 in
+7 −27
Original line number Diff line number Diff line
@@ -23,8 +23,6 @@ if [[ ! -r "${VM_IMAGE}" ]]; then
	exit 1
fi

DISKS_NUMBER=$(lspci -mm -n | grep 0108 | tr -d '"' | awk -F " " '{print "0000:"$1}' | wc -l)

WORKDIR=$(readlink -f $(dirname $0))

run_test "vhost_negative" $WORKDIR/other/negative.sh
@@ -50,31 +48,13 @@ if [ $RUN_NIGHTLY -eq 1 ]; then
	echo 'Running filesystem integrity suite with BLK...'
	run_test "vhost_blk_fs_integrity" $WORKDIR/integrity/integrity_start.sh --ctrl-type=spdk_vhost_blk --fs="xfs ntfs btrfs ext4"

	if [[ $DISKS_NUMBER -ge 2 ]]; then
		echo 'Running lvol integrity nightly suite with two cores and two controllers'
		run_test "vhost_scsi_2core_2ctrl" $WORKDIR/lvol/lvol_test.sh --fio-bin=$FIO_BIN \
			--ctrl-type=spdk_vhost_scsi --max-disks=2 --distribute-cores --vm-count=2
	echo 'Running lvol integrity nightly suite with multiple cores and two vhost controllers (vhost_scsi)'
	run_test "vhost_scsi_cores_2ctrl" $WORKDIR/lvol/lvol_test.sh --fio-bin=$FIO_BIN \
		--ctrl-type=spdk_vhost_scsi --distribute-cores --vm-count=2

		echo 'Running lvol integrity nightly suite with one core and two controllers'
		run_test "vhost_scsi_1core_2ctrl" $WORKDIR/lvol/lvol_test.sh --fio-bin=$FIO_BIN \
			--ctrl-type=spdk_vhost_scsi --max-disks=2 --vm-count=2
	fi
	echo 'Running lvol integrity nightly suite with one core and one controller'
	run_test "vhost_scsi_1core_1ctrl" $WORKDIR/lvol/lvol_test.sh --fio-bin=$FIO_BIN \
		--ctrl-type=spdk_vhost_scsi --max-disks=1

	if [[ $DISKS_NUMBER -ge 2 ]]; then
		echo 'Running lvol integrity nightly suite with two cores and two controllers'
		run_test "vhost_blk_2core_2ctrl" $WORKDIR/lvol/lvol_test.sh --fio-bin=$FIO_BIN \
			--ctrl-type=spdk_vhost_blk --max-disks=2 --distribute-cores --vm-count=2

		echo 'Running lvol integrity nightly suite with one core and two controllers'
		run_test "vhost_blk_1core_2ctrl" $WORKDIR/lvol/lvol_test.sh --fio-bin=$FIO_BIN \
			--ctrl-type=spdk_vhost_blk --max-disks=2 --vm-count=2
	fi
	echo 'Running lvol integrity nightly suite with one core and one controller'
	run_test "vhost_lvol_integrity_1core_1ctrl" $WORKDIR/lvol/lvol_test.sh --fio-bin=$FIO_BIN \
		--ctrl-type=spdk_vhost_blk --max-disks=1
	echo 'Running lvol integrity nightly suite with multiple cores and two vhost controllers (vhost_blk)'
	run_test "vhost_blk_cores_2ctrl" $WORKDIR/lvol/lvol_test.sh --fio-bin=$FIO_BIN \
		--ctrl-type=spdk_vhost_blk --distribute-cores --vm-count=2

	echo 'Running readonly tests suite...'
	run_test "vhost_readonly" $WORKDIR/readonly/readonly.sh --vm_image=$VM_IMAGE --disk=Nvme0n1 -x