summaryrefslogtreecommitdiffstats
path: root/src/spdk/test/vhost/common/common.sh
diff options
context:
space:
mode:
authorDaniel Baumann <daniel.baumann@progress-linux.org>2024-04-27 18:24:20 +0000
committerDaniel Baumann <daniel.baumann@progress-linux.org>2024-04-27 18:24:20 +0000
commit483eb2f56657e8e7f419ab1a4fab8dce9ade8609 (patch)
treee5d88d25d870d5dedacb6bbdbe2a966086a0a5cf /src/spdk/test/vhost/common/common.sh
parentInitial commit. (diff)
downloadceph-483eb2f56657e8e7f419ab1a4fab8dce9ade8609.tar.xz
ceph-483eb2f56657e8e7f419ab1a4fab8dce9ade8609.zip
Adding upstream version 14.2.21.upstream/14.2.21upstream
Signed-off-by: Daniel Baumann <daniel.baumann@progress-linux.org>
Diffstat (limited to 'src/spdk/test/vhost/common/common.sh')
-rw-r--r--src/spdk/test/vhost/common/common.sh1109
1 files changed, 1109 insertions, 0 deletions
diff --git a/src/spdk/test/vhost/common/common.sh b/src/spdk/test/vhost/common/common.sh
new file mode 100644
index 00000000..19c4be62
--- /dev/null
+++ b/src/spdk/test/vhost/common/common.sh
@@ -0,0 +1,1109 @@
+set -e
+
+: ${SPDK_VHOST_VERBOSE=false}
+: ${QEMU_PREFIX="/usr/local/qemu/spdk-2.12"}
+
+BASE_DIR=$(readlink -f $(dirname ${BASH_SOURCE[0]}))
+
+# Default running dir -> spdk/..
+[[ -z "$TEST_DIR" ]] && TEST_DIR=$BASE_DIR/../../../../
+
+TEST_DIR="$(mkdir -p $TEST_DIR && cd $TEST_DIR && echo $PWD)"
+SPDK_BUILD_DIR=$BASE_DIR/../../../
+
+SPDK_VHOST_SCSI_TEST_DIR=$TEST_DIR/vhost
+
+function message()
+{
+ if ! $SPDK_VHOST_VERBOSE; then
+ local verbose_out=""
+ elif [[ ${FUNCNAME[2]} == "source" ]]; then
+ local verbose_out=" (file $(basename ${BASH_SOURCE[1]}):${BASH_LINENO[1]})"
+ else
+ local verbose_out=" (function ${FUNCNAME[2]}:${BASH_LINENO[1]})"
+ fi
+
+ local msg_type="$1"
+ shift
+ echo -e "${msg_type}${verbose_out}: $@"
+}
+
+function fail()
+{
+ echo "===========" >&2
+ message "FAIL" "$@" >&2
+ echo "===========" >&2
+ exit 1
+}
+
+function error()
+{
+ echo "===========" >&2
+ message "ERROR" "$@" >&2
+ echo "===========" >&2
+ # Don't 'return 1' since the stack trace will be incomplete (why?) missing upper command.
+ false
+}
+
+function warning()
+{
+ message "WARN" "$@" >&2
+}
+
+function notice()
+{
+ message "INFO" "$@"
+}
+
+
+# SSH key file
+: ${SPDK_VHOST_SSH_KEY_FILE="$(readlink -e $HOME/.ssh/spdk_vhost_id_rsa)"}
+if [[ ! -r "$SPDK_VHOST_SSH_KEY_FILE" ]]; then
+ error "Could not find SSH key file $SPDK_VHOST_SSH_KEY_FILE"
+ exit 1
+fi
+echo "Using SSH key file $SPDK_VHOST_SSH_KEY_FILE"
+
+VM_BASE_DIR="$TEST_DIR/vms"
+
+
+mkdir -p $TEST_DIR
+
+#
+# Source config describing QEMU and VHOST cores and NUMA
+#
+source $(readlink -f $(dirname ${BASH_SOURCE[0]}))/autotest.config
+
+# Trace flag is optional, if it wasn't set earlier - disable it after sourcing
+# autotest_common.sh
+if [[ $- =~ x ]]; then
+ source $SPDK_BUILD_DIR/test/common/autotest_common.sh
+else
+ source $SPDK_BUILD_DIR/test/common/autotest_common.sh
+ set +x
+fi
+
+function get_vhost_dir()
+{
+ if [[ ! -z "$1" ]]; then
+ assert_number "$1"
+ local vhost_num=$1
+ else
+ local vhost_num=0
+ fi
+
+ echo "$SPDK_VHOST_SCSI_TEST_DIR${vhost_num}"
+}
+
+function spdk_vhost_list_all()
+{
+ shopt -s nullglob
+ local vhost_list="$(echo $SPDK_VHOST_SCSI_TEST_DIR[0-9]*)"
+ shopt -u nullglob
+
+ if [[ ! -z "$vhost_list" ]]; then
+ vhost_list="$(basename --multiple $vhost_list)"
+ echo "${vhost_list//vhost/}"
+ fi
+}
+
+function spdk_vhost_run()
+{
+ local param
+ local vhost_num=0
+ local vhost_conf_path=""
+ local memory=1024
+
+ for param in "$@"; do
+ case $param in
+ --vhost-num=*)
+ vhost_num="${param#*=}"
+ assert_number "$vhost_num"
+ ;;
+ --conf-path=*) local vhost_conf_path="${param#*=}" ;;
+ --json-path=*) local vhost_json_path="${param#*=}" ;;
+ --memory=*) local memory=${param#*=} ;;
+ --no-pci*) local no_pci="-u" ;;
+ *)
+ error "Invalid parameter '$param'"
+ return 1
+ ;;
+ esac
+ done
+
+ local vhost_dir="$(get_vhost_dir $vhost_num)"
+ local vhost_app="$SPDK_BUILD_DIR/app/vhost/vhost"
+ local vhost_log_file="$vhost_dir/vhost.log"
+ local vhost_pid_file="$vhost_dir/vhost.pid"
+ local vhost_socket="$vhost_dir/usvhost"
+ local vhost_conf_template="$vhost_conf_path/vhost.conf.in"
+ local vhost_conf_file="$vhost_conf_path/vhost.conf"
+ notice "starting vhost app in background"
+ [[ -r "$vhost_pid_file" ]] && spdk_vhost_kill $vhost_num
+ [[ -d $vhost_dir ]] && rm -f $vhost_dir/*
+ mkdir -p $vhost_dir
+
+ if [[ ! -x $vhost_app ]]; then
+ error "application not found: $vhost_app"
+ return 1
+ fi
+
+ local reactor_mask="vhost_${vhost_num}_reactor_mask"
+ reactor_mask="${!reactor_mask}"
+
+ local master_core="vhost_${vhost_num}_master_core"
+ master_core="${!master_core}"
+
+ if [[ -z "$reactor_mask" ]] || [[ -z "$master_core" ]]; then
+ error "Parameters vhost_${vhost_num}_reactor_mask or vhost_${vhost_num}_master_core not found in autotest.config file"
+ return 1
+ fi
+
+ local cmd="$vhost_app -m $reactor_mask -p $master_core -s $memory -r $vhost_dir/rpc.sock $no_pci"
+ if [[ -n "$vhost_conf_path" ]]; then
+ cp $vhost_conf_template $vhost_conf_file
+ $SPDK_BUILD_DIR/scripts/gen_nvme.sh >> $vhost_conf_file
+ cmd="$vhost_app -m $reactor_mask -p $master_core -c $vhost_conf_file -s $memory -r $vhost_dir/rpc.sock $no_pci"
+ fi
+
+ notice "Loging to: $vhost_log_file"
+ notice "Socket: $vhost_socket"
+ notice "Command: $cmd"
+
+ timing_enter vhost_start
+ cd $vhost_dir; $cmd &
+ vhost_pid=$!
+ echo $vhost_pid > $vhost_pid_file
+
+ notice "waiting for app to run..."
+ waitforlisten "$vhost_pid" "$vhost_dir/rpc.sock"
+ #do not generate nvmes if pci access is disabled
+ if [[ -z "$vhost_conf_path" ]] && [[ -z "$no_pci" ]]; then
+ $SPDK_BUILD_DIR/scripts/gen_nvme.sh "--json" | $SPDK_BUILD_DIR/scripts/rpc.py\
+ -s $vhost_dir/rpc.sock load_subsystem_config
+ fi
+
+ if [[ -n "$vhost_json_path" ]]; then
+ $SPDK_BUILD_DIR/scripts/rpc.py -s $vhost_dir/rpc.sock load_config < "$vhost_json_path/conf.json"
+ fi
+
+ notice "vhost started - pid=$vhost_pid"
+ timing_exit vhost_start
+
+ rm -f $vhost_conf_file
+}
+
+function spdk_vhost_kill()
+{
+ local rc=0
+ local vhost_num=0
+ if [[ ! -z "$1" ]]; then
+ vhost_num=$1
+ assert_number "$vhost_num"
+ fi
+
+ local vhost_pid_file="$(get_vhost_dir $vhost_num)/vhost.pid"
+
+ if [[ ! -r $vhost_pid_file ]]; then
+ warning "no vhost pid file found"
+ return 0
+ fi
+
+ timing_enter vhost_kill
+ local vhost_pid="$(cat $vhost_pid_file)"
+ notice "killing vhost (PID $vhost_pid) app"
+
+ if /bin/kill -INT $vhost_pid >/dev/null; then
+ notice "sent SIGINT to vhost app - waiting 60 seconds to exit"
+ for ((i=0; i<60; i++)); do
+ if /bin/kill -0 $vhost_pid; then
+ echo "."
+ sleep 1
+ else
+ break
+ fi
+ done
+ if /bin/kill -0 $vhost_pid; then
+ error "ERROR: vhost was NOT killed - sending SIGABRT"
+ /bin/kill -ABRT $vhost_pid
+ rm $vhost_pid_file
+ rc=1
+ else
+ while kill -0 $vhost_pid; do
+ echo "."
+ done
+ fi
+ elif /bin/kill -0 $vhost_pid; then
+ error "vhost NOT killed - you need to kill it manually"
+ rc=1
+ else
+ notice "vhost was no running"
+ fi
+
+ timing_exit vhost_kill
+ if [[ $rc == 0 ]]; then
+ rm $vhost_pid_file
+ fi
+
+ return $rc
+}
+
+###
+# Mgmt functions
+###
+
+function assert_number()
+{
+ [[ "$1" =~ [0-9]+ ]] && return 0
+
+ error "Invalid or missing paramter: need number but got '$1'"
+ return 1;
+}
+
+# Helper to validate VM number
+# param $1 VM number
+#
+function vm_num_is_valid()
+{
+ [[ "$1" =~ ^[0-9]+$ ]] && return 0
+
+ error "Invalid or missing paramter: vm number '$1'"
+ return 1;
+}
+
+
+# Print network socket for given VM number
+# param $1 virtual machine number
+#
+function vm_ssh_socket()
+{
+ vm_num_is_valid $1 || return 1
+ local vm_dir="$VM_BASE_DIR/$1"
+
+ cat $vm_dir/ssh_socket
+}
+
+function vm_fio_socket()
+{
+ vm_num_is_valid $1 || return 1
+ local vm_dir="$VM_BASE_DIR/$1"
+
+ cat $vm_dir/fio_socket
+}
+
+function vm_create_ssh_config()
+{
+ local ssh_config="$VM_BASE_DIR/ssh_config"
+ if [[ ! -f $ssh_config ]]; then
+ (
+ echo "Host *"
+ echo " ControlPersist=10m"
+ echo " ConnectTimeout=1"
+ echo " Compression=no"
+ echo " ControlMaster=auto"
+ echo " UserKnownHostsFile=/dev/null"
+ echo " StrictHostKeyChecking=no"
+ echo " User root"
+ echo " ControlPath=/tmp/%r@%h:%p.ssh"
+ echo ""
+ ) > $ssh_config
+ # Control path created at /tmp because of live migration test case 3.
+ # In case of using sshfs share for the test - control path cannot be
+ # on share because remote server will fail on ssh commands.
+ fi
+}
+
+# Execute ssh command on given VM
+# param $1 virtual machine number
+#
+function vm_ssh()
+{
+ vm_num_is_valid $1 || return 1
+ vm_create_ssh_config
+ local ssh_config="$VM_BASE_DIR/ssh_config"
+
+ local ssh_cmd="ssh -i $SPDK_VHOST_SSH_KEY_FILE -F $ssh_config \
+ -p $(vm_ssh_socket $1) $VM_SSH_OPTIONS 127.0.0.1"
+
+ shift
+ $ssh_cmd "$@"
+}
+
+# Execute scp command on given VM
+# param $1 virtual machine number
+#
+function vm_scp()
+{
+ vm_num_is_valid $1 || return 1
+ vm_create_ssh_config
+ local ssh_config="$VM_BASE_DIR/ssh_config"
+
+ local scp_cmd="scp -i $SPDK_VHOST_SSH_KEY_FILE -F $ssh_config \
+ -P $(vm_ssh_socket $1) "
+
+ shift
+ $scp_cmd "$@"
+}
+
+
+# check if specified VM is running
+# param $1 VM num
+function vm_is_running()
+{
+ vm_num_is_valid $1 || return 1
+ local vm_dir="$VM_BASE_DIR/$1"
+
+ if [[ ! -r $vm_dir/qemu.pid ]]; then
+ return 1
+ fi
+
+ local vm_pid="$(cat $vm_dir/qemu.pid)"
+
+ if /bin/kill -0 $vm_pid; then
+ return 0
+ else
+ if [[ $EUID -ne 0 ]]; then
+ warning "not root - assuming VM running since can't be checked"
+ return 0
+ fi
+
+ # not running - remove pid file
+ rm $vm_dir/qemu.pid
+ return 1
+ fi
+}
+
+# check if specified VM is running
+# param $1 VM num
+function vm_os_booted()
+{
+ vm_num_is_valid $1 || return 1
+ local vm_dir="$VM_BASE_DIR/$1"
+
+ if [[ ! -r $vm_dir/qemu.pid ]]; then
+ error "VM $1 is not running"
+ return 1
+ fi
+
+ if ! VM_SSH_OPTIONS="-o ControlMaster=no" vm_ssh $1 "true" 2>/dev/null; then
+ # Shutdown existing master. Ignore errors as it might not exist.
+ VM_SSH_OPTIONS="-O exit" vm_ssh $1 "true" 2>/dev/null
+ return 1
+ fi
+
+ return 0
+}
+
+
+# Shutdown given VM
+# param $1 virtual machine number
+# return non-zero in case of error.
+function vm_shutdown()
+{
+ vm_num_is_valid $1 || return 1
+ local vm_dir="$VM_BASE_DIR/$1"
+ if [[ ! -d "$vm_dir" ]]; then
+ error "VM$1 ($vm_dir) not exist - setup it first"
+ return 1
+ fi
+
+ if ! vm_is_running $1; then
+ notice "VM$1 ($vm_dir) is not running"
+ return 0
+ fi
+
+ # Temporarily disabling exit flag for next ssh command, since it will
+ # "fail" due to shutdown
+ notice "Shutting down virtual machine $vm_dir"
+ set +e
+ vm_ssh $1 "nohup sh -c 'shutdown -h -P now'" || true
+ notice "VM$1 is shutting down - wait a while to complete"
+ set -e
+}
+
+# Kill given VM
+# param $1 virtual machine number
+#
+function vm_kill()
+{
+ vm_num_is_valid $1 || return 1
+ local vm_dir="$VM_BASE_DIR/$1"
+
+ if [[ ! -r $vm_dir/qemu.pid ]]; then
+ return 0
+ fi
+
+ local vm_pid="$(cat $vm_dir/qemu.pid)"
+
+ notice "Killing virtual machine $vm_dir (pid=$vm_pid)"
+ # First kill should fail, second one must fail
+ if /bin/kill $vm_pid; then
+ notice "process $vm_pid killed"
+ rm $vm_dir/qemu.pid
+ elif vm_is_running $1; then
+ error "Process $vm_pid NOT killed"
+ return 1
+ fi
+}
+
+# List all VM numbers in VM_BASE_DIR
+#
+function vm_list_all()
+{
+ local vms="$(shopt -s nullglob; echo $VM_BASE_DIR/[0-9]*)"
+ if [[ ! -z "$vms" ]]; then
+ basename --multiple $vms
+ fi
+}
+
+# Kills all VM in $VM_BASE_DIR
+#
+function vm_kill_all()
+{
+ local vm
+ for vm in $(vm_list_all); do
+ vm_kill $vm
+ done
+}
+
+# Shutdown all VM in $VM_BASE_DIR
+#
+function vm_shutdown_all()
+{
+ local shell_restore_x="$( [[ "$-" =~ x ]] && echo 'set -x' )"
+ # XXX: temporally disable to debug shutdown issue
+ # set +x
+
+ local vms=$(vm_list_all)
+ local vm
+
+ for vm in $vms; do
+ vm_shutdown $vm
+ done
+
+ notice "Waiting for VMs to shutdown..."
+ local timeo=30
+ while [[ $timeo -gt 0 ]]; do
+ local all_vms_down=1
+ for vm in $vms; do
+ if vm_is_running $vm; then
+ all_vms_down=0
+ break
+ fi
+ done
+
+ if [[ $all_vms_down == 1 ]]; then
+ notice "All VMs successfully shut down"
+ $shell_restore_x
+ return 0
+ fi
+
+ ((timeo-=1))
+ sleep 1
+ done
+
+ $shell_restore_x
+ error "Timeout waiting for some VMs to shutdown"
+ return 1
+}
+
+function vm_setup()
+{
+ local shell_restore_x="$( [[ "$-" =~ x ]] && echo 'set -x' )"
+ local OPTIND optchar vm_num
+
+ local os=""
+ local os_mode=""
+ local qemu_args=""
+ local disk_type_g=NOT_DEFINED
+ local read_only="false"
+ local disks=""
+ local raw_cache=""
+ local vm_incoming=""
+ local vm_migrate_to=""
+ local force_vm=""
+ local guest_memory=1024
+ local queue_number=""
+ local vhost_dir="$(get_vhost_dir)"
+ while getopts ':-:' optchar; do
+ case "$optchar" in
+ -)
+ case "$OPTARG" in
+ os=*) local os="${OPTARG#*=}" ;;
+ os-mode=*) local os_mode="${OPTARG#*=}" ;;
+ qemu-args=*) local qemu_args="${qemu_args} ${OPTARG#*=}" ;;
+ disk-type=*) local disk_type_g="${OPTARG#*=}" ;;
+ read-only=*) local read_only="${OPTARG#*=}" ;;
+ disks=*) local disks="${OPTARG#*=}" ;;
+ raw-cache=*) local raw_cache=",cache${OPTARG#*=}" ;;
+ force=*) local force_vm=${OPTARG#*=} ;;
+ memory=*) local guest_memory=${OPTARG#*=} ;;
+ queue_num=*) local queue_number=${OPTARG#*=} ;;
+ incoming=*) local vm_incoming="${OPTARG#*=}" ;;
+ migrate-to=*) local vm_migrate_to="${OPTARG#*=}" ;;
+ vhost-num=*) local vhost_dir="$(get_vhost_dir ${OPTARG#*=})" ;;
+ spdk-boot=*) local boot_from="${OPTARG#*=}" ;;
+ *)
+ error "unknown argument $OPTARG"
+ return 1
+ esac
+ ;;
+ *)
+ error "vm_create Unknown param $OPTARG"
+ return 1
+ ;;
+ esac
+ done
+
+ # Find next directory we can use
+ if [[ ! -z $force_vm ]]; then
+ vm_num=$force_vm
+
+ vm_num_is_valid $vm_num || return 1
+ local vm_dir="$VM_BASE_DIR/$vm_num"
+ [[ -d $vm_dir ]] && warning "removing existing VM in '$vm_dir'"
+ else
+ local vm_dir=""
+
+ set +x
+ for (( i=0; i<=256; i++)); do
+ local vm_dir="$VM_BASE_DIR/$i"
+ [[ ! -d $vm_dir ]] && break
+ done
+ $shell_restore_x
+
+ vm_num=$i
+ fi
+
+ if [[ $i -eq 256 ]]; then
+ error "no free VM found. do some cleanup (256 VMs created, are you insane?)"
+ return 1
+ fi
+
+ if [[ ! -z "$vm_migrate_to" && ! -z "$vm_incoming" ]]; then
+ error "'--incoming' and '--migrate-to' cannot be used together"
+ return 1
+ elif [[ ! -z "$vm_incoming" ]]; then
+ if [[ ! -z "$os_mode" || ! -z "$os_img" ]]; then
+ error "'--incoming' can't be used together with '--os' nor '--os-mode'"
+ return 1
+ fi
+
+ os_mode="original"
+ os="$VM_BASE_DIR/$vm_incoming/os.qcow2"
+ elif [[ ! -z "$vm_migrate_to" ]]; then
+ [[ "$os_mode" != "backing" ]] && warning "Using 'backing' mode for OS since '--migrate-to' is used"
+ os_mode=backing
+ fi
+
+ notice "Creating new VM in $vm_dir"
+ mkdir -p $vm_dir
+
+ if [[ "$os_mode" == "backing" ]]; then
+ notice "Creating backing file for OS image file: $os"
+ if ! $QEMU_PREFIX/bin/qemu-img create -f qcow2 -b $os $vm_dir/os.qcow2; then
+ error "Failed to create OS backing file in '$vm_dir/os.qcow2' using '$os'"
+ return 1
+ fi
+
+ local os=$vm_dir/os.qcow2
+ elif [[ "$os_mode" == "original" ]]; then
+ warning "Using original OS image file: $os"
+ elif [[ "$os_mode" != "snapshot" ]]; then
+ if [[ -z "$os_mode" ]]; then
+ notice "No '--os-mode' parameter provided - using 'snapshot'"
+ os_mode="snapshot"
+ else
+ error "Invalid '--os-mode=$os_mode'"
+ return 1
+ fi
+ fi
+
+ # WARNING:
+ # each cmd+= must contain ' ${eol}' at the end
+ #
+ local eol="\\\\\n "
+ local qemu_mask_param="VM_${vm_num}_qemu_mask"
+ local qemu_numa_node_param="VM_${vm_num}_qemu_numa_node"
+
+ if [[ -z "${!qemu_mask_param}" ]] || [[ -z "${!qemu_numa_node_param}" ]]; then
+ error "Parameters ${qemu_mask_param} or ${qemu_numa_node_param} not found in autotest.config file"
+ return 1
+ fi
+
+ local task_mask=${!qemu_mask_param}
+
+ notice "TASK MASK: $task_mask"
+ local cmd="taskset -a -c $task_mask $QEMU_PREFIX/bin/qemu-system-x86_64 ${eol}"
+ local vm_socket_offset=$(( 10000 + 100 * vm_num ))
+
+ local ssh_socket=$(( vm_socket_offset + 0 ))
+ local fio_socket=$(( vm_socket_offset + 1 ))
+ local monitor_port=$(( vm_socket_offset + 2 ))
+ local migration_port=$(( vm_socket_offset + 3 ))
+ local gdbserver_socket=$(( vm_socket_offset + 4 ))
+ local vnc_socket=$(( 100 + vm_num ))
+ local qemu_pid_file="$vm_dir/qemu.pid"
+ local cpu_num=0
+
+ set +x
+ # cpu list for taskset can be comma separated or range
+ # or both at the same time, so first split on commas
+ cpu_list=$(echo $task_mask | tr "," "\n")
+ queue_number=0
+ for c in $cpu_list; do
+ # if range is detected - count how many cpus
+ if [[ $c =~ [0-9]+-[0-9]+ ]]; then
+ val=$(($c-1))
+ val=${val#-}
+ else
+ val=1
+ fi
+ cpu_num=$((cpu_num+val))
+ queue_number=$((queue_number+val))
+ done
+
+ if [ -z $queue_number ]; then
+ queue_number=$cpu_num
+ fi
+
+ $shell_restore_x
+
+ local node_num=${!qemu_numa_node_param}
+ local boot_disk_present=false
+ notice "NUMA NODE: $node_num"
+ cmd+="-m $guest_memory --enable-kvm -cpu host -smp $cpu_num -vga std -vnc :$vnc_socket -daemonize ${eol}"
+ cmd+="-object memory-backend-file,id=mem,size=${guest_memory}M,mem-path=/dev/hugepages,share=on,prealloc=yes,host-nodes=$node_num,policy=bind ${eol}"
+ [[ $os_mode == snapshot ]] && cmd+="-snapshot ${eol}"
+ [[ ! -z "$vm_incoming" ]] && cmd+=" -incoming tcp:0:$migration_port ${eol}"
+ cmd+="-monitor telnet:127.0.0.1:$monitor_port,server,nowait ${eol}"
+ cmd+="-numa node,memdev=mem ${eol}"
+ cmd+="-pidfile $qemu_pid_file ${eol}"
+ cmd+="-serial file:$vm_dir/serial.log ${eol}"
+ cmd+="-D $vm_dir/qemu.log ${eol}"
+ cmd+="-net user,hostfwd=tcp::$ssh_socket-:22,hostfwd=tcp::$fio_socket-:8765 ${eol}"
+ cmd+="-net nic ${eol}"
+ if [[ -z "$boot_from" ]]; then
+ cmd+="-drive file=$os,if=none,id=os_disk ${eol}"
+ cmd+="-device ide-hd,drive=os_disk,bootindex=0 ${eol}"
+ fi
+
+ if ( [[ $disks == '' ]] && [[ $disk_type_g == virtio* ]] ); then
+ disks=1
+ fi
+
+ for disk in ${disks//:/ }; do
+ if [[ $disk = *","* ]]; then
+ disk_type=${disk#*,}
+ disk=${disk%,*}
+ else
+ disk_type=$disk_type_g
+ fi
+
+ case $disk_type in
+ virtio)
+ local raw_name="RAWSCSI"
+ local raw_disk=$vm_dir/test.img
+
+ if [[ ! -z $disk ]]; then
+ [[ ! -b $disk ]] && touch $disk
+ local raw_disk=$(readlink -f $disk)
+ fi
+
+ # Create disk file if it not exist or it is smaller than 1G
+ if ( [[ -f $raw_disk ]] && [[ $(stat --printf="%s" $raw_disk) -lt $((1024 * 1024 * 1024)) ]] ) || \
+ [[ ! -e $raw_disk ]]; then
+ if [[ $raw_disk =~ /dev/.* ]]; then
+ error \
+ "ERROR: Virtio disk point to missing device ($raw_disk) -\n" \
+ " this is probably not what you want."
+ return 1
+ fi
+
+ notice "Creating Virtio disc $raw_disk"
+ dd if=/dev/zero of=$raw_disk bs=1024k count=1024
+ else
+ notice "Using existing image $raw_disk"
+ fi
+
+ cmd+="-device virtio-scsi-pci,num_queues=$queue_number ${eol}"
+ cmd+="-device scsi-hd,drive=hd$i,vendor=$raw_name ${eol}"
+ cmd+="-drive if=none,id=hd$i,file=$raw_disk,format=raw$raw_cache ${eol}"
+ ;;
+ spdk_vhost_scsi)
+ notice "using socket $vhost_dir/naa.$disk.$vm_num"
+ cmd+="-chardev socket,id=char_$disk,path=$vhost_dir/naa.$disk.$vm_num ${eol}"
+ cmd+="-device vhost-user-scsi-pci,id=scsi_$disk,num_queues=$queue_number,chardev=char_$disk"
+ if [[ "$disk" == "$boot_from" ]]; then
+ cmd+=",bootindex=0"
+ boot_disk_present=true
+ fi
+ cmd+=" ${eol}"
+ ;;
+ spdk_vhost_blk)
+ notice "using socket $vhost_dir/naa.$disk.$vm_num"
+ cmd+="-chardev socket,id=char_$disk,path=$vhost_dir/naa.$disk.$vm_num ${eol}"
+ cmd+="-device vhost-user-blk-pci,num-queues=$queue_number,chardev=char_$disk"
+ if [[ "$disk" == "$boot_from" ]]; then
+ cmd+=",bootindex=0"
+ boot_disk_present=true
+ fi
+ cmd+=" ${eol}"
+ ;;
+ kernel_vhost)
+ if [[ -z $disk ]]; then
+ error "need WWN for $disk_type"
+ return 1
+ elif [[ ! $disk =~ ^[[:alpha:]]{3}[.][[:xdigit:]]+$ ]]; then
+ error "$disk_type - disk(wnn)=$disk does not look like WNN number"
+ return 1
+ fi
+ notice "Using kernel vhost disk wwn=$disk"
+ cmd+=" -device vhost-scsi-pci,wwpn=$disk,num_queues=$queue_number ${eol}"
+ ;;
+ *)
+ error "unknown mode '$disk_type', use: virtio, spdk_vhost_scsi, spdk_vhost_blk or kernel_vhost"
+ return 1
+ esac
+ done
+
+ if [[ -n $boot_from ]] && [[ $boot_disk_present == false ]]; then
+ error "Boot from $boot_from is selected but device is not present"
+ return 1
+ fi
+
+ [[ ! -z $qemu_args ]] && cmd+=" $qemu_args ${eol}"
+ # remove last $eol
+ cmd="${cmd%\\\\\\n }"
+
+ notice "Saving to $vm_dir/run.sh"
+ (
+ echo '#!/bin/bash'
+ echo 'if [[ $EUID -ne 0 ]]; then '
+ echo ' echo "Go away user come back as root"'
+ echo ' exit 1'
+ echo 'fi';
+ echo
+ echo -e "qemu_cmd=\"$cmd\"";
+ echo
+ echo "echo 'Running VM in $vm_dir'"
+ echo "rm -f $qemu_pid_file"
+ echo '$qemu_cmd'
+ echo "echo 'Waiting for QEMU pid file'"
+ echo "sleep 1"
+ echo "[[ ! -f $qemu_pid_file ]] && sleep 1"
+ echo "[[ ! -f $qemu_pid_file ]] && echo 'ERROR: no qemu pid file found' && exit 1"
+ echo
+ echo "chmod +r $vm_dir/*"
+ echo
+ echo "echo '=== qemu.log ==='"
+ echo "cat $vm_dir/qemu.log"
+ echo "echo '=== qemu.log ==='"
+ echo '# EOF'
+ ) > $vm_dir/run.sh
+ chmod +x $vm_dir/run.sh
+
+ # Save generated sockets redirection
+ echo $ssh_socket > $vm_dir/ssh_socket
+ echo $fio_socket > $vm_dir/fio_socket
+ echo $monitor_port > $vm_dir/monitor_port
+
+ rm -f $vm_dir/migration_port
+ [[ -z $vm_incoming ]] || echo $migration_port > $vm_dir/migration_port
+
+ echo $gdbserver_socket > $vm_dir/gdbserver_socket
+ echo $vnc_socket >> $vm_dir/vnc_socket
+
+ [[ -z $vm_incoming ]] || ln -fs $VM_BASE_DIR/$vm_incoming $vm_dir/vm_incoming
+ [[ -z $vm_migrate_to ]] || ln -fs $VM_BASE_DIR/$vm_migrate_to $vm_dir/vm_migrate_to
+}
+
+function vm_run()
+{
+ local OPTIND optchar vm
+ local run_all=false
+ local vms_to_run=""
+
+ while getopts 'a-:' optchar; do
+ case "$optchar" in
+ a) run_all=true ;;
+ *)
+ error "Unknown param $OPTARG"
+ return 1
+ ;;
+ esac
+ done
+
+ if $run_all; then
+ vms_to_run="$(vm_list_all)"
+ else
+ shift $((OPTIND-1))
+ for vm in $@; do
+ vm_num_is_valid $1 || return 1
+ if [[ ! -x $VM_BASE_DIR/$vm/run.sh ]]; then
+ error "VM$vm not defined - setup it first"
+ return 1
+ fi
+ vms_to_run+=" $vm"
+ done
+ fi
+
+ for vm in $vms_to_run; do
+ if vm_is_running $vm; then
+ warning "VM$vm ($VM_BASE_DIR/$vm) already running"
+ continue
+ fi
+
+ notice "running $VM_BASE_DIR/$vm/run.sh"
+ if ! $VM_BASE_DIR/$vm/run.sh; then
+ error "FAILED to run vm $vm"
+ return 1
+ fi
+ done
+}
+
+# Wait for all created VMs to boot.
+# param $1 max wait time
+function vm_wait_for_boot()
+{
+ assert_number $1
+
+ local shell_restore_x="$( [[ "$-" =~ x ]] && echo 'set -x' )"
+ set +x
+
+ local all_booted=false
+ local timeout_time=$1
+ [[ $timeout_time -lt 10 ]] && timeout_time=10
+ local timeout_time=$(date -d "+$timeout_time seconds" +%s)
+
+ notice "Waiting for VMs to boot"
+ shift
+ if [[ "$@" == "" ]]; then
+ local vms_to_check="$VM_BASE_DIR/[0-9]*"
+ else
+ local vms_to_check=""
+ for vm in $@; do
+ vms_to_check+=" $VM_BASE_DIR/$vm"
+ done
+ fi
+
+ for vm in $vms_to_check; do
+ local vm_num=$(basename $vm)
+ local i=0
+ notice "waiting for VM$vm_num ($vm)"
+ while ! vm_os_booted $vm_num; do
+ if ! vm_is_running $vm_num; then
+
+ warning "VM $vm_num is not running"
+ warning "================"
+ warning "QEMU LOG:"
+ if [[ -r $vm/qemu.log ]]; then
+ cat $vm/qemu.log
+ else
+ warning "LOG not found"
+ fi
+
+ warning "VM LOG:"
+ if [[ -r $vm/serial.log ]]; then
+ cat $vm/serial.log
+ else
+ warning "LOG not found"
+ fi
+ warning "================"
+ $shell_restore_x
+ return 1
+ fi
+
+ if [[ $(date +%s) -gt $timeout_time ]]; then
+ warning "timeout waiting for machines to boot"
+ $shell_restore_x
+ return 1
+ fi
+ if (( i > 30 )); then
+ local i=0
+ echo
+ fi
+ echo -n "."
+ sleep 1
+ done
+ echo ""
+ notice "VM$vm_num ready"
+ #Change Timeout for stopping services to prevent lengthy powerdowns
+ vm_ssh $vm_num "echo 'DefaultTimeoutStopSec=10' >> /etc/systemd/system.conf; systemctl daemon-reexec"
+ done
+
+ notice "all VMs ready"
+ $shell_restore_x
+ return 0
+}
+
+function vm_start_fio_server()
+{
+ local OPTIND optchar
+ local readonly=''
+ while getopts ':-:' optchar; do
+ case "$optchar" in
+ -)
+ case "$OPTARG" in
+ fio-bin=*) local fio_bin="${OPTARG#*=}" ;;
+ readonly) local readonly="--readonly" ;;
+ *) error "Invalid argument '$OPTARG'" && return 1;;
+ esac
+ ;;
+ *) error "Invalid argument '$OPTARG'" && return 1;;
+ esac
+ done
+
+ shift $(( OPTIND - 1 ))
+ for vm_num in $@; do
+ notice "Starting fio server on VM$vm_num"
+ if [[ $fio_bin != "" ]]; then
+ cat $fio_bin | vm_ssh $vm_num 'cat > /root/fio; chmod +x /root/fio'
+ vm_ssh $vm_num /root/fio $readonly --eta=never --server --daemonize=/root/fio.pid
+ else
+ vm_ssh $vm_num fio $readonly --eta=never --server --daemonize=/root/fio.pid
+ fi
+ done
+}
+
+function vm_check_scsi_location()
+{
+ # Script to find wanted disc
+ local script='shopt -s nullglob; \
+ for entry in /sys/block/sd*; do \
+ disk_type="$(cat $entry/device/vendor)"; \
+ if [[ $disk_type == INTEL* ]] || [[ $disk_type == RAWSCSI* ]] || [[ $disk_type == LIO-ORG* ]]; then \
+ fname=$(basename $entry); \
+ echo -n " $fname"; \
+ fi; \
+ done'
+
+ SCSI_DISK="$(echo "$script" | vm_ssh $1 bash -s)"
+
+ if [[ -z "$SCSI_DISK" ]]; then
+ error "no test disk found!"
+ return 1
+ fi
+}
+
+# Script to perform scsi device reset on all disks in VM
+# param $1 VM num
+# param $2..$n Disks to perform reset on
+function vm_reset_scsi_devices()
+{
+ for disk in "${@:2}"; do
+ notice "VM$1 Performing device reset on disk $disk"
+ vm_ssh $1 sg_reset /dev/$disk -vNd
+ done
+}
+
+function vm_check_blk_location()
+{
+ local script='shopt -s nullglob; cd /sys/block; echo vd*'
+ SCSI_DISK="$(echo "$script" | vm_ssh $1 bash -s)"
+
+ if [[ -z "$SCSI_DISK" ]]; then
+ error "no blk test disk found!"
+ return 1
+ fi
+}
+
+function run_fio()
+{
+ local arg
+ local job_file=""
+ local fio_bin=""
+ local vms=()
+ local out=""
+ local fio_disks=""
+ local vm
+ local run_server_mode=true
+
+ for arg in $@; do
+ case "$arg" in
+ --job-file=*) local job_file="${arg#*=}" ;;
+ --fio-bin=*) local fio_bin="${arg#*=}" ;;
+ --vm=*) vms+=( "${arg#*=}" ) ;;
+ --out=*)
+ local out="${arg#*=}"
+ mkdir -p $out
+ ;;
+ --local) run_server_mode=false ;;
+ --json) json="--json" ;;
+ *)
+ error "Invalid argument '$arg'"
+ return 1
+ ;;
+ esac
+ done
+
+ if [[ ! -z "$fio_bin" && ! -r "$fio_bin" ]]; then
+ error "FIO binary '$fio_bin' does not exist"
+ return 1
+ fi
+
+ if [[ ! -r "$job_file" ]]; then
+ error "Fio job '$job_file' does not exist"
+ return 1
+ fi
+
+ local job_fname=$(basename "$job_file")
+ # prepare job file for each VM
+ for vm in ${vms[@]}; do
+ local vm_num=${vm%%:*}
+ local vmdisks=${vm#*:}
+
+ sed "s@filename=@filename=$vmdisks@" $job_file | vm_ssh $vm_num "cat > /root/$job_fname"
+ fio_disks+="127.0.0.1:$(vm_fio_socket $vm_num):$vmdisks,"
+
+ vm_ssh $vm_num cat /root/$job_fname
+ if ! $run_server_mode; then
+ if [[ ! -z "$fio_bin" ]]; then
+ cat $fio_bin | vm_ssh $vm_num 'cat > /root/fio; chmod +x /root/fio'
+ fi
+
+ notice "Running local fio on VM $vm_num"
+ vm_ssh $vm_num "nohup /root/fio /root/$job_fname 1>/root/$job_fname.out 2>/root/$job_fname.out </dev/null & echo \$! > /root/fio.pid"
+ fi
+ done
+
+ if ! $run_server_mode; then
+ # Give FIO time to run
+ sleep 0.5
+ return 0
+ fi
+
+ $SPDK_BUILD_DIR/test/vhost/common/run_fio.py --job-file=/root/$job_fname \
+ $([[ ! -z "$fio_bin" ]] && echo "--fio-bin=$fio_bin") \
+ --out=$out $json ${fio_disks%,}
+}
+
+# Shutdown or kill any running VM and SPDK APP.
+#
+function at_app_exit()
+{
+ local vhost_num
+
+ notice "APP EXITING"
+ notice "killing all VMs"
+ vm_kill_all
+ # Kill vhost application
+ notice "killing vhost app"
+
+ for vhost_num in $(spdk_vhost_list_all); do
+ spdk_vhost_kill $vhost_num
+ done
+
+ notice "EXIT DONE"
+}
+
+function error_exit()
+{
+ trap - ERR
+ print_backtrace
+ set +e
+ error "Error on $1 $2"
+
+ at_app_exit
+ exit 1
+}