#!/usr/bin/env bash rootdir=$(readlink -f $(dirname $0)) # In autotest_common.sh all tests are disabled by default. # If the configuration of tests is not provided, no tests will be carried out. if [[ ! -f $1 ]]; then echo "ERROR: SPDK test configuration not specified" exit 1 fi source "$1" source "$rootdir/test/common/autotest_common.sh" source "$rootdir/test/nvmf/common.sh" # always test with SPDK shared objects. export SPDK_LIB_DIR="$rootdir/build/lib" export DPDK_LIB_DIR="$rootdir/dpdk/build/lib" export LD_LIBRARY_PATH=$LD_LIBRARY_PATH:$SPDK_LIB_DIR:$DPDK_LIB_DIR if [ $EUID -ne 0 ]; then echo "$0 must be run as root" exit 1 fi if [ $(uname -s) = Linux ]; then # set core_pattern to a known value to avoid ABRT, systemd-coredump, etc. echo "core" > /proc/sys/kernel/core_pattern # Make sure that the hugepage state for our VM is fresh so we don't fail # hugepage allocation. Allow time for this action to complete. echo 1 > /proc/sys/vm/drop_caches sleep 3 # make sure nbd (network block device) driver is loaded if it is available # this ensures that when tests need to use nbd, it will be fully initialized modprobe nbd || true fi trap "process_core; autotest_cleanup; exit 1" SIGINT SIGTERM EXIT timing_enter autotest create_test_list src=$(readlink -f $(dirname $0)) out=$output_dir cd $src ./scripts/setup.sh status freebsd_update_contigmem_mod if hash lcov; then # setup output dir for unittest.sh export UT_COVERAGE=$out/ut_coverage export LCOV_OPTS=" --rc lcov_branch_coverage=1 --rc lcov_function_coverage=1 --rc genhtml_branch_coverage=1 --rc genhtml_function_coverage=1 --rc genhtml_legend=1 --rc geninfo_all_blocks=1 " export LCOV="lcov $LCOV_OPTS --no-external" # Print lcov version to log $LCOV -v # zero out coverage data $LCOV -q -c -i -t "Baseline" -d $src -o $out/cov_base.info fi # Make sure the disks are clean (no leftover partition tables) timing_enter cleanup # Remove old domain socket pathname just in case rm -f /var/tmp/spdk*.sock # Load the kernel driver ./scripts/setup.sh reset # Let the kernel discover any filesystems or partitions sleep 10 if [ $(uname -s) = Linux ]; then # OCSSD devices drivers don't support IO issues by kernel so # detect OCSSD devices and blacklist them (unbind from any driver). # If test scripts want to use this device it needs to do this explicitly. # # If some OCSSD device is bound to other driver than nvme we won't be able to # discover if it is OCSSD or not so load the kernel driver first. while IFS= read -r -d '' dev; do # Send Open Channel 2.0 Geometry opcode "0xe2" - not supported by NVMe device. if nvme admin-passthru $dev --namespace-id=1 --data-len=4096 --opcode=0xe2 --read > /dev/null; then bdf="$(basename $(readlink -e /sys/class/nvme/${dev#/dev/}/device))" echo "INFO: blacklisting OCSSD device: $dev ($bdf)" PCI_BLACKLIST+=" $bdf" OCSSD_PCI_DEVICES+=" $bdf" fi done < <(find /dev -maxdepth 1 -regex '/dev/nvme[0-9]+' -print0) export OCSSD_PCI_DEVICES # Now, bind blacklisted devices to pci-stub module. This will prevent # automatic grabbing these devices when we add device/vendor ID to # proper driver. if [[ -n "$PCI_BLACKLIST" ]]; then # shellcheck disable=SC2097,SC2098 PCI_WHITELIST="$PCI_BLACKLIST" \ PCI_BLACKLIST="" \ DRIVER_OVERRIDE="pci-stub" \ ./scripts/setup.sh # Export our blacklist so it will take effect during next setup.sh export PCI_BLACKLIST fi fi # Revert NVMe namespaces to default state nvme_namespace_revert # Delete all leftover lvols and gpt partitions # Matches both /dev/nvmeXnY on Linux and /dev/nvmeXnsY on BSD # Filter out nvme with partitions - the "p*" suffix for dev in $(ls /dev/nvme*n* | grep -v p || true); do dd if=/dev/zero of="$dev" bs=1M count=1 done sync timing_exit cleanup # set up huge pages timing_enter afterboot ./scripts/setup.sh timing_exit afterboot timing_enter nvmf_setup rdma_device_init timing_exit nvmf_setup if [[ $SPDK_TEST_CRYPTO -eq 1 || $SPDK_TEST_REDUCE -eq 1 ]]; then if grep -q '#define SPDK_CONFIG_IGB_UIO_DRIVER 1' $rootdir/include/spdk/config.h; then ./scripts/qat_setup.sh igb_uio else ./scripts/qat_setup.sh fi fi # Revert existing OPAL to factory settings that may have been left from earlier failed tests. # This ensures we won't hit any unexpected failures due to NVMe SSDs being locked. opal_revert_cleanup ##################### # Unit Tests ##################### if [ $SPDK_TEST_UNITTEST -eq 1 ]; then run_test "unittest" ./test/unit/unittest.sh run_test "env" test/env/env.sh fi if [ $SPDK_RUN_FUNCTIONAL_TEST -eq 1 ]; then timing_enter lib run_test "rpc" test/rpc/rpc.sh run_test "rpc_client" test/rpc_client/rpc_client.sh run_test "json_config" ./test/json_config/json_config.sh run_test "alias_rpc" test/json_config/alias_rpc/alias_rpc.sh run_test "spdkcli_tcp" test/spdkcli/tcp.sh run_test "dpdk_mem_utility" test/dpdk_memory_utility/test_dpdk_mem_info.sh run_test "event" test/event/event.sh if [ $SPDK_TEST_BLOCKDEV -eq 1 ]; then run_test "blockdev_general" test/bdev/blockdev.sh run_test "bdev_raid" test/bdev/bdev_raid.sh run_test "bdevperf_config" test/bdev/bdevperf/test_config.sh if [[ $(uname -s) == Linux ]]; then run_test "spdk_dd" test/dd/dd.sh fi fi if [ $SPDK_TEST_JSON -eq 1 ]; then run_test "test_converter" test/config_converter/test_converter.sh fi if [ $SPDK_TEST_NVME -eq 1 ]; then run_test "blockdev_nvme" test/bdev/blockdev.sh "nvme" run_test "blockdev_nvme_gpt" test/bdev/blockdev.sh "gpt" run_test "nvme" test/nvme/nvme.sh if [[ $SPDK_TEST_NVME_CLI -eq 1 ]]; then run_test "nvme_cli" test/nvme/spdk_nvme_cli.sh fi if [[ $SPDK_TEST_NVME_CUSE -eq 1 ]]; then run_test "nvme_cuse" test/nvme/cuse/nvme_cuse.sh fi run_test "nvme_rpc" test/nvme/nvme_rpc.sh # Only test hotplug without ASAN enabled. Since if it is # enabled, it catches SEGV earlier than our handler which # breaks the hotplug logic. if [ $SPDK_RUN_ASAN -eq 0 ]; then run_test "nvme_hotplug" test/nvme/hotplug.sh intel fi fi if [ $SPDK_TEST_IOAT -eq 1 ]; then run_test "ioat" test/ioat/ioat.sh fi timing_exit lib if [ $SPDK_TEST_ISCSI -eq 1 ]; then run_test "iscsi_tgt_posix" ./test/iscsi_tgt/iscsi_tgt.sh posix run_test "spdkcli_iscsi" ./test/spdkcli/iscsi.sh # Run raid spdkcli test under iSCSI since blockdev tests run on systems that can't run spdkcli yet run_test "spdkcli_raid" test/spdkcli/raid.sh fi if [ $SPDK_TEST_VPP -eq 1 ]; then run_test "iscsi_tgt_vpp" ./test/iscsi_tgt/iscsi_tgt.sh vpp fi if [ $SPDK_TEST_BLOBFS -eq 1 ]; then run_test "rocksdb" ./test/blobfs/rocksdb/rocksdb.sh run_test "blobstore" ./test/blobstore/blobstore.sh run_test "blobfs" ./test/blobfs/blobfs.sh run_test "hello_blob" $SPDK_EXAMPLE_DIR/hello_blob \ examples/blob/hello_world/hello_blob.json fi if [ $SPDK_TEST_NVMF -eq 1 ]; then # The NVMe-oF run test cases are split out like this so that the parser that compiles the # list of all tests can properly differentiate them. Please do not merge them into one line. if [ "$SPDK_TEST_NVMF_TRANSPORT" = "rdma" ]; then run_test "nvmf_rdma" ./test/nvmf/nvmf.sh --transport=$SPDK_TEST_NVMF_TRANSPORT run_test "spdkcli_nvmf_rdma" ./test/spdkcli/nvmf.sh elif [ "$SPDK_TEST_NVMF_TRANSPORT" = "tcp" ]; then run_test "nvmf_tcp" ./test/nvmf/nvmf.sh --transport=$SPDK_TEST_NVMF_TRANSPORT run_test "spdkcli_nvmf_tcp" ./test/spdkcli/nvmf.sh run_test "nvmf_identify_passthru" test/nvmf/target/identify_passthru.sh --transport=$SPDK_TEST_NVMF_TRANSPORT elif [ "$SPDK_TEST_NVMF_TRANSPORT" = "fc" ]; then run_test "nvmf_fc" ./test/nvmf/nvmf.sh --transport=$SPDK_TEST_NVMF_TRANSPORT run_test "spdkcli_nvmf_fc" ./test/spdkcli/nvmf.sh else echo "unknown NVMe transport, please specify rdma, tcp, or fc." exit 1 fi fi if [ $SPDK_TEST_VHOST -eq 1 ]; then run_test "vhost" ./test/vhost/vhost.sh fi if [ $SPDK_TEST_LVOL -eq 1 ]; then run_test "lvol2" ./test/lvol/lvol2.sh run_test "blob_io_wait" ./test/blobstore/blob_io_wait/blob_io_wait.sh fi if [ $SPDK_TEST_VHOST_INIT -eq 1 ]; then timing_enter vhost_initiator run_test "vhost_blockdev" ./test/vhost/initiator/blockdev.sh run_test "spdkcli_virtio" ./test/spdkcli/virtio.sh run_test "vhost_shared" ./test/vhost/shared/shared.sh run_test "vhost_fuzz" ./test/vhost/fuzz/fuzz.sh timing_exit vhost_initiator fi if [ $SPDK_TEST_PMDK -eq 1 ]; then run_test "blockdev_pmem" ./test/bdev/blockdev.sh "pmem" run_test "pmem" ./test/pmem/pmem.sh -x run_test "spdkcli_pmem" ./test/spdkcli/pmem.sh fi if [ $SPDK_TEST_RBD -eq 1 ]; then run_test "blockdev_rbd" ./test/bdev/blockdev.sh "rbd" run_test "spdkcli_rbd" ./test/spdkcli/rbd.sh fi if [ $SPDK_TEST_OCF -eq 1 ]; then run_test "ocf" ./test/ocf/ocf.sh fi if [ $SPDK_TEST_FTL -eq 1 ]; then run_test "ftl" ./test/ftl/ftl.sh fi if [ $SPDK_TEST_VMD -eq 1 ]; then run_test "vmd" ./test/vmd/vmd.sh fi if [ $SPDK_TEST_REDUCE -eq 1 ]; then run_test "compress_qat" ./test/compress/compress.sh "qat" run_test "compress_isal" ./test/compress/compress.sh "isal" fi if [ $SPDK_TEST_OPAL -eq 1 ]; then run_test "nvme_opal" ./test/nvme/nvme_opal.sh fi if [ $SPDK_TEST_CRYPTO -eq 1 ]; then run_test "blockdev_crypto_aesni" ./test/bdev/blockdev.sh "crypto_aesni" # Proceed with the test only if QAT devices are in place if [[ $(lspci -d:37c8) ]]; then run_test "blockdev_crypto_qat" ./test/bdev/blockdev.sh "crypto_qat" fi fi fi timing_enter cleanup autotest_cleanup timing_exit cleanup timing_exit autotest chmod a+r $output_dir/timing.txt trap - SIGINT SIGTERM EXIT # catch any stray core files process_core if hash lcov; then # generate coverage data and combine with baseline $LCOV -q -c -d $src -t "$(hostname)" -o $out/cov_test.info $LCOV -q -a $out/cov_base.info -a $out/cov_test.info -o $out/cov_total.info $LCOV -q -r $out/cov_total.info '*/dpdk/*' -o $out/cov_total.info $LCOV -q -r $out/cov_total.info '/usr/*' -o $out/cov_total.info git clean -f "*.gcda" rm -f cov_base.info cov_test.info OLD_STDOUT OLD_STDERR fi