summaryrefslogtreecommitdiffstats
path: root/src/spdk/lib/vhost/vhost_scsi.c
diff options
context:
space:
mode:
Diffstat (limited to 'src/spdk/lib/vhost/vhost_scsi.c')
-rw-r--r--src/spdk/lib/vhost/vhost_scsi.c1271
1 files changed, 1271 insertions, 0 deletions
diff --git a/src/spdk/lib/vhost/vhost_scsi.c b/src/spdk/lib/vhost/vhost_scsi.c
new file mode 100644
index 00000000..aefa4c45
--- /dev/null
+++ b/src/spdk/lib/vhost/vhost_scsi.c
@@ -0,0 +1,1271 @@
+/*-
+ * BSD LICENSE
+ *
+ * Copyright(c) Intel Corporation. All rights reserved.
+ * All rights reserved.
+ *
+ * Redistribution and use in source and binary forms, with or without
+ * modification, are permitted provided that the following conditions
+ * are met:
+ *
+ * * Redistributions of source code must retain the above copyright
+ * notice, this list of conditions and the following disclaimer.
+ * * Redistributions in binary form must reproduce the above copyright
+ * notice, this list of conditions and the following disclaimer in
+ * the documentation and/or other materials provided with the
+ * distribution.
+ * * Neither the name of Intel Corporation nor the names of its
+ * contributors may be used to endorse or promote products derived
+ * from this software without specific prior written permission.
+ *
+ * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
+ * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
+ * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR
+ * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
+ * OWNER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL,
+ * SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT
+ * LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
+ * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
+ * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
+ * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE
+ * OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
+ */
+
+#include "spdk/stdinc.h"
+
+#include <linux/virtio_scsi.h>
+
+#include "spdk/env.h"
+#include "spdk/thread.h"
+#include "spdk/scsi.h"
+#include "spdk/scsi_spec.h"
+#include "spdk/conf.h"
+#include "spdk/event.h"
+#include "spdk/util.h"
+#include "spdk/likely.h"
+
+#include "spdk/vhost.h"
+#include "vhost_internal.h"
+
+/* Features supported by SPDK VHOST lib. */
+#define SPDK_VHOST_SCSI_FEATURES (SPDK_VHOST_FEATURES | \
+ (1ULL << VIRTIO_SCSI_F_INOUT) | \
+ (1ULL << VIRTIO_SCSI_F_HOTPLUG) | \
+ (1ULL << VIRTIO_SCSI_F_CHANGE ) | \
+ (1ULL << VIRTIO_SCSI_F_T10_PI ))
+
+/* Features that are specified in VIRTIO SCSI but currently not supported:
+ * - Live migration not supported yet
+ * - T10 PI
+ */
+#define SPDK_VHOST_SCSI_DISABLED_FEATURES (SPDK_VHOST_DISABLED_FEATURES | \
+ (1ULL << VIRTIO_SCSI_F_T10_PI ))
+
+#define MGMT_POLL_PERIOD_US (1000 * 5)
+
+#define VIRTIO_SCSI_CONTROLQ 0
+#define VIRTIO_SCSI_EVENTQ 1
+#define VIRTIO_SCSI_REQUESTQ 2
+
+struct spdk_scsi_dev_vhost_state {
+ bool removed;
+ spdk_vhost_event_fn remove_cb;
+ void *remove_ctx;
+};
+
+struct spdk_vhost_scsi_dev {
+ struct spdk_vhost_dev vdev;
+ struct spdk_scsi_dev *scsi_dev[SPDK_VHOST_SCSI_CTRLR_MAX_DEVS];
+ struct spdk_scsi_dev_vhost_state scsi_dev_state[SPDK_VHOST_SCSI_CTRLR_MAX_DEVS];
+
+ struct spdk_poller *requestq_poller;
+ struct spdk_poller *mgmt_poller;
+ struct spdk_vhost_dev_destroy_ctx destroy_ctx;
+} __rte_cache_aligned;
+
+struct spdk_vhost_scsi_task {
+ struct spdk_scsi_task scsi;
+ struct iovec iovs[SPDK_VHOST_IOVS_MAX];
+
+ union {
+ struct virtio_scsi_cmd_resp *resp;
+ struct virtio_scsi_ctrl_tmf_resp *tmf_resp;
+ };
+
+ struct spdk_vhost_scsi_dev *svdev;
+ struct spdk_scsi_dev *scsi_dev;
+
+ /** Number of bytes that were written. */
+ uint32_t used_len;
+
+ int req_idx;
+
+ /* If set, the task is currently used for I/O processing. */
+ bool used;
+
+ struct spdk_vhost_virtqueue *vq;
+};
+
+static int spdk_vhost_scsi_start(struct spdk_vhost_dev *, void *);
+static int spdk_vhost_scsi_stop(struct spdk_vhost_dev *, void *);
+static void spdk_vhost_scsi_dump_info_json(struct spdk_vhost_dev *vdev,
+ struct spdk_json_write_ctx *w);
+static void spdk_vhost_scsi_write_config_json(struct spdk_vhost_dev *vdev,
+ struct spdk_json_write_ctx *w);
+static int spdk_vhost_scsi_dev_remove(struct spdk_vhost_dev *vdev);
+
+const struct spdk_vhost_dev_backend spdk_vhost_scsi_device_backend = {
+ .virtio_features = SPDK_VHOST_SCSI_FEATURES,
+ .disabled_features = SPDK_VHOST_SCSI_DISABLED_FEATURES,
+ .start_device = spdk_vhost_scsi_start,
+ .stop_device = spdk_vhost_scsi_stop,
+ .dump_info_json = spdk_vhost_scsi_dump_info_json,
+ .write_config_json = spdk_vhost_scsi_write_config_json,
+ .remove_device = spdk_vhost_scsi_dev_remove,
+};
+
+static void
+spdk_vhost_scsi_task_put(struct spdk_vhost_scsi_task *task)
+{
+ spdk_scsi_task_put(&task->scsi);
+}
+
+static void
+spdk_vhost_scsi_task_free_cb(struct spdk_scsi_task *scsi_task)
+{
+ struct spdk_vhost_scsi_task *task = SPDK_CONTAINEROF(scsi_task, struct spdk_vhost_scsi_task, scsi);
+
+ assert(task->svdev->vdev.task_cnt > 0);
+ task->svdev->vdev.task_cnt--;
+ task->used = false;
+}
+
+static void
+process_removed_devs(struct spdk_vhost_scsi_dev *svdev)
+{
+ struct spdk_scsi_dev *dev;
+ struct spdk_scsi_dev_vhost_state *state;
+ int i;
+
+ for (i = 0; i < SPDK_VHOST_SCSI_CTRLR_MAX_DEVS; ++i) {
+ dev = svdev->scsi_dev[i];
+ state = &svdev->scsi_dev_state[i];
+
+ if (dev && state->removed && !spdk_scsi_dev_has_pending_tasks(dev)) {
+ spdk_scsi_dev_free_io_channels(dev);
+ svdev->scsi_dev[i] = NULL;
+ spdk_scsi_dev_destruct(dev);
+ if (state->remove_cb) {
+ state->remove_cb(&svdev->vdev, state->remove_ctx);
+ state->remove_cb = NULL;
+ }
+ SPDK_INFOLOG(SPDK_LOG_VHOST, "%s: hot-detached device 'Dev %u'.\n",
+ svdev->vdev.name, i);
+ }
+ }
+}
+
+static void
+eventq_enqueue(struct spdk_vhost_scsi_dev *svdev, unsigned scsi_dev_num, uint32_t event,
+ uint32_t reason)
+{
+ struct spdk_vhost_virtqueue *vq;
+ struct vring_desc *desc, *desc_table;
+ struct virtio_scsi_event *desc_ev;
+ uint32_t desc_table_size, req_size = 0;
+ uint16_t req;
+ int rc;
+
+ assert(scsi_dev_num < SPDK_VHOST_SCSI_CTRLR_MAX_DEVS);
+ vq = &svdev->vdev.virtqueue[VIRTIO_SCSI_EVENTQ];
+
+ if (spdk_vhost_vq_avail_ring_get(vq, &req, 1) != 1) {
+ SPDK_ERRLOG("Controller %s: Failed to send virtio event (no avail ring entries?).\n",
+ svdev->vdev.name);
+ return;
+ }
+
+ rc = spdk_vhost_vq_get_desc(&svdev->vdev, vq, req, &desc, &desc_table, &desc_table_size);
+ if (rc != 0 || desc->len < sizeof(*desc_ev)) {
+ SPDK_ERRLOG("Controller %s: Invalid eventq descriptor at index %"PRIu16".\n",
+ svdev->vdev.name, req);
+ goto out;
+ }
+
+ desc_ev = spdk_vhost_gpa_to_vva(&svdev->vdev, desc->addr, sizeof(*desc_ev));
+ if (desc_ev == NULL) {
+ SPDK_ERRLOG("Controller %s: Eventq descriptor at index %"PRIu16" points to unmapped guest memory address %p.\n",
+ svdev->vdev.name, req, (void *)(uintptr_t)desc->addr);
+ goto out;
+ }
+
+ desc_ev->event = event;
+ desc_ev->lun[0] = 1;
+ desc_ev->lun[1] = scsi_dev_num;
+ /* virtio LUN id 0 can refer either to the entire device
+ * or actual LUN 0 (the only supported by vhost for now)
+ */
+ desc_ev->lun[2] = 0 >> 8;
+ desc_ev->lun[3] = 0 & 0xFF;
+ /* virtio doesn't specify any strict format for LUN id (bytes 2 and 3)
+ * current implementation relies on linux kernel sources
+ */
+ memset(&desc_ev->lun[4], 0, 4);
+ desc_ev->reason = reason;
+ req_size = sizeof(*desc_ev);
+
+out:
+ spdk_vhost_vq_used_ring_enqueue(&svdev->vdev, vq, req, req_size);
+}
+
+static void
+submit_completion(struct spdk_vhost_scsi_task *task)
+{
+ spdk_vhost_vq_used_ring_enqueue(&task->svdev->vdev, task->vq, task->req_idx,
+ task->used_len);
+ SPDK_DEBUGLOG(SPDK_LOG_VHOST_SCSI, "Finished task (%p) req_idx=%d\n", task, task->req_idx);
+
+ spdk_vhost_scsi_task_put(task);
+}
+
+static void
+spdk_vhost_scsi_task_mgmt_cpl(struct spdk_scsi_task *scsi_task)
+{
+ struct spdk_vhost_scsi_task *task = SPDK_CONTAINEROF(scsi_task, struct spdk_vhost_scsi_task, scsi);
+
+ submit_completion(task);
+}
+
+static void
+spdk_vhost_scsi_task_cpl(struct spdk_scsi_task *scsi_task)
+{
+ struct spdk_vhost_scsi_task *task = SPDK_CONTAINEROF(scsi_task, struct spdk_vhost_scsi_task, scsi);
+
+ /* The SCSI task has completed. Do final processing and then post
+ notification to the virtqueue's "used" ring.
+ */
+ task->resp->status = task->scsi.status;
+
+ if (task->scsi.status != SPDK_SCSI_STATUS_GOOD) {
+ memcpy(task->resp->sense, task->scsi.sense_data, task->scsi.sense_data_len);
+ task->resp->sense_len = task->scsi.sense_data_len;
+ SPDK_DEBUGLOG(SPDK_LOG_VHOST_SCSI, "Task (%p) req_idx=%d failed - status=%u\n", task, task->req_idx,
+ task->scsi.status);
+ }
+ assert(task->scsi.transfer_len == task->scsi.length);
+ task->resp->resid = task->scsi.length - task->scsi.data_transferred;
+
+ submit_completion(task);
+}
+
+static void
+task_submit(struct spdk_vhost_scsi_task *task)
+{
+ task->resp->response = VIRTIO_SCSI_S_OK;
+ spdk_scsi_dev_queue_task(task->scsi_dev, &task->scsi);
+}
+
+static void
+mgmt_task_submit(struct spdk_vhost_scsi_task *task, enum spdk_scsi_task_func func)
+{
+ task->tmf_resp->response = VIRTIO_SCSI_S_OK;
+ spdk_scsi_dev_queue_mgmt_task(task->scsi_dev, &task->scsi, func);
+}
+
+static void
+invalid_request(struct spdk_vhost_scsi_task *task)
+{
+ spdk_vhost_vq_used_ring_enqueue(&task->svdev->vdev, task->vq, task->req_idx,
+ task->used_len);
+ spdk_vhost_scsi_task_put(task);
+
+ SPDK_DEBUGLOG(SPDK_LOG_VHOST_SCSI, "Invalid request (status=%" PRIu8")\n",
+ task->resp ? task->resp->response : -1);
+}
+
+static int
+spdk_vhost_scsi_task_init_target(struct spdk_vhost_scsi_task *task, const __u8 *lun)
+{
+ struct spdk_scsi_dev *dev;
+ uint16_t lun_id = (((uint16_t)lun[2] << 8) | lun[3]) & 0x3FFF;
+
+ SPDK_TRACEDUMP(SPDK_LOG_VHOST_SCSI_QUEUE, "LUN", lun, 8);
+
+ /* First byte must be 1 and second is target */
+ if (lun[0] != 1 || lun[1] >= SPDK_VHOST_SCSI_CTRLR_MAX_DEVS) {
+ return -1;
+ }
+
+ dev = task->svdev->scsi_dev[lun[1]];
+ task->scsi_dev = dev;
+ if (dev == NULL || task->svdev->scsi_dev_state[lun[1]].removed) {
+ /* If dev has been hotdetached, return 0 to allow sending
+ * additional hotremove event via sense codes.
+ */
+ return task->svdev->scsi_dev_state[lun[1]].removed ? 0 : -1;
+ }
+
+ task->scsi.target_port = spdk_scsi_dev_find_port_by_id(task->scsi_dev, 0);
+ task->scsi.lun = spdk_scsi_dev_get_lun(dev, lun_id);
+ return 0;
+}
+
+static void
+process_ctrl_request(struct spdk_vhost_scsi_task *task)
+{
+ struct spdk_vhost_dev *vdev = &task->svdev->vdev;
+ struct vring_desc *desc, *desc_table;
+ struct virtio_scsi_ctrl_tmf_req *ctrl_req;
+ struct virtio_scsi_ctrl_an_resp *an_resp;
+ uint32_t desc_table_size, used_len = 0;
+ int rc;
+
+ spdk_scsi_task_construct(&task->scsi, spdk_vhost_scsi_task_mgmt_cpl, spdk_vhost_scsi_task_free_cb);
+ rc = spdk_vhost_vq_get_desc(vdev, task->vq, task->req_idx, &desc, &desc_table, &desc_table_size);
+ if (spdk_unlikely(rc != 0)) {
+ SPDK_ERRLOG("%s: Invalid controlq descriptor at index %d.\n",
+ vdev->name, task->req_idx);
+ goto out;
+ }
+
+ ctrl_req = spdk_vhost_gpa_to_vva(vdev, desc->addr, sizeof(*ctrl_req));
+ if (ctrl_req == NULL) {
+ SPDK_ERRLOG("%s: Invalid task management request at index %d.\n",
+ vdev->name, task->req_idx);
+ goto out;
+ }
+
+ SPDK_DEBUGLOG(SPDK_LOG_VHOST_SCSI_QUEUE,
+ "Processing controlq descriptor: desc %d/%p, desc_addr %p, len %d, flags %d, last_used_idx %d; kickfd %d; size %d\n",
+ task->req_idx, desc, (void *)desc->addr, desc->len, desc->flags, task->vq->vring.last_used_idx,
+ task->vq->vring.kickfd, task->vq->vring.size);
+ SPDK_TRACEDUMP(SPDK_LOG_VHOST_SCSI_QUEUE, "Request descriptor", (uint8_t *)ctrl_req,
+ desc->len);
+
+ spdk_vhost_scsi_task_init_target(task, ctrl_req->lun);
+
+ spdk_vhost_vring_desc_get_next(&desc, desc_table, desc_table_size);
+ if (spdk_unlikely(desc == NULL)) {
+ SPDK_ERRLOG("%s: No response descriptor for controlq request %d.\n",
+ vdev->name, task->req_idx);
+ goto out;
+ }
+
+ /* Process the TMF request */
+ switch (ctrl_req->type) {
+ case VIRTIO_SCSI_T_TMF:
+ task->tmf_resp = spdk_vhost_gpa_to_vva(vdev, desc->addr, sizeof(*task->tmf_resp));
+ if (spdk_unlikely(desc->len < sizeof(struct virtio_scsi_ctrl_tmf_resp) || task->tmf_resp == NULL)) {
+ SPDK_ERRLOG("%s: TMF response descriptor at index %d points to invalid guest memory region\n",
+ vdev->name, task->req_idx);
+ goto out;
+ }
+
+ /* Check if we are processing a valid request */
+ if (task->scsi_dev == NULL) {
+ task->tmf_resp->response = VIRTIO_SCSI_S_BAD_TARGET;
+ break;
+ }
+
+ switch (ctrl_req->subtype) {
+ case VIRTIO_SCSI_T_TMF_LOGICAL_UNIT_RESET:
+ /* Handle LUN reset */
+ SPDK_DEBUGLOG(SPDK_LOG_VHOST_SCSI_QUEUE, "LUN reset\n");
+
+ mgmt_task_submit(task, SPDK_SCSI_TASK_FUNC_LUN_RESET);
+ return;
+ default:
+ task->tmf_resp->response = VIRTIO_SCSI_S_ABORTED;
+ /* Unsupported command */
+ SPDK_DEBUGLOG(SPDK_LOG_VHOST_SCSI_QUEUE, "Unsupported TMF command %x\n", ctrl_req->subtype);
+ break;
+ }
+ break;
+ case VIRTIO_SCSI_T_AN_QUERY:
+ case VIRTIO_SCSI_T_AN_SUBSCRIBE: {
+ an_resp = spdk_vhost_gpa_to_vva(vdev, desc->addr, sizeof(*an_resp));
+ if (spdk_unlikely(desc->len < sizeof(struct virtio_scsi_ctrl_an_resp) || an_resp == NULL)) {
+ SPDK_WARNLOG("%s: Asynchronous response descriptor points to invalid guest memory region\n",
+ vdev->name);
+ goto out;
+ }
+
+ an_resp->response = VIRTIO_SCSI_S_ABORTED;
+ break;
+ }
+ default:
+ SPDK_DEBUGLOG(SPDK_LOG_VHOST_SCSI_QUEUE, "Unsupported control command %x\n", ctrl_req->type);
+ break;
+ }
+
+ used_len = sizeof(struct virtio_scsi_ctrl_tmf_resp);
+out:
+ spdk_vhost_vq_used_ring_enqueue(vdev, task->vq, task->req_idx, used_len);
+ spdk_vhost_scsi_task_put(task);
+}
+
+/*
+ * Process task's descriptor chain and setup data related fields.
+ * Return
+ * -1 if request is invalid and must be aborted,
+ * 0 if all data are set.
+ */
+static int
+task_data_setup(struct spdk_vhost_scsi_task *task,
+ struct virtio_scsi_cmd_req **req)
+{
+ struct spdk_vhost_dev *vdev = &task->svdev->vdev;
+ struct vring_desc *desc, *desc_table;
+ struct iovec *iovs = task->iovs;
+ uint16_t iovcnt = 0;
+ uint32_t desc_table_len, len = 0;
+ int rc;
+
+ spdk_scsi_task_construct(&task->scsi, spdk_vhost_scsi_task_cpl, spdk_vhost_scsi_task_free_cb);
+
+ rc = spdk_vhost_vq_get_desc(vdev, task->vq, task->req_idx, &desc, &desc_table, &desc_table_len);
+ /* First descriptor must be readable */
+ if (spdk_unlikely(rc != 0 || spdk_vhost_vring_desc_is_wr(desc) ||
+ desc->len < sizeof(struct virtio_scsi_cmd_req))) {
+ SPDK_WARNLOG("%s: invalid first (request) descriptor at index %"PRIu16".\n",
+ vdev->name, task->req_idx);
+ goto invalid_task;
+ }
+
+ *req = spdk_vhost_gpa_to_vva(vdev, desc->addr, sizeof(**req));
+ if (spdk_unlikely(*req == NULL)) {
+ SPDK_WARNLOG("%s: Request descriptor at index %d points to invalid guest memory region\n",
+ vdev->name, task->req_idx);
+ goto invalid_task;
+ }
+
+ /* Each request must have at least 2 descriptors (e.g. request and response) */
+ spdk_vhost_vring_desc_get_next(&desc, desc_table, desc_table_len);
+ if (desc == NULL) {
+ SPDK_WARNLOG("%s: Descriptor chain at index %d contains neither payload nor response buffer.\n",
+ vdev->name, task->req_idx);
+ goto invalid_task;
+ }
+ task->scsi.dxfer_dir = spdk_vhost_vring_desc_is_wr(desc) ? SPDK_SCSI_DIR_FROM_DEV :
+ SPDK_SCSI_DIR_TO_DEV;
+ task->scsi.iovs = iovs;
+
+ if (task->scsi.dxfer_dir == SPDK_SCSI_DIR_FROM_DEV) {
+ /*
+ * FROM_DEV (READ): [RD_req][WR_resp][WR_buf0]...[WR_bufN]
+ */
+ task->resp = spdk_vhost_gpa_to_vva(vdev, desc->addr, sizeof(*task->resp));
+ if (spdk_unlikely(desc->len < sizeof(struct virtio_scsi_cmd_resp) || task->resp == NULL)) {
+ SPDK_WARNLOG("%s: Response descriptor at index %d points to invalid guest memory region\n",
+ vdev->name, task->req_idx);
+ goto invalid_task;
+ }
+ rc = spdk_vhost_vring_desc_get_next(&desc, desc_table, desc_table_len);
+ if (spdk_unlikely(rc != 0)) {
+ SPDK_WARNLOG("%s: invalid descriptor chain at request index %d (descriptor id overflow?).\n",
+ vdev->name, task->req_idx);
+ goto invalid_task;
+ }
+
+ if (desc == NULL) {
+ /*
+ * TEST UNIT READY command and some others might not contain any payload and this is not an error.
+ */
+ SPDK_DEBUGLOG(SPDK_LOG_VHOST_SCSI_DATA,
+ "No payload descriptors for FROM DEV command req_idx=%"PRIu16".\n", task->req_idx);
+ SPDK_TRACEDUMP(SPDK_LOG_VHOST_SCSI_DATA, "CDB=", (*req)->cdb, VIRTIO_SCSI_CDB_SIZE);
+ task->used_len = sizeof(struct virtio_scsi_cmd_resp);
+ task->scsi.iovcnt = 1;
+ task->scsi.iovs[0].iov_len = 0;
+ task->scsi.length = 0;
+ task->scsi.transfer_len = 0;
+ return 0;
+ }
+
+ /* All remaining descriptors are data. */
+ while (desc) {
+ if (spdk_unlikely(!spdk_vhost_vring_desc_is_wr(desc))) {
+ SPDK_WARNLOG("FROM DEV cmd: descriptor nr %" PRIu16" in payload chain is read only.\n", iovcnt);
+ goto invalid_task;
+ }
+
+ if (spdk_unlikely(spdk_vhost_vring_desc_to_iov(vdev, iovs, &iovcnt, desc))) {
+ goto invalid_task;
+ }
+ len += desc->len;
+
+ rc = spdk_vhost_vring_desc_get_next(&desc, desc_table, desc_table_len);
+ if (spdk_unlikely(rc != 0)) {
+ SPDK_WARNLOG("%s: invalid payload in descriptor chain starting at index %d.\n",
+ vdev->name, task->req_idx);
+ goto invalid_task;
+ }
+ }
+
+ task->used_len = sizeof(struct virtio_scsi_cmd_resp) + len;
+ } else {
+ SPDK_DEBUGLOG(SPDK_LOG_VHOST_SCSI_DATA, "TO DEV");
+ /*
+ * TO_DEV (WRITE):[RD_req][RD_buf0]...[RD_bufN][WR_resp]
+ * No need to check descriptor WR flag as this is done while setting scsi.dxfer_dir.
+ */
+
+ /* Process descriptors up to response. */
+ while (!spdk_vhost_vring_desc_is_wr(desc)) {
+ if (spdk_unlikely(spdk_vhost_vring_desc_to_iov(vdev, iovs, &iovcnt, desc))) {
+ goto invalid_task;
+ }
+ len += desc->len;
+
+ spdk_vhost_vring_desc_get_next(&desc, desc_table, desc_table_len);
+ if (spdk_unlikely(desc == NULL)) {
+ SPDK_WARNLOG("TO_DEV cmd: no response descriptor.\n");
+ goto invalid_task;
+ }
+ }
+
+ task->resp = spdk_vhost_gpa_to_vva(vdev, desc->addr, sizeof(*task->resp));
+ if (spdk_unlikely(desc->len < sizeof(struct virtio_scsi_cmd_resp) || task->resp == NULL)) {
+ SPDK_WARNLOG("%s: Response descriptor at index %d points to invalid guest memory region\n",
+ vdev->name, task->req_idx);
+ goto invalid_task;
+ }
+
+ task->used_len = sizeof(struct virtio_scsi_cmd_resp);
+ }
+
+ task->scsi.iovcnt = iovcnt;
+ task->scsi.length = len;
+ task->scsi.transfer_len = len;
+ return 0;
+
+invalid_task:
+ SPDK_DEBUGLOG(SPDK_LOG_VHOST_SCSI_DATA, "%s: Invalid task at index %"PRIu16".\n",
+ vdev->name, task->req_idx);
+ return -1;
+}
+
+static int
+process_request(struct spdk_vhost_scsi_task *task)
+{
+ struct virtio_scsi_cmd_req *req;
+ int result;
+
+ result = task_data_setup(task, &req);
+ if (result) {
+ return result;
+ }
+
+ result = spdk_vhost_scsi_task_init_target(task, req->lun);
+ if (spdk_unlikely(result != 0)) {
+ task->resp->response = VIRTIO_SCSI_S_BAD_TARGET;
+ return -1;
+ }
+
+ task->scsi.cdb = req->cdb;
+ SPDK_TRACEDUMP(SPDK_LOG_VHOST_SCSI_DATA, "request CDB", req->cdb, VIRTIO_SCSI_CDB_SIZE);
+
+ if (spdk_unlikely(task->scsi.lun == NULL)) {
+ spdk_scsi_task_process_null_lun(&task->scsi);
+ task->resp->response = VIRTIO_SCSI_S_OK;
+ return 1;
+ }
+
+ return 0;
+}
+
+static void
+process_controlq(struct spdk_vhost_scsi_dev *svdev, struct spdk_vhost_virtqueue *vq)
+{
+ struct spdk_vhost_scsi_task *task;
+ uint16_t reqs[32];
+ uint16_t reqs_cnt, i;
+
+ reqs_cnt = spdk_vhost_vq_avail_ring_get(vq, reqs, SPDK_COUNTOF(reqs));
+ for (i = 0; i < reqs_cnt; i++) {
+ if (spdk_unlikely(reqs[i] >= vq->vring.size)) {
+ SPDK_ERRLOG("%s: invalid entry in avail ring. Buffer '%"PRIu16"' exceeds virtqueue size (%"PRIu16")\n",
+ svdev->vdev.name, reqs[i], vq->vring.size);
+ spdk_vhost_vq_used_ring_enqueue(&svdev->vdev, vq, reqs[i], 0);
+ continue;
+ }
+
+ task = &((struct spdk_vhost_scsi_task *)vq->tasks)[reqs[i]];
+ if (spdk_unlikely(task->used)) {
+ SPDK_ERRLOG("%s: invalid entry in avail ring. Buffer '%"PRIu16"' is still in use!\n",
+ svdev->vdev.name, reqs[i]);
+ spdk_vhost_vq_used_ring_enqueue(&svdev->vdev, vq, reqs[i], 0);
+ continue;
+ }
+
+ svdev->vdev.task_cnt++;
+ memset(&task->scsi, 0, sizeof(task->scsi));
+ task->tmf_resp = NULL;
+ task->used = true;
+ process_ctrl_request(task);
+ }
+}
+
+static void
+process_requestq(struct spdk_vhost_scsi_dev *svdev, struct spdk_vhost_virtqueue *vq)
+{
+ struct spdk_vhost_scsi_task *task;
+ uint16_t reqs[32];
+ uint16_t reqs_cnt, i;
+ int result;
+
+ reqs_cnt = spdk_vhost_vq_avail_ring_get(vq, reqs, SPDK_COUNTOF(reqs));
+ assert(reqs_cnt <= 32);
+
+ for (i = 0; i < reqs_cnt; i++) {
+ SPDK_DEBUGLOG(SPDK_LOG_VHOST_SCSI, "====== Starting processing request idx %"PRIu16"======\n",
+ reqs[i]);
+
+ if (spdk_unlikely(reqs[i] >= vq->vring.size)) {
+ SPDK_ERRLOG("%s: request idx '%"PRIu16"' exceeds virtqueue size (%"PRIu16").\n",
+ svdev->vdev.name, reqs[i], vq->vring.size);
+ spdk_vhost_vq_used_ring_enqueue(&svdev->vdev, vq, reqs[i], 0);
+ continue;
+ }
+
+ task = &((struct spdk_vhost_scsi_task *)vq->tasks)[reqs[i]];
+ if (spdk_unlikely(task->used)) {
+ SPDK_ERRLOG("%s: request with idx '%"PRIu16"' is already pending.\n",
+ svdev->vdev.name, reqs[i]);
+ spdk_vhost_vq_used_ring_enqueue(&svdev->vdev, vq, reqs[i], 0);
+ continue;
+ }
+
+ svdev->vdev.task_cnt++;
+ memset(&task->scsi, 0, sizeof(task->scsi));
+ task->resp = NULL;
+ task->used = true;
+ task->used_len = 0;
+ result = process_request(task);
+ if (likely(result == 0)) {
+ task_submit(task);
+ SPDK_DEBUGLOG(SPDK_LOG_VHOST_SCSI, "====== Task %p req_idx %d submitted ======\n", task,
+ task->req_idx);
+ } else if (result > 0) {
+ spdk_vhost_scsi_task_cpl(&task->scsi);
+ SPDK_DEBUGLOG(SPDK_LOG_VHOST_SCSI, "====== Task %p req_idx %d finished early ======\n", task,
+ task->req_idx);
+ } else {
+ invalid_request(task);
+ SPDK_DEBUGLOG(SPDK_LOG_VHOST_SCSI, "====== Task %p req_idx %d failed ======\n", task,
+ task->req_idx);
+ }
+ }
+}
+
+static int
+vdev_mgmt_worker(void *arg)
+{
+ struct spdk_vhost_scsi_dev *svdev = arg;
+
+ process_removed_devs(svdev);
+ spdk_vhost_vq_used_signal(&svdev->vdev, &svdev->vdev.virtqueue[VIRTIO_SCSI_EVENTQ]);
+
+ process_controlq(svdev, &svdev->vdev.virtqueue[VIRTIO_SCSI_CONTROLQ]);
+ spdk_vhost_vq_used_signal(&svdev->vdev, &svdev->vdev.virtqueue[VIRTIO_SCSI_CONTROLQ]);
+
+ return -1;
+}
+
+static int
+vdev_worker(void *arg)
+{
+ struct spdk_vhost_scsi_dev *svdev = arg;
+ uint32_t q_idx;
+
+ for (q_idx = VIRTIO_SCSI_REQUESTQ; q_idx < svdev->vdev.max_queues; q_idx++) {
+ process_requestq(svdev, &svdev->vdev.virtqueue[q_idx]);
+ }
+
+ spdk_vhost_dev_used_signal(&svdev->vdev);
+
+ return -1;
+}
+
+static struct spdk_vhost_scsi_dev *
+to_scsi_dev(struct spdk_vhost_dev *ctrlr)
+{
+ if (ctrlr == NULL) {
+ return NULL;
+ }
+
+ if (ctrlr->backend != &spdk_vhost_scsi_device_backend) {
+ SPDK_ERRLOG("%s: not a vhost-scsi device.\n", ctrlr->name);
+ return NULL;
+ }
+
+ return SPDK_CONTAINEROF(ctrlr, struct spdk_vhost_scsi_dev, vdev);
+}
+
+int
+spdk_vhost_scsi_dev_construct(const char *name, const char *cpumask)
+{
+ struct spdk_vhost_scsi_dev *svdev = spdk_dma_zmalloc(sizeof(struct spdk_vhost_scsi_dev),
+ SPDK_CACHE_LINE_SIZE, NULL);
+ int rc;
+
+ if (svdev == NULL) {
+ return -ENOMEM;
+ }
+
+ spdk_vhost_lock();
+ rc = spdk_vhost_dev_register(&svdev->vdev, name, cpumask,
+ &spdk_vhost_scsi_device_backend);
+
+ if (rc) {
+ spdk_dma_free(svdev);
+ }
+
+ spdk_vhost_unlock();
+ return rc;
+}
+
+static int
+spdk_vhost_scsi_dev_remove(struct spdk_vhost_dev *vdev)
+{
+ struct spdk_vhost_scsi_dev *svdev = to_scsi_dev(vdev);
+ int rc, i;
+
+ if (svdev == NULL) {
+ return -EINVAL;
+ }
+
+ for (i = 0; i < SPDK_VHOST_SCSI_CTRLR_MAX_DEVS; ++i) {
+ if (svdev->scsi_dev[i]) {
+ if (vdev->registered) {
+ SPDK_ERRLOG("Trying to remove non-empty controller: %s.\n", vdev->name);
+ return -EBUSY;
+ }
+
+ rc = spdk_vhost_scsi_dev_remove_tgt(vdev, i, NULL, NULL);
+ if (rc != 0) {
+ SPDK_ERRLOG("%s: failed to force-remove target %d\n", vdev->name, i);
+ return rc;
+ }
+ }
+ }
+
+ rc = spdk_vhost_dev_unregister(vdev);
+ if (rc != 0) {
+ return rc;
+ }
+
+ spdk_dma_free(svdev);
+ return 0;
+}
+
+struct spdk_scsi_dev *
+spdk_vhost_scsi_dev_get_tgt(struct spdk_vhost_dev *vdev, uint8_t num)
+{
+ struct spdk_vhost_scsi_dev *svdev;
+
+ assert(num < SPDK_VHOST_SCSI_CTRLR_MAX_DEVS);
+ svdev = to_scsi_dev(vdev);
+
+ return svdev ? svdev->scsi_dev[num] : NULL;
+}
+
+static void
+spdk_vhost_scsi_lun_hotremove(const struct spdk_scsi_lun *lun, void *arg)
+{
+ struct spdk_vhost_scsi_dev *svdev = arg;
+ const struct spdk_scsi_dev *scsi_dev;
+ unsigned scsi_dev_num;
+
+ assert(lun != NULL);
+ assert(svdev != NULL);
+ if (svdev->vdev.lcore != -1 &&
+ !spdk_vhost_dev_has_feature(&svdev->vdev, VIRTIO_SCSI_F_HOTPLUG)) {
+ SPDK_WARNLOG("%s: hotremove is not enabled for this controller.\n", svdev->vdev.name);
+ return;
+ }
+
+ scsi_dev = spdk_scsi_lun_get_dev(lun);
+ for (scsi_dev_num = 0; scsi_dev_num < SPDK_VHOST_SCSI_CTRLR_MAX_DEVS; scsi_dev_num++) {
+ if (svdev->scsi_dev[scsi_dev_num] == scsi_dev) {
+ break;
+ }
+ }
+
+ if (scsi_dev_num == SPDK_VHOST_SCSI_CTRLR_MAX_DEVS) {
+ /* The entire device has been already removed. */
+ return;
+ }
+
+ /* remove entire device */
+ spdk_vhost_scsi_dev_remove_tgt(&svdev->vdev, scsi_dev_num, NULL, NULL);
+}
+
+int
+spdk_vhost_scsi_dev_add_tgt(struct spdk_vhost_dev *vdev, unsigned scsi_tgt_num,
+ const char *bdev_name)
+{
+ struct spdk_vhost_scsi_dev *svdev;
+ char target_name[SPDK_SCSI_DEV_MAX_NAME];
+ int lun_id_list[1];
+ const char *bdev_names_list[1];
+
+ svdev = to_scsi_dev(vdev);
+ if (svdev == NULL) {
+ return -EINVAL;
+ }
+
+ if (scsi_tgt_num >= SPDK_VHOST_SCSI_CTRLR_MAX_DEVS) {
+ SPDK_ERRLOG("Controller %d target number too big (max %d)\n", scsi_tgt_num,
+ SPDK_VHOST_SCSI_CTRLR_MAX_DEVS);
+ return -EINVAL;
+ }
+
+ if (bdev_name == NULL) {
+ SPDK_ERRLOG("No lun name specified\n");
+ return -EINVAL;
+ }
+
+ if (svdev->scsi_dev[scsi_tgt_num] != NULL) {
+ SPDK_ERRLOG("Controller %s target %u already occupied\n", vdev->name, scsi_tgt_num);
+ return -EEXIST;
+ }
+
+ /*
+ * At this stage only one LUN per target
+ */
+ snprintf(target_name, sizeof(target_name), "Target %u", scsi_tgt_num);
+ lun_id_list[0] = 0;
+ bdev_names_list[0] = (char *)bdev_name;
+
+ svdev->scsi_dev_state[scsi_tgt_num].removed = false;
+ svdev->scsi_dev[scsi_tgt_num] = spdk_scsi_dev_construct(target_name, bdev_names_list, lun_id_list,
+ 1,
+ SPDK_SPC_PROTOCOL_IDENTIFIER_SAS, spdk_vhost_scsi_lun_hotremove, svdev);
+
+ if (svdev->scsi_dev[scsi_tgt_num] == NULL) {
+ SPDK_ERRLOG("Couldn't create spdk SCSI target '%s' using bdev '%s' in controller: %s\n",
+ target_name, bdev_name, vdev->name);
+ return -EINVAL;
+ }
+ spdk_scsi_dev_add_port(svdev->scsi_dev[scsi_tgt_num], 0, "vhost");
+
+ SPDK_INFOLOG(SPDK_LOG_VHOST, "Controller %s: defined target '%s' using bdev '%s'\n",
+ vdev->name, target_name, bdev_name);
+
+ if (vdev->lcore == -1) {
+ /* All done. */
+ return 0;
+ }
+
+ spdk_scsi_dev_allocate_io_channels(svdev->scsi_dev[scsi_tgt_num]);
+
+ if (spdk_vhost_dev_has_feature(vdev, VIRTIO_SCSI_F_HOTPLUG)) {
+ eventq_enqueue(svdev, scsi_tgt_num, VIRTIO_SCSI_T_TRANSPORT_RESET,
+ VIRTIO_SCSI_EVT_RESET_RESCAN);
+ } else {
+ SPDK_NOTICELOG("Device %s does not support hotplug. "
+ "Please restart the driver or perform a rescan.\n",
+ vdev->name);
+ }
+
+ return 0;
+}
+
+int
+spdk_vhost_scsi_dev_remove_tgt(struct spdk_vhost_dev *vdev, unsigned scsi_tgt_num,
+ spdk_vhost_event_fn cb_fn, void *cb_arg)
+{
+ struct spdk_vhost_scsi_dev *svdev;
+ struct spdk_scsi_dev *scsi_dev;
+ struct spdk_scsi_dev_vhost_state *scsi_dev_state;
+ int rc = 0;
+
+ if (scsi_tgt_num >= SPDK_VHOST_SCSI_CTRLR_MAX_DEVS) {
+ SPDK_ERRLOG("%s: invalid target number %d\n", vdev->name, scsi_tgt_num);
+ return -EINVAL;
+ }
+
+ svdev = to_scsi_dev(vdev);
+ if (svdev == NULL) {
+ return -ENODEV;
+ }
+
+ scsi_dev = svdev->scsi_dev[scsi_tgt_num];
+ if (scsi_dev == NULL) {
+ SPDK_ERRLOG("Controller %s target %u is not occupied\n", vdev->name, scsi_tgt_num);
+ return -ENODEV;
+ }
+
+ if (svdev->vdev.lcore == -1) {
+ /* controller is not in use, remove dev and exit */
+ svdev->scsi_dev[scsi_tgt_num] = NULL;
+ spdk_scsi_dev_destruct(scsi_dev);
+ if (cb_fn) {
+ rc = cb_fn(vdev, cb_arg);
+ }
+ SPDK_INFOLOG(SPDK_LOG_VHOST, "%s: removed target 'Target %u'\n",
+ vdev->name, scsi_tgt_num);
+ return rc;
+ }
+
+ if (!spdk_vhost_dev_has_feature(vdev, VIRTIO_SCSI_F_HOTPLUG)) {
+ SPDK_WARNLOG("%s: 'Target %u' is in use and hot-detach is not enabled for this controller.\n",
+ svdev->vdev.name, scsi_tgt_num);
+ return -ENOTSUP;
+ }
+
+ scsi_dev_state = &svdev->scsi_dev_state[scsi_tgt_num];
+ if (scsi_dev_state->removed) {
+ SPDK_WARNLOG("%s: 'Target %u' has been already marked to hotremove.\n", svdev->vdev.name,
+ scsi_tgt_num);
+ return -EBUSY;
+ }
+
+ scsi_dev_state->remove_cb = cb_fn;
+ scsi_dev_state->remove_ctx = cb_arg;
+ scsi_dev_state->removed = true;
+ eventq_enqueue(svdev, scsi_tgt_num, VIRTIO_SCSI_T_TRANSPORT_RESET, VIRTIO_SCSI_EVT_RESET_REMOVED);
+
+ SPDK_INFOLOG(SPDK_LOG_VHOST, "%s: queued 'Target %u' for hot-detach.\n", vdev->name, scsi_tgt_num);
+ return 0;
+}
+
+int
+spdk_vhost_scsi_controller_construct(void)
+{
+ struct spdk_conf_section *sp = spdk_conf_first_section(NULL);
+ struct spdk_vhost_dev *vdev;
+ int i, dev_num;
+ unsigned ctrlr_num = 0;
+ char *bdev_name, *tgt_num_str;
+ char *cpumask;
+ char *name;
+ char *tgt = NULL;
+
+ while (sp != NULL) {
+ if (!spdk_conf_section_match_prefix(sp, "VhostScsi")) {
+ sp = spdk_conf_next_section(sp);
+ continue;
+ }
+
+ if (sscanf(spdk_conf_section_get_name(sp), "VhostScsi%u", &ctrlr_num) != 1) {
+ SPDK_ERRLOG("Section '%s' has non-numeric suffix.\n",
+ spdk_conf_section_get_name(sp));
+ return -1;
+ }
+
+ name = spdk_conf_section_get_val(sp, "Name");
+ cpumask = spdk_conf_section_get_val(sp, "Cpumask");
+
+ if (spdk_vhost_scsi_dev_construct(name, cpumask) < 0) {
+ return -1;
+ }
+
+ vdev = spdk_vhost_dev_find(name);
+ assert(vdev);
+
+ for (i = 0; ; i++) {
+
+ tgt = spdk_conf_section_get_nval(sp, "Target", i);
+ if (tgt == NULL) {
+ break;
+ }
+
+ tgt_num_str = spdk_conf_section_get_nmval(sp, "Target", i, 0);
+ if (tgt_num_str == NULL) {
+ SPDK_ERRLOG("%s: Invalid or missing target number\n", name);
+ return -1;
+ }
+
+ dev_num = (int)strtol(tgt_num_str, NULL, 10);
+ bdev_name = spdk_conf_section_get_nmval(sp, "Target", i, 1);
+ if (bdev_name == NULL) {
+ SPDK_ERRLOG("%s: Invalid or missing bdev name for target %d\n", name, dev_num);
+ return -1;
+ } else if (spdk_conf_section_get_nmval(sp, "Target", i, 2)) {
+ SPDK_ERRLOG("%s: Only one LUN per vhost SCSI device supported\n", name);
+ return -1;
+ }
+
+ if (spdk_vhost_scsi_dev_add_tgt(vdev, dev_num, bdev_name) < 0) {
+ return -1;
+ }
+ }
+
+ sp = spdk_conf_next_section(sp);
+ }
+
+ return 0;
+}
+
+static void
+free_task_pool(struct spdk_vhost_scsi_dev *svdev)
+{
+ struct spdk_vhost_virtqueue *vq;
+ uint16_t i;
+
+ for (i = 0; i < svdev->vdev.max_queues; i++) {
+ vq = &svdev->vdev.virtqueue[i];
+ if (vq->tasks == NULL) {
+ continue;
+ }
+
+ spdk_dma_free(vq->tasks);
+ vq->tasks = NULL;
+ }
+}
+
+static int
+alloc_task_pool(struct spdk_vhost_scsi_dev *svdev)
+{
+ struct spdk_vhost_virtqueue *vq;
+ struct spdk_vhost_scsi_task *task;
+ uint32_t task_cnt;
+ uint16_t i;
+ uint32_t j;
+
+ for (i = 0; i < svdev->vdev.max_queues; i++) {
+ vq = &svdev->vdev.virtqueue[i];
+ if (vq->vring.desc == NULL) {
+ continue;
+ }
+
+ task_cnt = vq->vring.size;
+ if (task_cnt > SPDK_VHOST_MAX_VQ_SIZE) {
+ /* sanity check */
+ SPDK_ERRLOG("Controller %s: virtuque %"PRIu16" is too big. (size = %"PRIu32", max = %"PRIu32")\n",
+ svdev->vdev.name, i, task_cnt, SPDK_VHOST_MAX_VQ_SIZE);
+ free_task_pool(svdev);
+ return -1;
+ }
+ vq->tasks = spdk_dma_zmalloc(sizeof(struct spdk_vhost_scsi_task) * task_cnt,
+ SPDK_CACHE_LINE_SIZE, NULL);
+ if (vq->tasks == NULL) {
+ SPDK_ERRLOG("Controller %s: failed to allocate %"PRIu32" tasks for virtqueue %"PRIu16"\n",
+ svdev->vdev.name, task_cnt, i);
+ free_task_pool(svdev);
+ return -1;
+ }
+
+ for (j = 0; j < task_cnt; j++) {
+ task = &((struct spdk_vhost_scsi_task *)vq->tasks)[j];
+ task->svdev = svdev;
+ task->vq = vq;
+ task->req_idx = j;
+ }
+ }
+
+ return 0;
+}
+
+/*
+ * A new device is added to a data core. First the device is added to the main linked list
+ * and then allocated to a specific data core.
+ */
+static int
+spdk_vhost_scsi_start(struct spdk_vhost_dev *vdev, void *event_ctx)
+{
+ struct spdk_vhost_scsi_dev *svdev;
+ uint32_t i;
+ int rc;
+
+ svdev = to_scsi_dev(vdev);
+ if (svdev == NULL) {
+ SPDK_ERRLOG("Trying to start non-scsi controller as a scsi one.\n");
+ rc = -1;
+ goto out;
+ }
+
+ /* validate all I/O queues are in a contiguous index range */
+ for (i = VIRTIO_SCSI_REQUESTQ; i < vdev->max_queues; i++) {
+ if (vdev->virtqueue[i].vring.desc == NULL) {
+ SPDK_ERRLOG("%s: queue %"PRIu32" is empty\n", vdev->name, i);
+ rc = -1;
+ goto out;
+ }
+ }
+
+ rc = alloc_task_pool(svdev);
+ if (rc != 0) {
+ SPDK_ERRLOG("%s: failed to alloc task pool.\n", vdev->name);
+ goto out;
+ }
+
+ for (i = 0; i < SPDK_VHOST_SCSI_CTRLR_MAX_DEVS; i++) {
+ if (svdev->scsi_dev[i] == NULL) {
+ continue;
+ }
+ spdk_scsi_dev_allocate_io_channels(svdev->scsi_dev[i]);
+ }
+ SPDK_INFOLOG(SPDK_LOG_VHOST, "Started poller for vhost controller %s on lcore %d\n",
+ vdev->name, vdev->lcore);
+
+ svdev->requestq_poller = spdk_poller_register(vdev_worker, svdev, 0);
+ if (vdev->virtqueue[VIRTIO_SCSI_CONTROLQ].vring.desc &&
+ vdev->virtqueue[VIRTIO_SCSI_EVENTQ].vring.desc) {
+ svdev->mgmt_poller = spdk_poller_register(vdev_mgmt_worker, svdev,
+ MGMT_POLL_PERIOD_US);
+ }
+out:
+ spdk_vhost_dev_backend_event_done(event_ctx, rc);
+ return rc;
+}
+
+static int
+destroy_device_poller_cb(void *arg)
+{
+ struct spdk_vhost_scsi_dev *svdev = arg;
+ uint32_t i;
+
+ if (svdev->vdev.task_cnt > 0) {
+ return -1;
+ }
+
+
+ for (i = 0; i < svdev->vdev.max_queues; i++) {
+ spdk_vhost_vq_used_signal(&svdev->vdev, &svdev->vdev.virtqueue[i]);
+ }
+
+ for (i = 0; i < SPDK_VHOST_SCSI_CTRLR_MAX_DEVS; i++) {
+ if (svdev->scsi_dev[i] == NULL) {
+ continue;
+ }
+ spdk_scsi_dev_free_io_channels(svdev->scsi_dev[i]);
+ }
+
+ SPDK_INFOLOG(SPDK_LOG_VHOST, "Stopping poller for vhost controller %s\n", svdev->vdev.name);
+
+ free_task_pool(svdev);
+
+ spdk_poller_unregister(&svdev->destroy_ctx.poller);
+ spdk_vhost_dev_backend_event_done(svdev->destroy_ctx.event_ctx, 0);
+
+ return -1;
+}
+
+static int
+spdk_vhost_scsi_stop(struct spdk_vhost_dev *vdev, void *event_ctx)
+{
+ struct spdk_vhost_scsi_dev *svdev;
+
+ svdev = to_scsi_dev(vdev);
+ if (svdev == NULL) {
+ SPDK_ERRLOG("Trying to stop non-scsi controller as a scsi one.\n");
+ goto err;
+ }
+
+ svdev->destroy_ctx.event_ctx = event_ctx;
+ spdk_poller_unregister(&svdev->requestq_poller);
+ spdk_poller_unregister(&svdev->mgmt_poller);
+ svdev->destroy_ctx.poller = spdk_poller_register(destroy_device_poller_cb, svdev,
+ 1000);
+
+ return 0;
+
+err:
+ spdk_vhost_dev_backend_event_done(event_ctx, -1);
+ return -1;
+}
+
+static void
+spdk_vhost_scsi_dump_info_json(struct spdk_vhost_dev *vdev, struct spdk_json_write_ctx *w)
+{
+ struct spdk_scsi_dev *sdev;
+ struct spdk_scsi_lun *lun;
+ uint32_t dev_idx;
+ uint32_t lun_idx;
+
+ assert(vdev != NULL);
+ spdk_json_write_name(w, "scsi");
+ spdk_json_write_array_begin(w);
+ for (dev_idx = 0; dev_idx < SPDK_VHOST_SCSI_CTRLR_MAX_DEVS; dev_idx++) {
+ sdev = spdk_vhost_scsi_dev_get_tgt(vdev, dev_idx);
+ if (!sdev) {
+ continue;
+ }
+
+ spdk_json_write_object_begin(w);
+
+ spdk_json_write_name(w, "scsi_dev_num");
+ spdk_json_write_uint32(w, dev_idx);
+
+ spdk_json_write_name(w, "id");
+ spdk_json_write_int32(w, spdk_scsi_dev_get_id(sdev));
+
+ spdk_json_write_name(w, "target_name");
+ spdk_json_write_string(w, spdk_scsi_dev_get_name(sdev));
+
+ spdk_json_write_name(w, "luns");
+ spdk_json_write_array_begin(w);
+
+ for (lun_idx = 0; lun_idx < SPDK_SCSI_DEV_MAX_LUN; lun_idx++) {
+ lun = spdk_scsi_dev_get_lun(sdev, lun_idx);
+ if (!lun) {
+ continue;
+ }
+
+ spdk_json_write_object_begin(w);
+
+ spdk_json_write_name(w, "id");
+ spdk_json_write_int32(w, spdk_scsi_lun_get_id(lun));
+
+ spdk_json_write_name(w, "bdev_name");
+ spdk_json_write_string(w, spdk_scsi_lun_get_bdev_name(lun));
+
+ spdk_json_write_object_end(w);
+ }
+
+ spdk_json_write_array_end(w);
+ spdk_json_write_object_end(w);
+ }
+
+ spdk_json_write_array_end(w);
+}
+
+static void
+spdk_vhost_scsi_write_config_json(struct spdk_vhost_dev *vdev, struct spdk_json_write_ctx *w)
+{
+ struct spdk_vhost_scsi_dev *svdev;
+ struct spdk_scsi_lun *lun;
+ uint32_t i;
+
+ svdev = to_scsi_dev(vdev);
+ if (!svdev) {
+ return;
+ }
+
+ spdk_json_write_object_begin(w);
+ spdk_json_write_named_string(w, "method", "construct_vhost_scsi_controller");
+
+ spdk_json_write_named_object_begin(w, "params");
+ spdk_json_write_named_string(w, "ctrlr", vdev->name);
+ spdk_json_write_named_string(w, "cpumask", spdk_cpuset_fmt(vdev->cpumask));
+ spdk_json_write_object_end(w);
+
+ spdk_json_write_object_end(w);
+
+ for (i = 0; i < SPDK_COUNTOF(svdev->scsi_dev); i++) {
+ if (svdev->scsi_dev[i] == NULL || svdev->scsi_dev_state[i].removed) {
+ continue;
+ }
+
+ lun = spdk_scsi_dev_get_lun(svdev->scsi_dev[i], 0);
+
+ spdk_json_write_object_begin(w);
+ spdk_json_write_named_string(w, "method", "add_vhost_scsi_lun");
+
+ spdk_json_write_named_object_begin(w, "params");
+ spdk_json_write_named_string(w, "ctrlr", vdev->name);
+ spdk_json_write_named_uint32(w, "scsi_target_num", i);
+
+ spdk_json_write_named_string(w, "bdev_name", spdk_scsi_lun_get_bdev_name(lun));
+ spdk_json_write_object_end(w);
+
+ spdk_json_write_object_end(w);
+ }
+}
+
+SPDK_LOG_REGISTER_COMPONENT("vhost_scsi", SPDK_LOG_VHOST_SCSI)
+SPDK_LOG_REGISTER_COMPONENT("vhost_scsi_queue", SPDK_LOG_VHOST_SCSI_QUEUE)
+SPDK_LOG_REGISTER_COMPONENT("vhost_scsi_data", SPDK_LOG_VHOST_SCSI_DATA)