Page Menu
Home
FreeBSD
Search
Configure Global Search
Log In
Files
F145076649
D53469.diff
No One
Temporary
Actions
View File
Edit File
Delete File
View Transforms
Subscribe
Mute Notifications
Flag For Later
Award Token
Size
17 KB
Referenced Files
None
Subscribers
None
D53469.diff
View Options
diff --git a/usr.sbin/bhyve/pci_virtio_scsi.c b/usr.sbin/bhyve/pci_virtio_scsi.c
--- a/usr.sbin/bhyve/pci_virtio_scsi.c
+++ b/usr.sbin/bhyve/pci_virtio_scsi.c
@@ -102,13 +102,17 @@
uint32_t max_lun;
} __attribute__((packed));
+STAILQ_HEAD(pci_vtscsi_req_queue, pci_vtscsi_request);
+
struct pci_vtscsi_queue {
struct pci_vtscsi_softc * vsq_sc;
struct vqueue_info * vsq_vq;
- pthread_mutex_t vsq_mtx;
+ pthread_mutex_t vsq_rmtx;
+ pthread_mutex_t vsq_fmtx;
pthread_mutex_t vsq_qmtx;
pthread_cond_t vsq_cv;
- STAILQ_HEAD(, pci_vtscsi_request) vsq_requests;
+ struct pci_vtscsi_req_queue vsq_requests;
+ struct pci_vtscsi_req_queue vsq_free_requests;
LIST_HEAD(, pci_vtscsi_worker) vsq_workers;
};
@@ -124,8 +128,15 @@
struct iovec vsr_iov[VTSCSI_MAXSEG + SPLIT_IOV_ADDL_IOV];
struct iovec * vsr_iov_in;
struct iovec * vsr_iov_out;
+ struct iovec * vsr_data_iov_in;
+ struct iovec * vsr_data_iov_out;
+ struct pci_vtscsi_req_cmd_rd * vsr_cmd_rd;
+ struct pci_vtscsi_req_cmd_wr * vsr_cmd_wr;
+ union ctl_io * vsr_ctl_io;
size_t vsr_niov_in;
size_t vsr_niov_out;
+ size_t vsr_data_niov_in;
+ size_t vsr_data_niov_out;
uint32_t vsr_idx;
STAILQ_ENTRY(pci_vtscsi_request) vsr_link;
};
@@ -237,8 +248,21 @@
struct pci_vtscsi_ctrl_tmf *);
static void pci_vtscsi_an_handle(struct pci_vtscsi_softc *,
struct pci_vtscsi_ctrl_an *);
-static int pci_vtscsi_request_handle(struct pci_vtscsi_queue *, struct iovec *,
- size_t, struct iovec *, size_t);
+
+static struct pci_vtscsi_request *pci_vtscsi_alloc_request(
+ struct pci_vtscsi_softc *);
+static void pci_vtscsi_free_request(struct pci_vtscsi_request *);
+static struct pci_vtscsi_request *pci_vtscsi_get_request(
+ struct pci_vtscsi_req_queue *);
+static void pci_vtscsi_put_request(struct pci_vtscsi_req_queue *,
+ struct pci_vtscsi_request *);
+static void pci_vtscsi_queue_request(struct pci_vtscsi_softc *,
+ struct vqueue_info *);
+static void pci_vtscsi_return_request(struct pci_vtscsi_queue *,
+ struct pci_vtscsi_request *, int);
+static int pci_vtscsi_request_handle(struct pci_vtscsi_softc *,
+ struct pci_vtscsi_request *);
+
static void pci_vtscsi_controlq_notify(void *, struct vqueue_info *);
static void pci_vtscsi_eventq_notify(void *, struct vqueue_info *);
static void pci_vtscsi_requestq_notify(void *, struct vqueue_info *);
@@ -262,37 +286,33 @@
{
struct pci_vtscsi_worker *worker = (struct pci_vtscsi_worker *)arg;
struct pci_vtscsi_queue *q = worker->vsw_queue;
- struct pci_vtscsi_request *req;
+ struct pci_vtscsi_softc *sc = q->vsq_sc;
int iolen;
for (;;) {
- pthread_mutex_lock(&q->vsq_mtx);
+ struct pci_vtscsi_request *req;
- while (STAILQ_EMPTY(&q->vsq_requests)
- && !worker->vsw_exiting)
- pthread_cond_wait(&q->vsq_cv, &q->vsq_mtx);
+ pthread_mutex_lock(&q->vsq_rmtx);
- if (worker->vsw_exiting)
- break;
+ while (STAILQ_EMPTY(&q->vsq_requests) && !worker->vsw_exiting)
+ pthread_cond_wait(&q->vsq_cv, &q->vsq_rmtx);
- req = STAILQ_FIRST(&q->vsq_requests);
- STAILQ_REMOVE_HEAD(&q->vsq_requests, vsr_link);
+ if (worker->vsw_exiting) {
+ pthread_mutex_unlock(&q->vsq_rmtx);
+ return (NULL);
+ }
- pthread_mutex_unlock(&q->vsq_mtx);
- iolen = pci_vtscsi_request_handle(q, req->vsr_iov_in,
- req->vsr_niov_in, req->vsr_iov_out, req->vsr_niov_out);
+ req = pci_vtscsi_get_request(&q->vsq_requests);
+ pthread_mutex_unlock(&q->vsq_rmtx);
- pthread_mutex_lock(&q->vsq_qmtx);
- vq_relchain(q->vsq_vq, req->vsr_idx, iolen);
- vq_endchains(q->vsq_vq, 0);
- pthread_mutex_unlock(&q->vsq_qmtx);
+ DPRINTF("I/O request lun %d, data_niov_in %zu, data_niov_out "
+ "%zu", pci_vtscsi_get_lun(req->vsr_cmd_rd->lun),
+ req->vsr_data_niov_in, req->vsr_data_niov_out);
- DPRINTF("request <idx=%d> completed", req->vsr_idx);
- free(req);
- }
+ iolen = pci_vtscsi_request_handle(sc, req);
- pthread_mutex_unlock(&q->vsq_mtx);
- return (NULL);
+ pci_vtscsi_return_request(q, req, iolen);
+ }
}
static void
@@ -395,6 +415,14 @@
int err;
io = ctl_scsi_alloc_io(sc->vss_iid);
+ if (io == NULL) {
+ WPRINTF("failed to allocate ctl_io: err=%d (%s)",
+ errno, strerror(errno));
+
+ tmf->response = VIRTIO_SCSI_S_FAILURE;
+ return;
+ }
+
ctl_scsi_zero_io(io);
io->io_hdr.io_type = CTL_IO_TASK;
@@ -460,34 +488,112 @@
{
}
-static int
-pci_vtscsi_request_handle(struct pci_vtscsi_queue *q, struct iovec *iov_in,
- size_t niov_in, struct iovec *iov_out, size_t niov_out)
+static struct pci_vtscsi_request *
+pci_vtscsi_alloc_request(struct pci_vtscsi_softc *sc)
{
- struct pci_vtscsi_softc *sc = q->vsq_sc;
- struct pci_vtscsi_req_cmd_rd *cmd_rd = NULL;
- struct pci_vtscsi_req_cmd_wr *cmd_wr;
- struct iovec *data_iov_in, *data_iov_out;
- union ctl_io *io;
- size_t data_niov_in, data_niov_out;
- void *ext_data_ptr = NULL;
- uint32_t ext_data_len = 0, ext_sg_entries = 0;
- int err, nxferred;
+ struct pci_vtscsi_request *req;
+
+ req = calloc(1, sizeof(struct pci_vtscsi_request));
+ if (req == NULL)
+ goto alloc_fail;
+
+ req->vsr_cmd_rd = calloc(1, VTSCSI_IN_HEADER_LEN(sc));
+ if (req->vsr_cmd_rd == NULL)
+ goto alloc_fail;
+ req->vsr_cmd_wr = calloc(1, VTSCSI_OUT_HEADER_LEN(sc));
+ if (req->vsr_cmd_wr == NULL)
+ goto alloc_fail;
+
+ req->vsr_ctl_io = ctl_scsi_alloc_io(sc->vss_iid);
+ if (req->vsr_ctl_io == NULL)
+ goto alloc_fail;
+ ctl_scsi_zero_io(req->vsr_ctl_io);
+
+ return (req);
+
+alloc_fail:
+ EPRINTLN("failed to allocate request: %s", strerror(errno));
+
+ if (req != NULL)
+ pci_vtscsi_free_request(req);
+
+ return (NULL);
+}
+
+static void
+pci_vtscsi_free_request(struct pci_vtscsi_request *req)
+{
+ if (req->vsr_ctl_io != NULL)
+ ctl_scsi_free_io(req->vsr_ctl_io);
+ if (req->vsr_cmd_rd != NULL)
+ free(req->vsr_cmd_rd);
+ if (req->vsr_cmd_wr != NULL)
+ free(req->vsr_cmd_wr);
+
+ free(req);
+}
+
+static struct pci_vtscsi_request *
+pci_vtscsi_get_request(struct pci_vtscsi_req_queue *req_queue)
+{
+ struct pci_vtscsi_request *req;
+
+ assert(!STAILQ_EMPTY(req_queue));
+
+ req = STAILQ_FIRST(req_queue);
+ STAILQ_REMOVE_HEAD(req_queue, vsr_link);
+
+ return (req);
+}
+
+static void
+pci_vtscsi_put_request(struct pci_vtscsi_req_queue *req_queue,
+ struct pci_vtscsi_request *req)
+{
+ STAILQ_INSERT_TAIL(req_queue, req, vsr_link);
+}
+
+static void
+pci_vtscsi_queue_request(struct pci_vtscsi_softc *sc, struct vqueue_info *vq)
+{
+ struct pci_vtscsi_queue *q = &sc->vss_queues[vq->vq_num - 2];
+ struct pci_vtscsi_request *req;
+ struct vi_req vireq;
+ int n;
+
+ pthread_mutex_lock(&q->vsq_fmtx);
+ req = pci_vtscsi_get_request(&q->vsq_free_requests);
+ assert(req != NULL);
+ pthread_mutex_unlock(&q->vsq_fmtx);
+
+ n = vq_getchain(vq, req->vsr_iov, VTSCSI_MAXSEG, &vireq);
+ assert(n >= 1 && n <= VTSCSI_MAXSEG);
+
+ req->vsr_idx = vireq.idx;
+ req->vsr_queue = q;
+ req->vsr_iov_in = &req->vsr_iov[0];
+ req->vsr_niov_in = vireq.readable;
+ req->vsr_iov_out = &req->vsr_iov[vireq.readable];
+ req->vsr_niov_out = vireq.writable;
/*
* Make sure we got at least enough space for the VirtIO-SCSI
* command headers. If not, return this request immediately.
*/
- if (check_iov_len(iov_out, niov_out,
+ if (check_iov_len(req->vsr_iov_out, req->vsr_niov_out,
VTSCSI_OUT_HEADER_LEN(q->vsq_sc)) == false) {
WPRINTF("ignoring request with insufficient output");
- return (0);
+ req->vsr_cmd_wr->response = VIRTIO_SCSI_S_FAILURE;
+ pci_vtscsi_return_request(q, req, 1);
+ return;
}
- if (check_iov_len(iov_in, niov_in,
+ if (check_iov_len(req->vsr_iov_in, req->vsr_niov_in,
VTSCSI_IN_HEADER_LEN(q->vsq_sc)) == false) {
WPRINTF("ignoring request with incomplete header");
- return (0);
+ req->vsr_cmd_wr->response = VIRTIO_SCSI_S_FAILURE;
+ pci_vtscsi_return_request(q, req, 1);
+ return;
}
/*
@@ -500,8 +606,8 @@
* by one to make room for a new iovec covering the first part of the
* output data portion.
*/
- data_iov_out = split_iov(iov_out, &niov_out,
- VTSCSI_OUT_HEADER_LEN(q->vsq_sc), &data_niov_out);
+ req->vsr_data_iov_out = split_iov(req->vsr_iov_out, &req->vsr_niov_out,
+ VTSCSI_OUT_HEADER_LEN(q->vsq_sc), &req->vsr_data_niov_out);
/*
* Similarly, to not overwrite the first iovec of the output section,
@@ -509,43 +615,106 @@
* cover the entire iovec array (both input and the already split output
* sections).
*/
- niov_in += niov_out + data_niov_out;
+ req->vsr_niov_in += req->vsr_niov_out + req->vsr_data_niov_out;
- data_iov_in = split_iov(iov_in, &niov_in,
- VTSCSI_IN_HEADER_LEN(q->vsq_sc), &data_niov_in);
+ req->vsr_data_iov_in = split_iov(req->vsr_iov_in, &req->vsr_niov_in,
+ VTSCSI_IN_HEADER_LEN(q->vsq_sc), &req->vsr_data_niov_in);
/*
* And of course we now have to adjust data_niov_in accordingly.
*/
- data_niov_in -= niov_out + data_niov_out;
+ req->vsr_data_niov_in -= req->vsr_niov_out + req->vsr_data_niov_out;
- iov_to_buf(iov_in, niov_in, (void **)&cmd_rd);
+ /*
+ * iov_to_buf() realloc()s the buffer given as 3rd argument to the
+ * total size of all iovecs it will be copying. Since we've just
+ * truncated it in split_iov(), we know that the size will be
+ * VTSCSI_IN_HEADER_LEN(q->vsq_sc).
+ *
+ * Since we pre-allocated req->vsr_cmd_rd to this size, the realloc()
+ * should never fail.
+ *
+ * This will have to change if we begin allowing config space writes
+ * to change sense size.
+ */
+ assert(iov_to_buf(req->vsr_iov_in, req->vsr_niov_in,
+ (void **)&req->vsr_cmd_rd) == VTSCSI_IN_HEADER_LEN(q->vsq_sc));
- cmd_wr = calloc(1, VTSCSI_OUT_HEADER_LEN(sc));
- io = ctl_scsi_alloc_io(sc->vss_iid);
- ctl_scsi_zero_io(io);
+ pthread_mutex_lock(&q->vsq_rmtx);
+ pci_vtscsi_put_request(&q->vsq_requests, req);
+ pthread_cond_signal(&q->vsq_cv);
+ pthread_mutex_unlock(&q->vsq_rmtx);
+
+ DPRINTF("request <idx=%d> enqueued", vireq.idx);
+}
+
+static void
+pci_vtscsi_return_request(struct pci_vtscsi_queue *q,
+ struct pci_vtscsi_request *req, int iolen)
+{
+ void *cmd_rd = req->vsr_cmd_rd;
+ void *cmd_wr = req->vsr_cmd_wr;
+ void *ctl_io = req->vsr_ctl_io;
+ int idx = req->vsr_idx;
+
+ DPRINTF("request <idx=%d> completed, response %d", idx,
+ req->vsr_cmd_wr->response);
+
+ iolen += buf_to_iov(cmd_wr, VTSCSI_OUT_HEADER_LEN(q->vsq_sc),
+ req->vsr_iov_out, req->vsr_niov_out);
+
+ ctl_scsi_zero_io(req->vsr_ctl_io);
+
+ memset(cmd_rd, 0, VTSCSI_IN_HEADER_LEN(q->vsq_sc));
+ memset(cmd_wr, 0, VTSCSI_OUT_HEADER_LEN(q->vsq_sc));
+ memset(req, 0, sizeof(struct pci_vtscsi_request));
+
+ req->vsr_cmd_rd = cmd_rd;
+ req->vsr_cmd_wr = cmd_wr;
+ req->vsr_ctl_io = ctl_io;
+
+ pthread_mutex_lock(&q->vsq_fmtx);
+ pci_vtscsi_put_request(&q->vsq_free_requests, req);
+ pthread_mutex_unlock(&q->vsq_fmtx);
+
+ pthread_mutex_lock(&q->vsq_qmtx);
+ vq_relchain(q->vsq_vq, idx, iolen);
+ vq_endchains(q->vsq_vq, 0);
+ pthread_mutex_unlock(&q->vsq_qmtx);
+}
+
+static int
+pci_vtscsi_request_handle(struct pci_vtscsi_softc *sc,
+ struct pci_vtscsi_request *req)
+{
+ union ctl_io *io = req->vsr_ctl_io;
+ void *ext_data_ptr = NULL;
+ uint32_t ext_data_len = 0, ext_sg_entries = 0;
+ int err, nxferred;
io->io_hdr.nexus.initid = sc->vss_iid;
- io->io_hdr.nexus.targ_lun = pci_vtscsi_get_lun(cmd_rd->lun);
+ io->io_hdr.nexus.targ_lun = pci_vtscsi_get_lun(req->vsr_cmd_rd->lun);
io->io_hdr.io_type = CTL_IO_SCSI;
- if (data_niov_in > 0) {
- ext_data_ptr = (void *)data_iov_in;
- ext_sg_entries = data_niov_in;
- ext_data_len = count_iov(data_iov_in, data_niov_in);
+ if (req->vsr_data_niov_in > 0) {
+ ext_data_ptr = (void *)req->vsr_data_iov_in;
+ ext_sg_entries = req->vsr_data_niov_in;
+ ext_data_len = count_iov(req->vsr_data_iov_in,
+ req->vsr_data_niov_in);
io->io_hdr.flags |= CTL_FLAG_DATA_OUT;
- } else if (data_niov_out > 0) {
- ext_data_ptr = (void *)data_iov_out;
- ext_sg_entries = data_niov_out;
- ext_data_len = count_iov(data_iov_out, data_niov_out);
+ } else if (req->vsr_data_niov_out > 0) {
+ ext_data_ptr = (void *)req->vsr_data_iov_out;
+ ext_sg_entries = req->vsr_data_niov_out;
+ ext_data_len = count_iov(req->vsr_data_iov_out,
+ req->vsr_data_niov_out);
io->io_hdr.flags |= CTL_FLAG_DATA_IN;
}
io->scsiio.sense_len = sc->vss_config.sense_size;
- io->scsiio.tag_num = cmd_rd->id;
+ io->scsiio.tag_num = req->vsr_cmd_rd->id;
io->io_hdr.flags |= CTL_FLAG_USER_TAG;
- switch (cmd_rd->task_attr) {
+ switch (req->vsr_cmd_rd->task_attr) {
case VIRTIO_SCSI_S_ORDERED:
io->scsiio.tag_type = CTL_TAG_ORDERED;
break;
@@ -565,7 +734,7 @@
io->scsiio.ext_data_len = ext_data_len;
io->scsiio.ext_data_filled = 0;
io->scsiio.cdb_len = sc->vss_config.cdb_size;
- memcpy(io->scsiio.cdb, cmd_rd->cdb, sc->vss_config.cdb_size);
+ memcpy(io->scsiio.cdb, req->vsr_cmd_rd->cdb, sc->vss_config.cdb_size);
if (pci_vtscsi_debug) {
struct sbuf *sb = sbuf_new_auto();
@@ -578,22 +747,19 @@
err = ioctl(sc->vss_ctl_fd, CTL_IO, io);
if (err != 0) {
WPRINTF("CTL_IO: err=%d (%s)", errno, strerror(errno));
- cmd_wr->response = VIRTIO_SCSI_S_FAILURE;
+ req->vsr_cmd_wr->response = VIRTIO_SCSI_S_FAILURE;
} else {
- cmd_wr->sense_len = MIN(io->scsiio.sense_len,
- sc->vss_config.sense_size);
- cmd_wr->residual = ext_data_len - io->scsiio.ext_data_filled;
- cmd_wr->status = io->scsiio.scsi_status;
- cmd_wr->response = VIRTIO_SCSI_S_OK;
- memcpy(&cmd_wr->sense, &io->scsiio.sense_data,
- cmd_wr->sense_len);
+ req->vsr_cmd_wr->sense_len =
+ MIN(io->scsiio.sense_len, sc->vss_config.sense_size);
+ req->vsr_cmd_wr->residual = ext_data_len -
+ io->scsiio.ext_data_filled;
+ req->vsr_cmd_wr->status = io->scsiio.scsi_status;
+ req->vsr_cmd_wr->response = VIRTIO_SCSI_S_OK;
+ memcpy(&req->vsr_cmd_wr->sense, &io->scsiio.sense_data,
+ req->vsr_cmd_wr->sense_len);
}
- buf_to_iov(cmd_wr, VTSCSI_OUT_HEADER_LEN(sc), iov_out, niov_out);
- nxferred = VTSCSI_OUT_HEADER_LEN(sc) + io->scsiio.ext_data_filled;
- free(cmd_rd);
- free(cmd_wr);
- ctl_scsi_free_io(io);
+ nxferred = io->scsiio.ext_data_filled;
return (nxferred);
}
@@ -635,34 +801,8 @@
static void
pci_vtscsi_requestq_notify(void *vsc, struct vqueue_info *vq)
{
- struct pci_vtscsi_softc *sc;
- struct pci_vtscsi_queue *q;
- struct pci_vtscsi_request *req;
- struct vi_req vireq;
- int n;
-
- sc = vsc;
- q = &sc->vss_queues[vq->vq_num - 2];
-
while (vq_has_descs(vq)) {
- req = calloc(1, sizeof(struct pci_vtscsi_request));
-
- n = vq_getchain(vq, req->vsr_iov, VTSCSI_MAXSEG, &vireq);
- assert(n >= 1 && n <= VTSCSI_MAXSEG);
-
- req->vsr_idx = vireq.idx;
- req->vsr_queue = q;
- req->vsr_iov_in = &req->vsr_iov[0];
- req->vsr_niov_in = vireq.readable;
- req->vsr_iov_out = &req->vsr_iov[vireq.readable];
- req->vsr_niov_out = vireq.writable;
-
- pthread_mutex_lock(&q->vsq_mtx);
- STAILQ_INSERT_TAIL(&q->vsq_requests, req, vsr_link);
- pthread_cond_signal(&q->vsq_cv);
- pthread_mutex_unlock(&q->vsq_mtx);
-
- DPRINTF("request <idx=%d> enqueued", vireq.idx);
+ pci_vtscsi_queue_request(vsc, vq);
}
}
@@ -670,21 +810,36 @@
pci_vtscsi_init_queue(struct pci_vtscsi_softc *sc,
struct pci_vtscsi_queue *queue, int num)
{
- struct pci_vtscsi_worker *worker;
char tname[MAXCOMLEN + 1];
int i;
queue->vsq_sc = sc;
queue->vsq_vq = &sc->vss_vq[num + 2];
- pthread_mutex_init(&queue->vsq_mtx, NULL);
+ pthread_mutex_init(&queue->vsq_rmtx, NULL);
+ pthread_mutex_init(&queue->vsq_fmtx, NULL);
pthread_mutex_init(&queue->vsq_qmtx, NULL);
pthread_cond_init(&queue->vsq_cv, NULL);
STAILQ_INIT(&queue->vsq_requests);
+ STAILQ_INIT(&queue->vsq_free_requests);
LIST_INIT(&queue->vsq_workers);
+ for (i = 0; i < VTSCSI_RINGSZ; i++) {
+ struct pci_vtscsi_request *req;
+
+ req = pci_vtscsi_alloc_request(sc);
+ if (req == NULL)
+ return (-1);
+
+ pci_vtscsi_put_request(&queue->vsq_free_requests, req);
+ }
+
for (i = 0; i < VTSCSI_THR_PER_Q; i++) {
+ struct pci_vtscsi_worker *worker;
worker = calloc(1, sizeof(struct pci_vtscsi_worker));
+ if (worker == NULL)
+ return (-1);
+
worker->vsw_queue = queue;
pthread_create(&worker->vsw_thread, NULL, &pci_vtscsi_proc,
@@ -722,9 +877,13 @@
{
struct pci_vtscsi_softc *sc;
const char *devname, *value;
+ int err;
int i;
sc = calloc(1, sizeof(struct pci_vtscsi_softc));
+ if (sc == NULL)
+ return (-1);
+
value = get_config_value_node(nvl, "iid");
if (value != NULL)
sc->vss_iid = strtoul(value, NULL, 10);
@@ -753,6 +912,19 @@
vi_softc_linkup(&sc->vss_vs, &vtscsi_vi_consts, sc, pi, sc->vss_vq);
sc->vss_vs.vs_mtx = &sc->vss_mtx;
+ /*
+ * Perform a "reset" before we set up our queues.
+ *
+ * This will write the default config into vss_config, which is used
+ * by the rest of the driver to get the request header size. Note that
+ * if we ever allow the guest to override sense size through config
+ * space writes, pre-allocation of I/O requests will have to change
+ * accordingly.
+ */
+ pthread_mutex_lock(&sc->vss_mtx);
+ pci_vtscsi_reset(sc);
+ pthread_mutex_unlock(&sc->vss_mtx);
+
/* controlq */
sc->vss_vq[0].vq_qsize = VTSCSI_RINGSZ;
sc->vss_vq[0].vq_notify = pci_vtscsi_controlq_notify;
@@ -765,7 +937,10 @@
for (i = 2; i < VTSCSI_MAXQ; i++) {
sc->vss_vq[i].vq_qsize = VTSCSI_RINGSZ;
sc->vss_vq[i].vq_notify = pci_vtscsi_requestq_notify;
- pci_vtscsi_init_queue(sc, &sc->vss_queues[i - 2], i - 2);
+
+ err = pci_vtscsi_init_queue(sc, &sc->vss_queues[i - 2], i - 2);
+ if (err != 0)
+ return (err);
}
/* initialize config space */
File Metadata
Details
Attached
Mime Type
text/plain
Expires
Mon, Feb 16, 5:22 PM (22 h, 17 m)
Storage Engine
blob
Storage Format
Raw Data
Storage Handle
28628044
Default Alt Text
D53469.diff (17 KB)
Attached To
Mode
D53469: bhyve/virtio-scsi: Preallocate all I/O requests
Attached
Detach File
Event Timeline
Log In to Comment