2 * Copyright (c) 2009-2012,2016 Microsoft Corp.
3 * Copyright (c) 2012 NetApp Inc.
4 * Copyright (c) 2012 Citrix Inc.
7 * Redistribution and use in source and binary forms, with or without
8 * modification, are permitted provided that the following conditions
10 * 1. Redistributions of source code must retain the above copyright
11 * notice unmodified, this list of conditions, and the following
13 * 2. Redistributions in binary form must reproduce the above copyright
14 * notice, this list of conditions and the following disclaimer in the
15 * documentation and/or other materials provided with the distribution.
17 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR
18 * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
19 * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
20 * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT,
21 * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT
22 * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
23 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
24 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
25 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF
26 * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
30 * StorVSC driver for Hyper-V. This driver presents a SCSI HBA interface
31 * to the Comman Access Method (CAM) layer. CAM control blocks (CCBs) are
32 * converted into VSCSI protocol messages which are delivered to the parent
33 * partition StorVSP driver over the Hyper-V VMBUS.
35 #include <sys/cdefs.h>
36 __FBSDID("$FreeBSD$");
38 #include <sys/param.h>
40 #include <sys/condvar.h>
42 #include <sys/systm.h>
43 #include <sys/sockio.h>
45 #include <sys/malloc.h>
46 #include <sys/module.h>
47 #include <sys/kernel.h>
48 #include <sys/queue.h>
51 #include <sys/taskqueue.h>
53 #include <sys/mutex.h>
54 #include <sys/callout.h>
60 #include <sys/sglist.h>
61 #include <machine/bus.h>
62 #include <sys/bus_dma.h>
65 #include <cam/cam_ccb.h>
66 #include <cam/cam_periph.h>
67 #include <cam/cam_sim.h>
68 #include <cam/cam_xpt_sim.h>
69 #include <cam/cam_xpt_internal.h>
70 #include <cam/cam_debug.h>
71 #include <cam/scsi/scsi_all.h>
72 #include <cam/scsi/scsi_message.h>
74 #include <dev/hyperv/include/hyperv.h>
75 #include <dev/hyperv/include/vmbus.h>
77 #include "hv_vstorage.h"
80 #define STORVSC_RINGBUFFER_SIZE (20*PAGE_SIZE)
81 #define STORVSC_MAX_LUNS_PER_TARGET (64)
82 #define STORVSC_MAX_IO_REQUESTS (STORVSC_MAX_LUNS_PER_TARGET * 2)
83 #define BLKVSC_MAX_IDE_DISKS_PER_TARGET (1)
84 #define BLKVSC_MAX_IO_REQUESTS STORVSC_MAX_IO_REQUESTS
85 #define STORVSC_MAX_TARGETS (2)
87 #define VSTOR_PKT_SIZE (sizeof(struct vstor_packet) - vmscsi_size_delta)
89 #define HV_ALIGN(x, a) roundup2(x, a)
94 LIST_ENTRY(hv_sgl_node) link;
95 struct sglist *sgl_data;
98 struct hv_sgl_page_pool{
99 LIST_HEAD(, hv_sgl_node) in_use_sgl_list;
100 LIST_HEAD(, hv_sgl_node) free_sgl_list;
102 } g_hv_sgl_page_pool;
104 #define STORVSC_MAX_SG_PAGE_CNT STORVSC_MAX_IO_REQUESTS * VMBUS_CHAN_PRPLIST_MAX
106 enum storvsc_request_type {
112 struct hvs_gpa_range {
113 struct vmbus_gpa_range gpa_range;
114 uint64_t gpa_page[VMBUS_CHAN_PRPLIST_MAX];
117 struct hv_storvsc_request {
118 LIST_ENTRY(hv_storvsc_request) link;
119 struct vstor_packet vstor_packet;
121 struct hvs_gpa_range prp_list;
123 uint8_t sense_info_len;
126 struct storvsc_softc *softc;
127 struct callout callout;
128 struct sema synch_sema; /*Synchronize the request/response if needed */
129 struct sglist *bounce_sgl;
130 unsigned int bounce_sgl_count;
131 uint64_t not_aligned_seg_bits;
134 struct storvsc_softc {
135 struct hv_vmbus_channel *hs_chan;
136 LIST_HEAD(, hv_storvsc_request) hs_free_list;
138 struct storvsc_driver_props *hs_drv_props;
141 struct cam_sim *hs_sim;
142 struct cam_path *hs_path;
143 uint32_t hs_num_out_reqs;
144 boolean_t hs_destroy;
145 boolean_t hs_drain_notify;
146 struct sema hs_drain_sema;
147 struct hv_storvsc_request hs_init_req;
148 struct hv_storvsc_request hs_reset_req;
154 * HyperV storvsc timeout testing cases:
155 * a. IO returned after first timeout;
156 * b. IO returned after second timeout and queue freeze;
157 * c. IO returned while timer handler is running
158 * The first can be tested by "sg_senddiag -vv /dev/daX",
159 * and the second and third can be done by
160 * "sg_wr_mode -v -p 08 -c 0,1a -m 0,ff /dev/daX".
162 #define HVS_TIMEOUT_TEST 0
165 * Bus/adapter reset functionality on the Hyper-V host is
166 * buggy and it will be disabled until
167 * it can be further tested.
169 #define HVS_HOST_RESET 0
171 struct storvsc_driver_props {
174 uint8_t drv_max_luns_per_target;
175 uint8_t drv_max_ios_per_target;
176 uint32_t drv_ringbuffer_size;
179 enum hv_storage_type {
185 #define HS_MAX_ADAPTERS 10
187 #define HV_STORAGE_SUPPORTS_MULTI_CHANNEL 0x1
189 /* {ba6163d9-04a1-4d29-b605-72e2ffb1dc7f} */
190 static const struct hyperv_guid gStorVscDeviceType={
191 .hv_guid = {0xd9, 0x63, 0x61, 0xba, 0xa1, 0x04, 0x29, 0x4d,
192 0xb6, 0x05, 0x72, 0xe2, 0xff, 0xb1, 0xdc, 0x7f}
195 /* {32412632-86cb-44a2-9b5c-50d1417354f5} */
196 static const struct hyperv_guid gBlkVscDeviceType={
197 .hv_guid = {0x32, 0x26, 0x41, 0x32, 0xcb, 0x86, 0xa2, 0x44,
198 0x9b, 0x5c, 0x50, 0xd1, 0x41, 0x73, 0x54, 0xf5}
201 static struct storvsc_driver_props g_drv_props_table[] = {
202 {"blkvsc", "Hyper-V IDE Storage Interface",
203 BLKVSC_MAX_IDE_DISKS_PER_TARGET, BLKVSC_MAX_IO_REQUESTS,
204 STORVSC_RINGBUFFER_SIZE},
205 {"storvsc", "Hyper-V SCSI Storage Interface",
206 STORVSC_MAX_LUNS_PER_TARGET, STORVSC_MAX_IO_REQUESTS,
207 STORVSC_RINGBUFFER_SIZE}
211 * Sense buffer size changed in win8; have a run-time
212 * variable to track the size we should use.
214 static int sense_buffer_size = PRE_WIN8_STORVSC_SENSE_BUFFER_SIZE;
217 * The size of the vmscsi_request has changed in win8. The
218 * additional size is for the newly added elements in the
219 * structure. These elements are valid only when we are talking
221 * Track the correct size we need to apply.
223 static int vmscsi_size_delta;
225 * The storage protocol version is determined during the
226 * initial exchange with the host. It will indicate which
227 * storage functionality is available in the host.
229 static int vmstor_proto_version;
231 struct vmstor_proto {
233 int sense_buffer_size;
234 int vmscsi_size_delta;
237 static const struct vmstor_proto vmstor_proto_list[] = {
239 VMSTOR_PROTOCOL_VERSION_WIN10,
240 POST_WIN7_STORVSC_SENSE_BUFFER_SIZE,
244 VMSTOR_PROTOCOL_VERSION_WIN8_1,
245 POST_WIN7_STORVSC_SENSE_BUFFER_SIZE,
249 VMSTOR_PROTOCOL_VERSION_WIN8,
250 POST_WIN7_STORVSC_SENSE_BUFFER_SIZE,
254 VMSTOR_PROTOCOL_VERSION_WIN7,
255 PRE_WIN8_STORVSC_SENSE_BUFFER_SIZE,
256 sizeof(struct vmscsi_win8_extension),
259 VMSTOR_PROTOCOL_VERSION_WIN6,
260 PRE_WIN8_STORVSC_SENSE_BUFFER_SIZE,
261 sizeof(struct vmscsi_win8_extension),
265 /* static functions */
266 static int storvsc_probe(device_t dev);
267 static int storvsc_attach(device_t dev);
268 static int storvsc_detach(device_t dev);
269 static void storvsc_poll(struct cam_sim * sim);
270 static void storvsc_action(struct cam_sim * sim, union ccb * ccb);
271 static int create_storvsc_request(union ccb *ccb, struct hv_storvsc_request *reqp);
272 static void storvsc_free_request(struct storvsc_softc *sc, struct hv_storvsc_request *reqp);
273 static enum hv_storage_type storvsc_get_storage_type(device_t dev);
274 static void hv_storvsc_rescan_target(struct storvsc_softc *sc);
275 static void hv_storvsc_on_channel_callback(void *xchan);
276 static void hv_storvsc_on_iocompletion( struct storvsc_softc *sc,
277 struct vstor_packet *vstor_packet,
278 struct hv_storvsc_request *request);
279 static int hv_storvsc_connect_vsp(struct storvsc_softc *);
280 static void storvsc_io_done(struct hv_storvsc_request *reqp);
281 static void storvsc_copy_sgl_to_bounce_buf(struct sglist *bounce_sgl,
282 bus_dma_segment_t *orig_sgl,
283 unsigned int orig_sgl_count,
285 void storvsc_copy_from_bounce_buf_to_sgl(bus_dma_segment_t *dest_sgl,
286 unsigned int dest_sgl_count,
287 struct sglist* src_sgl,
290 static device_method_t storvsc_methods[] = {
291 /* Device interface */
292 DEVMETHOD(device_probe, storvsc_probe),
293 DEVMETHOD(device_attach, storvsc_attach),
294 DEVMETHOD(device_detach, storvsc_detach),
295 DEVMETHOD(device_shutdown, bus_generic_shutdown),
299 static driver_t storvsc_driver = {
300 "storvsc", storvsc_methods, sizeof(struct storvsc_softc),
303 static devclass_t storvsc_devclass;
304 DRIVER_MODULE(storvsc, vmbus, storvsc_driver, storvsc_devclass, 0, 0);
305 MODULE_VERSION(storvsc, 1);
306 MODULE_DEPEND(storvsc, vmbus, 1, 1, 1);
309 storvsc_subchan_attach(struct storvsc_softc *sc,
310 struct hv_vmbus_channel *new_channel)
312 struct vmstor_chan_props props;
315 memset(&props, 0, sizeof(props));
317 new_channel->hv_chan_priv1 = sc;
318 vmbus_channel_cpu_rr(new_channel);
319 ret = hv_vmbus_channel_open(new_channel,
320 sc->hs_drv_props->drv_ringbuffer_size,
321 sc->hs_drv_props->drv_ringbuffer_size,
323 sizeof(struct vmstor_chan_props),
324 hv_storvsc_on_channel_callback,
329 * @brief Send multi-channel creation request to host
331 * @param device a Hyper-V device pointer
332 * @param max_chans the max channels supported by vmbus
335 storvsc_send_multichannel_request(struct storvsc_softc *sc, int max_chans)
337 struct hv_vmbus_channel **subchan;
338 struct hv_storvsc_request *request;
339 struct vstor_packet *vstor_packet;
340 int request_channels_cnt = 0;
343 /* get multichannels count that need to create */
344 request_channels_cnt = MIN(max_chans, mp_ncpus);
346 request = &sc->hs_init_req;
348 /* request the host to create multi-channel */
349 memset(request, 0, sizeof(struct hv_storvsc_request));
351 sema_init(&request->synch_sema, 0, ("stor_synch_sema"));
353 vstor_packet = &request->vstor_packet;
355 vstor_packet->operation = VSTOR_OPERATION_CREATE_MULTI_CHANNELS;
356 vstor_packet->flags = REQUEST_COMPLETION_FLAG;
357 vstor_packet->u.multi_channels_cnt = request_channels_cnt;
359 ret = vmbus_chan_send(sc->hs_chan,
360 VMBUS_CHANPKT_TYPE_INBAND, VMBUS_CHANPKT_FLAG_RC,
361 vstor_packet, VSTOR_PKT_SIZE, (uint64_t)(uintptr_t)request);
363 /* wait for 5 seconds */
364 ret = sema_timedwait(&request->synch_sema, 5 * hz);
366 printf("Storvsc_error: create multi-channel timeout, %d\n",
371 if (vstor_packet->operation != VSTOR_OPERATION_COMPLETEIO ||
372 vstor_packet->status != 0) {
373 printf("Storvsc_error: create multi-channel invalid operation "
374 "(%d) or statue (%u)\n",
375 vstor_packet->operation, vstor_packet->status);
379 /* Wait for sub-channels setup to complete. */
380 subchan = vmbus_get_subchan(sc->hs_chan, request_channels_cnt);
382 /* Attach the sub-channels. */
383 for (i = 0; i < request_channels_cnt; ++i)
384 storvsc_subchan_attach(sc, subchan[i]);
386 /* Release the sub-channels. */
387 vmbus_rel_subchan(subchan, request_channels_cnt);
390 printf("Storvsc create multi-channel success!\n");
394 * @brief initialize channel connection to parent partition
396 * @param dev a Hyper-V device pointer
397 * @returns 0 on success, non-zero error on failure
400 hv_storvsc_channel_init(struct storvsc_softc *sc)
403 struct hv_storvsc_request *request;
404 struct vstor_packet *vstor_packet;
405 uint16_t max_chans = 0;
406 boolean_t support_multichannel = FALSE;
410 support_multichannel = FALSE;
412 request = &sc->hs_init_req;
413 memset(request, 0, sizeof(struct hv_storvsc_request));
414 vstor_packet = &request->vstor_packet;
418 * Initiate the vsc/vsp initialization protocol on the open channel
420 sema_init(&request->synch_sema, 0, ("stor_synch_sema"));
422 vstor_packet->operation = VSTOR_OPERATION_BEGININITIALIZATION;
423 vstor_packet->flags = REQUEST_COMPLETION_FLAG;
426 ret = vmbus_chan_send(sc->hs_chan,
427 VMBUS_CHANPKT_TYPE_INBAND, VMBUS_CHANPKT_FLAG_RC,
428 vstor_packet, VSTOR_PKT_SIZE, (uint64_t)(uintptr_t)request);
434 ret = sema_timedwait(&request->synch_sema, 5 * hz);
438 if (vstor_packet->operation != VSTOR_OPERATION_COMPLETEIO ||
439 vstor_packet->status != 0) {
443 for (i = 0; i < nitems(vmstor_proto_list); i++) {
444 /* reuse the packet for version range supported */
446 memset(vstor_packet, 0, sizeof(struct vstor_packet));
447 vstor_packet->operation = VSTOR_OPERATION_QUERYPROTOCOLVERSION;
448 vstor_packet->flags = REQUEST_COMPLETION_FLAG;
450 vstor_packet->u.version.major_minor =
451 vmstor_proto_list[i].proto_version;
453 /* revision is only significant for Windows guests */
454 vstor_packet->u.version.revision = 0;
456 ret = vmbus_chan_send(sc->hs_chan,
457 VMBUS_CHANPKT_TYPE_INBAND, VMBUS_CHANPKT_FLAG_RC,
458 vstor_packet, VSTOR_PKT_SIZE, (uint64_t)(uintptr_t)request);
464 ret = sema_timedwait(&request->synch_sema, 5 * hz);
469 if (vstor_packet->operation != VSTOR_OPERATION_COMPLETEIO) {
473 if (vstor_packet->status == 0) {
474 vmstor_proto_version =
475 vmstor_proto_list[i].proto_version;
477 vmstor_proto_list[i].sense_buffer_size;
479 vmstor_proto_list[i].vmscsi_size_delta;
484 if (vstor_packet->status != 0) {
489 * Query channel properties
491 memset(vstor_packet, 0, sizeof(struct vstor_packet));
492 vstor_packet->operation = VSTOR_OPERATION_QUERYPROPERTIES;
493 vstor_packet->flags = REQUEST_COMPLETION_FLAG;
495 ret = vmbus_chan_send(sc->hs_chan,
496 VMBUS_CHANPKT_TYPE_INBAND, VMBUS_CHANPKT_FLAG_RC,
497 vstor_packet, VSTOR_PKT_SIZE, (uint64_t)(uintptr_t)request);
503 ret = sema_timedwait(&request->synch_sema, 5 * hz);
508 /* TODO: Check returned version */
509 if (vstor_packet->operation != VSTOR_OPERATION_COMPLETEIO ||
510 vstor_packet->status != 0) {
514 /* multi-channels feature is supported by WIN8 and above version */
515 max_chans = vstor_packet->u.chan_props.max_channel_cnt;
516 version = VMBUS_GET_VERSION(device_get_parent(sc->hs_dev), sc->hs_dev);
517 if (version != VMBUS_VERSION_WIN7 && version != VMBUS_VERSION_WS2008 &&
518 (vstor_packet->u.chan_props.flags &
519 HV_STORAGE_SUPPORTS_MULTI_CHANNEL)) {
520 support_multichannel = TRUE;
523 memset(vstor_packet, 0, sizeof(struct vstor_packet));
524 vstor_packet->operation = VSTOR_OPERATION_ENDINITIALIZATION;
525 vstor_packet->flags = REQUEST_COMPLETION_FLAG;
527 ret = vmbus_chan_send(sc->hs_chan,
528 VMBUS_CHANPKT_TYPE_INBAND, VMBUS_CHANPKT_FLAG_RC,
529 vstor_packet, VSTOR_PKT_SIZE, (uint64_t)(uintptr_t)request);
536 ret = sema_timedwait(&request->synch_sema, 5 * hz);
541 if (vstor_packet->operation != VSTOR_OPERATION_COMPLETEIO ||
542 vstor_packet->status != 0)
546 * If multi-channel is supported, send multichannel create
549 if (support_multichannel)
550 storvsc_send_multichannel_request(sc, max_chans);
553 sema_destroy(&request->synch_sema);
558 * @brief Open channel connection to paraent partition StorVSP driver
560 * Open and initialize channel connection to parent partition StorVSP driver.
562 * @param pointer to a Hyper-V device
563 * @returns 0 on success, non-zero error on failure
566 hv_storvsc_connect_vsp(struct storvsc_softc *sc)
569 struct vmstor_chan_props props;
571 memset(&props, 0, sizeof(struct vmstor_chan_props));
576 KASSERT(sc->hs_chan->hv_chan_priv1 == sc, ("invalid chan priv1"));
577 vmbus_channel_cpu_rr(sc->hs_chan);
578 ret = hv_vmbus_channel_open(
580 sc->hs_drv_props->drv_ringbuffer_size,
581 sc->hs_drv_props->drv_ringbuffer_size,
583 sizeof(struct vmstor_chan_props),
584 hv_storvsc_on_channel_callback,
591 ret = hv_storvsc_channel_init(sc);
598 hv_storvsc_host_reset(struct storvsc_softc *sc)
602 struct hv_storvsc_request *request;
603 struct vstor_packet *vstor_packet;
605 request = &sc->hs_reset_req;
607 vstor_packet = &request->vstor_packet;
609 sema_init(&request->synch_sema, 0, "stor synch sema");
611 vstor_packet->operation = VSTOR_OPERATION_RESETBUS;
612 vstor_packet->flags = REQUEST_COMPLETION_FLAG;
614 ret = vmbus_chan_send(dev->channel,
615 VMBUS_CHANPKT_TYPE_INBAND, VMBUS_CHANPKT_FLAG_RC,
616 vstor_packet, VSTOR_PKT_SIZE,
617 (uint64_t)(uintptr_t)&sc->hs_reset_req);
623 ret = sema_timedwait(&request->synch_sema, 5 * hz); /* KYS 5 seconds */
631 * At this point, all outstanding requests in the adapter
632 * should have been flushed out and return to us
636 sema_destroy(&request->synch_sema);
639 #endif /* HVS_HOST_RESET */
642 * @brief Function to initiate an I/O request
644 * @param device Hyper-V device pointer
645 * @param request pointer to a request structure
646 * @returns 0 on success, non-zero error on failure
649 hv_storvsc_io_request(struct storvsc_softc *sc,
650 struct hv_storvsc_request *request)
652 struct vstor_packet *vstor_packet = &request->vstor_packet;
653 struct hv_vmbus_channel* outgoing_channel = NULL;
656 vstor_packet->flags |= REQUEST_COMPLETION_FLAG;
658 vstor_packet->u.vm_srb.length =
659 sizeof(struct vmscsi_req) - vmscsi_size_delta;
661 vstor_packet->u.vm_srb.sense_info_len = sense_buffer_size;
663 vstor_packet->u.vm_srb.transfer_len =
664 request->prp_list.gpa_range.gpa_len;
666 vstor_packet->operation = VSTOR_OPERATION_EXECUTESRB;
668 outgoing_channel = vmbus_select_outgoing_channel(sc->hs_chan);
670 mtx_unlock(&request->softc->hs_lock);
671 if (request->prp_list.gpa_range.gpa_len) {
672 ret = vmbus_chan_send_prplist(outgoing_channel,
673 &request->prp_list.gpa_range, request->prp_cnt,
674 vstor_packet, VSTOR_PKT_SIZE, (uint64_t)(uintptr_t)request);
676 ret = vmbus_chan_send(outgoing_channel,
677 VMBUS_CHANPKT_TYPE_INBAND, VMBUS_CHANPKT_FLAG_RC,
678 vstor_packet, VSTOR_PKT_SIZE, (uint64_t)(uintptr_t)request);
680 mtx_lock(&request->softc->hs_lock);
683 printf("Unable to send packet %p ret %d", vstor_packet, ret);
685 atomic_add_int(&sc->hs_num_out_reqs, 1);
693 * Process IO_COMPLETION_OPERATION and ready
694 * the result to be completed for upper layer
695 * processing by the CAM layer.
698 hv_storvsc_on_iocompletion(struct storvsc_softc *sc,
699 struct vstor_packet *vstor_packet,
700 struct hv_storvsc_request *request)
702 struct vmscsi_req *vm_srb;
704 vm_srb = &vstor_packet->u.vm_srb;
707 * Copy some fields of the host's response into the request structure,
708 * because the fields will be used later in storvsc_io_done().
710 request->vstor_packet.u.vm_srb.scsi_status = vm_srb->scsi_status;
711 request->vstor_packet.u.vm_srb.srb_status = vm_srb->srb_status;
712 request->vstor_packet.u.vm_srb.transfer_len = vm_srb->transfer_len;
714 if (((vm_srb->scsi_status & 0xFF) == SCSI_STATUS_CHECK_COND) &&
715 (vm_srb->srb_status & SRB_STATUS_AUTOSENSE_VALID)) {
716 /* Autosense data available */
718 KASSERT(vm_srb->sense_info_len <= request->sense_info_len,
719 ("vm_srb->sense_info_len <= "
720 "request->sense_info_len"));
722 memcpy(request->sense_data, vm_srb->u.sense_data,
723 vm_srb->sense_info_len);
725 request->sense_info_len = vm_srb->sense_info_len;
728 /* Complete request by passing to the CAM layer */
729 storvsc_io_done(request);
730 atomic_subtract_int(&sc->hs_num_out_reqs, 1);
731 if (sc->hs_drain_notify && (sc->hs_num_out_reqs == 0)) {
732 sema_post(&sc->hs_drain_sema);
737 hv_storvsc_rescan_target(struct storvsc_softc *sc)
740 target_id_t targetid;
743 pathid = cam_sim_path(sc->hs_sim);
744 targetid = CAM_TARGET_WILDCARD;
747 * Allocate a CCB and schedule a rescan.
749 ccb = xpt_alloc_ccb_nowait();
751 printf("unable to alloc CCB for rescan\n");
755 if (xpt_create_path(&ccb->ccb_h.path, NULL, pathid, targetid,
756 CAM_LUN_WILDCARD) != CAM_REQ_CMP) {
757 printf("unable to create path for rescan, pathid: %u,"
758 "targetid: %u\n", pathid, targetid);
763 if (targetid == CAM_TARGET_WILDCARD)
764 ccb->ccb_h.func_code = XPT_SCAN_BUS;
766 ccb->ccb_h.func_code = XPT_SCAN_TGT;
772 hv_storvsc_on_channel_callback(void *xchan)
775 hv_vmbus_channel *channel = xchan;
776 struct storvsc_softc *sc = channel->hv_chan_priv1;
777 uint32_t bytes_recvd;
779 uint8_t packet[roundup2(sizeof(struct vstor_packet), 8)];
780 struct hv_storvsc_request *request;
781 struct vstor_packet *vstor_packet;
783 ret = hv_vmbus_channel_recv_packet(
786 roundup2(VSTOR_PKT_SIZE, 8),
790 while ((ret == 0) && (bytes_recvd > 0)) {
791 request = (struct hv_storvsc_request *)(uintptr_t)request_id;
793 if ((request == &sc->hs_init_req) ||
794 (request == &sc->hs_reset_req)) {
795 memcpy(&request->vstor_packet, packet,
796 sizeof(struct vstor_packet));
797 sema_post(&request->synch_sema);
799 vstor_packet = (struct vstor_packet *)packet;
800 switch(vstor_packet->operation) {
801 case VSTOR_OPERATION_COMPLETEIO:
803 panic("VMBUS: storvsc received a "
804 "packet with NULL request id in "
805 "COMPLETEIO operation.");
807 hv_storvsc_on_iocompletion(sc,
808 vstor_packet, request);
810 case VSTOR_OPERATION_REMOVEDEVICE:
811 printf("VMBUS: storvsc operation %d not "
812 "implemented.\n", vstor_packet->operation);
813 /* TODO: implement */
815 case VSTOR_OPERATION_ENUMERATE_BUS:
816 hv_storvsc_rescan_target(sc);
822 ret = hv_vmbus_channel_recv_packet(
825 roundup2(VSTOR_PKT_SIZE, 8),
832 * @brief StorVSC probe function
834 * Device probe function. Returns 0 if the input device is a StorVSC
835 * device. Otherwise, a ENXIO is returned. If the input device is
836 * for BlkVSC (paravirtual IDE) device and this support is disabled in
837 * favor of the emulated ATA/IDE device, return ENXIO.
840 * @returns 0 on success, ENXIO if not a matcing StorVSC device
843 storvsc_probe(device_t dev)
845 int ata_disk_enable = 0;
848 switch (storvsc_get_storage_type(dev)) {
851 device_printf(dev, "DRIVER_BLKVSC-Emulated ATA/IDE probe\n");
852 if (!getenv_int("hw.ata.disk_enable", &ata_disk_enable)) {
855 "Enlightened ATA/IDE detected\n");
856 device_set_desc(dev, g_drv_props_table[DRIVER_BLKVSC].drv_desc);
857 ret = BUS_PROBE_DEFAULT;
858 } else if(bootverbose)
859 device_printf(dev, "Emulated ATA/IDE set (hw.ata.disk_enable set)\n");
863 device_printf(dev, "Enlightened SCSI device detected\n");
864 device_set_desc(dev, g_drv_props_table[DRIVER_STORVSC].drv_desc);
865 ret = BUS_PROBE_DEFAULT;
874 * @brief StorVSC attach function
876 * Function responsible for allocating per-device structures,
877 * setting up CAM interfaces and scanning for available LUNs to
878 * be used for SCSI device peripherals.
881 * @returns 0 on success or an error on failure
884 storvsc_attach(device_t dev)
886 enum hv_storage_type stor_type;
887 struct storvsc_softc *sc;
888 struct cam_devq *devq;
890 struct hv_storvsc_request *reqp;
891 struct root_hold_token *root_mount_token = NULL;
892 struct hv_sgl_node *sgl_node = NULL;
893 void *tmp_buff = NULL;
896 * We need to serialize storvsc attach calls.
898 root_mount_token = root_mount_hold("storvsc");
900 sc = device_get_softc(dev);
901 sc->hs_chan = vmbus_get_channel(dev);
902 sc->hs_chan->hv_chan_priv1 = sc;
904 stor_type = storvsc_get_storage_type(dev);
906 if (stor_type == DRIVER_UNKNOWN) {
911 /* fill in driver specific properties */
912 sc->hs_drv_props = &g_drv_props_table[stor_type];
914 /* fill in device specific properties */
915 sc->hs_unit = device_get_unit(dev);
918 LIST_INIT(&sc->hs_free_list);
919 mtx_init(&sc->hs_lock, "hvslck", NULL, MTX_DEF);
921 for (i = 0; i < sc->hs_drv_props->drv_max_ios_per_target; ++i) {
922 reqp = malloc(sizeof(struct hv_storvsc_request),
923 M_DEVBUF, M_WAITOK|M_ZERO);
926 LIST_INSERT_HEAD(&sc->hs_free_list, reqp, link);
929 /* create sg-list page pool */
930 if (FALSE == g_hv_sgl_page_pool.is_init) {
931 g_hv_sgl_page_pool.is_init = TRUE;
932 LIST_INIT(&g_hv_sgl_page_pool.in_use_sgl_list);
933 LIST_INIT(&g_hv_sgl_page_pool.free_sgl_list);
936 * Pre-create SG list, each SG list with
937 * VMBUS_CHAN_PRPLIST_MAX segments, each
938 * segment has one page buffer
940 for (i = 0; i < STORVSC_MAX_IO_REQUESTS; i++) {
941 sgl_node = malloc(sizeof(struct hv_sgl_node),
942 M_DEVBUF, M_WAITOK|M_ZERO);
945 sglist_alloc(VMBUS_CHAN_PRPLIST_MAX,
948 for (j = 0; j < VMBUS_CHAN_PRPLIST_MAX; j++) {
949 tmp_buff = malloc(PAGE_SIZE,
950 M_DEVBUF, M_WAITOK|M_ZERO);
952 sgl_node->sgl_data->sg_segs[j].ss_paddr =
953 (vm_paddr_t)tmp_buff;
956 LIST_INSERT_HEAD(&g_hv_sgl_page_pool.free_sgl_list,
961 sc->hs_destroy = FALSE;
962 sc->hs_drain_notify = FALSE;
963 sema_init(&sc->hs_drain_sema, 0, "Store Drain Sema");
965 ret = hv_storvsc_connect_vsp(sc);
971 * Create the device queue.
972 * Hyper-V maps each target to one SCSI HBA
974 devq = cam_simq_alloc(sc->hs_drv_props->drv_max_ios_per_target);
976 device_printf(dev, "Failed to alloc device queue\n");
981 sc->hs_sim = cam_sim_alloc(storvsc_action,
983 sc->hs_drv_props->drv_name,
987 sc->hs_drv_props->drv_max_ios_per_target,
990 if (sc->hs_sim == NULL) {
991 device_printf(dev, "Failed to alloc sim\n");
997 mtx_lock(&sc->hs_lock);
998 /* bus_id is set to 0, need to get it from VMBUS channel query? */
999 if (xpt_bus_register(sc->hs_sim, dev, 0) != CAM_SUCCESS) {
1000 cam_sim_free(sc->hs_sim, /*free_devq*/TRUE);
1001 mtx_unlock(&sc->hs_lock);
1002 device_printf(dev, "Unable to register SCSI bus\n");
1007 if (xpt_create_path(&sc->hs_path, /*periph*/NULL,
1008 cam_sim_path(sc->hs_sim),
1009 CAM_TARGET_WILDCARD, CAM_LUN_WILDCARD) != CAM_REQ_CMP) {
1010 xpt_bus_deregister(cam_sim_path(sc->hs_sim));
1011 cam_sim_free(sc->hs_sim, /*free_devq*/TRUE);
1012 mtx_unlock(&sc->hs_lock);
1013 device_printf(dev, "Unable to create path\n");
1018 mtx_unlock(&sc->hs_lock);
1020 root_mount_rel(root_mount_token);
1025 root_mount_rel(root_mount_token);
1026 while (!LIST_EMPTY(&sc->hs_free_list)) {
1027 reqp = LIST_FIRST(&sc->hs_free_list);
1028 LIST_REMOVE(reqp, link);
1029 free(reqp, M_DEVBUF);
1032 while (!LIST_EMPTY(&g_hv_sgl_page_pool.free_sgl_list)) {
1033 sgl_node = LIST_FIRST(&g_hv_sgl_page_pool.free_sgl_list);
1034 LIST_REMOVE(sgl_node, link);
1035 for (j = 0; j < VMBUS_CHAN_PRPLIST_MAX; j++) {
1037 (void*)sgl_node->sgl_data->sg_segs[j].ss_paddr) {
1038 free((void*)sgl_node->sgl_data->sg_segs[j].ss_paddr, M_DEVBUF);
1041 sglist_free(sgl_node->sgl_data);
1042 free(sgl_node, M_DEVBUF);
1049 * @brief StorVSC device detach function
1051 * This function is responsible for safely detaching a
1052 * StorVSC device. This includes waiting for inbound responses
1053 * to complete and freeing associated per-device structures.
1055 * @param dev a device
1056 * returns 0 on success
1059 storvsc_detach(device_t dev)
1061 struct storvsc_softc *sc = device_get_softc(dev);
1062 struct hv_storvsc_request *reqp = NULL;
1063 struct hv_sgl_node *sgl_node = NULL;
1066 sc->hs_destroy = TRUE;
1069 * At this point, all outbound traffic should be disabled. We
1070 * only allow inbound traffic (responses) to proceed so that
1071 * outstanding requests can be completed.
1074 sc->hs_drain_notify = TRUE;
1075 sema_wait(&sc->hs_drain_sema);
1076 sc->hs_drain_notify = FALSE;
1079 * Since we have already drained, we don't need to busy wait.
1080 * The call to close the channel will reset the callback
1081 * under the protection of the incoming channel lock.
1084 hv_vmbus_channel_close(sc->hs_chan);
1086 mtx_lock(&sc->hs_lock);
1087 while (!LIST_EMPTY(&sc->hs_free_list)) {
1088 reqp = LIST_FIRST(&sc->hs_free_list);
1089 LIST_REMOVE(reqp, link);
1091 free(reqp, M_DEVBUF);
1093 mtx_unlock(&sc->hs_lock);
1095 while (!LIST_EMPTY(&g_hv_sgl_page_pool.free_sgl_list)) {
1096 sgl_node = LIST_FIRST(&g_hv_sgl_page_pool.free_sgl_list);
1097 LIST_REMOVE(sgl_node, link);
1098 for (j = 0; j < VMBUS_CHAN_PRPLIST_MAX; j++){
1100 (void*)sgl_node->sgl_data->sg_segs[j].ss_paddr) {
1101 free((void*)sgl_node->sgl_data->sg_segs[j].ss_paddr, M_DEVBUF);
1104 sglist_free(sgl_node->sgl_data);
1105 free(sgl_node, M_DEVBUF);
1111 #if HVS_TIMEOUT_TEST
1113 * @brief unit test for timed out operations
1115 * This function provides unit testing capability to simulate
1116 * timed out operations. Recompilation with HV_TIMEOUT_TEST=1
1119 * @param reqp pointer to a request structure
1120 * @param opcode SCSI operation being performed
1121 * @param wait if 1, wait for I/O to complete
1124 storvsc_timeout_test(struct hv_storvsc_request *reqp,
1125 uint8_t opcode, int wait)
1128 union ccb *ccb = reqp->ccb;
1129 struct storvsc_softc *sc = reqp->softc;
1131 if (reqp->vstor_packet.vm_srb.cdb[0] != opcode) {
1136 mtx_lock(&reqp->event.mtx);
1138 ret = hv_storvsc_io_request(sc, reqp);
1141 mtx_unlock(&reqp->event.mtx);
1143 printf("%s: io_request failed with %d.\n",
1145 ccb->ccb_h.status = CAM_PROVIDE_FAIL;
1146 mtx_lock(&sc->hs_lock);
1147 storvsc_free_request(sc, reqp);
1149 mtx_unlock(&sc->hs_lock);
1154 xpt_print(ccb->ccb_h.path,
1155 "%u: %s: waiting for IO return.\n",
1157 ret = cv_timedwait(&reqp->event.cv, &reqp->event.mtx, 60*hz);
1158 mtx_unlock(&reqp->event.mtx);
1159 xpt_print(ccb->ccb_h.path, "%u: %s: %s.\n",
1160 ticks, __func__, (ret == 0)?
1161 "IO return detected" :
1162 "IO return not detected");
1164 * Now both the timer handler and io done are running
1165 * simultaneously. We want to confirm the io done always
1166 * finishes after the timer handler exits. So reqp used by
1167 * timer handler is not freed or stale. Do busy loop for
1168 * another 1/10 second to make sure io done does
1169 * wait for the timer handler to complete.
1172 mtx_lock(&sc->hs_lock);
1173 xpt_print(ccb->ccb_h.path,
1174 "%u: %s: finishing, queue frozen %d, "
1175 "ccb status 0x%x scsi_status 0x%x.\n",
1176 ticks, __func__, sc->hs_frozen,
1178 ccb->csio.scsi_status);
1179 mtx_unlock(&sc->hs_lock);
1182 #endif /* HVS_TIMEOUT_TEST */
1186 * @brief timeout handler for requests
1188 * This function is called as a result of a callout expiring.
1190 * @param arg pointer to a request
1193 storvsc_timeout(void *arg)
1195 struct hv_storvsc_request *reqp = arg;
1196 struct storvsc_softc *sc = reqp->softc;
1197 union ccb *ccb = reqp->ccb;
1199 if (reqp->retries == 0) {
1200 mtx_lock(&sc->hs_lock);
1201 xpt_print(ccb->ccb_h.path,
1202 "%u: IO timed out (req=0x%p), wait for another %u secs.\n",
1203 ticks, reqp, ccb->ccb_h.timeout / 1000);
1204 cam_error_print(ccb, CAM_ESF_ALL, CAM_EPF_ALL);
1205 mtx_unlock(&sc->hs_lock);
1208 callout_reset_sbt(&reqp->callout, SBT_1MS * ccb->ccb_h.timeout,
1209 0, storvsc_timeout, reqp, 0);
1210 #if HVS_TIMEOUT_TEST
1211 storvsc_timeout_test(reqp, SEND_DIAGNOSTIC, 0);
1216 mtx_lock(&sc->hs_lock);
1217 xpt_print(ccb->ccb_h.path,
1218 "%u: IO (reqp = 0x%p) did not return for %u seconds, %s.\n",
1219 ticks, reqp, ccb->ccb_h.timeout * (reqp->retries+1) / 1000,
1220 (sc->hs_frozen == 0)?
1221 "freezing the queue" : "the queue is already frozen");
1222 if (sc->hs_frozen == 0) {
1224 xpt_freeze_simq(xpt_path_sim(ccb->ccb_h.path), 1);
1226 mtx_unlock(&sc->hs_lock);
1228 #if HVS_TIMEOUT_TEST
1229 storvsc_timeout_test(reqp, MODE_SELECT_10, 1);
1235 * @brief StorVSC device poll function
1237 * This function is responsible for servicing requests when
1238 * interrupts are disabled (i.e when we are dumping core.)
1240 * @param sim a pointer to a CAM SCSI interface module
1243 storvsc_poll(struct cam_sim *sim)
1245 struct storvsc_softc *sc = cam_sim_softc(sim);
1247 mtx_assert(&sc->hs_lock, MA_OWNED);
1248 mtx_unlock(&sc->hs_lock);
1249 hv_storvsc_on_channel_callback(sc->hs_chan);
1250 mtx_lock(&sc->hs_lock);
1254 * @brief StorVSC device action function
1256 * This function is responsible for handling SCSI operations which
1257 * are passed from the CAM layer. The requests are in the form of
1258 * CAM control blocks which indicate the action being performed.
1259 * Not all actions require converting the request to a VSCSI protocol
1260 * message - these actions can be responded to by this driver.
1261 * Requests which are destined for a backend storage device are converted
1262 * to a VSCSI protocol message and sent on the channel connection associated
1265 * @param sim pointer to a CAM SCSI interface module
1266 * @param ccb pointer to a CAM control block
1269 storvsc_action(struct cam_sim *sim, union ccb *ccb)
1271 struct storvsc_softc *sc = cam_sim_softc(sim);
1274 mtx_assert(&sc->hs_lock, MA_OWNED);
1275 switch (ccb->ccb_h.func_code) {
1276 case XPT_PATH_INQ: {
1277 struct ccb_pathinq *cpi = &ccb->cpi;
1279 cpi->version_num = 1;
1280 cpi->hba_inquiry = PI_TAG_ABLE|PI_SDTR_ABLE;
1281 cpi->target_sprt = 0;
1282 cpi->hba_misc = PIM_NOBUSRESET;
1283 cpi->hba_eng_cnt = 0;
1284 cpi->max_target = STORVSC_MAX_TARGETS;
1285 cpi->max_lun = sc->hs_drv_props->drv_max_luns_per_target;
1286 cpi->initiator_id = cpi->max_target;
1287 cpi->bus_id = cam_sim_bus(sim);
1288 cpi->base_transfer_speed = 300000;
1289 cpi->transport = XPORT_SAS;
1290 cpi->transport_version = 0;
1291 cpi->protocol = PROTO_SCSI;
1292 cpi->protocol_version = SCSI_REV_SPC2;
1293 strncpy(cpi->sim_vid, "FreeBSD", SIM_IDLEN);
1294 strncpy(cpi->hba_vid, sc->hs_drv_props->drv_name, HBA_IDLEN);
1295 strncpy(cpi->dev_name, cam_sim_name(sim), DEV_IDLEN);
1296 cpi->unit_number = cam_sim_unit(sim);
1298 ccb->ccb_h.status = CAM_REQ_CMP;
1302 case XPT_GET_TRAN_SETTINGS: {
1303 struct ccb_trans_settings *cts = &ccb->cts;
1305 cts->transport = XPORT_SAS;
1306 cts->transport_version = 0;
1307 cts->protocol = PROTO_SCSI;
1308 cts->protocol_version = SCSI_REV_SPC2;
1310 /* enable tag queuing and disconnected mode */
1311 cts->proto_specific.valid = CTS_SCSI_VALID_TQ;
1312 cts->proto_specific.scsi.valid = CTS_SCSI_VALID_TQ;
1313 cts->proto_specific.scsi.flags = CTS_SCSI_FLAGS_TAG_ENB;
1314 cts->xport_specific.valid = CTS_SPI_VALID_DISC;
1315 cts->xport_specific.spi.flags = CTS_SPI_FLAGS_DISC_ENB;
1317 ccb->ccb_h.status = CAM_REQ_CMP;
1321 case XPT_SET_TRAN_SETTINGS: {
1322 ccb->ccb_h.status = CAM_REQ_CMP;
1326 case XPT_CALC_GEOMETRY:{
1327 cam_calc_geometry(&ccb->ccg, 1);
1332 case XPT_RESET_DEV:{
1334 if ((res = hv_storvsc_host_reset(sc)) != 0) {
1335 xpt_print(ccb->ccb_h.path,
1336 "hv_storvsc_host_reset failed with %d\n", res);
1337 ccb->ccb_h.status = CAM_PROVIDE_FAIL;
1341 ccb->ccb_h.status = CAM_REQ_CMP;
1345 xpt_print(ccb->ccb_h.path,
1346 "%s reset not supported.\n",
1347 (ccb->ccb_h.func_code == XPT_RESET_BUS)?
1349 ccb->ccb_h.status = CAM_REQ_INVALID;
1352 #endif /* HVS_HOST_RESET */
1355 case XPT_IMMED_NOTIFY: {
1356 struct hv_storvsc_request *reqp = NULL;
1358 if (ccb->csio.cdb_len == 0) {
1359 panic("cdl_len is 0\n");
1362 if (LIST_EMPTY(&sc->hs_free_list)) {
1363 ccb->ccb_h.status = CAM_REQUEUE_REQ;
1364 if (sc->hs_frozen == 0) {
1366 xpt_freeze_simq(sim, /* count*/1);
1372 reqp = LIST_FIRST(&sc->hs_free_list);
1373 LIST_REMOVE(reqp, link);
1375 bzero(reqp, sizeof(struct hv_storvsc_request));
1378 ccb->ccb_h.status |= CAM_SIM_QUEUED;
1379 if ((res = create_storvsc_request(ccb, reqp)) != 0) {
1380 ccb->ccb_h.status = CAM_REQ_INVALID;
1386 if (ccb->ccb_h.timeout != CAM_TIME_INFINITY) {
1387 callout_init(&reqp->callout, CALLOUT_MPSAFE);
1388 callout_reset_sbt(&reqp->callout,
1389 SBT_1MS * ccb->ccb_h.timeout, 0,
1390 storvsc_timeout, reqp, 0);
1391 #if HVS_TIMEOUT_TEST
1392 cv_init(&reqp->event.cv, "storvsc timeout cv");
1393 mtx_init(&reqp->event.mtx, "storvsc timeout mutex",
1395 switch (reqp->vstor_packet.vm_srb.cdb[0]) {
1396 case MODE_SELECT_10:
1397 case SEND_DIAGNOSTIC:
1398 /* To have timer send the request. */
1403 #endif /* HVS_TIMEOUT_TEST */
1407 if ((res = hv_storvsc_io_request(sc, reqp)) != 0) {
1408 xpt_print(ccb->ccb_h.path,
1409 "hv_storvsc_io_request failed with %d\n", res);
1410 ccb->ccb_h.status = CAM_PROVIDE_FAIL;
1411 storvsc_free_request(sc, reqp);
1419 ccb->ccb_h.status = CAM_REQ_INVALID;
1426 * @brief destroy bounce buffer
1428 * This function is responsible for destroy a Scatter/Gather list
1429 * that create by storvsc_create_bounce_buffer()
1431 * @param sgl- the Scatter/Gather need be destroy
1432 * @param sg_count- page count of the SG list.
1436 storvsc_destroy_bounce_buffer(struct sglist *sgl)
1438 struct hv_sgl_node *sgl_node = NULL;
1439 if (LIST_EMPTY(&g_hv_sgl_page_pool.in_use_sgl_list)) {
1440 printf("storvsc error: not enough in use sgl\n");
1443 sgl_node = LIST_FIRST(&g_hv_sgl_page_pool.in_use_sgl_list);
1444 LIST_REMOVE(sgl_node, link);
1445 sgl_node->sgl_data = sgl;
1446 LIST_INSERT_HEAD(&g_hv_sgl_page_pool.free_sgl_list, sgl_node, link);
1450 * @brief create bounce buffer
1452 * This function is responsible for create a Scatter/Gather list,
1453 * which hold several pages that can be aligned with page size.
1455 * @param seg_count- SG-list segments count
1456 * @param write - if WRITE_TYPE, set SG list page used size to 0,
1457 * otherwise set used size to page size.
1459 * return NULL if create failed
1461 static struct sglist *
1462 storvsc_create_bounce_buffer(uint16_t seg_count, int write)
1465 struct sglist *bounce_sgl = NULL;
1466 unsigned int buf_len = ((write == WRITE_TYPE) ? 0 : PAGE_SIZE);
1467 struct hv_sgl_node *sgl_node = NULL;
1469 /* get struct sglist from free_sgl_list */
1470 if (LIST_EMPTY(&g_hv_sgl_page_pool.free_sgl_list)) {
1471 printf("storvsc error: not enough free sgl\n");
1474 sgl_node = LIST_FIRST(&g_hv_sgl_page_pool.free_sgl_list);
1475 LIST_REMOVE(sgl_node, link);
1476 bounce_sgl = sgl_node->sgl_data;
1477 LIST_INSERT_HEAD(&g_hv_sgl_page_pool.in_use_sgl_list, sgl_node, link);
1479 bounce_sgl->sg_maxseg = seg_count;
1481 if (write == WRITE_TYPE)
1482 bounce_sgl->sg_nseg = 0;
1484 bounce_sgl->sg_nseg = seg_count;
1486 for (i = 0; i < seg_count; i++)
1487 bounce_sgl->sg_segs[i].ss_len = buf_len;
1493 * @brief copy data from SG list to bounce buffer
1495 * This function is responsible for copy data from one SG list's segments
1496 * to another SG list which used as bounce buffer.
1498 * @param bounce_sgl - the destination SG list
1499 * @param orig_sgl - the segment of the source SG list.
1500 * @param orig_sgl_count - the count of segments.
1501 * @param orig_sgl_count - indicate which segment need bounce buffer,
1506 storvsc_copy_sgl_to_bounce_buf(struct sglist *bounce_sgl,
1507 bus_dma_segment_t *orig_sgl,
1508 unsigned int orig_sgl_count,
1511 int src_sgl_idx = 0;
1513 for (src_sgl_idx = 0; src_sgl_idx < orig_sgl_count; src_sgl_idx++) {
1514 if (seg_bits & (1 << src_sgl_idx)) {
1515 memcpy((void*)bounce_sgl->sg_segs[src_sgl_idx].ss_paddr,
1516 (void*)orig_sgl[src_sgl_idx].ds_addr,
1517 orig_sgl[src_sgl_idx].ds_len);
1519 bounce_sgl->sg_segs[src_sgl_idx].ss_len =
1520 orig_sgl[src_sgl_idx].ds_len;
1526 * @brief copy data from SG list which used as bounce to another SG list
1528 * This function is responsible for copy data from one SG list with bounce
1529 * buffer to another SG list's segments.
1531 * @param dest_sgl - the destination SG list's segments
1532 * @param dest_sgl_count - the count of destination SG list's segment.
1533 * @param src_sgl - the source SG list.
1534 * @param seg_bits - indicate which segment used bounce buffer of src SG-list.
1538 storvsc_copy_from_bounce_buf_to_sgl(bus_dma_segment_t *dest_sgl,
1539 unsigned int dest_sgl_count,
1540 struct sglist* src_sgl,
1545 for (sgl_idx = 0; sgl_idx < dest_sgl_count; sgl_idx++) {
1546 if (seg_bits & (1 << sgl_idx)) {
1547 memcpy((void*)(dest_sgl[sgl_idx].ds_addr),
1548 (void*)(src_sgl->sg_segs[sgl_idx].ss_paddr),
1549 src_sgl->sg_segs[sgl_idx].ss_len);
1555 * @brief check SG list with bounce buffer or not
1557 * This function is responsible for check if need bounce buffer for SG list.
1559 * @param sgl - the SG list's segments
1560 * @param sg_count - the count of SG list's segment.
1561 * @param bits - segmengs number that need bounce buffer
1563 * return -1 if SG list needless bounce buffer
1566 storvsc_check_bounce_buffer_sgl(bus_dma_segment_t *sgl,
1567 unsigned int sg_count,
1572 uint64_t phys_addr = 0;
1573 uint64_t tmp_bits = 0;
1574 boolean_t found_hole = FALSE;
1575 boolean_t pre_aligned = TRUE;
1583 phys_addr = vtophys(sgl[0].ds_addr);
1584 offset = phys_addr - trunc_page(phys_addr);
1587 pre_aligned = FALSE;
1591 for (i = 1; i < sg_count; i++) {
1592 phys_addr = vtophys(sgl[i].ds_addr);
1593 offset = phys_addr - trunc_page(phys_addr);
1596 if (FALSE == pre_aligned){
1598 * This segment is aligned, if the previous
1599 * one is not aligned, find a hole
1607 if (phys_addr != vtophys(sgl[i-1].ds_addr +
1610 * Check whether connect to previous
1611 * segment,if not, find the hole
1618 pre_aligned = FALSE;
1631 * @brief Fill in a request structure based on a CAM control block
1633 * Fills in a request structure based on the contents of a CAM control
1634 * block. The request structure holds the payload information for
1635 * VSCSI protocol request.
1637 * @param ccb pointer to a CAM contorl block
1638 * @param reqp pointer to a request structure
1641 create_storvsc_request(union ccb *ccb, struct hv_storvsc_request *reqp)
1643 struct ccb_scsiio *csio = &ccb->csio;
1645 uint32_t bytes_to_copy = 0;
1646 uint32_t pfn_num = 0;
1648 uint64_t not_aligned_seg_bits = 0;
1649 struct hvs_gpa_range *prplist;
1651 /* refer to struct vmscsi_req for meanings of these two fields */
1652 reqp->vstor_packet.u.vm_srb.port =
1653 cam_sim_unit(xpt_path_sim(ccb->ccb_h.path));
1654 reqp->vstor_packet.u.vm_srb.path_id =
1655 cam_sim_bus(xpt_path_sim(ccb->ccb_h.path));
1657 reqp->vstor_packet.u.vm_srb.target_id = ccb->ccb_h.target_id;
1658 reqp->vstor_packet.u.vm_srb.lun = ccb->ccb_h.target_lun;
1660 reqp->vstor_packet.u.vm_srb.cdb_len = csio->cdb_len;
1661 if(ccb->ccb_h.flags & CAM_CDB_POINTER) {
1662 memcpy(&reqp->vstor_packet.u.vm_srb.u.cdb, csio->cdb_io.cdb_ptr,
1665 memcpy(&reqp->vstor_packet.u.vm_srb.u.cdb, csio->cdb_io.cdb_bytes,
1669 switch (ccb->ccb_h.flags & CAM_DIR_MASK) {
1671 reqp->vstor_packet.u.vm_srb.data_in = WRITE_TYPE;
1674 reqp->vstor_packet.u.vm_srb.data_in = READ_TYPE;
1677 reqp->vstor_packet.u.vm_srb.data_in = UNKNOWN_TYPE;
1680 reqp->vstor_packet.u.vm_srb.data_in = UNKNOWN_TYPE;
1684 reqp->sense_data = &csio->sense_data;
1685 reqp->sense_info_len = csio->sense_len;
1689 if (0 == csio->dxfer_len) {
1693 prplist = &reqp->prp_list;
1694 prplist->gpa_range.gpa_len = csio->dxfer_len;
1696 switch (ccb->ccb_h.flags & CAM_DATA_MASK) {
1697 case CAM_DATA_VADDR:
1699 bytes_to_copy = csio->dxfer_len;
1700 phys_addr = vtophys(csio->data_ptr);
1701 prplist->gpa_range.gpa_ofs = phys_addr & PAGE_MASK;
1703 while (bytes_to_copy != 0) {
1704 int bytes, page_offset;
1706 vtophys(&csio->data_ptr[prplist->gpa_range.gpa_len -
1708 pfn = phys_addr >> PAGE_SHIFT;
1709 prplist->gpa_page[pfn_num] = pfn;
1710 page_offset = phys_addr & PAGE_MASK;
1712 bytes = min(PAGE_SIZE - page_offset, bytes_to_copy);
1714 bytes_to_copy -= bytes;
1717 reqp->prp_cnt = pfn_num;
1727 bus_dma_segment_t *storvsc_sglist =
1728 (bus_dma_segment_t *)ccb->csio.data_ptr;
1729 u_int16_t storvsc_sg_count = ccb->csio.sglist_cnt;
1731 printf("Storvsc: get SG I/O operation, %d\n",
1732 reqp->vstor_packet.u.vm_srb.data_in);
1734 if (storvsc_sg_count > VMBUS_CHAN_PRPLIST_MAX){
1735 printf("Storvsc: %d segments is too much, "
1736 "only support %d segments\n",
1737 storvsc_sg_count, VMBUS_CHAN_PRPLIST_MAX);
1742 * We create our own bounce buffer function currently. Idealy
1743 * we should use BUS_DMA(9) framework. But with current BUS_DMA
1744 * code there is no callback API to check the page alignment of
1745 * middle segments before busdma can decide if a bounce buffer
1746 * is needed for particular segment. There is callback,
1747 * "bus_dma_filter_t *filter", but the parrameters are not
1748 * sufficient for storvsc driver.
1750 * Add page alignment check in BUS_DMA(9) callback. Once
1751 * this is complete, switch the following code to use
1752 * BUS_DMA(9) for storvsc bounce buffer support.
1754 /* check if we need to create bounce buffer */
1755 ret = storvsc_check_bounce_buffer_sgl(storvsc_sglist,
1756 storvsc_sg_count, ¬_aligned_seg_bits);
1759 storvsc_create_bounce_buffer(storvsc_sg_count,
1760 reqp->vstor_packet.u.vm_srb.data_in);
1761 if (NULL == reqp->bounce_sgl) {
1762 printf("Storvsc_error: "
1763 "create bounce buffer failed.\n");
1767 reqp->bounce_sgl_count = storvsc_sg_count;
1768 reqp->not_aligned_seg_bits = not_aligned_seg_bits;
1771 * if it is write, we need copy the original data
1774 if (WRITE_TYPE == reqp->vstor_packet.u.vm_srb.data_in) {
1775 storvsc_copy_sgl_to_bounce_buf(
1779 reqp->not_aligned_seg_bits);
1782 /* transfer virtual address to physical frame number */
1783 if (reqp->not_aligned_seg_bits & 0x1){
1785 vtophys(reqp->bounce_sgl->sg_segs[0].ss_paddr);
1788 vtophys(storvsc_sglist[0].ds_addr);
1790 prplist->gpa_range.gpa_ofs = phys_addr & PAGE_MASK;
1792 pfn = phys_addr >> PAGE_SHIFT;
1793 prplist->gpa_page[0] = pfn;
1795 for (i = 1; i < storvsc_sg_count; i++) {
1796 if (reqp->not_aligned_seg_bits & (1 << i)) {
1798 vtophys(reqp->bounce_sgl->sg_segs[i].ss_paddr);
1801 vtophys(storvsc_sglist[i].ds_addr);
1804 pfn = phys_addr >> PAGE_SHIFT;
1805 prplist->gpa_page[i] = pfn;
1809 phys_addr = vtophys(storvsc_sglist[0].ds_addr);
1811 prplist->gpa_range.gpa_ofs = phys_addr & PAGE_MASK;
1813 for (i = 0; i < storvsc_sg_count; i++) {
1814 phys_addr = vtophys(storvsc_sglist[i].ds_addr);
1815 pfn = phys_addr >> PAGE_SHIFT;
1816 prplist->gpa_page[i] = pfn;
1820 /* check the last segment cross boundary or not */
1821 offset = phys_addr & PAGE_MASK;
1823 /* Add one more PRP entry */
1825 vtophys(storvsc_sglist[i-1].ds_addr +
1826 PAGE_SIZE - offset);
1827 pfn = phys_addr >> PAGE_SHIFT;
1828 prplist->gpa_page[i] = pfn;
1832 reqp->bounce_sgl_count = 0;
1837 printf("Unknow flags: %d\n", ccb->ccb_h.flags);
1845 * @brief completion function before returning to CAM
1847 * I/O process has been completed and the result needs
1848 * to be passed to the CAM layer.
1849 * Free resources related to this request.
1851 * @param reqp pointer to a request structure
1854 storvsc_io_done(struct hv_storvsc_request *reqp)
1856 union ccb *ccb = reqp->ccb;
1857 struct ccb_scsiio *csio = &ccb->csio;
1858 struct storvsc_softc *sc = reqp->softc;
1859 struct vmscsi_req *vm_srb = &reqp->vstor_packet.u.vm_srb;
1860 bus_dma_segment_t *ori_sglist = NULL;
1861 int ori_sg_count = 0;
1862 /* destroy bounce buffer if it is used */
1863 if (reqp->bounce_sgl_count) {
1864 ori_sglist = (bus_dma_segment_t *)ccb->csio.data_ptr;
1865 ori_sg_count = ccb->csio.sglist_cnt;
1868 * If it is READ operation, we should copy back the data
1869 * to original SG list.
1871 if (READ_TYPE == reqp->vstor_packet.u.vm_srb.data_in) {
1872 storvsc_copy_from_bounce_buf_to_sgl(ori_sglist,
1875 reqp->not_aligned_seg_bits);
1878 storvsc_destroy_bounce_buffer(reqp->bounce_sgl);
1879 reqp->bounce_sgl_count = 0;
1882 if (reqp->retries > 0) {
1883 mtx_lock(&sc->hs_lock);
1884 #if HVS_TIMEOUT_TEST
1885 xpt_print(ccb->ccb_h.path,
1886 "%u: IO returned after timeout, "
1887 "waking up timer handler if any.\n", ticks);
1888 mtx_lock(&reqp->event.mtx);
1889 cv_signal(&reqp->event.cv);
1890 mtx_unlock(&reqp->event.mtx);
1893 xpt_print(ccb->ccb_h.path,
1894 "%u: IO returned after timeout, "
1895 "stopping timer if any.\n", ticks);
1896 mtx_unlock(&sc->hs_lock);
1901 * callout_drain() will wait for the timer handler to finish
1902 * if it is running. So we don't need any lock to synchronize
1903 * between this routine and the timer handler.
1904 * Note that we need to make sure reqp is not freed when timer
1905 * handler is using or will use it.
1907 if (ccb->ccb_h.timeout != CAM_TIME_INFINITY) {
1908 callout_drain(&reqp->callout);
1912 ccb->ccb_h.status &= ~CAM_SIM_QUEUED;
1913 ccb->ccb_h.status &= ~CAM_STATUS_MASK;
1914 if (vm_srb->scsi_status == SCSI_STATUS_OK) {
1915 const struct scsi_generic *cmd;
1917 if (vm_srb->srb_status != SRB_STATUS_SUCCESS) {
1918 if (vm_srb->srb_status == SRB_STATUS_INVALID_LUN) {
1919 xpt_print(ccb->ccb_h.path, "invalid LUN %d\n",
1922 xpt_print(ccb->ccb_h.path, "Unknown SRB flag: %d\n",
1923 vm_srb->srb_status);
1926 * If there are errors, for example, invalid LUN,
1927 * host will inform VM through SRB status.
1929 ccb->ccb_h.status |= CAM_SEL_TIMEOUT;
1931 ccb->ccb_h.status |= CAM_REQ_CMP;
1934 cmd = (const struct scsi_generic *)
1935 ((ccb->ccb_h.flags & CAM_CDB_POINTER) ?
1936 csio->cdb_io.cdb_ptr : csio->cdb_io.cdb_bytes);
1937 if (cmd->opcode == INQUIRY) {
1938 struct scsi_inquiry_data *inq_data =
1939 (struct scsi_inquiry_data *)csio->data_ptr;
1940 uint8_t *resp_buf = (uint8_t *)csio->data_ptr;
1941 int resp_xfer_len, resp_buf_len, data_len;
1943 /* Get the buffer length reported by host */
1944 resp_xfer_len = vm_srb->transfer_len;
1945 /* Get the available buffer length */
1946 resp_buf_len = resp_xfer_len >= 5 ? resp_buf[4] + 5 : 0;
1947 data_len = (resp_buf_len < resp_xfer_len) ?
1948 resp_buf_len : resp_xfer_len;
1950 if (bootverbose && data_len >= 5) {
1951 xpt_print(ccb->ccb_h.path, "storvsc inquiry "
1952 "(%d) [%x %x %x %x %x ... ]\n", data_len,
1953 resp_buf[0], resp_buf[1], resp_buf[2],
1954 resp_buf[3], resp_buf[4]);
1956 if (vm_srb->srb_status == SRB_STATUS_SUCCESS &&
1957 data_len > SHORT_INQUIRY_LENGTH) {
1960 cam_strvis(vendor, inq_data->vendor,
1961 sizeof(inq_data->vendor), sizeof(vendor));
1964 * XXX: Upgrade SPC2 to SPC3 if host is WIN8 or
1965 * WIN2012 R2 in order to support UNMAP feature.
1967 if (!strncmp(vendor, "Msft", 4) &&
1968 SID_ANSI_REV(inq_data) == SCSI_REV_SPC2 &&
1969 (vmstor_proto_version ==
1970 VMSTOR_PROTOCOL_VERSION_WIN8_1 ||
1971 vmstor_proto_version ==
1972 VMSTOR_PROTOCOL_VERSION_WIN8)) {
1973 inq_data->version = SCSI_REV_SPC3;
1975 xpt_print(ccb->ccb_h.path,
1983 mtx_lock(&sc->hs_lock);
1984 xpt_print(ccb->ccb_h.path,
1985 "storvsc scsi_status = %d\n",
1986 vm_srb->scsi_status);
1987 mtx_unlock(&sc->hs_lock);
1988 ccb->ccb_h.status |= CAM_SCSI_STATUS_ERROR;
1991 ccb->csio.scsi_status = (vm_srb->scsi_status & 0xFF);
1992 ccb->csio.resid = ccb->csio.dxfer_len - vm_srb->transfer_len;
1994 if (reqp->sense_info_len != 0) {
1995 csio->sense_resid = csio->sense_len - reqp->sense_info_len;
1996 ccb->ccb_h.status |= CAM_AUTOSNS_VALID;
1999 mtx_lock(&sc->hs_lock);
2000 if (reqp->softc->hs_frozen == 1) {
2001 xpt_print(ccb->ccb_h.path,
2002 "%u: storvsc unfreezing softc 0x%p.\n",
2003 ticks, reqp->softc);
2004 ccb->ccb_h.status |= CAM_RELEASE_SIMQ;
2005 reqp->softc->hs_frozen = 0;
2007 storvsc_free_request(sc, reqp);
2008 mtx_unlock(&sc->hs_lock);
2010 xpt_done_direct(ccb);
2014 * @brief Free a request structure
2016 * Free a request structure by returning it to the free list
2018 * @param sc pointer to a softc
2019 * @param reqp pointer to a request structure
2022 storvsc_free_request(struct storvsc_softc *sc, struct hv_storvsc_request *reqp)
2025 LIST_INSERT_HEAD(&sc->hs_free_list, reqp, link);
2029 * @brief Determine type of storage device from GUID
2031 * Using the type GUID, determine if this is a StorVSC (paravirtual
2032 * SCSI or BlkVSC (paravirtual IDE) device.
2034 * @param dev a device
2037 static enum hv_storage_type
2038 storvsc_get_storage_type(device_t dev)
2040 device_t parent = device_get_parent(dev);
2042 if (VMBUS_PROBE_GUID(parent, dev, &gBlkVscDeviceType) == 0)
2043 return DRIVER_BLKVSC;
2044 if (VMBUS_PROBE_GUID(parent, dev, &gStorVscDeviceType) == 0)
2045 return DRIVER_STORVSC;
2046 return DRIVER_UNKNOWN;