2 * Copyright (c) 2009 Yahoo! Inc.
3 * Copyright (c) 2011, 2012 LSI Corp.
6 * Redistribution and use in source and binary forms, with or without
7 * modification, are permitted provided that the following conditions
9 * 1. Redistributions of source code must retain the above copyright
10 * notice, this list of conditions and the following disclaimer.
11 * 2. Redistributions in binary form must reproduce the above copyright
12 * notice, this list of conditions and the following disclaimer in the
13 * documentation and/or other materials provided with the distribution.
15 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
16 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
17 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
18 * ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
19 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
20 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
21 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
22 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
23 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
24 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
27 * LSI MPT-Fusion Host Adapter FreeBSD
32 #include <sys/cdefs.h>
33 __FBSDID("$FreeBSD$");
35 /* Communications core for LSI MPT2 */
37 /* TODO Move headers to mpsvar */
38 #include <sys/types.h>
39 #include <sys/param.h>
40 #include <sys/systm.h>
41 #include <sys/kernel.h>
42 #include <sys/selinfo.h>
43 #include <sys/module.h>
47 #include <sys/malloc.h>
49 #include <sys/sysctl.h>
50 #include <sys/endian.h>
51 #include <sys/queue.h>
52 #include <sys/kthread.h>
53 #include <sys/taskqueue.h>
56 #include <machine/bus.h>
57 #include <machine/resource.h>
60 #include <machine/stdarg.h>
63 #include <cam/cam_ccb.h>
64 #include <cam/cam_xpt.h>
65 #include <cam/cam_debug.h>
66 #include <cam/cam_sim.h>
67 #include <cam/cam_xpt_sim.h>
68 #include <cam/cam_xpt_periph.h>
69 #include <cam/cam_periph.h>
70 #include <cam/scsi/scsi_all.h>
71 #include <cam/scsi/scsi_message.h>
72 #if __FreeBSD_version >= 900026
73 #include <cam/scsi/smp_all.h>
76 #include <dev/mps/mpi/mpi2_type.h>
77 #include <dev/mps/mpi/mpi2.h>
78 #include <dev/mps/mpi/mpi2_ioc.h>
79 #include <dev/mps/mpi/mpi2_sas.h>
80 #include <dev/mps/mpi/mpi2_cnfg.h>
81 #include <dev/mps/mpi/mpi2_init.h>
82 #include <dev/mps/mpi/mpi2_tool.h>
83 #include <dev/mps/mps_ioctl.h>
84 #include <dev/mps/mpsvar.h>
85 #include <dev/mps/mps_table.h>
86 #include <dev/mps/mps_sas.h>
88 #define MPSSAS_DISCOVERY_TIMEOUT 20
89 #define MPSSAS_MAX_DISCOVERY_TIMEOUTS 10 /* 200 seconds */
92 * static array to check SCSI OpCode for EEDP protection bits
94 #define PRO_R MPI2_SCSIIO_EEDPFLAGS_CHECK_REMOVE_OP
95 #define PRO_W MPI2_SCSIIO_EEDPFLAGS_INSERT_OP
96 #define PRO_V MPI2_SCSIIO_EEDPFLAGS_INSERT_OP
97 static uint8_t op_code_prot[256] = {
98 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
99 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
100 0, 0, 0, 0, 0, 0, 0, 0, PRO_R, 0, PRO_W, 0, 0, 0, PRO_W, PRO_V,
101 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
102 0, PRO_W, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
103 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
104 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
105 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
106 0, 0, 0, 0, 0, 0, 0, 0, PRO_R, 0, PRO_W, 0, 0, 0, PRO_W, PRO_V,
107 0, 0, 0, PRO_W, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
108 0, 0, 0, 0, 0, 0, 0, 0, PRO_R, 0, PRO_W, 0, 0, 0, PRO_W, PRO_V,
109 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
110 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
111 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
112 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0,
113 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0, 0
116 MALLOC_DEFINE(M_MPSSAS, "MPSSAS", "MPS SAS memory");
118 static void mpssas_discovery_timeout(void *data);
119 static void mpssas_remove_device(struct mps_softc *, struct mps_command *);
120 static void mpssas_remove_complete(struct mps_softc *, struct mps_command *);
121 static void mpssas_action(struct cam_sim *sim, union ccb *ccb);
122 static void mpssas_poll(struct cam_sim *sim);
123 static void mpssas_scsiio_timeout(void *data);
124 static void mpssas_abort_complete(struct mps_softc *sc, struct mps_command *cm);
125 static void mpssas_direct_drive_io(struct mpssas_softc *sassc,
126 struct mps_command *cm, union ccb *ccb);
127 static void mpssas_action_scsiio(struct mpssas_softc *, union ccb *);
128 static void mpssas_scsiio_complete(struct mps_softc *, struct mps_command *);
129 static void mpssas_action_resetdev(struct mpssas_softc *, union ccb *);
130 #if __FreeBSD_version >= 900026
131 static void mpssas_smpio_complete(struct mps_softc *sc, struct mps_command *cm);
132 static void mpssas_send_smpcmd(struct mpssas_softc *sassc, union ccb *ccb,
134 static void mpssas_action_smpio(struct mpssas_softc *sassc, union ccb *ccb);
135 #endif //FreeBSD_version >= 900026
136 static void mpssas_resetdev_complete(struct mps_softc *, struct mps_command *);
137 static int mpssas_send_abort(struct mps_softc *sc, struct mps_command *tm, struct mps_command *cm);
138 static int mpssas_send_reset(struct mps_softc *sc, struct mps_command *tm, uint8_t type);
139 static void mpssas_rescan(struct mpssas_softc *sassc, union ccb *ccb);
140 static void mpssas_rescan_done(struct cam_periph *periph, union ccb *done_ccb);
141 static void mpssas_scanner_thread(void *arg);
142 #if __FreeBSD_version >= 1000006
143 static void mpssas_async(void *callback_arg, uint32_t code,
144 struct cam_path *path, void *arg);
146 static void mpssas_check_eedp(struct mpssas_softc *sassc);
147 static void mpssas_read_cap_done(struct cam_periph *periph, union ccb *done_ccb);
149 static int mpssas_send_portenable(struct mps_softc *sc);
150 static void mpssas_portenable_complete(struct mps_softc *sc,
151 struct mps_command *cm);
153 struct mpssas_target *
154 mpssas_find_target_by_handle(struct mpssas_softc *sassc, int start, uint16_t handle)
156 struct mpssas_target *target;
159 for (i = start; i < sassc->sc->facts->MaxTargets; i++) {
160 target = &sassc->targets[i];
161 if (target->handle == handle)
168 /* we need to freeze the simq during attach and diag reset, to avoid failing
169 * commands before device handles have been found by discovery. Since
170 * discovery involves reading config pages and possibly sending commands,
171 * discovery actions may continue even after we receive the end of discovery
172 * event, so refcount discovery actions instead of assuming we can unfreeze
173 * the simq when we get the event.
176 mpssas_startup_increment(struct mpssas_softc *sassc)
178 if ((sassc->flags & MPSSAS_IN_STARTUP) != 0) {
179 if (sassc->startup_refcount++ == 0) {
180 /* just starting, freeze the simq */
181 mps_dprint(sassc->sc, MPS_INFO,
182 "%s freezing simq\n", __func__);
183 xpt_freeze_simq(sassc->sim, 1);
185 mps_dprint(sassc->sc, MPS_TRACE, "%s refcount %u\n", __func__,
186 sassc->startup_refcount);
191 mpssas_startup_decrement(struct mpssas_softc *sassc)
193 if ((sassc->flags & MPSSAS_IN_STARTUP) != 0) {
194 if (--sassc->startup_refcount == 0) {
195 /* finished all discovery-related actions, release
196 * the simq and rescan for the latest topology.
198 mps_dprint(sassc->sc, MPS_INFO,
199 "%s releasing simq\n", __func__);
200 sassc->flags &= ~MPSSAS_IN_STARTUP;
201 xpt_release_simq(sassc->sim, 1);
202 mpssas_rescan_target(sassc->sc, NULL);
204 mps_dprint(sassc->sc, MPS_TRACE, "%s refcount %u\n", __func__,
205 sassc->startup_refcount);
209 /* LSI's firmware requires us to stop sending commands when we're doing task
210 * management, so refcount the TMs and keep the simq frozen when any are in
214 mpssas_alloc_tm(struct mps_softc *sc)
216 struct mps_command *tm;
218 tm = mps_alloc_high_priority_command(sc);
220 if (sc->sassc->tm_count++ == 0) {
221 mps_printf(sc, "%s freezing simq\n", __func__);
222 xpt_freeze_simq(sc->sassc->sim, 1);
224 mps_dprint(sc, MPS_TRACE, "%s tm_count %u\n", __func__,
225 sc->sassc->tm_count);
231 mpssas_free_tm(struct mps_softc *sc, struct mps_command *tm)
236 /* if there are no TMs in use, we can release the simq. We use our
237 * own refcount so that it's easier for a diag reset to cleanup and
240 if (--sc->sassc->tm_count == 0) {
241 mps_printf(sc, "%s releasing simq\n", __func__);
242 xpt_release_simq(sc->sassc->sim, 1);
244 mps_dprint(sc, MPS_TRACE, "%s tm_count %u\n", __func__,
245 sc->sassc->tm_count);
247 mps_free_high_priority_command(sc, tm);
252 mpssas_rescan_target(struct mps_softc *sc, struct mpssas_target *targ)
254 struct mpssas_softc *sassc = sc->sassc;
256 target_id_t targetid;
259 pathid = cam_sim_path(sassc->sim);
261 targetid = CAM_TARGET_WILDCARD;
263 targetid = targ - sassc->targets;
266 * Allocate a CCB and schedule a rescan.
268 ccb = xpt_alloc_ccb_nowait();
270 mps_dprint(sc, MPS_FAULT, "unable to alloc CCB for rescan\n");
274 if (xpt_create_path(&ccb->ccb_h.path, xpt_periph, pathid,
275 targetid, CAM_LUN_WILDCARD) != CAM_REQ_CMP) {
276 mps_dprint(sc, MPS_FAULT, "unable to create path for rescan\n");
281 if (targetid == CAM_TARGET_WILDCARD)
282 ccb->ccb_h.func_code = XPT_SCAN_BUS;
284 ccb->ccb_h.func_code = XPT_SCAN_TGT;
286 mps_dprint(sc, MPS_TRACE, "%s targetid %u\n", __func__, targetid);
287 mpssas_rescan(sassc, ccb);
291 mpssas_log_command(struct mps_command *cm, const char *fmt, ...)
301 sbuf_new(&sb, str, sizeof(str), 0);
305 if (cm->cm_ccb != NULL) {
306 xpt_path_string(cm->cm_ccb->csio.ccb_h.path, path_str,
308 sbuf_cat(&sb, path_str);
309 if (cm->cm_ccb->ccb_h.func_code == XPT_SCSI_IO) {
310 scsi_command_string(&cm->cm_ccb->csio, &sb);
311 sbuf_printf(&sb, "length %d ",
312 cm->cm_ccb->csio.dxfer_len);
316 sbuf_printf(&sb, "(noperiph:%s%d:%u:%u:%u): ",
317 cam_sim_name(cm->cm_sc->sassc->sim),
318 cam_sim_unit(cm->cm_sc->sassc->sim),
319 cam_sim_bus(cm->cm_sc->sassc->sim),
320 cm->cm_targ ? cm->cm_targ->tid : 0xFFFFFFFF,
324 sbuf_printf(&sb, "SMID %u ", cm->cm_desc.Default.SMID);
325 sbuf_vprintf(&sb, fmt, ap);
327 printf("%s", sbuf_data(&sb));
334 mpssas_remove_volume(struct mps_softc *sc, struct mps_command *tm)
336 MPI2_SCSI_TASK_MANAGE_REPLY *reply;
337 struct mpssas_target *targ;
340 mps_dprint(sc, MPS_INFO, "%s\n", __func__);
342 reply = (MPI2_SCSI_TASK_MANAGE_REPLY *)tm->cm_reply;
343 handle = (uint16_t)(uintptr_t)tm->cm_complete_data;
347 /* XXX retry the remove after the diag reset completes? */
348 mps_printf(sc, "%s NULL reply reseting device 0x%04x\n",
350 mpssas_free_tm(sc, tm);
354 if (reply->IOCStatus != MPI2_IOCSTATUS_SUCCESS) {
355 mps_printf(sc, "IOCStatus = 0x%x while resetting device 0x%x\n",
356 reply->IOCStatus, handle);
357 mpssas_free_tm(sc, tm);
361 mps_printf(sc, "Reset aborted %u commands\n", reply->TerminationCount);
362 mps_free_reply(sc, tm->cm_reply_data);
363 tm->cm_reply = NULL; /* Ensures the reply won't get re-freed */
365 mps_printf(sc, "clearing target %u handle 0x%04x\n", targ->tid, handle);
368 * Don't clear target if remove fails because things will get confusing.
369 * Leave the devname and sasaddr intact so that we know to avoid reusing
370 * this target id if possible, and so we can assign the same target id
371 * to this device if it comes back in the future.
373 if (reply->IOCStatus == MPI2_IOCSTATUS_SUCCESS) {
376 targ->encl_handle = 0x0;
377 targ->encl_slot = 0x0;
378 targ->exp_dev_handle = 0x0;
380 targ->linkrate = 0x0;
385 mpssas_free_tm(sc, tm);
390 * No Need to call "MPI2_SAS_OP_REMOVE_DEVICE" For Volume removal.
391 * Otherwise Volume Delete is same as Bare Drive Removal.
394 mpssas_prepare_volume_remove(struct mpssas_softc *sassc, uint16_t handle)
396 MPI2_SCSI_TASK_MANAGE_REQUEST *req;
397 struct mps_softc *sc;
398 struct mps_command *cm;
399 struct mpssas_target *targ = NULL;
401 mps_dprint(sassc->sc, MPS_INFO, "%s\n", __func__);
406 * If this is a WD controller, determine if the disk should be exposed
407 * to the OS or not. If disk should be exposed, return from this
408 * function without doing anything.
410 if (sc->WD_available && (sc->WD_hide_expose ==
411 MPS_WD_EXPOSE_ALWAYS)) {
416 targ = mpssas_find_target_by_handle(sassc, 0, handle);
418 /* FIXME: what is the action? */
419 /* We don't know about this device? */
420 printf("%s %d : invalid handle 0x%x \n", __func__,__LINE__, handle);
424 targ->flags |= MPSSAS_TARGET_INREMOVAL;
426 cm = mpssas_alloc_tm(sc);
428 mps_printf(sc, "%s: command alloc failure\n", __func__);
432 mpssas_rescan_target(sc, targ);
434 req = (MPI2_SCSI_TASK_MANAGE_REQUEST *)cm->cm_req;
435 req->DevHandle = targ->handle;
436 req->Function = MPI2_FUNCTION_SCSI_TASK_MGMT;
437 req->TaskType = MPI2_SCSITASKMGMT_TASKTYPE_TARGET_RESET;
439 /* SAS Hard Link Reset / SATA Link Reset */
440 req->MsgFlags = MPI2_SCSITASKMGMT_MSGFLAGS_LINK_RESET;
444 cm->cm_desc.HighPriority.RequestFlags =
445 MPI2_REQ_DESCRIPT_FLAGS_HIGH_PRIORITY;
446 cm->cm_complete = mpssas_remove_volume;
447 cm->cm_complete_data = (void *)(uintptr_t)handle;
448 mps_map_command(sc, cm);
452 * The MPT2 firmware performs debounce on the link to avoid transient link
453 * errors and false removals. When it does decide that link has been lost
454 * and a device need to go away, it expects that the host will perform a
455 * target reset and then an op remove. The reset has the side-effect of
456 * aborting any outstanding requests for the device, which is required for
457 * the op-remove to succeed. It's not clear if the host should check for
458 * the device coming back alive after the reset.
461 mpssas_prepare_remove(struct mpssas_softc *sassc, uint16_t handle)
463 MPI2_SCSI_TASK_MANAGE_REQUEST *req;
464 struct mps_softc *sc;
465 struct mps_command *cm;
466 struct mpssas_target *targ = NULL;
468 mps_dprint(sassc->sc, MPS_INFO, "%s\n", __func__);
472 targ = mpssas_find_target_by_handle(sassc, 0, handle);
474 /* FIXME: what is the action? */
475 /* We don't know about this device? */
476 printf("%s %d : invalid handle 0x%x \n", __func__,__LINE__, handle);
480 targ->flags |= MPSSAS_TARGET_INREMOVAL;
482 cm = mpssas_alloc_tm(sc);
484 mps_printf(sc, "%s: command alloc failure\n", __func__);
488 mpssas_rescan_target(sc, targ);
490 req = (MPI2_SCSI_TASK_MANAGE_REQUEST *)cm->cm_req;
491 memset(req, 0, sizeof(*req));
492 req->DevHandle = htole16(targ->handle);
493 req->Function = MPI2_FUNCTION_SCSI_TASK_MGMT;
494 req->TaskType = MPI2_SCSITASKMGMT_TASKTYPE_TARGET_RESET;
496 /* SAS Hard Link Reset / SATA Link Reset */
497 req->MsgFlags = MPI2_SCSITASKMGMT_MSGFLAGS_LINK_RESET;
501 cm->cm_desc.HighPriority.RequestFlags = MPI2_REQ_DESCRIPT_FLAGS_HIGH_PRIORITY;
502 cm->cm_complete = mpssas_remove_device;
503 cm->cm_complete_data = (void *)(uintptr_t)handle;
504 mps_map_command(sc, cm);
508 mpssas_remove_device(struct mps_softc *sc, struct mps_command *tm)
510 MPI2_SCSI_TASK_MANAGE_REPLY *reply;
511 MPI2_SAS_IOUNIT_CONTROL_REQUEST *req;
512 struct mpssas_target *targ;
513 struct mps_command *next_cm;
516 mps_dprint(sc, MPS_INFO, "%s\n", __func__);
518 reply = (MPI2_SCSI_TASK_MANAGE_REPLY *)tm->cm_reply;
519 handle = (uint16_t)(uintptr_t)tm->cm_complete_data;
523 * Currently there should be no way we can hit this case. It only
524 * happens when we have a failure to allocate chain frames, and
525 * task management commands don't have S/G lists.
527 if ((tm->cm_flags & MPS_CM_FLAGS_ERROR_MASK) != 0) {
528 mps_printf(sc, "%s: cm_flags = %#x for remove of handle %#04x! "
529 "This should not happen!\n", __func__, tm->cm_flags,
531 mpssas_free_tm(sc, tm);
536 /* XXX retry the remove after the diag reset completes? */
537 mps_printf(sc, "%s NULL reply reseting device 0x%04x\n",
539 mpssas_free_tm(sc, tm);
543 if (le16toh(reply->IOCStatus) != MPI2_IOCSTATUS_SUCCESS) {
544 mps_printf(sc, "IOCStatus = 0x%x while resetting device 0x%x\n",
545 le16toh(reply->IOCStatus), handle);
546 mpssas_free_tm(sc, tm);
550 mps_dprint(sc, MPS_INFO, "Reset aborted %u commands\n",
551 le32toh(reply->TerminationCount));
552 mps_free_reply(sc, tm->cm_reply_data);
553 tm->cm_reply = NULL; /* Ensures the reply won't get re-freed */
555 /* Reuse the existing command */
556 req = (MPI2_SAS_IOUNIT_CONTROL_REQUEST *)tm->cm_req;
557 memset(req, 0, sizeof(*req));
558 req->Function = MPI2_FUNCTION_SAS_IO_UNIT_CONTROL;
559 req->Operation = MPI2_SAS_OP_REMOVE_DEVICE;
560 req->DevHandle = htole16(handle);
562 tm->cm_desc.Default.RequestFlags = MPI2_REQ_DESCRIPT_FLAGS_DEFAULT_TYPE;
563 tm->cm_complete = mpssas_remove_complete;
564 tm->cm_complete_data = (void *)(uintptr_t)handle;
566 mps_map_command(sc, tm);
568 mps_dprint(sc, MPS_INFO, "clearing target %u handle 0x%04x\n",
570 TAILQ_FOREACH_SAFE(tm, &targ->commands, cm_link, next_cm) {
573 mps_dprint(sc, MPS_INFO, "Completing missed command %p\n", tm);
574 ccb = tm->cm_complete_data;
575 ccb->ccb_h.status = CAM_DEV_NOT_THERE;
576 mpssas_scsiio_complete(sc, tm);
581 mpssas_remove_complete(struct mps_softc *sc, struct mps_command *tm)
583 MPI2_SAS_IOUNIT_CONTROL_REPLY *reply;
585 struct mpssas_target *targ;
586 struct mpssas_lun *lun;
588 mps_dprint(sc, MPS_INFO, "%s\n", __func__);
590 reply = (MPI2_SAS_IOUNIT_CONTROL_REPLY *)tm->cm_reply;
591 handle = (uint16_t)(uintptr_t)tm->cm_complete_data;
594 * Currently there should be no way we can hit this case. It only
595 * happens when we have a failure to allocate chain frames, and
596 * task management commands don't have S/G lists.
598 if ((tm->cm_flags & MPS_CM_FLAGS_ERROR_MASK) != 0) {
599 mps_printf(sc, "%s: cm_flags = %#x for remove of handle %#04x! "
600 "This should not happen!\n", __func__, tm->cm_flags,
602 mpssas_free_tm(sc, tm);
607 /* most likely a chip reset */
608 mps_printf(sc, "%s NULL reply removing device 0x%04x\n",
610 mpssas_free_tm(sc, tm);
614 mps_printf(sc, "%s on handle 0x%04x, IOCStatus= 0x%x\n", __func__,
615 handle, le16toh(reply->IOCStatus));
618 * Don't clear target if remove fails because things will get confusing.
619 * Leave the devname and sasaddr intact so that we know to avoid reusing
620 * this target id if possible, and so we can assign the same target id
621 * to this device if it comes back in the future.
623 if (le16toh(reply->IOCStatus) == MPI2_IOCSTATUS_SUCCESS) {
626 targ->encl_handle = 0x0;
627 targ->encl_slot = 0x0;
628 targ->exp_dev_handle = 0x0;
630 targ->linkrate = 0x0;
634 while(!SLIST_EMPTY(&targ->luns)) {
635 lun = SLIST_FIRST(&targ->luns);
636 SLIST_REMOVE_HEAD(&targ->luns, lun_link);
642 mpssas_free_tm(sc, tm);
646 mpssas_register_events(struct mps_softc *sc)
648 u32 events[MPI2_EVENT_NOTIFY_EVENTMASK_WORDS];
651 setbit(events, MPI2_EVENT_SAS_DEVICE_STATUS_CHANGE);
652 setbit(events, MPI2_EVENT_SAS_DISCOVERY);
653 setbit(events, MPI2_EVENT_SAS_BROADCAST_PRIMITIVE);
654 setbit(events, MPI2_EVENT_SAS_INIT_DEVICE_STATUS_CHANGE);
655 setbit(events, MPI2_EVENT_SAS_INIT_TABLE_OVERFLOW);
656 setbit(events, MPI2_EVENT_SAS_TOPOLOGY_CHANGE_LIST);
657 setbit(events, MPI2_EVENT_SAS_ENCL_DEVICE_STATUS_CHANGE);
658 setbit(events, MPI2_EVENT_IR_CONFIGURATION_CHANGE_LIST);
659 setbit(events, MPI2_EVENT_IR_VOLUME);
660 setbit(events, MPI2_EVENT_IR_PHYSICAL_DISK);
661 setbit(events, MPI2_EVENT_IR_OPERATION_STATUS);
662 setbit(events, MPI2_EVENT_LOG_ENTRY_ADDED);
664 mps_register_events(sc, events, mpssas_evt_handler, NULL,
665 &sc->sassc->mpssas_eh);
671 mps_attach_sas(struct mps_softc *sc)
673 struct mpssas_softc *sassc;
674 #if __FreeBSD_version >= 1000006
679 mps_dprint(sc, MPS_TRACE, "%s\n", __func__);
681 sassc = malloc(sizeof(struct mpssas_softc), M_MPT2, M_WAITOK|M_ZERO);
683 device_printf(sc->mps_dev, "Cannot allocate memory %s %d\n",
687 sassc->targets = malloc(sizeof(struct mpssas_target) *
688 sc->facts->MaxTargets, M_MPT2, M_WAITOK|M_ZERO);
689 if(!sassc->targets) {
690 device_printf(sc->mps_dev, "Cannot allocate memory %s %d\n",
698 if ((sassc->devq = cam_simq_alloc(sc->num_reqs)) == NULL) {
699 mps_dprint(sc, MPS_FAULT, "Cannot allocate SIMQ\n");
704 unit = device_get_unit(sc->mps_dev);
705 sassc->sim = cam_sim_alloc(mpssas_action, mpssas_poll, "mps", sassc,
706 unit, &sc->mps_mtx, sc->num_reqs, sc->num_reqs, sassc->devq);
707 if (sassc->sim == NULL) {
708 mps_dprint(sc, MPS_FAULT, "Cannot allocate SIM\n");
713 TAILQ_INIT(&sassc->ev_queue);
715 /* Initialize taskqueue for Event Handling */
716 TASK_INIT(&sassc->ev_task, 0, mpssas_firmware_event_work, sc);
717 sassc->ev_tq = taskqueue_create("mps_taskq", M_NOWAIT | M_ZERO,
718 taskqueue_thread_enqueue, &sassc->ev_tq);
720 /* Run the task queue with lowest priority */
721 taskqueue_start_threads(&sassc->ev_tq, 1, 255, "%s taskq",
722 device_get_nameunit(sc->mps_dev));
724 TAILQ_INIT(&sassc->ccb_scanq);
725 error = mps_kproc_create(mpssas_scanner_thread, sassc,
726 &sassc->rescan_thread, 0, 0, "mps_scan%d", unit);
728 mps_printf(sc, "Error %d starting rescan thread\n", error);
733 sassc->flags |= MPSSAS_SCANTHREAD;
736 * XXX There should be a bus for every port on the adapter, but since
737 * we're just going to fake the topology for now, we'll pretend that
738 * everything is just a target on a single bus.
740 if ((error = xpt_bus_register(sassc->sim, sc->mps_dev, 0)) != 0) {
741 mps_dprint(sc, MPS_FAULT, "Error %d registering SCSI bus\n",
748 * Assume that discovery events will start right away. Freezing
749 * the simq will prevent the CAM boottime scanner from running
750 * before discovery is complete.
752 sassc->flags |= MPSSAS_IN_STARTUP | MPSSAS_IN_DISCOVERY;
753 xpt_freeze_simq(sassc->sim, 1);
754 sc->sassc->startup_refcount = 0;
756 callout_init(&sassc->discovery_callout, 1 /*mpsafe*/);
757 sassc->discovery_timeouts = 0;
761 #if __FreeBSD_version >= 1000006
762 status = xpt_register_async(AC_ADVINFO_CHANGED, mpssas_async, sc, NULL);
763 if (status != CAM_REQ_CMP) {
764 mps_printf(sc, "Error %#x registering async handler for "
765 "AC_ADVINFO_CHANGED events\n", status);
771 mpssas_register_events(sc);
779 mps_detach_sas(struct mps_softc *sc)
781 struct mpssas_softc *sassc;
782 struct mpssas_lun *lun, *lun_tmp;
783 struct mpssas_target *targ;
786 mps_dprint(sc, MPS_INFO, "%s\n", __func__);
788 if (sc->sassc == NULL)
792 mps_deregister_events(sc, sassc->mpssas_eh);
795 * Drain and free the event handling taskqueue with the lock
796 * unheld so that any parallel processing tasks drain properly
797 * without deadlocking.
799 if (sassc->ev_tq != NULL)
800 taskqueue_free(sassc->ev_tq);
802 /* Make sure CAM doesn't wedge if we had to bail out early. */
805 /* Deregister our async handler */
806 #if __FreeBSD_version >= 1000006
807 xpt_register_async(0, mpssas_async, sc, NULL);
810 if (sassc->flags & MPSSAS_IN_STARTUP)
811 xpt_release_simq(sassc->sim, 1);
813 if (sassc->sim != NULL) {
814 xpt_bus_deregister(cam_sim_path(sassc->sim));
815 cam_sim_free(sassc->sim, FALSE);
818 if (sassc->flags & MPSSAS_SCANTHREAD) {
819 sassc->flags |= MPSSAS_SHUTDOWN;
820 wakeup(&sassc->ccb_scanq);
822 if (sassc->flags & MPSSAS_SCANTHREAD) {
823 msleep(&sassc->flags, &sc->mps_mtx, PRIBIO,
824 "mps_shutdown", 30 * hz);
829 mps_dprint(sc, MPS_INFO, "%s:%d\n", __func__,__LINE__);
830 if (sassc->devq != NULL)
831 cam_simq_free(sassc->devq);
833 for(i=0; i< sc->facts->MaxTargets ;i++) {
834 targ = &sassc->targets[i];
835 SLIST_FOREACH_SAFE(lun, &targ->luns, lun_link, lun_tmp) {
839 free(sassc->targets, M_MPT2);
847 mpssas_discovery_end(struct mpssas_softc *sassc)
849 struct mps_softc *sc = sassc->sc;
851 mps_dprint(sc, MPS_TRACE, "%s\n", __func__);
853 if (sassc->flags & MPSSAS_DISCOVERY_TIMEOUT_PENDING)
854 callout_stop(&sassc->discovery_callout);
859 mpssas_discovery_timeout(void *data)
861 struct mpssas_softc *sassc = data;
862 struct mps_softc *sc;
865 mps_dprint(sc, MPS_TRACE, "%s\n", __func__);
869 "Timeout waiting for discovery, interrupts may not be working!\n");
870 sassc->flags &= ~MPSSAS_DISCOVERY_TIMEOUT_PENDING;
872 /* Poll the hardware for events in case interrupts aren't working */
875 mps_printf(sassc->sc,
876 "Finished polling after discovery timeout at %d\n", ticks);
878 if ((sassc->flags & MPSSAS_IN_DISCOVERY) == 0) {
879 mpssas_discovery_end(sassc);
881 if (sassc->discovery_timeouts < MPSSAS_MAX_DISCOVERY_TIMEOUTS) {
882 sassc->flags |= MPSSAS_DISCOVERY_TIMEOUT_PENDING;
883 callout_reset(&sassc->discovery_callout,
884 MPSSAS_DISCOVERY_TIMEOUT * hz,
885 mpssas_discovery_timeout, sassc);
886 sassc->discovery_timeouts++;
888 mps_dprint(sassc->sc, MPS_FAULT,
889 "Discovery timed out, continuing.\n");
890 sassc->flags &= ~MPSSAS_IN_DISCOVERY;
891 mpssas_discovery_end(sassc);
899 mpssas_action(struct cam_sim *sim, union ccb *ccb)
901 struct mpssas_softc *sassc;
903 sassc = cam_sim_softc(sim);
905 mps_dprint(sassc->sc, MPS_TRACE, "%s func 0x%x\n", __func__,
906 ccb->ccb_h.func_code);
907 mtx_assert(&sassc->sc->mps_mtx, MA_OWNED);
909 switch (ccb->ccb_h.func_code) {
912 struct ccb_pathinq *cpi = &ccb->cpi;
914 cpi->version_num = 1;
915 cpi->hba_inquiry = PI_SDTR_ABLE|PI_TAG_ABLE|PI_WIDE_16;
916 cpi->target_sprt = 0;
917 cpi->hba_misc = PIM_NOBUSRESET;
918 cpi->hba_eng_cnt = 0;
919 cpi->max_target = sassc->sc->facts->MaxTargets - 1;
921 cpi->initiator_id = 255;
922 strncpy(cpi->sim_vid, "FreeBSD", SIM_IDLEN);
923 strncpy(cpi->hba_vid, "LSILogic", HBA_IDLEN);
924 strncpy(cpi->dev_name, cam_sim_name(sim), DEV_IDLEN);
925 cpi->unit_number = cam_sim_unit(sim);
926 cpi->bus_id = cam_sim_bus(sim);
927 cpi->base_transfer_speed = 150000;
928 cpi->transport = XPORT_SAS;
929 cpi->transport_version = 0;
930 cpi->protocol = PROTO_SCSI;
931 cpi->protocol_version = SCSI_REV_SPC;
932 #if __FreeBSD_version >= 800001
934 * XXX KDM where does this number come from?
936 cpi->maxio = 256 * 1024;
938 cpi->ccb_h.status = CAM_REQ_CMP;
941 case XPT_GET_TRAN_SETTINGS:
943 struct ccb_trans_settings *cts;
944 struct ccb_trans_settings_sas *sas;
945 struct ccb_trans_settings_scsi *scsi;
946 struct mpssas_target *targ;
949 sas = &cts->xport_specific.sas;
950 scsi = &cts->proto_specific.scsi;
952 targ = &sassc->targets[cts->ccb_h.target_id];
953 if (targ->handle == 0x0) {
954 cts->ccb_h.status = CAM_SEL_TIMEOUT;
958 cts->protocol_version = SCSI_REV_SPC2;
959 cts->transport = XPORT_SAS;
960 cts->transport_version = 0;
962 sas->valid = CTS_SAS_VALID_SPEED;
963 switch (targ->linkrate) {
965 sas->bitrate = 150000;
968 sas->bitrate = 300000;
971 sas->bitrate = 600000;
977 cts->protocol = PROTO_SCSI;
978 scsi->valid = CTS_SCSI_VALID_TQ;
979 scsi->flags = CTS_SCSI_FLAGS_TAG_ENB;
981 cts->ccb_h.status = CAM_REQ_CMP;
984 case XPT_CALC_GEOMETRY:
985 cam_calc_geometry(&ccb->ccg, /*extended*/1);
986 ccb->ccb_h.status = CAM_REQ_CMP;
989 mps_printf(sassc->sc, "mpssas_action XPT_RESET_DEV\n");
990 mpssas_action_resetdev(sassc, ccb);
995 mps_printf(sassc->sc, "mpssas_action faking success for "
997 ccb->ccb_h.status = CAM_REQ_CMP;
1000 mpssas_action_scsiio(sassc, ccb);
1002 #if __FreeBSD_version >= 900026
1004 mpssas_action_smpio(sassc, ccb);
1008 ccb->ccb_h.status = CAM_FUNC_NOTAVAIL;
1016 mpssas_announce_reset(struct mps_softc *sc, uint32_t ac_code,
1017 target_id_t target_id, lun_id_t lun_id)
1019 path_id_t path_id = cam_sim_path(sc->sassc->sim);
1020 struct cam_path *path;
1022 mps_printf(sc, "%s code %x target %d lun %d\n", __func__,
1023 ac_code, target_id, lun_id);
1025 if (xpt_create_path(&path, NULL,
1026 path_id, target_id, lun_id) != CAM_REQ_CMP) {
1027 mps_printf(sc, "unable to create path for reset "
1032 xpt_async(ac_code, path, NULL);
1033 xpt_free_path(path);
1037 mpssas_complete_all_commands(struct mps_softc *sc)
1039 struct mps_command *cm;
1043 mps_printf(sc, "%s\n", __func__);
1044 mtx_assert(&sc->mps_mtx, MA_OWNED);
1046 /* complete all commands with a NULL reply */
1047 for (i = 1; i < sc->num_reqs; i++) {
1048 cm = &sc->commands[i];
1049 cm->cm_reply = NULL;
1052 if (cm->cm_flags & MPS_CM_FLAGS_POLLED)
1053 cm->cm_flags |= MPS_CM_FLAGS_COMPLETE;
1055 if (cm->cm_complete != NULL) {
1056 mpssas_log_command(cm,
1057 "completing cm %p state %x ccb %p for diag reset\n",
1058 cm, cm->cm_state, cm->cm_ccb);
1060 cm->cm_complete(sc, cm);
1064 if (cm->cm_flags & MPS_CM_FLAGS_WAKEUP) {
1065 mpssas_log_command(cm,
1066 "waking up cm %p state %x ccb %p for diag reset\n",
1067 cm, cm->cm_state, cm->cm_ccb);
1072 if ((completed == 0) && (cm->cm_state != MPS_CM_STATE_FREE)) {
1073 /* this should never happen, but if it does, log */
1074 mpssas_log_command(cm,
1075 "cm %p state %x flags 0x%x ccb %p during diag "
1076 "reset\n", cm, cm->cm_state, cm->cm_flags,
1083 mpssas_handle_reinit(struct mps_softc *sc)
1087 /* Go back into startup mode and freeze the simq, so that CAM
1088 * doesn't send any commands until after we've rediscovered all
1089 * targets and found the proper device handles for them.
1091 * After the reset, portenable will trigger discovery, and after all
1092 * discovery-related activities have finished, the simq will be
1095 mps_printf(sc, "%s startup\n", __func__);
1096 sc->sassc->flags |= MPSSAS_IN_STARTUP;
1097 sc->sassc->flags |= MPSSAS_IN_DISCOVERY;
1098 xpt_freeze_simq(sc->sassc->sim, 1);
1100 /* notify CAM of a bus reset */
1101 mpssas_announce_reset(sc, AC_BUS_RESET, CAM_TARGET_WILDCARD,
1104 /* complete and cleanup after all outstanding commands */
1105 mpssas_complete_all_commands(sc);
1107 mps_printf(sc, "%s startup %u tm %u after command completion\n",
1108 __func__, sc->sassc->startup_refcount, sc->sassc->tm_count);
1111 * The simq was explicitly frozen above, so set the refcount to 0.
1112 * The simq will be explicitly released after port enable completes.
1114 sc->sassc->startup_refcount = 0;
1116 /* zero all the target handles, since they may change after the
1117 * reset, and we have to rediscover all the targets and use the new
1120 for (i = 0; i < sc->facts->MaxTargets; i++) {
1121 if (sc->sassc->targets[i].outstanding != 0)
1122 mps_printf(sc, "target %u outstanding %u\n",
1123 i, sc->sassc->targets[i].outstanding);
1124 sc->sassc->targets[i].handle = 0x0;
1125 sc->sassc->targets[i].exp_dev_handle = 0x0;
1126 sc->sassc->targets[i].outstanding = 0;
1127 sc->sassc->targets[i].flags = MPSSAS_TARGET_INDIAGRESET;
1131 mpssas_tm_timeout(void *data)
1133 struct mps_command *tm = data;
1134 struct mps_softc *sc = tm->cm_sc;
1136 mtx_assert(&sc->mps_mtx, MA_OWNED);
1138 mpssas_log_command(tm, "task mgmt %p timed out\n", tm);
1143 mpssas_logical_unit_reset_complete(struct mps_softc *sc, struct mps_command *tm)
1145 MPI2_SCSI_TASK_MANAGE_REPLY *reply;
1146 MPI2_SCSI_TASK_MANAGE_REQUEST *req;
1147 unsigned int cm_count = 0;
1148 struct mps_command *cm;
1149 struct mpssas_target *targ;
1151 callout_stop(&tm->cm_callout);
1153 req = (MPI2_SCSI_TASK_MANAGE_REQUEST *)tm->cm_req;
1154 reply = (MPI2_SCSI_TASK_MANAGE_REPLY *)tm->cm_reply;
1158 * Currently there should be no way we can hit this case. It only
1159 * happens when we have a failure to allocate chain frames, and
1160 * task management commands don't have S/G lists.
1162 if ((tm->cm_flags & MPS_CM_FLAGS_ERROR_MASK) != 0) {
1163 mps_printf(sc, "%s: cm_flags = %#x for LUN reset! "
1164 "This should not happen!\n", __func__, tm->cm_flags);
1165 mpssas_free_tm(sc, tm);
1169 if (reply == NULL) {
1170 mpssas_log_command(tm, "NULL reset reply for tm %p\n", tm);
1171 if ((sc->mps_flags & MPS_FLAGS_DIAGRESET) != 0) {
1172 /* this completion was due to a reset, just cleanup */
1173 targ->flags &= ~MPSSAS_TARGET_INRESET;
1175 mpssas_free_tm(sc, tm);
1178 /* we should have gotten a reply. */
1184 mpssas_log_command(tm,
1185 "logical unit reset status 0x%x code 0x%x count %u\n",
1186 le16toh(reply->IOCStatus), le32toh(reply->ResponseCode),
1187 le32toh(reply->TerminationCount));
1189 /* See if there are any outstanding commands for this LUN.
1190 * This could be made more efficient by using a per-LU data
1191 * structure of some sort.
1193 TAILQ_FOREACH(cm, &targ->commands, cm_link) {
1194 if (cm->cm_lun == tm->cm_lun)
1198 if (cm_count == 0) {
1199 mpssas_log_command(tm,
1200 "logical unit %u finished recovery after reset\n",
1203 mpssas_announce_reset(sc, AC_SENT_BDR, tm->cm_targ->tid,
1206 /* we've finished recovery for this logical unit. check and
1207 * see if some other logical unit has a timedout command
1208 * that needs to be processed.
1210 cm = TAILQ_FIRST(&targ->timedout_commands);
1212 mpssas_send_abort(sc, tm, cm);
1216 mpssas_free_tm(sc, tm);
1220 /* if we still have commands for this LUN, the reset
1221 * effectively failed, regardless of the status reported.
1222 * Escalate to a target reset.
1224 mpssas_log_command(tm,
1225 "logical unit reset complete for tm %p, but still have %u command(s)\n",
1227 mpssas_send_reset(sc, tm,
1228 MPI2_SCSITASKMGMT_TASKTYPE_TARGET_RESET);
1233 mpssas_target_reset_complete(struct mps_softc *sc, struct mps_command *tm)
1235 MPI2_SCSI_TASK_MANAGE_REPLY *reply;
1236 MPI2_SCSI_TASK_MANAGE_REQUEST *req;
1237 struct mpssas_target *targ;
1239 callout_stop(&tm->cm_callout);
1241 req = (MPI2_SCSI_TASK_MANAGE_REQUEST *)tm->cm_req;
1242 reply = (MPI2_SCSI_TASK_MANAGE_REPLY *)tm->cm_reply;
1246 * Currently there should be no way we can hit this case. It only
1247 * happens when we have a failure to allocate chain frames, and
1248 * task management commands don't have S/G lists.
1250 if ((tm->cm_flags & MPS_CM_FLAGS_ERROR_MASK) != 0) {
1251 mps_printf(sc, "%s: cm_flags = %#x for target reset! "
1252 "This should not happen!\n", __func__, tm->cm_flags);
1253 mpssas_free_tm(sc, tm);
1257 if (reply == NULL) {
1258 mpssas_log_command(tm, "NULL reset reply for tm %p\n", tm);
1259 if ((sc->mps_flags & MPS_FLAGS_DIAGRESET) != 0) {
1260 /* this completion was due to a reset, just cleanup */
1261 targ->flags &= ~MPSSAS_TARGET_INRESET;
1263 mpssas_free_tm(sc, tm);
1266 /* we should have gotten a reply. */
1272 mpssas_log_command(tm,
1273 "target reset status 0x%x code 0x%x count %u\n",
1274 le16toh(reply->IOCStatus), le32toh(reply->ResponseCode),
1275 le32toh(reply->TerminationCount));
1277 targ->flags &= ~MPSSAS_TARGET_INRESET;
1279 if (targ->outstanding == 0) {
1280 /* we've finished recovery for this target and all
1281 * of its logical units.
1283 mpssas_log_command(tm,
1284 "recovery finished after target reset\n");
1286 mpssas_announce_reset(sc, AC_SENT_BDR, tm->cm_targ->tid,
1290 mpssas_free_tm(sc, tm);
1293 /* after a target reset, if this target still has
1294 * outstanding commands, the reset effectively failed,
1295 * regardless of the status reported. escalate.
1297 mpssas_log_command(tm,
1298 "target reset complete for tm %p, but still have %u command(s)\n",
1299 tm, targ->outstanding);
1304 #define MPS_RESET_TIMEOUT 30
1307 mpssas_send_reset(struct mps_softc *sc, struct mps_command *tm, uint8_t type)
1309 MPI2_SCSI_TASK_MANAGE_REQUEST *req;
1310 struct mpssas_target *target;
1313 target = tm->cm_targ;
1314 if (target->handle == 0) {
1315 mps_printf(sc, "%s null devhandle for target_id %d\n",
1316 __func__, target->tid);
1320 req = (MPI2_SCSI_TASK_MANAGE_REQUEST *)tm->cm_req;
1321 req->DevHandle = htole16(target->handle);
1322 req->Function = MPI2_FUNCTION_SCSI_TASK_MGMT;
1323 req->TaskType = type;
1325 if (type == MPI2_SCSITASKMGMT_TASKTYPE_LOGICAL_UNIT_RESET) {
1326 /* XXX Need to handle invalid LUNs */
1327 MPS_SET_LUN(req->LUN, tm->cm_lun);
1328 tm->cm_targ->logical_unit_resets++;
1329 mpssas_log_command(tm, "sending logical unit reset\n");
1330 tm->cm_complete = mpssas_logical_unit_reset_complete;
1332 else if (type == MPI2_SCSITASKMGMT_TASKTYPE_TARGET_RESET) {
1333 /* Target reset method = SAS Hard Link Reset / SATA Link Reset */
1334 req->MsgFlags = MPI2_SCSITASKMGMT_MSGFLAGS_LINK_RESET;
1335 tm->cm_targ->target_resets++;
1336 tm->cm_targ->flags |= MPSSAS_TARGET_INRESET;
1337 mpssas_log_command(tm, "sending target reset\n");
1338 tm->cm_complete = mpssas_target_reset_complete;
1341 mps_printf(sc, "unexpected reset type 0x%x\n", type);
1346 tm->cm_desc.HighPriority.RequestFlags = MPI2_REQ_DESCRIPT_FLAGS_HIGH_PRIORITY;
1347 tm->cm_complete_data = (void *)tm;
1349 callout_reset(&tm->cm_callout, MPS_RESET_TIMEOUT * hz,
1350 mpssas_tm_timeout, tm);
1352 err = mps_map_command(sc, tm);
1354 mpssas_log_command(tm,
1355 "error %d sending reset type %u\n",
1363 mpssas_abort_complete(struct mps_softc *sc, struct mps_command *tm)
1365 struct mps_command *cm;
1366 MPI2_SCSI_TASK_MANAGE_REPLY *reply;
1367 MPI2_SCSI_TASK_MANAGE_REQUEST *req;
1368 struct mpssas_target *targ;
1370 callout_stop(&tm->cm_callout);
1372 req = (MPI2_SCSI_TASK_MANAGE_REQUEST *)tm->cm_req;
1373 reply = (MPI2_SCSI_TASK_MANAGE_REPLY *)tm->cm_reply;
1377 * Currently there should be no way we can hit this case. It only
1378 * happens when we have a failure to allocate chain frames, and
1379 * task management commands don't have S/G lists.
1381 if ((tm->cm_flags & MPS_CM_FLAGS_ERROR_MASK) != 0) {
1382 mpssas_log_command(tm,
1383 "cm_flags = %#x for abort %p TaskMID %u!\n",
1384 tm->cm_flags, tm, le16toh(req->TaskMID));
1385 mpssas_free_tm(sc, tm);
1389 if (reply == NULL) {
1390 mpssas_log_command(tm,
1391 "NULL abort reply for tm %p TaskMID %u\n",
1392 tm, le16toh(req->TaskMID));
1393 if ((sc->mps_flags & MPS_FLAGS_DIAGRESET) != 0) {
1394 /* this completion was due to a reset, just cleanup */
1396 mpssas_free_tm(sc, tm);
1399 /* we should have gotten a reply. */
1405 mpssas_log_command(tm,
1406 "abort TaskMID %u status 0x%x code 0x%x count %u\n",
1407 le16toh(req->TaskMID),
1408 le16toh(reply->IOCStatus), le32toh(reply->ResponseCode),
1409 le32toh(reply->TerminationCount));
1411 cm = TAILQ_FIRST(&tm->cm_targ->timedout_commands);
1413 /* if there are no more timedout commands, we're done with
1414 * error recovery for this target.
1416 mpssas_log_command(tm,
1417 "finished recovery after aborting TaskMID %u\n",
1418 le16toh(req->TaskMID));
1421 mpssas_free_tm(sc, tm);
1423 else if (le16toh(req->TaskMID) != cm->cm_desc.Default.SMID) {
1424 /* abort success, but we have more timedout commands to abort */
1425 mpssas_log_command(tm,
1426 "continuing recovery after aborting TaskMID %u\n",
1427 le16toh(req->TaskMID));
1429 mpssas_send_abort(sc, tm, cm);
1432 /* we didn't get a command completion, so the abort
1433 * failed as far as we're concerned. escalate.
1435 mpssas_log_command(tm,
1436 "abort failed for TaskMID %u tm %p\n",
1437 le16toh(req->TaskMID), tm);
1439 mpssas_send_reset(sc, tm,
1440 MPI2_SCSITASKMGMT_TASKTYPE_LOGICAL_UNIT_RESET);
1444 #define MPS_ABORT_TIMEOUT 5
1447 mpssas_send_abort(struct mps_softc *sc, struct mps_command *tm, struct mps_command *cm)
1449 MPI2_SCSI_TASK_MANAGE_REQUEST *req;
1450 struct mpssas_target *targ;
1454 if (targ->handle == 0) {
1455 mps_printf(sc, "%s null devhandle for target_id %d\n",
1456 __func__, cm->cm_ccb->ccb_h.target_id);
1460 req = (MPI2_SCSI_TASK_MANAGE_REQUEST *)tm->cm_req;
1461 req->DevHandle = htole16(targ->handle);
1462 req->Function = MPI2_FUNCTION_SCSI_TASK_MGMT;
1463 req->TaskType = MPI2_SCSITASKMGMT_TASKTYPE_ABORT_TASK;
1465 /* XXX Need to handle invalid LUNs */
1466 MPS_SET_LUN(req->LUN, cm->cm_ccb->ccb_h.target_lun);
1468 req->TaskMID = htole16(cm->cm_desc.Default.SMID);
1471 tm->cm_desc.HighPriority.RequestFlags = MPI2_REQ_DESCRIPT_FLAGS_HIGH_PRIORITY;
1472 tm->cm_complete = mpssas_abort_complete;
1473 tm->cm_complete_data = (void *)tm;
1474 tm->cm_targ = cm->cm_targ;
1475 tm->cm_lun = cm->cm_lun;
1477 callout_reset(&tm->cm_callout, MPS_ABORT_TIMEOUT * hz,
1478 mpssas_tm_timeout, tm);
1482 err = mps_map_command(sc, tm);
1484 mpssas_log_command(tm,
1485 "error %d sending abort for cm %p SMID %u\n",
1486 err, cm, req->TaskMID);
1492 mpssas_scsiio_timeout(void *data)
1494 struct mps_softc *sc;
1495 struct mps_command *cm;
1496 struct mpssas_target *targ;
1498 cm = (struct mps_command *)data;
1501 mtx_assert(&sc->mps_mtx, MA_OWNED);
1503 mps_printf(sc, "%s checking sc %p cm %p\n", __func__, sc, cm);
1506 * Run the interrupt handler to make sure it's not pending. This
1507 * isn't perfect because the command could have already completed
1508 * and been re-used, though this is unlikely.
1510 mps_intr_locked(sc);
1511 if (cm->cm_state == MPS_CM_STATE_FREE) {
1512 mps_printf(sc, "SCSI command %p sc %p almost timed out\n", cm, sc);
1516 if (cm->cm_ccb == NULL) {
1517 mps_printf(sc, "command timeout with NULL ccb\n");
1521 mpssas_log_command(cm, "command timeout cm %p ccb %p\n",
1527 /* XXX first, check the firmware state, to see if it's still
1528 * operational. if not, do a diag reset.
1531 cm->cm_ccb->ccb_h.status = CAM_CMD_TIMEOUT;
1532 cm->cm_state = MPS_CM_STATE_TIMEDOUT;
1533 TAILQ_INSERT_TAIL(&targ->timedout_commands, cm, cm_recovery);
1535 if (targ->tm != NULL) {
1536 /* target already in recovery, just queue up another
1537 * timedout command to be processed later.
1539 mps_printf(sc, "queued timedout cm %p for processing by tm %p\n",
1542 else if ((targ->tm = mpssas_alloc_tm(sc)) != NULL) {
1543 mps_printf(sc, "timedout cm %p allocated tm %p\n",
1546 /* start recovery by aborting the first timedout command */
1547 mpssas_send_abort(sc, targ->tm, cm);
1550 /* XXX queue this target up for recovery once a TM becomes
1551 * available. The firmware only has a limited number of
1552 * HighPriority credits for the high priority requests used
1553 * for task management, and we ran out.
1555 * Isilon: don't worry about this for now, since we have
1556 * more credits than disks in an enclosure, and limit
1557 * ourselves to one TM per target for recovery.
1559 mps_printf(sc, "timedout cm %p failed to allocate a tm\n",
1566 mpssas_action_scsiio(struct mpssas_softc *sassc, union ccb *ccb)
1568 MPI2_SCSI_IO_REQUEST *req;
1569 struct ccb_scsiio *csio;
1570 struct mps_softc *sc;
1571 struct mpssas_target *targ;
1572 struct mpssas_lun *lun;
1573 struct mps_command *cm;
1574 uint8_t i, lba_byte, *ref_tag_addr;
1575 uint16_t eedp_flags;
1576 uint32_t mpi_control;
1579 mtx_assert(&sc->mps_mtx, MA_OWNED);
1582 targ = &sassc->targets[csio->ccb_h.target_id];
1583 mps_dprint(sc, MPS_TRACE, "%s ccb %p target flag %x\n", __func__, ccb, targ->flags);
1584 if (targ->handle == 0x0) {
1585 mps_dprint(sc, MPS_TRACE, "%s NULL handle for target %u\n",
1586 __func__, csio->ccb_h.target_id);
1587 csio->ccb_h.status = CAM_SEL_TIMEOUT;
1591 if (targ->flags & MPS_TARGET_FLAGS_RAID_COMPONENT) {
1592 mps_dprint(sc, MPS_TRACE, "%s Raid component no SCSI IO supported %u\n",
1593 __func__, csio->ccb_h.target_id);
1594 csio->ccb_h.status = CAM_TID_INVALID;
1599 * If devinfo is 0 this will be a volume. In that case don't tell CAM
1600 * that the volume has timed out. We want volumes to be enumerated
1601 * until they are deleted/removed, not just failed.
1603 if (targ->flags & MPSSAS_TARGET_INREMOVAL) {
1604 if (targ->devinfo == 0)
1605 csio->ccb_h.status = CAM_REQ_CMP;
1607 csio->ccb_h.status = CAM_SEL_TIMEOUT;
1612 if ((sc->mps_flags & MPS_FLAGS_SHUTDOWN) != 0) {
1613 mps_dprint(sc, MPS_TRACE, "%s shutting down\n", __func__);
1614 csio->ccb_h.status = CAM_TID_INVALID;
1619 cm = mps_alloc_command(sc);
1621 if ((sassc->flags & MPSSAS_QUEUE_FROZEN) == 0) {
1622 xpt_freeze_simq(sassc->sim, 1);
1623 sassc->flags |= MPSSAS_QUEUE_FROZEN;
1625 ccb->ccb_h.status &= ~CAM_SIM_QUEUED;
1626 ccb->ccb_h.status |= CAM_REQUEUE_REQ;
1631 req = (MPI2_SCSI_IO_REQUEST *)cm->cm_req;
1632 bzero(req, sizeof(*req));
1633 req->DevHandle = htole16(targ->handle);
1634 req->Function = MPI2_FUNCTION_SCSI_IO_REQUEST;
1636 req->SenseBufferLowAddress = htole32(cm->cm_sense_busaddr);
1637 req->SenseBufferLength = MPS_SENSE_LEN;
1639 req->ChainOffset = 0;
1640 req->SGLOffset0 = 24; /* 32bit word offset to the SGL */
1645 req->DataLength = htole32(csio->dxfer_len);
1646 req->BidirectionalDataLength = 0;
1647 req->IoFlags = htole16(csio->cdb_len);
1650 /* Note: BiDirectional transfers are not supported */
1651 switch (csio->ccb_h.flags & CAM_DIR_MASK) {
1653 mpi_control = MPI2_SCSIIO_CONTROL_READ;
1654 cm->cm_flags |= MPS_CM_FLAGS_DATAIN;
1657 mpi_control = MPI2_SCSIIO_CONTROL_WRITE;
1658 cm->cm_flags |= MPS_CM_FLAGS_DATAOUT;
1662 mpi_control = MPI2_SCSIIO_CONTROL_NODATATRANSFER;
1666 if (csio->cdb_len == 32)
1667 mpi_control |= 4 << MPI2_SCSIIO_CONTROL_ADDCDBLEN_SHIFT;
1669 * It looks like the hardware doesn't require an explicit tag
1670 * number for each transaction. SAM Task Management not supported
1673 switch (csio->tag_action) {
1674 case MSG_HEAD_OF_Q_TAG:
1675 mpi_control |= MPI2_SCSIIO_CONTROL_HEADOFQ;
1677 case MSG_ORDERED_Q_TAG:
1678 mpi_control |= MPI2_SCSIIO_CONTROL_ORDEREDQ;
1681 mpi_control |= MPI2_SCSIIO_CONTROL_ACAQ;
1683 case CAM_TAG_ACTION_NONE:
1684 case MSG_SIMPLE_Q_TAG:
1686 mpi_control |= MPI2_SCSIIO_CONTROL_SIMPLEQ;
1689 mpi_control |= sc->mapping_table[csio->ccb_h.target_id].TLR_bits;
1690 req->Control = htole32(mpi_control);
1691 if (MPS_SET_LUN(req->LUN, csio->ccb_h.target_lun) != 0) {
1692 mps_free_command(sc, cm);
1693 ccb->ccb_h.status = CAM_LUN_INVALID;
1698 if (csio->ccb_h.flags & CAM_CDB_POINTER)
1699 bcopy(csio->cdb_io.cdb_ptr, &req->CDB.CDB32[0], csio->cdb_len);
1701 bcopy(csio->cdb_io.cdb_bytes, &req->CDB.CDB32[0],csio->cdb_len);
1702 req->IoFlags = htole16(csio->cdb_len);
1705 * Check if EEDP is supported and enabled. If it is then check if the
1706 * SCSI opcode could be using EEDP. If so, make sure the LUN exists and
1707 * is formatted for EEDP support. If all of this is true, set CDB up
1708 * for EEDP transfer.
1710 eedp_flags = op_code_prot[req->CDB.CDB32[0]];
1711 if (sc->eedp_enabled && eedp_flags) {
1712 SLIST_FOREACH(lun, &targ->luns, lun_link) {
1713 if (lun->lun_id == csio->ccb_h.target_lun) {
1718 if ((lun != NULL) && (lun->eedp_formatted)) {
1719 req->EEDPBlockSize = htole16(lun->eedp_block_size);
1720 eedp_flags |= (MPI2_SCSIIO_EEDPFLAGS_INC_PRI_REFTAG |
1721 MPI2_SCSIIO_EEDPFLAGS_CHECK_REFTAG |
1722 MPI2_SCSIIO_EEDPFLAGS_CHECK_GUARD);
1723 req->EEDPFlags = htole16(eedp_flags);
1726 * If CDB less than 32, fill in Primary Ref Tag with
1727 * low 4 bytes of LBA. If CDB is 32, tag stuff is
1728 * already there. Also, set protection bit. FreeBSD
1729 * currently does not support CDBs bigger than 16, but
1730 * the code doesn't hurt, and will be here for the
1733 if (csio->cdb_len != 32) {
1734 lba_byte = (csio->cdb_len == 16) ? 6 : 2;
1735 ref_tag_addr = (uint8_t *)&req->CDB.EEDP32.
1736 PrimaryReferenceTag;
1737 for (i = 0; i < 4; i++) {
1739 req->CDB.CDB32[lba_byte + i];
1742 req->CDB.EEDP32.PrimaryReferenceTag =
1743 htole32(req->CDB.EEDP32.PrimaryReferenceTag);
1744 req->CDB.EEDP32.PrimaryApplicationTagMask =
1746 req->CDB.CDB32[1] = (req->CDB.CDB32[1] & 0x1F) |
1750 MPI2_SCSIIO_EEDPFLAGS_INC_PRI_APPTAG;
1751 req->EEDPFlags = htole16(eedp_flags);
1752 req->CDB.CDB32[10] = (req->CDB.CDB32[10] &
1758 cm->cm_data = csio->data_ptr;
1759 cm->cm_length = csio->dxfer_len;
1760 cm->cm_sge = &req->SGL;
1761 cm->cm_sglsize = (32 - 24) * 4;
1762 cm->cm_desc.SCSIIO.RequestFlags = MPI2_REQ_DESCRIPT_FLAGS_SCSI_IO;
1763 cm->cm_desc.SCSIIO.DevHandle = htole16(targ->handle);
1764 cm->cm_complete = mpssas_scsiio_complete;
1765 cm->cm_complete_data = ccb;
1767 cm->cm_lun = csio->ccb_h.target_lun;
1771 * If HBA is a WD and the command is not for a retry, try to build a
1772 * direct I/O message. If failed, or the command is for a retry, send
1773 * the I/O to the IR volume itself.
1775 if (sc->WD_valid_config) {
1776 if (ccb->ccb_h.status != MPS_WD_RETRY) {
1777 mpssas_direct_drive_io(sassc, cm, ccb);
1779 ccb->ccb_h.status = CAM_REQ_INPROG;
1783 callout_reset(&cm->cm_callout, (ccb->ccb_h.timeout * hz) / 1000,
1784 mpssas_scsiio_timeout, cm);
1787 targ->outstanding++;
1788 TAILQ_INSERT_TAIL(&targ->commands, cm, cm_link);
1790 if ((sc->mps_debug & MPS_TRACE) != 0)
1791 mpssas_log_command(cm, "%s cm %p ccb %p outstanding %u\n",
1792 __func__, cm, ccb, targ->outstanding);
1794 mps_map_command(sc, cm);
1799 mps_response_code(struct mps_softc *sc, u8 response_code)
1803 switch (response_code) {
1804 case MPI2_SCSITASKMGMT_RSP_TM_COMPLETE:
1805 desc = "task management request completed";
1807 case MPI2_SCSITASKMGMT_RSP_INVALID_FRAME:
1808 desc = "invalid frame";
1810 case MPI2_SCSITASKMGMT_RSP_TM_NOT_SUPPORTED:
1811 desc = "task management request not supported";
1813 case MPI2_SCSITASKMGMT_RSP_TM_FAILED:
1814 desc = "task management request failed";
1816 case MPI2_SCSITASKMGMT_RSP_TM_SUCCEEDED:
1817 desc = "task management request succeeded";
1819 case MPI2_SCSITASKMGMT_RSP_TM_INVALID_LUN:
1820 desc = "invalid lun";
1823 desc = "overlapped tag attempted";
1825 case MPI2_SCSITASKMGMT_RSP_IO_QUEUED_ON_IOC:
1826 desc = "task queued, however not sent to target";
1832 mps_dprint(sc, MPS_INFO, "response_code(0x%01x): %s\n",
1833 response_code, desc);
1836 * mps_sc_failed_io_info - translated non-succesfull SCSI_IO request
1839 mps_sc_failed_io_info(struct mps_softc *sc, struct ccb_scsiio *csio,
1840 Mpi2SCSIIOReply_t *mpi_reply)
1844 u16 ioc_status = le16toh(mpi_reply->IOCStatus) &
1845 MPI2_IOCSTATUS_MASK;
1846 u8 scsi_state = mpi_reply->SCSIState;
1847 u8 scsi_status = mpi_reply->SCSIStatus;
1848 char *desc_ioc_state = NULL;
1849 char *desc_scsi_status = NULL;
1850 char *desc_scsi_state = sc->tmp_string;
1851 u32 log_info = le32toh(mpi_reply->IOCLogInfo);
1853 if (log_info == 0x31170000)
1856 switch (ioc_status) {
1857 case MPI2_IOCSTATUS_SUCCESS:
1858 desc_ioc_state = "success";
1860 case MPI2_IOCSTATUS_INVALID_FUNCTION:
1861 desc_ioc_state = "invalid function";
1863 case MPI2_IOCSTATUS_SCSI_RECOVERED_ERROR:
1864 desc_ioc_state = "scsi recovered error";
1866 case MPI2_IOCSTATUS_SCSI_INVALID_DEVHANDLE:
1867 desc_ioc_state = "scsi invalid dev handle";
1869 case MPI2_IOCSTATUS_SCSI_DEVICE_NOT_THERE:
1870 desc_ioc_state = "scsi device not there";
1872 case MPI2_IOCSTATUS_SCSI_DATA_OVERRUN:
1873 desc_ioc_state = "scsi data overrun";
1875 case MPI2_IOCSTATUS_SCSI_DATA_UNDERRUN:
1876 desc_ioc_state = "scsi data underrun";
1878 case MPI2_IOCSTATUS_SCSI_IO_DATA_ERROR:
1879 desc_ioc_state = "scsi io data error";
1881 case MPI2_IOCSTATUS_SCSI_PROTOCOL_ERROR:
1882 desc_ioc_state = "scsi protocol error";
1884 case MPI2_IOCSTATUS_SCSI_TASK_TERMINATED:
1885 desc_ioc_state = "scsi task terminated";
1887 case MPI2_IOCSTATUS_SCSI_RESIDUAL_MISMATCH:
1888 desc_ioc_state = "scsi residual mismatch";
1890 case MPI2_IOCSTATUS_SCSI_TASK_MGMT_FAILED:
1891 desc_ioc_state = "scsi task mgmt failed";
1893 case MPI2_IOCSTATUS_SCSI_IOC_TERMINATED:
1894 desc_ioc_state = "scsi ioc terminated";
1896 case MPI2_IOCSTATUS_SCSI_EXT_TERMINATED:
1897 desc_ioc_state = "scsi ext terminated";
1899 case MPI2_IOCSTATUS_EEDP_GUARD_ERROR:
1900 desc_ioc_state = "eedp guard error";
1902 case MPI2_IOCSTATUS_EEDP_REF_TAG_ERROR:
1903 desc_ioc_state = "eedp ref tag error";
1905 case MPI2_IOCSTATUS_EEDP_APP_TAG_ERROR:
1906 desc_ioc_state = "eedp app tag error";
1909 desc_ioc_state = "unknown";
1913 switch (scsi_status) {
1914 case MPI2_SCSI_STATUS_GOOD:
1915 desc_scsi_status = "good";
1917 case MPI2_SCSI_STATUS_CHECK_CONDITION:
1918 desc_scsi_status = "check condition";
1920 case MPI2_SCSI_STATUS_CONDITION_MET:
1921 desc_scsi_status = "condition met";
1923 case MPI2_SCSI_STATUS_BUSY:
1924 desc_scsi_status = "busy";
1926 case MPI2_SCSI_STATUS_INTERMEDIATE:
1927 desc_scsi_status = "intermediate";
1929 case MPI2_SCSI_STATUS_INTERMEDIATE_CONDMET:
1930 desc_scsi_status = "intermediate condmet";
1932 case MPI2_SCSI_STATUS_RESERVATION_CONFLICT:
1933 desc_scsi_status = "reservation conflict";
1935 case MPI2_SCSI_STATUS_COMMAND_TERMINATED:
1936 desc_scsi_status = "command terminated";
1938 case MPI2_SCSI_STATUS_TASK_SET_FULL:
1939 desc_scsi_status = "task set full";
1941 case MPI2_SCSI_STATUS_ACA_ACTIVE:
1942 desc_scsi_status = "aca active";
1944 case MPI2_SCSI_STATUS_TASK_ABORTED:
1945 desc_scsi_status = "task aborted";
1948 desc_scsi_status = "unknown";
1952 desc_scsi_state[0] = '\0';
1954 desc_scsi_state = " ";
1955 if (scsi_state & MPI2_SCSI_STATE_RESPONSE_INFO_VALID)
1956 strcat(desc_scsi_state, "response info ");
1957 if (scsi_state & MPI2_SCSI_STATE_TERMINATED)
1958 strcat(desc_scsi_state, "state terminated ");
1959 if (scsi_state & MPI2_SCSI_STATE_NO_SCSI_STATUS)
1960 strcat(desc_scsi_state, "no status ");
1961 if (scsi_state & MPI2_SCSI_STATE_AUTOSENSE_FAILED)
1962 strcat(desc_scsi_state, "autosense failed ");
1963 if (scsi_state & MPI2_SCSI_STATE_AUTOSENSE_VALID)
1964 strcat(desc_scsi_state, "autosense valid ");
1966 mps_dprint(sc, MPS_INFO, "\thandle(0x%04x), ioc_status(%s)(0x%04x), \n",
1967 le16toh(mpi_reply->DevHandle),
1968 desc_ioc_state, ioc_status);
1969 /* We can add more detail about underflow data here
1972 mps_dprint(sc, MPS_INFO, "\tscsi_status(%s)(0x%02x), "
1973 "scsi_state(%s)(0x%02x)\n", desc_scsi_status,
1974 scsi_status, desc_scsi_state, scsi_state);
1976 if (sc->mps_debug & MPS_INFO &&
1977 scsi_state & MPI2_SCSI_STATE_AUTOSENSE_VALID) {
1978 mps_dprint(sc, MPS_INFO, "-> Sense Buffer Data : Start :\n");
1979 scsi_sense_print(csio);
1980 mps_dprint(sc, MPS_INFO, "-> Sense Buffer Data : End :\n");
1983 if (scsi_state & MPI2_SCSI_STATE_RESPONSE_INFO_VALID) {
1984 response_info = le32toh(mpi_reply->ResponseInfo);
1985 response_bytes = (u8 *)&response_info;
1986 mps_response_code(sc,response_bytes[0]);
1991 mpssas_scsiio_complete(struct mps_softc *sc, struct mps_command *cm)
1993 MPI2_SCSI_IO_REPLY *rep;
1995 struct ccb_scsiio *csio;
1996 struct mpssas_softc *sassc;
1997 struct scsi_vpd_supported_page_list *vpd_list = NULL;
1998 u8 *TLR_bits, TLR_on;
2002 mps_dprint(sc, MPS_TRACE,
2003 "%s cm %p SMID %u ccb %p reply %p outstanding %u\n",
2004 __func__, cm, cm->cm_desc.Default.SMID, cm->cm_ccb, cm->cm_reply,
2005 cm->cm_targ->outstanding);
2007 callout_stop(&cm->cm_callout);
2008 mtx_assert(&sc->mps_mtx, MA_OWNED);
2011 ccb = cm->cm_complete_data;
2013 rep = (MPI2_SCSI_IO_REPLY *)cm->cm_reply;
2015 * XXX KDM if the chain allocation fails, does it matter if we do
2016 * the sync and unload here? It is simpler to do it in every case,
2017 * assuming it doesn't cause problems.
2019 if (cm->cm_data != NULL) {
2020 if (cm->cm_flags & MPS_CM_FLAGS_DATAIN)
2021 dir = BUS_DMASYNC_POSTREAD;
2022 else if (cm->cm_flags & MPS_CM_FLAGS_DATAOUT)
2023 dir = BUS_DMASYNC_POSTWRITE;;
2024 bus_dmamap_sync(sc->buffer_dmat, cm->cm_dmamap, dir);
2025 bus_dmamap_unload(sc->buffer_dmat, cm->cm_dmamap);
2028 cm->cm_targ->completed++;
2029 cm->cm_targ->outstanding--;
2030 TAILQ_REMOVE(&cm->cm_targ->commands, cm, cm_link);
2032 if (cm->cm_state == MPS_CM_STATE_TIMEDOUT) {
2033 TAILQ_REMOVE(&cm->cm_targ->timedout_commands, cm, cm_recovery);
2034 if (cm->cm_reply != NULL)
2035 mpssas_log_command(cm,
2036 "completed timedout cm %p ccb %p during recovery "
2037 "ioc %x scsi %x state %x xfer %u\n",
2039 le16toh(rep->IOCStatus), rep->SCSIStatus, rep->SCSIState,
2040 le32toh(rep->TransferCount));
2042 mpssas_log_command(cm,
2043 "completed timedout cm %p ccb %p during recovery\n",
2045 } else if (cm->cm_targ->tm != NULL) {
2046 if (cm->cm_reply != NULL)
2047 mpssas_log_command(cm,
2048 "completed cm %p ccb %p during recovery "
2049 "ioc %x scsi %x state %x xfer %u\n",
2051 le16toh(rep->IOCStatus), rep->SCSIStatus, rep->SCSIState,
2052 le32toh(rep->TransferCount));
2054 mpssas_log_command(cm,
2055 "completed cm %p ccb %p during recovery\n",
2057 } else if ((sc->mps_flags & MPS_FLAGS_DIAGRESET) != 0) {
2058 mpssas_log_command(cm,
2059 "reset completed cm %p ccb %p\n",
2063 if ((cm->cm_flags & MPS_CM_FLAGS_ERROR_MASK) != 0) {
2065 * We ran into an error after we tried to map the command,
2066 * so we're getting a callback without queueing the command
2067 * to the hardware. So we set the status here, and it will
2068 * be retained below. We'll go through the "fast path",
2069 * because there can be no reply when we haven't actually
2070 * gone out to the hardware.
2072 ccb->ccb_h.status |= CAM_REQUEUE_REQ;
2075 * Currently the only error included in the mask is
2076 * MPS_CM_FLAGS_CHAIN_FAILED, which means we're out of
2077 * chain frames. We need to freeze the queue until we get
2078 * a command that completed without this error, which will
2079 * hopefully have some chain frames attached that we can
2080 * use. If we wanted to get smarter about it, we would
2081 * only unfreeze the queue in this condition when we're
2082 * sure that we're getting some chain frames back. That's
2083 * probably unnecessary.
2085 if ((sassc->flags & MPSSAS_QUEUE_FROZEN) == 0) {
2086 xpt_freeze_simq(sassc->sim, 1);
2087 sassc->flags |= MPSSAS_QUEUE_FROZEN;
2088 mps_dprint(sc, MPS_INFO, "Error sending command, "
2089 "freezing SIM queue\n");
2093 /* Take the fast path to completion */
2094 if (cm->cm_reply == NULL) {
2095 if ((ccb->ccb_h.status & CAM_STATUS_MASK) == CAM_REQ_INPROG) {
2096 if ((sc->mps_flags & MPS_FLAGS_DIAGRESET) != 0)
2097 ccb->ccb_h.status = CAM_SCSI_BUS_RESET;
2099 ccb->ccb_h.status = CAM_REQ_CMP;
2100 ccb->csio.scsi_status = SCSI_STATUS_OK;
2102 if (sassc->flags & MPSSAS_QUEUE_FROZEN) {
2103 ccb->ccb_h.status |= CAM_RELEASE_SIMQ;
2104 sassc->flags &= ~MPSSAS_QUEUE_FROZEN;
2105 mps_dprint(sc, MPS_INFO,
2106 "Unfreezing SIM queue\n");
2111 * There are two scenarios where the status won't be
2112 * CAM_REQ_CMP. The first is if MPS_CM_FLAGS_ERROR_MASK is
2113 * set, the second is in the MPS_FLAGS_DIAGRESET above.
2115 if ((ccb->ccb_h.status & CAM_STATUS_MASK) != CAM_REQ_CMP) {
2117 * Freeze the dev queue so that commands are
2118 * executed in the correct order with after error
2121 ccb->ccb_h.status |= CAM_DEV_QFRZN;
2122 xpt_freeze_devq(ccb->ccb_h.path, /*count*/ 1);
2124 mps_free_command(sc, cm);
2129 if (sc->mps_debug & MPS_TRACE)
2130 mpssas_log_command(cm,
2131 "ioc %x scsi %x state %x xfer %u\n",
2132 le16toh(rep->IOCStatus), rep->SCSIStatus, rep->SCSIState,
2133 le32toh(rep->TransferCount));
2136 * If this is a Direct Drive I/O, reissue the I/O to the original IR
2137 * Volume if an error occurred (normal I/O retry). Use the original
2138 * CCB, but set a flag that this will be a retry so that it's sent to
2139 * the original volume. Free the command but reuse the CCB.
2141 if (cm->cm_flags & MPS_CM_FLAGS_DD_IO) {
2142 mps_free_command(sc, cm);
2143 ccb->ccb_h.status = MPS_WD_RETRY;
2144 mpssas_action_scsiio(sassc, ccb);
2148 switch (le16toh(rep->IOCStatus) & MPI2_IOCSTATUS_MASK) {
2149 case MPI2_IOCSTATUS_SCSI_DATA_UNDERRUN:
2150 csio->resid = cm->cm_length - le32toh(rep->TransferCount);
2152 case MPI2_IOCSTATUS_SUCCESS:
2153 case MPI2_IOCSTATUS_SCSI_RECOVERED_ERROR:
2155 if ((le16toh(rep->IOCStatus) & MPI2_IOCSTATUS_MASK) ==
2156 MPI2_IOCSTATUS_SCSI_RECOVERED_ERROR)
2157 mpssas_log_command(cm, "recovered error\n");
2159 /* Completion failed at the transport level. */
2160 if (rep->SCSIState & (MPI2_SCSI_STATE_NO_SCSI_STATUS |
2161 MPI2_SCSI_STATE_TERMINATED)) {
2162 ccb->ccb_h.status = CAM_REQ_CMP_ERR;
2166 /* In a modern packetized environment, an autosense failure
2167 * implies that there's not much else that can be done to
2168 * recover the command.
2170 if (rep->SCSIState & MPI2_SCSI_STATE_AUTOSENSE_FAILED) {
2171 ccb->ccb_h.status = CAM_AUTOSENSE_FAIL;
2176 * CAM doesn't care about SAS Response Info data, but if this is
2177 * the state check if TLR should be done. If not, clear the
2178 * TLR_bits for the target.
2180 if ((rep->SCSIState & MPI2_SCSI_STATE_RESPONSE_INFO_VALID) &&
2181 ((le32toh(rep->ResponseInfo) & MPI2_SCSI_RI_MASK_REASONCODE) ==
2182 MPS_SCSI_RI_INVALID_FRAME)) {
2183 sc->mapping_table[csio->ccb_h.target_id].TLR_bits =
2184 (u8)MPI2_SCSIIO_CONTROL_NO_TLR;
2188 * Intentionally override the normal SCSI status reporting
2189 * for these two cases. These are likely to happen in a
2190 * multi-initiator environment, and we want to make sure that
2191 * CAM retries these commands rather than fail them.
2193 if ((rep->SCSIStatus == MPI2_SCSI_STATUS_COMMAND_TERMINATED) ||
2194 (rep->SCSIStatus == MPI2_SCSI_STATUS_TASK_ABORTED)) {
2195 ccb->ccb_h.status = CAM_REQ_ABORTED;
2199 /* Handle normal status and sense */
2200 csio->scsi_status = rep->SCSIStatus;
2201 if (rep->SCSIStatus == MPI2_SCSI_STATUS_GOOD)
2202 ccb->ccb_h.status = CAM_REQ_CMP;
2204 ccb->ccb_h.status = CAM_SCSI_STATUS_ERROR;
2206 if (rep->SCSIState & MPI2_SCSI_STATE_AUTOSENSE_VALID) {
2207 int sense_len, returned_sense_len;
2209 returned_sense_len = min(le32toh(rep->SenseCount),
2210 sizeof(struct scsi_sense_data));
2211 if (returned_sense_len < ccb->csio.sense_len)
2212 ccb->csio.sense_resid = ccb->csio.sense_len -
2215 ccb->csio.sense_resid = 0;
2217 sense_len = min(returned_sense_len,
2218 ccb->csio.sense_len - ccb->csio.sense_resid);
2219 bzero(&ccb->csio.sense_data,
2220 sizeof(ccb->csio.sense_data));
2221 bcopy(cm->cm_sense, &ccb->csio.sense_data, sense_len);
2222 ccb->ccb_h.status |= CAM_AUTOSNS_VALID;
2226 * Check if this is an INQUIRY command. If it's a VPD inquiry,
2227 * and it's page code 0 (Supported Page List), and there is
2228 * inquiry data, and this is for a sequential access device, and
2229 * the device is an SSP target, and TLR is supported by the
2230 * controller, turn the TLR_bits value ON if page 0x90 is
2233 if ((csio->cdb_io.cdb_bytes[0] == INQUIRY) &&
2234 (csio->cdb_io.cdb_bytes[1] & SI_EVPD) &&
2235 (csio->cdb_io.cdb_bytes[2] == SVPD_SUPPORTED_PAGE_LIST) &&
2236 (csio->data_ptr != NULL) && (((uint8_t *)cm->cm_data)[0] ==
2237 T_SEQUENTIAL) && (sc->control_TLR) &&
2238 (sc->mapping_table[csio->ccb_h.target_id].device_info &
2239 MPI2_SAS_DEVICE_INFO_SSP_TARGET)) {
2240 vpd_list = (struct scsi_vpd_supported_page_list *)
2242 TLR_bits = &sc->mapping_table[csio->ccb_h.target_id].
2244 *TLR_bits = (u8)MPI2_SCSIIO_CONTROL_NO_TLR;
2245 TLR_on = (u8)MPI2_SCSIIO_CONTROL_TLR_ON;
2246 alloc_len = ((u16)csio->cdb_io.cdb_bytes[3] << 8) +
2247 csio->cdb_io.cdb_bytes[4];
2248 for (i = 0; i < MIN(vpd_list->length, alloc_len); i++) {
2249 if (vpd_list->list[i] == 0x90) {
2256 case MPI2_IOCSTATUS_SCSI_INVALID_DEVHANDLE:
2257 case MPI2_IOCSTATUS_SCSI_DEVICE_NOT_THERE:
2259 * If devinfo is 0 this will be a volume. In that case don't
2260 * tell CAM that the volume is not there. We want volumes to
2261 * be enumerated until they are deleted/removed, not just
2264 if (cm->cm_targ->devinfo == 0)
2265 ccb->ccb_h.status = CAM_REQ_CMP;
2267 ccb->ccb_h.status = CAM_DEV_NOT_THERE;
2269 case MPI2_IOCSTATUS_INVALID_SGL:
2270 mps_print_scsiio_cmd(sc, cm);
2271 ccb->ccb_h.status = CAM_UNREC_HBA_ERROR;
2273 case MPI2_IOCSTATUS_SCSI_TASK_TERMINATED:
2275 * This is one of the responses that comes back when an I/O
2276 * has been aborted. If it is because of a timeout that we
2277 * initiated, just set the status to CAM_CMD_TIMEOUT.
2278 * Otherwise set it to CAM_REQ_ABORTED. The effect on the
2279 * command is the same (it gets retried, subject to the
2280 * retry counter), the only difference is what gets printed
2283 if (cm->cm_state == MPS_CM_STATE_TIMEDOUT)
2284 ccb->ccb_h.status = CAM_CMD_TIMEOUT;
2286 ccb->ccb_h.status = CAM_REQ_ABORTED;
2288 case MPI2_IOCSTATUS_SCSI_DATA_OVERRUN:
2289 /* resid is ignored for this condition */
2291 ccb->ccb_h.status = CAM_DATA_RUN_ERR;
2293 case MPI2_IOCSTATUS_SCSI_IOC_TERMINATED:
2294 case MPI2_IOCSTATUS_SCSI_EXT_TERMINATED:
2296 * Since these are generally external (i.e. hopefully
2297 * transient transport-related) errors, retry these without
2298 * decrementing the retry count.
2300 ccb->ccb_h.status = CAM_REQUEUE_REQ;
2301 mpssas_log_command(cm,
2302 "terminated ioc %x scsi %x state %x xfer %u\n",
2303 le16toh(rep->IOCStatus), rep->SCSIStatus, rep->SCSIState,
2304 le32toh(rep->TransferCount));
2306 case MPI2_IOCSTATUS_INVALID_FUNCTION:
2307 case MPI2_IOCSTATUS_INTERNAL_ERROR:
2308 case MPI2_IOCSTATUS_INVALID_VPID:
2309 case MPI2_IOCSTATUS_INVALID_FIELD:
2310 case MPI2_IOCSTATUS_INVALID_STATE:
2311 case MPI2_IOCSTATUS_OP_STATE_NOT_SUPPORTED:
2312 case MPI2_IOCSTATUS_SCSI_IO_DATA_ERROR:
2313 case MPI2_IOCSTATUS_SCSI_PROTOCOL_ERROR:
2314 case MPI2_IOCSTATUS_SCSI_RESIDUAL_MISMATCH:
2315 case MPI2_IOCSTATUS_SCSI_TASK_MGMT_FAILED:
2317 mpssas_log_command(cm,
2318 "completed ioc %x scsi %x state %x xfer %u\n",
2319 le16toh(rep->IOCStatus), rep->SCSIStatus, rep->SCSIState,
2320 le32toh(rep->TransferCount));
2321 csio->resid = cm->cm_length;
2322 ccb->ccb_h.status = CAM_REQ_CMP_ERR;
2326 mps_sc_failed_io_info(sc,csio,rep);
2328 if (sassc->flags & MPSSAS_QUEUE_FROZEN) {
2329 ccb->ccb_h.status |= CAM_RELEASE_SIMQ;
2330 sassc->flags &= ~MPSSAS_QUEUE_FROZEN;
2331 mps_dprint(sc, MPS_INFO, "Command completed, "
2332 "unfreezing SIM queue\n");
2335 if ((ccb->ccb_h.status & CAM_STATUS_MASK) != CAM_REQ_CMP) {
2336 ccb->ccb_h.status |= CAM_DEV_QFRZN;
2337 xpt_freeze_devq(ccb->ccb_h.path, /*count*/ 1);
2340 mps_free_command(sc, cm);
2344 /* All Request reached here are Endian safe */
2346 mpssas_direct_drive_io(struct mpssas_softc *sassc, struct mps_command *cm,
2348 pMpi2SCSIIORequest_t pIO_req;
2349 struct mps_softc *sc = sassc->sc;
2351 uint32_t physLBA, stripe_offset, stripe_unit;
2352 uint32_t io_size, column;
2353 uint8_t *ptrLBA, lba_idx, physLBA_byte, *CDB;
2356 * If this is a valid SCSI command (Read6, Read10, Read16, Write6,
2357 * Write10, or Write16), build a direct I/O message. Otherwise, the I/O
2358 * will be sent to the IR volume itself. Since Read6 and Write6 are a
2359 * bit different than the 10/16 CDBs, handle them separately.
2361 pIO_req = (pMpi2SCSIIORequest_t)cm->cm_req;
2362 CDB = pIO_req->CDB.CDB32;
2365 * Handle 6 byte CDBs.
2367 if ((pIO_req->DevHandle == sc->DD_dev_handle) && ((CDB[0] == READ_6) ||
2368 (CDB[0] == WRITE_6))) {
2370 * Get the transfer size in blocks.
2372 io_size = (cm->cm_length >> sc->DD_block_exponent);
2375 * Get virtual LBA given in the CDB.
2377 virtLBA = ((uint64_t)(CDB[1] & 0x1F) << 16) |
2378 ((uint64_t)CDB[2] << 8) | (uint64_t)CDB[3];
2381 * Check that LBA range for I/O does not exceed volume's
2384 if ((virtLBA + (uint64_t)io_size - 1) <=
2387 * Check if the I/O crosses a stripe boundary. If not,
2388 * translate the virtual LBA to a physical LBA and set
2389 * the DevHandle for the PhysDisk to be used. If it
2390 * does cross a boundry, do normal I/O. To get the
2391 * right DevHandle to use, get the map number for the
2392 * column, then use that map number to look up the
2393 * DevHandle of the PhysDisk.
2395 stripe_offset = (uint32_t)virtLBA &
2396 (sc->DD_stripe_size - 1);
2397 if ((stripe_offset + io_size) <= sc->DD_stripe_size) {
2398 physLBA = (uint32_t)virtLBA >>
2399 sc->DD_stripe_exponent;
2400 stripe_unit = physLBA / sc->DD_num_phys_disks;
2401 column = physLBA % sc->DD_num_phys_disks;
2402 pIO_req->DevHandle =
2403 htole16(sc->DD_column_map[column].dev_handle);
2404 /* ???? Is this endian safe*/
2405 cm->cm_desc.SCSIIO.DevHandle =
2408 physLBA = (stripe_unit <<
2409 sc->DD_stripe_exponent) + stripe_offset;
2410 ptrLBA = &pIO_req->CDB.CDB32[1];
2411 physLBA_byte = (uint8_t)(physLBA >> 16);
2412 *ptrLBA = physLBA_byte;
2413 ptrLBA = &pIO_req->CDB.CDB32[2];
2414 physLBA_byte = (uint8_t)(physLBA >> 8);
2415 *ptrLBA = physLBA_byte;
2416 ptrLBA = &pIO_req->CDB.CDB32[3];
2417 physLBA_byte = (uint8_t)physLBA;
2418 *ptrLBA = physLBA_byte;
2421 * Set flag that Direct Drive I/O is
2424 cm->cm_flags |= MPS_CM_FLAGS_DD_IO;
2431 * Handle 10, 12 or 16 byte CDBs.
2433 if ((pIO_req->DevHandle == sc->DD_dev_handle) && ((CDB[0] == READ_10) ||
2434 (CDB[0] == WRITE_10) || (CDB[0] == READ_16) ||
2435 (CDB[0] == WRITE_16) || (CDB[0] == READ_12) ||
2436 (CDB[0] == WRITE_12))) {
2438 * For 16-byte CDB's, verify that the upper 4 bytes of the CDB
2439 * are 0. If not, this is accessing beyond 2TB so handle it in
2440 * the else section. 10-byte and 12-byte CDB's are OK.
2441 * FreeBSD sends very rare 12 byte READ/WRITE, but driver is
2442 * ready to accept 12byte CDB for Direct IOs.
2444 if ((CDB[0] == READ_10 || CDB[0] == WRITE_10) ||
2445 (CDB[0] == READ_12 || CDB[0] == WRITE_12) ||
2446 !(CDB[2] | CDB[3] | CDB[4] | CDB[5])) {
2448 * Get the transfer size in blocks.
2450 io_size = (cm->cm_length >> sc->DD_block_exponent);
2453 * Get virtual LBA. Point to correct lower 4 bytes of
2454 * LBA in the CDB depending on command.
2456 lba_idx = ((CDB[0] == READ_12) ||
2457 (CDB[0] == WRITE_12) ||
2458 (CDB[0] == READ_10) ||
2459 (CDB[0] == WRITE_10))? 2 : 6;
2460 virtLBA = ((uint64_t)CDB[lba_idx] << 24) |
2461 ((uint64_t)CDB[lba_idx + 1] << 16) |
2462 ((uint64_t)CDB[lba_idx + 2] << 8) |
2463 (uint64_t)CDB[lba_idx + 3];
2466 * Check that LBA range for I/O does not exceed volume's
2469 if ((virtLBA + (uint64_t)io_size - 1) <=
2472 * Check if the I/O crosses a stripe boundary.
2473 * If not, translate the virtual LBA to a
2474 * physical LBA and set the DevHandle for the
2475 * PhysDisk to be used. If it does cross a
2476 * boundry, do normal I/O. To get the right
2477 * DevHandle to use, get the map number for the
2478 * column, then use that map number to look up
2479 * the DevHandle of the PhysDisk.
2481 stripe_offset = (uint32_t)virtLBA &
2482 (sc->DD_stripe_size - 1);
2483 if ((stripe_offset + io_size) <=
2484 sc->DD_stripe_size) {
2485 physLBA = (uint32_t)virtLBA >>
2486 sc->DD_stripe_exponent;
2487 stripe_unit = physLBA /
2488 sc->DD_num_phys_disks;
2490 sc->DD_num_phys_disks;
2491 pIO_req->DevHandle =
2492 htole16(sc->DD_column_map[column].
2494 cm->cm_desc.SCSIIO.DevHandle =
2497 physLBA = (stripe_unit <<
2498 sc->DD_stripe_exponent) +
2501 &pIO_req->CDB.CDB32[lba_idx];
2502 physLBA_byte = (uint8_t)(physLBA >> 24);
2503 *ptrLBA = physLBA_byte;
2505 &pIO_req->CDB.CDB32[lba_idx + 1];
2506 physLBA_byte = (uint8_t)(physLBA >> 16);
2507 *ptrLBA = physLBA_byte;
2509 &pIO_req->CDB.CDB32[lba_idx + 2];
2510 physLBA_byte = (uint8_t)(physLBA >> 8);
2511 *ptrLBA = physLBA_byte;
2513 &pIO_req->CDB.CDB32[lba_idx + 3];
2514 physLBA_byte = (uint8_t)physLBA;
2515 *ptrLBA = physLBA_byte;
2518 * Set flag that Direct Drive I/O is
2521 cm->cm_flags |= MPS_CM_FLAGS_DD_IO;
2526 * 16-byte CDB and the upper 4 bytes of the CDB are not
2527 * 0. Get the transfer size in blocks.
2529 io_size = (cm->cm_length >> sc->DD_block_exponent);
2534 virtLBA = ((uint64_t)CDB[2] << 54) |
2535 ((uint64_t)CDB[3] << 48) |
2536 ((uint64_t)CDB[4] << 40) |
2537 ((uint64_t)CDB[5] << 32) |
2538 ((uint64_t)CDB[6] << 24) |
2539 ((uint64_t)CDB[7] << 16) |
2540 ((uint64_t)CDB[8] << 8) |
2544 * Check that LBA range for I/O does not exceed volume's
2547 if ((virtLBA + (uint64_t)io_size - 1) <=
2550 * Check if the I/O crosses a stripe boundary.
2551 * If not, translate the virtual LBA to a
2552 * physical LBA and set the DevHandle for the
2553 * PhysDisk to be used. If it does cross a
2554 * boundry, do normal I/O. To get the right
2555 * DevHandle to use, get the map number for the
2556 * column, then use that map number to look up
2557 * the DevHandle of the PhysDisk.
2559 stripe_offset = (uint32_t)virtLBA &
2560 (sc->DD_stripe_size - 1);
2561 if ((stripe_offset + io_size) <=
2562 sc->DD_stripe_size) {
2563 physLBA = (uint32_t)(virtLBA >>
2564 sc->DD_stripe_exponent);
2565 stripe_unit = physLBA /
2566 sc->DD_num_phys_disks;
2568 sc->DD_num_phys_disks;
2569 pIO_req->DevHandle =
2570 htole16(sc->DD_column_map[column].
2572 cm->cm_desc.SCSIIO.DevHandle =
2575 physLBA = (stripe_unit <<
2576 sc->DD_stripe_exponent) +
2580 * Set upper 4 bytes of LBA to 0. We
2581 * assume that the phys disks are less
2582 * than 2 TB's in size. Then, set the
2585 pIO_req->CDB.CDB32[2] = 0;
2586 pIO_req->CDB.CDB32[3] = 0;
2587 pIO_req->CDB.CDB32[4] = 0;
2588 pIO_req->CDB.CDB32[5] = 0;
2589 ptrLBA = &pIO_req->CDB.CDB32[6];
2590 physLBA_byte = (uint8_t)(physLBA >> 24);
2591 *ptrLBA = physLBA_byte;
2592 ptrLBA = &pIO_req->CDB.CDB32[7];
2593 physLBA_byte = (uint8_t)(physLBA >> 16);
2594 *ptrLBA = physLBA_byte;
2595 ptrLBA = &pIO_req->CDB.CDB32[8];
2596 physLBA_byte = (uint8_t)(physLBA >> 8);
2597 *ptrLBA = physLBA_byte;
2598 ptrLBA = &pIO_req->CDB.CDB32[9];
2599 physLBA_byte = (uint8_t)physLBA;
2600 *ptrLBA = physLBA_byte;
2603 * Set flag that Direct Drive I/O is
2606 cm->cm_flags |= MPS_CM_FLAGS_DD_IO;
2613 #if __FreeBSD_version >= 900026
2615 mpssas_smpio_complete(struct mps_softc *sc, struct mps_command *cm)
2617 MPI2_SMP_PASSTHROUGH_REPLY *rpl;
2618 MPI2_SMP_PASSTHROUGH_REQUEST *req;
2622 ccb = cm->cm_complete_data;
2625 * Currently there should be no way we can hit this case. It only
2626 * happens when we have a failure to allocate chain frames, and SMP
2627 * commands require two S/G elements only. That should be handled
2628 * in the standard request size.
2630 if ((cm->cm_flags & MPS_CM_FLAGS_ERROR_MASK) != 0) {
2631 mps_printf(sc, "%s: cm_flags = %#x on SMP request!\n",
2632 __func__, cm->cm_flags);
2633 ccb->ccb_h.status = CAM_REQ_CMP_ERR;
2637 rpl = (MPI2_SMP_PASSTHROUGH_REPLY *)cm->cm_reply;
2639 mps_dprint(sc, MPS_INFO, "%s: NULL cm_reply!\n", __func__);
2640 ccb->ccb_h.status = CAM_REQ_CMP_ERR;
2644 req = (MPI2_SMP_PASSTHROUGH_REQUEST *)cm->cm_req;
2645 sasaddr = le32toh(req->SASAddress.Low);
2646 sasaddr |= ((uint64_t)(le32toh(req->SASAddress.High))) << 32;
2648 if ((le16toh(rpl->IOCStatus) & MPI2_IOCSTATUS_MASK) != MPI2_IOCSTATUS_SUCCESS ||
2649 rpl->SASStatus != MPI2_SASSTATUS_SUCCESS) {
2650 mps_dprint(sc, MPS_INFO, "%s: IOCStatus %04x SASStatus %02x\n",
2651 __func__, le16toh(rpl->IOCStatus), rpl->SASStatus);
2652 ccb->ccb_h.status = CAM_REQ_CMP_ERR;
2656 mps_dprint(sc, MPS_INFO, "%s: SMP request to SAS address "
2657 "%#jx completed successfully\n", __func__,
2658 (uintmax_t)sasaddr);
2660 if (ccb->smpio.smp_response[2] == SMP_FR_ACCEPTED)
2661 ccb->ccb_h.status = CAM_REQ_CMP;
2663 ccb->ccb_h.status = CAM_SMP_STATUS_ERROR;
2667 * We sync in both directions because we had DMAs in the S/G list
2668 * in both directions.
2670 bus_dmamap_sync(sc->buffer_dmat, cm->cm_dmamap,
2671 BUS_DMASYNC_POSTREAD | BUS_DMASYNC_POSTWRITE);
2672 bus_dmamap_unload(sc->buffer_dmat, cm->cm_dmamap);
2673 mps_free_command(sc, cm);
2678 mpssas_send_smpcmd(struct mpssas_softc *sassc, union ccb *ccb, uint64_t sasaddr)
2680 struct mps_command *cm;
2681 uint8_t *request, *response;
2682 MPI2_SMP_PASSTHROUGH_REQUEST *req;
2683 struct mps_softc *sc;
2692 * XXX We don't yet support physical addresses here.
2694 if (ccb->ccb_h.flags & (CAM_DATA_PHYS|CAM_SG_LIST_PHYS)) {
2695 mps_printf(sc, "%s: physical addresses not supported\n",
2697 ccb->ccb_h.status = CAM_REQ_INVALID;
2703 * If the user wants to send an S/G list, check to make sure they
2704 * have single buffers.
2706 if (ccb->ccb_h.flags & CAM_SCATTER_VALID) {
2708 * The chip does not support more than one buffer for the
2709 * request or response.
2711 if ((ccb->smpio.smp_request_sglist_cnt > 1)
2712 || (ccb->smpio.smp_response_sglist_cnt > 1)) {
2713 mps_printf(sc, "%s: multiple request or response "
2714 "buffer segments not supported for SMP\n",
2716 ccb->ccb_h.status = CAM_REQ_INVALID;
2722 * The CAM_SCATTER_VALID flag was originally implemented
2723 * for the XPT_SCSI_IO CCB, which only has one data pointer.
2724 * We have two. So, just take that flag to mean that we
2725 * might have S/G lists, and look at the S/G segment count
2726 * to figure out whether that is the case for each individual
2729 if (ccb->smpio.smp_request_sglist_cnt != 0) {
2730 bus_dma_segment_t *req_sg;
2732 req_sg = (bus_dma_segment_t *)ccb->smpio.smp_request;
2733 request = (uint8_t *)req_sg[0].ds_addr;
2735 request = ccb->smpio.smp_request;
2737 if (ccb->smpio.smp_response_sglist_cnt != 0) {
2738 bus_dma_segment_t *rsp_sg;
2740 rsp_sg = (bus_dma_segment_t *)ccb->smpio.smp_response;
2741 response = (uint8_t *)rsp_sg[0].ds_addr;
2743 response = ccb->smpio.smp_response;
2745 request = ccb->smpio.smp_request;
2746 response = ccb->smpio.smp_response;
2749 cm = mps_alloc_command(sc);
2751 mps_printf(sc, "%s: cannot allocate command\n", __func__);
2752 ccb->ccb_h.status = CAM_RESRC_UNAVAIL;
2757 req = (MPI2_SMP_PASSTHROUGH_REQUEST *)cm->cm_req;
2758 bzero(req, sizeof(*req));
2759 req->Function = MPI2_FUNCTION_SMP_PASSTHROUGH;
2761 /* Allow the chip to use any route to this SAS address. */
2762 req->PhysicalPort = 0xff;
2764 req->RequestDataLength = htole16(ccb->smpio.smp_request_len);
2766 MPI2_SGLFLAGS_SYSTEM_ADDRESS_SPACE | MPI2_SGLFLAGS_SGL_TYPE_MPI;
2768 mps_dprint(sc, MPS_INFO, "%s: sending SMP request to SAS "
2769 "address %#jx\n", __func__, (uintmax_t)sasaddr);
2771 mpi_init_sge(cm, req, &req->SGL);
2774 * Set up a uio to pass into mps_map_command(). This allows us to
2775 * do one map command, and one busdma call in there.
2777 cm->cm_uio.uio_iov = cm->cm_iovec;
2778 cm->cm_uio.uio_iovcnt = 2;
2779 cm->cm_uio.uio_segflg = UIO_SYSSPACE;
2782 * The read/write flag isn't used by busdma, but set it just in
2783 * case. This isn't exactly accurate, either, since we're going in
2786 cm->cm_uio.uio_rw = UIO_WRITE;
2788 cm->cm_iovec[0].iov_base = request;
2789 cm->cm_iovec[0].iov_len = le16toh(req->RequestDataLength);
2790 cm->cm_iovec[1].iov_base = response;
2791 cm->cm_iovec[1].iov_len = ccb->smpio.smp_response_len;
2793 cm->cm_uio.uio_resid = cm->cm_iovec[0].iov_len +
2794 cm->cm_iovec[1].iov_len;
2797 * Trigger a warning message in mps_data_cb() for the user if we
2798 * wind up exceeding two S/G segments. The chip expects one
2799 * segment for the request and another for the response.
2801 cm->cm_max_segs = 2;
2803 cm->cm_desc.Default.RequestFlags = MPI2_REQ_DESCRIPT_FLAGS_DEFAULT_TYPE;
2804 cm->cm_complete = mpssas_smpio_complete;
2805 cm->cm_complete_data = ccb;
2808 * Tell the mapping code that we're using a uio, and that this is
2809 * an SMP passthrough request. There is a little special-case
2810 * logic there (in mps_data_cb()) to handle the bidirectional
2813 cm->cm_flags |= MPS_CM_FLAGS_USE_UIO | MPS_CM_FLAGS_SMP_PASS |
2814 MPS_CM_FLAGS_DATAIN | MPS_CM_FLAGS_DATAOUT;
2816 /* The chip data format is little endian. */
2817 req->SASAddress.High = htole32(sasaddr >> 32);
2818 req->SASAddress.Low = htole32(sasaddr);
2821 * XXX Note that we don't have a timeout/abort mechanism here.
2822 * From the manual, it looks like task management requests only
2823 * work for SCSI IO and SATA passthrough requests. We may need to
2824 * have a mechanism to retry requests in the event of a chip reset
2825 * at least. Hopefully the chip will insure that any errors short
2826 * of that are relayed back to the driver.
2828 error = mps_map_command(sc, cm);
2829 if ((error != 0) && (error != EINPROGRESS)) {
2830 mps_printf(sc, "%s: error %d returned from mps_map_command()\n",
2838 mps_free_command(sc, cm);
2839 ccb->ccb_h.status = CAM_RESRC_UNAVAIL;
2846 mpssas_action_smpio(struct mpssas_softc *sassc, union ccb *ccb)
2848 struct mps_softc *sc;
2849 struct mpssas_target *targ;
2850 uint64_t sasaddr = 0;
2855 * Make sure the target exists.
2857 targ = &sassc->targets[ccb->ccb_h.target_id];
2858 if (targ->handle == 0x0) {
2859 mps_printf(sc, "%s: target %d does not exist!\n", __func__,
2860 ccb->ccb_h.target_id);
2861 ccb->ccb_h.status = CAM_SEL_TIMEOUT;
2867 * If this device has an embedded SMP target, we'll talk to it
2869 * figure out what the expander's address is.
2871 if ((targ->devinfo & MPI2_SAS_DEVICE_INFO_SMP_TARGET) != 0)
2872 sasaddr = targ->sasaddr;
2875 * If we don't have a SAS address for the expander yet, try
2876 * grabbing it from the page 0x83 information cached in the
2877 * transport layer for this target. LSI expanders report the
2878 * expander SAS address as the port-associated SAS address in
2879 * Inquiry VPD page 0x83. Maxim expanders don't report it in page
2882 * XXX KDM disable this for now, but leave it commented out so that
2883 * it is obvious that this is another possible way to get the SAS
2886 * The parent handle method below is a little more reliable, and
2887 * the other benefit is that it works for devices other than SES
2888 * devices. So you can send a SMP request to a da(4) device and it
2889 * will get routed to the expander that device is attached to.
2890 * (Assuming the da(4) device doesn't contain an SMP target...)
2894 sasaddr = xpt_path_sas_addr(ccb->ccb_h.path);
2898 * If we still don't have a SAS address for the expander, look for
2899 * the parent device of this device, which is probably the expander.
2902 #ifdef OLD_MPS_PROBE
2903 struct mpssas_target *parent_target;
2906 if (targ->parent_handle == 0x0) {
2907 mps_printf(sc, "%s: handle %d does not have a valid "
2908 "parent handle!\n", __func__, targ->handle);
2909 ccb->ccb_h.status = CAM_REQ_INVALID;
2912 #ifdef OLD_MPS_PROBE
2913 parent_target = mpssas_find_target_by_handle(sassc, 0,
2914 targ->parent_handle);
2916 if (parent_target == NULL) {
2917 mps_printf(sc, "%s: handle %d does not have a valid "
2918 "parent target!\n", __func__, targ->handle);
2919 ccb->ccb_h.status = CAM_REQ_INVALID;
2923 if ((parent_target->devinfo &
2924 MPI2_SAS_DEVICE_INFO_SMP_TARGET) == 0) {
2925 mps_printf(sc, "%s: handle %d parent %d does not "
2926 "have an SMP target!\n", __func__,
2927 targ->handle, parent_target->handle);
2928 ccb->ccb_h.status = CAM_REQ_INVALID;
2933 sasaddr = parent_target->sasaddr;
2934 #else /* OLD_MPS_PROBE */
2935 if ((targ->parent_devinfo &
2936 MPI2_SAS_DEVICE_INFO_SMP_TARGET) == 0) {
2937 mps_printf(sc, "%s: handle %d parent %d does not "
2938 "have an SMP target!\n", __func__,
2939 targ->handle, targ->parent_handle);
2940 ccb->ccb_h.status = CAM_REQ_INVALID;
2944 if (targ->parent_sasaddr == 0x0) {
2945 mps_printf(sc, "%s: handle %d parent handle %d does "
2946 "not have a valid SAS address!\n",
2947 __func__, targ->handle, targ->parent_handle);
2948 ccb->ccb_h.status = CAM_REQ_INVALID;
2952 sasaddr = targ->parent_sasaddr;
2953 #endif /* OLD_MPS_PROBE */
2958 mps_printf(sc, "%s: unable to find SAS address for handle %d\n",
2959 __func__, targ->handle);
2960 ccb->ccb_h.status = CAM_REQ_INVALID;
2963 mpssas_send_smpcmd(sassc, ccb, sasaddr);
2971 #endif //__FreeBSD_version >= 900026
2974 mpssas_action_resetdev(struct mpssas_softc *sassc, union ccb *ccb)
2976 MPI2_SCSI_TASK_MANAGE_REQUEST *req;
2977 struct mps_softc *sc;
2978 struct mps_command *tm;
2979 struct mpssas_target *targ;
2981 mps_dprint(sassc->sc, MPS_TRACE, __func__);
2982 mtx_assert(&sassc->sc->mps_mtx, MA_OWNED);
2985 tm = mps_alloc_command(sc);
2987 mps_printf(sc, "comand alloc failure in mpssas_action_resetdev\n");
2988 ccb->ccb_h.status = CAM_RESRC_UNAVAIL;
2993 targ = &sassc->targets[ccb->ccb_h.target_id];
2994 req = (MPI2_SCSI_TASK_MANAGE_REQUEST *)tm->cm_req;
2995 req->DevHandle = htole16(targ->handle);
2996 req->Function = MPI2_FUNCTION_SCSI_TASK_MGMT;
2997 req->TaskType = MPI2_SCSITASKMGMT_TASKTYPE_TARGET_RESET;
2999 /* SAS Hard Link Reset / SATA Link Reset */
3000 req->MsgFlags = MPI2_SCSITASKMGMT_MSGFLAGS_LINK_RESET;
3003 tm->cm_desc.HighPriority.RequestFlags = MPI2_REQ_DESCRIPT_FLAGS_HIGH_PRIORITY;
3004 tm->cm_complete = mpssas_resetdev_complete;
3005 tm->cm_complete_data = ccb;
3007 mps_map_command(sc, tm);
3011 mpssas_resetdev_complete(struct mps_softc *sc, struct mps_command *tm)
3013 MPI2_SCSI_TASK_MANAGE_REPLY *resp;
3016 mps_dprint(sc, MPS_TRACE, __func__);
3017 mtx_assert(&sc->mps_mtx, MA_OWNED);
3019 resp = (MPI2_SCSI_TASK_MANAGE_REPLY *)tm->cm_reply;
3020 ccb = tm->cm_complete_data;
3023 * Currently there should be no way we can hit this case. It only
3024 * happens when we have a failure to allocate chain frames, and
3025 * task management commands don't have S/G lists.
3027 if ((tm->cm_flags & MPS_CM_FLAGS_ERROR_MASK) != 0) {
3028 MPI2_SCSI_TASK_MANAGE_REQUEST *req;
3030 req = (MPI2_SCSI_TASK_MANAGE_REQUEST *)tm->cm_req;
3032 mps_printf(sc, "%s: cm_flags = %#x for reset of handle %#04x! "
3033 "This should not happen!\n", __func__, tm->cm_flags,
3035 ccb->ccb_h.status = CAM_REQ_CMP_ERR;
3039 printf("%s: IOCStatus = 0x%x ResponseCode = 0x%x\n", __func__,
3040 le16toh(resp->IOCStatus), le32toh(resp->ResponseCode));
3042 if (le32toh(resp->ResponseCode) == MPI2_SCSITASKMGMT_RSP_TM_COMPLETE) {
3043 ccb->ccb_h.status = CAM_REQ_CMP;
3044 mpssas_announce_reset(sc, AC_SENT_BDR, tm->cm_targ->tid,
3048 ccb->ccb_h.status = CAM_REQ_CMP_ERR;
3052 mpssas_free_tm(sc, tm);
3057 mpssas_poll(struct cam_sim *sim)
3059 struct mpssas_softc *sassc;
3061 sassc = cam_sim_softc(sim);
3063 if (sassc->sc->mps_debug & MPS_TRACE) {
3064 /* frequent debug messages during a panic just slow
3065 * everything down too much.
3067 mps_printf(sassc->sc, "%s clearing MPS_TRACE\n", __func__);
3068 sassc->sc->mps_debug &= ~MPS_TRACE;
3071 mps_intr_locked(sassc->sc);
3075 mpssas_rescan_done(struct cam_periph *periph, union ccb *done_ccb)
3077 struct mpssas_softc *sassc;
3080 if (done_ccb == NULL)
3083 sassc = (struct mpssas_softc *)done_ccb->ccb_h.ppriv_ptr1;
3085 mtx_assert(&sassc->sc->mps_mtx, MA_OWNED);
3087 xpt_path_string(done_ccb->ccb_h.path, path_str, sizeof(path_str));
3088 mps_dprint(sassc->sc, MPS_INFO, "Completing rescan for %s\n", path_str);
3090 xpt_free_path(done_ccb->ccb_h.path);
3091 xpt_free_ccb(done_ccb);
3093 #if __FreeBSD_version < 1000006
3095 * Before completing scan, get EEDP stuff for all of the existing
3098 mpssas_check_eedp(sassc);
3103 /* thread to handle bus rescans */
3105 mpssas_scanner_thread(void *arg)
3107 struct mpssas_softc *sassc;
3108 struct mps_softc *sc;
3111 sassc = (struct mpssas_softc *)arg;
3114 mps_dprint(sc, MPS_TRACE, "%s\n", __func__);
3118 /* Sleep for 1 second and check the queue status*/
3119 msleep(&sassc->ccb_scanq, &sc->mps_mtx, PRIBIO,
3120 "mps_scanq", 1 * hz);
3121 if (sassc->flags & MPSSAS_SHUTDOWN) {
3122 mps_dprint(sc, MPS_TRACE, "Scanner shutting down\n");
3127 ccb = (union ccb *)TAILQ_FIRST(&sassc->ccb_scanq);
3131 TAILQ_REMOVE(&sassc->ccb_scanq, &ccb->ccb_h, sim_links.tqe);
3133 if (sassc->flags & MPSSAS_SHUTDOWN) {
3134 mps_dprint(sc, MPS_TRACE, "Scanner shutting down\n");
3140 sassc->flags &= ~MPSSAS_SCANTHREAD;
3141 wakeup(&sassc->flags);
3143 mps_dprint(sc, MPS_TRACE, "Scanner exiting\n");
3148 * This function will send READ_CAP_16 to find out EEDP protection mode.
3149 * It will check inquiry data before sending READ_CAP_16.
3150 * Callback for READ_CAP_16 is "mpssas_read_cap_done".
3151 * This is insternal scsi command and we need to take care release of devq, if
3152 * CAM_DEV_QFRZN is set. Driver needs to release devq if it has frozen any.
3153 * xpt_release_devq is called from mpssas_read_cap_done.
3155 * All other commands will be handled by periph layer and there it will
3156 * check for "CAM_DEV_QFRZN" and release of devq will be done.
3159 mpssas_rescan(struct mpssas_softc *sassc, union ccb *ccb)
3163 mps_dprint(sassc->sc, MPS_TRACE, "%s\n", __func__);
3165 mtx_assert(&sassc->sc->mps_mtx, MA_OWNED);
3170 xpt_path_string(ccb->ccb_h.path, path_str, sizeof(path_str));
3171 mps_dprint(sassc->sc, MPS_INFO, "Queueing rescan for %s\n", path_str);
3173 /* Prepare request */
3174 ccb->ccb_h.ppriv_ptr1 = sassc;
3175 ccb->ccb_h.cbfcnp = mpssas_rescan_done;
3176 xpt_setup_ccb(&ccb->ccb_h, ccb->ccb_h.path, MPS_PRIORITY_XPT);
3177 TAILQ_INSERT_TAIL(&sassc->ccb_scanq, &ccb->ccb_h, sim_links.tqe);
3178 wakeup(&sassc->ccb_scanq);
3181 #if __FreeBSD_version >= 1000006
3183 mpssas_async(void *callback_arg, uint32_t code, struct cam_path *path,
3186 struct mps_softc *sc;
3188 sc = (struct mps_softc *)callback_arg;
3191 case AC_ADVINFO_CHANGED: {
3192 struct mpssas_target *target;
3193 struct mpssas_softc *sassc;
3194 struct scsi_read_capacity_data_long rcap_buf;
3195 struct ccb_dev_advinfo cdai;
3196 struct mpssas_lun *lun;
3201 buftype = (uintptr_t)arg;
3207 * We're only interested in read capacity data changes.
3209 if (buftype != CDAI_TYPE_RCAPLONG)
3213 * We're only interested in devices that are attached to
3216 if (xpt_path_path_id(path) != sassc->sim->path_id)
3220 * We should have a handle for this, but check to make sure.
3222 target = &sassc->targets[xpt_path_target_id(path)];
3223 if (target->handle == 0)
3226 lunid = xpt_path_lun_id(path);
3228 SLIST_FOREACH(lun, &target->luns, lun_link) {
3229 if (lun->lun_id == lunid) {
3235 if (found_lun == 0) {
3236 lun = malloc(sizeof(struct mpssas_lun), M_MPT2,
3239 mps_dprint(sc, MPS_FAULT, "Unable to alloc "
3240 "LUN for EEDP support.\n");
3243 lun->lun_id = lunid;
3244 SLIST_INSERT_HEAD(&target->luns, lun, lun_link);
3247 bzero(&rcap_buf, sizeof(rcap_buf));
3248 xpt_setup_ccb(&cdai.ccb_h, path, CAM_PRIORITY_NORMAL);
3249 cdai.ccb_h.func_code = XPT_DEV_ADVINFO;
3250 cdai.ccb_h.flags = CAM_DIR_IN;
3251 cdai.buftype = CDAI_TYPE_RCAPLONG;
3253 cdai.bufsiz = sizeof(rcap_buf);
3254 cdai.buf = (uint8_t *)&rcap_buf;
3255 xpt_action((union ccb *)&cdai);
3256 if ((cdai.ccb_h.status & CAM_DEV_QFRZN) != 0)
3257 cam_release_devq(cdai.ccb_h.path,
3260 if (((cdai.ccb_h.status & CAM_STATUS_MASK) == CAM_REQ_CMP)
3261 && (rcap_buf.prot & SRC16_PROT_EN)) {
3262 lun->eedp_formatted = TRUE;
3263 lun->eedp_block_size = scsi_4btoul(rcap_buf.length);
3265 lun->eedp_formatted = FALSE;
3266 lun->eedp_block_size = 0;
3274 #else /* __FreeBSD_version >= 1000006 */
3277 mpssas_check_eedp(struct mpssas_softc *sassc)
3279 struct mps_softc *sc = sassc->sc;
3280 struct ccb_scsiio *csio;
3281 struct scsi_read_capacity_16 *scsi_cmd;
3282 struct scsi_read_capacity_eedp *rcap_buf;
3284 path_id_t pathid = cam_sim_path(sassc->sim);
3285 target_id_t targetid;
3287 struct cam_periph *found_periph;
3288 struct mpssas_target *target;
3289 struct mpssas_lun *lun;
3291 struct ccb_getdev cgd;
3295 * Issue a READ CAPACITY 16 command to each LUN of each target. This
3296 * info is used to determine if the LUN is formatted for EEDP support.
3298 for (targetid = 0; targetid < sc->facts->MaxTargets; targetid++) {
3299 target = &sassc->targets[targetid];
3300 if (target->handle == 0x0) {
3306 ccb = xpt_alloc_ccb_nowait();
3308 mps_dprint(sc, MPS_FAULT, "Unable to alloc CCB "
3309 "for EEDP support.\n");
3313 if (xpt_create_path(&ccb->ccb_h.path, xpt_periph,
3314 pathid, targetid, lunid) != CAM_REQ_CMP) {
3315 mps_dprint(sc, MPS_FAULT, "Unable to create "
3316 "path for EEDP support\n");
3322 * If a periph is returned, the LUN exists. Create an
3323 * entry in the target's LUN list.
3325 if ((found_periph = cam_periph_find(ccb->ccb_h.path,
3328 * If LUN is already in list, don't create a new
3332 SLIST_FOREACH(lun, &target->luns, lun_link) {
3333 if (lun->lun_id == lunid) {
3339 lun = malloc(sizeof(struct mpssas_lun),
3340 M_MPT2, M_NOWAIT | M_ZERO);
3342 mps_dprint(sc, MPS_FAULT,
3343 "Unable to alloc LUN for "
3345 xpt_free_path(ccb->ccb_h.path);
3349 lun->lun_id = lunid;
3350 SLIST_INSERT_HEAD(&target->luns, lun,
3354 /* Before Issuing READ CAPACITY 16,
3355 * check Device type.
3357 xpt_setup_ccb(&cgd.ccb_h, ccb->ccb_h.path,
3358 CAM_PRIORITY_NORMAL);
3359 cgd.ccb_h.func_code = XPT_GDEV_TYPE;
3360 xpt_action((union ccb *)&cgd);
3363 * If this flag is set in the inquiry data,
3364 * the device supports protection information,
3365 * and must support the 16 byte read
3366 * capacity command, otherwise continue without
3367 * sending read cap 16
3370 xpt_path_string(ccb->ccb_h.path, path_str,
3373 if ((cgd.inq_data.spc3_flags &
3374 SPC3_SID_PROTECT) == 0) {
3375 xpt_free_path(ccb->ccb_h.path);
3380 mps_dprint(sc, MPS_INFO,
3381 "Sending read cap: path %s"
3382 " handle %d\n", path_str, target->handle );
3385 * Issue a READ CAPACITY 16 command for the LUN.
3386 * The mpssas_read_cap_done function will load
3387 * the read cap info into the LUN struct.
3390 malloc(sizeof(struct scsi_read_capacity_eedp),
3391 M_MPT2, M_NOWAIT| M_ZERO);
3392 if (rcap_buf == NULL) {
3393 mps_dprint(sc, MPS_FAULT, "Unable to alloc read "
3394 "capacity buffer for EEDP support.\n");
3395 xpt_free_path(ccb->ccb_h.path);
3400 csio->ccb_h.func_code = XPT_SCSI_IO;
3401 csio->ccb_h.flags = CAM_DIR_IN;
3402 csio->ccb_h.retry_count = 4;
3403 csio->ccb_h.cbfcnp = mpssas_read_cap_done;
3404 csio->ccb_h.timeout = 60000;
3405 csio->data_ptr = (uint8_t *)rcap_buf;
3406 csio->dxfer_len = sizeof(struct
3407 scsi_read_capacity_eedp);
3408 csio->sense_len = MPS_SENSE_LEN;
3409 csio->cdb_len = sizeof(*scsi_cmd);
3410 csio->tag_action = MSG_SIMPLE_Q_TAG;
3412 scsi_cmd = (struct scsi_read_capacity_16 *)
3413 &csio->cdb_io.cdb_bytes;
3414 bzero(scsi_cmd, sizeof(*scsi_cmd));
3415 scsi_cmd->opcode = 0x9E;
3416 scsi_cmd->service_action = SRC16_SERVICE_ACTION;
3417 ((uint8_t *)scsi_cmd)[13] = sizeof(struct
3418 scsi_read_capacity_eedp);
3421 * Set the path, target and lun IDs for the READ
3424 ccb->ccb_h.path_id =
3425 xpt_path_path_id(ccb->ccb_h.path);
3426 ccb->ccb_h.target_id =
3427 xpt_path_target_id(ccb->ccb_h.path);
3428 ccb->ccb_h.target_lun =
3429 xpt_path_lun_id(ccb->ccb_h.path);
3431 ccb->ccb_h.ppriv_ptr1 = sassc;
3434 xpt_free_path(ccb->ccb_h.path);
3437 } while (found_periph);
3443 mpssas_read_cap_done(struct cam_periph *periph, union ccb *done_ccb)
3445 struct mpssas_softc *sassc;
3446 struct mpssas_target *target;
3447 struct mpssas_lun *lun;
3448 struct scsi_read_capacity_eedp *rcap_buf;
3450 if (done_ccb == NULL)
3453 /* Driver need to release devq, it Scsi command is
3454 * generated by driver internally.
3455 * Currently there is a single place where driver
3456 * calls scsi command internally. In future if driver
3457 * calls more scsi command internally, it needs to release
3458 * devq internally, since those command will not go back to
3461 if ((done_ccb->ccb_h.status & CAM_DEV_QFRZN) ) {
3462 done_ccb->ccb_h.status &= ~CAM_DEV_QFRZN;
3463 xpt_release_devq(done_ccb->ccb_h.path,
3464 /*count*/ 1, /*run_queue*/TRUE);
3467 rcap_buf = (struct scsi_read_capacity_eedp *)done_ccb->csio.data_ptr;
3470 * Get the LUN ID for the path and look it up in the LUN list for the
3473 sassc = (struct mpssas_softc *)done_ccb->ccb_h.ppriv_ptr1;
3474 target = &sassc->targets[done_ccb->ccb_h.target_id];
3475 SLIST_FOREACH(lun, &target->luns, lun_link) {
3476 if (lun->lun_id != done_ccb->ccb_h.target_lun)
3480 * Got the LUN in the target's LUN list. Fill it in
3481 * with EEDP info. If the READ CAP 16 command had some
3482 * SCSI error (common if command is not supported), mark
3483 * the lun as not supporting EEDP and set the block size
3486 if (((done_ccb->ccb_h.status & CAM_STATUS_MASK) != CAM_REQ_CMP)
3487 || (done_ccb->csio.scsi_status != SCSI_STATUS_OK)) {
3488 lun->eedp_formatted = FALSE;
3489 lun->eedp_block_size = 0;
3493 if (rcap_buf->protect & 0x01) {
3494 lun->eedp_formatted = TRUE;
3495 lun->eedp_block_size = scsi_4btoul(rcap_buf->length);
3500 // Finished with this CCB and path.
3501 free(rcap_buf, M_MPT2);
3502 xpt_free_path(done_ccb->ccb_h.path);
3503 xpt_free_ccb(done_ccb);
3505 #endif /* __FreeBSD_version >= 1000006 */
3508 mpssas_startup(struct mps_softc *sc)
3510 struct mpssas_softc *sassc;
3513 * Send the port enable message and set the wait_for_port_enable flag.
3514 * This flag helps to keep the simq frozen until all discovery events
3518 mpssas_startup_increment(sassc);
3519 sc->wait_for_port_enable = 1;
3520 mpssas_send_portenable(sc);
3525 mpssas_send_portenable(struct mps_softc *sc)
3527 MPI2_PORT_ENABLE_REQUEST *request;
3528 struct mps_command *cm;
3530 mps_dprint(sc, MPS_TRACE, "%s\n", __func__);
3532 if ((cm = mps_alloc_command(sc)) == NULL)
3534 request = (MPI2_PORT_ENABLE_REQUEST *)cm->cm_req;
3535 request->Function = MPI2_FUNCTION_PORT_ENABLE;
3536 request->MsgFlags = 0;
3538 cm->cm_desc.Default.RequestFlags = MPI2_REQ_DESCRIPT_FLAGS_DEFAULT_TYPE;
3539 cm->cm_complete = mpssas_portenable_complete;
3543 mps_map_command(sc, cm);
3544 mps_dprint(sc, MPS_TRACE,
3545 "mps_send_portenable finished cm %p req %p complete %p\n",
3546 cm, cm->cm_req, cm->cm_complete);
3551 mpssas_portenable_complete(struct mps_softc *sc, struct mps_command *cm)
3553 MPI2_PORT_ENABLE_REPLY *reply;
3554 struct mpssas_softc *sassc;
3556 mps_dprint(sc, MPS_TRACE, "%s\n", __func__);
3560 * Currently there should be no way we can hit this case. It only
3561 * happens when we have a failure to allocate chain frames, and
3562 * port enable commands don't have S/G lists.
3564 if ((cm->cm_flags & MPS_CM_FLAGS_ERROR_MASK) != 0) {
3565 mps_printf(sc, "%s: cm_flags = %#x for port enable! "
3566 "This should not happen!\n", __func__, cm->cm_flags);
3569 reply = (MPI2_PORT_ENABLE_REPLY *)cm->cm_reply;
3571 mps_dprint(sc, MPS_FAULT, "Portenable NULL reply\n");
3572 else if (le16toh(reply->IOCStatus & MPI2_IOCSTATUS_MASK) !=
3573 MPI2_IOCSTATUS_SUCCESS)
3574 mps_dprint(sc, MPS_FAULT, "Portenable failed\n");
3576 mps_free_command(sc, cm);
3577 if (sc->mps_ich.ich_arg != NULL) {
3578 mps_dprint(sc, MPS_INFO, "disestablish config intrhook\n");
3579 config_intrhook_disestablish(&sc->mps_ich);
3580 sc->mps_ich.ich_arg = NULL;
3584 * Get WarpDrive info after discovery is complete but before the scan
3585 * starts. At this point, all devices are ready to be exposed to the
3586 * OS. If devices should be hidden instead, take them out of the
3587 * 'targets' array before the scan. The devinfo for a disk will have
3588 * some info and a volume's will be 0. Use that to remove disks.
3590 mps_wd_config_pages(sc);
3593 * Done waiting for port enable to complete. Decrement the refcount.
3594 * If refcount is 0, discovery is complete and a rescan of the bus can
3595 * take place. Since the simq was explicitly frozen before port
3596 * enable, it must be explicitly released here to keep the
3597 * freeze/release count in sync.
3599 sc->wait_for_port_enable = 0;
3600 sc->port_enable_complete = 1;
3601 wakeup(&sc->port_enable_complete);
3602 mpssas_startup_decrement(sassc);
3603 xpt_release_simq(sassc->sim, 1);