2 * Copyright (c) 2018 Microsemi Corporation.
5 * Redistribution and use in source and binary forms, with or without
6 * modification, are permitted provided that the following conditions
8 * 1. Redistributions of source code must retain the above copyright
9 * notice, this list of conditions and the following disclaimer.
10 * 2. Redistributions in binary form must reproduce the above copyright
11 * notice, this list of conditions and the following disclaimer in the
12 * documentation and/or other materials provided with the distribution.
14 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
15 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
16 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
17 * ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
18 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
19 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
20 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
21 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
22 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
23 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
29 * CAM interface for smartpqi driver
32 #include "smartpqi_includes.h"
35 * Set cam sim properties of the smartpqi adapter.
37 static void update_sim_properties(struct cam_sim *sim, struct ccb_pathinq *cpi)
40 pqisrc_softstate_t *softs = (struct pqisrc_softstate *)
45 cpi->hba_inquiry = PI_SDTR_ABLE|PI_TAG_ABLE|PI_WIDE_16;
47 cpi->hba_misc = PIM_NOBUSRESET | PIM_UNMAPPED;
49 cpi->max_lun = PQI_MAX_MULTILUN;
50 cpi->max_target = 1088;
51 cpi->maxio = (softs->pqi_cap.max_sg_elem - 1) * PAGE_SIZE;
52 cpi->initiator_id = 255;
53 strncpy(cpi->sim_vid, "FreeBSD", SIM_IDLEN);
54 strncpy(cpi->hba_vid, "Microsemi", HBA_IDLEN);
55 strncpy(cpi->dev_name, cam_sim_name(sim), DEV_IDLEN);
56 cpi->unit_number = cam_sim_unit(sim);
57 cpi->bus_id = cam_sim_bus(sim);
58 cpi->base_transfer_speed = 1200000; /* Base bus speed in KB/sec */
59 cpi->protocol = PROTO_SCSI;
60 cpi->protocol_version = SCSI_REV_SPC4;
61 cpi->transport = XPORT_SPI;
62 cpi->transport_version = 2;
63 cpi->ccb_h.status = CAM_REQ_CMP;
69 * Get transport settings of the smartpqi adapter
71 static void get_transport_settings(struct pqisrc_softstate *softs,
72 struct ccb_trans_settings *cts)
74 struct ccb_trans_settings_scsi *scsi = &cts->proto_specific.scsi;
75 struct ccb_trans_settings_sas *sas = &cts->xport_specific.sas;
76 struct ccb_trans_settings_spi *spi = &cts->xport_specific.spi;
80 cts->protocol = PROTO_SCSI;
81 cts->protocol_version = SCSI_REV_SPC4;
82 cts->transport = XPORT_SPI;
83 cts->transport_version = 2;
84 spi->valid = CTS_SPI_VALID_DISC;
85 spi->flags = CTS_SPI_FLAGS_DISC_ENB;
86 scsi->valid = CTS_SCSI_VALID_TQ;
87 scsi->flags = CTS_SCSI_FLAGS_TAG_ENB;
88 sas->valid = CTS_SAS_VALID_SPEED;
89 cts->ccb_h.status = CAM_REQ_CMP;
95 * Add the target to CAM layer and rescan, when a new device is found
97 void os_add_device(pqisrc_softstate_t *softs, pqi_scsi_dev_t *device) {
102 if(softs->os_specific.sim_registered) {
103 if ((ccb = xpt_alloc_ccb_nowait()) == NULL) {
104 DBG_ERR("rescan failed (can't allocate CCB)\n");
108 if (xpt_create_path(&ccb->ccb_h.path, NULL,
109 cam_sim_path(softs->os_specific.sim),
110 device->target, device->lun) != CAM_REQ_CMP) {
111 DBG_ERR("rescan failed (can't create path)\n");
122 * Remove the device from CAM layer when deleted or hot removed
124 void os_remove_device(pqisrc_softstate_t *softs,
125 pqi_scsi_dev_t *device) {
126 struct cam_path *tmppath;
130 if(softs->os_specific.sim_registered) {
131 if (xpt_create_path(&tmppath, NULL,
132 cam_sim_path(softs->os_specific.sim),
133 device->target, device->lun) != CAM_REQ_CMP) {
134 DBG_ERR("unable to create path for async event");
137 xpt_async(AC_LOST_DEVICE, tmppath, NULL);
138 xpt_free_path(tmppath);
139 pqisrc_free_device(softs, device);
147 * Function to release the frozen simq
149 static void pqi_release_camq( rcb_t *rcb )
151 pqisrc_softstate_t *softs;
152 struct ccb_scsiio *csio;
154 csio = (struct ccb_scsiio *)&rcb->cm_ccb->csio;
159 if (softs->os_specific.pqi_flags & PQI_FLAG_BUSY) {
160 softs->os_specific.pqi_flags &= ~PQI_FLAG_BUSY;
161 if (csio->ccb_h.status & CAM_RELEASE_SIMQ)
162 xpt_release_simq(xpt_path_sim(csio->ccb_h.path), 0);
164 csio->ccb_h.status |= CAM_RELEASE_SIMQ;
171 * Function to dma-unmap the completed request
173 static void pqi_unmap_request(void *arg)
175 pqisrc_softstate_t *softs;
178 DBG_IO("IN rcb = %p\n", arg);
183 if (!(rcb->cm_flags & PQI_CMD_MAPPED))
186 if (rcb->bcount != 0 ) {
187 if (rcb->data_dir == SOP_DATA_DIR_FROM_DEVICE)
188 bus_dmamap_sync(softs->os_specific.pqi_buffer_dmat,
190 BUS_DMASYNC_POSTREAD);
191 if (rcb->data_dir == SOP_DATA_DIR_TO_DEVICE)
192 bus_dmamap_sync(softs->os_specific.pqi_buffer_dmat,
194 BUS_DMASYNC_POSTWRITE);
195 bus_dmamap_unload(softs->os_specific.pqi_buffer_dmat,
198 rcb->cm_flags &= ~PQI_CMD_MAPPED;
200 if(rcb->sgt && rcb->nseg)
201 os_mem_free(rcb->softs, (void*)rcb->sgt,
202 rcb->nseg*sizeof(sgt_t));
204 pqisrc_put_tag(&softs->taglist, rcb->tag);
210 * Construct meaningful LD name for volume here.
213 smartpqi_fix_ld_inquiry(pqisrc_softstate_t *softs, struct ccb_scsiio *csio)
215 struct scsi_inquiry_data *inq = NULL;
217 pqi_scsi_dev_t *device = NULL;
221 cdb = (csio->ccb_h.flags & CAM_CDB_POINTER) ?
222 (uint8_t *)csio->cdb_io.cdb_ptr : csio->cdb_io.cdb_bytes;
223 if(cdb[0] == INQUIRY &&
224 (cdb[1] & SI_EVPD) == 0 &&
225 (csio->ccb_h.flags & CAM_DIR_MASK) == CAM_DIR_IN &&
226 csio->dxfer_len >= SHORT_INQUIRY_LENGTH) {
228 inq = (struct scsi_inquiry_data *)csio->data_ptr;
230 device = softs->device_list[csio->ccb_h.target_id][csio->ccb_h.target_lun];
232 /* Let the disks be probed and dealt with via CAM. Only for LD
233 let it fall through and inquiry be tweaked */
234 if( !device || !pqisrc_is_logical_device(device) ||
235 (device->devtype != DISK_DEVICE) ||
236 pqisrc_is_external_raid_device(device)) {
240 strncpy(inq->vendor, "MSCC",
242 strncpy(inq->product,
243 pqisrc_raidlevel_to_string(device->raid_level),
245 strncpy(inq->revision, device->volume_offline?"OFF":"OK",
253 * Handle completion of a command - pass results back through the CCB
256 os_io_response_success(rcb_t *rcb)
258 struct ccb_scsiio *csio;
260 DBG_IO("IN rcb = %p\n", rcb);
263 panic("rcb is null");
265 csio = (struct ccb_scsiio *)&rcb->cm_ccb->csio;
268 panic("csio is null");
270 rcb->status = REQUEST_SUCCESS;
271 csio->ccb_h.status = CAM_REQ_CMP;
273 smartpqi_fix_ld_inquiry(rcb->softs, csio);
274 pqi_release_camq(rcb);
275 pqi_unmap_request(rcb);
276 xpt_done((union ccb *)csio);
282 * Error response handling for raid IO
284 void os_raid_response_error(rcb_t *rcb, raid_path_error_info_elem_t *err_info)
286 struct ccb_scsiio *csio;
287 pqisrc_softstate_t *softs;
291 csio = (struct ccb_scsiio *)&rcb->cm_ccb->csio;
294 panic("csio is null");
298 ASSERT(err_info != NULL);
299 csio->scsi_status = err_info->status;
300 csio->ccb_h.status = CAM_REQ_CMP_ERR;
302 if (csio->ccb_h.func_code == XPT_SCSI_IO) {
304 * Handle specific SCSI status values.
306 switch(csio->scsi_status) {
307 case PQI_RAID_STATUS_QUEUE_FULL:
308 csio->ccb_h.status = CAM_REQ_CMP;
309 DBG_ERR("Queue Full error");
311 /* check condition, sense data included */
312 case PQI_RAID_STATUS_CHECK_CONDITION:
314 uint16_t sense_data_len =
315 LE_16(err_info->sense_data_len);
316 uint8_t *sense_data = NULL;
318 sense_data = err_info->data;
319 memset(&csio->sense_data, 0, csio->sense_len);
320 sense_data_len = (sense_data_len >
325 memcpy(&csio->sense_data, sense_data,
327 if (csio->sense_len > sense_data_len)
328 csio->sense_resid = csio->sense_len
331 csio->sense_resid = 0;
332 csio->ccb_h.status = CAM_SCSI_STATUS_ERROR
339 case PQI_RAID_DATA_IN_OUT_UNDERFLOW:
342 resid = rcb->bcount-err_info->data_out_transferred;
344 csio->ccb_h.status = CAM_REQ_CMP;
348 csio->ccb_h.status = CAM_REQ_CMP;
353 if (softs->os_specific.pqi_flags & PQI_FLAG_BUSY) {
354 softs->os_specific.pqi_flags &= ~PQI_FLAG_BUSY;
355 if (csio->ccb_h.status & CAM_RELEASE_SIMQ)
356 xpt_release_simq(xpt_path_sim(csio->ccb_h.path), 0);
358 csio->ccb_h.status |= CAM_RELEASE_SIMQ;
361 pqi_unmap_request(rcb);
362 xpt_done((union ccb *)csio);
369 * Error response handling for aio.
371 void os_aio_response_error(rcb_t *rcb, aio_path_error_info_elem_t *err_info)
373 struct ccb_scsiio *csio;
374 pqisrc_softstate_t *softs;
379 panic("rcb is null");
381 rcb->status = REQUEST_SUCCESS;
382 csio = (struct ccb_scsiio *)&rcb->cm_ccb->csio;
384 panic("csio is null");
388 switch (err_info->service_resp) {
389 case PQI_AIO_SERV_RESPONSE_COMPLETE:
390 csio->ccb_h.status = err_info->status;
392 case PQI_AIO_SERV_RESPONSE_FAILURE:
393 switch(err_info->status) {
394 case PQI_AIO_STATUS_IO_ABORTED:
395 csio->ccb_h.status = CAM_REQ_ABORTED;
396 DBG_WARN_BTL(rcb->dvp, "IO aborted\n");
398 case PQI_AIO_STATUS_UNDERRUN:
399 csio->ccb_h.status = CAM_REQ_CMP;
401 LE_32(err_info->resd_count);
403 case PQI_AIO_STATUS_OVERRUN:
404 csio->ccb_h.status = CAM_REQ_CMP;
406 case PQI_AIO_STATUS_AIO_PATH_DISABLED:
407 DBG_WARN_BTL(rcb->dvp,"AIO Path Disabled\n");
408 rcb->dvp->offload_enabled = false;
409 csio->ccb_h.status |= CAM_REQUEUE_REQ;
411 case PQI_AIO_STATUS_IO_ERROR:
412 case PQI_AIO_STATUS_IO_NO_DEVICE:
413 case PQI_AIO_STATUS_INVALID_DEVICE:
415 DBG_WARN_BTL(rcb->dvp,"IO Error/Invalid/No device\n");
416 csio->ccb_h.status |=
417 CAM_SCSI_STATUS_ERROR;
421 case PQI_AIO_SERV_RESPONSE_TMF_COMPLETE:
422 case PQI_AIO_SERV_RESPONSE_TMF_SUCCEEDED:
423 csio->ccb_h.status = CAM_REQ_CMP;
425 case PQI_AIO_SERV_RESPONSE_TMF_REJECTED:
426 case PQI_AIO_SERV_RESPONSE_TMF_INCORRECT_LUN:
427 DBG_WARN_BTL(rcb->dvp,"TMF rejected/Incorrect Lun\n");
428 csio->ccb_h.status |= CAM_SCSI_STATUS_ERROR;
431 DBG_WARN_BTL(rcb->dvp,"Scsi Status Error\n");
432 csio->ccb_h.status |= CAM_SCSI_STATUS_ERROR;
435 if(err_info->data_pres == DATA_PRESENT_SENSE_DATA ) {
436 csio->scsi_status = PQI_AIO_STATUS_CHECK_CONDITION;
437 uint8_t *sense_data = NULL;
438 unsigned sense_data_len = LE_16(err_info->data_len);
440 sense_data = err_info->data;
441 DBG_ERR_BTL(rcb->dvp, "SCSI_STATUS_CHECK_COND sense size %u\n",
443 memset(&csio->sense_data, 0, csio->sense_len);
445 memcpy(&csio->sense_data, sense_data, ((sense_data_len >
446 csio->sense_len) ? csio->sense_len : sense_data_len));
447 if (csio->sense_len > sense_data_len)
448 csio->sense_resid = csio->sense_len - sense_data_len;
450 csio->sense_resid = 0;
451 csio->ccb_h.status = CAM_SCSI_STATUS_ERROR | CAM_AUTOSNS_VALID;
454 smartpqi_fix_ld_inquiry(softs, csio);
455 pqi_release_camq(rcb);
456 pqi_unmap_request(rcb);
457 xpt_done((union ccb *)csio);
462 pqi_freeze_ccb(union ccb *ccb)
464 if ((ccb->ccb_h.status & CAM_DEV_QFRZN) == 0) {
465 ccb->ccb_h.status |= CAM_DEV_QFRZN;
466 xpt_freeze_devq(ccb->ccb_h.path, 1);
471 * Command-mapping helper function - populate this command's s/g table.
474 pqi_request_map_helper(void *arg, bus_dma_segment_t *segs, int nseg, int error)
476 pqisrc_softstate_t *softs;
482 if( error || nseg > softs->pqi_cap.max_sg_elem )
484 rcb->cm_ccb->ccb_h.status = CAM_RESRC_UNAVAIL;
485 pqi_freeze_ccb(rcb->cm_ccb);
486 DBG_ERR_BTL(rcb->dvp, "map failed err = %d or nseg(%d) > sgelem(%d)\n",
487 error, nseg, softs->pqi_cap.max_sg_elem);
488 pqi_unmap_request(rcb);
489 xpt_done((union ccb *)rcb->cm_ccb);
493 rcb->sgt = os_mem_alloc(softs, nseg * sizeof(rcb_t));
494 if (rcb->sgt == NULL) {
495 rcb->cm_ccb->ccb_h.status = CAM_RESRC_UNAVAIL;
496 pqi_freeze_ccb(rcb->cm_ccb);
497 DBG_ERR_BTL(rcb->dvp, "os_mem_alloc() failed; nseg = %d\n", nseg);
498 pqi_unmap_request(rcb);
499 xpt_done((union ccb *)rcb->cm_ccb);
504 for (int i = 0; i < nseg; i++) {
505 rcb->sgt[i].addr = segs[i].ds_addr;
506 rcb->sgt[i].len = segs[i].ds_len;
507 rcb->sgt[i].flags = 0;
510 if (rcb->data_dir == SOP_DATA_DIR_FROM_DEVICE)
511 bus_dmamap_sync(softs->os_specific.pqi_buffer_dmat,
512 rcb->cm_datamap, BUS_DMASYNC_PREREAD);
513 if (rcb->data_dir == SOP_DATA_DIR_TO_DEVICE)
514 bus_dmamap_sync(softs->os_specific.pqi_buffer_dmat,
515 rcb->cm_datamap, BUS_DMASYNC_PREWRITE);
517 /* Call IO functions depending on pd or ld */
518 rcb->status = REQUEST_PENDING;
520 error = pqisrc_build_send_io(softs, rcb);
523 rcb->req_pending = false;
524 rcb->cm_ccb->ccb_h.status = CAM_RESRC_UNAVAIL;
525 pqi_freeze_ccb(rcb->cm_ccb);
526 DBG_ERR_BTL(rcb->dvp, "Build IO failed, error = %d\n", error);
527 pqi_unmap_request(rcb);
528 xpt_done((union ccb *)rcb->cm_ccb);
534 * Function to dma-map the request buffer
536 static int pqi_map_request( rcb_t *rcb )
538 pqisrc_softstate_t *softs = rcb->softs;
539 int error = PQI_STATUS_SUCCESS;
540 union ccb *ccb = rcb->cm_ccb;
544 /* check that mapping is necessary */
545 if (rcb->cm_flags & PQI_CMD_MAPPED)
547 rcb->cm_flags |= PQI_CMD_MAPPED;
550 error = bus_dmamap_load_ccb(softs->os_specific.pqi_buffer_dmat,
551 rcb->cm_datamap, ccb, pqi_request_map_helper, rcb, 0);
553 DBG_ERR_BTL(rcb->dvp, "bus_dmamap_load_ccb failed = %d count = %d\n",
559 * Set up the command to go to the controller. If there are no
560 * data buffers associated with the command then it can bypass
563 /* Call IO functions depending on pd or ld */
564 rcb->status = REQUEST_PENDING;
566 error = pqisrc_build_send_io(softs, rcb);
570 DBG_FUNC("OUT error = %d\n", error);
576 * Function to clear the request control block
578 void os_reset_rcb( rcb_t *rcb )
580 rcb->error_info = NULL;
583 rcb->tag = INVALID_ELEM;
593 rcb->encrypt_enable = false;
594 rcb->ioaccel_handle = 0;
596 rcb->req_pending = false;
600 * Callback function for the lun rescan
602 static void smartpqi_lunrescan_cb(struct cam_periph *periph, union ccb *ccb)
604 xpt_free_path(ccb->ccb_h.path);
610 * Function to rescan the lun
612 static void smartpqi_lun_rescan(struct pqisrc_softstate *softs, int target,
615 union ccb *ccb = NULL;
616 cam_status status = 0;
617 struct cam_path *path = NULL;
621 ccb = xpt_alloc_ccb_nowait();
622 status = xpt_create_path(&path, NULL,
623 cam_sim_path(softs->os_specific.sim), target, lun);
624 if (status != CAM_REQ_CMP) {
625 DBG_ERR("xpt_create_path status(%d) != CAM_REQ_CMP \n",
631 bzero(ccb, sizeof(union ccb));
632 xpt_setup_ccb(&ccb->ccb_h, path, 5);
633 ccb->ccb_h.func_code = XPT_SCAN_LUN;
634 ccb->ccb_h.cbfcnp = smartpqi_lunrescan_cb;
635 ccb->crcn.flags = CAM_FLAG_NONE;
643 * Function to rescan the lun under each target
645 void smartpqi_target_rescan(struct pqisrc_softstate *softs)
647 int target = 0, lun = 0;
651 for(target = 0; target < PQI_MAX_DEVICES; target++){
652 for(lun = 0; lun < PQI_MAX_MULTILUN; lun++){
653 if(softs->device_list[target][lun]){
654 smartpqi_lun_rescan(softs, target, lun);
663 * Set the mode of tagged command queueing for the current task.
665 uint8_t os_get_task_attr(rcb_t *rcb)
667 union ccb *ccb = rcb->cm_ccb;
668 uint8_t tag_action = SOP_TASK_ATTRIBUTE_SIMPLE;
670 switch(ccb->csio.tag_action) {
671 case MSG_HEAD_OF_Q_TAG:
672 tag_action = SOP_TASK_ATTRIBUTE_HEAD_OF_QUEUE;
674 case MSG_ORDERED_Q_TAG:
675 tag_action = SOP_TASK_ATTRIBUTE_ORDERED;
677 case MSG_SIMPLE_Q_TAG:
679 tag_action = SOP_TASK_ATTRIBUTE_SIMPLE;
686 * Complete all outstanding commands
688 void os_complete_outstanding_cmds_nodevice(pqisrc_softstate_t *softs)
694 for (tag = 1; tag < softs->max_outstanding_io; tag++) {
695 rcb_t *prcb = &softs->rcb[tag];
696 if(prcb->req_pending && prcb->cm_ccb ) {
697 prcb->req_pending = false;
698 prcb->cm_ccb->ccb_h.status = CAM_REQ_ABORTED | CAM_REQ_CMP;
699 xpt_done((union ccb *)prcb->cm_ccb);
708 * IO handling functionality entry point
710 static int pqisrc_io_start(struct cam_sim *sim, union ccb *ccb)
713 uint32_t tag, no_transfer = 0;
714 pqisrc_softstate_t *softs = (struct pqisrc_softstate *)
716 int32_t error = PQI_STATUS_FAILURE;
721 if( softs->device_list[ccb->ccb_h.target_id][ccb->ccb_h.target_lun] == NULL ) {
722 ccb->ccb_h.status = CAM_DEV_NOT_THERE;
723 DBG_INFO("Device = %d not there\n", ccb->ccb_h.target_id);
724 return PQI_STATUS_FAILURE;
727 dvp = softs->device_list[ccb->ccb_h.target_id][ccb->ccb_h.target_lun];
728 /* Check controller state */
729 if (IN_PQI_RESET(softs)) {
730 ccb->ccb_h.status = CAM_SCSI_BUS_RESET
731 | CAM_BUSY | CAM_REQ_INPROG;
732 DBG_WARN("Device = %d BUSY/IN_RESET\n", ccb->ccb_h.target_id);
735 /* Check device state */
736 if (pqisrc_ctrl_offline(softs) || DEV_GONE(dvp)) {
737 ccb->ccb_h.status = CAM_DEV_NOT_THERE | CAM_REQ_CMP;
738 DBG_WARN("Device = %d GONE/OFFLINE\n", ccb->ccb_h.target_id);
741 /* Check device reset */
742 if (dvp->reset_in_progress) {
743 ccb->ccb_h.status = CAM_SCSI_BUSY | CAM_REQ_INPROG | CAM_BUSY;
744 DBG_WARN("Device %d reset returned busy\n", ccb->ccb_h.target_id);
748 if (dvp->expose_device == false) {
749 ccb->ccb_h.status = CAM_DEV_NOT_THERE;
750 DBG_INFO("Device = %d not exposed\n", ccb->ccb_h.target_id);
754 tag = pqisrc_get_tag(&softs->taglist);
755 if( tag == INVALID_ELEM ) {
756 DBG_ERR("Get Tag failed\n");
757 xpt_freeze_simq(softs->os_specific.sim, 1);
758 softs->os_specific.pqi_flags |= PQI_FLAG_BUSY;
759 ccb->ccb_h.status |= (CAM_REQUEUE_REQ | CAM_RELEASE_SIMQ);
760 return PQI_STATUS_FAILURE;
763 DBG_IO("tag = %d &softs->taglist : %p\n", tag, &softs->taglist);
765 rcb = &softs->rcb[tag];
769 rcb->cmdlen = ccb->csio.cdb_len;
770 ccb->ccb_h.sim_priv.entries[0].ptr = rcb;
772 switch (ccb->ccb_h.flags & CAM_DIR_MASK) {
774 rcb->data_dir = SOP_DATA_DIR_FROM_DEVICE;
777 rcb->data_dir = SOP_DATA_DIR_TO_DEVICE;
783 DBG_ERR("Unknown Dir\n");
787 rcb->dvp = softs->device_list[ccb->ccb_h.target_id][ccb->ccb_h.target_lun];
790 rcb->cm_data = (void *)ccb->csio.data_ptr;
791 rcb->bcount = ccb->csio.dxfer_len;
797 * Submit the request to the adapter.
799 * Note that this may fail if we're unable to map the request (and
800 * if we ever learn a transport layer other than simple, may fail
801 * if the adapter rejects the command).
803 if ((error = pqi_map_request(rcb)) != 0) {
804 rcb->req_pending = false;
805 xpt_freeze_simq(softs->os_specific.sim, 1);
806 ccb->ccb_h.status |= CAM_RELEASE_SIMQ;
807 if (error == EINPROGRESS) {
808 DBG_WARN("In Progress on %d\n", ccb->ccb_h.target_id);
811 ccb->ccb_h.status |= CAM_REQUEUE_REQ;
812 DBG_WARN("Requeue req error = %d target = %d\n", error,
813 ccb->ccb_h.target_id);
814 pqi_unmap_request(rcb);
818 DBG_FUNC("OUT error = %d\n", error);
823 * Abort a task, task management functionality
826 pqisrc_scsi_abort_task(pqisrc_softstate_t *softs, union ccb *ccb)
828 rcb_t *rcb = ccb->ccb_h.sim_priv.entries[0].ptr;
829 uint32_t abort_tag = rcb->tag;
831 int rval = PQI_STATUS_SUCCESS;
836 qid = (uint16_t)rcb->resp_qid;
838 tag = pqisrc_get_tag(&softs->taglist);
839 rcb = &softs->rcb[tag];
843 rval = pqisrc_send_tmf(softs, rcb->dvp, rcb, abort_tag,
844 SOP_TASK_MANAGEMENT_FUNCTION_ABORT_TASK);
846 if (PQI_STATUS_SUCCESS == rval) {
848 if (REQUEST_SUCCESS == rval) {
849 ccb->ccb_h.status = CAM_REQ_ABORTED;
852 pqisrc_put_tag(&softs->taglist, abort_tag);
853 pqisrc_put_tag(&softs->taglist,rcb->tag);
855 DBG_FUNC("OUT rval = %d\n", rval);
861 * Abort a taskset, task management functionality
864 pqisrc_scsi_abort_task_set(pqisrc_softstate_t *softs, union ccb *ccb)
868 int rval = PQI_STATUS_SUCCESS;
872 tag = pqisrc_get_tag(&softs->taglist);
873 rcb = &softs->rcb[tag];
876 rval = pqisrc_send_tmf(softs, rcb->dvp, rcb, 0,
877 SOP_TASK_MANAGEMENT_FUNCTION_ABORT_TASK_SET);
879 if (rval == PQI_STATUS_SUCCESS) {
883 pqisrc_put_tag(&softs->taglist,rcb->tag);
885 DBG_FUNC("OUT rval = %d\n", rval);
891 * Target reset task management functionality
894 pqisrc_target_reset( pqisrc_softstate_t *softs, union ccb *ccb)
896 pqi_scsi_dev_t *devp = softs->device_list[ccb->ccb_h.target_id][ccb->ccb_h.target_lun];
899 int rval = PQI_STATUS_SUCCESS;
904 DBG_ERR("bad target t%d\n", ccb->ccb_h.target_id);
908 tag = pqisrc_get_tag(&softs->taglist);
909 rcb = &softs->rcb[tag];
912 devp->reset_in_progress = true;
913 rval = pqisrc_send_tmf(softs, devp, rcb, 0,
914 SOP_TASK_MANAGEMENT_LUN_RESET);
915 if (PQI_STATUS_SUCCESS == rval) {
918 devp->reset_in_progress = false;
919 pqisrc_put_tag(&softs->taglist,rcb->tag);
921 DBG_FUNC("OUT rval = %d\n", rval);
923 return ((rval == REQUEST_SUCCESS) ?
924 PQI_STATUS_SUCCESS : PQI_STATUS_FAILURE);
928 * cam entry point of the smartpqi module.
930 static void smartpqi_cam_action(struct cam_sim *sim, union ccb *ccb)
932 struct pqisrc_softstate *softs = cam_sim_softc(sim);
933 struct ccb_hdr *ccb_h = &ccb->ccb_h;
937 switch (ccb_h->func_code) {
940 if(!pqisrc_io_start(sim, ccb)) {
945 case XPT_CALC_GEOMETRY:
947 struct ccb_calc_geometry *ccg;
949 if (ccg->block_size == 0) {
950 ccb->ccb_h.status &= ~CAM_SIM_QUEUED;
951 ccb->ccb_h.status = CAM_REQ_INVALID;
954 cam_calc_geometry(ccg, /* extended */ 1);
955 ccb->ccb_h.status = CAM_REQ_CMP;
960 update_sim_properties(sim, &ccb->cpi);
961 ccb->ccb_h.status = CAM_REQ_CMP;
964 case XPT_GET_TRAN_SETTINGS:
965 get_transport_settings(softs, &ccb->cts);
966 ccb->ccb_h.status = CAM_REQ_CMP;
969 if(pqisrc_scsi_abort_task(softs, ccb)) {
970 ccb->ccb_h.status = CAM_REQ_CMP_ERR;
972 DBG_ERR("Abort task failed on %d\n",
973 ccb->ccb_h.target_id);
978 if (pqisrc_scsi_abort_task_set(softs, ccb)) {
979 ccb->ccb_h.status = CAM_REQ_CMP_ERR;
980 DBG_ERR("Abort task set failed on %d\n",
981 ccb->ccb_h.target_id);
987 if(pqisrc_target_reset(softs, ccb)) {
988 ccb->ccb_h.status = CAM_REQ_CMP_ERR;
989 DBG_ERR("Target reset failed on %d\n",
990 ccb->ccb_h.target_id);
994 ccb->ccb_h.status = CAM_REQ_CMP;
998 ccb->ccb_h.status = CAM_REQ_CMP;
1000 case XPT_SET_TRAN_SETTINGS:
1001 ccb->ccb_h.status = CAM_FUNC_NOTAVAIL;
1004 DBG_WARN("UNSUPPORTED FUNC CODE\n");
1005 ccb->ccb_h.status = CAM_FUNC_NOTAVAIL;
1014 * Function to poll the response, when interrupts are unavailable
1015 * This also serves supporting crash dump.
1017 static void smartpqi_poll(struct cam_sim *sim)
1019 struct pqisrc_softstate *softs = cam_sim_softc(sim);
1022 for (i = 1; i < softs->intr_count; i++ )
1023 pqisrc_process_response_queue(softs, i);
1027 * Function to adjust the queue depth of a device
1029 void smartpqi_adjust_queue_depth(struct cam_path *path, uint32_t queue_depth)
1031 struct ccb_relsim crs;
1035 xpt_setup_ccb(&crs.ccb_h, path, 5);
1036 crs.ccb_h.func_code = XPT_REL_SIMQ;
1037 crs.ccb_h.flags = CAM_DEV_QFREEZE;
1038 crs.release_flags = RELSIM_ADJUST_OPENINGS;
1039 crs.openings = queue_depth;
1040 xpt_action((union ccb *)&crs);
1041 if(crs.ccb_h.status != CAM_REQ_CMP) {
1042 printf("XPT_REL_SIMQ failed stat=%d\n", crs.ccb_h.status);
1049 * Function to register async callback for setting queue depth
1052 smartpqi_async(void *callback_arg, u_int32_t code,
1053 struct cam_path *path, void *arg)
1055 struct pqisrc_softstate *softs;
1056 softs = (struct pqisrc_softstate*)callback_arg;
1061 case AC_FOUND_DEVICE:
1063 struct ccb_getdev *cgd;
1064 cgd = (struct ccb_getdev *)arg;
1068 uint32_t t_id = cgd->ccb_h.target_id;
1070 if (t_id <= (PQI_CTLR_INDEX - 1)) {
1071 if (softs != NULL) {
1072 pqi_scsi_dev_t *dvp = softs->device_list[t_id][cgd->ccb_h.target_lun];
1073 smartpqi_adjust_queue_depth(path,
1087 * Function to register sim with CAM layer for smartpqi driver
1089 int register_sim(struct pqisrc_softstate *softs, int card_index)
1092 int max_transactions;
1093 union ccb *ccb = NULL;
1094 cam_status status = 0;
1095 struct ccb_setasync csa;
1096 struct cam_sim *sim;
1100 max_transactions = softs->max_io_for_scsi_ml;
1101 softs->os_specific.devq = cam_simq_alloc(max_transactions);
1102 if (softs->os_specific.devq == NULL) {
1103 DBG_ERR("cam_simq_alloc failed txns = %d\n",
1105 return PQI_STATUS_FAILURE;
1108 sim = cam_sim_alloc(smartpqi_cam_action, \
1109 smartpqi_poll, "smartpqi", softs, \
1110 card_index, &softs->os_specific.cam_lock, \
1111 1, max_transactions, softs->os_specific.devq);
1113 DBG_ERR("cam_sim_alloc failed txns = %d\n",
1115 cam_simq_free(softs->os_specific.devq);
1116 return PQI_STATUS_FAILURE;
1119 softs->os_specific.sim = sim;
1120 mtx_lock(&softs->os_specific.cam_lock);
1121 status = xpt_bus_register(sim, softs->os_specific.pqi_dev, 0);
1122 if (status != CAM_SUCCESS) {
1123 DBG_ERR("xpt_bus_register failed status=%d\n", status);
1124 cam_sim_free(softs->os_specific.sim, FALSE);
1125 cam_simq_free(softs->os_specific.devq);
1126 mtx_unlock(&softs->os_specific.cam_lock);
1127 return PQI_STATUS_FAILURE;
1130 softs->os_specific.sim_registered = TRUE;
1131 ccb = xpt_alloc_ccb_nowait();
1133 DBG_ERR("xpt_create_path failed\n");
1134 return PQI_STATUS_FAILURE;
1137 if (xpt_create_path(&ccb->ccb_h.path, NULL,
1138 cam_sim_path(softs->os_specific.sim),
1139 CAM_TARGET_WILDCARD,
1140 CAM_LUN_WILDCARD) != CAM_REQ_CMP) {
1141 DBG_ERR("xpt_create_path failed\n");
1143 xpt_bus_deregister(cam_sim_path(softs->os_specific.sim));
1144 cam_sim_free(softs->os_specific.sim, TRUE);
1145 mtx_unlock(&softs->os_specific.cam_lock);
1146 return PQI_STATUS_FAILURE;
1149 * Callback to set the queue depth per target which is
1150 * derived from the FW.
1152 softs->os_specific.path = ccb->ccb_h.path;
1153 xpt_setup_ccb(&csa.ccb_h, softs->os_specific.path, 5);
1154 csa.ccb_h.func_code = XPT_SASYNC_CB;
1155 csa.event_enable = AC_FOUND_DEVICE;
1156 csa.callback = smartpqi_async;
1157 csa.callback_arg = softs;
1158 xpt_action((union ccb *)&csa);
1159 if (csa.ccb_h.status != CAM_REQ_CMP) {
1160 DBG_ERR("Unable to register smartpqi_aysnc handler: %d!\n",
1164 mtx_unlock(&softs->os_specific.cam_lock);
1170 * Function to deregister smartpqi sim from cam layer
1172 void deregister_sim(struct pqisrc_softstate *softs)
1174 struct ccb_setasync csa;
1178 if (softs->os_specific.mtx_init) {
1179 mtx_lock(&softs->os_specific.cam_lock);
1183 xpt_setup_ccb(&csa.ccb_h, softs->os_specific.path, 5);
1184 csa.ccb_h.func_code = XPT_SASYNC_CB;
1185 csa.event_enable = 0;
1186 csa.callback = smartpqi_async;
1187 csa.callback_arg = softs;
1188 xpt_action((union ccb *)&csa);
1189 xpt_free_path(softs->os_specific.path);
1191 xpt_release_simq(softs->os_specific.sim, 0);
1193 xpt_bus_deregister(cam_sim_path(softs->os_specific.sim));
1194 softs->os_specific.sim_registered = FALSE;
1196 if (softs->os_specific.sim) {
1197 cam_sim_free(softs->os_specific.sim, FALSE);
1198 softs->os_specific.sim = NULL;
1200 if (softs->os_specific.mtx_init) {
1201 mtx_unlock(&softs->os_specific.cam_lock);
1203 if (softs->os_specific.devq != NULL) {
1204 cam_simq_free(softs->os_specific.devq);
1206 if (softs->os_specific.mtx_init) {
1207 mtx_destroy(&softs->os_specific.cam_lock);
1208 softs->os_specific.mtx_init = FALSE;
1211 mtx_destroy(&softs->os_specific.map_lock);