2 * Copyright (c) 2010, LSI Corp.
4 * Author : Manjunath Ranganathaiah
5 * Support: freebsdraid@lsi.com
7 * Redistribution and use in source and binary forms, with or without
8 * modification, are permitted provided that the following conditions
11 * 1. Redistributions of source code must retain the above copyright
12 * notice, this list of conditions and the following disclaimer.
13 * 2. Redistributions in binary form must reproduce the above copyright
14 * notice, this list of conditions and the following disclaimer in
15 * the documentation and/or other materials provided with the
17 * 3. Neither the name of the <ORGANIZATION> nor the names of its
18 * contributors may be used to endorse or promote products derived
19 * from this software without specific prior written permission.
21 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
22 * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
23 * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS
24 * FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE
25 * COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT,
26 * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING,
27 * BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES;
28 * LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER
29 * CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
30 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN
31 * ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
32 * POSSIBILITY OF SUCH DAMAGE.
35 #include <sys/cdefs.h>
36 __FBSDID("$FreeBSD$");
38 #include <dev/tws/tws.h>
39 #include <dev/tws/tws_services.h>
40 #include <dev/tws/tws_hdm.h>
43 #include <cam/cam_ccb.h>
45 MALLOC_DEFINE(M_TWS, "twsbuf", "buffers used by tws driver");
46 int tws_queue_depth = TWS_MAX_REQS;
47 int tws_enable_msi = 0;
48 int tws_enable_msix = 0;
53 extern int tws_cam_attach(struct tws_softc *sc);
54 extern void tws_cam_detach(struct tws_softc *sc);
55 extern int tws_init_ctlr(struct tws_softc *sc);
56 extern boolean tws_ctlr_ready(struct tws_softc *sc);
57 extern void tws_turn_off_interrupts(struct tws_softc *sc);
58 extern void tws_q_insert_tail(struct tws_softc *sc, struct tws_request *req,
60 extern struct tws_request *tws_q_remove_request(struct tws_softc *sc,
61 struct tws_request *req, u_int8_t q_type );
62 extern struct tws_request *tws_q_remove_head(struct tws_softc *sc,
64 extern boolean tws_get_response(struct tws_softc *sc, u_int16_t *req_id);
65 extern boolean tws_ctlr_reset(struct tws_softc *sc);
66 extern void tws_intr(void *arg);
67 extern int tws_use_32bit_sgls;
70 struct tws_request *tws_get_request(struct tws_softc *sc, u_int16_t type);
71 int tws_init_connect(struct tws_softc *sc, u_int16_t mc);
72 void tws_send_event(struct tws_softc *sc, u_int8_t event);
73 uint8_t tws_get_state(struct tws_softc *sc);
74 void tws_release_request(struct tws_request *req);
78 /* Function prototypes */
79 static d_open_t tws_open;
80 static d_close_t tws_close;
81 static d_read_t tws_read;
82 static d_write_t tws_write;
83 extern d_ioctl_t tws_ioctl;
85 static int tws_init(struct tws_softc *sc);
86 static void tws_dmamap_cmds_load_cbfn(void *arg, bus_dma_segment_t *segs,
89 static int tws_init_reqs(struct tws_softc *sc, u_int32_t dma_mem_size);
90 static int tws_init_aen_q(struct tws_softc *sc);
91 static int tws_init_trace_q(struct tws_softc *sc);
92 static int tws_setup_irq(struct tws_softc *sc);
93 int tws_setup_intr(struct tws_softc *sc, int irqs);
94 int tws_teardown_intr(struct tws_softc *sc);
97 /* Character device entry points */
99 static struct cdevsw tws_cdevsw = {
100 .d_version = D_VERSION,
102 .d_close = tws_close,
104 .d_write = tws_write,
105 .d_ioctl = tws_ioctl,
110 * In the cdevsw routines, we find our softc by using the si_drv1 member
111 * of struct cdev. We set this variable to point to our softc in our
112 * attach routine when we create the /dev entry.
116 tws_open(struct cdev *dev, int oflags, int devtype, struct thread *td)
118 struct tws_softc *sc = dev->si_drv1;
121 TWS_TRACE_DEBUG(sc, "entry", dev, oflags);
126 tws_close(struct cdev *dev, int fflag, int devtype, struct thread *td)
128 struct tws_softc *sc = dev->si_drv1;
131 TWS_TRACE_DEBUG(sc, "entry", dev, fflag);
136 tws_read(struct cdev *dev, struct uio *uio, int ioflag)
138 struct tws_softc *sc = dev->si_drv1;
141 TWS_TRACE_DEBUG(sc, "entry", dev, ioflag);
146 tws_write(struct cdev *dev, struct uio *uio, int ioflag)
148 struct tws_softc *sc = dev->si_drv1;
151 TWS_TRACE_DEBUG(sc, "entry", dev, ioflag);
155 /* PCI Support Functions */
158 * Compare the device ID of this device against the IDs that this driver
159 * supports. If there is a match, set the description and return success.
162 tws_probe(device_t dev)
164 static u_int8_t first_ctlr = 1;
166 if ((pci_get_vendor(dev) == TWS_VENDOR_ID) &&
167 (pci_get_device(dev) == TWS_DEVICE_ID)) {
168 device_set_desc(dev, "LSI 3ware SAS/SATA Storage Controller");
170 printf("LSI 3ware device driver for SAS/SATA storage "
171 "controllers, version: %s\n", TWS_DRIVER_VERSION_STRING);
175 return(BUS_PROBE_DEFAULT);
180 /* Attach function is only called if the probe is successful. */
183 tws_attach(device_t dev)
185 struct tws_softc *sc = device_get_softc(dev);
190 /* Look up our softc and initialize its fields. */
192 sc->device_id = pci_get_device(dev);
193 sc->subvendor_id = pci_get_subvendor(dev);
194 sc->subdevice_id = pci_get_subdevice(dev);
196 /* Intialize mutexes */
197 mtx_init( &sc->q_lock, "tws_q_lock", NULL, MTX_DEF);
198 mtx_init( &sc->sim_lock, "tws_sim_lock", NULL, MTX_DEF);
199 mtx_init( &sc->gen_lock, "tws_gen_lock", NULL, MTX_DEF);
200 mtx_init( &sc->io_lock, "tws_io_lock", NULL, MTX_DEF | MTX_RECURSE);
201 callout_init(&sc->stats_timer, 1);
203 if ( tws_init_trace_q(sc) == FAILURE )
204 printf("trace init failure\n");
205 /* send init event */
206 mtx_lock(&sc->gen_lock);
207 tws_send_event(sc, TWS_INIT_START);
208 mtx_unlock(&sc->gen_lock);
211 #if _BYTE_ORDER == _BIG_ENDIAN
212 TWS_TRACE(sc, "BIG endian", 0, 0);
214 /* sysctl context setup */
215 sysctl_ctx_init(&sc->tws_clist);
216 sc->tws_oidp = SYSCTL_ADD_NODE(&sc->tws_clist,
217 SYSCTL_STATIC_CHILDREN(_hw), OID_AUTO,
218 device_get_nameunit(dev),
220 if ( sc->tws_oidp == NULL ) {
221 tws_log(sc, SYSCTL_TREE_NODE_ADD);
224 SYSCTL_ADD_STRING(&sc->tws_clist, SYSCTL_CHILDREN(sc->tws_oidp),
225 OID_AUTO, "driver_version", CTLFLAG_RD,
226 TWS_DRIVER_VERSION_STRING, 0, "TWS driver version");
228 pci_enable_busmaster(dev);
230 bar = pci_read_config(dev, TWS_PCI_BAR0, 4);
231 TWS_TRACE_DEBUG(sc, "bar0 ", bar, 0);
232 bar = pci_read_config(dev, TWS_PCI_BAR1, 4);
233 bar = bar & ~TWS_BIT2;
234 TWS_TRACE_DEBUG(sc, "bar1 ", bar, 0);
236 /* MFA base address is BAR2 register used for
237 * push mode. Firmware will evatualy move to
238 * pull mode during witch this needs to change
240 #ifndef TWS_PULL_MODE_ENABLE
241 sc->mfa_base = (u_int64_t)pci_read_config(dev, TWS_PCI_BAR2, 4);
242 sc->mfa_base = sc->mfa_base & ~TWS_BIT2;
243 TWS_TRACE_DEBUG(sc, "bar2 ", sc->mfa_base, 0);
246 /* allocate MMIO register space */
247 sc->reg_res_id = TWS_PCI_BAR1; /* BAR1 offset */
248 if ((sc->reg_res = bus_alloc_resource_any(dev, SYS_RES_MEMORY,
249 &(sc->reg_res_id), RF_ACTIVE))
251 tws_log(sc, ALLOC_MEMORY_RES);
254 sc->bus_tag = rman_get_bustag(sc->reg_res);
255 sc->bus_handle = rman_get_bushandle(sc->reg_res);
257 #ifndef TWS_PULL_MODE_ENABLE
258 /* Allocate bus space for inbound mfa */
259 sc->mfa_res_id = TWS_PCI_BAR2; /* BAR2 offset */
260 if ((sc->mfa_res = bus_alloc_resource_any(dev, SYS_RES_MEMORY,
261 &(sc->mfa_res_id), RF_ACTIVE))
263 tws_log(sc, ALLOC_MEMORY_RES);
266 sc->bus_mfa_tag = rman_get_bustag(sc->mfa_res);
267 sc->bus_mfa_handle = rman_get_bushandle(sc->mfa_res);
270 /* Allocate and register our interrupt. */
271 sc->intr_type = TWS_INTx; /* default */
273 if ( tws_enable_msi )
274 sc->intr_type = TWS_MSI;
275 if ( tws_setup_irq(sc) == FAILURE ) {
276 tws_log(sc, ALLOC_MEMORY_RES);
281 * Create a /dev entry for this device. The kernel will assign us
282 * a major number automatically. We use the unit number of this
283 * device as the minor number and name the character device
286 sc->tws_cdev = make_dev(&tws_cdevsw, device_get_unit(dev),
287 UID_ROOT, GID_OPERATOR, S_IRUSR | S_IWUSR, "tws%u",
288 device_get_unit(dev));
289 sc->tws_cdev->si_drv1 = sc;
291 if ( tws_init(sc) == FAILURE ) {
292 tws_log(sc, TWS_INIT_FAILURE);
295 if ( tws_init_ctlr(sc) == FAILURE ) {
296 tws_log(sc, TWS_CTLR_INIT_FAILURE);
299 if ((error = tws_cam_attach(sc))) {
300 tws_log(sc, TWS_CAM_ATTACH);
303 /* send init complete event */
304 mtx_lock(&sc->gen_lock);
305 tws_send_event(sc, TWS_INIT_COMPLETE);
306 mtx_unlock(&sc->gen_lock);
308 TWS_TRACE_DEBUG(sc, "attached successfully", 0, sc->device_id);
312 tws_teardown_intr(sc);
313 destroy_dev(sc->tws_cdev);
314 if (sc->dma_mem_phys)
315 bus_dmamap_unload(sc->cmd_tag, sc->cmd_map);
317 bus_dmamem_free(sc->cmd_tag, sc->dma_mem, sc->cmd_map);
319 bus_dma_tag_destroy(sc->cmd_tag);
321 for(i=0;i<sc->irqs;i++) {
322 if ( sc->irq_res[i] ){
323 if (bus_release_resource(sc->tws_dev,
324 SYS_RES_IRQ, sc->irq_res_id[i], sc->irq_res[i]))
325 TWS_TRACE(sc, "bus irq res", 0, 0);
328 #ifndef TWS_PULL_MODE_ENABLE
332 if (bus_release_resource(sc->tws_dev,
333 SYS_RES_MEMORY, sc->mfa_res_id, sc->mfa_res))
334 TWS_TRACE(sc, "bus release ", 0, sc->mfa_res_id);
337 if (bus_release_resource(sc->tws_dev,
338 SYS_RES_MEMORY, sc->reg_res_id, sc->reg_res))
339 TWS_TRACE(sc, "bus release2 ", 0, sc->reg_res_id);
342 mtx_destroy(&sc->q_lock);
343 mtx_destroy(&sc->sim_lock);
344 mtx_destroy(&sc->gen_lock);
345 mtx_destroy(&sc->io_lock);
346 sysctl_ctx_free(&sc->tws_clist);
353 tws_detach(device_t dev)
355 struct tws_softc *sc = device_get_softc(dev);
359 TWS_TRACE_DEBUG(sc, "entry", 0, 0);
361 mtx_lock(&sc->gen_lock);
362 tws_send_event(sc, TWS_UNINIT_START);
363 mtx_unlock(&sc->gen_lock);
365 /* needs to disable interrupt before detaching from cam */
366 tws_turn_off_interrupts(sc);
367 /* clear door bell */
368 tws_write_reg(sc, TWS_I2O0_HOBDBC, ~0, 4);
369 reg = tws_read_reg(sc, TWS_I2O0_HIMASK, 4);
370 TWS_TRACE_DEBUG(sc, "turn-off-intr", reg, 0);
371 sc->obfl_q_overrun = false;
372 tws_init_connect(sc, 1);
374 /* Teardown the state in our softc created in our attach routine. */
375 /* Disconnect the interrupt handler. */
376 tws_teardown_intr(sc);
378 /* Release irq resource */
379 for(i=0;i<sc->irqs;i++) {
380 if ( sc->irq_res[i] ){
381 if (bus_release_resource(sc->tws_dev,
382 SYS_RES_IRQ, sc->irq_res_id[i], sc->irq_res[i]))
383 TWS_TRACE(sc, "bus release irq resource",
384 i, sc->irq_res_id[i]);
387 if ( sc->intr_type == TWS_MSI ) {
388 pci_release_msi(sc->tws_dev);
393 if (sc->dma_mem_phys)
394 bus_dmamap_unload(sc->cmd_tag, sc->cmd_map);
396 bus_dmamem_free(sc->cmd_tag, sc->dma_mem, sc->cmd_map);
398 bus_dma_tag_destroy(sc->cmd_tag);
400 /* Release memory resource */
402 if (bus_release_resource(sc->tws_dev,
403 SYS_RES_MEMORY, sc->mfa_res_id, sc->mfa_res))
404 TWS_TRACE(sc, "bus release mem resource", 0, sc->mfa_res_id);
407 if (bus_release_resource(sc->tws_dev,
408 SYS_RES_MEMORY, sc->reg_res_id, sc->reg_res))
409 TWS_TRACE(sc, "bus release mem resource", 0, sc->reg_res_id);
412 for ( i=0; i< tws_queue_depth; i++) {
413 if (sc->reqs[i].dma_map)
414 bus_dmamap_destroy(sc->data_tag, sc->reqs[i].dma_map);
415 callout_drain(&sc->reqs[i].timeout);
418 callout_drain(&sc->stats_timer);
419 free(sc->reqs, M_TWS);
420 free(sc->sense_bufs, M_TWS);
421 free(sc->scan_ccb, M_TWS);
422 if (sc->ioctl_data_mem)
423 bus_dmamem_free(sc->data_tag, sc->ioctl_data_mem, sc->ioctl_data_map);
425 bus_dma_tag_destroy(sc->data_tag);
426 free(sc->aen_q.q, M_TWS);
427 free(sc->trace_q.q, M_TWS);
428 mtx_destroy(&sc->q_lock);
429 mtx_destroy(&sc->sim_lock);
430 mtx_destroy(&sc->gen_lock);
431 mtx_destroy(&sc->io_lock);
432 destroy_dev(sc->tws_cdev);
433 sysctl_ctx_free(&sc->tws_clist);
438 tws_setup_intr(struct tws_softc *sc, int irqs)
442 for(i=0;i<irqs;i++) {
443 if (!(sc->intr_handle[i])) {
444 if ((error = bus_setup_intr(sc->tws_dev, sc->irq_res[i],
445 INTR_TYPE_CAM | INTR_MPSAFE,
446 #if (__FreeBSD_version >= 700000)
449 tws_intr, sc, &sc->intr_handle[i]))) {
450 tws_log(sc, SETUP_INTR_RES);
461 tws_teardown_intr(struct tws_softc *sc)
465 for(i=0;i<sc->irqs;i++) {
466 if (sc->intr_handle[i]) {
467 error = bus_teardown_intr(sc->tws_dev,
468 sc->irq_res[i], sc->intr_handle[i]);
469 sc->intr_handle[i] = NULL;
477 tws_setup_irq(struct tws_softc *sc)
481 switch(sc->intr_type) {
484 sc->irq_res_id[0] = 0;
485 sc->irq_res[0] = bus_alloc_resource_any(sc->tws_dev, SYS_RES_IRQ,
486 &sc->irq_res_id[0], RF_SHAREABLE | RF_ACTIVE);
487 if ( ! sc->irq_res[0] )
489 if ( tws_setup_intr(sc, sc->irqs) == FAILURE )
491 device_printf(sc->tws_dev, "Using legacy INTx\n");
495 sc->irq_res_id[0] = 1;
497 if (pci_alloc_msi(sc->tws_dev, &messages) != 0 ) {
498 TWS_TRACE(sc, "pci alloc msi fail", 0, messages);
501 sc->irq_res[0] = bus_alloc_resource_any(sc->tws_dev, SYS_RES_IRQ,
502 &sc->irq_res_id[0], RF_SHAREABLE | RF_ACTIVE);
504 if ( !sc->irq_res[0] )
506 if ( tws_setup_intr(sc, sc->irqs) == FAILURE )
508 device_printf(sc->tws_dev, "Using MSI\n");
517 tws_init(struct tws_softc *sc)
520 u_int32_t max_sg_elements;
521 u_int32_t dma_mem_size;
526 if ( tws_queue_depth > TWS_MAX_REQS )
527 tws_queue_depth = TWS_MAX_REQS;
528 if (tws_queue_depth < TWS_RESERVED_REQS+1)
529 tws_queue_depth = TWS_RESERVED_REQS+1;
530 sc->is64bit = (sizeof(bus_addr_t) == 8) ? true : false;
531 max_sg_elements = (sc->is64bit && !tws_use_32bit_sgls) ?
532 TWS_MAX_64BIT_SG_ELEMENTS :
533 TWS_MAX_32BIT_SG_ELEMENTS;
534 dma_mem_size = (sizeof(struct tws_command_packet) * tws_queue_depth) +
536 if ( bus_dma_tag_create(bus_get_dma_tag(sc->tws_dev), /* PCI parent */
537 TWS_ALIGNMENT, /* alignment */
539 BUS_SPACE_MAXADDR_32BIT, /* lowaddr */
540 BUS_SPACE_MAXADDR, /* highaddr */
541 NULL, NULL, /* filter, filterarg */
542 BUS_SPACE_MAXSIZE, /* maxsize */
543 max_sg_elements, /* numsegs */
544 BUS_SPACE_MAXSIZE, /* maxsegsize */
546 NULL, NULL, /* lockfunc, lockfuncarg */
547 &sc->parent_tag /* tag */
549 TWS_TRACE_DEBUG(sc, "DMA parent tag Create fail", max_sg_elements,
553 /* In bound message frame requires 16byte alignment.
554 * Outbound MF's can live with 4byte alignment - for now just
557 if ( bus_dma_tag_create(sc->parent_tag, /* parent */
558 TWS_IN_MF_ALIGNMENT, /* alignment */
560 BUS_SPACE_MAXADDR_32BIT, /* lowaddr */
561 BUS_SPACE_MAXADDR, /* highaddr */
562 NULL, NULL, /* filter, filterarg */
563 dma_mem_size, /* maxsize */
565 BUS_SPACE_MAXSIZE, /* maxsegsize */
567 NULL, NULL, /* lockfunc, lockfuncarg */
568 &sc->cmd_tag /* tag */
570 TWS_TRACE_DEBUG(sc, "DMA cmd tag Create fail", max_sg_elements, sc->is64bit);
574 if (bus_dmamem_alloc(sc->cmd_tag, &sc->dma_mem,
575 BUS_DMA_NOWAIT, &sc->cmd_map)) {
576 TWS_TRACE_DEBUG(sc, "DMA mem alloc fail", max_sg_elements, sc->is64bit);
580 /* if bus_dmamem_alloc succeeds then bus_dmamap_load will succeed */
582 error = bus_dmamap_load(sc->cmd_tag, sc->cmd_map, sc->dma_mem,
583 dma_mem_size, tws_dmamap_cmds_load_cbfn,
584 &sc->dma_mem_phys, 0);
587 * Create a dma tag for data buffers; size will be the maximum
588 * possible I/O size (128kB).
590 if (bus_dma_tag_create(sc->parent_tag, /* parent */
591 TWS_ALIGNMENT, /* alignment */
593 BUS_SPACE_MAXADDR_32BIT,/* lowaddr */
594 BUS_SPACE_MAXADDR, /* highaddr */
595 NULL, NULL, /* filter, filterarg */
596 TWS_MAX_IO_SIZE, /* maxsize */
597 max_sg_elements, /* nsegments */
598 TWS_MAX_IO_SIZE, /* maxsegsize */
599 BUS_DMA_ALLOCNOW, /* flags */
600 busdma_lock_mutex, /* lockfunc */
601 &sc->io_lock, /* lockfuncarg */
602 &sc->data_tag /* tag */)) {
603 TWS_TRACE_DEBUG(sc, "DMA cmd tag Create fail", max_sg_elements, sc->is64bit);
607 sc->reqs = malloc(sizeof(struct tws_request) * tws_queue_depth, M_TWS,
609 sc->sense_bufs = malloc(sizeof(struct tws_sense) * tws_queue_depth, M_TWS,
611 sc->scan_ccb = malloc(sizeof(union ccb), M_TWS, M_WAITOK | M_ZERO);
612 if (bus_dmamem_alloc(sc->data_tag, (void **)&sc->ioctl_data_mem,
613 (BUS_DMA_NOWAIT | BUS_DMA_ZERO), &sc->ioctl_data_map)) {
614 device_printf(sc->tws_dev, "Cannot allocate ioctl data mem\n");
618 if ( !tws_ctlr_ready(sc) )
619 if( !tws_ctlr_reset(sc) )
622 bzero(&sc->stats, sizeof(struct tws_stats));
624 tws_turn_off_interrupts(sc);
627 * enable pull mode by setting bit1 .
628 * setting bit0 to 1 will enable interrupt coalesing
632 #ifdef TWS_PULL_MODE_ENABLE
634 reg = tws_read_reg(sc, TWS_I2O0_CTL, 4);
635 TWS_TRACE_DEBUG(sc, "i20 ctl", reg, TWS_I2O0_CTL);
636 tws_write_reg(sc, TWS_I2O0_CTL, reg | TWS_BIT1, 4);
640 TWS_TRACE_DEBUG(sc, "dma_mem_phys", sc->dma_mem_phys, TWS_I2O0_CTL);
641 if ( tws_init_reqs(sc, dma_mem_size) == FAILURE )
643 if ( tws_init_aen_q(sc) == FAILURE )
651 tws_init_aen_q(struct tws_softc *sc)
656 sc->aen_q.overflow=0;
657 sc->aen_q.q = malloc(sizeof(struct tws_event_packet)*sc->aen_q.depth,
658 M_TWS, M_WAITOK | M_ZERO);
663 tws_init_trace_q(struct tws_softc *sc)
667 sc->trace_q.depth=256;
668 sc->trace_q.overflow=0;
669 sc->trace_q.q = malloc(sizeof(struct tws_trace_rec)*sc->trace_q.depth,
670 M_TWS, M_WAITOK | M_ZERO);
675 tws_init_reqs(struct tws_softc *sc, u_int32_t dma_mem_size)
678 struct tws_command_packet *cmd_buf;
679 cmd_buf = (struct tws_command_packet *)sc->dma_mem;
682 bzero(cmd_buf, dma_mem_size);
683 TWS_TRACE_DEBUG(sc, "phy cmd", sc->dma_mem_phys, 0);
684 mtx_lock(&sc->q_lock);
685 for ( i=0; i< tws_queue_depth; i++)
687 if (bus_dmamap_create(sc->data_tag, 0, &sc->reqs[i].dma_map)) {
688 /* log a ENOMEM failure msg here */
689 mtx_unlock(&sc->q_lock);
692 sc->reqs[i].cmd_pkt = &cmd_buf[i];
694 sc->sense_bufs[i].hdr = &cmd_buf[i].hdr ;
695 sc->sense_bufs[i].hdr_pkt_phy = sc->dma_mem_phys +
696 (i * sizeof(struct tws_command_packet));
698 sc->reqs[i].cmd_pkt_phy = sc->dma_mem_phys +
699 sizeof(struct tws_command_header) +
700 (i * sizeof(struct tws_command_packet));
701 sc->reqs[i].request_id = i;
704 sc->reqs[i].cmd_pkt->hdr.header_desc.size_header = 128;
706 callout_init(&sc->reqs[i].timeout, 1);
707 sc->reqs[i].state = TWS_REQ_STATE_FREE;
708 if ( i >= TWS_RESERVED_REQS )
709 tws_q_insert_tail(sc, &sc->reqs[i], TWS_FREE_Q);
711 mtx_unlock(&sc->q_lock);
716 tws_dmamap_cmds_load_cbfn(void *arg, bus_dma_segment_t *segs,
720 /* printf("command load done \n"); */
722 *((bus_addr_t *)arg) = segs[0].ds_addr;
726 tws_send_event(struct tws_softc *sc, u_int8_t event)
728 mtx_assert(&sc->gen_lock, MA_OWNED);
729 TWS_TRACE_DEBUG(sc, "received event ", 0, event);
733 sc->tws_state = TWS_INIT;
736 case TWS_INIT_COMPLETE:
737 if (sc->tws_state != TWS_INIT) {
738 device_printf(sc->tws_dev, "invalid state transition %d => TWS_ONLINE\n", sc->tws_state);
740 sc->tws_state = TWS_ONLINE;
744 case TWS_RESET_START:
745 /* We can transition to reset state from any state except reset*/
746 if (sc->tws_state != TWS_RESET) {
747 sc->tws_prev_state = sc->tws_state;
748 sc->tws_state = TWS_RESET;
752 case TWS_RESET_COMPLETE:
753 if (sc->tws_state != TWS_RESET) {
754 device_printf(sc->tws_dev, "invalid state transition %d => %d (previous state)\n", sc->tws_state, sc->tws_prev_state);
756 sc->tws_state = sc->tws_prev_state;
760 case TWS_SCAN_FAILURE:
761 if (sc->tws_state != TWS_ONLINE) {
762 device_printf(sc->tws_dev, "invalid state transition %d => TWS_OFFLINE\n", sc->tws_state);
764 sc->tws_state = TWS_OFFLINE;
768 case TWS_UNINIT_START:
769 if ((sc->tws_state != TWS_ONLINE) && (sc->tws_state != TWS_OFFLINE)) {
770 device_printf(sc->tws_dev, "invalid state transition %d => TWS_UNINIT\n", sc->tws_state);
772 sc->tws_state = TWS_UNINIT;
780 tws_get_state(struct tws_softc *sc)
783 return((u_int8_t)sc->tws_state);
787 /* Called during system shutdown after sync. */
790 tws_shutdown(device_t dev)
793 struct tws_softc *sc = device_get_softc(dev);
795 TWS_TRACE_DEBUG(sc, "entry", 0, 0);
797 tws_turn_off_interrupts(sc);
798 tws_init_connect(sc, 1);
804 * Device suspend routine.
807 tws_suspend(device_t dev)
809 struct tws_softc *sc = device_get_softc(dev);
812 TWS_TRACE_DEBUG(sc, "entry", 0, 0);
817 * Device resume routine.
820 tws_resume(device_t dev)
823 struct tws_softc *sc = device_get_softc(dev);
826 TWS_TRACE_DEBUG(sc, "entry", 0, 0);
832 tws_get_request(struct tws_softc *sc, u_int16_t type)
834 struct mtx *my_mutex = ((type == TWS_REQ_TYPE_SCSI_IO) ? &sc->q_lock : &sc->gen_lock);
835 struct tws_request *r = NULL;
839 if (type == TWS_REQ_TYPE_SCSI_IO) {
840 r = tws_q_remove_head(sc, TWS_FREE_Q);
842 if ( sc->reqs[type].state == TWS_REQ_STATE_FREE ) {
848 bzero(&r->cmd_pkt->cmd, sizeof(struct tws_command_apache));
852 r->flags = TWS_DIR_UNKNOWN;
853 r->error_code = TWS_REQ_RET_INVALID;
856 callout_stop(&r->timeout);
857 r->next = r->prev = NULL;
859 r->state = ((type == TWS_REQ_TYPE_SCSI_IO) ? TWS_REQ_STATE_TRAN : TWS_REQ_STATE_BUSY);
862 mtx_unlock(my_mutex);
868 tws_release_request(struct tws_request *req)
871 struct tws_softc *sc = req->sc;
873 TWS_TRACE_DEBUG(sc, "entry", sc, 0);
874 mtx_lock(&sc->q_lock);
875 tws_q_insert_tail(sc, req, TWS_FREE_Q);
876 mtx_unlock(&sc->q_lock);
879 static device_method_t tws_methods[] = {
880 /* Device interface */
881 DEVMETHOD(device_probe, tws_probe),
882 DEVMETHOD(device_attach, tws_attach),
883 DEVMETHOD(device_detach, tws_detach),
884 DEVMETHOD(device_shutdown, tws_shutdown),
885 DEVMETHOD(device_suspend, tws_suspend),
886 DEVMETHOD(device_resume, tws_resume),
891 static driver_t tws_driver = {
894 sizeof(struct tws_softc)
898 static devclass_t tws_devclass;
900 /* DEFINE_CLASS_0(tws, tws_driver, tws_methods, sizeof(struct tws_softc)); */
901 DRIVER_MODULE(tws, pci, tws_driver, tws_devclass, 0, 0);
902 MODULE_DEPEND(tws, cam, 1, 1, 1);
903 MODULE_DEPEND(tws, pci, 1, 1, 1);
905 TUNABLE_INT("hw.tws.queue_depth", &tws_queue_depth);
906 TUNABLE_INT("hw.tws.enable_msi", &tws_enable_msi);