3 * Generic defines for LSI '909 FC adapters.
6 * Copyright (c) 2000, 2001 by Greg Ansley
8 * Redistribution and use in source and binary forms, with or without
9 * modification, are permitted provided that the following conditions
11 * 1. Redistributions of source code must retain the above copyright
12 * notice immediately at the beginning of the file, without modification,
13 * this list of conditions, and the following disclaimer.
14 * 2. The name of the author may not be used to endorse or promote products
15 * derived from this software without specific prior written permission.
17 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
18 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
19 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
20 * ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE FOR
21 * ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
22 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
23 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
24 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
25 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
26 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
30 * Copyright (c) 2002, 2006 by Matthew Jacob
31 * All rights reserved.
33 * Redistribution and use in source and binary forms, with or without
34 * modification, are permitted provided that the following conditions are
36 * 1. Redistributions of source code must retain the above copyright
37 * notice, this list of conditions and the following disclaimer.
38 * 2. Redistributions in binary form must reproduce at minimum a disclaimer
39 * substantially similar to the "NO WARRANTY" disclaimer below
40 * ("Disclaimer") and any redistribution must be conditioned upon including
41 * a substantially similar Disclaimer requirement for further binary
43 * 3. Neither the names of the above listed copyright holders nor the names
44 * of any contributors may be used to endorse or promote products derived
45 * from this software without specific prior written permission.
47 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
48 * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
49 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
50 * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE
51 * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
52 * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
53 * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
54 * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
55 * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
56 * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF THE COPYRIGHT
57 * OWNER OR CONTRIBUTOR IS ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
59 * Support from Chris Ellsworth in order to make SAS adapters work
60 * is gratefully acknowledged.
63 * Support from LSI-Logic has also gone a great deal toward making this a
64 * workable subsystem and is gratefully acknowledged.
67 * Copyright (c) 2004, Avid Technology, Inc. and its contributors.
68 * Copyright (c) 2004, 2005 Justin T. Gibbs
69 * Copyright (c) 2005, WHEEL Sp. z o.o.
70 * All rights reserved.
72 * Redistribution and use in source and binary forms, with or without
73 * modification, are permitted provided that the following conditions are
75 * 1. Redistributions of source code must retain the above copyright
76 * notice, this list of conditions and the following disclaimer.
77 * 2. Redistributions in binary form must reproduce at minimum a disclaimer
78 * substantially similar to the "NO WARRANTY" disclaimer below
79 * ("Disclaimer") and any redistribution must be conditioned upon including
80 * a substantially similar Disclaimer requirement for further binary
82 * 3. Neither the names of the above listed copyright holders nor the names
83 * of any contributors may be used to endorse or promote products derived
84 * from this software without specific prior written permission.
86 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
87 * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
88 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
89 * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE
90 * LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
91 * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
92 * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
93 * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
94 * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
95 * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF THE COPYRIGHT
96 * OWNER OR CONTRIBUTOR IS ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
102 /********************************* OS Includes ********************************/
103 #include <sys/types.h>
104 #include <sys/param.h>
105 #include <sys/systm.h>
106 #include <sys/endian.h>
107 #include <sys/eventhandler.h>
108 #if __FreeBSD_version < 500000
109 #include <sys/kernel.h>
110 #include <sys/queue.h>
111 #include <sys/malloc.h>
112 #include <sys/devicestat.h>
114 #include <sys/lock.h>
115 #include <sys/kernel.h>
116 #include <sys/queue.h>
117 #include <sys/malloc.h>
118 #include <sys/mutex.h>
119 #include <sys/condvar.h>
121 #include <sys/proc.h>
123 #include <sys/module.h>
125 #include <machine/cpu.h>
126 #include <machine/resource.h>
128 #if __FreeBSD_version < 500000
129 #include <machine/bus.h>
130 #include <machine/clock.h>
134 #include <dev/ofw/openfirm.h>
135 #include <machine/ofw_machdep.h>
138 #include <sys/rman.h>
140 #if __FreeBSD_version < 500000
141 #include <pci/pcireg.h>
142 #include <pci/pcivar.h>
144 #include <dev/pci/pcireg.h>
145 #include <dev/pci/pcivar.h>
148 #include <machine/bus.h>
151 /**************************** Register Definitions ****************************/
152 #include <dev/mpt/mpt_reg.h>
154 /******************************* MPI Definitions ******************************/
155 #include <dev/mpt/mpilib/mpi_type.h>
156 #include <dev/mpt/mpilib/mpi.h>
157 #include <dev/mpt/mpilib/mpi_cnfg.h>
158 #include <dev/mpt/mpilib/mpi_ioc.h>
159 #include <dev/mpt/mpilib/mpi_raid.h>
161 /* XXX For mpt_debug.c */
162 #include <dev/mpt/mpilib/mpi_init.h>
164 #define MPT_S64_2_SCALAR(y) ((((int64_t)y.High) << 32) | (y.Low))
165 #define MPT_U64_2_SCALAR(y) ((((uint64_t)y.High) << 32) | (y.Low))
167 /****************************** Misc Definitions ******************************/
168 /* #define MPT_TEST_MULTIPATH 1 */
170 #define MPT_FAIL (0x10000)
172 #define NUM_ELEMENTS(array) (sizeof(array) / sizeof(*array))
174 #define MPT_ROLE_NONE 0
175 #define MPT_ROLE_INITIATOR 1
176 #define MPT_ROLE_TARGET 2
177 #define MPT_ROLE_BOTH 3
178 #define MPT_ROLE_DEFAULT MPT_ROLE_INITIATOR
180 #define MPT_INI_ID_NONE -1
182 /**************************** Forward Declarations ****************************/
184 struct mpt_personality;
185 typedef struct req_entry request_t;
187 /************************* Personality Module Support *************************/
188 typedef int mpt_load_handler_t(struct mpt_personality *);
189 typedef int mpt_probe_handler_t(struct mpt_softc *);
190 typedef int mpt_attach_handler_t(struct mpt_softc *);
191 typedef int mpt_enable_handler_t(struct mpt_softc *);
192 typedef void mpt_ready_handler_t(struct mpt_softc *);
193 typedef int mpt_event_handler_t(struct mpt_softc *, request_t *,
194 MSG_EVENT_NOTIFY_REPLY *);
195 typedef void mpt_reset_handler_t(struct mpt_softc *, int /*type*/);
196 /* XXX Add return value and use for veto? */
197 typedef void mpt_shutdown_handler_t(struct mpt_softc *);
198 typedef void mpt_detach_handler_t(struct mpt_softc *);
199 typedef int mpt_unload_handler_t(struct mpt_personality *);
201 struct mpt_personality
204 uint32_t id; /* Assigned identifier. */
205 u_int use_count; /* Instances using personality*/
206 mpt_load_handler_t *load; /* configure personailty */
207 #define MPT_PERS_FIRST_HANDLER(pers) (&(pers)->load)
208 mpt_probe_handler_t *probe; /* configure personailty */
209 mpt_attach_handler_t *attach; /* initialize device instance */
210 mpt_enable_handler_t *enable; /* enable device */
211 mpt_ready_handler_t *ready; /* final open for business */
212 mpt_event_handler_t *event; /* Handle MPI event. */
213 mpt_reset_handler_t *reset; /* Re-init after reset. */
214 mpt_shutdown_handler_t *shutdown; /* Shutdown instance. */
215 mpt_detach_handler_t *detach; /* release device instance */
216 mpt_unload_handler_t *unload; /* Shutdown personality */
217 #define MPT_PERS_LAST_HANDLER(pers) (&(pers)->unload)
220 int mpt_modevent(module_t, int, void *);
222 /* Maximum supported number of personalities. */
223 #define MPT_MAX_PERSONALITIES (15)
225 #define MPT_PERSONALITY_DEPEND(name, dep, vmin, vpref, vmax) \
226 MODULE_DEPEND(name, dep, vmin, vpref, vmax)
228 #define DECLARE_MPT_PERSONALITY(name, order) \
229 static moduledata_t name##_mod = { \
230 #name, mpt_modevent, &name##_personality \
232 DECLARE_MODULE(name, name##_mod, SI_SUB_DRIVERS, order); \
233 MODULE_VERSION(name, 1); \
234 MPT_PERSONALITY_DEPEND(name, mpt_core, 1, 1, 1)
236 /******************************* Bus DMA Support ******************************/
237 /* XXX Need to update bus_dmamap_sync to take a range argument. */
238 #define bus_dmamap_sync_range(dma_tag, dmamap, offset, len, op) \
239 bus_dmamap_sync(dma_tag, dmamap, op)
241 #if __FreeBSD_version < 600000
242 #define bus_get_dma_tag(x) NULL
244 #if __FreeBSD_version >= 501102
245 #define mpt_dma_tag_create(mpt, parent_tag, alignment, boundary, \
246 lowaddr, highaddr, filter, filterarg, \
247 maxsize, nsegments, maxsegsz, flags, \
249 bus_dma_tag_create(parent_tag, alignment, boundary, \
250 lowaddr, highaddr, filter, filterarg, \
251 maxsize, nsegments, maxsegsz, flags, \
252 busdma_lock_mutex, &(mpt)->mpt_lock, \
255 #define mpt_dma_tag_create(mpt, parent_tag, alignment, boundary, \
256 lowaddr, highaddr, filter, filterarg, \
257 maxsize, nsegments, maxsegsz, flags, \
259 bus_dma_tag_create(parent_tag, alignment, boundary, \
260 lowaddr, highaddr, filter, filterarg, \
261 maxsize, nsegments, maxsegsz, flags, \
265 struct mpt_map_info {
266 struct mpt_softc *mpt;
271 void mpt_map_rquest(void *, bus_dma_segment_t *, int, int);
272 /* **************************** NewBUS interrupt Crock ************************/
273 #if __FreeBSD_version < 700031
274 #define mpt_setup_intr(d, i, f, U, if, ifa, hp) \
275 bus_setup_intr(d, i, f, if, ifa, hp)
277 #define mpt_setup_intr bus_setup_intr
280 /* **************************** NewBUS CAM Support ****************************/
281 #if __FreeBSD_version < 700049
282 #define mpt_xpt_bus_register(sim, parent, bus) \
283 xpt_bus_register(sim, bus)
285 #define mpt_xpt_bus_register xpt_bus_register
288 /**************************** Kernel Thread Support ***************************/
289 #if __FreeBSD_version > 800001
290 #define mpt_kthread_create(func, farg, proc_ptr, flags, stackpgs, fmtstr, arg) \
291 kproc_create(func, farg, proc_ptr, flags, stackpgs, fmtstr, arg)
292 #define mpt_kthread_exit(status) \
294 #elif __FreeBSD_version > 500005
295 #define mpt_kthread_create(func, farg, proc_ptr, flags, stackpgs, fmtstr, arg) \
296 kthread_create(func, farg, proc_ptr, flags, stackpgs, fmtstr, arg)
297 #define mpt_kthread_exit(status) \
300 #define mpt_kthread_create(func, farg, proc_ptr, flags, stackpgs, fmtstr, arg) \
301 kthread_create(func, farg, proc_ptr, fmtstr, arg)
302 #define mpt_kthread_exit(status) \
306 /********************************** Endianess *********************************/
307 #define MPT_2_HOST64(ptr, tag) ptr->tag = le64toh(ptr->tag)
308 #define MPT_2_HOST32(ptr, tag) ptr->tag = le32toh(ptr->tag)
309 #define MPT_2_HOST16(ptr, tag) ptr->tag = le16toh(ptr->tag)
311 #define HOST_2_MPT64(ptr, tag) ptr->tag = htole64(ptr->tag)
312 #define HOST_2_MPT32(ptr, tag) ptr->tag = htole32(ptr->tag)
313 #define HOST_2_MPT16(ptr, tag) ptr->tag = htole16(ptr->tag)
315 #if _BYTE_ORDER == _BIG_ENDIAN
316 void mpt2host_sge_simple_union(SGE_SIMPLE_UNION *);
317 void mpt2host_iocfacts_reply(MSG_IOC_FACTS_REPLY *);
318 void mpt2host_portfacts_reply(MSG_PORT_FACTS_REPLY *);
319 void mpt2host_config_page_ioc2(CONFIG_PAGE_IOC_2 *);
320 void mpt2host_config_page_ioc3(CONFIG_PAGE_IOC_3 *);
321 void mpt2host_config_page_scsi_port_0(CONFIG_PAGE_SCSI_PORT_0 *);
322 void mpt2host_config_page_scsi_port_1(CONFIG_PAGE_SCSI_PORT_1 *);
323 void host2mpt_config_page_scsi_port_1(CONFIG_PAGE_SCSI_PORT_1 *);
324 void mpt2host_config_page_scsi_port_2(CONFIG_PAGE_SCSI_PORT_2 *);
325 void mpt2host_config_page_scsi_device_0(CONFIG_PAGE_SCSI_DEVICE_0 *);
326 void mpt2host_config_page_scsi_device_1(CONFIG_PAGE_SCSI_DEVICE_1 *);
327 void host2mpt_config_page_scsi_device_1(CONFIG_PAGE_SCSI_DEVICE_1 *);
328 void mpt2host_config_page_fc_port_0(CONFIG_PAGE_FC_PORT_0 *);
329 void mpt2host_config_page_fc_port_1(CONFIG_PAGE_FC_PORT_1 *);
330 void host2mpt_config_page_fc_port_1(CONFIG_PAGE_FC_PORT_1 *);
331 void mpt2host_config_page_raid_vol_0(CONFIG_PAGE_RAID_VOL_0 *);
332 void mpt2host_config_page_raid_phys_disk_0(CONFIG_PAGE_RAID_PHYS_DISK_0 *);
333 void mpt2host_mpi_raid_vol_indicator(MPI_RAID_VOL_INDICATOR *);
335 #define mpt2host_sge_simple_union(x) do { ; } while (0)
336 #define mpt2host_iocfacts_reply(x) do { ; } while (0)
337 #define mpt2host_portfacts_reply(x) do { ; } while (0)
338 #define mpt2host_config_page_ioc2(x) do { ; } while (0)
339 #define mpt2host_config_page_ioc3(x) do { ; } while (0)
340 #define mpt2host_config_page_scsi_port_0(x) do { ; } while (0)
341 #define mpt2host_config_page_scsi_port_1(x) do { ; } while (0)
342 #define host2mpt_config_page_scsi_port_1(x) do { ; } while (0)
343 #define mpt2host_config_page_scsi_port_2(x) do { ; } while (0)
344 #define mpt2host_config_page_scsi_device_0(x) do { ; } while (0)
345 #define mpt2host_config_page_scsi_device_1(x) do { ; } while (0)
346 #define host2mpt_config_page_scsi_device_1(x) do { ; } while (0)
347 #define mpt2host_config_page_fc_port_0(x) do { ; } while (0)
348 #define mpt2host_config_page_fc_port_1(x) do { ; } while (0)
349 #define host2mpt_config_page_fc_port_1(x) do { ; } while (0)
350 #define mpt2host_config_page_raid_vol_0(x) do { ; } while (0)
351 #define mpt2host_config_page_raid_phys_disk_0(x) \
353 #define mpt2host_mpi_raid_vol_indicator(x) do { ; } while (0)
356 /**************************** MPI Transaction State ***************************/
358 REQ_STATE_NIL = 0x00,
359 REQ_STATE_FREE = 0x01,
360 REQ_STATE_ALLOCATED = 0x02,
361 REQ_STATE_QUEUED = 0x04,
362 REQ_STATE_DONE = 0x08,
363 REQ_STATE_TIMEDOUT = 0x10,
364 REQ_STATE_NEED_WAKEUP = 0x20,
365 REQ_STATE_LOCKED = 0x80, /* can't be freed */
366 REQ_STATE_MASK = 0xFF
370 TAILQ_ENTRY(req_entry) links; /* Pointer to next in list */
371 mpt_req_state_t state; /* Request State Information */
372 uint16_t index; /* Index of this entry */
373 uint16_t IOCStatus; /* Completion status */
374 uint16_t ResponseCode; /* TMF Reponse Code */
375 uint16_t serno; /* serial number */
376 union ccb *ccb; /* CAM request */
377 void *req_vbuf; /* Virtual Address of Entry */
378 void *sense_vbuf; /* Virtual Address of sense data */
379 bus_addr_t req_pbuf; /* Physical Address of Entry */
380 bus_addr_t sense_pbuf; /* Physical Address of sense data */
381 bus_dmamap_t dmap; /* DMA map for data buffers */
382 struct req_entry *chain; /* for SGE overallocations */
383 struct callout callout; /* Timeout for the request */
386 typedef struct mpt_config_params {
397 /**************************** MPI Target State Info ***************************/
400 uint32_t reply_desc; /* current reply descriptor */
401 uint32_t resid; /* current data residual */
402 uint32_t bytes_xfered; /* current relative offset */
403 union ccb *ccb; /* pointer to currently active ccb */
404 request_t *req; /* pointer to currently active assist request */
414 TGT_STATE_SETTING_UP_FOR_DATA,
415 TGT_STATE_MOVING_DATA,
416 TGT_STATE_MOVING_DATA_AND_STATUS,
417 TGT_STATE_SENDING_STATUS
422 * When we get an incoming command it has its own tag which is called the
423 * IoIndex. This is the value we gave that particular command buffer when
424 * we originally assigned it. It's just a number, really. The FC card uses
425 * it as an RX_ID. We can use it to index into mpt->tgt_cmd_ptrs, which
426 * contains pointers the request_t structures related to that IoIndex.
428 * What *we* do is construct a tag out of the index for the target command
429 * which owns the incoming ATIO plus a rolling sequence number.
431 #define MPT_MAKE_TAGID(mpt, req, ioindex) \
432 ((ioindex << 18) | (((mpt->sequence++) & 0x3f) << 12) | (req->index & 0xfff))
435 #define MPT_TAG_2_REQ(a, b) mpt_tag_2_req(a, (uint32_t) b)
437 #define MPT_TAG_2_REQ(mpt, tag) mpt->tgt_cmd_ptrs[tag >> 18]
440 #define MPT_TGT_STATE(mpt, req) ((mpt_tgt_state_t *) \
441 (&((uint8_t *)req->req_vbuf)[MPT_RQSL(mpt) - sizeof (mpt_tgt_state_t)]))
443 STAILQ_HEAD(mpt_hdr_stailq, ccb_hdr);
444 #define MPT_MAX_LUNS 256
446 struct mpt_hdr_stailq atios;
447 struct mpt_hdr_stailq inots;
450 #define MPT_MAX_ELS 64
452 /**************************** Handler Registration ****************************/
454 * Global table of registered reply handlers. The
455 * handler is indicated by byte 3 of the request
456 * index submitted to the IOC. This allows the
457 * driver core to perform generic processing without
458 * any knowledge of per-personality behavior.
460 * MPT_NUM_REPLY_HANDLERS must be a power of 2
461 * to allow the easy generation of a mask.
463 * The handler offsets used by the core are hard coded
464 * allowing faster code generation when assigning a handler
465 * to a request. All "personalities" must use the
466 * the handler registration mechanism.
468 * The IOC handlers that are rarely executed are placed
469 * at the tail of the table to make it more likely that
470 * all commonly executed handlers fit in a single cache
473 #define MPT_NUM_REPLY_HANDLERS (32)
474 #define MPT_REPLY_HANDLER_EVENTS MPT_CBI_TO_HID(0)
475 #define MPT_REPLY_HANDLER_CONFIG MPT_CBI_TO_HID(MPT_NUM_REPLY_HANDLERS-1)
476 #define MPT_REPLY_HANDLER_HANDSHAKE MPT_CBI_TO_HID(MPT_NUM_REPLY_HANDLERS-2)
477 typedef int mpt_reply_handler_t(struct mpt_softc *mpt, request_t *request,
478 uint32_t reply_desc, MSG_DEFAULT_REPLY *reply_frame);
480 mpt_reply_handler_t *reply_handler;
490 struct mpt_handler_record
492 LIST_ENTRY(mpt_handler_record) links;
493 mpt_handler_t handler;
496 LIST_HEAD(mpt_handler_list, mpt_handler_record);
499 * The handler_id is currently unused but would contain the
500 * handler ID used in the MsgContext field to allow direction
501 * of replies to the handler. Registrations that don't require
502 * a handler id can pass in NULL for the handler_id.
504 * Deregistrations for handlers without a handler id should
505 * pass in MPT_HANDLER_ID_NONE.
507 #define MPT_HANDLER_ID_NONE (0xFFFFFFFF)
508 int mpt_register_handler(struct mpt_softc *, mpt_handler_type,
509 mpt_handler_t, uint32_t *);
510 int mpt_deregister_handler(struct mpt_softc *, mpt_handler_type,
511 mpt_handler_t, uint32_t);
513 /******************* Per-Controller Instance Data Structures ******************/
514 TAILQ_HEAD(req_queue, req_entry);
516 /* Structure for saving proper values for modifyable PCI config registers */
519 uint16_t LatencyTimer_LineSize;
521 uint32_t Mem0_BAR[2];
522 uint32_t Mem1_BAR[2];
530 MPT_RVF_ACTIVE = 0x1,
531 MPT_RVF_ANNOUNCED = 0x2,
532 MPT_RVF_UP2DATE = 0x4,
533 MPT_RVF_REFERENCED = 0x8,
534 MPT_RVF_WCE_CHANGED = 0x10
535 } mpt_raid_volume_flags;
537 struct mpt_raid_volume {
538 CONFIG_PAGE_RAID_VOL_0 *config_page;
539 MPI_RAID_VOL_INDICATOR sync_progress;
540 mpt_raid_volume_flags flags;
541 u_int quiesced_disks;
546 MPT_RDF_ACTIVE = 0x01,
547 MPT_RDF_ANNOUNCED = 0x02,
548 MPT_RDF_UP2DATE = 0x04,
549 MPT_RDF_REFERENCED = 0x08,
550 MPT_RDF_QUIESCING = 0x10,
551 MPT_RDF_QUIESCED = 0x20
552 } mpt_raid_disk_flags;
554 struct mpt_raid_disk {
555 CONFIG_PAGE_RAID_PHYS_DISK_0 config_page;
556 struct mpt_raid_volume *volume;
558 u_int pass_thru_active;
559 mpt_raid_disk_flags flags;
562 struct mpt_evtf_record {
563 MSG_EVENT_NOTIFY_REPLY reply;
565 LIST_ENTRY(mpt_evtf_record) links;
568 LIST_HEAD(mpt_evtf_list, mpt_evtf_record);
570 struct mptsas_devinfo {
572 uint16_t parent_dev_handle;
573 uint16_t enclosure_handle;
576 uint8_t physical_port;
579 uint64_t sas_address;
580 uint32_t device_info;
583 struct mptsas_phyinfo {
587 uint8_t negotiated_link_rate;
588 uint8_t hw_link_rate;
589 uint8_t programmed_link_rate;
590 uint8_t sas_port_add_phy;
591 struct mptsas_devinfo identify;
592 struct mptsas_devinfo attached;
595 struct mptsas_portinfo {
597 struct mptsas_phyinfo *phy_info;
602 #if __FreeBSD_version < 500000
603 uint32_t mpt_islocked;
609 uint32_t mpt_pers_mask;
631 u_int role; /* role: none, ini, target, both */
634 #ifdef MPT_TEST_MULTIPATH
641 MSG_IOC_FACTS_REPLY ioc_facts;
646 MSG_PORT_FACTS_REPLY * port_facts;
647 #define mpt_max_tgtcmds port_facts[0].MaxPostedCmdBuffers
650 * Device Configuration Information
654 CONFIG_PAGE_SCSI_PORT_0 _port_page0;
655 CONFIG_PAGE_SCSI_PORT_1 _port_page1;
656 CONFIG_PAGE_SCSI_PORT_2 _port_page2;
657 CONFIG_PAGE_SCSI_DEVICE_0 _dev_page0[16];
658 CONFIG_PAGE_SCSI_DEVICE_1 _dev_page1[16];
660 uint16_t _tag_enable;
661 uint16_t _disc_enable;
663 #define mpt_port_page0 cfg.spi._port_page0
664 #define mpt_port_page1 cfg.spi._port_page1
665 #define mpt_port_page2 cfg.spi._port_page2
666 #define mpt_dev_page0 cfg.spi._dev_page0
667 #define mpt_dev_page1 cfg.spi._dev_page1
668 #define mpt_ini_id cfg.spi._ini_id
669 #define mpt_tag_enable cfg.spi._tag_enable
670 #define mpt_disc_enable cfg.spi._disc_enable
672 CONFIG_PAGE_FC_PORT_0 _port_page0;
673 uint32_t _port_speed;
674 #define mpt_fcport_page0 cfg.fc._port_page0
675 #define mpt_fcport_speed cfg.fc._port_speed
678 #if __FreeBSD_version >= 500000
680 * Device config information stored up for sysctl to access
684 unsigned int initiator_id;
693 /* Controller Info for RAID information */
694 CONFIG_PAGE_IOC_2 * ioc_page2;
695 CONFIG_PAGE_IOC_3 * ioc_page3;
698 struct mpt_raid_volume* raid_volumes;
699 struct mpt_raid_disk* raid_disks;
700 u_int raid_max_volumes;
701 u_int raid_max_disks;
702 u_int raid_page0_len;
705 u_int raid_resync_rate;
706 u_int raid_mwce_setting;
707 u_int raid_queue_depth;
708 u_int raid_nonopt_volumes;
709 struct proc *raid_thread;
710 struct callout raid_timer;
716 struct resource * pci_irq; /* Interrupt map for chip */
717 void * ih; /* Interupt handle */
718 struct mpt_pci_cfg pci_cfg; /* saved PCI conf registers */
723 struct resource * pci_reg; /* Register map for chip */
724 int pci_mem_rid; /* Resource ID */
725 bus_space_tag_t pci_st; /* Bus tag for registers */
726 bus_space_handle_t pci_sh; /* Bus handle for registers */
727 /* PIO versions of above. */
729 struct resource * pci_pio_reg;
730 bus_space_tag_t pci_pio_st;
731 bus_space_handle_t pci_pio_sh;
733 bus_dma_tag_t parent_dmat; /* DMA tag for parent PCI bus */
734 bus_dma_tag_t reply_dmat; /* DMA tag for reply memory */
735 bus_dmamap_t reply_dmap; /* DMA map for reply memory */
736 uint8_t *reply; /* KVA of reply memory */
737 bus_addr_t reply_phys; /* BusAddr of reply memory */
739 bus_dma_tag_t buffer_dmat; /* DMA tag for buffers */
740 bus_dma_tag_t request_dmat; /* DMA tag for request memroy */
741 bus_dmamap_t request_dmap; /* DMA map for request memroy */
742 uint8_t *request; /* KVA of Request memory */
743 bus_addr_t request_phys; /* BusAddr of request memory */
745 uint32_t max_seg_cnt; /* calculated after IOC facts */
746 uint32_t max_cam_seg_cnt;/* calculated from MAXPHYS*/
749 * Hardware management
754 * CAM && Software Management
756 request_t *request_pool;
757 struct req_queue request_free_list;
758 struct req_queue request_pending_list;
759 struct req_queue request_timeout_list;
763 struct cam_path *path;
765 struct cam_sim *phydisk_sim;
766 struct cam_path *phydisk_path;
768 struct proc *recovery_thread;
772 * Deferred frame acks due to resource shortage.
774 struct mpt_evtf_list ack_frames;
777 * Target Mode Support
779 uint32_t scsi_tgt_handler_id;
780 request_t ** tgt_cmd_ptrs;
781 request_t ** els_cmd_ptrs; /* FC only */
784 * *snork*- this is chosen to be here *just in case* somebody
785 * forgets to point to it exactly and we index off of trt with
788 tgt_resource_t trt_wildcard; /* wildcard luns */
789 tgt_resource_t trt[MPT_MAX_LUNS];
790 uint16_t tgt_cmds_allocated;
791 uint16_t els_cmds_allocated; /* FC only */
793 uint16_t timeouts; /* timeout count */
794 uint16_t success; /* successes afer timeout */
795 uint16_t sequence; /* Sequence Number */
799 /* Paired port in some dual adapters configurations */
800 struct mpt_softc * mpt2;
802 /* FW Image management */
803 uint32_t fw_image_size;
805 bus_dma_tag_t fw_dmat; /* DMA tag for firmware image */
806 bus_dmamap_t fw_dmap; /* DMA map for firmware image */
807 bus_addr_t fw_phys; /* BusAddr of firmware image */
810 struct mptsas_portinfo *sas_portinfo;
812 /* Shutdown Event Handler. */
815 /* Userland management interface. */
818 TAILQ_ENTRY(mpt_softc) links;
821 static __inline void mpt_assign_serno(struct mpt_softc *, request_t *);
824 mpt_assign_serno(struct mpt_softc *mpt, request_t *req)
826 if ((req->serno = mpt->sequence++) == 0) {
827 req->serno = mpt->sequence++;
831 /***************************** Locking Primitives *****************************/
832 #if __FreeBSD_version < 500000
833 #define MPT_IFLAGS INTR_TYPE_CAM
834 #define MPT_LOCK(mpt) mpt_lockspl(mpt)
835 #define MPT_UNLOCK(mpt) mpt_unlockspl(mpt)
836 #define MPT_OWNED(mpt) mpt->mpt_islocked
837 #define MPT_LOCK_ASSERT(mpt)
838 #define MPTLOCK_2_CAMLOCK MPT_UNLOCK
839 #define CAMLOCK_2_MPTLOCK MPT_LOCK
840 #define MPT_LOCK_SETUP(mpt)
841 #define MPT_LOCK_DESTROY(mpt)
843 static __inline void mpt_lockspl(struct mpt_softc *mpt);
844 static __inline void mpt_unlockspl(struct mpt_softc *mpt);
847 mpt_lockspl(struct mpt_softc *mpt)
852 if (mpt->mpt_islocked++ == 0) {
853 mpt->mpt_splsaved = s;
856 panic("Recursed lock with mask: 0x%x\n", s);
861 mpt_unlockspl(struct mpt_softc *mpt)
863 if (mpt->mpt_islocked) {
864 if (--mpt->mpt_islocked == 0) {
865 splx(mpt->mpt_splsaved);
868 panic("Negative lock count\n");
872 mpt_sleep(struct mpt_softc *mpt, void *ident, int priority,
873 const char *wmesg, int timo)
879 KASSERT(mpt->mpt_islocked <= 1, ("Invalid lock count on tsleep"));
880 saved_cnt = mpt->mpt_islocked;
881 saved_spl = mpt->mpt_splsaved;
882 mpt->mpt_islocked = 0;
883 error = tsleep(ident, priority, wmesg, timo);
884 KASSERT(mpt->mpt_islocked == 0, ("Invalid lock count on wakeup"));
885 mpt->mpt_islocked = saved_cnt;
886 mpt->mpt_splsaved = saved_spl;
890 #define mpt_req_timeout(req, ticks, func, arg) \
891 callout_reset(&(req)->callout, (ticks), (func), (arg));
892 #define mpt_req_untimeout(req, func, arg) \
893 callout_stop(&(req)->callout)
894 #define mpt_callout_init(mpt, c) \
896 #define mpt_callout_drain(mpt, c) \
901 #define MPT_IFLAGS INTR_TYPE_CAM | INTR_ENTROPY | INTR_MPSAFE
902 #define MPT_LOCK_SETUP(mpt) \
903 mtx_init(&mpt->mpt_lock, "mpt", NULL, MTX_DEF); \
904 mpt->mpt_locksetup = 1
905 #define MPT_LOCK_DESTROY(mpt) \
906 if (mpt->mpt_locksetup) { \
907 mtx_destroy(&mpt->mpt_lock); \
908 mpt->mpt_locksetup = 0; \
911 #define MPT_LOCK(mpt) mtx_lock(&(mpt)->mpt_lock)
912 #define MPT_UNLOCK(mpt) mtx_unlock(&(mpt)->mpt_lock)
913 #define MPT_OWNED(mpt) mtx_owned(&(mpt)->mpt_lock)
914 #define MPT_LOCK_ASSERT(mpt) mtx_assert(&(mpt)->mpt_lock, MA_OWNED)
915 #define MPTLOCK_2_CAMLOCK(mpt)
916 #define CAMLOCK_2_MPTLOCK(mpt)
917 #define mpt_sleep(mpt, ident, priority, wmesg, timo) \
918 msleep(ident, &(mpt)->mpt_lock, priority, wmesg, timo)
919 #define mpt_req_timeout(req, ticks, func, arg) \
920 callout_reset(&(req)->callout, (ticks), (func), (arg))
921 #define mpt_req_untimeout(req, func, arg) \
922 callout_stop(&(req)->callout)
923 #define mpt_callout_init(mpt, c) \
924 callout_init_mtx(c, &(mpt)->mpt_lock, 0)
925 #define mpt_callout_drain(mpt, c) \
930 #define MPT_IFLAGS INTR_TYPE_CAM | INTR_ENTROPY
931 #define MPT_LOCK_SETUP(mpt) do { } while (0)
932 #define MPT_LOCK_DESTROY(mpt) do { } while (0)
933 #define MPT_LOCK_ASSERT(mpt) mtx_assert(&Giant, MA_OWNED)
934 #define MPT_LOCK(mpt) mtx_lock(&Giant)
935 #define MPT_UNLOCK(mpt) mtx_unlock(&Giant)
936 #define MPTLOCK_2_CAMLOCK(mpt)
937 #define CAMLOCK_2_MPTLOCK(mpt)
939 #define mpt_req_timeout(req, ticks, func, arg) \
940 callout_reset(&(req)->callout, (ticks), (func), (arg))
941 #define mpt_req_untimeout(req, func, arg) \
942 callout_stop(&(req)->callout)
943 #define mpt_callout_init(mpt, c) \
945 #define mpt_callout_drain(mpt, c) \
949 mpt_sleep(struct mpt_softc *, void *, int, const char *, int);
952 mpt_sleep(struct mpt_softc *mpt, void *i, int p, const char *w, int t)
955 r = tsleep(i, p, w, t);
961 /******************************* Register Access ******************************/
962 static __inline void mpt_write(struct mpt_softc *, size_t, uint32_t);
963 static __inline uint32_t mpt_read(struct mpt_softc *, int);
964 static __inline void mpt_pio_write(struct mpt_softc *, size_t, uint32_t);
965 static __inline uint32_t mpt_pio_read(struct mpt_softc *, int);
968 mpt_write(struct mpt_softc *mpt, size_t offset, uint32_t val)
970 bus_space_write_4(mpt->pci_st, mpt->pci_sh, offset, val);
973 static __inline uint32_t
974 mpt_read(struct mpt_softc *mpt, int offset)
976 return (bus_space_read_4(mpt->pci_st, mpt->pci_sh, offset));
980 * Some operations (e.g. diagnostic register writes while the ARM proccessor
981 * is disabled), must be performed using "PCI pio" operations. On non-PCI
982 * busses, these operations likely map to normal register accesses.
985 mpt_pio_write(struct mpt_softc *mpt, size_t offset, uint32_t val)
987 bus_space_write_4(mpt->pci_pio_st, mpt->pci_pio_sh, offset, val);
990 static __inline uint32_t
991 mpt_pio_read(struct mpt_softc *mpt, int offset)
993 return (bus_space_read_4(mpt->pci_pio_st, mpt->pci_pio_sh, offset));
995 /*********************** Reply Frame/Request Management ***********************/
996 /* Max MPT Reply we are willing to accept (must be power of 2) */
997 #define MPT_REPLY_SIZE 256
1000 * Must be less than 16384 in order for target mode to work
1002 #define MPT_MAX_REQUESTS(mpt) 512
1003 #define MPT_REQUEST_AREA 512
1004 #define MPT_SENSE_SIZE 32 /* included in MPT_REQUEST_AREA */
1005 #define MPT_REQ_MEM_SIZE(mpt) (MPT_MAX_REQUESTS(mpt) * MPT_REQUEST_AREA)
1007 #define MPT_CONTEXT_CB_SHIFT (16)
1008 #define MPT_CBI(handle) (handle >> MPT_CONTEXT_CB_SHIFT)
1009 #define MPT_CBI_TO_HID(cbi) ((cbi) << MPT_CONTEXT_CB_SHIFT)
1010 #define MPT_CONTEXT_TO_CBI(x) \
1011 (((x) >> MPT_CONTEXT_CB_SHIFT) & (MPT_NUM_REPLY_HANDLERS - 1))
1012 #define MPT_CONTEXT_REQI_MASK 0xFFFF
1013 #define MPT_CONTEXT_TO_REQI(x) ((x) & MPT_CONTEXT_REQI_MASK)
1016 * Convert a 32bit physical address returned from IOC to an
1017 * offset into our reply frame memory or the kvm address needed
1018 * to access the data. The returned address is only the low
1019 * 32 bits, so mask our base physical address accordingly.
1021 #define MPT_REPLY_BADDR(x) \
1023 #define MPT_REPLY_OTOV(m, i) \
1024 ((void *)(&m->reply[i]))
1026 #define MPT_DUMP_REPLY_FRAME(mpt, reply_frame) \
1028 if (mpt->verbose > MPT_PRT_DEBUG) \
1029 mpt_dump_reply_frame(mpt, reply_frame); \
1032 static __inline uint32_t mpt_pop_reply_queue(struct mpt_softc *mpt);
1033 static __inline void mpt_free_reply(struct mpt_softc *mpt, uint32_t ptr);
1036 * Give the reply buffer back to the IOC after we have
1037 * finished processing it.
1039 static __inline void
1040 mpt_free_reply(struct mpt_softc *mpt, uint32_t ptr)
1042 mpt_write(mpt, MPT_OFFSET_REPLY_Q, ptr);
1045 /* Get a reply from the IOC */
1046 static __inline uint32_t
1047 mpt_pop_reply_queue(struct mpt_softc *mpt)
1049 return mpt_read(mpt, MPT_OFFSET_REPLY_Q);
1053 mpt_complete_request_chain(struct mpt_softc *, struct req_queue *, u_int);
1055 /************************** Scatter Gather Managment **************************/
1056 /* MPT_RQSL- size of request frame, in bytes */
1057 #define MPT_RQSL(mpt) (mpt->ioc_facts.RequestFrameSize << 2)
1059 /* MPT_NSGL- how many SG entries can fit in a request frame size */
1060 #define MPT_NSGL(mpt) (MPT_RQSL(mpt) / sizeof (SGE_IO_UNION))
1062 /* MPT_NRFM- how many request frames can fit in each request alloc we make */
1063 #define MPT_NRFM(mpt) (MPT_REQUEST_AREA / MPT_RQSL(mpt))
1066 * MPT_NSGL_FIRST- # of SG elements that can fit after
1067 * an I/O request but still within the request frame.
1068 * Do this safely based upon SGE_IO_UNION.
1070 * Note that the first element is *within* the SCSI request.
1072 #define MPT_NSGL_FIRST(mpt) \
1073 ((MPT_RQSL(mpt) - sizeof (MSG_SCSI_IO_REQUEST) + sizeof (SGE_IO_UNION)) / \
1074 sizeof (SGE_IO_UNION))
1076 /***************************** IOC Initialization *****************************/
1077 int mpt_reset(struct mpt_softc *, int /*reinit*/);
1079 /****************************** Debugging ************************************/
1080 typedef struct mpt_decode_entry {
1084 } mpt_decode_entry_t;
1086 int mpt_decode_value(mpt_decode_entry_t *table, u_int num_entries,
1087 const char *name, u_int value, u_int *cur_column,
1090 void mpt_dump_data(struct mpt_softc *, const char *, void *, int);
1091 void mpt_dump_request(struct mpt_softc *, request_t *);
1099 MPT_PRT_NEGOTIATION,
1108 #if __FreeBSD_version > 500000
1109 #define mpt_lprt(mpt, level, ...) \
1111 if (level <= (mpt)->verbose) \
1112 mpt_prt(mpt, __VA_ARGS__); \
1115 #define mpt_lprtc(mpt, level, ...) \
1117 if (level <= (mpt)->verbose) \
1118 mpt_prtc(mpt, __VA_ARGS__); \
1121 void mpt_lprt(struct mpt_softc *, int, const char *, ...)
1123 void mpt_lprtc(struct mpt_softc *, int, const char *, ...)
1126 void mpt_prt(struct mpt_softc *, const char *, ...)
1128 void mpt_prtc(struct mpt_softc *, const char *, ...)
1131 /**************************** Target Mode Related ***************************/
1132 static __inline int mpt_cdblen(uint8_t, int);
1134 mpt_cdblen(uint8_t cdb0, int maxlen)
1136 int group = cdb0 >> 5;
1150 static __inline request_t * mpt_tag_2_req(struct mpt_softc *, uint32_t);
1151 static __inline request_t *
1152 mpt_tag_2_req(struct mpt_softc *mpt, uint32_t tag)
1154 uint16_t rtg = (tag >> 18);
1155 KASSERT(rtg < mpt->tgt_cmds_allocated, ("bad tag %d\n", tag));
1156 KASSERT(mpt->tgt_cmd_ptrs, ("no cmd backpointer array"));
1157 KASSERT(mpt->tgt_cmd_ptrs[rtg], ("no cmd backpointer"));
1158 return (mpt->tgt_cmd_ptrs[rtg]);
1163 mpt_req_on_free_list(struct mpt_softc *, request_t *);
1165 mpt_req_on_pending_list(struct mpt_softc *, request_t *);
1168 * Is request on freelist?
1171 mpt_req_on_free_list(struct mpt_softc *mpt, request_t *req)
1175 TAILQ_FOREACH(lrq, &mpt->request_free_list, links) {
1184 * Is request on pending list?
1187 mpt_req_on_pending_list(struct mpt_softc *mpt, request_t *req)
1191 TAILQ_FOREACH(lrq, &mpt->request_pending_list, links) {
1200 static __inline void
1201 mpt_req_spcl(struct mpt_softc *, request_t *, const char *, int);
1202 static __inline void
1203 mpt_req_not_spcl(struct mpt_softc *, request_t *, const char *, int);
1206 * Make sure that req *is* part of one of the special lists
1208 static __inline void
1209 mpt_req_spcl(struct mpt_softc *mpt, request_t *req, const char *s, int line)
1212 for (i = 0; i < mpt->els_cmds_allocated; i++) {
1213 if (req == mpt->els_cmd_ptrs[i]) {
1217 for (i = 0; i < mpt->tgt_cmds_allocated; i++) {
1218 if (req == mpt->tgt_cmd_ptrs[i]) {
1222 panic("%s(%d): req %p:%u function %x not in els or tgt ptrs\n",
1223 s, line, req, req->serno,
1224 ((PTR_MSG_REQUEST_HEADER)req->req_vbuf)->Function);
1228 * Make sure that req is *not* part of one of the special lists.
1230 static __inline void
1231 mpt_req_not_spcl(struct mpt_softc *mpt, request_t *req, const char *s, int line)
1234 for (i = 0; i < mpt->els_cmds_allocated; i++) {
1235 KASSERT(req != mpt->els_cmd_ptrs[i],
1236 ("%s(%d): req %p:%u func %x in els ptrs at ioindex %d\n",
1237 s, line, req, req->serno,
1238 ((PTR_MSG_REQUEST_HEADER)req->req_vbuf)->Function, i));
1240 for (i = 0; i < mpt->tgt_cmds_allocated; i++) {
1241 KASSERT(req != mpt->tgt_cmd_ptrs[i],
1242 ("%s(%d): req %p:%u func %x in tgt ptrs at ioindex %d\n",
1243 s, line, req, req->serno,
1244 ((PTR_MSG_REQUEST_HEADER)req->req_vbuf)->Function, i));
1250 * Task Management Types, purely for internal consumption
1253 MPT_ABORT_TASK_SET=1234,
1258 MPT_NIL_TMT_VALUE=5678
1261 /**************************** Unclassified Routines ***************************/
1262 void mpt_send_cmd(struct mpt_softc *mpt, request_t *req);
1263 int mpt_recv_handshake_reply(struct mpt_softc *mpt,
1264 size_t reply_len, void *reply);
1265 int mpt_wait_req(struct mpt_softc *mpt, request_t *req,
1266 mpt_req_state_t state, mpt_req_state_t mask,
1267 int sleep_ok, int time_ms);
1268 void mpt_enable_ints(struct mpt_softc *mpt);
1269 void mpt_disable_ints(struct mpt_softc *mpt);
1270 int mpt_attach(struct mpt_softc *mpt);
1271 int mpt_shutdown(struct mpt_softc *mpt);
1272 int mpt_detach(struct mpt_softc *mpt);
1273 int mpt_send_handshake_cmd(struct mpt_softc *mpt,
1274 size_t len, void *cmd);
1275 request_t * mpt_get_request(struct mpt_softc *mpt, int sleep_ok);
1276 void mpt_free_request(struct mpt_softc *mpt, request_t *req);
1277 void mpt_intr(void *arg);
1278 void mpt_check_doorbell(struct mpt_softc *mpt);
1279 void mpt_dump_reply_frame(struct mpt_softc *mpt,
1280 MSG_DEFAULT_REPLY *reply_frame);
1282 void mpt_set_config_regs(struct mpt_softc *);
1283 int mpt_issue_cfg_req(struct mpt_softc */*mpt*/, request_t */*req*/,
1285 bus_addr_t /*addr*/, bus_size_t/*len*/,
1286 int /*sleep_ok*/, int /*timeout_ms*/);
1287 int mpt_read_extcfg_header(struct mpt_softc *mpt, int PageVersion,
1288 int PageNumber, uint32_t PageAddress,
1290 CONFIG_EXTENDED_PAGE_HEADER *rslt,
1291 int sleep_ok, int timeout_ms);
1292 int mpt_read_extcfg_page(struct mpt_softc *mpt, int Action,
1293 uint32_t PageAddress,
1294 CONFIG_EXTENDED_PAGE_HEADER *hdr,
1295 void *buf, size_t len, int sleep_ok,
1297 int mpt_read_cfg_header(struct mpt_softc *, int /*PageType*/,
1299 uint32_t /*PageAddress*/,
1300 CONFIG_PAGE_HEADER *,
1301 int /*sleep_ok*/, int /*timeout_ms*/);
1302 int mpt_read_cfg_page(struct mpt_softc *t, int /*Action*/,
1303 uint32_t /*PageAddress*/,
1304 CONFIG_PAGE_HEADER *, size_t /*len*/,
1305 int /*sleep_ok*/, int /*timeout_ms*/);
1306 int mpt_write_cfg_page(struct mpt_softc *, int /*Action*/,
1307 uint32_t /*PageAddress*/,
1308 CONFIG_PAGE_HEADER *, size_t /*len*/,
1309 int /*sleep_ok*/, int /*timeout_ms*/);
1311 mpt_read_cur_cfg_page(struct mpt_softc *mpt, uint32_t PageAddress,
1312 CONFIG_PAGE_HEADER *hdr, size_t len,
1313 int sleep_ok, int timeout_ms)
1315 return (mpt_read_cfg_page(mpt, MPI_CONFIG_ACTION_PAGE_READ_CURRENT,
1316 PageAddress, hdr, len, sleep_ok, timeout_ms));
1320 mpt_write_cur_cfg_page(struct mpt_softc *mpt, uint32_t PageAddress,
1321 CONFIG_PAGE_HEADER *hdr, size_t len, int sleep_ok,
1324 return (mpt_write_cfg_page(mpt, MPI_CONFIG_ACTION_PAGE_WRITE_CURRENT,
1325 PageAddress, hdr, len, sleep_ok,
1328 /* mpt_debug.c functions */
1329 void mpt_print_reply(void *vmsg);
1330 void mpt_print_db(uint32_t mb);
1331 void mpt_print_config_reply(void *vmsg);
1332 char *mpt_ioc_diag(uint32_t diag);
1333 void mpt_req_state(mpt_req_state_t state);
1334 void mpt_print_config_request(void *vmsg);
1335 void mpt_print_request(void *vmsg);
1336 void mpt_print_scsi_io_request(MSG_SCSI_IO_REQUEST *msg);
1337 void mpt_dump_sgl(SGE_IO_UNION *se, int offset);
1338 #endif /* _MPT_H_ */