1 /**************************************************************************
2 SPDX-License-Identifier: BSD-2-Clause-FreeBSD
4 Copyright (c) 2007-2009, Chelsio Inc.
7 Redistribution and use in source and binary forms, with or without
8 modification, are permitted provided that the following conditions are met:
10 1. Redistributions of source code must retain the above copyright notice,
11 this list of conditions and the following disclaimer.
13 2. Neither the name of the Chelsio Corporation nor the names of its
14 contributors may be used to endorse or promote products derived from
15 this software without specific prior written permission.
17 THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
18 AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
19 IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
20 ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE
21 LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
22 CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
23 SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
24 INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
25 CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
26 ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
27 POSSIBILITY OF SUCH DAMAGE.
31 ***************************************************************************/
34 #ifndef _CXGB_ADAPTER_H_
35 #define _CXGB_ADAPTER_H_
38 #include <sys/mutex.h>
41 #include <sys/socket.h>
42 #include <sys/sockio.h>
43 #include <sys/condvar.h>
44 #include <sys/buf_ring.h>
45 #include <sys/taskqueue.h>
47 #include <net/ethernet.h>
49 #include <net/if_var.h>
50 #include <net/if_media.h>
51 #include <net/if_dl.h>
52 #include <netinet/in.h>
53 #include <netinet/tcp_lro.h>
55 #include <machine/bus.h>
56 #include <machine/resource.h>
58 #include <dev/pci/pcireg.h>
59 #include <dev/pci/pcivar.h>
61 #include <cxgb_osdep.h>
65 extern int cxgb_debug;
68 #define MTX_INIT(lock, lockname, class, flags) \
70 printf("initializing %s at %s:%d\n", lockname, __FILE__, __LINE__); \
71 mtx_init((lock), lockname, class, flags); \
74 #define MTX_DESTROY(lock) \
76 printf("destroying %s at %s:%d\n", (lock)->lock_object.lo_name, __FILE__, __LINE__); \
77 mtx_destroy((lock)); \
81 #define MTX_INIT mtx_init
82 #define MTX_DESTROY mtx_destroy
92 struct adapter *adapter;
96 const struct port_type_info *port_type;
99 struct timeval last_refreshed;
100 struct link_config link_config;
101 struct ifmedia media;
110 uint8_t hw_addr[ETHER_ADDR_LEN];
111 struct callout link_check_ch;
112 struct task link_check_task;
113 struct task timer_reclaim_task;
114 struct cdev *port_cdev;
116 #define PORT_LOCK_NAME_LEN 32
117 #define PORT_NAME_LEN 32
118 char lockbuf[PORT_LOCK_NAME_LEN];
119 char namebuf[PORT_NAME_LEN];
120 } __aligned(L1_CACHE_BYTES);
124 FULL_INIT_DONE = (1 << 0),
125 USING_MSI = (1 << 1),
126 USING_MSIX = (1 << 2),
127 QUEUES_BOUND = (1 << 3),
128 FW_UPTODATE = (1 << 4),
129 TPS_UPTODATE = (1 << 5),
130 CXGB_SHUTDOWN = (1 << 6),
131 CXGB_OFLD_INIT = (1 << 7),
132 TP_PARITY_INIT = (1 << 8),
133 CXGB_BUSY = (1 << 9),
134 TOM_INIT_DONE = (1 << 10),
139 #define IS_DOOMED(p) (p->flags & DOOMED)
140 #define SET_DOOMED(p) do {p->flags |= DOOMED;} while (0)
141 #define IS_BUSY(sc) (sc->flags & CXGB_BUSY)
142 #define SET_BUSY(sc) do {sc->flags |= CXGB_BUSY;} while (0)
143 #define CLR_BUSY(sc) do {sc->flags &= ~CXGB_BUSY;} while (0)
145 #define FL_Q_SIZE 4096
146 #define JUMBO_Q_SIZE 1024
147 #define RSPQ_Q_SIZE 2048
148 #define TX_ETH_Q_SIZE 1024
149 #define TX_OFLD_Q_SIZE 1024
150 #define TX_CTRL_Q_SIZE 256
158 * work request size in bytes
160 #define WR_LEN (WR_FLITS * 8)
161 #define PIO_LEN (WR_LEN - sizeof(struct cpl_tx_pkt_lso))
164 unsigned short enabled;
165 struct lro_ctrl ctrl;
168 #define RX_BUNDLE_SIZE 8
178 uint32_t holdoff_tmr;
179 uint32_t next_holdoff;
181 uint32_t async_notif;
183 uint32_t offload_pkts;
185 uint32_t unhandled_irqs;
188 bus_addr_t phys_addr;
189 bus_dma_tag_t desc_tag;
190 bus_dmamap_t desc_map;
192 struct t3_mbuf_hdr rspq_mh;
193 struct rsp_desc *desc;
195 #define RSPQ_NAME_LEN 32
196 char lockbuf[RSPQ_NAME_LEN];
197 uint32_t rspq_dump_start;
198 uint32_t rspq_dump_count;
212 bus_addr_t phys_addr;
215 bus_dma_tag_t desc_tag;
216 bus_dmamap_t desc_map;
217 bus_dma_tag_t entry_tag;
219 struct rx_desc *desc;
220 struct rx_sw_desc *sdesc;
227 #define TXQ_TRANSMITTING 0x1
241 struct tx_desc *desc;
242 struct tx_sw_desc *sdesc;
244 bus_addr_t phys_addr;
245 struct task qresume_task;
246 struct task qreclaim_task;
250 bus_dma_tag_t desc_tag;
251 bus_dmamap_t desc_map;
252 bus_dma_tag_t entry_tag;
255 struct buf_ring *txq_mr;
256 struct ifaltq *txq_ifq;
257 struct callout txq_timer;
258 struct callout txq_watchdog;
259 uint64_t txq_coalesced;
260 uint32_t txq_skipped;
261 uint32_t txq_enqueued;
262 uint32_t txq_dump_start;
263 uint32_t txq_dump_count;
264 uint64_t txq_direct_packets;
265 uint64_t txq_direct_bytes;
267 struct sg_ent txq_sgl[TX_MAX_SEGS / 2 + 1];
270 #define SGE_PSTAT_MAX (SGE_PSTAT_VLANINS+1)
272 #define QS_EXITING 0x1
273 #define QS_RUNNING 0x2
275 #define QS_FLUSHING 0x8
276 #define QS_TIMEOUT 0x10
279 struct sge_rspq rspq;
280 struct sge_fl fl[SGE_RXQ_PER_SET];
281 struct lro_state lro;
282 struct sge_txq txq[SGE_TXQ_PER_SET];
283 uint32_t txq_stopped; /* which Tx queues are stopped */
284 struct port_info *port;
285 struct adapter *adap;
286 int idx; /* qset # */
291 #define QS_NAME_LEN 32
292 char namebuf[QS_NAME_LEN];
296 struct sge_qset qs[SGE_QSETS];
302 typedef int (*cpl_handler_t)(struct sge_qset *, struct rsp_desc *,
306 SLIST_ENTRY(adapter) link;
310 /* PCI register resources */
312 struct resource *regs_res;
314 struct resource *udbs_res;
315 bus_space_handle_t bh;
321 bus_dma_tag_t parent_dmat;
322 bus_dma_tag_t rx_dmat;
323 bus_dma_tag_t rx_jumbo_dmat;
324 bus_dma_tag_t tx_dmat;
326 /* Interrupt resources */
327 struct resource *irq_res;
331 uint32_t msix_regs_rid;
332 struct resource *msix_regs_res;
334 struct resource *msix_irq_res[SGE_QSETS];
335 int msix_irq_rid[SGE_QSETS];
336 void *msix_intr_tag[SGE_QSETS];
337 uint8_t rxpkt_map[8]; /* maps RX_PKT interface values to port ids */
338 uint8_t rrss_map[SGE_QSETS]; /* revers RSS map table */
339 uint16_t rspq_map[RSS_TABLE_SIZE]; /* maps 7-bit cookie to qidx */
341 uint8_t fill[SGE_QSETS];
345 #define tunq_fill u.fill
346 #define tunq_coalesce u.coalesce
348 struct filter_info *filters;
351 struct task slow_intr_task;
352 struct task tick_task;
353 struct taskqueue *tq;
354 struct callout cxgb_tick_ch;
355 struct callout sge_timer_ch;
357 /* Register lock for use by the hardware layer */
358 struct mtx mdio_lock;
359 struct mtx elmer_lock;
361 /* Bookkeeping for the hardware layer */
362 struct adapter_params params;
363 unsigned int slow_intr_mask;
364 unsigned long irq_stats[IRQ_NUM_STATS];
373 struct port_info port[MAX_NPORTS];
374 device_t portdev[MAX_NPORTS];
380 char port_types[MAX_NPORTS + 1];
381 uint32_t open_device_map;
386 driver_intr_t *cxgb_intr;
389 #define ADAPTER_LOCK_NAME_LEN 32
390 char lockbuf[ADAPTER_LOCK_NAME_LEN];
391 char reglockbuf[ADAPTER_LOCK_NAME_LEN];
392 char mdiolockbuf[ADAPTER_LOCK_NAME_LEN];
393 char elmerlockbuf[ADAPTER_LOCK_NAME_LEN];
398 #define NUM_CPL_HANDLERS 0xa7
399 cpl_handler_t cpl_handler[NUM_CPL_HANDLERS] __aligned(CACHE_LINE_SIZE);
406 struct port_info *port;
409 #define MDIO_LOCK(adapter) mtx_lock(&(adapter)->mdio_lock)
410 #define MDIO_UNLOCK(adapter) mtx_unlock(&(adapter)->mdio_lock)
411 #define ELMR_LOCK(adapter) mtx_lock(&(adapter)->elmer_lock)
412 #define ELMR_UNLOCK(adapter) mtx_unlock(&(adapter)->elmer_lock)
415 #define PORT_LOCK(port) mtx_lock(&(port)->lock);
416 #define PORT_UNLOCK(port) mtx_unlock(&(port)->lock);
417 #define PORT_LOCK_INIT(port, name) mtx_init(&(port)->lock, name, 0, MTX_DEF)
418 #define PORT_LOCK_DEINIT(port) mtx_destroy(&(port)->lock)
419 #define PORT_LOCK_ASSERT_NOTOWNED(port) mtx_assert(&(port)->lock, MA_NOTOWNED)
420 #define PORT_LOCK_ASSERT_OWNED(port) mtx_assert(&(port)->lock, MA_OWNED)
422 #define ADAPTER_LOCK(adap) mtx_lock(&(adap)->lock);
423 #define ADAPTER_UNLOCK(adap) mtx_unlock(&(adap)->lock);
424 #define ADAPTER_LOCK_INIT(adap, name) mtx_init(&(adap)->lock, name, 0, MTX_DEF)
425 #define ADAPTER_LOCK_DEINIT(adap) mtx_destroy(&(adap)->lock)
426 #define ADAPTER_LOCK_ASSERT_NOTOWNED(adap) mtx_assert(&(adap)->lock, MA_NOTOWNED)
427 #define ADAPTER_LOCK_ASSERT_OWNED(adap) mtx_assert(&(adap)->lock, MA_OWNED)
430 static __inline uint32_t
431 t3_read_reg(adapter_t *adapter, uint32_t reg_addr)
433 return (bus_space_read_4(adapter->bt, adapter->bh, reg_addr));
437 t3_write_reg(adapter_t *adapter, uint32_t reg_addr, uint32_t val)
439 bus_space_write_4(adapter->bt, adapter->bh, reg_addr, val);
443 t3_os_pci_read_config_4(adapter_t *adapter, int reg, uint32_t *val)
445 *val = pci_read_config(adapter->dev, reg, 4);
449 t3_os_pci_write_config_4(adapter_t *adapter, int reg, uint32_t val)
451 pci_write_config(adapter->dev, reg, val, 4);
455 t3_os_pci_read_config_2(adapter_t *adapter, int reg, uint16_t *val)
457 *val = pci_read_config(adapter->dev, reg, 2);
461 t3_os_pci_write_config_2(adapter_t *adapter, int reg, uint16_t val)
463 pci_write_config(adapter->dev, reg, val, 2);
466 static __inline uint8_t *
467 t3_get_next_mcaddr(struct t3_rx_mode *rm)
469 uint8_t *macaddr = NULL;
470 struct ifnet *ifp = rm->port->ifp;
471 struct ifmultiaddr *ifma;
475 CK_STAILQ_FOREACH(ifma, &ifp->if_multiaddrs, ifma_link) {
476 if (ifma->ifma_addr->sa_family != AF_LINK)
479 macaddr = LLADDR((struct sockaddr_dl *)ifma->ifma_addr);
484 if_maddr_runlock(ifp);
491 t3_init_rx_mode(struct t3_rx_mode *rm, struct port_info *port)
497 static __inline struct port_info *
498 adap2pinfo(struct adapter *adap, int idx)
500 return &adap->port[idx];
503 int t3_os_find_pci_capability(adapter_t *adapter, int cap);
504 int t3_os_pci_save_state(struct adapter *adapter);
505 int t3_os_pci_restore_state(struct adapter *adapter);
506 void t3_os_link_intr(struct port_info *);
507 void t3_os_link_changed(adapter_t *adapter, int port_id, int link_status,
508 int speed, int duplex, int fc, int mac_was_reset);
509 void t3_os_phymod_changed(struct adapter *adap, int port_id);
510 void t3_sge_err_intr_handler(adapter_t *adapter);
512 int t3_offload_tx(struct adapter *, struct mbuf *);
514 void t3_os_set_hw_addr(adapter_t *adapter, int port_idx, u8 hw_addr[]);
515 int t3_mgmt_tx(adapter_t *adap, struct mbuf *m);
516 int t3_register_cpl_handler(struct adapter *, int, cpl_handler_t);
518 int t3_sge_alloc(struct adapter *);
519 int t3_sge_free(struct adapter *);
520 int t3_sge_alloc_qset(adapter_t *, uint32_t, int, int, const struct qset_params *,
521 int, struct port_info *);
522 void t3_free_sge_resources(adapter_t *, int);
523 void t3_sge_start(adapter_t *);
524 void t3_sge_stop(adapter_t *);
525 void t3b_intr(void *data);
526 void t3_intr_msi(void *data);
527 void t3_intr_msix(void *data);
529 int t3_sge_init_adapter(adapter_t *);
530 int t3_sge_reset_adapter(adapter_t *);
531 int t3_sge_init_port(struct port_info *);
532 void t3_free_tx_desc(struct sge_qset *qs, int n, int qid);
534 void t3_rx_eth(struct adapter *adap, struct mbuf *m, int ethpad);
536 void t3_add_attach_sysctls(adapter_t *sc);
537 void t3_add_configured_sysctls(adapter_t *sc);
538 int t3_get_desc(const struct sge_qset *qs, unsigned int qnum, unsigned int idx,
539 unsigned char *data);
540 void t3_update_qset_coalesce(struct sge_qset *qs, const struct qset_params *p);
543 * XXX figure out how we can return this to being private to sge
545 #define desc_reclaimable(q) ((int)((q)->processed - (q)->cleaned - TX_MAX_DESC))
547 #define container_of(p, stype, field) ((stype *)(((uint8_t *)(p)) - offsetof(stype, field)))
549 static __inline struct sge_qset *
550 fl_to_qset(struct sge_fl *q, int qidx)
552 return container_of(q, struct sge_qset, fl[qidx]);
555 static __inline struct sge_qset *
556 rspq_to_qset(struct sge_rspq *q)
558 return container_of(q, struct sge_qset, rspq);
561 static __inline struct sge_qset *
562 txq_to_qset(struct sge_txq *q, int qidx)
564 return container_of(q, struct sge_qset, txq[qidx]);
569 #define OFFLOAD_DEVMAP_BIT (1 << MAX_NPORTS)
570 static inline int offload_running(adapter_t *adapter)
572 return isset(&adapter->open_device_map, OFFLOAD_DEVMAP_BIT);
575 void cxgb_tx_watchdog(void *arg);
576 int cxgb_transmit(struct ifnet *ifp, struct mbuf *m);
577 void cxgb_qflush(struct ifnet *ifp);
578 void t3_iterate(void (*)(struct adapter *, void *), void *);
579 void cxgb_refresh_stats(struct port_info *);
582 int cxgb_netdump_encap(struct sge_qset *qs, struct mbuf **m);
583 int cxgb_netdump_poll_rx(adapter_t *adap, struct sge_qset *qs);
584 int cxgb_netdump_poll_tx(struct sge_qset *qs);