2 * Copyright (c) 2007-2009 Sam Leffler, Errno Consulting
3 * Copyright (c) 2007-2008 Marvell Semiconductor, Inc.
6 * Redistribution and use in source and binary forms, with or without
7 * modification, are permitted provided that the following conditions
9 * 1. Redistributions of source code must retain the above copyright
10 * notice, this list of conditions and the following disclaimer,
11 * without modification.
12 * 2. Redistributions in binary form must reproduce at minimum a disclaimer
13 * similar to the "NO WARRANTY" disclaimer below ("Disclaimer") and any
14 * redistribution must be conditioned upon including a substantially
15 * similar Disclaimer requirement for further binary redistribution.
18 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
19 * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
20 * LIMITED TO, THE IMPLIED WARRANTIES OF NONINFRINGEMENT, MERCHANTIBILITY
21 * AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL
22 * THE COPYRIGHT HOLDERS OR CONTRIBUTORS BE LIABLE FOR SPECIAL, EXEMPLARY,
23 * OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
24 * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
25 * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER
26 * IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
27 * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF
28 * THE POSSIBILITY OF SUCH DAMAGES.
31 #include <sys/cdefs.h>
32 __FBSDID("$FreeBSD$");
35 * Driver for the Marvell 88W8363 Wireless LAN controller.
42 #include <sys/param.h>
43 #include <sys/systm.h>
44 #include <sys/sysctl.h>
46 #include <sys/malloc.h>
48 #include <sys/mutex.h>
49 #include <sys/kernel.h>
50 #include <sys/socket.h>
51 #include <sys/sockio.h>
52 #include <sys/errno.h>
53 #include <sys/callout.h>
55 #include <sys/endian.h>
56 #include <sys/kthread.h>
57 #include <sys/taskqueue.h>
59 #include <machine/bus.h>
62 #include <net/if_var.h>
63 #include <net/if_dl.h>
64 #include <net/if_media.h>
65 #include <net/if_types.h>
66 #include <net/if_arp.h>
67 #include <net/ethernet.h>
68 #include <net/if_llc.h>
72 #include <net80211/ieee80211_var.h>
73 #include <net80211/ieee80211_input.h>
74 #include <net80211/ieee80211_regdomain.h>
77 #include <netinet/in.h>
78 #include <netinet/if_ether.h>
81 #include <dev/mwl/if_mwlvar.h>
82 #include <dev/mwl/mwldiag.h>
84 /* idiomatic shorthands: MS = mask+shift, SM = shift+mask */
85 #define MS(v,x) (((v) & x) >> x##_S)
86 #define SM(v,x) (((v) << x##_S) & x)
88 static struct ieee80211vap *mwl_vap_create(struct ieee80211com *,
89 const char [IFNAMSIZ], int, enum ieee80211_opmode, int,
90 const uint8_t [IEEE80211_ADDR_LEN],
91 const uint8_t [IEEE80211_ADDR_LEN]);
92 static void mwl_vap_delete(struct ieee80211vap *);
93 static int mwl_setupdma(struct mwl_softc *);
94 static int mwl_hal_reset(struct mwl_softc *sc);
95 static int mwl_init(struct mwl_softc *);
96 static void mwl_parent(struct ieee80211com *);
97 static int mwl_reset(struct ieee80211vap *, u_long);
98 static void mwl_stop(struct mwl_softc *);
99 static void mwl_start(struct mwl_softc *);
100 static int mwl_transmit(struct ieee80211com *, struct mbuf *);
101 static int mwl_raw_xmit(struct ieee80211_node *, struct mbuf *,
102 const struct ieee80211_bpf_params *);
103 static int mwl_media_change(struct ifnet *);
104 static void mwl_watchdog(void *);
105 static int mwl_ioctl(struct ieee80211com *, u_long, void *);
106 static void mwl_radar_proc(void *, int);
107 static void mwl_chanswitch_proc(void *, int);
108 static void mwl_bawatchdog_proc(void *, int);
109 static int mwl_key_alloc(struct ieee80211vap *,
110 struct ieee80211_key *,
111 ieee80211_keyix *, ieee80211_keyix *);
112 static int mwl_key_delete(struct ieee80211vap *,
113 const struct ieee80211_key *);
114 static int mwl_key_set(struct ieee80211vap *,
115 const struct ieee80211_key *);
116 static int _mwl_key_set(struct ieee80211vap *,
117 const struct ieee80211_key *,
118 const uint8_t mac[IEEE80211_ADDR_LEN]);
119 static int mwl_mode_init(struct mwl_softc *);
120 static void mwl_update_mcast(struct ieee80211com *);
121 static void mwl_update_promisc(struct ieee80211com *);
122 static void mwl_updateslot(struct ieee80211com *);
123 static int mwl_beacon_setup(struct ieee80211vap *);
124 static void mwl_beacon_update(struct ieee80211vap *, int);
125 #ifdef MWL_HOST_PS_SUPPORT
126 static void mwl_update_ps(struct ieee80211vap *, int);
127 static int mwl_set_tim(struct ieee80211_node *, int);
129 static int mwl_dma_setup(struct mwl_softc *);
130 static void mwl_dma_cleanup(struct mwl_softc *);
131 static struct ieee80211_node *mwl_node_alloc(struct ieee80211vap *,
132 const uint8_t [IEEE80211_ADDR_LEN]);
133 static void mwl_node_cleanup(struct ieee80211_node *);
134 static void mwl_node_drain(struct ieee80211_node *);
135 static void mwl_node_getsignal(const struct ieee80211_node *,
137 static void mwl_node_getmimoinfo(const struct ieee80211_node *,
138 struct ieee80211_mimo_info *);
139 static int mwl_rxbuf_init(struct mwl_softc *, struct mwl_rxbuf *);
140 static void mwl_rx_proc(void *, int);
141 static void mwl_txq_init(struct mwl_softc *sc, struct mwl_txq *, int);
142 static int mwl_tx_setup(struct mwl_softc *, int, int);
143 static int mwl_wme_update(struct ieee80211com *);
144 static void mwl_tx_cleanupq(struct mwl_softc *, struct mwl_txq *);
145 static void mwl_tx_cleanup(struct mwl_softc *);
146 static uint16_t mwl_calcformat(uint8_t rate, const struct ieee80211_node *);
147 static int mwl_tx_start(struct mwl_softc *, struct ieee80211_node *,
148 struct mwl_txbuf *, struct mbuf *);
149 static void mwl_tx_proc(void *, int);
150 static int mwl_chan_set(struct mwl_softc *, struct ieee80211_channel *);
151 static void mwl_draintxq(struct mwl_softc *);
152 static void mwl_cleartxq(struct mwl_softc *, struct ieee80211vap *);
153 static int mwl_recv_action(struct ieee80211_node *,
154 const struct ieee80211_frame *,
155 const uint8_t *, const uint8_t *);
156 static int mwl_addba_request(struct ieee80211_node *,
157 struct ieee80211_tx_ampdu *, int dialogtoken,
158 int baparamset, int batimeout);
159 static int mwl_addba_response(struct ieee80211_node *,
160 struct ieee80211_tx_ampdu *, int status,
161 int baparamset, int batimeout);
162 static void mwl_addba_stop(struct ieee80211_node *,
163 struct ieee80211_tx_ampdu *);
164 static int mwl_startrecv(struct mwl_softc *);
165 static MWL_HAL_APMODE mwl_getapmode(const struct ieee80211vap *,
166 struct ieee80211_channel *);
167 static int mwl_setapmode(struct ieee80211vap *, struct ieee80211_channel*);
168 static void mwl_scan_start(struct ieee80211com *);
169 static void mwl_scan_end(struct ieee80211com *);
170 static void mwl_set_channel(struct ieee80211com *);
171 static int mwl_peerstadb(struct ieee80211_node *,
172 int aid, int staid, MWL_HAL_PEERINFO *pi);
173 static int mwl_localstadb(struct ieee80211vap *);
174 static int mwl_newstate(struct ieee80211vap *, enum ieee80211_state, int);
175 static int allocstaid(struct mwl_softc *sc, int aid);
176 static void delstaid(struct mwl_softc *sc, int staid);
177 static void mwl_newassoc(struct ieee80211_node *, int);
178 static void mwl_agestations(void *);
179 static int mwl_setregdomain(struct ieee80211com *,
180 struct ieee80211_regdomain *, int,
181 struct ieee80211_channel []);
182 static void mwl_getradiocaps(struct ieee80211com *, int, int *,
183 struct ieee80211_channel []);
184 static int mwl_getchannels(struct mwl_softc *);
186 static void mwl_sysctlattach(struct mwl_softc *);
187 static void mwl_announce(struct mwl_softc *);
189 SYSCTL_NODE(_hw, OID_AUTO, mwl, CTLFLAG_RD, 0, "Marvell driver parameters");
191 static int mwl_rxdesc = MWL_RXDESC; /* # rx desc's to allocate */
192 SYSCTL_INT(_hw_mwl, OID_AUTO, rxdesc, CTLFLAG_RW, &mwl_rxdesc,
193 0, "rx descriptors allocated");
194 static int mwl_rxbuf = MWL_RXBUF; /* # rx buffers to allocate */
195 SYSCTL_INT(_hw_mwl, OID_AUTO, rxbuf, CTLFLAG_RWTUN, &mwl_rxbuf,
196 0, "rx buffers allocated");
197 static int mwl_txbuf = MWL_TXBUF; /* # tx buffers to allocate */
198 SYSCTL_INT(_hw_mwl, OID_AUTO, txbuf, CTLFLAG_RWTUN, &mwl_txbuf,
199 0, "tx buffers allocated");
200 static int mwl_txcoalesce = 8; /* # tx packets to q before poking f/w*/
201 SYSCTL_INT(_hw_mwl, OID_AUTO, txcoalesce, CTLFLAG_RWTUN, &mwl_txcoalesce,
202 0, "tx buffers to send at once");
203 static int mwl_rxquota = MWL_RXBUF; /* # max buffers to process */
204 SYSCTL_INT(_hw_mwl, OID_AUTO, rxquota, CTLFLAG_RWTUN, &mwl_rxquota,
205 0, "max rx buffers to process per interrupt");
206 static int mwl_rxdmalow = 3; /* # min buffers for wakeup */
207 SYSCTL_INT(_hw_mwl, OID_AUTO, rxdmalow, CTLFLAG_RWTUN, &mwl_rxdmalow,
208 0, "min free rx buffers before restarting traffic");
211 static int mwl_debug = 0;
212 SYSCTL_INT(_hw_mwl, OID_AUTO, debug, CTLFLAG_RWTUN, &mwl_debug,
213 0, "control debugging printfs");
215 MWL_DEBUG_XMIT = 0x00000001, /* basic xmit operation */
216 MWL_DEBUG_XMIT_DESC = 0x00000002, /* xmit descriptors */
217 MWL_DEBUG_RECV = 0x00000004, /* basic recv operation */
218 MWL_DEBUG_RECV_DESC = 0x00000008, /* recv descriptors */
219 MWL_DEBUG_RESET = 0x00000010, /* reset processing */
220 MWL_DEBUG_BEACON = 0x00000020, /* beacon handling */
221 MWL_DEBUG_INTR = 0x00000040, /* ISR */
222 MWL_DEBUG_TX_PROC = 0x00000080, /* tx ISR proc */
223 MWL_DEBUG_RX_PROC = 0x00000100, /* rx ISR proc */
224 MWL_DEBUG_KEYCACHE = 0x00000200, /* key cache management */
225 MWL_DEBUG_STATE = 0x00000400, /* 802.11 state transitions */
226 MWL_DEBUG_NODE = 0x00000800, /* node management */
227 MWL_DEBUG_RECV_ALL = 0x00001000, /* trace all frames (beacons) */
228 MWL_DEBUG_TSO = 0x00002000, /* TSO processing */
229 MWL_DEBUG_AMPDU = 0x00004000, /* BA stream handling */
230 MWL_DEBUG_ANY = 0xffffffff
232 #define IS_BEACON(wh) \
233 ((wh->i_fc[0] & (IEEE80211_FC0_TYPE_MASK|IEEE80211_FC0_SUBTYPE_MASK)) == \
234 (IEEE80211_FC0_TYPE_MGT|IEEE80211_FC0_SUBTYPE_BEACON))
235 #define IFF_DUMPPKTS_RECV(sc, wh) \
236 ((sc->sc_debug & MWL_DEBUG_RECV) && \
237 ((sc->sc_debug & MWL_DEBUG_RECV_ALL) || !IS_BEACON(wh)))
238 #define IFF_DUMPPKTS_XMIT(sc) \
239 (sc->sc_debug & MWL_DEBUG_XMIT)
241 #define DPRINTF(sc, m, fmt, ...) do { \
242 if (sc->sc_debug & (m)) \
243 printf(fmt, __VA_ARGS__); \
245 #define KEYPRINTF(sc, hk, mac) do { \
246 if (sc->sc_debug & MWL_DEBUG_KEYCACHE) \
247 mwl_keyprint(sc, __func__, hk, mac); \
249 static void mwl_printrxbuf(const struct mwl_rxbuf *bf, u_int ix);
250 static void mwl_printtxbuf(const struct mwl_txbuf *bf, u_int qnum, u_int ix);
252 #define IFF_DUMPPKTS_RECV(sc, wh) 0
253 #define IFF_DUMPPKTS_XMIT(sc) 0
254 #define DPRINTF(sc, m, fmt, ...) do { (void )sc; } while (0)
255 #define KEYPRINTF(sc, k, mac) do { (void )sc; } while (0)
258 static MALLOC_DEFINE(M_MWLDEV, "mwldev", "mwl driver dma buffers");
261 * Each packet has fixed front matter: a 2-byte length
262 * of the payload, followed by a 4-address 802.11 header
263 * (regardless of the actual header and always w/o any
264 * QoS header). The payload then follows.
268 struct ieee80211_frame_addr4 wh;
272 * Read/Write shorthands for accesses to BAR 0. Note
273 * that all BAR 1 operations are done in the "hal" and
274 * there should be no reference to them here.
277 static __inline uint32_t
278 RD4(struct mwl_softc *sc, bus_size_t off)
280 return bus_space_read_4(sc->sc_io0t, sc->sc_io0h, off);
285 WR4(struct mwl_softc *sc, bus_size_t off, uint32_t val)
287 bus_space_write_4(sc->sc_io0t, sc->sc_io0h, off, val);
291 mwl_attach(uint16_t devid, struct mwl_softc *sc)
293 struct ieee80211com *ic = &sc->sc_ic;
297 DPRINTF(sc, MWL_DEBUG_ANY, "%s: devid 0x%x\n", __func__, devid);
300 * Setup the RX free list lock early, so it can be consistently
305 mh = mwl_hal_attach(sc->sc_dev, devid,
306 sc->sc_io1h, sc->sc_io1t, sc->sc_dmat);
308 device_printf(sc->sc_dev, "unable to attach HAL\n");
314 * Load firmware so we can get setup. We arbitrarily
315 * pick station firmware; we'll re-load firmware as
316 * needed so setting up the wrong mode isn't a big deal.
318 if (mwl_hal_fwload(mh, NULL) != 0) {
319 device_printf(sc->sc_dev, "unable to setup builtin firmware\n");
323 if (mwl_hal_gethwspecs(mh, &sc->sc_hwspecs) != 0) {
324 device_printf(sc->sc_dev, "unable to fetch h/w specs\n");
328 error = mwl_getchannels(sc);
332 sc->sc_txantenna = 0; /* h/w default */
333 sc->sc_rxantenna = 0; /* h/w default */
334 sc->sc_invalid = 0; /* ready to go, enable int handling */
335 sc->sc_ageinterval = MWL_AGEINTERVAL;
338 * Allocate tx+rx descriptors and populate the lists.
339 * We immediately push the information to the firmware
340 * as otherwise it gets upset.
342 error = mwl_dma_setup(sc);
344 device_printf(sc->sc_dev, "failed to setup descriptors: %d\n",
348 error = mwl_setupdma(sc); /* push to firmware */
349 if (error != 0) /* NB: mwl_setupdma prints msg */
352 callout_init(&sc->sc_timer, 1);
353 callout_init_mtx(&sc->sc_watchdog, &sc->sc_mtx, 0);
354 mbufq_init(&sc->sc_snd, ifqmaxlen);
356 sc->sc_tq = taskqueue_create("mwl_taskq", M_NOWAIT,
357 taskqueue_thread_enqueue, &sc->sc_tq);
358 taskqueue_start_threads(&sc->sc_tq, 1, PI_NET,
359 "%s taskq", device_get_nameunit(sc->sc_dev));
361 TASK_INIT(&sc->sc_rxtask, 0, mwl_rx_proc, sc);
362 TASK_INIT(&sc->sc_radartask, 0, mwl_radar_proc, sc);
363 TASK_INIT(&sc->sc_chanswitchtask, 0, mwl_chanswitch_proc, sc);
364 TASK_INIT(&sc->sc_bawatchdogtask, 0, mwl_bawatchdog_proc, sc);
366 /* NB: insure BK queue is the lowest priority h/w queue */
367 if (!mwl_tx_setup(sc, WME_AC_BK, MWL_WME_AC_BK)) {
368 device_printf(sc->sc_dev,
369 "unable to setup xmit queue for %s traffic!\n",
370 ieee80211_wme_acnames[WME_AC_BK]);
374 if (!mwl_tx_setup(sc, WME_AC_BE, MWL_WME_AC_BE) ||
375 !mwl_tx_setup(sc, WME_AC_VI, MWL_WME_AC_VI) ||
376 !mwl_tx_setup(sc, WME_AC_VO, MWL_WME_AC_VO)) {
378 * Not enough hardware tx queues to properly do WME;
379 * just punt and assign them all to the same h/w queue.
380 * We could do a better job of this if, for example,
381 * we allocate queues when we switch from station to
384 if (sc->sc_ac2q[WME_AC_VI] != NULL)
385 mwl_tx_cleanupq(sc, sc->sc_ac2q[WME_AC_VI]);
386 if (sc->sc_ac2q[WME_AC_BE] != NULL)
387 mwl_tx_cleanupq(sc, sc->sc_ac2q[WME_AC_BE]);
388 sc->sc_ac2q[WME_AC_BE] = sc->sc_ac2q[WME_AC_BK];
389 sc->sc_ac2q[WME_AC_VI] = sc->sc_ac2q[WME_AC_BK];
390 sc->sc_ac2q[WME_AC_VO] = sc->sc_ac2q[WME_AC_BK];
392 TASK_INIT(&sc->sc_txtask, 0, mwl_tx_proc, sc);
395 ic->ic_name = device_get_nameunit(sc->sc_dev);
396 /* XXX not right but it's not used anywhere important */
397 ic->ic_phytype = IEEE80211_T_OFDM;
398 ic->ic_opmode = IEEE80211_M_STA;
400 IEEE80211_C_STA /* station mode supported */
401 | IEEE80211_C_HOSTAP /* hostap mode */
402 | IEEE80211_C_MONITOR /* monitor mode */
404 | IEEE80211_C_IBSS /* ibss, nee adhoc, mode */
405 | IEEE80211_C_AHDEMO /* adhoc demo mode */
407 | IEEE80211_C_MBSS /* mesh point link mode */
408 | IEEE80211_C_WDS /* WDS supported */
409 | IEEE80211_C_SHPREAMBLE /* short preamble supported */
410 | IEEE80211_C_SHSLOT /* short slot time supported */
411 | IEEE80211_C_WME /* WME/WMM supported */
412 | IEEE80211_C_BURST /* xmit bursting supported */
413 | IEEE80211_C_WPA /* capable of WPA1+WPA2 */
414 | IEEE80211_C_BGSCAN /* capable of bg scanning */
415 | IEEE80211_C_TXFRAG /* handle tx frags */
416 | IEEE80211_C_TXPMGT /* capable of txpow mgt */
417 | IEEE80211_C_DFS /* DFS supported */
421 IEEE80211_HTCAP_SMPS_ENA /* SM PS mode enabled */
422 | IEEE80211_HTCAP_CHWIDTH40 /* 40MHz channel width */
423 | IEEE80211_HTCAP_SHORTGI20 /* short GI in 20MHz */
424 | IEEE80211_HTCAP_SHORTGI40 /* short GI in 40MHz */
425 | IEEE80211_HTCAP_RXSTBC_2STREAM/* 1-2 spatial streams */
426 #if MWL_AGGR_SIZE == 7935
427 | IEEE80211_HTCAP_MAXAMSDU_7935 /* max A-MSDU length */
429 | IEEE80211_HTCAP_MAXAMSDU_3839 /* max A-MSDU length */
432 | IEEE80211_HTCAP_PSMP /* PSMP supported */
433 | IEEE80211_HTCAP_40INTOLERANT /* 40MHz intolerant */
435 /* s/w capabilities */
436 | IEEE80211_HTC_HT /* HT operation */
437 | IEEE80211_HTC_AMPDU /* tx A-MPDU */
438 | IEEE80211_HTC_AMSDU /* tx A-MSDU */
439 | IEEE80211_HTC_SMPS /* SMPS available */
443 * Mark h/w crypto support.
444 * XXX no way to query h/w support.
446 ic->ic_cryptocaps |= IEEE80211_CRYPTO_WEP
447 | IEEE80211_CRYPTO_AES_CCM
448 | IEEE80211_CRYPTO_TKIP
449 | IEEE80211_CRYPTO_TKIPMIC
452 * Transmit requires space in the packet for a special
453 * format transmit record and optional padding between
454 * this record and the payload. Ask the net80211 layer
455 * to arrange this when encapsulating packets so we can
456 * add it efficiently.
458 ic->ic_headroom = sizeof(struct mwltxrec) -
459 sizeof(struct ieee80211_frame);
461 IEEE80211_ADDR_COPY(ic->ic_macaddr, sc->sc_hwspecs.macAddr);
463 /* call MI attach routine. */
464 ieee80211_ifattach(ic);
465 ic->ic_setregdomain = mwl_setregdomain;
466 ic->ic_getradiocaps = mwl_getradiocaps;
467 /* override default methods */
468 ic->ic_raw_xmit = mwl_raw_xmit;
469 ic->ic_newassoc = mwl_newassoc;
470 ic->ic_updateslot = mwl_updateslot;
471 ic->ic_update_mcast = mwl_update_mcast;
472 ic->ic_update_promisc = mwl_update_promisc;
473 ic->ic_wme.wme_update = mwl_wme_update;
474 ic->ic_transmit = mwl_transmit;
475 ic->ic_ioctl = mwl_ioctl;
476 ic->ic_parent = mwl_parent;
478 ic->ic_node_alloc = mwl_node_alloc;
479 sc->sc_node_cleanup = ic->ic_node_cleanup;
480 ic->ic_node_cleanup = mwl_node_cleanup;
481 sc->sc_node_drain = ic->ic_node_drain;
482 ic->ic_node_drain = mwl_node_drain;
483 ic->ic_node_getsignal = mwl_node_getsignal;
484 ic->ic_node_getmimoinfo = mwl_node_getmimoinfo;
486 ic->ic_scan_start = mwl_scan_start;
487 ic->ic_scan_end = mwl_scan_end;
488 ic->ic_set_channel = mwl_set_channel;
490 sc->sc_recv_action = ic->ic_recv_action;
491 ic->ic_recv_action = mwl_recv_action;
492 sc->sc_addba_request = ic->ic_addba_request;
493 ic->ic_addba_request = mwl_addba_request;
494 sc->sc_addba_response = ic->ic_addba_response;
495 ic->ic_addba_response = mwl_addba_response;
496 sc->sc_addba_stop = ic->ic_addba_stop;
497 ic->ic_addba_stop = mwl_addba_stop;
499 ic->ic_vap_create = mwl_vap_create;
500 ic->ic_vap_delete = mwl_vap_delete;
502 ieee80211_radiotap_attach(ic,
503 &sc->sc_tx_th.wt_ihdr, sizeof(sc->sc_tx_th),
504 MWL_TX_RADIOTAP_PRESENT,
505 &sc->sc_rx_th.wr_ihdr, sizeof(sc->sc_rx_th),
506 MWL_RX_RADIOTAP_PRESENT);
508 * Setup dynamic sysctl's now that country code and
509 * regdomain are available from the hal.
511 mwl_sysctlattach(sc);
514 ieee80211_announce(ic);
522 MWL_RXFREE_DESTROY(sc);
528 mwl_detach(struct mwl_softc *sc)
530 struct ieee80211com *ic = &sc->sc_ic;
536 * NB: the order of these is important:
537 * o call the 802.11 layer before detaching the hal to
538 * insure callbacks into the driver to delete global
539 * key cache entries can be handled
540 * o reclaim the tx queue data structures after calling
541 * the 802.11 layer as we'll get called back to reclaim
542 * node state and potentially want to use them
543 * o to cleanup the tx queues the hal is called, so detach
545 * Other than that, it's straightforward...
547 ieee80211_ifdetach(ic);
548 callout_drain(&sc->sc_watchdog);
550 MWL_RXFREE_DESTROY(sc);
552 mwl_hal_detach(sc->sc_mh);
553 mbufq_drain(&sc->sc_snd);
559 * MAC address handling for multiple BSS on the same radio.
560 * The first vap uses the MAC address from the EEPROM. For
561 * subsequent vap's we set the U/L bit (bit 1) in the MAC
562 * address and use the next six bits as an index.
565 assign_address(struct mwl_softc *sc, uint8_t mac[IEEE80211_ADDR_LEN], int clone)
569 if (clone && mwl_hal_ismbsscapable(sc->sc_mh)) {
570 /* NB: we only do this if h/w supports multiple bssid */
571 for (i = 0; i < 32; i++)
572 if ((sc->sc_bssidmask & (1<<i)) == 0)
575 mac[0] |= (i << 2)|0x2;
578 sc->sc_bssidmask |= 1<<i;
584 reclaim_address(struct mwl_softc *sc, const uint8_t mac[IEEE80211_ADDR_LEN])
587 if (i != 0 || --sc->sc_nbssid0 == 0)
588 sc->sc_bssidmask &= ~(1<<i);
591 static struct ieee80211vap *
592 mwl_vap_create(struct ieee80211com *ic, const char name[IFNAMSIZ], int unit,
593 enum ieee80211_opmode opmode, int flags,
594 const uint8_t bssid[IEEE80211_ADDR_LEN],
595 const uint8_t mac0[IEEE80211_ADDR_LEN])
597 struct mwl_softc *sc = ic->ic_softc;
598 struct mwl_hal *mh = sc->sc_mh;
599 struct ieee80211vap *vap, *apvap;
600 struct mwl_hal_vap *hvap;
602 uint8_t mac[IEEE80211_ADDR_LEN];
604 IEEE80211_ADDR_COPY(mac, mac0);
606 case IEEE80211_M_HOSTAP:
607 case IEEE80211_M_MBSS:
608 if ((flags & IEEE80211_CLONE_MACADDR) == 0)
609 assign_address(sc, mac, flags & IEEE80211_CLONE_BSSID);
610 hvap = mwl_hal_newvap(mh, MWL_HAL_AP, mac);
612 if ((flags & IEEE80211_CLONE_MACADDR) == 0)
613 reclaim_address(sc, mac);
617 case IEEE80211_M_STA:
618 if ((flags & IEEE80211_CLONE_MACADDR) == 0)
619 assign_address(sc, mac, flags & IEEE80211_CLONE_BSSID);
620 hvap = mwl_hal_newvap(mh, MWL_HAL_STA, mac);
622 if ((flags & IEEE80211_CLONE_MACADDR) == 0)
623 reclaim_address(sc, mac);
626 /* no h/w beacon miss support; always use s/w */
627 flags |= IEEE80211_CLONE_NOBEACONS;
629 case IEEE80211_M_WDS:
630 hvap = NULL; /* NB: we use associated AP vap */
631 if (sc->sc_napvaps == 0)
632 return NULL; /* no existing AP vap */
634 case IEEE80211_M_MONITOR:
637 case IEEE80211_M_IBSS:
638 case IEEE80211_M_AHDEMO:
643 mvp = malloc(sizeof(struct mwl_vap), M_80211_VAP, M_WAITOK | M_ZERO);
645 if (opmode == IEEE80211_M_WDS) {
647 * WDS vaps must have an associated AP vap; find one.
650 TAILQ_FOREACH(apvap, &ic->ic_vaps, iv_next)
651 if (apvap->iv_opmode == IEEE80211_M_HOSTAP) {
652 mvp->mv_ap_hvap = MWL_VAP(apvap)->mv_hvap;
655 KASSERT(mvp->mv_ap_hvap != NULL, ("no ap vap"));
658 ieee80211_vap_setup(ic, vap, name, unit, opmode, flags, bssid);
659 /* override with driver methods */
660 mvp->mv_newstate = vap->iv_newstate;
661 vap->iv_newstate = mwl_newstate;
662 vap->iv_max_keyix = 0; /* XXX */
663 vap->iv_key_alloc = mwl_key_alloc;
664 vap->iv_key_delete = mwl_key_delete;
665 vap->iv_key_set = mwl_key_set;
666 #ifdef MWL_HOST_PS_SUPPORT
667 if (opmode == IEEE80211_M_HOSTAP || opmode == IEEE80211_M_MBSS) {
668 vap->iv_update_ps = mwl_update_ps;
669 mvp->mv_set_tim = vap->iv_set_tim;
670 vap->iv_set_tim = mwl_set_tim;
673 vap->iv_reset = mwl_reset;
674 vap->iv_update_beacon = mwl_beacon_update;
676 /* override max aid so sta's cannot assoc when we're out of sta id's */
677 vap->iv_max_aid = MWL_MAXSTAID;
678 /* override default A-MPDU rx parameters */
679 vap->iv_ampdu_rxmax = IEEE80211_HTCAP_MAXRXAMPDU_64K;
680 vap->iv_ampdu_density = IEEE80211_HTCAP_MPDUDENSITY_4;
683 ieee80211_vap_attach(vap, mwl_media_change, ieee80211_media_status,
686 switch (vap->iv_opmode) {
687 case IEEE80211_M_HOSTAP:
688 case IEEE80211_M_MBSS:
689 case IEEE80211_M_STA:
691 * Setup sta db entry for local address.
694 if (vap->iv_opmode == IEEE80211_M_HOSTAP ||
695 vap->iv_opmode == IEEE80211_M_MBSS)
700 case IEEE80211_M_WDS:
707 * Setup overall operating mode.
710 ic->ic_opmode = IEEE80211_M_HOSTAP;
711 else if (sc->sc_nstavaps)
712 ic->ic_opmode = IEEE80211_M_STA;
714 ic->ic_opmode = opmode;
720 mwl_vap_delete(struct ieee80211vap *vap)
722 struct mwl_vap *mvp = MWL_VAP(vap);
723 struct mwl_softc *sc = vap->iv_ic->ic_softc;
724 struct mwl_hal *mh = sc->sc_mh;
725 struct mwl_hal_vap *hvap = mvp->mv_hvap;
726 enum ieee80211_opmode opmode = vap->iv_opmode;
728 /* XXX disallow ap vap delete if WDS still present */
729 if (sc->sc_running) {
730 /* quiesce h/w while we remove the vap */
731 mwl_hal_intrset(mh, 0); /* disable interrupts */
733 ieee80211_vap_detach(vap);
735 case IEEE80211_M_HOSTAP:
736 case IEEE80211_M_MBSS:
737 case IEEE80211_M_STA:
738 KASSERT(hvap != NULL, ("no hal vap handle"));
739 (void) mwl_hal_delstation(hvap, vap->iv_myaddr);
740 mwl_hal_delvap(hvap);
741 if (opmode == IEEE80211_M_HOSTAP || opmode == IEEE80211_M_MBSS)
745 /* XXX don't do it for IEEE80211_CLONE_MACADDR */
746 reclaim_address(sc, vap->iv_myaddr);
748 case IEEE80211_M_WDS:
754 mwl_cleartxq(sc, vap);
755 free(mvp, M_80211_VAP);
757 mwl_hal_intrset(mh, sc->sc_imask);
761 mwl_suspend(struct mwl_softc *sc)
770 mwl_resume(struct mwl_softc *sc)
775 if (sc->sc_ic.ic_nrunning > 0)
776 error = mwl_init(sc);
780 ieee80211_start_all(&sc->sc_ic); /* start all vap's */
784 mwl_shutdown(void *arg)
786 struct mwl_softc *sc = arg;
794 * Interrupt handler. Most of the actual processing is deferred.
799 struct mwl_softc *sc = arg;
800 struct mwl_hal *mh = sc->sc_mh;
803 if (sc->sc_invalid) {
805 * The hardware is not ready/present, don't touch anything.
806 * Note this can happen early on if the IRQ is shared.
808 DPRINTF(sc, MWL_DEBUG_ANY, "%s: invalid; ignored\n", __func__);
812 * Figure out the reason(s) for the interrupt.
814 mwl_hal_getisr(mh, &status); /* NB: clears ISR too */
815 if (status == 0) /* must be a shared irq */
818 DPRINTF(sc, MWL_DEBUG_INTR, "%s: status 0x%x imask 0x%x\n",
819 __func__, status, sc->sc_imask);
820 if (status & MACREG_A2HRIC_BIT_RX_RDY)
821 taskqueue_enqueue(sc->sc_tq, &sc->sc_rxtask);
822 if (status & MACREG_A2HRIC_BIT_TX_DONE)
823 taskqueue_enqueue(sc->sc_tq, &sc->sc_txtask);
824 if (status & MACREG_A2HRIC_BIT_BA_WATCHDOG)
825 taskqueue_enqueue(sc->sc_tq, &sc->sc_bawatchdogtask);
826 if (status & MACREG_A2HRIC_BIT_OPC_DONE)
828 if (status & MACREG_A2HRIC_BIT_MAC_EVENT) {
831 if (status & MACREG_A2HRIC_BIT_ICV_ERROR) {
833 sc->sc_stats.mst_rx_badtkipicv++;
835 if (status & MACREG_A2HRIC_BIT_QUEUE_EMPTY) {
836 /* 11n aggregation queue is empty, re-fill */
839 if (status & MACREG_A2HRIC_BIT_QUEUE_FULL) {
842 if (status & MACREG_A2HRIC_BIT_RADAR_DETECT) {
843 /* radar detected, process event */
844 taskqueue_enqueue(sc->sc_tq, &sc->sc_radartask);
846 if (status & MACREG_A2HRIC_BIT_CHAN_SWITCH) {
847 /* DFS channel switch */
848 taskqueue_enqueue(sc->sc_tq, &sc->sc_chanswitchtask);
853 mwl_radar_proc(void *arg, int pending)
855 struct mwl_softc *sc = arg;
856 struct ieee80211com *ic = &sc->sc_ic;
858 DPRINTF(sc, MWL_DEBUG_ANY, "%s: radar detected, pending %u\n",
861 sc->sc_stats.mst_radardetect++;
862 /* XXX stop h/w BA streams? */
865 ieee80211_dfs_notify_radar(ic, ic->ic_curchan);
866 IEEE80211_UNLOCK(ic);
870 mwl_chanswitch_proc(void *arg, int pending)
872 struct mwl_softc *sc = arg;
873 struct ieee80211com *ic = &sc->sc_ic;
875 DPRINTF(sc, MWL_DEBUG_ANY, "%s: channel switch notice, pending %u\n",
879 sc->sc_csapending = 0;
880 ieee80211_csa_completeswitch(ic);
881 IEEE80211_UNLOCK(ic);
885 mwl_bawatchdog(const MWL_HAL_BASTREAM *sp)
887 struct ieee80211_node *ni = sp->data[0];
889 /* send DELBA and drop the stream */
890 ieee80211_ampdu_stop(ni, sp->data[1], IEEE80211_REASON_UNSPECIFIED);
894 mwl_bawatchdog_proc(void *arg, int pending)
896 struct mwl_softc *sc = arg;
897 struct mwl_hal *mh = sc->sc_mh;
898 const MWL_HAL_BASTREAM *sp;
901 sc->sc_stats.mst_bawatchdog++;
903 if (mwl_hal_getwatchdogbitmap(mh, &bitmap) != 0) {
904 DPRINTF(sc, MWL_DEBUG_AMPDU,
905 "%s: could not get bitmap\n", __func__);
906 sc->sc_stats.mst_bawatchdog_failed++;
909 DPRINTF(sc, MWL_DEBUG_AMPDU, "%s: bitmap 0x%x\n", __func__, bitmap);
910 if (bitmap == 0xff) {
912 /* disable all ba streams */
913 for (bitmap = 0; bitmap < 8; bitmap++) {
914 sp = mwl_hal_bastream_lookup(mh, bitmap);
921 DPRINTF(sc, MWL_DEBUG_AMPDU,
922 "%s: no BA streams found\n", __func__);
923 sc->sc_stats.mst_bawatchdog_empty++;
925 } else if (bitmap != 0xaa) {
926 /* disable a single ba stream */
927 sp = mwl_hal_bastream_lookup(mh, bitmap);
931 DPRINTF(sc, MWL_DEBUG_AMPDU,
932 "%s: no BA stream %d\n", __func__, bitmap);
933 sc->sc_stats.mst_bawatchdog_notfound++;
939 * Convert net80211 channel to a HAL channel.
942 mwl_mapchan(MWL_HAL_CHANNEL *hc, const struct ieee80211_channel *chan)
944 hc->channel = chan->ic_ieee;
946 *(uint32_t *)&hc->channelFlags = 0;
947 if (IEEE80211_IS_CHAN_2GHZ(chan))
948 hc->channelFlags.FreqBand = MWL_FREQ_BAND_2DOT4GHZ;
949 else if (IEEE80211_IS_CHAN_5GHZ(chan))
950 hc->channelFlags.FreqBand = MWL_FREQ_BAND_5GHZ;
951 if (IEEE80211_IS_CHAN_HT40(chan)) {
952 hc->channelFlags.ChnlWidth = MWL_CH_40_MHz_WIDTH;
953 if (IEEE80211_IS_CHAN_HT40U(chan))
954 hc->channelFlags.ExtChnlOffset = MWL_EXT_CH_ABOVE_CTRL_CH;
956 hc->channelFlags.ExtChnlOffset = MWL_EXT_CH_BELOW_CTRL_CH;
958 hc->channelFlags.ChnlWidth = MWL_CH_20_MHz_WIDTH;
959 /* XXX 10MHz channels */
963 * Inform firmware of our tx/rx dma setup. The BAR 0
964 * writes below are for compatibility with older firmware.
965 * For current firmware we send this information with a
966 * cmd block via mwl_hal_sethwdma.
969 mwl_setupdma(struct mwl_softc *sc)
973 sc->sc_hwdma.rxDescRead = sc->sc_rxdma.dd_desc_paddr;
974 WR4(sc, sc->sc_hwspecs.rxDescRead, sc->sc_hwdma.rxDescRead);
975 WR4(sc, sc->sc_hwspecs.rxDescWrite, sc->sc_hwdma.rxDescRead);
977 for (i = 0; i < MWL_NUM_TX_QUEUES-MWL_NUM_ACK_QUEUES; i++) {
978 struct mwl_txq *txq = &sc->sc_txq[i];
979 sc->sc_hwdma.wcbBase[i] = txq->dma.dd_desc_paddr;
980 WR4(sc, sc->sc_hwspecs.wcbBase[i], sc->sc_hwdma.wcbBase[i]);
982 sc->sc_hwdma.maxNumTxWcb = mwl_txbuf;
983 sc->sc_hwdma.maxNumWCB = MWL_NUM_TX_QUEUES-MWL_NUM_ACK_QUEUES;
985 error = mwl_hal_sethwdma(sc->sc_mh, &sc->sc_hwdma);
987 device_printf(sc->sc_dev,
988 "unable to setup tx/rx dma; hal status %u\n", error);
995 * Inform firmware of tx rate parameters.
996 * Called after a channel change.
999 mwl_setcurchanrates(struct mwl_softc *sc)
1001 struct ieee80211com *ic = &sc->sc_ic;
1002 const struct ieee80211_rateset *rs;
1003 MWL_HAL_TXRATE rates;
1005 memset(&rates, 0, sizeof(rates));
1006 rs = ieee80211_get_suprates(ic, ic->ic_curchan);
1007 /* rate used to send management frames */
1008 rates.MgtRate = rs->rs_rates[0] & IEEE80211_RATE_VAL;
1009 /* rate used to send multicast frames */
1010 rates.McastRate = rates.MgtRate;
1012 return mwl_hal_settxrate_auto(sc->sc_mh, &rates);
1016 * Inform firmware of tx rate parameters. Called whenever
1017 * user-settable params change and after a channel change.
1020 mwl_setrates(struct ieee80211vap *vap)
1022 struct mwl_vap *mvp = MWL_VAP(vap);
1023 struct ieee80211_node *ni = vap->iv_bss;
1024 const struct ieee80211_txparam *tp = ni->ni_txparms;
1025 MWL_HAL_TXRATE rates;
1027 KASSERT(vap->iv_state == IEEE80211_S_RUN, ("state %d", vap->iv_state));
1030 * Update the h/w rate map.
1031 * NB: 0x80 for MCS is passed through unchanged
1033 memset(&rates, 0, sizeof(rates));
1034 /* rate used to send management frames */
1035 rates.MgtRate = tp->mgmtrate;
1036 /* rate used to send multicast frames */
1037 rates.McastRate = tp->mcastrate;
1039 /* while here calculate EAPOL fixed rate cookie */
1040 mvp->mv_eapolformat = htole16(mwl_calcformat(rates.MgtRate, ni));
1042 return mwl_hal_settxrate(mvp->mv_hvap,
1043 tp->ucastrate != IEEE80211_FIXED_RATE_NONE ?
1044 RATE_FIXED : RATE_AUTO, &rates);
1048 * Setup a fixed xmit rate cookie for EAPOL frames.
1051 mwl_seteapolformat(struct ieee80211vap *vap)
1053 struct mwl_vap *mvp = MWL_VAP(vap);
1054 struct ieee80211_node *ni = vap->iv_bss;
1055 enum ieee80211_phymode mode;
1058 KASSERT(vap->iv_state == IEEE80211_S_RUN, ("state %d", vap->iv_state));
1060 mode = ieee80211_chan2mode(ni->ni_chan);
1062 * Use legacy rates when operating a mixed HT+non-HT bss.
1063 * NB: this may violate POLA for sta and wds vap's.
1065 if (mode == IEEE80211_MODE_11NA &&
1066 (vap->iv_flags_ht & IEEE80211_FHT_PUREN) == 0)
1067 rate = vap->iv_txparms[IEEE80211_MODE_11A].mgmtrate;
1068 else if (mode == IEEE80211_MODE_11NG &&
1069 (vap->iv_flags_ht & IEEE80211_FHT_PUREN) == 0)
1070 rate = vap->iv_txparms[IEEE80211_MODE_11G].mgmtrate;
1072 rate = vap->iv_txparms[mode].mgmtrate;
1074 mvp->mv_eapolformat = htole16(mwl_calcformat(rate, ni));
1078 * Map SKU+country code to region code for radar bin'ing.
1081 mwl_map2regioncode(const struct ieee80211_regdomain *rd)
1083 switch (rd->regdomain) {
1086 return DOMAIN_CODE_FCC;
1088 return DOMAIN_CODE_IC;
1092 if (rd->country == CTRY_SPAIN)
1093 return DOMAIN_CODE_SPAIN;
1094 if (rd->country == CTRY_FRANCE || rd->country == CTRY_FRANCE2)
1095 return DOMAIN_CODE_FRANCE;
1096 /* XXX force 1.3.1 radar type */
1097 return DOMAIN_CODE_ETSI_131;
1099 return DOMAIN_CODE_MKK;
1101 return DOMAIN_CODE_DGT; /* Taiwan */
1105 return DOMAIN_CODE_AUS; /* Australia */
1108 return DOMAIN_CODE_FCC; /* XXX? */
1112 mwl_hal_reset(struct mwl_softc *sc)
1114 struct ieee80211com *ic = &sc->sc_ic;
1115 struct mwl_hal *mh = sc->sc_mh;
1117 mwl_hal_setantenna(mh, WL_ANTENNATYPE_RX, sc->sc_rxantenna);
1118 mwl_hal_setantenna(mh, WL_ANTENNATYPE_TX, sc->sc_txantenna);
1119 mwl_hal_setradio(mh, 1, WL_AUTO_PREAMBLE);
1120 mwl_hal_setwmm(sc->sc_mh, (ic->ic_flags & IEEE80211_F_WME) != 0);
1121 mwl_chan_set(sc, ic->ic_curchan);
1122 /* NB: RF/RA performance tuned for indoor mode */
1123 mwl_hal_setrateadaptmode(mh, 0);
1124 mwl_hal_setoptimizationlevel(mh,
1125 (ic->ic_flags & IEEE80211_F_BURST) != 0);
1127 mwl_hal_setregioncode(mh, mwl_map2regioncode(&ic->ic_regdomain));
1129 mwl_hal_setaggampduratemode(mh, 1, 80); /* XXX */
1130 mwl_hal_setcfend(mh, 0); /* XXX */
1136 mwl_init(struct mwl_softc *sc)
1138 struct mwl_hal *mh = sc->sc_mh;
1141 MWL_LOCK_ASSERT(sc);
1144 * Stop anything previously setup. This is safe
1145 * whether this is the first time through or not.
1150 * Push vap-independent state to the firmware.
1152 if (!mwl_hal_reset(sc)) {
1153 device_printf(sc->sc_dev, "unable to reset hardware\n");
1158 * Setup recv (once); transmit is already good to go.
1160 error = mwl_startrecv(sc);
1162 device_printf(sc->sc_dev, "unable to start recv logic\n");
1167 * Enable interrupts.
1169 sc->sc_imask = MACREG_A2HRIC_BIT_RX_RDY
1170 | MACREG_A2HRIC_BIT_TX_DONE
1171 | MACREG_A2HRIC_BIT_OPC_DONE
1173 | MACREG_A2HRIC_BIT_MAC_EVENT
1175 | MACREG_A2HRIC_BIT_ICV_ERROR
1176 | MACREG_A2HRIC_BIT_RADAR_DETECT
1177 | MACREG_A2HRIC_BIT_CHAN_SWITCH
1179 | MACREG_A2HRIC_BIT_QUEUE_EMPTY
1181 | MACREG_A2HRIC_BIT_BA_WATCHDOG
1182 | MACREQ_A2HRIC_BIT_TX_ACK
1186 mwl_hal_intrset(mh, sc->sc_imask);
1187 callout_reset(&sc->sc_watchdog, hz, mwl_watchdog, sc);
1193 mwl_stop(struct mwl_softc *sc)
1196 MWL_LOCK_ASSERT(sc);
1197 if (sc->sc_running) {
1199 * Shutdown the hardware and driver.
1202 callout_stop(&sc->sc_watchdog);
1203 sc->sc_tx_timer = 0;
1209 mwl_reset_vap(struct ieee80211vap *vap, int state)
1211 struct mwl_hal_vap *hvap = MWL_VAP(vap)->mv_hvap;
1212 struct ieee80211com *ic = vap->iv_ic;
1214 if (state == IEEE80211_S_RUN)
1217 mwl_hal_setrtsthreshold(hvap, vap->iv_rtsthreshold);
1218 /* XXX auto? 20/40 split? */
1219 mwl_hal_sethtgi(hvap, (vap->iv_flags_ht &
1220 (IEEE80211_FHT_SHORTGI20|IEEE80211_FHT_SHORTGI40)) ? 1 : 0);
1221 mwl_hal_setnprot(hvap, ic->ic_htprotmode == IEEE80211_PROT_NONE ?
1222 HTPROTECT_NONE : HTPROTECT_AUTO);
1223 /* XXX txpower cap */
1225 /* re-setup beacons */
1226 if (state == IEEE80211_S_RUN &&
1227 (vap->iv_opmode == IEEE80211_M_HOSTAP ||
1228 vap->iv_opmode == IEEE80211_M_MBSS ||
1229 vap->iv_opmode == IEEE80211_M_IBSS)) {
1230 mwl_setapmode(vap, vap->iv_bss->ni_chan);
1231 mwl_hal_setnprotmode(hvap,
1232 MS(ic->ic_curhtprotmode, IEEE80211_HTINFO_OPMODE));
1233 return mwl_beacon_setup(vap);
1239 * Reset the hardware w/o losing operational state.
1240 * Used to to reset or reload hardware state for a vap.
1243 mwl_reset(struct ieee80211vap *vap, u_long cmd)
1245 struct mwl_hal_vap *hvap = MWL_VAP(vap)->mv_hvap;
1248 if (hvap != NULL) { /* WDS, MONITOR, etc. */
1249 struct ieee80211com *ic = vap->iv_ic;
1250 struct mwl_softc *sc = ic->ic_softc;
1251 struct mwl_hal *mh = sc->sc_mh;
1253 /* XXX handle DWDS sta vap change */
1254 /* XXX do we need to disable interrupts? */
1255 mwl_hal_intrset(mh, 0); /* disable interrupts */
1256 error = mwl_reset_vap(vap, vap->iv_state);
1257 mwl_hal_intrset(mh, sc->sc_imask);
1263 * Allocate a tx buffer for sending a frame. The
1264 * packet is assumed to have the WME AC stored so
1265 * we can use it to select the appropriate h/w queue.
1267 static struct mwl_txbuf *
1268 mwl_gettxbuf(struct mwl_softc *sc, struct mwl_txq *txq)
1270 struct mwl_txbuf *bf;
1273 * Grab a TX buffer and associated resources.
1276 bf = STAILQ_FIRST(&txq->free);
1278 STAILQ_REMOVE_HEAD(&txq->free, bf_list);
1281 MWL_TXQ_UNLOCK(txq);
1283 DPRINTF(sc, MWL_DEBUG_XMIT,
1284 "%s: out of xmit buffers on q %d\n", __func__, txq->qnum);
1289 * Return a tx buffer to the queue it came from. Note there
1290 * are two cases because we must preserve the order of buffers
1291 * as it reflects the fixed order of descriptors in memory
1292 * (the firmware pre-fetches descriptors so we cannot reorder).
1295 mwl_puttxbuf_head(struct mwl_txq *txq, struct mwl_txbuf *bf)
1300 STAILQ_INSERT_HEAD(&txq->free, bf, bf_list);
1302 MWL_TXQ_UNLOCK(txq);
1306 mwl_puttxbuf_tail(struct mwl_txq *txq, struct mwl_txbuf *bf)
1311 STAILQ_INSERT_TAIL(&txq->free, bf, bf_list);
1313 MWL_TXQ_UNLOCK(txq);
1317 mwl_transmit(struct ieee80211com *ic, struct mbuf *m)
1319 struct mwl_softc *sc = ic->ic_softc;
1323 if (!sc->sc_running) {
1327 error = mbufq_enqueue(&sc->sc_snd, m);
1338 mwl_start(struct mwl_softc *sc)
1340 struct ieee80211_node *ni;
1341 struct mwl_txbuf *bf;
1343 struct mwl_txq *txq = NULL; /* XXX silence gcc */
1346 MWL_LOCK_ASSERT(sc);
1347 if (!sc->sc_running || sc->sc_invalid)
1350 while ((m = mbufq_dequeue(&sc->sc_snd)) != NULL) {
1352 * Grab the node for the destination.
1354 ni = (struct ieee80211_node *) m->m_pkthdr.rcvif;
1355 KASSERT(ni != NULL, ("no node"));
1356 m->m_pkthdr.rcvif = NULL; /* committed, clear ref */
1358 * Grab a TX buffer and associated resources.
1359 * We honor the classification by the 802.11 layer.
1361 txq = sc->sc_ac2q[M_WME_GETAC(m)];
1362 bf = mwl_gettxbuf(sc, txq);
1365 ieee80211_free_node(ni);
1366 #ifdef MWL_TX_NODROP
1367 sc->sc_stats.mst_tx_qstop++;
1370 DPRINTF(sc, MWL_DEBUG_XMIT,
1371 "%s: tail drop on q %d\n", __func__, txq->qnum);
1372 sc->sc_stats.mst_tx_qdrop++;
1374 #endif /* MWL_TX_NODROP */
1378 * Pass the frame to the h/w for transmission.
1380 if (mwl_tx_start(sc, ni, bf, m)) {
1381 if_inc_counter(ni->ni_vap->iv_ifp,
1382 IFCOUNTER_OERRORS, 1);
1383 mwl_puttxbuf_head(txq, bf);
1384 ieee80211_free_node(ni);
1388 if (nqueued >= mwl_txcoalesce) {
1390 * Poke the firmware to process queued frames;
1391 * see below about (lack of) locking.
1394 mwl_hal_txstart(sc->sc_mh, 0/*XXX*/);
1399 * NB: We don't need to lock against tx done because
1400 * this just prods the firmware to check the transmit
1401 * descriptors. The firmware will also start fetching
1402 * descriptors by itself if it notices new ones are
1403 * present when it goes to deliver a tx done interrupt
1404 * to the host. So if we race with tx done processing
1405 * it's ok. Delivering the kick here rather than in
1406 * mwl_tx_start is an optimization to avoid poking the
1407 * firmware for each packet.
1409 * NB: the queue id isn't used so 0 is ok.
1411 mwl_hal_txstart(sc->sc_mh, 0/*XXX*/);
1416 mwl_raw_xmit(struct ieee80211_node *ni, struct mbuf *m,
1417 const struct ieee80211_bpf_params *params)
1419 struct ieee80211com *ic = ni->ni_ic;
1420 struct mwl_softc *sc = ic->ic_softc;
1421 struct mwl_txbuf *bf;
1422 struct mwl_txq *txq;
1424 if (!sc->sc_running || sc->sc_invalid) {
1429 * Grab a TX buffer and associated resources.
1430 * Note that we depend on the classification
1431 * by the 802.11 layer to get to the right h/w
1432 * queue. Management frames must ALWAYS go on
1433 * queue 1 but we cannot just force that here
1434 * because we may receive non-mgt frames.
1436 txq = sc->sc_ac2q[M_WME_GETAC(m)];
1437 bf = mwl_gettxbuf(sc, txq);
1439 sc->sc_stats.mst_tx_qstop++;
1444 * Pass the frame to the h/w for transmission.
1446 if (mwl_tx_start(sc, ni, bf, m)) {
1447 mwl_puttxbuf_head(txq, bf);
1449 return EIO; /* XXX */
1452 * NB: We don't need to lock against tx done because
1453 * this just prods the firmware to check the transmit
1454 * descriptors. The firmware will also start fetching
1455 * descriptors by itself if it notices new ones are
1456 * present when it goes to deliver a tx done interrupt
1457 * to the host. So if we race with tx done processing
1458 * it's ok. Delivering the kick here rather than in
1459 * mwl_tx_start is an optimization to avoid poking the
1460 * firmware for each packet.
1462 * NB: the queue id isn't used so 0 is ok.
1464 mwl_hal_txstart(sc->sc_mh, 0/*XXX*/);
1469 mwl_media_change(struct ifnet *ifp)
1471 struct ieee80211vap *vap = ifp->if_softc;
1474 error = ieee80211_media_change(ifp);
1475 /* NB: only the fixed rate can change and that doesn't need a reset */
1476 if (error == ENETRESET) {
1485 mwl_keyprint(struct mwl_softc *sc, const char *tag,
1486 const MWL_HAL_KEYVAL *hk, const uint8_t mac[IEEE80211_ADDR_LEN])
1488 static const char *ciphers[] = {
1495 printf("%s: [%u] %-7s", tag, hk->keyIndex, ciphers[hk->keyTypeId]);
1496 for (i = 0, n = hk->keyLen; i < n; i++)
1497 printf(" %02x", hk->key.aes[i]);
1498 printf(" mac %s", ether_sprintf(mac));
1499 if (hk->keyTypeId == KEY_TYPE_ID_TKIP) {
1500 printf(" %s", "rxmic");
1501 for (i = 0; i < sizeof(hk->key.tkip.rxMic); i++)
1502 printf(" %02x", hk->key.tkip.rxMic[i]);
1504 for (i = 0; i < sizeof(hk->key.tkip.txMic); i++)
1505 printf(" %02x", hk->key.tkip.txMic[i]);
1507 printf(" flags 0x%x\n", hk->keyFlags);
1512 * Allocate a key cache slot for a unicast key. The
1513 * firmware handles key allocation and every station is
1514 * guaranteed key space so we are always successful.
1517 mwl_key_alloc(struct ieee80211vap *vap, struct ieee80211_key *k,
1518 ieee80211_keyix *keyix, ieee80211_keyix *rxkeyix)
1520 struct mwl_softc *sc = vap->iv_ic->ic_softc;
1522 if (k->wk_keyix != IEEE80211_KEYIX_NONE ||
1523 (k->wk_flags & IEEE80211_KEY_GROUP)) {
1524 if (!(&vap->iv_nw_keys[0] <= k &&
1525 k < &vap->iv_nw_keys[IEEE80211_WEP_NKID])) {
1526 /* should not happen */
1527 DPRINTF(sc, MWL_DEBUG_KEYCACHE,
1528 "%s: bogus group key\n", __func__);
1531 /* give the caller what they requested */
1532 *keyix = *rxkeyix = ieee80211_crypto_get_key_wepidx(vap, k);
1535 * Firmware handles key allocation.
1537 *keyix = *rxkeyix = 0;
1543 * Delete a key entry allocated by mwl_key_alloc.
1546 mwl_key_delete(struct ieee80211vap *vap, const struct ieee80211_key *k)
1548 struct mwl_softc *sc = vap->iv_ic->ic_softc;
1549 struct mwl_hal_vap *hvap = MWL_VAP(vap)->mv_hvap;
1551 const uint8_t bcastaddr[IEEE80211_ADDR_LEN] =
1552 { 0xff, 0xff, 0xff, 0xff, 0xff, 0xff };
1555 if (vap->iv_opmode != IEEE80211_M_WDS) {
1556 /* XXX monitor mode? */
1557 DPRINTF(sc, MWL_DEBUG_KEYCACHE,
1558 "%s: no hvap for opmode %d\n", __func__,
1562 hvap = MWL_VAP(vap)->mv_ap_hvap;
1565 DPRINTF(sc, MWL_DEBUG_KEYCACHE, "%s: delete key %u\n",
1566 __func__, k->wk_keyix);
1568 memset(&hk, 0, sizeof(hk));
1569 hk.keyIndex = k->wk_keyix;
1570 switch (k->wk_cipher->ic_cipher) {
1571 case IEEE80211_CIPHER_WEP:
1572 hk.keyTypeId = KEY_TYPE_ID_WEP;
1574 case IEEE80211_CIPHER_TKIP:
1575 hk.keyTypeId = KEY_TYPE_ID_TKIP;
1577 case IEEE80211_CIPHER_AES_CCM:
1578 hk.keyTypeId = KEY_TYPE_ID_AES;
1581 /* XXX should not happen */
1582 DPRINTF(sc, MWL_DEBUG_KEYCACHE, "%s: unknown cipher %d\n",
1583 __func__, k->wk_cipher->ic_cipher);
1586 return (mwl_hal_keyreset(hvap, &hk, bcastaddr) == 0); /*XXX*/
1590 addgroupflags(MWL_HAL_KEYVAL *hk, const struct ieee80211_key *k)
1592 if (k->wk_flags & IEEE80211_KEY_GROUP) {
1593 if (k->wk_flags & IEEE80211_KEY_XMIT)
1594 hk->keyFlags |= KEY_FLAG_TXGROUPKEY;
1595 if (k->wk_flags & IEEE80211_KEY_RECV)
1596 hk->keyFlags |= KEY_FLAG_RXGROUPKEY;
1603 * Set the key cache contents for the specified key. Key cache
1604 * slot(s) must already have been allocated by mwl_key_alloc.
1607 mwl_key_set(struct ieee80211vap *vap, const struct ieee80211_key *k)
1609 return (_mwl_key_set(vap, k, k->wk_macaddr));
1613 _mwl_key_set(struct ieee80211vap *vap, const struct ieee80211_key *k,
1614 const uint8_t mac[IEEE80211_ADDR_LEN])
1616 #define GRPXMIT (IEEE80211_KEY_XMIT | IEEE80211_KEY_GROUP)
1617 /* NB: static wep keys are marked GROUP+tx/rx; GTK will be tx or rx */
1618 #define IEEE80211_IS_STATICKEY(k) \
1619 (((k)->wk_flags & (GRPXMIT|IEEE80211_KEY_RECV)) == \
1620 (GRPXMIT|IEEE80211_KEY_RECV))
1621 struct mwl_softc *sc = vap->iv_ic->ic_softc;
1622 struct mwl_hal_vap *hvap = MWL_VAP(vap)->mv_hvap;
1623 const struct ieee80211_cipher *cip = k->wk_cipher;
1624 const uint8_t *macaddr;
1627 KASSERT((k->wk_flags & IEEE80211_KEY_SWCRYPT) == 0,
1628 ("s/w crypto set?"));
1631 if (vap->iv_opmode != IEEE80211_M_WDS) {
1632 /* XXX monitor mode? */
1633 DPRINTF(sc, MWL_DEBUG_KEYCACHE,
1634 "%s: no hvap for opmode %d\n", __func__,
1638 hvap = MWL_VAP(vap)->mv_ap_hvap;
1640 memset(&hk, 0, sizeof(hk));
1641 hk.keyIndex = k->wk_keyix;
1642 switch (cip->ic_cipher) {
1643 case IEEE80211_CIPHER_WEP:
1644 hk.keyTypeId = KEY_TYPE_ID_WEP;
1645 hk.keyLen = k->wk_keylen;
1646 if (k->wk_keyix == vap->iv_def_txkey)
1647 hk.keyFlags = KEY_FLAG_WEP_TXKEY;
1648 if (!IEEE80211_IS_STATICKEY(k)) {
1649 /* NB: WEP is never used for the PTK */
1650 (void) addgroupflags(&hk, k);
1653 case IEEE80211_CIPHER_TKIP:
1654 hk.keyTypeId = KEY_TYPE_ID_TKIP;
1655 hk.key.tkip.tsc.high = (uint32_t)(k->wk_keytsc >> 16);
1656 hk.key.tkip.tsc.low = (uint16_t)k->wk_keytsc;
1657 hk.keyFlags = KEY_FLAG_TSC_VALID | KEY_FLAG_MICKEY_VALID;
1658 hk.keyLen = k->wk_keylen + IEEE80211_MICBUF_SIZE;
1659 if (!addgroupflags(&hk, k))
1660 hk.keyFlags |= KEY_FLAG_PAIRWISE;
1662 case IEEE80211_CIPHER_AES_CCM:
1663 hk.keyTypeId = KEY_TYPE_ID_AES;
1664 hk.keyLen = k->wk_keylen;
1665 if (!addgroupflags(&hk, k))
1666 hk.keyFlags |= KEY_FLAG_PAIRWISE;
1669 /* XXX should not happen */
1670 DPRINTF(sc, MWL_DEBUG_KEYCACHE, "%s: unknown cipher %d\n",
1671 __func__, k->wk_cipher->ic_cipher);
1675 * NB: tkip mic keys get copied here too; the layout
1676 * just happens to match that in ieee80211_key.
1678 memcpy(hk.key.aes, k->wk_key, hk.keyLen);
1681 * Locate address of sta db entry for writing key;
1682 * the convention unfortunately is somewhat different
1683 * than how net80211, hostapd, and wpa_supplicant think.
1685 if (vap->iv_opmode == IEEE80211_M_STA) {
1687 * NB: keys plumbed before the sta reaches AUTH state
1688 * will be discarded or written to the wrong sta db
1689 * entry because iv_bss is meaningless. This is ok
1690 * (right now) because we handle deferred plumbing of
1691 * WEP keys when the sta reaches AUTH state.
1693 macaddr = vap->iv_bss->ni_bssid;
1694 if ((k->wk_flags & IEEE80211_KEY_GROUP) == 0) {
1695 /* XXX plumb to local sta db too for static key wep */
1696 mwl_hal_keyset(hvap, &hk, vap->iv_myaddr);
1698 } else if (vap->iv_opmode == IEEE80211_M_WDS &&
1699 vap->iv_state != IEEE80211_S_RUN) {
1701 * Prior to RUN state a WDS vap will not it's BSS node
1702 * setup so we will plumb the key to the wrong mac
1703 * address (it'll be our local address). Workaround
1704 * this for the moment by grabbing the correct address.
1706 macaddr = vap->iv_des_bssid;
1707 } else if ((k->wk_flags & GRPXMIT) == GRPXMIT)
1708 macaddr = vap->iv_myaddr;
1711 KEYPRINTF(sc, &hk, macaddr);
1712 return (mwl_hal_keyset(hvap, &hk, macaddr) == 0);
1713 #undef IEEE80211_IS_STATICKEY
1718 * Set the multicast filter contents into the hardware.
1719 * XXX f/w has no support; just defer to the os.
1722 mwl_setmcastfilter(struct mwl_softc *sc)
1725 struct ether_multi *enm;
1726 struct ether_multistep estep;
1727 uint8_t macs[IEEE80211_ADDR_LEN*MWL_HAL_MCAST_MAX];/* XXX stack use */
1733 ETHER_FIRST_MULTI(estep, &sc->sc_ec, enm);
1734 while (enm != NULL) {
1735 /* XXX Punt on ranges. */
1736 if (nmc == MWL_HAL_MCAST_MAX ||
1737 !IEEE80211_ADDR_EQ(enm->enm_addrlo, enm->enm_addrhi)) {
1738 ifp->if_flags |= IFF_ALLMULTI;
1741 IEEE80211_ADDR_COPY(mp, enm->enm_addrlo);
1742 mp += IEEE80211_ADDR_LEN, nmc++;
1743 ETHER_NEXT_MULTI(estep, enm);
1745 ifp->if_flags &= ~IFF_ALLMULTI;
1746 mwl_hal_setmcast(sc->sc_mh, nmc, macs);
1751 mwl_mode_init(struct mwl_softc *sc)
1753 struct ieee80211com *ic = &sc->sc_ic;
1754 struct mwl_hal *mh = sc->sc_mh;
1756 mwl_hal_setpromisc(mh, ic->ic_promisc > 0);
1757 mwl_setmcastfilter(sc);
1763 * Callback from the 802.11 layer after a multicast state change.
1766 mwl_update_mcast(struct ieee80211com *ic)
1768 struct mwl_softc *sc = ic->ic_softc;
1770 mwl_setmcastfilter(sc);
1774 * Callback from the 802.11 layer after a promiscuous mode change.
1775 * Note this interface does not check the operating mode as this
1776 * is an internal callback and we are expected to honor the current
1777 * state (e.g. this is used for setting the interface in promiscuous
1778 * mode when operating in hostap mode to do ACS).
1781 mwl_update_promisc(struct ieee80211com *ic)
1783 struct mwl_softc *sc = ic->ic_softc;
1785 mwl_hal_setpromisc(sc->sc_mh, ic->ic_promisc > 0);
1789 * Callback from the 802.11 layer to update the slot time
1790 * based on the current setting. We use it to notify the
1791 * firmware of ERP changes and the f/w takes care of things
1792 * like slot time and preamble.
1795 mwl_updateslot(struct ieee80211com *ic)
1797 struct mwl_softc *sc = ic->ic_softc;
1798 struct mwl_hal *mh = sc->sc_mh;
1801 /* NB: can be called early; suppress needless cmds */
1802 if (!sc->sc_running)
1806 * Calculate the ERP flags. The firwmare will use
1807 * this to carry out the appropriate measures.
1810 if (IEEE80211_IS_CHAN_ANYG(ic->ic_curchan)) {
1811 if ((ic->ic_flags & IEEE80211_F_SHSLOT) == 0)
1812 prot |= IEEE80211_ERP_NON_ERP_PRESENT;
1813 if (ic->ic_flags & IEEE80211_F_USEPROT)
1814 prot |= IEEE80211_ERP_USE_PROTECTION;
1815 if (ic->ic_flags & IEEE80211_F_USEBARKER)
1816 prot |= IEEE80211_ERP_LONG_PREAMBLE;
1819 DPRINTF(sc, MWL_DEBUG_RESET,
1820 "%s: chan %u MHz/flags 0x%x %s slot, (prot 0x%x ic_flags 0x%x)\n",
1821 __func__, ic->ic_curchan->ic_freq, ic->ic_curchan->ic_flags,
1822 ic->ic_flags & IEEE80211_F_SHSLOT ? "short" : "long", prot,
1825 mwl_hal_setgprot(mh, prot);
1829 * Setup the beacon frame.
1832 mwl_beacon_setup(struct ieee80211vap *vap)
1834 struct mwl_hal_vap *hvap = MWL_VAP(vap)->mv_hvap;
1835 struct ieee80211_node *ni = vap->iv_bss;
1838 m = ieee80211_beacon_alloc(ni);
1841 mwl_hal_setbeacon(hvap, mtod(m, const void *), m->m_len);
1848 * Update the beacon frame in response to a change.
1851 mwl_beacon_update(struct ieee80211vap *vap, int item)
1853 struct mwl_hal_vap *hvap = MWL_VAP(vap)->mv_hvap;
1854 struct ieee80211com *ic = vap->iv_ic;
1856 KASSERT(hvap != NULL, ("no beacon"));
1858 case IEEE80211_BEACON_ERP:
1861 case IEEE80211_BEACON_HTINFO:
1862 mwl_hal_setnprotmode(hvap,
1863 MS(ic->ic_curhtprotmode, IEEE80211_HTINFO_OPMODE));
1865 case IEEE80211_BEACON_CAPS:
1866 case IEEE80211_BEACON_WME:
1867 case IEEE80211_BEACON_APPIE:
1868 case IEEE80211_BEACON_CSA:
1870 case IEEE80211_BEACON_TIM:
1871 /* NB: firmware always forms TIM */
1874 /* XXX retain beacon frame and update */
1875 mwl_beacon_setup(vap);
1879 mwl_load_cb(void *arg, bus_dma_segment_t *segs, int nsegs, int error)
1881 bus_addr_t *paddr = (bus_addr_t*) arg;
1882 KASSERT(error == 0, ("error %u on bus_dma callback", error));
1883 *paddr = segs->ds_addr;
1886 #ifdef MWL_HOST_PS_SUPPORT
1888 * Handle power save station occupancy changes.
1891 mwl_update_ps(struct ieee80211vap *vap, int nsta)
1893 struct mwl_vap *mvp = MWL_VAP(vap);
1895 if (nsta == 0 || mvp->mv_last_ps_sta == 0)
1896 mwl_hal_setpowersave_bss(mvp->mv_hvap, nsta);
1897 mvp->mv_last_ps_sta = nsta;
1901 * Handle associated station power save state changes.
1904 mwl_set_tim(struct ieee80211_node *ni, int set)
1906 struct ieee80211vap *vap = ni->ni_vap;
1907 struct mwl_vap *mvp = MWL_VAP(vap);
1909 if (mvp->mv_set_tim(ni, set)) { /* NB: state change */
1910 mwl_hal_setpowersave_sta(mvp->mv_hvap,
1911 IEEE80211_AID(ni->ni_associd), set);
1916 #endif /* MWL_HOST_PS_SUPPORT */
1919 mwl_desc_setup(struct mwl_softc *sc, const char *name,
1920 struct mwl_descdma *dd,
1921 int nbuf, size_t bufsize, int ndesc, size_t descsize)
1926 DPRINTF(sc, MWL_DEBUG_RESET,
1927 "%s: %s DMA: %u bufs (%ju) %u desc/buf (%ju)\n",
1928 __func__, name, nbuf, (uintmax_t) bufsize,
1929 ndesc, (uintmax_t) descsize);
1932 dd->dd_desc_len = nbuf * ndesc * descsize;
1935 * Setup DMA descriptor area.
1937 error = bus_dma_tag_create(bus_get_dma_tag(sc->sc_dev), /* parent */
1938 PAGE_SIZE, 0, /* alignment, bounds */
1939 BUS_SPACE_MAXADDR_32BIT, /* lowaddr */
1940 BUS_SPACE_MAXADDR, /* highaddr */
1941 NULL, NULL, /* filter, filterarg */
1942 dd->dd_desc_len, /* maxsize */
1944 dd->dd_desc_len, /* maxsegsize */
1945 BUS_DMA_ALLOCNOW, /* flags */
1946 NULL, /* lockfunc */
1950 device_printf(sc->sc_dev, "cannot allocate %s DMA tag\n", dd->dd_name);
1954 /* allocate descriptors */
1955 error = bus_dmamem_alloc(dd->dd_dmat, (void**) &dd->dd_desc,
1956 BUS_DMA_NOWAIT | BUS_DMA_COHERENT,
1959 device_printf(sc->sc_dev, "unable to alloc memory for %u %s descriptors, "
1960 "error %u\n", nbuf * ndesc, dd->dd_name, error);
1964 error = bus_dmamap_load(dd->dd_dmat, dd->dd_dmamap,
1965 dd->dd_desc, dd->dd_desc_len,
1966 mwl_load_cb, &dd->dd_desc_paddr,
1969 device_printf(sc->sc_dev, "unable to map %s descriptors, error %u\n",
1970 dd->dd_name, error);
1975 memset(ds, 0, dd->dd_desc_len);
1976 DPRINTF(sc, MWL_DEBUG_RESET,
1977 "%s: %s DMA map: %p (%lu) -> 0x%jx (%lu)\n",
1978 __func__, dd->dd_name, ds, (u_long) dd->dd_desc_len,
1979 (uintmax_t) dd->dd_desc_paddr, /*XXX*/ (u_long) dd->dd_desc_len);
1983 bus_dmamem_free(dd->dd_dmat, dd->dd_desc, dd->dd_dmamap);
1985 bus_dma_tag_destroy(dd->dd_dmat);
1986 memset(dd, 0, sizeof(*dd));
1992 mwl_desc_cleanup(struct mwl_softc *sc, struct mwl_descdma *dd)
1994 bus_dmamap_unload(dd->dd_dmat, dd->dd_dmamap);
1995 bus_dmamem_free(dd->dd_dmat, dd->dd_desc, dd->dd_dmamap);
1996 bus_dma_tag_destroy(dd->dd_dmat);
1998 memset(dd, 0, sizeof(*dd));
2002 * Construct a tx q's free list. The order of entries on
2003 * the list must reflect the physical layout of tx descriptors
2004 * because the firmware pre-fetches descriptors.
2006 * XXX might be better to use indices into the buffer array.
2009 mwl_txq_reset(struct mwl_softc *sc, struct mwl_txq *txq)
2011 struct mwl_txbuf *bf;
2014 bf = txq->dma.dd_bufptr;
2015 STAILQ_INIT(&txq->free);
2016 for (i = 0; i < mwl_txbuf; i++, bf++)
2017 STAILQ_INSERT_TAIL(&txq->free, bf, bf_list);
2021 #define DS2PHYS(_dd, _ds) \
2022 ((_dd)->dd_desc_paddr + ((caddr_t)(_ds) - (caddr_t)(_dd)->dd_desc))
2025 mwl_txdma_setup(struct mwl_softc *sc, struct mwl_txq *txq)
2027 int error, bsize, i;
2028 struct mwl_txbuf *bf;
2029 struct mwl_txdesc *ds;
2031 error = mwl_desc_setup(sc, "tx", &txq->dma,
2032 mwl_txbuf, sizeof(struct mwl_txbuf),
2033 MWL_TXDESC, sizeof(struct mwl_txdesc));
2037 /* allocate and setup tx buffers */
2038 bsize = mwl_txbuf * sizeof(struct mwl_txbuf);
2039 bf = malloc(bsize, M_MWLDEV, M_NOWAIT | M_ZERO);
2041 device_printf(sc->sc_dev, "malloc of %u tx buffers failed\n",
2045 txq->dma.dd_bufptr = bf;
2047 ds = txq->dma.dd_desc;
2048 for (i = 0; i < mwl_txbuf; i++, bf++, ds += MWL_TXDESC) {
2050 bf->bf_daddr = DS2PHYS(&txq->dma, ds);
2051 error = bus_dmamap_create(sc->sc_dmat, BUS_DMA_NOWAIT,
2054 device_printf(sc->sc_dev, "unable to create dmamap for tx "
2055 "buffer %u, error %u\n", i, error);
2059 mwl_txq_reset(sc, txq);
2064 mwl_txdma_cleanup(struct mwl_softc *sc, struct mwl_txq *txq)
2066 struct mwl_txbuf *bf;
2069 bf = txq->dma.dd_bufptr;
2070 for (i = 0; i < mwl_txbuf; i++, bf++) {
2071 KASSERT(bf->bf_m == NULL, ("mbuf on free list"));
2072 KASSERT(bf->bf_node == NULL, ("node on free list"));
2073 if (bf->bf_dmamap != NULL)
2074 bus_dmamap_destroy(sc->sc_dmat, bf->bf_dmamap);
2076 STAILQ_INIT(&txq->free);
2078 if (txq->dma.dd_bufptr != NULL) {
2079 free(txq->dma.dd_bufptr, M_MWLDEV);
2080 txq->dma.dd_bufptr = NULL;
2082 if (txq->dma.dd_desc_len != 0)
2083 mwl_desc_cleanup(sc, &txq->dma);
2087 mwl_rxdma_setup(struct mwl_softc *sc)
2089 int error, jumbosize, bsize, i;
2090 struct mwl_rxbuf *bf;
2091 struct mwl_jumbo *rbuf;
2092 struct mwl_rxdesc *ds;
2095 error = mwl_desc_setup(sc, "rx", &sc->sc_rxdma,
2096 mwl_rxdesc, sizeof(struct mwl_rxbuf),
2097 1, sizeof(struct mwl_rxdesc));
2102 * Receive is done to a private pool of jumbo buffers.
2103 * This allows us to attach to mbuf's and avoid re-mapping
2104 * memory on each rx we post. We allocate a large chunk
2105 * of memory and manage it in the driver. The mbuf free
2106 * callback method is used to reclaim frames after sending
2107 * them up the stack. By default we allocate 2x the number of
2108 * rx descriptors configured so we have some slop to hold
2109 * us while frames are processed.
2111 if (mwl_rxbuf < 2*mwl_rxdesc) {
2112 device_printf(sc->sc_dev,
2113 "too few rx dma buffers (%d); increasing to %d\n",
2114 mwl_rxbuf, 2*mwl_rxdesc);
2115 mwl_rxbuf = 2*mwl_rxdesc;
2117 jumbosize = roundup(MWL_AGGR_SIZE, PAGE_SIZE);
2118 sc->sc_rxmemsize = mwl_rxbuf*jumbosize;
2120 error = bus_dma_tag_create(sc->sc_dmat, /* parent */
2121 PAGE_SIZE, 0, /* alignment, bounds */
2122 BUS_SPACE_MAXADDR_32BIT, /* lowaddr */
2123 BUS_SPACE_MAXADDR, /* highaddr */
2124 NULL, NULL, /* filter, filterarg */
2125 sc->sc_rxmemsize, /* maxsize */
2127 sc->sc_rxmemsize, /* maxsegsize */
2128 BUS_DMA_ALLOCNOW, /* flags */
2129 NULL, /* lockfunc */
2133 device_printf(sc->sc_dev, "could not create rx DMA tag\n");
2137 error = bus_dmamem_alloc(sc->sc_rxdmat, (void**) &sc->sc_rxmem,
2138 BUS_DMA_NOWAIT | BUS_DMA_COHERENT,
2141 device_printf(sc->sc_dev, "could not alloc %ju bytes of rx DMA memory\n",
2142 (uintmax_t) sc->sc_rxmemsize);
2146 error = bus_dmamap_load(sc->sc_rxdmat, sc->sc_rxmap,
2147 sc->sc_rxmem, sc->sc_rxmemsize,
2148 mwl_load_cb, &sc->sc_rxmem_paddr,
2151 device_printf(sc->sc_dev, "could not load rx DMA map\n");
2156 * Allocate rx buffers and set them up.
2158 bsize = mwl_rxdesc * sizeof(struct mwl_rxbuf);
2159 bf = malloc(bsize, M_MWLDEV, M_NOWAIT | M_ZERO);
2161 device_printf(sc->sc_dev, "malloc of %u rx buffers failed\n", bsize);
2164 sc->sc_rxdma.dd_bufptr = bf;
2166 STAILQ_INIT(&sc->sc_rxbuf);
2167 ds = sc->sc_rxdma.dd_desc;
2168 for (i = 0; i < mwl_rxdesc; i++, bf++, ds++) {
2170 bf->bf_daddr = DS2PHYS(&sc->sc_rxdma, ds);
2171 /* pre-assign dma buffer */
2172 bf->bf_data = ((uint8_t *)sc->sc_rxmem) + (i*jumbosize);
2173 /* NB: tail is intentional to preserve descriptor order */
2174 STAILQ_INSERT_TAIL(&sc->sc_rxbuf, bf, bf_list);
2178 * Place remainder of dma memory buffers on the free list.
2180 SLIST_INIT(&sc->sc_rxfree);
2181 for (; i < mwl_rxbuf; i++) {
2182 data = ((uint8_t *)sc->sc_rxmem) + (i*jumbosize);
2183 rbuf = MWL_JUMBO_DATA2BUF(data);
2184 SLIST_INSERT_HEAD(&sc->sc_rxfree, rbuf, next);
2192 mwl_rxdma_cleanup(struct mwl_softc *sc)
2194 if (sc->sc_rxmem_paddr != 0) {
2195 bus_dmamap_unload(sc->sc_rxdmat, sc->sc_rxmap);
2196 sc->sc_rxmem_paddr = 0;
2198 if (sc->sc_rxmem != NULL) {
2199 bus_dmamem_free(sc->sc_rxdmat, sc->sc_rxmem, sc->sc_rxmap);
2200 sc->sc_rxmem = NULL;
2202 if (sc->sc_rxdma.dd_bufptr != NULL) {
2203 free(sc->sc_rxdma.dd_bufptr, M_MWLDEV);
2204 sc->sc_rxdma.dd_bufptr = NULL;
2206 if (sc->sc_rxdma.dd_desc_len != 0)
2207 mwl_desc_cleanup(sc, &sc->sc_rxdma);
2211 mwl_dma_setup(struct mwl_softc *sc)
2215 error = mwl_rxdma_setup(sc);
2217 mwl_rxdma_cleanup(sc);
2221 for (i = 0; i < MWL_NUM_TX_QUEUES; i++) {
2222 error = mwl_txdma_setup(sc, &sc->sc_txq[i]);
2224 mwl_dma_cleanup(sc);
2232 mwl_dma_cleanup(struct mwl_softc *sc)
2236 for (i = 0; i < MWL_NUM_TX_QUEUES; i++)
2237 mwl_txdma_cleanup(sc, &sc->sc_txq[i]);
2238 mwl_rxdma_cleanup(sc);
2241 static struct ieee80211_node *
2242 mwl_node_alloc(struct ieee80211vap *vap, const uint8_t mac[IEEE80211_ADDR_LEN])
2244 struct ieee80211com *ic = vap->iv_ic;
2245 struct mwl_softc *sc = ic->ic_softc;
2246 const size_t space = sizeof(struct mwl_node);
2247 struct mwl_node *mn;
2249 mn = malloc(space, M_80211_NODE, M_NOWAIT|M_ZERO);
2254 DPRINTF(sc, MWL_DEBUG_NODE, "%s: mn %p\n", __func__, mn);
2255 return &mn->mn_node;
2259 mwl_node_cleanup(struct ieee80211_node *ni)
2261 struct ieee80211com *ic = ni->ni_ic;
2262 struct mwl_softc *sc = ic->ic_softc;
2263 struct mwl_node *mn = MWL_NODE(ni);
2265 DPRINTF(sc, MWL_DEBUG_NODE, "%s: ni %p ic %p staid %d\n",
2266 __func__, ni, ni->ni_ic, mn->mn_staid);
2268 if (mn->mn_staid != 0) {
2269 struct ieee80211vap *vap = ni->ni_vap;
2271 if (mn->mn_hvap != NULL) {
2272 if (vap->iv_opmode == IEEE80211_M_STA)
2273 mwl_hal_delstation(mn->mn_hvap, vap->iv_myaddr);
2275 mwl_hal_delstation(mn->mn_hvap, ni->ni_macaddr);
2278 * NB: legacy WDS peer sta db entry is installed using
2279 * the associate ap's hvap; use it again to delete it.
2280 * XXX can vap be NULL?
2282 else if (vap->iv_opmode == IEEE80211_M_WDS &&
2283 MWL_VAP(vap)->mv_ap_hvap != NULL)
2284 mwl_hal_delstation(MWL_VAP(vap)->mv_ap_hvap,
2286 delstaid(sc, mn->mn_staid);
2289 sc->sc_node_cleanup(ni);
2293 * Reclaim rx dma buffers from packets sitting on the ampdu
2294 * reorder queue for a station. We replace buffers with a
2295 * system cluster (if available).
2298 mwl_ampdu_rxdma_reclaim(struct ieee80211_rx_ampdu *rap)
2305 n = rap->rxa_qframes;
2306 for (i = 0; i < rap->rxa_wnd && n > 0; i++) {
2311 /* our dma buffers have a well-known free routine */
2312 if ((m->m_flags & M_EXT) == 0 ||
2313 m->m_ext.ext_free != mwl_ext_free)
2316 * Try to allocate a cluster and move the data.
2318 off = m->m_data - m->m_ext.ext_buf;
2319 if (off + m->m_pkthdr.len > MCLBYTES) {
2320 /* XXX no AMSDU for now */
2323 cl = pool_cache_get_paddr(&mclpool_cache, 0,
2324 &m->m_ext.ext_paddr);
2327 * Copy the existing data to the cluster, remove
2328 * the rx dma buffer, and attach the cluster in
2329 * its place. Note we preserve the offset to the
2330 * data so frames being bridged can still prepend
2331 * their headers without adding another mbuf.
2333 memcpy((caddr_t) cl + off, m->m_data, m->m_pkthdr.len);
2335 MEXTADD(m, cl, MCLBYTES, 0, NULL, &mclpool_cache);
2336 /* setup mbuf like _MCLGET does */
2337 m->m_flags |= M_CLUSTER | M_EXT_RW;
2338 _MOWNERREF(m, M_EXT | M_CLUSTER);
2339 /* NB: m_data is clobbered by MEXTADDR, adjust */
2347 * Callback to reclaim resources. We first let the
2348 * net80211 layer do it's thing, then if we are still
2349 * blocked by a lack of rx dma buffers we walk the ampdu
2350 * reorder q's to reclaim buffers by copying to a system
2354 mwl_node_drain(struct ieee80211_node *ni)
2356 struct ieee80211com *ic = ni->ni_ic;
2357 struct mwl_softc *sc = ic->ic_softc;
2358 struct mwl_node *mn = MWL_NODE(ni);
2360 DPRINTF(sc, MWL_DEBUG_NODE, "%s: ni %p vap %p staid %d\n",
2361 __func__, ni, ni->ni_vap, mn->mn_staid);
2363 /* NB: call up first to age out ampdu q's */
2364 sc->sc_node_drain(ni);
2366 /* XXX better to not check low water mark? */
2367 if (sc->sc_rxblocked && mn->mn_staid != 0 &&
2368 (ni->ni_flags & IEEE80211_NODE_HT)) {
2371 * Walk the reorder q and reclaim rx dma buffers by copying
2372 * the packet contents into clusters.
2374 for (tid = 0; tid < WME_NUM_TID; tid++) {
2375 struct ieee80211_rx_ampdu *rap;
2377 rap = &ni->ni_rx_ampdu[tid];
2378 if ((rap->rxa_flags & IEEE80211_AGGR_XCHGPEND) == 0)
2380 if (rap->rxa_qframes)
2381 mwl_ampdu_rxdma_reclaim(rap);
2387 mwl_node_getsignal(const struct ieee80211_node *ni, int8_t *rssi, int8_t *noise)
2389 *rssi = ni->ni_ic->ic_node_getrssi(ni);
2390 #ifdef MWL_ANT_INFO_SUPPORT
2392 /* XXX need to smooth data */
2393 *noise = -MWL_NODE_CONST(ni)->mn_ai.nf;
2395 *noise = -95; /* XXX */
2398 *noise = -95; /* XXX */
2403 * Convert Hardware per-antenna rssi info to common format:
2404 * Let a1, a2, a3 represent the amplitudes per chain
2405 * Let amax represent max[a1, a2, a3]
2406 * Rssi1_dBm = RSSI_dBm + 20*log10(a1/amax)
2407 * Rssi1_dBm = RSSI_dBm + 20*log10(a1) - 20*log10(amax)
2408 * We store a table that is 4*20*log10(idx) - the extra 4 is to store or
2409 * maintain some extra precision.
2411 * Values are stored in .5 db format capped at 127.
2414 mwl_node_getmimoinfo(const struct ieee80211_node *ni,
2415 struct ieee80211_mimo_info *mi)
2417 #define CVT(_dst, _src) do { \
2418 (_dst) = rssi + ((logdbtbl[_src] - logdbtbl[rssi_max]) >> 2); \
2419 (_dst) = (_dst) > 64 ? 127 : ((_dst) << 1); \
2421 static const int8_t logdbtbl[32] = {
2422 0, 0, 24, 38, 48, 56, 62, 68,
2423 72, 76, 80, 83, 86, 89, 92, 94,
2424 96, 98, 100, 102, 104, 106, 107, 109,
2425 110, 112, 113, 115, 116, 117, 118, 119
2427 const struct mwl_node *mn = MWL_NODE_CONST(ni);
2428 uint8_t rssi = mn->mn_ai.rsvd1/2; /* XXX */
2431 rssi_max = mn->mn_ai.rssi_a;
2432 if (mn->mn_ai.rssi_b > rssi_max)
2433 rssi_max = mn->mn_ai.rssi_b;
2434 if (mn->mn_ai.rssi_c > rssi_max)
2435 rssi_max = mn->mn_ai.rssi_c;
2437 CVT(mi->ch[0].rssi[0], mn->mn_ai.rssi_a);
2438 CVT(mi->ch[1].rssi[0], mn->mn_ai.rssi_b);
2439 CVT(mi->ch[2].rssi[0], mn->mn_ai.rssi_c);
2441 mi->ch[0].noise[0] = mn->mn_ai.nf_a;
2442 mi->ch[1].noise[0] = mn->mn_ai.nf_b;
2443 mi->ch[2].noise[0] = mn->mn_ai.nf_c;
2447 static __inline void *
2448 mwl_getrxdma(struct mwl_softc *sc)
2450 struct mwl_jumbo *buf;
2454 * Allocate from jumbo pool.
2456 MWL_RXFREE_LOCK(sc);
2457 buf = SLIST_FIRST(&sc->sc_rxfree);
2459 DPRINTF(sc, MWL_DEBUG_ANY,
2460 "%s: out of rx dma buffers\n", __func__);
2461 sc->sc_stats.mst_rx_nodmabuf++;
2464 SLIST_REMOVE_HEAD(&sc->sc_rxfree, next);
2466 data = MWL_JUMBO_BUF2DATA(buf);
2468 MWL_RXFREE_UNLOCK(sc);
2472 static __inline void
2473 mwl_putrxdma(struct mwl_softc *sc, void *data)
2475 struct mwl_jumbo *buf;
2477 /* XXX bounds check data */
2478 MWL_RXFREE_LOCK(sc);
2479 buf = MWL_JUMBO_DATA2BUF(data);
2480 SLIST_INSERT_HEAD(&sc->sc_rxfree, buf, next);
2482 MWL_RXFREE_UNLOCK(sc);
2486 mwl_rxbuf_init(struct mwl_softc *sc, struct mwl_rxbuf *bf)
2488 struct mwl_rxdesc *ds;
2491 if (bf->bf_data == NULL) {
2492 bf->bf_data = mwl_getrxdma(sc);
2493 if (bf->bf_data == NULL) {
2494 /* mark descriptor to be skipped */
2495 ds->RxControl = EAGLE_RXD_CTRL_OS_OWN;
2496 /* NB: don't need PREREAD */
2497 MWL_RXDESC_SYNC(sc, ds, BUS_DMASYNC_PREWRITE);
2498 sc->sc_stats.mst_rxbuf_failed++;
2503 * NB: DMA buffer contents is known to be unmodified
2504 * so there's no need to flush the data cache.
2512 ds->Status = EAGLE_RXD_STATUS_IDLE;
2514 ds->PktLen = htole16(MWL_AGGR_SIZE);
2516 ds->pPhysBuffData = htole32(MWL_JUMBO_DMA_ADDR(sc, bf->bf_data));
2517 /* NB: don't touch pPhysNext, set once */
2518 ds->RxControl = EAGLE_RXD_CTRL_DRIVER_OWN;
2519 MWL_RXDESC_SYNC(sc, ds, BUS_DMASYNC_PREREAD | BUS_DMASYNC_PREWRITE);
2525 mwl_ext_free(struct mbuf *m, void *data, void *arg)
2527 struct mwl_softc *sc = arg;
2529 /* XXX bounds check data */
2530 mwl_putrxdma(sc, data);
2532 * If we were previously blocked by a lack of rx dma buffers
2533 * check if we now have enough to restart rx interrupt handling.
2534 * NB: we know we are called at splvm which is above splnet.
2536 if (sc->sc_rxblocked && sc->sc_nrxfree > mwl_rxdmalow) {
2537 sc->sc_rxblocked = 0;
2538 mwl_hal_intrset(sc->sc_mh, sc->sc_imask);
2542 struct mwl_frame_bar {
2545 u_int8_t i_ra[IEEE80211_ADDR_LEN];
2546 u_int8_t i_ta[IEEE80211_ADDR_LEN];
2551 * Like ieee80211_anyhdrsize, but handles BAR frames
2552 * specially so the logic below to piece the 802.11
2553 * header together works.
2556 mwl_anyhdrsize(const void *data)
2558 const struct ieee80211_frame *wh = data;
2560 if ((wh->i_fc[0]&IEEE80211_FC0_TYPE_MASK) == IEEE80211_FC0_TYPE_CTL) {
2561 switch (wh->i_fc[0] & IEEE80211_FC0_SUBTYPE_MASK) {
2562 case IEEE80211_FC0_SUBTYPE_CTS:
2563 case IEEE80211_FC0_SUBTYPE_ACK:
2564 return sizeof(struct ieee80211_frame_ack);
2565 case IEEE80211_FC0_SUBTYPE_BAR:
2566 return sizeof(struct mwl_frame_bar);
2568 return sizeof(struct ieee80211_frame_min);
2570 return ieee80211_hdrsize(data);
2574 mwl_handlemicerror(struct ieee80211com *ic, const uint8_t *data)
2576 const struct ieee80211_frame *wh;
2577 struct ieee80211_node *ni;
2579 wh = (const struct ieee80211_frame *)(data + sizeof(uint16_t));
2580 ni = ieee80211_find_rxnode(ic, (const struct ieee80211_frame_min *) wh);
2582 ieee80211_notify_michael_failure(ni->ni_vap, wh, 0);
2583 ieee80211_free_node(ni);
2588 * Convert hardware signal strength to rssi. The value
2589 * provided by the device has the noise floor added in;
2590 * we need to compensate for this but we don't have that
2591 * so we use a fixed value.
2593 * The offset of 8 is good for both 2.4 and 5GHz. The LNA
2594 * offset is already set as part of the initial gain. This
2595 * will give at least +/- 3dB for 2.4GHz and +/- 5dB for 5GHz.
2598 cvtrssi(uint8_t ssi)
2600 int rssi = (int) ssi + 8;
2601 /* XXX hack guess until we have a real noise floor */
2602 rssi = 2*(87 - rssi); /* NB: .5 dBm units */
2603 return (rssi < 0 ? 0 : rssi > 127 ? 127 : rssi);
2607 mwl_rx_proc(void *arg, int npending)
2609 struct mwl_softc *sc = arg;
2610 struct ieee80211com *ic = &sc->sc_ic;
2611 struct mwl_rxbuf *bf;
2612 struct mwl_rxdesc *ds;
2614 struct ieee80211_qosframe *wh;
2615 struct ieee80211_qosframe_addr4 *wh4;
2616 struct ieee80211_node *ni;
2617 struct mwl_node *mn;
2618 int off, len, hdrlen, pktlen, rssi, ntodo;
2619 uint8_t *data, status;
2623 DPRINTF(sc, MWL_DEBUG_RX_PROC, "%s: pending %u rdptr 0x%x wrptr 0x%x\n",
2624 __func__, npending, RD4(sc, sc->sc_hwspecs.rxDescRead),
2625 RD4(sc, sc->sc_hwspecs.rxDescWrite));
2628 for (ntodo = mwl_rxquota; ntodo > 0; ntodo--) {
2630 bf = STAILQ_FIRST(&sc->sc_rxbuf);
2635 * If data allocation failed previously there
2636 * will be no buffer; try again to re-populate it.
2637 * Note the firmware will not advance to the next
2638 * descriptor with a dma buffer so we must mimic
2639 * this or we'll get out of sync.
2641 DPRINTF(sc, MWL_DEBUG_ANY,
2642 "%s: rx buf w/o dma memory\n", __func__);
2643 (void) mwl_rxbuf_init(sc, bf);
2644 sc->sc_stats.mst_rx_dmabufmissing++;
2647 MWL_RXDESC_SYNC(sc, ds,
2648 BUS_DMASYNC_POSTREAD | BUS_DMASYNC_POSTWRITE);
2649 if (ds->RxControl != EAGLE_RXD_CTRL_DMA_OWN)
2652 if (sc->sc_debug & MWL_DEBUG_RECV_DESC)
2653 mwl_printrxbuf(bf, 0);
2655 status = ds->Status;
2656 if (status & EAGLE_RXD_STATUS_DECRYPT_ERR_MASK) {
2657 counter_u64_add(ic->ic_ierrors, 1);
2658 sc->sc_stats.mst_rx_crypto++;
2660 * NB: Check EAGLE_RXD_STATUS_GENERAL_DECRYPT_ERR
2661 * for backwards compatibility.
2663 if (status != EAGLE_RXD_STATUS_GENERAL_DECRYPT_ERR &&
2664 (status & EAGLE_RXD_STATUS_TKIP_MIC_DECRYPT_ERR)) {
2666 * MIC error, notify upper layers.
2668 bus_dmamap_sync(sc->sc_rxdmat, sc->sc_rxmap,
2669 BUS_DMASYNC_POSTREAD);
2670 mwl_handlemicerror(ic, data);
2671 sc->sc_stats.mst_rx_tkipmic++;
2673 /* XXX too painful to tap packets */
2677 * Sync the data buffer.
2679 len = le16toh(ds->PktLen);
2680 bus_dmamap_sync(sc->sc_rxdmat, sc->sc_rxmap, BUS_DMASYNC_POSTREAD);
2682 * The 802.11 header is provided all or in part at the front;
2683 * use it to calculate the true size of the header that we'll
2684 * construct below. We use this to figure out where to copy
2685 * payload prior to constructing the header.
2687 hdrlen = mwl_anyhdrsize(data + sizeof(uint16_t));
2688 off = sizeof(uint16_t) + sizeof(struct ieee80211_frame_addr4);
2690 /* calculate rssi early so we can re-use for each aggregate */
2691 rssi = cvtrssi(ds->RSSI);
2693 pktlen = hdrlen + (len - off);
2695 * NB: we know our frame is at least as large as
2696 * IEEE80211_MIN_LEN because there is a 4-address
2697 * frame at the front. Hence there's no need to
2698 * vet the packet length. If the frame in fact
2699 * is too small it should be discarded at the
2704 * Attach dma buffer to an mbuf. We tried
2705 * doing this based on the packet size (i.e.
2706 * copying small packets) but it turns out to
2707 * be a net loss. The tradeoff might be system
2708 * dependent (cache architecture is important).
2710 MGETHDR(m, M_NOWAIT, MT_DATA);
2712 DPRINTF(sc, MWL_DEBUG_ANY,
2713 "%s: no rx mbuf\n", __func__);
2714 sc->sc_stats.mst_rx_nombuf++;
2718 * Acquire the replacement dma buffer before
2719 * processing the frame. If we're out of dma
2720 * buffers we disable rx interrupts and wait
2721 * for the free pool to reach mlw_rxdmalow buffers
2722 * before starting to do work again. If the firmware
2723 * runs out of descriptors then it will toss frames
2724 * which is better than our doing it as that can
2725 * starve our processing. It is also important that
2726 * we always process rx'd frames in case they are
2727 * A-MPDU as otherwise the host's view of the BA
2728 * window may get out of sync with the firmware.
2730 newdata = mwl_getrxdma(sc);
2731 if (newdata == NULL) {
2732 /* NB: stat+msg in mwl_getrxdma */
2734 /* disable RX interrupt and mark state */
2735 mwl_hal_intrset(sc->sc_mh,
2736 sc->sc_imask &~ MACREG_A2HRIC_BIT_RX_RDY);
2737 sc->sc_rxblocked = 1;
2738 ieee80211_drain(ic);
2739 /* XXX check rxblocked and immediately start again? */
2742 bf->bf_data = newdata;
2744 * Attach the dma buffer to the mbuf;
2745 * mwl_rxbuf_init will re-setup the rx
2746 * descriptor using the replacement dma
2747 * buffer we just installed above.
2749 MEXTADD(m, data, MWL_AGGR_SIZE, mwl_ext_free,
2750 data, sc, 0, EXT_NET_DRV);
2751 m->m_data += off - hdrlen;
2752 m->m_pkthdr.len = m->m_len = pktlen;
2753 /* NB: dma buffer assumed read-only */
2756 * Piece 802.11 header together.
2758 wh = mtod(m, struct ieee80211_qosframe *);
2759 /* NB: don't need to do this sometimes but ... */
2760 /* XXX special case so we can memcpy after m_devget? */
2761 ovbcopy(data + sizeof(uint16_t), wh, hdrlen);
2762 if (IEEE80211_QOS_HAS_SEQ(wh)) {
2763 if (IEEE80211_IS_DSTODS(wh)) {
2765 struct ieee80211_qosframe_addr4*);
2766 *(uint16_t *)wh4->i_qos = ds->QosCtrl;
2768 *(uint16_t *)wh->i_qos = ds->QosCtrl;
2772 * The f/w strips WEP header but doesn't clear
2773 * the WEP bit; mark the packet with M_WEP so
2774 * net80211 will treat the data as decrypted.
2775 * While here also clear the PWR_MGT bit since
2776 * power save is handled by the firmware and
2777 * passing this up will potentially cause the
2778 * upper layer to put a station in power save
2779 * (except when configured with MWL_HOST_PS_SUPPORT).
2781 if (wh->i_fc[1] & IEEE80211_FC1_PROTECTED)
2782 m->m_flags |= M_WEP;
2783 #ifdef MWL_HOST_PS_SUPPORT
2784 wh->i_fc[1] &= ~IEEE80211_FC1_PROTECTED;
2786 wh->i_fc[1] &= ~(IEEE80211_FC1_PROTECTED |
2787 IEEE80211_FC1_PWR_MGT);
2790 if (ieee80211_radiotap_active(ic)) {
2791 struct mwl_rx_radiotap_header *tap = &sc->sc_rx_th;
2794 tap->wr_rate = ds->Rate;
2795 tap->wr_antsignal = rssi + nf;
2796 tap->wr_antnoise = nf;
2798 if (IFF_DUMPPKTS_RECV(sc, wh)) {
2799 ieee80211_dump_pkt(ic, mtod(m, caddr_t),
2800 len, ds->Rate, rssi);
2803 ni = ieee80211_find_rxnode(ic,
2804 (const struct ieee80211_frame_min *) wh);
2807 #ifdef MWL_ANT_INFO_SUPPORT
2808 mn->mn_ai.rssi_a = ds->ai.rssi_a;
2809 mn->mn_ai.rssi_b = ds->ai.rssi_b;
2810 mn->mn_ai.rssi_c = ds->ai.rssi_c;
2811 mn->mn_ai.rsvd1 = rssi;
2813 /* tag AMPDU aggregates for reorder processing */
2814 if (ni->ni_flags & IEEE80211_NODE_HT)
2815 m->m_flags |= M_AMPDU;
2816 (void) ieee80211_input(ni, m, rssi, nf);
2817 ieee80211_free_node(ni);
2819 (void) ieee80211_input_all(ic, m, rssi, nf);
2821 /* NB: ignore ENOMEM so we process more descriptors */
2822 (void) mwl_rxbuf_init(sc, bf);
2823 bf = STAILQ_NEXT(bf, bf_list);
2828 if (mbufq_first(&sc->sc_snd) != NULL) {
2829 /* NB: kick fw; the tx thread may have been preempted */
2830 mwl_hal_txstart(sc->sc_mh, 0);
2836 mwl_txq_init(struct mwl_softc *sc, struct mwl_txq *txq, int qnum)
2838 struct mwl_txbuf *bf, *bn;
2839 struct mwl_txdesc *ds;
2841 MWL_TXQ_LOCK_INIT(sc, txq);
2843 txq->txpri = 0; /* XXX */
2845 /* NB: q setup by mwl_txdma_setup XXX */
2846 STAILQ_INIT(&txq->free);
2848 STAILQ_FOREACH(bf, &txq->free, bf_list) {
2852 bn = STAILQ_NEXT(bf, bf_list);
2854 bn = STAILQ_FIRST(&txq->free);
2855 ds->pPhysNext = htole32(bn->bf_daddr);
2857 STAILQ_INIT(&txq->active);
2861 * Setup a hardware data transmit queue for the specified
2862 * access control. We record the mapping from ac's
2863 * to h/w queues for use by mwl_tx_start.
2866 mwl_tx_setup(struct mwl_softc *sc, int ac, int mvtype)
2868 struct mwl_txq *txq;
2870 if (ac >= nitems(sc->sc_ac2q)) {
2871 device_printf(sc->sc_dev, "AC %u out of range, max %zu!\n",
2872 ac, nitems(sc->sc_ac2q));
2875 if (mvtype >= MWL_NUM_TX_QUEUES) {
2876 device_printf(sc->sc_dev, "mvtype %u out of range, max %u!\n",
2877 mvtype, MWL_NUM_TX_QUEUES);
2880 txq = &sc->sc_txq[mvtype];
2881 mwl_txq_init(sc, txq, mvtype);
2882 sc->sc_ac2q[ac] = txq;
2887 * Update WME parameters for a transmit queue.
2890 mwl_txq_update(struct mwl_softc *sc, int ac)
2892 #define MWL_EXPONENT_TO_VALUE(v) ((1<<v)-1)
2893 struct ieee80211com *ic = &sc->sc_ic;
2894 struct mwl_txq *txq = sc->sc_ac2q[ac];
2895 struct wmeParams *wmep = &ic->ic_wme.wme_chanParams.cap_wmeParams[ac];
2896 struct mwl_hal *mh = sc->sc_mh;
2897 int aifs, cwmin, cwmax, txoplim;
2899 aifs = wmep->wmep_aifsn;
2900 /* XXX in sta mode need to pass log values for cwmin/max */
2901 cwmin = MWL_EXPONENT_TO_VALUE(wmep->wmep_logcwmin);
2902 cwmax = MWL_EXPONENT_TO_VALUE(wmep->wmep_logcwmax);
2903 txoplim = wmep->wmep_txopLimit; /* NB: units of 32us */
2905 if (mwl_hal_setedcaparams(mh, txq->qnum, cwmin, cwmax, aifs, txoplim)) {
2906 device_printf(sc->sc_dev, "unable to update hardware queue "
2907 "parameters for %s traffic!\n",
2908 ieee80211_wme_acnames[ac]);
2912 #undef MWL_EXPONENT_TO_VALUE
2916 * Callback from the 802.11 layer to update WME parameters.
2919 mwl_wme_update(struct ieee80211com *ic)
2921 struct mwl_softc *sc = ic->ic_softc;
2923 return !mwl_txq_update(sc, WME_AC_BE) ||
2924 !mwl_txq_update(sc, WME_AC_BK) ||
2925 !mwl_txq_update(sc, WME_AC_VI) ||
2926 !mwl_txq_update(sc, WME_AC_VO) ? EIO : 0;
2930 * Reclaim resources for a setup queue.
2933 mwl_tx_cleanupq(struct mwl_softc *sc, struct mwl_txq *txq)
2936 MWL_TXQ_LOCK_DESTROY(txq);
2940 * Reclaim all tx queue resources.
2943 mwl_tx_cleanup(struct mwl_softc *sc)
2947 for (i = 0; i < MWL_NUM_TX_QUEUES; i++)
2948 mwl_tx_cleanupq(sc, &sc->sc_txq[i]);
2952 mwl_tx_dmasetup(struct mwl_softc *sc, struct mwl_txbuf *bf, struct mbuf *m0)
2958 * Load the DMA map so any coalescing is done. This
2959 * also calculates the number of descriptors we need.
2961 error = bus_dmamap_load_mbuf_sg(sc->sc_dmat, bf->bf_dmamap, m0,
2962 bf->bf_segs, &bf->bf_nseg,
2964 if (error == EFBIG) {
2965 /* XXX packet requires too many descriptors */
2966 bf->bf_nseg = MWL_TXDESC+1;
2967 } else if (error != 0) {
2968 sc->sc_stats.mst_tx_busdma++;
2973 * Discard null packets and check for packets that
2974 * require too many TX descriptors. We try to convert
2975 * the latter to a cluster.
2977 if (error == EFBIG) { /* too many desc's, linearize */
2978 sc->sc_stats.mst_tx_linear++;
2980 m = m_collapse(m0, M_NOWAIT, MWL_TXDESC);
2982 m = m_defrag(m0, M_NOWAIT);
2986 sc->sc_stats.mst_tx_nombuf++;
2990 error = bus_dmamap_load_mbuf_sg(sc->sc_dmat, bf->bf_dmamap, m0,
2991 bf->bf_segs, &bf->bf_nseg,
2994 sc->sc_stats.mst_tx_busdma++;
2998 KASSERT(bf->bf_nseg <= MWL_TXDESC,
2999 ("too many segments after defrag; nseg %u", bf->bf_nseg));
3000 } else if (bf->bf_nseg == 0) { /* null packet, discard */
3001 sc->sc_stats.mst_tx_nodata++;
3005 DPRINTF(sc, MWL_DEBUG_XMIT, "%s: m %p len %u\n",
3006 __func__, m0, m0->m_pkthdr.len);
3007 bus_dmamap_sync(sc->sc_dmat, bf->bf_dmamap, BUS_DMASYNC_PREWRITE);
3014 mwl_cvtlegacyrate(int rate)
3035 * Calculate fixed tx rate information per client state;
3036 * this value is suitable for writing to the Format field
3037 * of a tx descriptor.
3040 mwl_calcformat(uint8_t rate, const struct ieee80211_node *ni)
3044 fmt = SM(3, EAGLE_TXD_ANTENNA)
3045 | (IEEE80211_IS_CHAN_HT40D(ni->ni_chan) ?
3046 EAGLE_TXD_EXTCHAN_LO : EAGLE_TXD_EXTCHAN_HI);
3047 if (rate & IEEE80211_RATE_MCS) { /* HT MCS */
3048 fmt |= EAGLE_TXD_FORMAT_HT
3049 /* NB: 0x80 implicitly stripped from ucastrate */
3050 | SM(rate, EAGLE_TXD_RATE);
3051 /* XXX short/long GI may be wrong; re-check */
3052 if (IEEE80211_IS_CHAN_HT40(ni->ni_chan)) {
3053 fmt |= EAGLE_TXD_CHW_40
3054 | (ni->ni_htcap & IEEE80211_HTCAP_SHORTGI40 ?
3055 EAGLE_TXD_GI_SHORT : EAGLE_TXD_GI_LONG);
3057 fmt |= EAGLE_TXD_CHW_20
3058 | (ni->ni_htcap & IEEE80211_HTCAP_SHORTGI20 ?
3059 EAGLE_TXD_GI_SHORT : EAGLE_TXD_GI_LONG);
3061 } else { /* legacy rate */
3062 fmt |= EAGLE_TXD_FORMAT_LEGACY
3063 | SM(mwl_cvtlegacyrate(rate), EAGLE_TXD_RATE)
3065 /* XXX iv_flags & IEEE80211_F_SHPREAMBLE? */
3066 | (ni->ni_capinfo & IEEE80211_CAPINFO_SHORT_PREAMBLE ?
3067 EAGLE_TXD_PREAMBLE_SHORT : EAGLE_TXD_PREAMBLE_LONG);
3073 mwl_tx_start(struct mwl_softc *sc, struct ieee80211_node *ni, struct mwl_txbuf *bf,
3076 struct ieee80211com *ic = &sc->sc_ic;
3077 struct ieee80211vap *vap = ni->ni_vap;
3078 int error, iswep, ismcast;
3079 int hdrlen, copyhdrlen, pktlen;
3080 struct mwl_txdesc *ds;
3081 struct mwl_txq *txq;
3082 struct ieee80211_frame *wh;
3083 struct mwltxrec *tr;
3084 struct mwl_node *mn;
3090 wh = mtod(m0, struct ieee80211_frame *);
3091 iswep = wh->i_fc[1] & IEEE80211_FC1_PROTECTED;
3092 ismcast = IEEE80211_IS_MULTICAST(wh->i_addr1);
3093 hdrlen = ieee80211_anyhdrsize(wh);
3094 copyhdrlen = hdrlen;
3095 pktlen = m0->m_pkthdr.len;
3096 if (IEEE80211_QOS_HAS_SEQ(wh)) {
3097 if (IEEE80211_IS_DSTODS(wh)) {
3099 (((struct ieee80211_qosframe_addr4 *) wh)->i_qos);
3100 copyhdrlen -= sizeof(qos);
3103 (((struct ieee80211_qosframe *) wh)->i_qos);
3108 const struct ieee80211_cipher *cip;
3109 struct ieee80211_key *k;
3112 * Construct the 802.11 header+trailer for an encrypted
3113 * frame. The only reason this can fail is because of an
3114 * unknown or unsupported cipher/key type.
3116 * NB: we do this even though the firmware will ignore
3117 * what we've done for WEP and TKIP as we need the
3118 * ExtIV filled in for CCMP and this also adjusts
3119 * the headers which simplifies our work below.
3121 k = ieee80211_crypto_encap(ni, m0);
3124 * This can happen when the key is yanked after the
3125 * frame was queued. Just discard the frame; the
3126 * 802.11 layer counts failures and provides
3127 * debugging/diagnostics.
3133 * Adjust the packet length for the crypto additions
3134 * done during encap and any other bits that the f/w
3135 * will add later on.
3138 pktlen += cip->ic_header + cip->ic_miclen + cip->ic_trailer;
3140 /* packet header may have moved, reset our local pointer */
3141 wh = mtod(m0, struct ieee80211_frame *);
3144 if (ieee80211_radiotap_active_vap(vap)) {
3145 sc->sc_tx_th.wt_flags = 0; /* XXX */
3147 sc->sc_tx_th.wt_flags |= IEEE80211_RADIOTAP_F_WEP;
3149 sc->sc_tx_th.wt_rate = ds->DataRate;
3151 sc->sc_tx_th.wt_txpower = ni->ni_txpower;
3152 sc->sc_tx_th.wt_antenna = sc->sc_txantenna;
3154 ieee80211_radiotap_tx(vap, m0);
3157 * Copy up/down the 802.11 header; the firmware requires
3158 * we present a 2-byte payload length followed by a
3159 * 4-address header (w/o QoS), followed (optionally) by
3160 * any WEP/ExtIV header (but only filled in for CCMP).
3161 * We are assured the mbuf has sufficient headroom to
3162 * prepend in-place by the setup of ic_headroom in
3165 if (hdrlen < sizeof(struct mwltxrec)) {
3166 const int space = sizeof(struct mwltxrec) - hdrlen;
3167 if (M_LEADINGSPACE(m0) < space) {
3168 /* NB: should never happen */
3169 device_printf(sc->sc_dev,
3170 "not enough headroom, need %d found %zd, "
3171 "m_flags 0x%x m_len %d\n",
3172 space, M_LEADINGSPACE(m0), m0->m_flags, m0->m_len);
3173 ieee80211_dump_pkt(ic,
3174 mtod(m0, const uint8_t *), m0->m_len, 0, -1);
3176 sc->sc_stats.mst_tx_noheadroom++;
3179 M_PREPEND(m0, space, M_NOWAIT);
3181 tr = mtod(m0, struct mwltxrec *);
3182 if (wh != (struct ieee80211_frame *) &tr->wh)
3183 ovbcopy(wh, &tr->wh, hdrlen);
3185 * Note: the "firmware length" is actually the length
3186 * of the fully formed "802.11 payload". That is, it's
3187 * everything except for the 802.11 header. In particular
3188 * this includes all crypto material including the MIC!
3190 tr->fwlen = htole16(pktlen - hdrlen);
3193 * Load the DMA map so any coalescing is done. This
3194 * also calculates the number of descriptors we need.
3196 error = mwl_tx_dmasetup(sc, bf, m0);
3198 /* NB: stat collected in mwl_tx_dmasetup */
3199 DPRINTF(sc, MWL_DEBUG_XMIT,
3200 "%s: unable to setup dma\n", __func__);
3203 bf->bf_node = ni; /* NB: held reference */
3204 m0 = bf->bf_m; /* NB: may have changed */
3205 tr = mtod(m0, struct mwltxrec *);
3206 wh = (struct ieee80211_frame *)&tr->wh;
3209 * Formulate tx descriptor.
3214 ds->QosCtrl = qos; /* NB: already little-endian */
3217 * NB: multiframes should be zero because the descriptors
3218 * are initialized to zero. This should handle the case
3219 * where the driver is built with MWL_TXDESC=1 but we are
3220 * using firmware with multi-segment support.
3222 ds->PktPtr = htole32(bf->bf_segs[0].ds_addr);
3223 ds->PktLen = htole16(bf->bf_segs[0].ds_len);
3225 ds->multiframes = htole32(bf->bf_nseg);
3226 ds->PktLen = htole16(m0->m_pkthdr.len);
3227 for (i = 0; i < bf->bf_nseg; i++) {
3228 ds->PktPtrArray[i] = htole32(bf->bf_segs[i].ds_addr);
3229 ds->PktLenArray[i] = htole16(bf->bf_segs[i].ds_len);
3232 /* NB: pPhysNext, DataRate, and SapPktInfo setup once, don't touch */
3235 ds->ack_wcb_addr = 0;
3239 * Select transmit rate.
3241 switch (wh->i_fc[0] & IEEE80211_FC0_TYPE_MASK) {
3242 case IEEE80211_FC0_TYPE_MGT:
3243 sc->sc_stats.mst_tx_mgmt++;
3245 case IEEE80211_FC0_TYPE_CTL:
3246 /* NB: assign to BE q to avoid bursting */
3247 ds->TxPriority = MWL_WME_AC_BE;
3249 case IEEE80211_FC0_TYPE_DATA:
3251 const struct ieee80211_txparam *tp = ni->ni_txparms;
3253 * EAPOL frames get forced to a fixed rate and w/o
3254 * aggregation; otherwise check for any fixed rate
3255 * for the client (may depend on association state).
3257 if (m0->m_flags & M_EAPOL) {
3258 const struct mwl_vap *mvp = MWL_VAP_CONST(vap);
3259 ds->Format = mvp->mv_eapolformat;
3261 EAGLE_TXD_FIXED_RATE | EAGLE_TXD_DONT_AGGR);
3262 } else if (tp->ucastrate != IEEE80211_FIXED_RATE_NONE) {
3263 /* XXX pre-calculate per node */
3264 ds->Format = htole16(
3265 mwl_calcformat(tp->ucastrate, ni));
3266 ds->pad = htole16(EAGLE_TXD_FIXED_RATE);
3268 /* NB: EAPOL frames will never have qos set */
3270 ds->TxPriority = txq->qnum;
3272 else if (mwl_bastream_match(&mn->mn_ba[3], qos))
3273 ds->TxPriority = mn->mn_ba[3].txq;
3276 else if (mwl_bastream_match(&mn->mn_ba[2], qos))
3277 ds->TxPriority = mn->mn_ba[2].txq;
3280 else if (mwl_bastream_match(&mn->mn_ba[1], qos))
3281 ds->TxPriority = mn->mn_ba[1].txq;
3284 else if (mwl_bastream_match(&mn->mn_ba[0], qos))
3285 ds->TxPriority = mn->mn_ba[0].txq;
3288 ds->TxPriority = txq->qnum;
3290 ds->TxPriority = txq->qnum;
3293 device_printf(sc->sc_dev, "bogus frame type 0x%x (%s)\n",
3294 wh->i_fc[0] & IEEE80211_FC0_TYPE_MASK, __func__);
3295 sc->sc_stats.mst_tx_badframetype++;
3300 if (IFF_DUMPPKTS_XMIT(sc))
3301 ieee80211_dump_pkt(ic,
3302 mtod(m0, const uint8_t *)+sizeof(uint16_t),
3303 m0->m_len - sizeof(uint16_t), ds->DataRate, -1);
3306 ds->Status = htole32(EAGLE_TXD_STATUS_FW_OWNED);
3307 STAILQ_INSERT_TAIL(&txq->active, bf, bf_list);
3308 MWL_TXDESC_SYNC(txq, ds, BUS_DMASYNC_PREREAD | BUS_DMASYNC_PREWRITE);
3310 sc->sc_tx_timer = 5;
3311 MWL_TXQ_UNLOCK(txq);
3317 mwl_cvtlegacyrix(int rix)
3319 static const int ieeerates[] =
3320 { 2, 4, 11, 22, 44, 12, 18, 24, 36, 48, 72, 96, 108 };
3321 return (rix < nitems(ieeerates) ? ieeerates[rix] : 0);
3325 * Process completed xmit descriptors from the specified queue.
3328 mwl_tx_processq(struct mwl_softc *sc, struct mwl_txq *txq)
3330 #define EAGLE_TXD_STATUS_MCAST \
3331 (EAGLE_TXD_STATUS_MULTICAST_TX | EAGLE_TXD_STATUS_BROADCAST_TX)
3332 struct ieee80211com *ic = &sc->sc_ic;
3333 struct mwl_txbuf *bf;
3334 struct mwl_txdesc *ds;
3335 struct ieee80211_node *ni;
3336 struct mwl_node *an;
3340 DPRINTF(sc, MWL_DEBUG_TX_PROC, "%s: tx queue %u\n", __func__, txq->qnum);
3341 for (nreaped = 0;; nreaped++) {
3343 bf = STAILQ_FIRST(&txq->active);
3345 MWL_TXQ_UNLOCK(txq);
3349 MWL_TXDESC_SYNC(txq, ds,
3350 BUS_DMASYNC_POSTREAD | BUS_DMASYNC_POSTWRITE);
3351 if (ds->Status & htole32(EAGLE_TXD_STATUS_FW_OWNED)) {
3352 MWL_TXQ_UNLOCK(txq);
3355 STAILQ_REMOVE_HEAD(&txq->active, bf_list);
3356 MWL_TXQ_UNLOCK(txq);
3359 if (sc->sc_debug & MWL_DEBUG_XMIT_DESC)
3360 mwl_printtxbuf(bf, txq->qnum, nreaped);
3365 status = le32toh(ds->Status);
3366 if (status & EAGLE_TXD_STATUS_OK) {
3367 uint16_t Format = le16toh(ds->Format);
3368 uint8_t txant = MS(Format, EAGLE_TXD_ANTENNA);
3370 sc->sc_stats.mst_ant_tx[txant]++;
3371 if (status & EAGLE_TXD_STATUS_OK_RETRY)
3372 sc->sc_stats.mst_tx_retries++;
3373 if (status & EAGLE_TXD_STATUS_OK_MORE_RETRY)
3374 sc->sc_stats.mst_tx_mretries++;
3375 if (txq->qnum >= MWL_WME_AC_VO)
3376 ic->ic_wme.wme_hipri_traffic++;
3377 ni->ni_txrate = MS(Format, EAGLE_TXD_RATE);
3378 if ((Format & EAGLE_TXD_FORMAT_HT) == 0) {
3379 ni->ni_txrate = mwl_cvtlegacyrix(
3382 ni->ni_txrate |= IEEE80211_RATE_MCS;
3383 sc->sc_stats.mst_tx_rate = ni->ni_txrate;
3385 if (status & EAGLE_TXD_STATUS_FAILED_LINK_ERROR)
3386 sc->sc_stats.mst_tx_linkerror++;
3387 if (status & EAGLE_TXD_STATUS_FAILED_XRETRY)
3388 sc->sc_stats.mst_tx_xretries++;
3389 if (status & EAGLE_TXD_STATUS_FAILED_AGING)
3390 sc->sc_stats.mst_tx_aging++;
3391 if (bf->bf_m->m_flags & M_FF)
3392 sc->sc_stats.mst_ff_txerr++;
3394 if (bf->bf_m->m_flags & M_TXCB)
3395 /* XXX strip fw len in case header inspected */
3396 m_adj(bf->bf_m, sizeof(uint16_t));
3397 ieee80211_tx_complete(ni, bf->bf_m,
3398 (status & EAGLE_TXD_STATUS_OK) == 0);
3401 ds->Status = htole32(EAGLE_TXD_STATUS_IDLE);
3403 bus_dmamap_sync(sc->sc_dmat, bf->bf_dmamap,
3404 BUS_DMASYNC_POSTWRITE);
3405 bus_dmamap_unload(sc->sc_dmat, bf->bf_dmamap);
3407 mwl_puttxbuf_tail(txq, bf);
3410 #undef EAGLE_TXD_STATUS_MCAST
3414 * Deferred processing of transmit interrupt; special-cased
3415 * for four hardware queues, 0-3.
3418 mwl_tx_proc(void *arg, int npending)
3420 struct mwl_softc *sc = arg;
3424 * Process each active queue.
3427 if (!STAILQ_EMPTY(&sc->sc_txq[0].active))
3428 nreaped += mwl_tx_processq(sc, &sc->sc_txq[0]);
3429 if (!STAILQ_EMPTY(&sc->sc_txq[1].active))
3430 nreaped += mwl_tx_processq(sc, &sc->sc_txq[1]);
3431 if (!STAILQ_EMPTY(&sc->sc_txq[2].active))
3432 nreaped += mwl_tx_processq(sc, &sc->sc_txq[2]);
3433 if (!STAILQ_EMPTY(&sc->sc_txq[3].active))
3434 nreaped += mwl_tx_processq(sc, &sc->sc_txq[3]);
3437 sc->sc_tx_timer = 0;
3438 if (mbufq_first(&sc->sc_snd) != NULL) {
3439 /* NB: kick fw; the tx thread may have been preempted */
3440 mwl_hal_txstart(sc->sc_mh, 0);
3447 mwl_tx_draintxq(struct mwl_softc *sc, struct mwl_txq *txq)
3449 struct ieee80211_node *ni;
3450 struct mwl_txbuf *bf;
3454 * NB: this assumes output has been stopped and
3455 * we do not need to block mwl_tx_tasklet
3457 for (ix = 0;; ix++) {
3459 bf = STAILQ_FIRST(&txq->active);
3461 MWL_TXQ_UNLOCK(txq);
3464 STAILQ_REMOVE_HEAD(&txq->active, bf_list);
3465 MWL_TXQ_UNLOCK(txq);
3467 if (sc->sc_debug & MWL_DEBUG_RESET) {
3468 struct ieee80211com *ic = &sc->sc_ic;
3469 const struct mwltxrec *tr =
3470 mtod(bf->bf_m, const struct mwltxrec *);
3471 mwl_printtxbuf(bf, txq->qnum, ix);
3472 ieee80211_dump_pkt(ic, (const uint8_t *)&tr->wh,
3473 bf->bf_m->m_len - sizeof(tr->fwlen), 0, -1);
3475 #endif /* MWL_DEBUG */
3476 bus_dmamap_unload(sc->sc_dmat, bf->bf_dmamap);
3480 * Reclaim node reference.
3482 ieee80211_free_node(ni);
3486 mwl_puttxbuf_tail(txq, bf);
3491 * Drain the transmit queues and reclaim resources.
3494 mwl_draintxq(struct mwl_softc *sc)
3498 for (i = 0; i < MWL_NUM_TX_QUEUES; i++)
3499 mwl_tx_draintxq(sc, &sc->sc_txq[i]);
3500 sc->sc_tx_timer = 0;
3505 * Reset the transmit queues to a pristine state after a fw download.
3508 mwl_resettxq(struct mwl_softc *sc)
3512 for (i = 0; i < MWL_NUM_TX_QUEUES; i++)
3513 mwl_txq_reset(sc, &sc->sc_txq[i]);
3515 #endif /* MWL_DIAGAPI */
3518 * Clear the transmit queues of any frames submitted for the
3519 * specified vap. This is done when the vap is deleted so we
3520 * don't potentially reference the vap after it is gone.
3521 * Note we cannot remove the frames; we only reclaim the node
3525 mwl_cleartxq(struct mwl_softc *sc, struct ieee80211vap *vap)
3527 struct mwl_txq *txq;
3528 struct mwl_txbuf *bf;
3531 for (i = 0; i < MWL_NUM_TX_QUEUES; i++) {
3532 txq = &sc->sc_txq[i];
3534 STAILQ_FOREACH(bf, &txq->active, bf_list) {
3535 struct ieee80211_node *ni = bf->bf_node;
3536 if (ni != NULL && ni->ni_vap == vap) {
3538 ieee80211_free_node(ni);
3541 MWL_TXQ_UNLOCK(txq);
3546 mwl_recv_action(struct ieee80211_node *ni, const struct ieee80211_frame *wh,
3547 const uint8_t *frm, const uint8_t *efrm)
3549 struct mwl_softc *sc = ni->ni_ic->ic_softc;
3550 const struct ieee80211_action *ia;
3552 ia = (const struct ieee80211_action *) frm;
3553 if (ia->ia_category == IEEE80211_ACTION_CAT_HT &&
3554 ia->ia_action == IEEE80211_ACTION_HT_MIMOPWRSAVE) {
3555 const struct ieee80211_action_ht_mimopowersave *mps =
3556 (const struct ieee80211_action_ht_mimopowersave *) ia;
3558 mwl_hal_setmimops(sc->sc_mh, ni->ni_macaddr,
3559 mps->am_control & IEEE80211_A_HT_MIMOPWRSAVE_ENA,
3560 MS(mps->am_control, IEEE80211_A_HT_MIMOPWRSAVE_MODE));
3563 return sc->sc_recv_action(ni, wh, frm, efrm);
3567 mwl_addba_request(struct ieee80211_node *ni, struct ieee80211_tx_ampdu *tap,
3568 int dialogtoken, int baparamset, int batimeout)
3570 struct mwl_softc *sc = ni->ni_ic->ic_softc;
3571 struct ieee80211vap *vap = ni->ni_vap;
3572 struct mwl_node *mn = MWL_NODE(ni);
3573 struct mwl_bastate *bas;
3575 bas = tap->txa_private;
3577 const MWL_HAL_BASTREAM *sp;
3579 * Check for a free BA stream slot.
3582 if (mn->mn_ba[3].bastream == NULL)
3583 bas = &mn->mn_ba[3];
3587 if (mn->mn_ba[2].bastream == NULL)
3588 bas = &mn->mn_ba[2];
3592 if (mn->mn_ba[1].bastream == NULL)
3593 bas = &mn->mn_ba[1];
3597 if (mn->mn_ba[0].bastream == NULL)
3598 bas = &mn->mn_ba[0];
3602 /* sta already has max BA streams */
3603 /* XXX assign BA stream to highest priority tid */
3604 DPRINTF(sc, MWL_DEBUG_AMPDU,
3605 "%s: already has max bastreams\n", __func__);
3606 sc->sc_stats.mst_ampdu_reject++;
3609 /* NB: no held reference to ni */
3610 sp = mwl_hal_bastream_alloc(MWL_VAP(vap)->mv_hvap,
3611 (baparamset & IEEE80211_BAPS_POLICY_IMMEDIATE) != 0,
3612 ni->ni_macaddr, tap->txa_tid, ni->ni_htparam,
3616 * No available stream, return 0 so no
3617 * a-mpdu aggregation will be done.
3619 DPRINTF(sc, MWL_DEBUG_AMPDU,
3620 "%s: no bastream available\n", __func__);
3621 sc->sc_stats.mst_ampdu_nostream++;
3624 DPRINTF(sc, MWL_DEBUG_AMPDU, "%s: alloc bastream %p\n",
3626 /* NB: qos is left zero so we won't match in mwl_tx_start */
3628 tap->txa_private = bas;
3630 /* fetch current seq# from the firmware; if available */
3631 if (mwl_hal_bastream_get_seqno(sc->sc_mh, bas->bastream,
3632 vap->iv_opmode == IEEE80211_M_STA ? vap->iv_myaddr : ni->ni_macaddr,
3633 &tap->txa_start) != 0)
3635 return sc->sc_addba_request(ni, tap, dialogtoken, baparamset, batimeout);
3639 mwl_addba_response(struct ieee80211_node *ni, struct ieee80211_tx_ampdu *tap,
3640 int code, int baparamset, int batimeout)
3642 struct mwl_softc *sc = ni->ni_ic->ic_softc;
3643 struct mwl_bastate *bas;
3645 bas = tap->txa_private;
3647 /* XXX should not happen */
3648 DPRINTF(sc, MWL_DEBUG_AMPDU,
3649 "%s: no BA stream allocated, TID %d\n",
3650 __func__, tap->txa_tid);
3651 sc->sc_stats.mst_addba_nostream++;
3654 if (code == IEEE80211_STATUS_SUCCESS) {
3655 struct ieee80211vap *vap = ni->ni_vap;
3659 * Tell the firmware to setup the BA stream;
3660 * we know resources are available because we
3661 * pre-allocated one before forming the request.
3663 bufsiz = MS(baparamset, IEEE80211_BAPS_BUFSIZ);
3665 bufsiz = IEEE80211_AGGR_BAWMAX;
3666 error = mwl_hal_bastream_create(MWL_VAP(vap)->mv_hvap,
3667 bas->bastream, bufsiz, bufsiz, tap->txa_start);
3670 * Setup failed, return immediately so no a-mpdu
3671 * aggregation will be done.
3673 mwl_hal_bastream_destroy(sc->sc_mh, bas->bastream);
3674 mwl_bastream_free(bas);
3675 tap->txa_private = NULL;
3677 DPRINTF(sc, MWL_DEBUG_AMPDU,
3678 "%s: create failed, error %d, bufsiz %d TID %d "
3679 "htparam 0x%x\n", __func__, error, bufsiz,
3680 tap->txa_tid, ni->ni_htparam);
3681 sc->sc_stats.mst_bacreate_failed++;
3684 /* NB: cache txq to avoid ptr indirect */
3685 mwl_bastream_setup(bas, tap->txa_tid, bas->bastream->txq);
3686 DPRINTF(sc, MWL_DEBUG_AMPDU,
3687 "%s: bastream %p assigned to txq %d TID %d bufsiz %d "
3688 "htparam 0x%x\n", __func__, bas->bastream,
3689 bas->txq, tap->txa_tid, bufsiz, ni->ni_htparam);
3692 * Other side NAK'd us; return the resources.
3694 DPRINTF(sc, MWL_DEBUG_AMPDU,
3695 "%s: request failed with code %d, destroy bastream %p\n",
3696 __func__, code, bas->bastream);
3697 mwl_hal_bastream_destroy(sc->sc_mh, bas->bastream);
3698 mwl_bastream_free(bas);
3699 tap->txa_private = NULL;
3701 /* NB: firmware sends BAR so we don't need to */
3702 return sc->sc_addba_response(ni, tap, code, baparamset, batimeout);
3706 mwl_addba_stop(struct ieee80211_node *ni, struct ieee80211_tx_ampdu *tap)
3708 struct mwl_softc *sc = ni->ni_ic->ic_softc;
3709 struct mwl_bastate *bas;
3711 bas = tap->txa_private;
3713 DPRINTF(sc, MWL_DEBUG_AMPDU, "%s: destroy bastream %p\n",
3714 __func__, bas->bastream);
3715 mwl_hal_bastream_destroy(sc->sc_mh, bas->bastream);
3716 mwl_bastream_free(bas);
3717 tap->txa_private = NULL;
3719 sc->sc_addba_stop(ni, tap);
3723 * Setup the rx data structures. This should only be
3724 * done once or we may get out of sync with the firmware.
3727 mwl_startrecv(struct mwl_softc *sc)
3729 if (!sc->sc_recvsetup) {
3730 struct mwl_rxbuf *bf, *prev;
3731 struct mwl_rxdesc *ds;
3734 STAILQ_FOREACH(bf, &sc->sc_rxbuf, bf_list) {
3735 int error = mwl_rxbuf_init(sc, bf);
3737 DPRINTF(sc, MWL_DEBUG_RECV,
3738 "%s: mwl_rxbuf_init failed %d\n",
3744 ds->pPhysNext = htole32(bf->bf_daddr);
3751 htole32(STAILQ_FIRST(&sc->sc_rxbuf)->bf_daddr);
3753 sc->sc_recvsetup = 1;
3755 mwl_mode_init(sc); /* set filters, etc. */
3759 static MWL_HAL_APMODE
3760 mwl_getapmode(const struct ieee80211vap *vap, struct ieee80211_channel *chan)
3762 MWL_HAL_APMODE mode;
3764 if (IEEE80211_IS_CHAN_HT(chan)) {
3765 if (vap->iv_flags_ht & IEEE80211_FHT_PUREN)
3766 mode = AP_MODE_N_ONLY;
3767 else if (IEEE80211_IS_CHAN_5GHZ(chan))
3768 mode = AP_MODE_AandN;
3769 else if (vap->iv_flags & IEEE80211_F_PUREG)
3770 mode = AP_MODE_GandN;
3772 mode = AP_MODE_BandGandN;
3773 } else if (IEEE80211_IS_CHAN_ANYG(chan)) {
3774 if (vap->iv_flags & IEEE80211_F_PUREG)
3775 mode = AP_MODE_G_ONLY;
3777 mode = AP_MODE_MIXED;
3778 } else if (IEEE80211_IS_CHAN_B(chan))
3779 mode = AP_MODE_B_ONLY;
3780 else if (IEEE80211_IS_CHAN_A(chan))
3781 mode = AP_MODE_A_ONLY;
3783 mode = AP_MODE_MIXED; /* XXX should not happen? */
3788 mwl_setapmode(struct ieee80211vap *vap, struct ieee80211_channel *chan)
3790 struct mwl_hal_vap *hvap = MWL_VAP(vap)->mv_hvap;
3791 return mwl_hal_setapmode(hvap, mwl_getapmode(vap, chan));
3795 * Set/change channels.
3798 mwl_chan_set(struct mwl_softc *sc, struct ieee80211_channel *chan)
3800 struct mwl_hal *mh = sc->sc_mh;
3801 struct ieee80211com *ic = &sc->sc_ic;
3802 MWL_HAL_CHANNEL hchan;
3805 DPRINTF(sc, MWL_DEBUG_RESET, "%s: chan %u MHz/flags 0x%x\n",
3806 __func__, chan->ic_freq, chan->ic_flags);
3809 * Convert to a HAL channel description with
3810 * the flags constrained to reflect the current
3813 mwl_mapchan(&hchan, chan);
3814 mwl_hal_intrset(mh, 0); /* disable interrupts */
3816 mwl_draintxq(sc); /* clear pending tx frames */
3818 mwl_hal_setchannel(mh, &hchan);
3820 * Tx power is cap'd by the regulatory setting and
3821 * possibly a user-set limit. We pass the min of
3822 * these to the hal to apply them to the cal data
3826 maxtxpow = 2*chan->ic_maxregpower;
3827 if (maxtxpow > ic->ic_txpowlimit)
3828 maxtxpow = ic->ic_txpowlimit;
3829 mwl_hal_settxpower(mh, &hchan, maxtxpow / 2);
3830 /* NB: potentially change mcast/mgt rates */
3831 mwl_setcurchanrates(sc);
3834 * Update internal state.
3836 sc->sc_tx_th.wt_chan_freq = htole16(chan->ic_freq);
3837 sc->sc_rx_th.wr_chan_freq = htole16(chan->ic_freq);
3838 if (IEEE80211_IS_CHAN_A(chan)) {
3839 sc->sc_tx_th.wt_chan_flags = htole16(IEEE80211_CHAN_A);
3840 sc->sc_rx_th.wr_chan_flags = htole16(IEEE80211_CHAN_A);
3841 } else if (IEEE80211_IS_CHAN_ANYG(chan)) {
3842 sc->sc_tx_th.wt_chan_flags = htole16(IEEE80211_CHAN_G);
3843 sc->sc_rx_th.wr_chan_flags = htole16(IEEE80211_CHAN_G);
3845 sc->sc_tx_th.wt_chan_flags = htole16(IEEE80211_CHAN_B);
3846 sc->sc_rx_th.wr_chan_flags = htole16(IEEE80211_CHAN_B);
3848 sc->sc_curchan = hchan;
3849 mwl_hal_intrset(mh, sc->sc_imask);
3855 mwl_scan_start(struct ieee80211com *ic)
3857 struct mwl_softc *sc = ic->ic_softc;
3859 DPRINTF(sc, MWL_DEBUG_STATE, "%s\n", __func__);
3863 mwl_scan_end(struct ieee80211com *ic)
3865 struct mwl_softc *sc = ic->ic_softc;
3867 DPRINTF(sc, MWL_DEBUG_STATE, "%s\n", __func__);
3871 mwl_set_channel(struct ieee80211com *ic)
3873 struct mwl_softc *sc = ic->ic_softc;
3875 (void) mwl_chan_set(sc, ic->ic_curchan);
3879 * Handle a channel switch request. We inform the firmware
3880 * and mark the global state to suppress various actions.
3881 * NB: we issue only one request to the fw; we may be called
3882 * multiple times if there are multiple vap's.
3885 mwl_startcsa(struct ieee80211vap *vap)
3887 struct ieee80211com *ic = vap->iv_ic;
3888 struct mwl_softc *sc = ic->ic_softc;
3889 MWL_HAL_CHANNEL hchan;
3891 if (sc->sc_csapending)
3894 mwl_mapchan(&hchan, ic->ic_csa_newchan);
3895 /* 1 =>'s quiet channel */
3896 mwl_hal_setchannelswitchie(sc->sc_mh, &hchan, 1, ic->ic_csa_count);
3897 sc->sc_csapending = 1;
3901 * Plumb any static WEP key for the station. This is
3902 * necessary as we must propagate the key from the
3903 * global key table of the vap to each sta db entry.
3906 mwl_setanywepkey(struct ieee80211vap *vap, const uint8_t mac[IEEE80211_ADDR_LEN])
3908 if ((vap->iv_flags & (IEEE80211_F_PRIVACY|IEEE80211_F_WPA)) ==
3909 IEEE80211_F_PRIVACY &&
3910 vap->iv_def_txkey != IEEE80211_KEYIX_NONE &&
3911 vap->iv_nw_keys[vap->iv_def_txkey].wk_keyix != IEEE80211_KEYIX_NONE)
3912 (void) _mwl_key_set(vap, &vap->iv_nw_keys[vap->iv_def_txkey],
3917 mwl_peerstadb(struct ieee80211_node *ni, int aid, int staid, MWL_HAL_PEERINFO *pi)
3919 #define WME(ie) ((const struct ieee80211_wme_info *) ie)
3920 struct ieee80211vap *vap = ni->ni_vap;
3921 struct mwl_hal_vap *hvap;
3924 if (vap->iv_opmode == IEEE80211_M_WDS) {
3926 * WDS vap's do not have a f/w vap; instead they piggyback
3927 * on an AP vap and we must install the sta db entry and
3928 * crypto state using that AP's handle (the WDS vap has none).
3930 hvap = MWL_VAP(vap)->mv_ap_hvap;
3932 hvap = MWL_VAP(vap)->mv_hvap;
3933 error = mwl_hal_newstation(hvap, ni->ni_macaddr,
3935 ni->ni_flags & (IEEE80211_NODE_QOS | IEEE80211_NODE_HT),
3936 ni->ni_ies.wme_ie != NULL ? WME(ni->ni_ies.wme_ie)->wme_info : 0);
3939 * Setup security for this station. For sta mode this is
3940 * needed even though do the same thing on transition to
3941 * AUTH state because the call to mwl_hal_newstation
3942 * clobbers the crypto state we setup.
3944 mwl_setanywepkey(vap, ni->ni_macaddr);
3951 mwl_setglobalkeys(struct ieee80211vap *vap)
3953 struct ieee80211_key *wk;
3955 wk = &vap->iv_nw_keys[0];
3956 for (; wk < &vap->iv_nw_keys[IEEE80211_WEP_NKID]; wk++)
3957 if (wk->wk_keyix != IEEE80211_KEYIX_NONE)
3958 (void) _mwl_key_set(vap, wk, vap->iv_myaddr);
3962 * Convert a legacy rate set to a firmware bitmask.
3965 get_rate_bitmap(const struct ieee80211_rateset *rs)
3971 for (i = 0; i < rs->rs_nrates; i++)
3972 switch (rs->rs_rates[i] & IEEE80211_RATE_VAL) {
3973 case 2: rates |= 0x001; break;
3974 case 4: rates |= 0x002; break;
3975 case 11: rates |= 0x004; break;
3976 case 22: rates |= 0x008; break;
3977 case 44: rates |= 0x010; break;
3978 case 12: rates |= 0x020; break;
3979 case 18: rates |= 0x040; break;
3980 case 24: rates |= 0x080; break;
3981 case 36: rates |= 0x100; break;
3982 case 48: rates |= 0x200; break;
3983 case 72: rates |= 0x400; break;
3984 case 96: rates |= 0x800; break;
3985 case 108: rates |= 0x1000; break;
3991 * Construct an HT firmware bitmask from an HT rate set.
3994 get_htrate_bitmap(const struct ieee80211_htrateset *rs)
4000 for (i = 0; i < rs->rs_nrates; i++) {
4001 if (rs->rs_rates[i] < 16)
4002 rates |= 1<<rs->rs_rates[i];
4008 * Craft station database entry for station.
4009 * NB: use host byte order here, the hal handles byte swapping.
4011 static MWL_HAL_PEERINFO *
4012 mkpeerinfo(MWL_HAL_PEERINFO *pi, const struct ieee80211_node *ni)
4014 const struct ieee80211vap *vap = ni->ni_vap;
4016 memset(pi, 0, sizeof(*pi));
4017 pi->LegacyRateBitMap = get_rate_bitmap(&ni->ni_rates);
4018 pi->CapInfo = ni->ni_capinfo;
4019 if (ni->ni_flags & IEEE80211_NODE_HT) {
4020 /* HT capabilities, etc */
4021 pi->HTCapabilitiesInfo = ni->ni_htcap;
4022 /* XXX pi.HTCapabilitiesInfo */
4023 pi->MacHTParamInfo = ni->ni_htparam;
4024 pi->HTRateBitMap = get_htrate_bitmap(&ni->ni_htrates);
4025 pi->AddHtInfo.ControlChan = ni->ni_htctlchan;
4026 pi->AddHtInfo.AddChan = ni->ni_ht2ndchan;
4027 pi->AddHtInfo.OpMode = ni->ni_htopmode;
4028 pi->AddHtInfo.stbc = ni->ni_htstbc;
4030 /* constrain according to local configuration */
4031 if ((vap->iv_flags_ht & IEEE80211_FHT_SHORTGI40) == 0)
4032 pi->HTCapabilitiesInfo &= ~IEEE80211_HTCAP_SHORTGI40;
4033 if ((vap->iv_flags_ht & IEEE80211_FHT_SHORTGI20) == 0)
4034 pi->HTCapabilitiesInfo &= ~IEEE80211_HTCAP_SHORTGI20;
4035 if (ni->ni_chw != 40)
4036 pi->HTCapabilitiesInfo &= ~IEEE80211_HTCAP_CHWIDTH40;
4042 * Re-create the local sta db entry for a vap to ensure
4043 * up to date WME state is pushed to the firmware. Because
4044 * this resets crypto state this must be followed by a
4045 * reload of any keys in the global key table.
4048 mwl_localstadb(struct ieee80211vap *vap)
4050 #define WME(ie) ((const struct ieee80211_wme_info *) ie)
4051 struct mwl_hal_vap *hvap = MWL_VAP(vap)->mv_hvap;
4052 struct ieee80211_node *bss;
4053 MWL_HAL_PEERINFO pi;
4056 switch (vap->iv_opmode) {
4057 case IEEE80211_M_STA:
4059 error = mwl_hal_newstation(hvap, vap->iv_myaddr, 0, 0,
4060 vap->iv_state == IEEE80211_S_RUN ?
4061 mkpeerinfo(&pi, bss) : NULL,
4062 (bss->ni_flags & (IEEE80211_NODE_QOS | IEEE80211_NODE_HT)),
4063 bss->ni_ies.wme_ie != NULL ?
4064 WME(bss->ni_ies.wme_ie)->wme_info : 0);
4066 mwl_setglobalkeys(vap);
4068 case IEEE80211_M_HOSTAP:
4069 case IEEE80211_M_MBSS:
4070 error = mwl_hal_newstation(hvap, vap->iv_myaddr,
4071 0, 0, NULL, vap->iv_flags & IEEE80211_F_WME, 0);
4073 mwl_setglobalkeys(vap);
4084 mwl_newstate(struct ieee80211vap *vap, enum ieee80211_state nstate, int arg)
4086 struct mwl_vap *mvp = MWL_VAP(vap);
4087 struct mwl_hal_vap *hvap = mvp->mv_hvap;
4088 struct ieee80211com *ic = vap->iv_ic;
4089 struct ieee80211_node *ni = NULL;
4090 struct mwl_softc *sc = ic->ic_softc;
4091 struct mwl_hal *mh = sc->sc_mh;
4092 enum ieee80211_state ostate = vap->iv_state;
4095 DPRINTF(sc, MWL_DEBUG_STATE, "%s: %s: %s -> %s\n",
4096 vap->iv_ifp->if_xname, __func__,
4097 ieee80211_state_name[ostate], ieee80211_state_name[nstate]);
4099 callout_stop(&sc->sc_timer);
4101 * Clear current radar detection state.
4103 if (ostate == IEEE80211_S_CAC) {
4104 /* stop quiet mode radar detection */
4105 mwl_hal_setradardetection(mh, DR_CHK_CHANNEL_AVAILABLE_STOP);
4106 } else if (sc->sc_radarena) {
4107 /* stop in-service radar detection */
4108 mwl_hal_setradardetection(mh, DR_DFS_DISABLE);
4109 sc->sc_radarena = 0;
4112 * Carry out per-state actions before doing net80211 work.
4114 if (nstate == IEEE80211_S_INIT) {
4115 /* NB: only ap+sta vap's have a fw entity */
4118 } else if (nstate == IEEE80211_S_SCAN) {
4119 mwl_hal_start(hvap);
4120 /* NB: this disables beacon frames */
4121 mwl_hal_setinframode(hvap);
4122 } else if (nstate == IEEE80211_S_AUTH) {
4124 * Must create a sta db entry in case a WEP key needs to
4125 * be plumbed. This entry will be overwritten if we
4126 * associate; otherwise it will be reclaimed on node free.
4129 MWL_NODE(ni)->mn_hvap = hvap;
4130 (void) mwl_peerstadb(ni, 0, 0, NULL);
4131 } else if (nstate == IEEE80211_S_CSA) {
4132 /* XXX move to below? */
4133 if (vap->iv_opmode == IEEE80211_M_HOSTAP ||
4134 vap->iv_opmode == IEEE80211_M_MBSS)
4136 } else if (nstate == IEEE80211_S_CAC) {
4137 /* XXX move to below? */
4138 /* stop ap xmit and enable quiet mode radar detection */
4139 mwl_hal_setradardetection(mh, DR_CHK_CHANNEL_AVAILABLE_START);
4143 * Invoke the parent method to do net80211 work.
4145 error = mvp->mv_newstate(vap, nstate, arg);
4148 * Carry out work that must be done after net80211 runs;
4149 * this work requires up to date state (e.g. iv_bss).
4151 if (error == 0 && nstate == IEEE80211_S_RUN) {
4152 /* NB: collect bss node again, it may have changed */
4155 DPRINTF(sc, MWL_DEBUG_STATE,
4156 "%s: %s(RUN): iv_flags 0x%08x bintvl %d bssid %s "
4157 "capinfo 0x%04x chan %d\n",
4158 vap->iv_ifp->if_xname, __func__, vap->iv_flags,
4159 ni->ni_intval, ether_sprintf(ni->ni_bssid), ni->ni_capinfo,
4160 ieee80211_chan2ieee(ic, ic->ic_curchan));
4163 * Recreate local sta db entry to update WME/HT state.
4165 mwl_localstadb(vap);
4166 switch (vap->iv_opmode) {
4167 case IEEE80211_M_HOSTAP:
4168 case IEEE80211_M_MBSS:
4169 if (ostate == IEEE80211_S_CAC) {
4170 /* enable in-service radar detection */
4171 mwl_hal_setradardetection(mh,
4172 DR_IN_SERVICE_MONITOR_START);
4173 sc->sc_radarena = 1;
4176 * Allocate and setup the beacon frame
4177 * (and related state).
4179 error = mwl_reset_vap(vap, IEEE80211_S_RUN);
4181 DPRINTF(sc, MWL_DEBUG_STATE,
4182 "%s: beacon setup failed, error %d\n",
4186 /* NB: must be after setting up beacon */
4187 mwl_hal_start(hvap);
4189 case IEEE80211_M_STA:
4190 DPRINTF(sc, MWL_DEBUG_STATE, "%s: %s: aid 0x%x\n",
4191 vap->iv_ifp->if_xname, __func__, ni->ni_associd);
4193 * Set state now that we're associated.
4195 mwl_hal_setassocid(hvap, ni->ni_bssid, ni->ni_associd);
4197 mwl_hal_setrtsthreshold(hvap, vap->iv_rtsthreshold);
4198 if ((vap->iv_flags & IEEE80211_F_DWDS) &&
4199 sc->sc_ndwdsvaps++ == 0)
4200 mwl_hal_setdwds(mh, 1);
4202 case IEEE80211_M_WDS:
4203 DPRINTF(sc, MWL_DEBUG_STATE, "%s: %s: bssid %s\n",
4204 vap->iv_ifp->if_xname, __func__,
4205 ether_sprintf(ni->ni_bssid));
4206 mwl_seteapolformat(vap);
4212 * Set CS mode according to operating channel;
4213 * this mostly an optimization for 5GHz.
4215 * NB: must follow mwl_hal_start which resets csmode
4217 if (IEEE80211_IS_CHAN_5GHZ(ic->ic_bsschan))
4218 mwl_hal_setcsmode(mh, CSMODE_AGGRESSIVE);
4220 mwl_hal_setcsmode(mh, CSMODE_AUTO_ENA);
4222 * Start timer to prod firmware.
4224 if (sc->sc_ageinterval != 0)
4225 callout_reset(&sc->sc_timer, sc->sc_ageinterval*hz,
4226 mwl_agestations, sc);
4227 } else if (nstate == IEEE80211_S_SLEEP) {
4228 /* XXX set chip in power save */
4229 } else if ((vap->iv_flags & IEEE80211_F_DWDS) &&
4230 --sc->sc_ndwdsvaps == 0)
4231 mwl_hal_setdwds(mh, 0);
4237 * Manage station id's; these are separate from AID's
4238 * as AID's may have values out of the range of possible
4239 * station id's acceptable to the firmware.
4242 allocstaid(struct mwl_softc *sc, int aid)
4246 if (!(0 < aid && aid < MWL_MAXSTAID) || isset(sc->sc_staid, aid)) {
4247 /* NB: don't use 0 */
4248 for (staid = 1; staid < MWL_MAXSTAID; staid++)
4249 if (isclr(sc->sc_staid, staid))
4253 setbit(sc->sc_staid, staid);
4258 delstaid(struct mwl_softc *sc, int staid)
4260 clrbit(sc->sc_staid, staid);
4264 * Setup driver-specific state for a newly associated node.
4265 * Note that we're called also on a re-associate, the isnew
4266 * param tells us if this is the first time or not.
4269 mwl_newassoc(struct ieee80211_node *ni, int isnew)
4271 struct ieee80211vap *vap = ni->ni_vap;
4272 struct mwl_softc *sc = vap->iv_ic->ic_softc;
4273 struct mwl_node *mn = MWL_NODE(ni);
4274 MWL_HAL_PEERINFO pi;
4278 aid = IEEE80211_AID(ni->ni_associd);
4280 mn->mn_staid = allocstaid(sc, aid);
4281 mn->mn_hvap = MWL_VAP(vap)->mv_hvap;
4284 /* XXX reset BA stream? */
4286 DPRINTF(sc, MWL_DEBUG_NODE, "%s: mac %s isnew %d aid %d staid %d\n",
4287 __func__, ether_sprintf(ni->ni_macaddr), isnew, aid, mn->mn_staid);
4288 error = mwl_peerstadb(ni, aid, mn->mn_staid, mkpeerinfo(&pi, ni));
4290 DPRINTF(sc, MWL_DEBUG_NODE,
4291 "%s: error %d creating sta db entry\n",
4293 /* XXX how to deal with error? */
4298 * Periodically poke the firmware to age out station state
4299 * (power save queues, pending tx aggregates).
4302 mwl_agestations(void *arg)
4304 struct mwl_softc *sc = arg;
4306 mwl_hal_setkeepalive(sc->sc_mh);
4307 if (sc->sc_ageinterval != 0) /* NB: catch dynamic changes */
4308 callout_schedule(&sc->sc_timer, sc->sc_ageinterval*hz);
4311 static const struct mwl_hal_channel *
4312 findhalchannel(const MWL_HAL_CHANNELINFO *ci, int ieee)
4316 for (i = 0; i < ci->nchannels; i++) {
4317 const struct mwl_hal_channel *hc = &ci->channels[i];
4318 if (hc->ieee == ieee)
4325 mwl_setregdomain(struct ieee80211com *ic, struct ieee80211_regdomain *rd,
4326 int nchan, struct ieee80211_channel chans[])
4328 struct mwl_softc *sc = ic->ic_softc;
4329 struct mwl_hal *mh = sc->sc_mh;
4330 const MWL_HAL_CHANNELINFO *ci;
4333 for (i = 0; i < nchan; i++) {
4334 struct ieee80211_channel *c = &chans[i];
4335 const struct mwl_hal_channel *hc;
4337 if (IEEE80211_IS_CHAN_2GHZ(c)) {
4338 mwl_hal_getchannelinfo(mh, MWL_FREQ_BAND_2DOT4GHZ,
4339 IEEE80211_IS_CHAN_HT40(c) ?
4340 MWL_CH_40_MHz_WIDTH : MWL_CH_20_MHz_WIDTH, &ci);
4341 } else if (IEEE80211_IS_CHAN_5GHZ(c)) {
4342 mwl_hal_getchannelinfo(mh, MWL_FREQ_BAND_5GHZ,
4343 IEEE80211_IS_CHAN_HT40(c) ?
4344 MWL_CH_40_MHz_WIDTH : MWL_CH_20_MHz_WIDTH, &ci);
4346 device_printf(sc->sc_dev,
4347 "%s: channel %u freq %u/0x%x not 2.4/5GHz\n",
4348 __func__, c->ic_ieee, c->ic_freq, c->ic_flags);
4352 * Verify channel has cal data and cap tx power.
4354 hc = findhalchannel(ci, c->ic_ieee);
4356 if (c->ic_maxpower > 2*hc->maxTxPow)
4357 c->ic_maxpower = 2*hc->maxTxPow;
4360 if (IEEE80211_IS_CHAN_HT40(c)) {
4362 * Look for the extension channel since the
4363 * hal table only has the primary channel.
4365 hc = findhalchannel(ci, c->ic_extieee);
4367 if (c->ic_maxpower > 2*hc->maxTxPow)
4368 c->ic_maxpower = 2*hc->maxTxPow;
4372 device_printf(sc->sc_dev,
4373 "%s: no cal data for channel %u ext %u freq %u/0x%x\n",
4374 __func__, c->ic_ieee, c->ic_extieee,
4375 c->ic_freq, c->ic_flags);
4383 #define IEEE80211_CHAN_HTG (IEEE80211_CHAN_HT|IEEE80211_CHAN_G)
4384 #define IEEE80211_CHAN_HTA (IEEE80211_CHAN_HT|IEEE80211_CHAN_A)
4387 addht40channels(struct ieee80211_channel chans[], int maxchans, int *nchans,
4388 const MWL_HAL_CHANNELINFO *ci, int flags)
4392 for (i = 0; i < ci->nchannels; i++) {
4393 const struct mwl_hal_channel *hc = &ci->channels[i];
4395 error = ieee80211_add_channel_ht40(chans, maxchans, nchans,
4396 hc->ieee, hc->maxTxPow, flags);
4397 if (error != 0 && error != ENOENT)
4403 addchannels(struct ieee80211_channel chans[], int maxchans, int *nchans,
4404 const MWL_HAL_CHANNELINFO *ci, const uint8_t bands[])
4409 for (i = 0; i < ci->nchannels && error == 0; i++) {
4410 const struct mwl_hal_channel *hc = &ci->channels[i];
4412 error = ieee80211_add_channel(chans, maxchans, nchans,
4413 hc->ieee, hc->freq, hc->maxTxPow, 0, bands);
4418 getchannels(struct mwl_softc *sc, int maxchans, int *nchans,
4419 struct ieee80211_channel chans[])
4421 const MWL_HAL_CHANNELINFO *ci;
4422 uint8_t bands[IEEE80211_MODE_BYTES];
4425 * Use the channel info from the hal to craft the
4426 * channel list. Note that we pass back an unsorted
4427 * list; the caller is required to sort it for us
4431 if (mwl_hal_getchannelinfo(sc->sc_mh,
4432 MWL_FREQ_BAND_2DOT4GHZ, MWL_CH_20_MHz_WIDTH, &ci) == 0) {
4433 memset(bands, 0, sizeof(bands));
4434 setbit(bands, IEEE80211_MODE_11B);
4435 setbit(bands, IEEE80211_MODE_11G);
4436 setbit(bands, IEEE80211_MODE_11NG);
4437 addchannels(chans, maxchans, nchans, ci, bands);
4439 if (mwl_hal_getchannelinfo(sc->sc_mh,
4440 MWL_FREQ_BAND_5GHZ, MWL_CH_20_MHz_WIDTH, &ci) == 0) {
4441 memset(bands, 0, sizeof(bands));
4442 setbit(bands, IEEE80211_MODE_11A);
4443 setbit(bands, IEEE80211_MODE_11NA);
4444 addchannels(chans, maxchans, nchans, ci, bands);
4446 if (mwl_hal_getchannelinfo(sc->sc_mh,
4447 MWL_FREQ_BAND_2DOT4GHZ, MWL_CH_40_MHz_WIDTH, &ci) == 0)
4448 addht40channels(chans, maxchans, nchans, ci, IEEE80211_CHAN_HTG);
4449 if (mwl_hal_getchannelinfo(sc->sc_mh,
4450 MWL_FREQ_BAND_5GHZ, MWL_CH_40_MHz_WIDTH, &ci) == 0)
4451 addht40channels(chans, maxchans, nchans, ci, IEEE80211_CHAN_HTA);
4455 mwl_getradiocaps(struct ieee80211com *ic,
4456 int maxchans, int *nchans, struct ieee80211_channel chans[])
4458 struct mwl_softc *sc = ic->ic_softc;
4460 getchannels(sc, maxchans, nchans, chans);
4464 mwl_getchannels(struct mwl_softc *sc)
4466 struct ieee80211com *ic = &sc->sc_ic;
4469 * Use the channel info from the hal to craft the
4470 * channel list for net80211. Note that we pass up
4471 * an unsorted list; net80211 will sort it for us.
4473 memset(ic->ic_channels, 0, sizeof(ic->ic_channels));
4475 getchannels(sc, IEEE80211_CHAN_MAX, &ic->ic_nchans, ic->ic_channels);
4477 ic->ic_regdomain.regdomain = SKU_DEBUG;
4478 ic->ic_regdomain.country = CTRY_DEFAULT;
4479 ic->ic_regdomain.location = 'I';
4480 ic->ic_regdomain.isocc[0] = ' '; /* XXX? */
4481 ic->ic_regdomain.isocc[1] = ' ';
4482 return (ic->ic_nchans == 0 ? EIO : 0);
4484 #undef IEEE80211_CHAN_HTA
4485 #undef IEEE80211_CHAN_HTG
4489 mwl_printrxbuf(const struct mwl_rxbuf *bf, u_int ix)
4491 const struct mwl_rxdesc *ds = bf->bf_desc;
4492 uint32_t status = le32toh(ds->Status);
4494 printf("R[%2u] (DS.V:%p DS.P:0x%jx) NEXT:%08x DATA:%08x RC:%02x%s\n"
4495 " STAT:%02x LEN:%04x RSSI:%02x CHAN:%02x RATE:%02x QOS:%04x HT:%04x\n",
4496 ix, ds, (uintmax_t)bf->bf_daddr, le32toh(ds->pPhysNext),
4497 le32toh(ds->pPhysBuffData), ds->RxControl,
4498 ds->RxControl != EAGLE_RXD_CTRL_DRIVER_OWN ?
4499 "" : (status & EAGLE_RXD_STATUS_OK) ? " *" : " !",
4500 ds->Status, le16toh(ds->PktLen), ds->RSSI, ds->Channel,
4501 ds->Rate, le16toh(ds->QosCtrl), le16toh(ds->HtSig2));
4505 mwl_printtxbuf(const struct mwl_txbuf *bf, u_int qnum, u_int ix)
4507 const struct mwl_txdesc *ds = bf->bf_desc;
4508 uint32_t status = le32toh(ds->Status);
4510 printf("Q%u[%3u]", qnum, ix);
4511 printf(" (DS.V:%p DS.P:0x%jx)\n", ds, (uintmax_t)bf->bf_daddr);
4512 printf(" NEXT:%08x DATA:%08x LEN:%04x STAT:%08x%s\n",
4513 le32toh(ds->pPhysNext),
4514 le32toh(ds->PktPtr), le16toh(ds->PktLen), status,
4515 status & EAGLE_TXD_STATUS_USED ?
4516 "" : (status & 3) != 0 ? " *" : " !");
4517 printf(" RATE:%02x PRI:%x QOS:%04x SAP:%08x FORMAT:%04x\n",
4518 ds->DataRate, ds->TxPriority, le16toh(ds->QosCtrl),
4519 le32toh(ds->SapPktInfo), le16toh(ds->Format));
4521 printf(" MULTIFRAMES:%u LEN:%04x %04x %04x %04x %04x %04x\n"
4522 , le32toh(ds->multiframes)
4523 , le16toh(ds->PktLenArray[0]), le16toh(ds->PktLenArray[1])
4524 , le16toh(ds->PktLenArray[2]), le16toh(ds->PktLenArray[3])
4525 , le16toh(ds->PktLenArray[4]), le16toh(ds->PktLenArray[5])
4527 printf(" DATA:%08x %08x %08x %08x %08x %08x\n"
4528 , le32toh(ds->PktPtrArray[0]), le32toh(ds->PktPtrArray[1])
4529 , le32toh(ds->PktPtrArray[2]), le32toh(ds->PktPtrArray[3])
4530 , le32toh(ds->PktPtrArray[4]), le32toh(ds->PktPtrArray[5])
4534 { const uint8_t *cp = (const uint8_t *) ds;
4536 for (i = 0; i < sizeof(struct mwl_txdesc); i++) {
4537 printf("%02x ", cp[i]);
4538 if (((i+1) % 16) == 0)
4545 #endif /* MWL_DEBUG */
4549 mwl_txq_dump(struct mwl_txq *txq)
4551 struct mwl_txbuf *bf;
4555 STAILQ_FOREACH(bf, &txq->active, bf_list) {
4556 struct mwl_txdesc *ds = bf->bf_desc;
4557 MWL_TXDESC_SYNC(txq, ds,
4558 BUS_DMASYNC_POSTREAD | BUS_DMASYNC_POSTWRITE);
4560 mwl_printtxbuf(bf, txq->qnum, i);
4564 MWL_TXQ_UNLOCK(txq);
4569 mwl_watchdog(void *arg)
4571 struct mwl_softc *sc = arg;
4573 callout_reset(&sc->sc_watchdog, hz, mwl_watchdog, sc);
4574 if (sc->sc_tx_timer == 0 || --sc->sc_tx_timer > 0)
4577 if (sc->sc_running && !sc->sc_invalid) {
4578 if (mwl_hal_setkeepalive(sc->sc_mh))
4579 device_printf(sc->sc_dev,
4580 "transmit timeout (firmware hung?)\n");
4582 device_printf(sc->sc_dev,
4583 "transmit timeout\n");
4586 mwl_txq_dump(&sc->sc_txq[0]);/*XXX*/
4588 counter_u64_add(sc->sc_ic.ic_oerrors, 1);
4589 sc->sc_stats.mst_watchdog++;
4595 * Diagnostic interface to the HAL. This is used by various
4596 * tools to do things like retrieve register contents for
4597 * debugging. The mechanism is intentionally opaque so that
4598 * it can change frequently w/o concern for compatibility.
4601 mwl_ioctl_diag(struct mwl_softc *sc, struct mwl_diag *md)
4603 struct mwl_hal *mh = sc->sc_mh;
4604 u_int id = md->md_id & MWL_DIAG_ID;
4605 void *indata = NULL;
4606 void *outdata = NULL;
4607 u_int32_t insize = md->md_in_size;
4608 u_int32_t outsize = md->md_out_size;
4611 if (md->md_id & MWL_DIAG_IN) {
4615 indata = malloc(insize, M_TEMP, M_NOWAIT);
4616 if (indata == NULL) {
4620 error = copyin(md->md_in_data, indata, insize);
4624 if (md->md_id & MWL_DIAG_DYN) {
4626 * Allocate a buffer for the results (otherwise the HAL
4627 * returns a pointer to a buffer where we can read the
4628 * results). Note that we depend on the HAL leaving this
4629 * pointer for us to use below in reclaiming the buffer;
4630 * may want to be more defensive.
4632 outdata = malloc(outsize, M_TEMP, M_NOWAIT);
4633 if (outdata == NULL) {
4638 if (mwl_hal_getdiagstate(mh, id, indata, insize, &outdata, &outsize)) {
4639 if (outsize < md->md_out_size)
4640 md->md_out_size = outsize;
4641 if (outdata != NULL)
4642 error = copyout(outdata, md->md_out_data,
4648 if ((md->md_id & MWL_DIAG_IN) && indata != NULL)
4649 free(indata, M_TEMP);
4650 if ((md->md_id & MWL_DIAG_DYN) && outdata != NULL)
4651 free(outdata, M_TEMP);
4656 mwl_ioctl_reset(struct mwl_softc *sc, struct mwl_diag *md)
4658 struct mwl_hal *mh = sc->sc_mh;
4661 MWL_LOCK_ASSERT(sc);
4663 if (md->md_id == 0 && mwl_hal_fwload(mh, NULL) != 0) {
4664 device_printf(sc->sc_dev, "unable to load firmware\n");
4667 if (mwl_hal_gethwspecs(mh, &sc->sc_hwspecs) != 0) {
4668 device_printf(sc->sc_dev, "unable to fetch h/w specs\n");
4671 error = mwl_setupdma(sc);
4673 /* NB: mwl_setupdma prints a msg */
4677 * Reset tx/rx data structures; after reload we must
4678 * re-start the driver's notion of the next xmit/recv.
4680 mwl_draintxq(sc); /* clear pending frames */
4681 mwl_resettxq(sc); /* rebuild tx q lists */
4682 sc->sc_rxnext = NULL; /* force rx to start at the list head */
4685 #endif /* MWL_DIAGAPI */
4688 mwl_parent(struct ieee80211com *ic)
4690 struct mwl_softc *sc = ic->ic_softc;
4694 if (ic->ic_nrunning > 0) {
4695 if (sc->sc_running) {
4697 * To avoid rescanning another access point,
4698 * do not call mwl_init() here. Instead,
4699 * only reflect promisc mode settings.
4704 * Beware of being called during attach/detach
4705 * to reset promiscuous mode. In that case we
4706 * will still be marked UP but not RUNNING.
4707 * However trying to re-init the interface
4708 * is the wrong thing to do as we've already
4709 * torn down much of our state. There's
4710 * probably a better way to deal with this.
4712 if (!sc->sc_invalid) {
4713 mwl_init(sc); /* XXX lose error */
4721 ieee80211_start_all(ic);
4725 mwl_ioctl(struct ieee80211com *ic, u_long cmd, void *data)
4727 struct mwl_softc *sc = ic->ic_softc;
4728 struct ifreq *ifr = data;
4733 mwl_hal_gethwstats(sc->sc_mh, &sc->sc_stats.hw_stats);
4735 /* NB: embed these numbers to get a consistent view */
4736 sc->sc_stats.mst_tx_packets =
4737 ifp->if_get_counter(ifp, IFCOUNTER_OPACKETS);
4738 sc->sc_stats.mst_rx_packets =
4739 ifp->if_get_counter(ifp, IFCOUNTER_IPACKETS);
4742 * NB: Drop the softc lock in case of a page fault;
4743 * we'll accept any potential inconsisentcy in the
4744 * statistics. The alternative is to copy the data
4745 * to a local structure.
4747 return (copyout(&sc->sc_stats,
4748 ifr->ifr_data, sizeof (sc->sc_stats)));
4751 /* XXX check privs */
4752 return mwl_ioctl_diag(sc, (struct mwl_diag *) ifr);
4754 /* XXX check privs */
4756 error = mwl_ioctl_reset(sc,(struct mwl_diag *) ifr);
4759 #endif /* MWL_DIAGAPI */
4769 mwl_sysctl_debug(SYSCTL_HANDLER_ARGS)
4771 struct mwl_softc *sc = arg1;
4774 debug = sc->sc_debug | (mwl_hal_getdebug(sc->sc_mh) << 24);
4775 error = sysctl_handle_int(oidp, &debug, 0, req);
4776 if (error || !req->newptr)
4778 mwl_hal_setdebug(sc->sc_mh, debug >> 24);
4779 sc->sc_debug = debug & 0x00ffffff;
4782 #endif /* MWL_DEBUG */
4785 mwl_sysctlattach(struct mwl_softc *sc)
4788 struct sysctl_ctx_list *ctx = device_get_sysctl_ctx(sc->sc_dev);
4789 struct sysctl_oid *tree = device_get_sysctl_tree(sc->sc_dev);
4791 sc->sc_debug = mwl_debug;
4792 SYSCTL_ADD_PROC(ctx, SYSCTL_CHILDREN(tree), OID_AUTO,
4793 "debug", CTLTYPE_INT | CTLFLAG_RW, sc, 0,
4794 mwl_sysctl_debug, "I", "control debugging printfs");
4799 * Announce various information on device/driver attach.
4802 mwl_announce(struct mwl_softc *sc)
4805 device_printf(sc->sc_dev, "Rev A%d hardware, v%d.%d.%d.%d firmware (regioncode %d)\n",
4806 sc->sc_hwspecs.hwVersion,
4807 (sc->sc_hwspecs.fwReleaseNumber>>24) & 0xff,
4808 (sc->sc_hwspecs.fwReleaseNumber>>16) & 0xff,
4809 (sc->sc_hwspecs.fwReleaseNumber>>8) & 0xff,
4810 (sc->sc_hwspecs.fwReleaseNumber>>0) & 0xff,
4811 sc->sc_hwspecs.regionCode);
4812 sc->sc_fwrelease = sc->sc_hwspecs.fwReleaseNumber;
4816 for (i = 0; i <= WME_AC_VO; i++) {
4817 struct mwl_txq *txq = sc->sc_ac2q[i];
4818 device_printf(sc->sc_dev, "Use hw queue %u for %s traffic\n",
4819 txq->qnum, ieee80211_wme_acnames[i]);
4822 if (bootverbose || mwl_rxdesc != MWL_RXDESC)
4823 device_printf(sc->sc_dev, "using %u rx descriptors\n", mwl_rxdesc);
4824 if (bootverbose || mwl_rxbuf != MWL_RXBUF)
4825 device_printf(sc->sc_dev, "using %u rx buffers\n", mwl_rxbuf);
4826 if (bootverbose || mwl_txbuf != MWL_TXBUF)
4827 device_printf(sc->sc_dev, "using %u tx buffers\n", mwl_txbuf);
4828 if (bootverbose && mwl_hal_ismbsscapable(sc->sc_mh))
4829 device_printf(sc->sc_dev, "multi-bss support\n");
4830 #ifdef MWL_TX_NODROP
4832 device_printf(sc->sc_dev, "no tx drop\n");