2 * SPDX-License-Identifier: BSD-2-Clause-FreeBSD
4 * Copyright (c) 2007-2009 Sam Leffler, Errno Consulting
5 * Copyright (c) 2007-2008 Marvell Semiconductor, Inc.
8 * Redistribution and use in source and binary forms, with or without
9 * modification, are permitted provided that the following conditions
11 * 1. Redistributions of source code must retain the above copyright
12 * notice, this list of conditions and the following disclaimer,
13 * without modification.
14 * 2. Redistributions in binary form must reproduce at minimum a disclaimer
15 * similar to the "NO WARRANTY" disclaimer below ("Disclaimer") and any
16 * redistribution must be conditioned upon including a substantially
17 * similar Disclaimer requirement for further binary redistribution.
20 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
21 * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
22 * LIMITED TO, THE IMPLIED WARRANTIES OF NONINFRINGEMENT, MERCHANTIBILITY
23 * AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL
24 * THE COPYRIGHT HOLDERS OR CONTRIBUTORS BE LIABLE FOR SPECIAL, EXEMPLARY,
25 * OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
26 * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
27 * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER
28 * IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
29 * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF
30 * THE POSSIBILITY OF SUCH DAMAGES.
33 #include <sys/cdefs.h>
34 __FBSDID("$FreeBSD$");
37 * Driver for the Marvell 88W8363 Wireless LAN controller.
44 #include <sys/param.h>
45 #include <sys/systm.h>
46 #include <sys/sysctl.h>
48 #include <sys/malloc.h>
50 #include <sys/mutex.h>
51 #include <sys/kernel.h>
52 #include <sys/socket.h>
53 #include <sys/sockio.h>
54 #include <sys/errno.h>
55 #include <sys/callout.h>
57 #include <sys/endian.h>
58 #include <sys/kthread.h>
59 #include <sys/taskqueue.h>
61 #include <machine/bus.h>
64 #include <net/if_var.h>
65 #include <net/if_dl.h>
66 #include <net/if_media.h>
67 #include <net/if_types.h>
68 #include <net/if_arp.h>
69 #include <net/ethernet.h>
70 #include <net/if_llc.h>
74 #include <net80211/ieee80211_var.h>
75 #include <net80211/ieee80211_input.h>
76 #include <net80211/ieee80211_regdomain.h>
79 #include <netinet/in.h>
80 #include <netinet/if_ether.h>
83 #include <dev/mwl/if_mwlvar.h>
84 #include <dev/mwl/mwldiag.h>
86 /* idiomatic shorthands: MS = mask+shift, SM = shift+mask */
87 #define MS(v,x) (((v) & x) >> x##_S)
88 #define SM(v,x) (((v) << x##_S) & x)
90 static struct ieee80211vap *mwl_vap_create(struct ieee80211com *,
91 const char [IFNAMSIZ], int, enum ieee80211_opmode, int,
92 const uint8_t [IEEE80211_ADDR_LEN],
93 const uint8_t [IEEE80211_ADDR_LEN]);
94 static void mwl_vap_delete(struct ieee80211vap *);
95 static int mwl_setupdma(struct mwl_softc *);
96 static int mwl_hal_reset(struct mwl_softc *sc);
97 static int mwl_init(struct mwl_softc *);
98 static void mwl_parent(struct ieee80211com *);
99 static int mwl_reset(struct ieee80211vap *, u_long);
100 static void mwl_stop(struct mwl_softc *);
101 static void mwl_start(struct mwl_softc *);
102 static int mwl_transmit(struct ieee80211com *, struct mbuf *);
103 static int mwl_raw_xmit(struct ieee80211_node *, struct mbuf *,
104 const struct ieee80211_bpf_params *);
105 static int mwl_media_change(struct ifnet *);
106 static void mwl_watchdog(void *);
107 static int mwl_ioctl(struct ieee80211com *, u_long, void *);
108 static void mwl_radar_proc(void *, int);
109 static void mwl_chanswitch_proc(void *, int);
110 static void mwl_bawatchdog_proc(void *, int);
111 static int mwl_key_alloc(struct ieee80211vap *,
112 struct ieee80211_key *,
113 ieee80211_keyix *, ieee80211_keyix *);
114 static int mwl_key_delete(struct ieee80211vap *,
115 const struct ieee80211_key *);
116 static int mwl_key_set(struct ieee80211vap *,
117 const struct ieee80211_key *);
118 static int _mwl_key_set(struct ieee80211vap *,
119 const struct ieee80211_key *,
120 const uint8_t mac[IEEE80211_ADDR_LEN]);
121 static int mwl_mode_init(struct mwl_softc *);
122 static void mwl_update_mcast(struct ieee80211com *);
123 static void mwl_update_promisc(struct ieee80211com *);
124 static void mwl_updateslot(struct ieee80211com *);
125 static int mwl_beacon_setup(struct ieee80211vap *);
126 static void mwl_beacon_update(struct ieee80211vap *, int);
127 #ifdef MWL_HOST_PS_SUPPORT
128 static void mwl_update_ps(struct ieee80211vap *, int);
129 static int mwl_set_tim(struct ieee80211_node *, int);
131 static int mwl_dma_setup(struct mwl_softc *);
132 static void mwl_dma_cleanup(struct mwl_softc *);
133 static struct ieee80211_node *mwl_node_alloc(struct ieee80211vap *,
134 const uint8_t [IEEE80211_ADDR_LEN]);
135 static void mwl_node_cleanup(struct ieee80211_node *);
136 static void mwl_node_drain(struct ieee80211_node *);
137 static void mwl_node_getsignal(const struct ieee80211_node *,
139 static void mwl_node_getmimoinfo(const struct ieee80211_node *,
140 struct ieee80211_mimo_info *);
141 static int mwl_rxbuf_init(struct mwl_softc *, struct mwl_rxbuf *);
142 static void mwl_rx_proc(void *, int);
143 static void mwl_txq_init(struct mwl_softc *sc, struct mwl_txq *, int);
144 static int mwl_tx_setup(struct mwl_softc *, int, int);
145 static int mwl_wme_update(struct ieee80211com *);
146 static void mwl_tx_cleanupq(struct mwl_softc *, struct mwl_txq *);
147 static void mwl_tx_cleanup(struct mwl_softc *);
148 static uint16_t mwl_calcformat(uint8_t rate, const struct ieee80211_node *);
149 static int mwl_tx_start(struct mwl_softc *, struct ieee80211_node *,
150 struct mwl_txbuf *, struct mbuf *);
151 static void mwl_tx_proc(void *, int);
152 static int mwl_chan_set(struct mwl_softc *, struct ieee80211_channel *);
153 static void mwl_draintxq(struct mwl_softc *);
154 static void mwl_cleartxq(struct mwl_softc *, struct ieee80211vap *);
155 static int mwl_recv_action(struct ieee80211_node *,
156 const struct ieee80211_frame *,
157 const uint8_t *, const uint8_t *);
158 static int mwl_addba_request(struct ieee80211_node *,
159 struct ieee80211_tx_ampdu *, int dialogtoken,
160 int baparamset, int batimeout);
161 static int mwl_addba_response(struct ieee80211_node *,
162 struct ieee80211_tx_ampdu *, int status,
163 int baparamset, int batimeout);
164 static void mwl_addba_stop(struct ieee80211_node *,
165 struct ieee80211_tx_ampdu *);
166 static int mwl_startrecv(struct mwl_softc *);
167 static MWL_HAL_APMODE mwl_getapmode(const struct ieee80211vap *,
168 struct ieee80211_channel *);
169 static int mwl_setapmode(struct ieee80211vap *, struct ieee80211_channel*);
170 static void mwl_scan_start(struct ieee80211com *);
171 static void mwl_scan_end(struct ieee80211com *);
172 static void mwl_set_channel(struct ieee80211com *);
173 static int mwl_peerstadb(struct ieee80211_node *,
174 int aid, int staid, MWL_HAL_PEERINFO *pi);
175 static int mwl_localstadb(struct ieee80211vap *);
176 static int mwl_newstate(struct ieee80211vap *, enum ieee80211_state, int);
177 static int allocstaid(struct mwl_softc *sc, int aid);
178 static void delstaid(struct mwl_softc *sc, int staid);
179 static void mwl_newassoc(struct ieee80211_node *, int);
180 static void mwl_agestations(void *);
181 static int mwl_setregdomain(struct ieee80211com *,
182 struct ieee80211_regdomain *, int,
183 struct ieee80211_channel []);
184 static void mwl_getradiocaps(struct ieee80211com *, int, int *,
185 struct ieee80211_channel []);
186 static int mwl_getchannels(struct mwl_softc *);
188 static void mwl_sysctlattach(struct mwl_softc *);
189 static void mwl_announce(struct mwl_softc *);
191 SYSCTL_NODE(_hw, OID_AUTO, mwl, CTLFLAG_RD, 0, "Marvell driver parameters");
193 static int mwl_rxdesc = MWL_RXDESC; /* # rx desc's to allocate */
194 SYSCTL_INT(_hw_mwl, OID_AUTO, rxdesc, CTLFLAG_RW, &mwl_rxdesc,
195 0, "rx descriptors allocated");
196 static int mwl_rxbuf = MWL_RXBUF; /* # rx buffers to allocate */
197 SYSCTL_INT(_hw_mwl, OID_AUTO, rxbuf, CTLFLAG_RWTUN, &mwl_rxbuf,
198 0, "rx buffers allocated");
199 static int mwl_txbuf = MWL_TXBUF; /* # tx buffers to allocate */
200 SYSCTL_INT(_hw_mwl, OID_AUTO, txbuf, CTLFLAG_RWTUN, &mwl_txbuf,
201 0, "tx buffers allocated");
202 static int mwl_txcoalesce = 8; /* # tx packets to q before poking f/w*/
203 SYSCTL_INT(_hw_mwl, OID_AUTO, txcoalesce, CTLFLAG_RWTUN, &mwl_txcoalesce,
204 0, "tx buffers to send at once");
205 static int mwl_rxquota = MWL_RXBUF; /* # max buffers to process */
206 SYSCTL_INT(_hw_mwl, OID_AUTO, rxquota, CTLFLAG_RWTUN, &mwl_rxquota,
207 0, "max rx buffers to process per interrupt");
208 static int mwl_rxdmalow = 3; /* # min buffers for wakeup */
209 SYSCTL_INT(_hw_mwl, OID_AUTO, rxdmalow, CTLFLAG_RWTUN, &mwl_rxdmalow,
210 0, "min free rx buffers before restarting traffic");
213 static int mwl_debug = 0;
214 SYSCTL_INT(_hw_mwl, OID_AUTO, debug, CTLFLAG_RWTUN, &mwl_debug,
215 0, "control debugging printfs");
217 MWL_DEBUG_XMIT = 0x00000001, /* basic xmit operation */
218 MWL_DEBUG_XMIT_DESC = 0x00000002, /* xmit descriptors */
219 MWL_DEBUG_RECV = 0x00000004, /* basic recv operation */
220 MWL_DEBUG_RECV_DESC = 0x00000008, /* recv descriptors */
221 MWL_DEBUG_RESET = 0x00000010, /* reset processing */
222 MWL_DEBUG_BEACON = 0x00000020, /* beacon handling */
223 MWL_DEBUG_INTR = 0x00000040, /* ISR */
224 MWL_DEBUG_TX_PROC = 0x00000080, /* tx ISR proc */
225 MWL_DEBUG_RX_PROC = 0x00000100, /* rx ISR proc */
226 MWL_DEBUG_KEYCACHE = 0x00000200, /* key cache management */
227 MWL_DEBUG_STATE = 0x00000400, /* 802.11 state transitions */
228 MWL_DEBUG_NODE = 0x00000800, /* node management */
229 MWL_DEBUG_RECV_ALL = 0x00001000, /* trace all frames (beacons) */
230 MWL_DEBUG_TSO = 0x00002000, /* TSO processing */
231 MWL_DEBUG_AMPDU = 0x00004000, /* BA stream handling */
232 MWL_DEBUG_ANY = 0xffffffff
234 #define IS_BEACON(wh) \
235 ((wh->i_fc[0] & (IEEE80211_FC0_TYPE_MASK|IEEE80211_FC0_SUBTYPE_MASK)) == \
236 (IEEE80211_FC0_TYPE_MGT|IEEE80211_FC0_SUBTYPE_BEACON))
237 #define IFF_DUMPPKTS_RECV(sc, wh) \
238 ((sc->sc_debug & MWL_DEBUG_RECV) && \
239 ((sc->sc_debug & MWL_DEBUG_RECV_ALL) || !IS_BEACON(wh)))
240 #define IFF_DUMPPKTS_XMIT(sc) \
241 (sc->sc_debug & MWL_DEBUG_XMIT)
243 #define DPRINTF(sc, m, fmt, ...) do { \
244 if (sc->sc_debug & (m)) \
245 printf(fmt, __VA_ARGS__); \
247 #define KEYPRINTF(sc, hk, mac) do { \
248 if (sc->sc_debug & MWL_DEBUG_KEYCACHE) \
249 mwl_keyprint(sc, __func__, hk, mac); \
251 static void mwl_printrxbuf(const struct mwl_rxbuf *bf, u_int ix);
252 static void mwl_printtxbuf(const struct mwl_txbuf *bf, u_int qnum, u_int ix);
254 #define IFF_DUMPPKTS_RECV(sc, wh) 0
255 #define IFF_DUMPPKTS_XMIT(sc) 0
256 #define DPRINTF(sc, m, fmt, ...) do { (void )sc; } while (0)
257 #define KEYPRINTF(sc, k, mac) do { (void )sc; } while (0)
260 static MALLOC_DEFINE(M_MWLDEV, "mwldev", "mwl driver dma buffers");
263 * Each packet has fixed front matter: a 2-byte length
264 * of the payload, followed by a 4-address 802.11 header
265 * (regardless of the actual header and always w/o any
266 * QoS header). The payload then follows.
270 struct ieee80211_frame_addr4 wh;
274 * Read/Write shorthands for accesses to BAR 0. Note
275 * that all BAR 1 operations are done in the "hal" and
276 * there should be no reference to them here.
279 static __inline uint32_t
280 RD4(struct mwl_softc *sc, bus_size_t off)
282 return bus_space_read_4(sc->sc_io0t, sc->sc_io0h, off);
287 WR4(struct mwl_softc *sc, bus_size_t off, uint32_t val)
289 bus_space_write_4(sc->sc_io0t, sc->sc_io0h, off, val);
293 mwl_attach(uint16_t devid, struct mwl_softc *sc)
295 struct ieee80211com *ic = &sc->sc_ic;
299 DPRINTF(sc, MWL_DEBUG_ANY, "%s: devid 0x%x\n", __func__, devid);
302 * Setup the RX free list lock early, so it can be consistently
307 mh = mwl_hal_attach(sc->sc_dev, devid,
308 sc->sc_io1h, sc->sc_io1t, sc->sc_dmat);
310 device_printf(sc->sc_dev, "unable to attach HAL\n");
316 * Load firmware so we can get setup. We arbitrarily
317 * pick station firmware; we'll re-load firmware as
318 * needed so setting up the wrong mode isn't a big deal.
320 if (mwl_hal_fwload(mh, NULL) != 0) {
321 device_printf(sc->sc_dev, "unable to setup builtin firmware\n");
325 if (mwl_hal_gethwspecs(mh, &sc->sc_hwspecs) != 0) {
326 device_printf(sc->sc_dev, "unable to fetch h/w specs\n");
330 error = mwl_getchannels(sc);
334 sc->sc_txantenna = 0; /* h/w default */
335 sc->sc_rxantenna = 0; /* h/w default */
336 sc->sc_invalid = 0; /* ready to go, enable int handling */
337 sc->sc_ageinterval = MWL_AGEINTERVAL;
340 * Allocate tx+rx descriptors and populate the lists.
341 * We immediately push the information to the firmware
342 * as otherwise it gets upset.
344 error = mwl_dma_setup(sc);
346 device_printf(sc->sc_dev, "failed to setup descriptors: %d\n",
350 error = mwl_setupdma(sc); /* push to firmware */
351 if (error != 0) /* NB: mwl_setupdma prints msg */
354 callout_init(&sc->sc_timer, 1);
355 callout_init_mtx(&sc->sc_watchdog, &sc->sc_mtx, 0);
356 mbufq_init(&sc->sc_snd, ifqmaxlen);
358 sc->sc_tq = taskqueue_create("mwl_taskq", M_NOWAIT,
359 taskqueue_thread_enqueue, &sc->sc_tq);
360 taskqueue_start_threads(&sc->sc_tq, 1, PI_NET,
361 "%s taskq", device_get_nameunit(sc->sc_dev));
363 TASK_INIT(&sc->sc_rxtask, 0, mwl_rx_proc, sc);
364 TASK_INIT(&sc->sc_radartask, 0, mwl_radar_proc, sc);
365 TASK_INIT(&sc->sc_chanswitchtask, 0, mwl_chanswitch_proc, sc);
366 TASK_INIT(&sc->sc_bawatchdogtask, 0, mwl_bawatchdog_proc, sc);
368 /* NB: insure BK queue is the lowest priority h/w queue */
369 if (!mwl_tx_setup(sc, WME_AC_BK, MWL_WME_AC_BK)) {
370 device_printf(sc->sc_dev,
371 "unable to setup xmit queue for %s traffic!\n",
372 ieee80211_wme_acnames[WME_AC_BK]);
376 if (!mwl_tx_setup(sc, WME_AC_BE, MWL_WME_AC_BE) ||
377 !mwl_tx_setup(sc, WME_AC_VI, MWL_WME_AC_VI) ||
378 !mwl_tx_setup(sc, WME_AC_VO, MWL_WME_AC_VO)) {
380 * Not enough hardware tx queues to properly do WME;
381 * just punt and assign them all to the same h/w queue.
382 * We could do a better job of this if, for example,
383 * we allocate queues when we switch from station to
386 if (sc->sc_ac2q[WME_AC_VI] != NULL)
387 mwl_tx_cleanupq(sc, sc->sc_ac2q[WME_AC_VI]);
388 if (sc->sc_ac2q[WME_AC_BE] != NULL)
389 mwl_tx_cleanupq(sc, sc->sc_ac2q[WME_AC_BE]);
390 sc->sc_ac2q[WME_AC_BE] = sc->sc_ac2q[WME_AC_BK];
391 sc->sc_ac2q[WME_AC_VI] = sc->sc_ac2q[WME_AC_BK];
392 sc->sc_ac2q[WME_AC_VO] = sc->sc_ac2q[WME_AC_BK];
394 TASK_INIT(&sc->sc_txtask, 0, mwl_tx_proc, sc);
397 ic->ic_name = device_get_nameunit(sc->sc_dev);
398 /* XXX not right but it's not used anywhere important */
399 ic->ic_phytype = IEEE80211_T_OFDM;
400 ic->ic_opmode = IEEE80211_M_STA;
402 IEEE80211_C_STA /* station mode supported */
403 | IEEE80211_C_HOSTAP /* hostap mode */
404 | IEEE80211_C_MONITOR /* monitor mode */
406 | IEEE80211_C_IBSS /* ibss, nee adhoc, mode */
407 | IEEE80211_C_AHDEMO /* adhoc demo mode */
409 | IEEE80211_C_MBSS /* mesh point link mode */
410 | IEEE80211_C_WDS /* WDS supported */
411 | IEEE80211_C_SHPREAMBLE /* short preamble supported */
412 | IEEE80211_C_SHSLOT /* short slot time supported */
413 | IEEE80211_C_WME /* WME/WMM supported */
414 | IEEE80211_C_BURST /* xmit bursting supported */
415 | IEEE80211_C_WPA /* capable of WPA1+WPA2 */
416 | IEEE80211_C_BGSCAN /* capable of bg scanning */
417 | IEEE80211_C_TXFRAG /* handle tx frags */
418 | IEEE80211_C_TXPMGT /* capable of txpow mgt */
419 | IEEE80211_C_DFS /* DFS supported */
423 IEEE80211_HTCAP_SMPS_ENA /* SM PS mode enabled */
424 | IEEE80211_HTCAP_CHWIDTH40 /* 40MHz channel width */
425 | IEEE80211_HTCAP_SHORTGI20 /* short GI in 20MHz */
426 | IEEE80211_HTCAP_SHORTGI40 /* short GI in 40MHz */
427 | IEEE80211_HTCAP_RXSTBC_2STREAM/* 1-2 spatial streams */
428 #if MWL_AGGR_SIZE == 7935
429 | IEEE80211_HTCAP_MAXAMSDU_7935 /* max A-MSDU length */
431 | IEEE80211_HTCAP_MAXAMSDU_3839 /* max A-MSDU length */
434 | IEEE80211_HTCAP_PSMP /* PSMP supported */
435 | IEEE80211_HTCAP_40INTOLERANT /* 40MHz intolerant */
437 /* s/w capabilities */
438 | IEEE80211_HTC_HT /* HT operation */
439 | IEEE80211_HTC_AMPDU /* tx A-MPDU */
440 | IEEE80211_HTC_AMSDU /* tx A-MSDU */
441 | IEEE80211_HTC_SMPS /* SMPS available */
445 * Mark h/w crypto support.
446 * XXX no way to query h/w support.
448 ic->ic_cryptocaps |= IEEE80211_CRYPTO_WEP
449 | IEEE80211_CRYPTO_AES_CCM
450 | IEEE80211_CRYPTO_TKIP
451 | IEEE80211_CRYPTO_TKIPMIC
454 * Transmit requires space in the packet for a special
455 * format transmit record and optional padding between
456 * this record and the payload. Ask the net80211 layer
457 * to arrange this when encapsulating packets so we can
458 * add it efficiently.
460 ic->ic_headroom = sizeof(struct mwltxrec) -
461 sizeof(struct ieee80211_frame);
463 IEEE80211_ADDR_COPY(ic->ic_macaddr, sc->sc_hwspecs.macAddr);
465 /* call MI attach routine. */
466 ieee80211_ifattach(ic);
467 ic->ic_setregdomain = mwl_setregdomain;
468 ic->ic_getradiocaps = mwl_getradiocaps;
469 /* override default methods */
470 ic->ic_raw_xmit = mwl_raw_xmit;
471 ic->ic_newassoc = mwl_newassoc;
472 ic->ic_updateslot = mwl_updateslot;
473 ic->ic_update_mcast = mwl_update_mcast;
474 ic->ic_update_promisc = mwl_update_promisc;
475 ic->ic_wme.wme_update = mwl_wme_update;
476 ic->ic_transmit = mwl_transmit;
477 ic->ic_ioctl = mwl_ioctl;
478 ic->ic_parent = mwl_parent;
480 ic->ic_node_alloc = mwl_node_alloc;
481 sc->sc_node_cleanup = ic->ic_node_cleanup;
482 ic->ic_node_cleanup = mwl_node_cleanup;
483 sc->sc_node_drain = ic->ic_node_drain;
484 ic->ic_node_drain = mwl_node_drain;
485 ic->ic_node_getsignal = mwl_node_getsignal;
486 ic->ic_node_getmimoinfo = mwl_node_getmimoinfo;
488 ic->ic_scan_start = mwl_scan_start;
489 ic->ic_scan_end = mwl_scan_end;
490 ic->ic_set_channel = mwl_set_channel;
492 sc->sc_recv_action = ic->ic_recv_action;
493 ic->ic_recv_action = mwl_recv_action;
494 sc->sc_addba_request = ic->ic_addba_request;
495 ic->ic_addba_request = mwl_addba_request;
496 sc->sc_addba_response = ic->ic_addba_response;
497 ic->ic_addba_response = mwl_addba_response;
498 sc->sc_addba_stop = ic->ic_addba_stop;
499 ic->ic_addba_stop = mwl_addba_stop;
501 ic->ic_vap_create = mwl_vap_create;
502 ic->ic_vap_delete = mwl_vap_delete;
504 ieee80211_radiotap_attach(ic,
505 &sc->sc_tx_th.wt_ihdr, sizeof(sc->sc_tx_th),
506 MWL_TX_RADIOTAP_PRESENT,
507 &sc->sc_rx_th.wr_ihdr, sizeof(sc->sc_rx_th),
508 MWL_RX_RADIOTAP_PRESENT);
510 * Setup dynamic sysctl's now that country code and
511 * regdomain are available from the hal.
513 mwl_sysctlattach(sc);
516 ieee80211_announce(ic);
524 MWL_RXFREE_DESTROY(sc);
530 mwl_detach(struct mwl_softc *sc)
532 struct ieee80211com *ic = &sc->sc_ic;
538 * NB: the order of these is important:
539 * o call the 802.11 layer before detaching the hal to
540 * insure callbacks into the driver to delete global
541 * key cache entries can be handled
542 * o reclaim the tx queue data structures after calling
543 * the 802.11 layer as we'll get called back to reclaim
544 * node state and potentially want to use them
545 * o to cleanup the tx queues the hal is called, so detach
547 * Other than that, it's straightforward...
549 ieee80211_ifdetach(ic);
550 callout_drain(&sc->sc_watchdog);
552 MWL_RXFREE_DESTROY(sc);
554 mwl_hal_detach(sc->sc_mh);
555 mbufq_drain(&sc->sc_snd);
561 * MAC address handling for multiple BSS on the same radio.
562 * The first vap uses the MAC address from the EEPROM. For
563 * subsequent vap's we set the U/L bit (bit 1) in the MAC
564 * address and use the next six bits as an index.
567 assign_address(struct mwl_softc *sc, uint8_t mac[IEEE80211_ADDR_LEN], int clone)
571 if (clone && mwl_hal_ismbsscapable(sc->sc_mh)) {
572 /* NB: we only do this if h/w supports multiple bssid */
573 for (i = 0; i < 32; i++)
574 if ((sc->sc_bssidmask & (1<<i)) == 0)
577 mac[0] |= (i << 2)|0x2;
580 sc->sc_bssidmask |= 1<<i;
586 reclaim_address(struct mwl_softc *sc, const uint8_t mac[IEEE80211_ADDR_LEN])
589 if (i != 0 || --sc->sc_nbssid0 == 0)
590 sc->sc_bssidmask &= ~(1<<i);
593 static struct ieee80211vap *
594 mwl_vap_create(struct ieee80211com *ic, const char name[IFNAMSIZ], int unit,
595 enum ieee80211_opmode opmode, int flags,
596 const uint8_t bssid[IEEE80211_ADDR_LEN],
597 const uint8_t mac0[IEEE80211_ADDR_LEN])
599 struct mwl_softc *sc = ic->ic_softc;
600 struct mwl_hal *mh = sc->sc_mh;
601 struct ieee80211vap *vap, *apvap;
602 struct mwl_hal_vap *hvap;
604 uint8_t mac[IEEE80211_ADDR_LEN];
606 IEEE80211_ADDR_COPY(mac, mac0);
608 case IEEE80211_M_HOSTAP:
609 case IEEE80211_M_MBSS:
610 if ((flags & IEEE80211_CLONE_MACADDR) == 0)
611 assign_address(sc, mac, flags & IEEE80211_CLONE_BSSID);
612 hvap = mwl_hal_newvap(mh, MWL_HAL_AP, mac);
614 if ((flags & IEEE80211_CLONE_MACADDR) == 0)
615 reclaim_address(sc, mac);
619 case IEEE80211_M_STA:
620 if ((flags & IEEE80211_CLONE_MACADDR) == 0)
621 assign_address(sc, mac, flags & IEEE80211_CLONE_BSSID);
622 hvap = mwl_hal_newvap(mh, MWL_HAL_STA, mac);
624 if ((flags & IEEE80211_CLONE_MACADDR) == 0)
625 reclaim_address(sc, mac);
628 /* no h/w beacon miss support; always use s/w */
629 flags |= IEEE80211_CLONE_NOBEACONS;
631 case IEEE80211_M_WDS:
632 hvap = NULL; /* NB: we use associated AP vap */
633 if (sc->sc_napvaps == 0)
634 return NULL; /* no existing AP vap */
636 case IEEE80211_M_MONITOR:
639 case IEEE80211_M_IBSS:
640 case IEEE80211_M_AHDEMO:
645 mvp = malloc(sizeof(struct mwl_vap), M_80211_VAP, M_WAITOK | M_ZERO);
647 if (opmode == IEEE80211_M_WDS) {
649 * WDS vaps must have an associated AP vap; find one.
652 TAILQ_FOREACH(apvap, &ic->ic_vaps, iv_next)
653 if (apvap->iv_opmode == IEEE80211_M_HOSTAP) {
654 mvp->mv_ap_hvap = MWL_VAP(apvap)->mv_hvap;
657 KASSERT(mvp->mv_ap_hvap != NULL, ("no ap vap"));
660 ieee80211_vap_setup(ic, vap, name, unit, opmode, flags, bssid);
661 /* override with driver methods */
662 mvp->mv_newstate = vap->iv_newstate;
663 vap->iv_newstate = mwl_newstate;
664 vap->iv_max_keyix = 0; /* XXX */
665 vap->iv_key_alloc = mwl_key_alloc;
666 vap->iv_key_delete = mwl_key_delete;
667 vap->iv_key_set = mwl_key_set;
668 #ifdef MWL_HOST_PS_SUPPORT
669 if (opmode == IEEE80211_M_HOSTAP || opmode == IEEE80211_M_MBSS) {
670 vap->iv_update_ps = mwl_update_ps;
671 mvp->mv_set_tim = vap->iv_set_tim;
672 vap->iv_set_tim = mwl_set_tim;
675 vap->iv_reset = mwl_reset;
676 vap->iv_update_beacon = mwl_beacon_update;
678 /* override max aid so sta's cannot assoc when we're out of sta id's */
679 vap->iv_max_aid = MWL_MAXSTAID;
680 /* override default A-MPDU rx parameters */
681 vap->iv_ampdu_rxmax = IEEE80211_HTCAP_MAXRXAMPDU_64K;
682 vap->iv_ampdu_density = IEEE80211_HTCAP_MPDUDENSITY_4;
685 ieee80211_vap_attach(vap, mwl_media_change, ieee80211_media_status,
688 switch (vap->iv_opmode) {
689 case IEEE80211_M_HOSTAP:
690 case IEEE80211_M_MBSS:
691 case IEEE80211_M_STA:
693 * Setup sta db entry for local address.
696 if (vap->iv_opmode == IEEE80211_M_HOSTAP ||
697 vap->iv_opmode == IEEE80211_M_MBSS)
702 case IEEE80211_M_WDS:
709 * Setup overall operating mode.
712 ic->ic_opmode = IEEE80211_M_HOSTAP;
713 else if (sc->sc_nstavaps)
714 ic->ic_opmode = IEEE80211_M_STA;
716 ic->ic_opmode = opmode;
722 mwl_vap_delete(struct ieee80211vap *vap)
724 struct mwl_vap *mvp = MWL_VAP(vap);
725 struct mwl_softc *sc = vap->iv_ic->ic_softc;
726 struct mwl_hal *mh = sc->sc_mh;
727 struct mwl_hal_vap *hvap = mvp->mv_hvap;
728 enum ieee80211_opmode opmode = vap->iv_opmode;
730 /* XXX disallow ap vap delete if WDS still present */
731 if (sc->sc_running) {
732 /* quiesce h/w while we remove the vap */
733 mwl_hal_intrset(mh, 0); /* disable interrupts */
735 ieee80211_vap_detach(vap);
737 case IEEE80211_M_HOSTAP:
738 case IEEE80211_M_MBSS:
739 case IEEE80211_M_STA:
740 KASSERT(hvap != NULL, ("no hal vap handle"));
741 (void) mwl_hal_delstation(hvap, vap->iv_myaddr);
742 mwl_hal_delvap(hvap);
743 if (opmode == IEEE80211_M_HOSTAP || opmode == IEEE80211_M_MBSS)
747 /* XXX don't do it for IEEE80211_CLONE_MACADDR */
748 reclaim_address(sc, vap->iv_myaddr);
750 case IEEE80211_M_WDS:
756 mwl_cleartxq(sc, vap);
757 free(mvp, M_80211_VAP);
759 mwl_hal_intrset(mh, sc->sc_imask);
763 mwl_suspend(struct mwl_softc *sc)
772 mwl_resume(struct mwl_softc *sc)
777 if (sc->sc_ic.ic_nrunning > 0)
778 error = mwl_init(sc);
782 ieee80211_start_all(&sc->sc_ic); /* start all vap's */
786 mwl_shutdown(void *arg)
788 struct mwl_softc *sc = arg;
796 * Interrupt handler. Most of the actual processing is deferred.
801 struct mwl_softc *sc = arg;
802 struct mwl_hal *mh = sc->sc_mh;
805 if (sc->sc_invalid) {
807 * The hardware is not ready/present, don't touch anything.
808 * Note this can happen early on if the IRQ is shared.
810 DPRINTF(sc, MWL_DEBUG_ANY, "%s: invalid; ignored\n", __func__);
814 * Figure out the reason(s) for the interrupt.
816 mwl_hal_getisr(mh, &status); /* NB: clears ISR too */
817 if (status == 0) /* must be a shared irq */
820 DPRINTF(sc, MWL_DEBUG_INTR, "%s: status 0x%x imask 0x%x\n",
821 __func__, status, sc->sc_imask);
822 if (status & MACREG_A2HRIC_BIT_RX_RDY)
823 taskqueue_enqueue(sc->sc_tq, &sc->sc_rxtask);
824 if (status & MACREG_A2HRIC_BIT_TX_DONE)
825 taskqueue_enqueue(sc->sc_tq, &sc->sc_txtask);
826 if (status & MACREG_A2HRIC_BIT_BA_WATCHDOG)
827 taskqueue_enqueue(sc->sc_tq, &sc->sc_bawatchdogtask);
828 if (status & MACREG_A2HRIC_BIT_OPC_DONE)
830 if (status & MACREG_A2HRIC_BIT_MAC_EVENT) {
833 if (status & MACREG_A2HRIC_BIT_ICV_ERROR) {
835 sc->sc_stats.mst_rx_badtkipicv++;
837 if (status & MACREG_A2HRIC_BIT_QUEUE_EMPTY) {
838 /* 11n aggregation queue is empty, re-fill */
841 if (status & MACREG_A2HRIC_BIT_QUEUE_FULL) {
844 if (status & MACREG_A2HRIC_BIT_RADAR_DETECT) {
845 /* radar detected, process event */
846 taskqueue_enqueue(sc->sc_tq, &sc->sc_radartask);
848 if (status & MACREG_A2HRIC_BIT_CHAN_SWITCH) {
849 /* DFS channel switch */
850 taskqueue_enqueue(sc->sc_tq, &sc->sc_chanswitchtask);
855 mwl_radar_proc(void *arg, int pending)
857 struct mwl_softc *sc = arg;
858 struct ieee80211com *ic = &sc->sc_ic;
860 DPRINTF(sc, MWL_DEBUG_ANY, "%s: radar detected, pending %u\n",
863 sc->sc_stats.mst_radardetect++;
864 /* XXX stop h/w BA streams? */
867 ieee80211_dfs_notify_radar(ic, ic->ic_curchan);
868 IEEE80211_UNLOCK(ic);
872 mwl_chanswitch_proc(void *arg, int pending)
874 struct mwl_softc *sc = arg;
875 struct ieee80211com *ic = &sc->sc_ic;
877 DPRINTF(sc, MWL_DEBUG_ANY, "%s: channel switch notice, pending %u\n",
881 sc->sc_csapending = 0;
882 ieee80211_csa_completeswitch(ic);
883 IEEE80211_UNLOCK(ic);
887 mwl_bawatchdog(const MWL_HAL_BASTREAM *sp)
889 struct ieee80211_node *ni = sp->data[0];
891 /* send DELBA and drop the stream */
892 ieee80211_ampdu_stop(ni, sp->data[1], IEEE80211_REASON_UNSPECIFIED);
896 mwl_bawatchdog_proc(void *arg, int pending)
898 struct mwl_softc *sc = arg;
899 struct mwl_hal *mh = sc->sc_mh;
900 const MWL_HAL_BASTREAM *sp;
903 sc->sc_stats.mst_bawatchdog++;
905 if (mwl_hal_getwatchdogbitmap(mh, &bitmap) != 0) {
906 DPRINTF(sc, MWL_DEBUG_AMPDU,
907 "%s: could not get bitmap\n", __func__);
908 sc->sc_stats.mst_bawatchdog_failed++;
911 DPRINTF(sc, MWL_DEBUG_AMPDU, "%s: bitmap 0x%x\n", __func__, bitmap);
912 if (bitmap == 0xff) {
914 /* disable all ba streams */
915 for (bitmap = 0; bitmap < 8; bitmap++) {
916 sp = mwl_hal_bastream_lookup(mh, bitmap);
923 DPRINTF(sc, MWL_DEBUG_AMPDU,
924 "%s: no BA streams found\n", __func__);
925 sc->sc_stats.mst_bawatchdog_empty++;
927 } else if (bitmap != 0xaa) {
928 /* disable a single ba stream */
929 sp = mwl_hal_bastream_lookup(mh, bitmap);
933 DPRINTF(sc, MWL_DEBUG_AMPDU,
934 "%s: no BA stream %d\n", __func__, bitmap);
935 sc->sc_stats.mst_bawatchdog_notfound++;
941 * Convert net80211 channel to a HAL channel.
944 mwl_mapchan(MWL_HAL_CHANNEL *hc, const struct ieee80211_channel *chan)
946 hc->channel = chan->ic_ieee;
948 *(uint32_t *)&hc->channelFlags = 0;
949 if (IEEE80211_IS_CHAN_2GHZ(chan))
950 hc->channelFlags.FreqBand = MWL_FREQ_BAND_2DOT4GHZ;
951 else if (IEEE80211_IS_CHAN_5GHZ(chan))
952 hc->channelFlags.FreqBand = MWL_FREQ_BAND_5GHZ;
953 if (IEEE80211_IS_CHAN_HT40(chan)) {
954 hc->channelFlags.ChnlWidth = MWL_CH_40_MHz_WIDTH;
955 if (IEEE80211_IS_CHAN_HT40U(chan))
956 hc->channelFlags.ExtChnlOffset = MWL_EXT_CH_ABOVE_CTRL_CH;
958 hc->channelFlags.ExtChnlOffset = MWL_EXT_CH_BELOW_CTRL_CH;
960 hc->channelFlags.ChnlWidth = MWL_CH_20_MHz_WIDTH;
961 /* XXX 10MHz channels */
965 * Inform firmware of our tx/rx dma setup. The BAR 0
966 * writes below are for compatibility with older firmware.
967 * For current firmware we send this information with a
968 * cmd block via mwl_hal_sethwdma.
971 mwl_setupdma(struct mwl_softc *sc)
975 sc->sc_hwdma.rxDescRead = sc->sc_rxdma.dd_desc_paddr;
976 WR4(sc, sc->sc_hwspecs.rxDescRead, sc->sc_hwdma.rxDescRead);
977 WR4(sc, sc->sc_hwspecs.rxDescWrite, sc->sc_hwdma.rxDescRead);
979 for (i = 0; i < MWL_NUM_TX_QUEUES-MWL_NUM_ACK_QUEUES; i++) {
980 struct mwl_txq *txq = &sc->sc_txq[i];
981 sc->sc_hwdma.wcbBase[i] = txq->dma.dd_desc_paddr;
982 WR4(sc, sc->sc_hwspecs.wcbBase[i], sc->sc_hwdma.wcbBase[i]);
984 sc->sc_hwdma.maxNumTxWcb = mwl_txbuf;
985 sc->sc_hwdma.maxNumWCB = MWL_NUM_TX_QUEUES-MWL_NUM_ACK_QUEUES;
987 error = mwl_hal_sethwdma(sc->sc_mh, &sc->sc_hwdma);
989 device_printf(sc->sc_dev,
990 "unable to setup tx/rx dma; hal status %u\n", error);
997 * Inform firmware of tx rate parameters.
998 * Called after a channel change.
1001 mwl_setcurchanrates(struct mwl_softc *sc)
1003 struct ieee80211com *ic = &sc->sc_ic;
1004 const struct ieee80211_rateset *rs;
1005 MWL_HAL_TXRATE rates;
1007 memset(&rates, 0, sizeof(rates));
1008 rs = ieee80211_get_suprates(ic, ic->ic_curchan);
1009 /* rate used to send management frames */
1010 rates.MgtRate = rs->rs_rates[0] & IEEE80211_RATE_VAL;
1011 /* rate used to send multicast frames */
1012 rates.McastRate = rates.MgtRate;
1014 return mwl_hal_settxrate_auto(sc->sc_mh, &rates);
1018 * Inform firmware of tx rate parameters. Called whenever
1019 * user-settable params change and after a channel change.
1022 mwl_setrates(struct ieee80211vap *vap)
1024 struct mwl_vap *mvp = MWL_VAP(vap);
1025 struct ieee80211_node *ni = vap->iv_bss;
1026 const struct ieee80211_txparam *tp = ni->ni_txparms;
1027 MWL_HAL_TXRATE rates;
1029 KASSERT(vap->iv_state == IEEE80211_S_RUN, ("state %d", vap->iv_state));
1032 * Update the h/w rate map.
1033 * NB: 0x80 for MCS is passed through unchanged
1035 memset(&rates, 0, sizeof(rates));
1036 /* rate used to send management frames */
1037 rates.MgtRate = tp->mgmtrate;
1038 /* rate used to send multicast frames */
1039 rates.McastRate = tp->mcastrate;
1041 /* while here calculate EAPOL fixed rate cookie */
1042 mvp->mv_eapolformat = htole16(mwl_calcformat(rates.MgtRate, ni));
1044 return mwl_hal_settxrate(mvp->mv_hvap,
1045 tp->ucastrate != IEEE80211_FIXED_RATE_NONE ?
1046 RATE_FIXED : RATE_AUTO, &rates);
1050 * Setup a fixed xmit rate cookie for EAPOL frames.
1053 mwl_seteapolformat(struct ieee80211vap *vap)
1055 struct mwl_vap *mvp = MWL_VAP(vap);
1056 struct ieee80211_node *ni = vap->iv_bss;
1057 enum ieee80211_phymode mode;
1060 KASSERT(vap->iv_state == IEEE80211_S_RUN, ("state %d", vap->iv_state));
1062 mode = ieee80211_chan2mode(ni->ni_chan);
1064 * Use legacy rates when operating a mixed HT+non-HT bss.
1065 * NB: this may violate POLA for sta and wds vap's.
1067 if (mode == IEEE80211_MODE_11NA &&
1068 (vap->iv_flags_ht & IEEE80211_FHT_PUREN) == 0)
1069 rate = vap->iv_txparms[IEEE80211_MODE_11A].mgmtrate;
1070 else if (mode == IEEE80211_MODE_11NG &&
1071 (vap->iv_flags_ht & IEEE80211_FHT_PUREN) == 0)
1072 rate = vap->iv_txparms[IEEE80211_MODE_11G].mgmtrate;
1074 rate = vap->iv_txparms[mode].mgmtrate;
1076 mvp->mv_eapolformat = htole16(mwl_calcformat(rate, ni));
1080 * Map SKU+country code to region code for radar bin'ing.
1083 mwl_map2regioncode(const struct ieee80211_regdomain *rd)
1085 switch (rd->regdomain) {
1088 return DOMAIN_CODE_FCC;
1090 return DOMAIN_CODE_IC;
1094 if (rd->country == CTRY_SPAIN)
1095 return DOMAIN_CODE_SPAIN;
1096 if (rd->country == CTRY_FRANCE || rd->country == CTRY_FRANCE2)
1097 return DOMAIN_CODE_FRANCE;
1098 /* XXX force 1.3.1 radar type */
1099 return DOMAIN_CODE_ETSI_131;
1101 return DOMAIN_CODE_MKK;
1103 return DOMAIN_CODE_DGT; /* Taiwan */
1107 return DOMAIN_CODE_AUS; /* Australia */
1110 return DOMAIN_CODE_FCC; /* XXX? */
1114 mwl_hal_reset(struct mwl_softc *sc)
1116 struct ieee80211com *ic = &sc->sc_ic;
1117 struct mwl_hal *mh = sc->sc_mh;
1119 mwl_hal_setantenna(mh, WL_ANTENNATYPE_RX, sc->sc_rxantenna);
1120 mwl_hal_setantenna(mh, WL_ANTENNATYPE_TX, sc->sc_txantenna);
1121 mwl_hal_setradio(mh, 1, WL_AUTO_PREAMBLE);
1122 mwl_hal_setwmm(sc->sc_mh, (ic->ic_flags & IEEE80211_F_WME) != 0);
1123 mwl_chan_set(sc, ic->ic_curchan);
1124 /* NB: RF/RA performance tuned for indoor mode */
1125 mwl_hal_setrateadaptmode(mh, 0);
1126 mwl_hal_setoptimizationlevel(mh,
1127 (ic->ic_flags & IEEE80211_F_BURST) != 0);
1129 mwl_hal_setregioncode(mh, mwl_map2regioncode(&ic->ic_regdomain));
1131 mwl_hal_setaggampduratemode(mh, 1, 80); /* XXX */
1132 mwl_hal_setcfend(mh, 0); /* XXX */
1138 mwl_init(struct mwl_softc *sc)
1140 struct mwl_hal *mh = sc->sc_mh;
1143 MWL_LOCK_ASSERT(sc);
1146 * Stop anything previously setup. This is safe
1147 * whether this is the first time through or not.
1152 * Push vap-independent state to the firmware.
1154 if (!mwl_hal_reset(sc)) {
1155 device_printf(sc->sc_dev, "unable to reset hardware\n");
1160 * Setup recv (once); transmit is already good to go.
1162 error = mwl_startrecv(sc);
1164 device_printf(sc->sc_dev, "unable to start recv logic\n");
1169 * Enable interrupts.
1171 sc->sc_imask = MACREG_A2HRIC_BIT_RX_RDY
1172 | MACREG_A2HRIC_BIT_TX_DONE
1173 | MACREG_A2HRIC_BIT_OPC_DONE
1175 | MACREG_A2HRIC_BIT_MAC_EVENT
1177 | MACREG_A2HRIC_BIT_ICV_ERROR
1178 | MACREG_A2HRIC_BIT_RADAR_DETECT
1179 | MACREG_A2HRIC_BIT_CHAN_SWITCH
1181 | MACREG_A2HRIC_BIT_QUEUE_EMPTY
1183 | MACREG_A2HRIC_BIT_BA_WATCHDOG
1184 | MACREQ_A2HRIC_BIT_TX_ACK
1188 mwl_hal_intrset(mh, sc->sc_imask);
1189 callout_reset(&sc->sc_watchdog, hz, mwl_watchdog, sc);
1195 mwl_stop(struct mwl_softc *sc)
1198 MWL_LOCK_ASSERT(sc);
1199 if (sc->sc_running) {
1201 * Shutdown the hardware and driver.
1204 callout_stop(&sc->sc_watchdog);
1205 sc->sc_tx_timer = 0;
1211 mwl_reset_vap(struct ieee80211vap *vap, int state)
1213 struct mwl_hal_vap *hvap = MWL_VAP(vap)->mv_hvap;
1214 struct ieee80211com *ic = vap->iv_ic;
1216 if (state == IEEE80211_S_RUN)
1219 mwl_hal_setrtsthreshold(hvap, vap->iv_rtsthreshold);
1220 /* XXX auto? 20/40 split? */
1221 mwl_hal_sethtgi(hvap, (vap->iv_flags_ht &
1222 (IEEE80211_FHT_SHORTGI20|IEEE80211_FHT_SHORTGI40)) ? 1 : 0);
1223 mwl_hal_setnprot(hvap, ic->ic_htprotmode == IEEE80211_PROT_NONE ?
1224 HTPROTECT_NONE : HTPROTECT_AUTO);
1225 /* XXX txpower cap */
1227 /* re-setup beacons */
1228 if (state == IEEE80211_S_RUN &&
1229 (vap->iv_opmode == IEEE80211_M_HOSTAP ||
1230 vap->iv_opmode == IEEE80211_M_MBSS ||
1231 vap->iv_opmode == IEEE80211_M_IBSS)) {
1232 mwl_setapmode(vap, vap->iv_bss->ni_chan);
1233 mwl_hal_setnprotmode(hvap,
1234 MS(ic->ic_curhtprotmode, IEEE80211_HTINFO_OPMODE));
1235 return mwl_beacon_setup(vap);
1241 * Reset the hardware w/o losing operational state.
1242 * Used to reset or reload hardware state for a vap.
1245 mwl_reset(struct ieee80211vap *vap, u_long cmd)
1247 struct mwl_hal_vap *hvap = MWL_VAP(vap)->mv_hvap;
1250 if (hvap != NULL) { /* WDS, MONITOR, etc. */
1251 struct ieee80211com *ic = vap->iv_ic;
1252 struct mwl_softc *sc = ic->ic_softc;
1253 struct mwl_hal *mh = sc->sc_mh;
1255 /* XXX handle DWDS sta vap change */
1256 /* XXX do we need to disable interrupts? */
1257 mwl_hal_intrset(mh, 0); /* disable interrupts */
1258 error = mwl_reset_vap(vap, vap->iv_state);
1259 mwl_hal_intrset(mh, sc->sc_imask);
1265 * Allocate a tx buffer for sending a frame. The
1266 * packet is assumed to have the WME AC stored so
1267 * we can use it to select the appropriate h/w queue.
1269 static struct mwl_txbuf *
1270 mwl_gettxbuf(struct mwl_softc *sc, struct mwl_txq *txq)
1272 struct mwl_txbuf *bf;
1275 * Grab a TX buffer and associated resources.
1278 bf = STAILQ_FIRST(&txq->free);
1280 STAILQ_REMOVE_HEAD(&txq->free, bf_list);
1283 MWL_TXQ_UNLOCK(txq);
1285 DPRINTF(sc, MWL_DEBUG_XMIT,
1286 "%s: out of xmit buffers on q %d\n", __func__, txq->qnum);
1291 * Return a tx buffer to the queue it came from. Note there
1292 * are two cases because we must preserve the order of buffers
1293 * as it reflects the fixed order of descriptors in memory
1294 * (the firmware pre-fetches descriptors so we cannot reorder).
1297 mwl_puttxbuf_head(struct mwl_txq *txq, struct mwl_txbuf *bf)
1302 STAILQ_INSERT_HEAD(&txq->free, bf, bf_list);
1304 MWL_TXQ_UNLOCK(txq);
1308 mwl_puttxbuf_tail(struct mwl_txq *txq, struct mwl_txbuf *bf)
1313 STAILQ_INSERT_TAIL(&txq->free, bf, bf_list);
1315 MWL_TXQ_UNLOCK(txq);
1319 mwl_transmit(struct ieee80211com *ic, struct mbuf *m)
1321 struct mwl_softc *sc = ic->ic_softc;
1325 if (!sc->sc_running) {
1329 error = mbufq_enqueue(&sc->sc_snd, m);
1340 mwl_start(struct mwl_softc *sc)
1342 struct ieee80211_node *ni;
1343 struct mwl_txbuf *bf;
1345 struct mwl_txq *txq = NULL; /* XXX silence gcc */
1348 MWL_LOCK_ASSERT(sc);
1349 if (!sc->sc_running || sc->sc_invalid)
1352 while ((m = mbufq_dequeue(&sc->sc_snd)) != NULL) {
1354 * Grab the node for the destination.
1356 ni = (struct ieee80211_node *) m->m_pkthdr.rcvif;
1357 KASSERT(ni != NULL, ("no node"));
1358 m->m_pkthdr.rcvif = NULL; /* committed, clear ref */
1360 * Grab a TX buffer and associated resources.
1361 * We honor the classification by the 802.11 layer.
1363 txq = sc->sc_ac2q[M_WME_GETAC(m)];
1364 bf = mwl_gettxbuf(sc, txq);
1367 ieee80211_free_node(ni);
1368 #ifdef MWL_TX_NODROP
1369 sc->sc_stats.mst_tx_qstop++;
1372 DPRINTF(sc, MWL_DEBUG_XMIT,
1373 "%s: tail drop on q %d\n", __func__, txq->qnum);
1374 sc->sc_stats.mst_tx_qdrop++;
1376 #endif /* MWL_TX_NODROP */
1380 * Pass the frame to the h/w for transmission.
1382 if (mwl_tx_start(sc, ni, bf, m)) {
1383 if_inc_counter(ni->ni_vap->iv_ifp,
1384 IFCOUNTER_OERRORS, 1);
1385 mwl_puttxbuf_head(txq, bf);
1386 ieee80211_free_node(ni);
1390 if (nqueued >= mwl_txcoalesce) {
1392 * Poke the firmware to process queued frames;
1393 * see below about (lack of) locking.
1396 mwl_hal_txstart(sc->sc_mh, 0/*XXX*/);
1401 * NB: We don't need to lock against tx done because
1402 * this just prods the firmware to check the transmit
1403 * descriptors. The firmware will also start fetching
1404 * descriptors by itself if it notices new ones are
1405 * present when it goes to deliver a tx done interrupt
1406 * to the host. So if we race with tx done processing
1407 * it's ok. Delivering the kick here rather than in
1408 * mwl_tx_start is an optimization to avoid poking the
1409 * firmware for each packet.
1411 * NB: the queue id isn't used so 0 is ok.
1413 mwl_hal_txstart(sc->sc_mh, 0/*XXX*/);
1418 mwl_raw_xmit(struct ieee80211_node *ni, struct mbuf *m,
1419 const struct ieee80211_bpf_params *params)
1421 struct ieee80211com *ic = ni->ni_ic;
1422 struct mwl_softc *sc = ic->ic_softc;
1423 struct mwl_txbuf *bf;
1424 struct mwl_txq *txq;
1426 if (!sc->sc_running || sc->sc_invalid) {
1431 * Grab a TX buffer and associated resources.
1432 * Note that we depend on the classification
1433 * by the 802.11 layer to get to the right h/w
1434 * queue. Management frames must ALWAYS go on
1435 * queue 1 but we cannot just force that here
1436 * because we may receive non-mgt frames.
1438 txq = sc->sc_ac2q[M_WME_GETAC(m)];
1439 bf = mwl_gettxbuf(sc, txq);
1441 sc->sc_stats.mst_tx_qstop++;
1446 * Pass the frame to the h/w for transmission.
1448 if (mwl_tx_start(sc, ni, bf, m)) {
1449 mwl_puttxbuf_head(txq, bf);
1451 return EIO; /* XXX */
1454 * NB: We don't need to lock against tx done because
1455 * this just prods the firmware to check the transmit
1456 * descriptors. The firmware will also start fetching
1457 * descriptors by itself if it notices new ones are
1458 * present when it goes to deliver a tx done interrupt
1459 * to the host. So if we race with tx done processing
1460 * it's ok. Delivering the kick here rather than in
1461 * mwl_tx_start is an optimization to avoid poking the
1462 * firmware for each packet.
1464 * NB: the queue id isn't used so 0 is ok.
1466 mwl_hal_txstart(sc->sc_mh, 0/*XXX*/);
1471 mwl_media_change(struct ifnet *ifp)
1473 struct ieee80211vap *vap = ifp->if_softc;
1476 error = ieee80211_media_change(ifp);
1477 /* NB: only the fixed rate can change and that doesn't need a reset */
1478 if (error == ENETRESET) {
1487 mwl_keyprint(struct mwl_softc *sc, const char *tag,
1488 const MWL_HAL_KEYVAL *hk, const uint8_t mac[IEEE80211_ADDR_LEN])
1490 static const char *ciphers[] = {
1497 printf("%s: [%u] %-7s", tag, hk->keyIndex, ciphers[hk->keyTypeId]);
1498 for (i = 0, n = hk->keyLen; i < n; i++)
1499 printf(" %02x", hk->key.aes[i]);
1500 printf(" mac %s", ether_sprintf(mac));
1501 if (hk->keyTypeId == KEY_TYPE_ID_TKIP) {
1502 printf(" %s", "rxmic");
1503 for (i = 0; i < sizeof(hk->key.tkip.rxMic); i++)
1504 printf(" %02x", hk->key.tkip.rxMic[i]);
1506 for (i = 0; i < sizeof(hk->key.tkip.txMic); i++)
1507 printf(" %02x", hk->key.tkip.txMic[i]);
1509 printf(" flags 0x%x\n", hk->keyFlags);
1514 * Allocate a key cache slot for a unicast key. The
1515 * firmware handles key allocation and every station is
1516 * guaranteed key space so we are always successful.
1519 mwl_key_alloc(struct ieee80211vap *vap, struct ieee80211_key *k,
1520 ieee80211_keyix *keyix, ieee80211_keyix *rxkeyix)
1522 struct mwl_softc *sc = vap->iv_ic->ic_softc;
1524 if (k->wk_keyix != IEEE80211_KEYIX_NONE ||
1525 (k->wk_flags & IEEE80211_KEY_GROUP)) {
1526 if (!(&vap->iv_nw_keys[0] <= k &&
1527 k < &vap->iv_nw_keys[IEEE80211_WEP_NKID])) {
1528 /* should not happen */
1529 DPRINTF(sc, MWL_DEBUG_KEYCACHE,
1530 "%s: bogus group key\n", __func__);
1533 /* give the caller what they requested */
1534 *keyix = *rxkeyix = ieee80211_crypto_get_key_wepidx(vap, k);
1537 * Firmware handles key allocation.
1539 *keyix = *rxkeyix = 0;
1545 * Delete a key entry allocated by mwl_key_alloc.
1548 mwl_key_delete(struct ieee80211vap *vap, const struct ieee80211_key *k)
1550 struct mwl_softc *sc = vap->iv_ic->ic_softc;
1551 struct mwl_hal_vap *hvap = MWL_VAP(vap)->mv_hvap;
1553 const uint8_t bcastaddr[IEEE80211_ADDR_LEN] =
1554 { 0xff, 0xff, 0xff, 0xff, 0xff, 0xff };
1557 if (vap->iv_opmode != IEEE80211_M_WDS) {
1558 /* XXX monitor mode? */
1559 DPRINTF(sc, MWL_DEBUG_KEYCACHE,
1560 "%s: no hvap for opmode %d\n", __func__,
1564 hvap = MWL_VAP(vap)->mv_ap_hvap;
1567 DPRINTF(sc, MWL_DEBUG_KEYCACHE, "%s: delete key %u\n",
1568 __func__, k->wk_keyix);
1570 memset(&hk, 0, sizeof(hk));
1571 hk.keyIndex = k->wk_keyix;
1572 switch (k->wk_cipher->ic_cipher) {
1573 case IEEE80211_CIPHER_WEP:
1574 hk.keyTypeId = KEY_TYPE_ID_WEP;
1576 case IEEE80211_CIPHER_TKIP:
1577 hk.keyTypeId = KEY_TYPE_ID_TKIP;
1579 case IEEE80211_CIPHER_AES_CCM:
1580 hk.keyTypeId = KEY_TYPE_ID_AES;
1583 /* XXX should not happen */
1584 DPRINTF(sc, MWL_DEBUG_KEYCACHE, "%s: unknown cipher %d\n",
1585 __func__, k->wk_cipher->ic_cipher);
1588 return (mwl_hal_keyreset(hvap, &hk, bcastaddr) == 0); /*XXX*/
1592 addgroupflags(MWL_HAL_KEYVAL *hk, const struct ieee80211_key *k)
1594 if (k->wk_flags & IEEE80211_KEY_GROUP) {
1595 if (k->wk_flags & IEEE80211_KEY_XMIT)
1596 hk->keyFlags |= KEY_FLAG_TXGROUPKEY;
1597 if (k->wk_flags & IEEE80211_KEY_RECV)
1598 hk->keyFlags |= KEY_FLAG_RXGROUPKEY;
1605 * Set the key cache contents for the specified key. Key cache
1606 * slot(s) must already have been allocated by mwl_key_alloc.
1609 mwl_key_set(struct ieee80211vap *vap, const struct ieee80211_key *k)
1611 return (_mwl_key_set(vap, k, k->wk_macaddr));
1615 _mwl_key_set(struct ieee80211vap *vap, const struct ieee80211_key *k,
1616 const uint8_t mac[IEEE80211_ADDR_LEN])
1618 #define GRPXMIT (IEEE80211_KEY_XMIT | IEEE80211_KEY_GROUP)
1619 /* NB: static wep keys are marked GROUP+tx/rx; GTK will be tx or rx */
1620 #define IEEE80211_IS_STATICKEY(k) \
1621 (((k)->wk_flags & (GRPXMIT|IEEE80211_KEY_RECV)) == \
1622 (GRPXMIT|IEEE80211_KEY_RECV))
1623 struct mwl_softc *sc = vap->iv_ic->ic_softc;
1624 struct mwl_hal_vap *hvap = MWL_VAP(vap)->mv_hvap;
1625 const struct ieee80211_cipher *cip = k->wk_cipher;
1626 const uint8_t *macaddr;
1629 KASSERT((k->wk_flags & IEEE80211_KEY_SWCRYPT) == 0,
1630 ("s/w crypto set?"));
1633 if (vap->iv_opmode != IEEE80211_M_WDS) {
1634 /* XXX monitor mode? */
1635 DPRINTF(sc, MWL_DEBUG_KEYCACHE,
1636 "%s: no hvap for opmode %d\n", __func__,
1640 hvap = MWL_VAP(vap)->mv_ap_hvap;
1642 memset(&hk, 0, sizeof(hk));
1643 hk.keyIndex = k->wk_keyix;
1644 switch (cip->ic_cipher) {
1645 case IEEE80211_CIPHER_WEP:
1646 hk.keyTypeId = KEY_TYPE_ID_WEP;
1647 hk.keyLen = k->wk_keylen;
1648 if (k->wk_keyix == vap->iv_def_txkey)
1649 hk.keyFlags = KEY_FLAG_WEP_TXKEY;
1650 if (!IEEE80211_IS_STATICKEY(k)) {
1651 /* NB: WEP is never used for the PTK */
1652 (void) addgroupflags(&hk, k);
1655 case IEEE80211_CIPHER_TKIP:
1656 hk.keyTypeId = KEY_TYPE_ID_TKIP;
1657 hk.key.tkip.tsc.high = (uint32_t)(k->wk_keytsc >> 16);
1658 hk.key.tkip.tsc.low = (uint16_t)k->wk_keytsc;
1659 hk.keyFlags = KEY_FLAG_TSC_VALID | KEY_FLAG_MICKEY_VALID;
1660 hk.keyLen = k->wk_keylen + IEEE80211_MICBUF_SIZE;
1661 if (!addgroupflags(&hk, k))
1662 hk.keyFlags |= KEY_FLAG_PAIRWISE;
1664 case IEEE80211_CIPHER_AES_CCM:
1665 hk.keyTypeId = KEY_TYPE_ID_AES;
1666 hk.keyLen = k->wk_keylen;
1667 if (!addgroupflags(&hk, k))
1668 hk.keyFlags |= KEY_FLAG_PAIRWISE;
1671 /* XXX should not happen */
1672 DPRINTF(sc, MWL_DEBUG_KEYCACHE, "%s: unknown cipher %d\n",
1673 __func__, k->wk_cipher->ic_cipher);
1677 * NB: tkip mic keys get copied here too; the layout
1678 * just happens to match that in ieee80211_key.
1680 memcpy(hk.key.aes, k->wk_key, hk.keyLen);
1683 * Locate address of sta db entry for writing key;
1684 * the convention unfortunately is somewhat different
1685 * than how net80211, hostapd, and wpa_supplicant think.
1687 if (vap->iv_opmode == IEEE80211_M_STA) {
1689 * NB: keys plumbed before the sta reaches AUTH state
1690 * will be discarded or written to the wrong sta db
1691 * entry because iv_bss is meaningless. This is ok
1692 * (right now) because we handle deferred plumbing of
1693 * WEP keys when the sta reaches AUTH state.
1695 macaddr = vap->iv_bss->ni_bssid;
1696 if ((k->wk_flags & IEEE80211_KEY_GROUP) == 0) {
1697 /* XXX plumb to local sta db too for static key wep */
1698 mwl_hal_keyset(hvap, &hk, vap->iv_myaddr);
1700 } else if (vap->iv_opmode == IEEE80211_M_WDS &&
1701 vap->iv_state != IEEE80211_S_RUN) {
1703 * Prior to RUN state a WDS vap will not it's BSS node
1704 * setup so we will plumb the key to the wrong mac
1705 * address (it'll be our local address). Workaround
1706 * this for the moment by grabbing the correct address.
1708 macaddr = vap->iv_des_bssid;
1709 } else if ((k->wk_flags & GRPXMIT) == GRPXMIT)
1710 macaddr = vap->iv_myaddr;
1713 KEYPRINTF(sc, &hk, macaddr);
1714 return (mwl_hal_keyset(hvap, &hk, macaddr) == 0);
1715 #undef IEEE80211_IS_STATICKEY
1720 * Set the multicast filter contents into the hardware.
1721 * XXX f/w has no support; just defer to the os.
1724 mwl_setmcastfilter(struct mwl_softc *sc)
1727 struct ether_multi *enm;
1728 struct ether_multistep estep;
1729 uint8_t macs[IEEE80211_ADDR_LEN*MWL_HAL_MCAST_MAX];/* XXX stack use */
1735 ETHER_FIRST_MULTI(estep, &sc->sc_ec, enm);
1736 while (enm != NULL) {
1737 /* XXX Punt on ranges. */
1738 if (nmc == MWL_HAL_MCAST_MAX ||
1739 !IEEE80211_ADDR_EQ(enm->enm_addrlo, enm->enm_addrhi)) {
1740 ifp->if_flags |= IFF_ALLMULTI;
1743 IEEE80211_ADDR_COPY(mp, enm->enm_addrlo);
1744 mp += IEEE80211_ADDR_LEN, nmc++;
1745 ETHER_NEXT_MULTI(estep, enm);
1747 ifp->if_flags &= ~IFF_ALLMULTI;
1748 mwl_hal_setmcast(sc->sc_mh, nmc, macs);
1753 mwl_mode_init(struct mwl_softc *sc)
1755 struct ieee80211com *ic = &sc->sc_ic;
1756 struct mwl_hal *mh = sc->sc_mh;
1758 mwl_hal_setpromisc(mh, ic->ic_promisc > 0);
1759 mwl_setmcastfilter(sc);
1765 * Callback from the 802.11 layer after a multicast state change.
1768 mwl_update_mcast(struct ieee80211com *ic)
1770 struct mwl_softc *sc = ic->ic_softc;
1772 mwl_setmcastfilter(sc);
1776 * Callback from the 802.11 layer after a promiscuous mode change.
1777 * Note this interface does not check the operating mode as this
1778 * is an internal callback and we are expected to honor the current
1779 * state (e.g. this is used for setting the interface in promiscuous
1780 * mode when operating in hostap mode to do ACS).
1783 mwl_update_promisc(struct ieee80211com *ic)
1785 struct mwl_softc *sc = ic->ic_softc;
1787 mwl_hal_setpromisc(sc->sc_mh, ic->ic_promisc > 0);
1791 * Callback from the 802.11 layer to update the slot time
1792 * based on the current setting. We use it to notify the
1793 * firmware of ERP changes and the f/w takes care of things
1794 * like slot time and preamble.
1797 mwl_updateslot(struct ieee80211com *ic)
1799 struct mwl_softc *sc = ic->ic_softc;
1800 struct mwl_hal *mh = sc->sc_mh;
1803 /* NB: can be called early; suppress needless cmds */
1804 if (!sc->sc_running)
1808 * Calculate the ERP flags. The firwmare will use
1809 * this to carry out the appropriate measures.
1812 if (IEEE80211_IS_CHAN_ANYG(ic->ic_curchan)) {
1813 if ((ic->ic_flags & IEEE80211_F_SHSLOT) == 0)
1814 prot |= IEEE80211_ERP_NON_ERP_PRESENT;
1815 if (ic->ic_flags & IEEE80211_F_USEPROT)
1816 prot |= IEEE80211_ERP_USE_PROTECTION;
1817 if (ic->ic_flags & IEEE80211_F_USEBARKER)
1818 prot |= IEEE80211_ERP_LONG_PREAMBLE;
1821 DPRINTF(sc, MWL_DEBUG_RESET,
1822 "%s: chan %u MHz/flags 0x%x %s slot, (prot 0x%x ic_flags 0x%x)\n",
1823 __func__, ic->ic_curchan->ic_freq, ic->ic_curchan->ic_flags,
1824 ic->ic_flags & IEEE80211_F_SHSLOT ? "short" : "long", prot,
1827 mwl_hal_setgprot(mh, prot);
1831 * Setup the beacon frame.
1834 mwl_beacon_setup(struct ieee80211vap *vap)
1836 struct mwl_hal_vap *hvap = MWL_VAP(vap)->mv_hvap;
1837 struct ieee80211_node *ni = vap->iv_bss;
1840 m = ieee80211_beacon_alloc(ni);
1843 mwl_hal_setbeacon(hvap, mtod(m, const void *), m->m_len);
1850 * Update the beacon frame in response to a change.
1853 mwl_beacon_update(struct ieee80211vap *vap, int item)
1855 struct mwl_hal_vap *hvap = MWL_VAP(vap)->mv_hvap;
1856 struct ieee80211com *ic = vap->iv_ic;
1858 KASSERT(hvap != NULL, ("no beacon"));
1860 case IEEE80211_BEACON_ERP:
1863 case IEEE80211_BEACON_HTINFO:
1864 mwl_hal_setnprotmode(hvap,
1865 MS(ic->ic_curhtprotmode, IEEE80211_HTINFO_OPMODE));
1867 case IEEE80211_BEACON_CAPS:
1868 case IEEE80211_BEACON_WME:
1869 case IEEE80211_BEACON_APPIE:
1870 case IEEE80211_BEACON_CSA:
1872 case IEEE80211_BEACON_TIM:
1873 /* NB: firmware always forms TIM */
1876 /* XXX retain beacon frame and update */
1877 mwl_beacon_setup(vap);
1881 mwl_load_cb(void *arg, bus_dma_segment_t *segs, int nsegs, int error)
1883 bus_addr_t *paddr = (bus_addr_t*) arg;
1884 KASSERT(error == 0, ("error %u on bus_dma callback", error));
1885 *paddr = segs->ds_addr;
1888 #ifdef MWL_HOST_PS_SUPPORT
1890 * Handle power save station occupancy changes.
1893 mwl_update_ps(struct ieee80211vap *vap, int nsta)
1895 struct mwl_vap *mvp = MWL_VAP(vap);
1897 if (nsta == 0 || mvp->mv_last_ps_sta == 0)
1898 mwl_hal_setpowersave_bss(mvp->mv_hvap, nsta);
1899 mvp->mv_last_ps_sta = nsta;
1903 * Handle associated station power save state changes.
1906 mwl_set_tim(struct ieee80211_node *ni, int set)
1908 struct ieee80211vap *vap = ni->ni_vap;
1909 struct mwl_vap *mvp = MWL_VAP(vap);
1911 if (mvp->mv_set_tim(ni, set)) { /* NB: state change */
1912 mwl_hal_setpowersave_sta(mvp->mv_hvap,
1913 IEEE80211_AID(ni->ni_associd), set);
1918 #endif /* MWL_HOST_PS_SUPPORT */
1921 mwl_desc_setup(struct mwl_softc *sc, const char *name,
1922 struct mwl_descdma *dd,
1923 int nbuf, size_t bufsize, int ndesc, size_t descsize)
1928 DPRINTF(sc, MWL_DEBUG_RESET,
1929 "%s: %s DMA: %u bufs (%ju) %u desc/buf (%ju)\n",
1930 __func__, name, nbuf, (uintmax_t) bufsize,
1931 ndesc, (uintmax_t) descsize);
1934 dd->dd_desc_len = nbuf * ndesc * descsize;
1937 * Setup DMA descriptor area.
1939 error = bus_dma_tag_create(bus_get_dma_tag(sc->sc_dev), /* parent */
1940 PAGE_SIZE, 0, /* alignment, bounds */
1941 BUS_SPACE_MAXADDR_32BIT, /* lowaddr */
1942 BUS_SPACE_MAXADDR, /* highaddr */
1943 NULL, NULL, /* filter, filterarg */
1944 dd->dd_desc_len, /* maxsize */
1946 dd->dd_desc_len, /* maxsegsize */
1947 BUS_DMA_ALLOCNOW, /* flags */
1948 NULL, /* lockfunc */
1952 device_printf(sc->sc_dev, "cannot allocate %s DMA tag\n", dd->dd_name);
1956 /* allocate descriptors */
1957 error = bus_dmamem_alloc(dd->dd_dmat, (void**) &dd->dd_desc,
1958 BUS_DMA_NOWAIT | BUS_DMA_COHERENT,
1961 device_printf(sc->sc_dev, "unable to alloc memory for %u %s descriptors, "
1962 "error %u\n", nbuf * ndesc, dd->dd_name, error);
1966 error = bus_dmamap_load(dd->dd_dmat, dd->dd_dmamap,
1967 dd->dd_desc, dd->dd_desc_len,
1968 mwl_load_cb, &dd->dd_desc_paddr,
1971 device_printf(sc->sc_dev, "unable to map %s descriptors, error %u\n",
1972 dd->dd_name, error);
1977 memset(ds, 0, dd->dd_desc_len);
1978 DPRINTF(sc, MWL_DEBUG_RESET,
1979 "%s: %s DMA map: %p (%lu) -> 0x%jx (%lu)\n",
1980 __func__, dd->dd_name, ds, (u_long) dd->dd_desc_len,
1981 (uintmax_t) dd->dd_desc_paddr, /*XXX*/ (u_long) dd->dd_desc_len);
1985 bus_dmamem_free(dd->dd_dmat, dd->dd_desc, dd->dd_dmamap);
1987 bus_dma_tag_destroy(dd->dd_dmat);
1988 memset(dd, 0, sizeof(*dd));
1994 mwl_desc_cleanup(struct mwl_softc *sc, struct mwl_descdma *dd)
1996 bus_dmamap_unload(dd->dd_dmat, dd->dd_dmamap);
1997 bus_dmamem_free(dd->dd_dmat, dd->dd_desc, dd->dd_dmamap);
1998 bus_dma_tag_destroy(dd->dd_dmat);
2000 memset(dd, 0, sizeof(*dd));
2004 * Construct a tx q's free list. The order of entries on
2005 * the list must reflect the physical layout of tx descriptors
2006 * because the firmware pre-fetches descriptors.
2008 * XXX might be better to use indices into the buffer array.
2011 mwl_txq_reset(struct mwl_softc *sc, struct mwl_txq *txq)
2013 struct mwl_txbuf *bf;
2016 bf = txq->dma.dd_bufptr;
2017 STAILQ_INIT(&txq->free);
2018 for (i = 0; i < mwl_txbuf; i++, bf++)
2019 STAILQ_INSERT_TAIL(&txq->free, bf, bf_list);
2023 #define DS2PHYS(_dd, _ds) \
2024 ((_dd)->dd_desc_paddr + ((caddr_t)(_ds) - (caddr_t)(_dd)->dd_desc))
2027 mwl_txdma_setup(struct mwl_softc *sc, struct mwl_txq *txq)
2029 int error, bsize, i;
2030 struct mwl_txbuf *bf;
2031 struct mwl_txdesc *ds;
2033 error = mwl_desc_setup(sc, "tx", &txq->dma,
2034 mwl_txbuf, sizeof(struct mwl_txbuf),
2035 MWL_TXDESC, sizeof(struct mwl_txdesc));
2039 /* allocate and setup tx buffers */
2040 bsize = mwl_txbuf * sizeof(struct mwl_txbuf);
2041 bf = malloc(bsize, M_MWLDEV, M_NOWAIT | M_ZERO);
2043 device_printf(sc->sc_dev, "malloc of %u tx buffers failed\n",
2047 txq->dma.dd_bufptr = bf;
2049 ds = txq->dma.dd_desc;
2050 for (i = 0; i < mwl_txbuf; i++, bf++, ds += MWL_TXDESC) {
2052 bf->bf_daddr = DS2PHYS(&txq->dma, ds);
2053 error = bus_dmamap_create(sc->sc_dmat, BUS_DMA_NOWAIT,
2056 device_printf(sc->sc_dev, "unable to create dmamap for tx "
2057 "buffer %u, error %u\n", i, error);
2061 mwl_txq_reset(sc, txq);
2066 mwl_txdma_cleanup(struct mwl_softc *sc, struct mwl_txq *txq)
2068 struct mwl_txbuf *bf;
2071 bf = txq->dma.dd_bufptr;
2072 for (i = 0; i < mwl_txbuf; i++, bf++) {
2073 KASSERT(bf->bf_m == NULL, ("mbuf on free list"));
2074 KASSERT(bf->bf_node == NULL, ("node on free list"));
2075 if (bf->bf_dmamap != NULL)
2076 bus_dmamap_destroy(sc->sc_dmat, bf->bf_dmamap);
2078 STAILQ_INIT(&txq->free);
2080 if (txq->dma.dd_bufptr != NULL) {
2081 free(txq->dma.dd_bufptr, M_MWLDEV);
2082 txq->dma.dd_bufptr = NULL;
2084 if (txq->dma.dd_desc_len != 0)
2085 mwl_desc_cleanup(sc, &txq->dma);
2089 mwl_rxdma_setup(struct mwl_softc *sc)
2091 int error, jumbosize, bsize, i;
2092 struct mwl_rxbuf *bf;
2093 struct mwl_jumbo *rbuf;
2094 struct mwl_rxdesc *ds;
2097 error = mwl_desc_setup(sc, "rx", &sc->sc_rxdma,
2098 mwl_rxdesc, sizeof(struct mwl_rxbuf),
2099 1, sizeof(struct mwl_rxdesc));
2104 * Receive is done to a private pool of jumbo buffers.
2105 * This allows us to attach to mbuf's and avoid re-mapping
2106 * memory on each rx we post. We allocate a large chunk
2107 * of memory and manage it in the driver. The mbuf free
2108 * callback method is used to reclaim frames after sending
2109 * them up the stack. By default we allocate 2x the number of
2110 * rx descriptors configured so we have some slop to hold
2111 * us while frames are processed.
2113 if (mwl_rxbuf < 2*mwl_rxdesc) {
2114 device_printf(sc->sc_dev,
2115 "too few rx dma buffers (%d); increasing to %d\n",
2116 mwl_rxbuf, 2*mwl_rxdesc);
2117 mwl_rxbuf = 2*mwl_rxdesc;
2119 jumbosize = roundup(MWL_AGGR_SIZE, PAGE_SIZE);
2120 sc->sc_rxmemsize = mwl_rxbuf*jumbosize;
2122 error = bus_dma_tag_create(sc->sc_dmat, /* parent */
2123 PAGE_SIZE, 0, /* alignment, bounds */
2124 BUS_SPACE_MAXADDR_32BIT, /* lowaddr */
2125 BUS_SPACE_MAXADDR, /* highaddr */
2126 NULL, NULL, /* filter, filterarg */
2127 sc->sc_rxmemsize, /* maxsize */
2129 sc->sc_rxmemsize, /* maxsegsize */
2130 BUS_DMA_ALLOCNOW, /* flags */
2131 NULL, /* lockfunc */
2135 device_printf(sc->sc_dev, "could not create rx DMA tag\n");
2139 error = bus_dmamem_alloc(sc->sc_rxdmat, (void**) &sc->sc_rxmem,
2140 BUS_DMA_NOWAIT | BUS_DMA_COHERENT,
2143 device_printf(sc->sc_dev, "could not alloc %ju bytes of rx DMA memory\n",
2144 (uintmax_t) sc->sc_rxmemsize);
2148 error = bus_dmamap_load(sc->sc_rxdmat, sc->sc_rxmap,
2149 sc->sc_rxmem, sc->sc_rxmemsize,
2150 mwl_load_cb, &sc->sc_rxmem_paddr,
2153 device_printf(sc->sc_dev, "could not load rx DMA map\n");
2158 * Allocate rx buffers and set them up.
2160 bsize = mwl_rxdesc * sizeof(struct mwl_rxbuf);
2161 bf = malloc(bsize, M_MWLDEV, M_NOWAIT | M_ZERO);
2163 device_printf(sc->sc_dev, "malloc of %u rx buffers failed\n", bsize);
2166 sc->sc_rxdma.dd_bufptr = bf;
2168 STAILQ_INIT(&sc->sc_rxbuf);
2169 ds = sc->sc_rxdma.dd_desc;
2170 for (i = 0; i < mwl_rxdesc; i++, bf++, ds++) {
2172 bf->bf_daddr = DS2PHYS(&sc->sc_rxdma, ds);
2173 /* pre-assign dma buffer */
2174 bf->bf_data = ((uint8_t *)sc->sc_rxmem) + (i*jumbosize);
2175 /* NB: tail is intentional to preserve descriptor order */
2176 STAILQ_INSERT_TAIL(&sc->sc_rxbuf, bf, bf_list);
2180 * Place remainder of dma memory buffers on the free list.
2182 SLIST_INIT(&sc->sc_rxfree);
2183 for (; i < mwl_rxbuf; i++) {
2184 data = ((uint8_t *)sc->sc_rxmem) + (i*jumbosize);
2185 rbuf = MWL_JUMBO_DATA2BUF(data);
2186 SLIST_INSERT_HEAD(&sc->sc_rxfree, rbuf, next);
2194 mwl_rxdma_cleanup(struct mwl_softc *sc)
2196 if (sc->sc_rxmem_paddr != 0) {
2197 bus_dmamap_unload(sc->sc_rxdmat, sc->sc_rxmap);
2198 sc->sc_rxmem_paddr = 0;
2200 if (sc->sc_rxmem != NULL) {
2201 bus_dmamem_free(sc->sc_rxdmat, sc->sc_rxmem, sc->sc_rxmap);
2202 sc->sc_rxmem = NULL;
2204 if (sc->sc_rxdma.dd_bufptr != NULL) {
2205 free(sc->sc_rxdma.dd_bufptr, M_MWLDEV);
2206 sc->sc_rxdma.dd_bufptr = NULL;
2208 if (sc->sc_rxdma.dd_desc_len != 0)
2209 mwl_desc_cleanup(sc, &sc->sc_rxdma);
2213 mwl_dma_setup(struct mwl_softc *sc)
2217 error = mwl_rxdma_setup(sc);
2219 mwl_rxdma_cleanup(sc);
2223 for (i = 0; i < MWL_NUM_TX_QUEUES; i++) {
2224 error = mwl_txdma_setup(sc, &sc->sc_txq[i]);
2226 mwl_dma_cleanup(sc);
2234 mwl_dma_cleanup(struct mwl_softc *sc)
2238 for (i = 0; i < MWL_NUM_TX_QUEUES; i++)
2239 mwl_txdma_cleanup(sc, &sc->sc_txq[i]);
2240 mwl_rxdma_cleanup(sc);
2243 static struct ieee80211_node *
2244 mwl_node_alloc(struct ieee80211vap *vap, const uint8_t mac[IEEE80211_ADDR_LEN])
2246 struct ieee80211com *ic = vap->iv_ic;
2247 struct mwl_softc *sc = ic->ic_softc;
2248 const size_t space = sizeof(struct mwl_node);
2249 struct mwl_node *mn;
2251 mn = malloc(space, M_80211_NODE, M_NOWAIT|M_ZERO);
2256 DPRINTF(sc, MWL_DEBUG_NODE, "%s: mn %p\n", __func__, mn);
2257 return &mn->mn_node;
2261 mwl_node_cleanup(struct ieee80211_node *ni)
2263 struct ieee80211com *ic = ni->ni_ic;
2264 struct mwl_softc *sc = ic->ic_softc;
2265 struct mwl_node *mn = MWL_NODE(ni);
2267 DPRINTF(sc, MWL_DEBUG_NODE, "%s: ni %p ic %p staid %d\n",
2268 __func__, ni, ni->ni_ic, mn->mn_staid);
2270 if (mn->mn_staid != 0) {
2271 struct ieee80211vap *vap = ni->ni_vap;
2273 if (mn->mn_hvap != NULL) {
2274 if (vap->iv_opmode == IEEE80211_M_STA)
2275 mwl_hal_delstation(mn->mn_hvap, vap->iv_myaddr);
2277 mwl_hal_delstation(mn->mn_hvap, ni->ni_macaddr);
2280 * NB: legacy WDS peer sta db entry is installed using
2281 * the associate ap's hvap; use it again to delete it.
2282 * XXX can vap be NULL?
2284 else if (vap->iv_opmode == IEEE80211_M_WDS &&
2285 MWL_VAP(vap)->mv_ap_hvap != NULL)
2286 mwl_hal_delstation(MWL_VAP(vap)->mv_ap_hvap,
2288 delstaid(sc, mn->mn_staid);
2291 sc->sc_node_cleanup(ni);
2295 * Reclaim rx dma buffers from packets sitting on the ampdu
2296 * reorder queue for a station. We replace buffers with a
2297 * system cluster (if available).
2300 mwl_ampdu_rxdma_reclaim(struct ieee80211_rx_ampdu *rap)
2307 n = rap->rxa_qframes;
2308 for (i = 0; i < rap->rxa_wnd && n > 0; i++) {
2313 /* our dma buffers have a well-known free routine */
2314 if ((m->m_flags & M_EXT) == 0 ||
2315 m->m_ext.ext_free != mwl_ext_free)
2318 * Try to allocate a cluster and move the data.
2320 off = m->m_data - m->m_ext.ext_buf;
2321 if (off + m->m_pkthdr.len > MCLBYTES) {
2322 /* XXX no AMSDU for now */
2325 cl = pool_cache_get_paddr(&mclpool_cache, 0,
2326 &m->m_ext.ext_paddr);
2329 * Copy the existing data to the cluster, remove
2330 * the rx dma buffer, and attach the cluster in
2331 * its place. Note we preserve the offset to the
2332 * data so frames being bridged can still prepend
2333 * their headers without adding another mbuf.
2335 memcpy((caddr_t) cl + off, m->m_data, m->m_pkthdr.len);
2337 MEXTADD(m, cl, MCLBYTES, 0, NULL, &mclpool_cache);
2338 /* setup mbuf like _MCLGET does */
2339 m->m_flags |= M_CLUSTER | M_EXT_RW;
2340 _MOWNERREF(m, M_EXT | M_CLUSTER);
2341 /* NB: m_data is clobbered by MEXTADDR, adjust */
2349 * Callback to reclaim resources. We first let the
2350 * net80211 layer do it's thing, then if we are still
2351 * blocked by a lack of rx dma buffers we walk the ampdu
2352 * reorder q's to reclaim buffers by copying to a system
2356 mwl_node_drain(struct ieee80211_node *ni)
2358 struct ieee80211com *ic = ni->ni_ic;
2359 struct mwl_softc *sc = ic->ic_softc;
2360 struct mwl_node *mn = MWL_NODE(ni);
2362 DPRINTF(sc, MWL_DEBUG_NODE, "%s: ni %p vap %p staid %d\n",
2363 __func__, ni, ni->ni_vap, mn->mn_staid);
2365 /* NB: call up first to age out ampdu q's */
2366 sc->sc_node_drain(ni);
2368 /* XXX better to not check low water mark? */
2369 if (sc->sc_rxblocked && mn->mn_staid != 0 &&
2370 (ni->ni_flags & IEEE80211_NODE_HT)) {
2373 * Walk the reorder q and reclaim rx dma buffers by copying
2374 * the packet contents into clusters.
2376 for (tid = 0; tid < WME_NUM_TID; tid++) {
2377 struct ieee80211_rx_ampdu *rap;
2379 rap = &ni->ni_rx_ampdu[tid];
2380 if ((rap->rxa_flags & IEEE80211_AGGR_XCHGPEND) == 0)
2382 if (rap->rxa_qframes)
2383 mwl_ampdu_rxdma_reclaim(rap);
2389 mwl_node_getsignal(const struct ieee80211_node *ni, int8_t *rssi, int8_t *noise)
2391 *rssi = ni->ni_ic->ic_node_getrssi(ni);
2392 #ifdef MWL_ANT_INFO_SUPPORT
2394 /* XXX need to smooth data */
2395 *noise = -MWL_NODE_CONST(ni)->mn_ai.nf;
2397 *noise = -95; /* XXX */
2400 *noise = -95; /* XXX */
2405 * Convert Hardware per-antenna rssi info to common format:
2406 * Let a1, a2, a3 represent the amplitudes per chain
2407 * Let amax represent max[a1, a2, a3]
2408 * Rssi1_dBm = RSSI_dBm + 20*log10(a1/amax)
2409 * Rssi1_dBm = RSSI_dBm + 20*log10(a1) - 20*log10(amax)
2410 * We store a table that is 4*20*log10(idx) - the extra 4 is to store or
2411 * maintain some extra precision.
2413 * Values are stored in .5 db format capped at 127.
2416 mwl_node_getmimoinfo(const struct ieee80211_node *ni,
2417 struct ieee80211_mimo_info *mi)
2419 #define CVT(_dst, _src) do { \
2420 (_dst) = rssi + ((logdbtbl[_src] - logdbtbl[rssi_max]) >> 2); \
2421 (_dst) = (_dst) > 64 ? 127 : ((_dst) << 1); \
2423 static const int8_t logdbtbl[32] = {
2424 0, 0, 24, 38, 48, 56, 62, 68,
2425 72, 76, 80, 83, 86, 89, 92, 94,
2426 96, 98, 100, 102, 104, 106, 107, 109,
2427 110, 112, 113, 115, 116, 117, 118, 119
2429 const struct mwl_node *mn = MWL_NODE_CONST(ni);
2430 uint8_t rssi = mn->mn_ai.rsvd1/2; /* XXX */
2433 rssi_max = mn->mn_ai.rssi_a;
2434 if (mn->mn_ai.rssi_b > rssi_max)
2435 rssi_max = mn->mn_ai.rssi_b;
2436 if (mn->mn_ai.rssi_c > rssi_max)
2437 rssi_max = mn->mn_ai.rssi_c;
2439 CVT(mi->ch[0].rssi[0], mn->mn_ai.rssi_a);
2440 CVT(mi->ch[1].rssi[0], mn->mn_ai.rssi_b);
2441 CVT(mi->ch[2].rssi[0], mn->mn_ai.rssi_c);
2443 mi->ch[0].noise[0] = mn->mn_ai.nf_a;
2444 mi->ch[1].noise[0] = mn->mn_ai.nf_b;
2445 mi->ch[2].noise[0] = mn->mn_ai.nf_c;
2449 static __inline void *
2450 mwl_getrxdma(struct mwl_softc *sc)
2452 struct mwl_jumbo *buf;
2456 * Allocate from jumbo pool.
2458 MWL_RXFREE_LOCK(sc);
2459 buf = SLIST_FIRST(&sc->sc_rxfree);
2461 DPRINTF(sc, MWL_DEBUG_ANY,
2462 "%s: out of rx dma buffers\n", __func__);
2463 sc->sc_stats.mst_rx_nodmabuf++;
2466 SLIST_REMOVE_HEAD(&sc->sc_rxfree, next);
2468 data = MWL_JUMBO_BUF2DATA(buf);
2470 MWL_RXFREE_UNLOCK(sc);
2474 static __inline void
2475 mwl_putrxdma(struct mwl_softc *sc, void *data)
2477 struct mwl_jumbo *buf;
2479 /* XXX bounds check data */
2480 MWL_RXFREE_LOCK(sc);
2481 buf = MWL_JUMBO_DATA2BUF(data);
2482 SLIST_INSERT_HEAD(&sc->sc_rxfree, buf, next);
2484 MWL_RXFREE_UNLOCK(sc);
2488 mwl_rxbuf_init(struct mwl_softc *sc, struct mwl_rxbuf *bf)
2490 struct mwl_rxdesc *ds;
2493 if (bf->bf_data == NULL) {
2494 bf->bf_data = mwl_getrxdma(sc);
2495 if (bf->bf_data == NULL) {
2496 /* mark descriptor to be skipped */
2497 ds->RxControl = EAGLE_RXD_CTRL_OS_OWN;
2498 /* NB: don't need PREREAD */
2499 MWL_RXDESC_SYNC(sc, ds, BUS_DMASYNC_PREWRITE);
2500 sc->sc_stats.mst_rxbuf_failed++;
2505 * NB: DMA buffer contents is known to be unmodified
2506 * so there's no need to flush the data cache.
2514 ds->Status = EAGLE_RXD_STATUS_IDLE;
2516 ds->PktLen = htole16(MWL_AGGR_SIZE);
2518 ds->pPhysBuffData = htole32(MWL_JUMBO_DMA_ADDR(sc, bf->bf_data));
2519 /* NB: don't touch pPhysNext, set once */
2520 ds->RxControl = EAGLE_RXD_CTRL_DRIVER_OWN;
2521 MWL_RXDESC_SYNC(sc, ds, BUS_DMASYNC_PREREAD | BUS_DMASYNC_PREWRITE);
2527 mwl_ext_free(struct mbuf *m)
2529 struct mwl_softc *sc = m->m_ext.ext_arg1;
2531 /* XXX bounds check data */
2532 mwl_putrxdma(sc, m->m_ext.ext_buf);
2534 * If we were previously blocked by a lack of rx dma buffers
2535 * check if we now have enough to restart rx interrupt handling.
2536 * NB: we know we are called at splvm which is above splnet.
2538 if (sc->sc_rxblocked && sc->sc_nrxfree > mwl_rxdmalow) {
2539 sc->sc_rxblocked = 0;
2540 mwl_hal_intrset(sc->sc_mh, sc->sc_imask);
2544 struct mwl_frame_bar {
2547 u_int8_t i_ra[IEEE80211_ADDR_LEN];
2548 u_int8_t i_ta[IEEE80211_ADDR_LEN];
2553 * Like ieee80211_anyhdrsize, but handles BAR frames
2554 * specially so the logic below to piece the 802.11
2555 * header together works.
2558 mwl_anyhdrsize(const void *data)
2560 const struct ieee80211_frame *wh = data;
2562 if ((wh->i_fc[0]&IEEE80211_FC0_TYPE_MASK) == IEEE80211_FC0_TYPE_CTL) {
2563 switch (wh->i_fc[0] & IEEE80211_FC0_SUBTYPE_MASK) {
2564 case IEEE80211_FC0_SUBTYPE_CTS:
2565 case IEEE80211_FC0_SUBTYPE_ACK:
2566 return sizeof(struct ieee80211_frame_ack);
2567 case IEEE80211_FC0_SUBTYPE_BAR:
2568 return sizeof(struct mwl_frame_bar);
2570 return sizeof(struct ieee80211_frame_min);
2572 return ieee80211_hdrsize(data);
2576 mwl_handlemicerror(struct ieee80211com *ic, const uint8_t *data)
2578 const struct ieee80211_frame *wh;
2579 struct ieee80211_node *ni;
2581 wh = (const struct ieee80211_frame *)(data + sizeof(uint16_t));
2582 ni = ieee80211_find_rxnode(ic, (const struct ieee80211_frame_min *) wh);
2584 ieee80211_notify_michael_failure(ni->ni_vap, wh, 0);
2585 ieee80211_free_node(ni);
2590 * Convert hardware signal strength to rssi. The value
2591 * provided by the device has the noise floor added in;
2592 * we need to compensate for this but we don't have that
2593 * so we use a fixed value.
2595 * The offset of 8 is good for both 2.4 and 5GHz. The LNA
2596 * offset is already set as part of the initial gain. This
2597 * will give at least +/- 3dB for 2.4GHz and +/- 5dB for 5GHz.
2600 cvtrssi(uint8_t ssi)
2602 int rssi = (int) ssi + 8;
2603 /* XXX hack guess until we have a real noise floor */
2604 rssi = 2*(87 - rssi); /* NB: .5 dBm units */
2605 return (rssi < 0 ? 0 : rssi > 127 ? 127 : rssi);
2609 mwl_rx_proc(void *arg, int npending)
2611 struct mwl_softc *sc = arg;
2612 struct ieee80211com *ic = &sc->sc_ic;
2613 struct mwl_rxbuf *bf;
2614 struct mwl_rxdesc *ds;
2616 struct ieee80211_qosframe *wh;
2617 struct ieee80211_node *ni;
2618 struct mwl_node *mn;
2619 int off, len, hdrlen, pktlen, rssi, ntodo;
2620 uint8_t *data, status;
2624 DPRINTF(sc, MWL_DEBUG_RX_PROC, "%s: pending %u rdptr 0x%x wrptr 0x%x\n",
2625 __func__, npending, RD4(sc, sc->sc_hwspecs.rxDescRead),
2626 RD4(sc, sc->sc_hwspecs.rxDescWrite));
2629 for (ntodo = mwl_rxquota; ntodo > 0; ntodo--) {
2631 bf = STAILQ_FIRST(&sc->sc_rxbuf);
2636 * If data allocation failed previously there
2637 * will be no buffer; try again to re-populate it.
2638 * Note the firmware will not advance to the next
2639 * descriptor with a dma buffer so we must mimic
2640 * this or we'll get out of sync.
2642 DPRINTF(sc, MWL_DEBUG_ANY,
2643 "%s: rx buf w/o dma memory\n", __func__);
2644 (void) mwl_rxbuf_init(sc, bf);
2645 sc->sc_stats.mst_rx_dmabufmissing++;
2648 MWL_RXDESC_SYNC(sc, ds,
2649 BUS_DMASYNC_POSTREAD | BUS_DMASYNC_POSTWRITE);
2650 if (ds->RxControl != EAGLE_RXD_CTRL_DMA_OWN)
2653 if (sc->sc_debug & MWL_DEBUG_RECV_DESC)
2654 mwl_printrxbuf(bf, 0);
2656 status = ds->Status;
2657 if (status & EAGLE_RXD_STATUS_DECRYPT_ERR_MASK) {
2658 counter_u64_add(ic->ic_ierrors, 1);
2659 sc->sc_stats.mst_rx_crypto++;
2661 * NB: Check EAGLE_RXD_STATUS_GENERAL_DECRYPT_ERR
2662 * for backwards compatibility.
2664 if (status != EAGLE_RXD_STATUS_GENERAL_DECRYPT_ERR &&
2665 (status & EAGLE_RXD_STATUS_TKIP_MIC_DECRYPT_ERR)) {
2667 * MIC error, notify upper layers.
2669 bus_dmamap_sync(sc->sc_rxdmat, sc->sc_rxmap,
2670 BUS_DMASYNC_POSTREAD);
2671 mwl_handlemicerror(ic, data);
2672 sc->sc_stats.mst_rx_tkipmic++;
2674 /* XXX too painful to tap packets */
2678 * Sync the data buffer.
2680 len = le16toh(ds->PktLen);
2681 bus_dmamap_sync(sc->sc_rxdmat, sc->sc_rxmap, BUS_DMASYNC_POSTREAD);
2683 * The 802.11 header is provided all or in part at the front;
2684 * use it to calculate the true size of the header that we'll
2685 * construct below. We use this to figure out where to copy
2686 * payload prior to constructing the header.
2688 hdrlen = mwl_anyhdrsize(data + sizeof(uint16_t));
2689 off = sizeof(uint16_t) + sizeof(struct ieee80211_frame_addr4);
2691 /* calculate rssi early so we can re-use for each aggregate */
2692 rssi = cvtrssi(ds->RSSI);
2694 pktlen = hdrlen + (len - off);
2696 * NB: we know our frame is at least as large as
2697 * IEEE80211_MIN_LEN because there is a 4-address
2698 * frame at the front. Hence there's no need to
2699 * vet the packet length. If the frame in fact
2700 * is too small it should be discarded at the
2705 * Attach dma buffer to an mbuf. We tried
2706 * doing this based on the packet size (i.e.
2707 * copying small packets) but it turns out to
2708 * be a net loss. The tradeoff might be system
2709 * dependent (cache architecture is important).
2711 MGETHDR(m, M_NOWAIT, MT_DATA);
2713 DPRINTF(sc, MWL_DEBUG_ANY,
2714 "%s: no rx mbuf\n", __func__);
2715 sc->sc_stats.mst_rx_nombuf++;
2719 * Acquire the replacement dma buffer before
2720 * processing the frame. If we're out of dma
2721 * buffers we disable rx interrupts and wait
2722 * for the free pool to reach mlw_rxdmalow buffers
2723 * before starting to do work again. If the firmware
2724 * runs out of descriptors then it will toss frames
2725 * which is better than our doing it as that can
2726 * starve our processing. It is also important that
2727 * we always process rx'd frames in case they are
2728 * A-MPDU as otherwise the host's view of the BA
2729 * window may get out of sync with the firmware.
2731 newdata = mwl_getrxdma(sc);
2732 if (newdata == NULL) {
2733 /* NB: stat+msg in mwl_getrxdma */
2735 /* disable RX interrupt and mark state */
2736 mwl_hal_intrset(sc->sc_mh,
2737 sc->sc_imask &~ MACREG_A2HRIC_BIT_RX_RDY);
2738 sc->sc_rxblocked = 1;
2739 ieee80211_drain(ic);
2740 /* XXX check rxblocked and immediately start again? */
2743 bf->bf_data = newdata;
2745 * Attach the dma buffer to the mbuf;
2746 * mwl_rxbuf_init will re-setup the rx
2747 * descriptor using the replacement dma
2748 * buffer we just installed above.
2750 m_extadd(m, data, MWL_AGGR_SIZE, mwl_ext_free, sc, NULL, 0,
2752 m->m_data += off - hdrlen;
2753 m->m_pkthdr.len = m->m_len = pktlen;
2754 /* NB: dma buffer assumed read-only */
2757 * Piece 802.11 header together.
2759 wh = mtod(m, struct ieee80211_qosframe *);
2760 /* NB: don't need to do this sometimes but ... */
2761 /* XXX special case so we can memcpy after m_devget? */
2762 ovbcopy(data + sizeof(uint16_t), wh, hdrlen);
2763 if (IEEE80211_QOS_HAS_SEQ(wh))
2764 *(uint16_t *)ieee80211_getqos(wh) = ds->QosCtrl;
2766 * The f/w strips WEP header but doesn't clear
2767 * the WEP bit; mark the packet with M_WEP so
2768 * net80211 will treat the data as decrypted.
2769 * While here also clear the PWR_MGT bit since
2770 * power save is handled by the firmware and
2771 * passing this up will potentially cause the
2772 * upper layer to put a station in power save
2773 * (except when configured with MWL_HOST_PS_SUPPORT).
2775 if (wh->i_fc[1] & IEEE80211_FC1_PROTECTED)
2776 m->m_flags |= M_WEP;
2777 #ifdef MWL_HOST_PS_SUPPORT
2778 wh->i_fc[1] &= ~IEEE80211_FC1_PROTECTED;
2780 wh->i_fc[1] &= ~(IEEE80211_FC1_PROTECTED |
2781 IEEE80211_FC1_PWR_MGT);
2784 if (ieee80211_radiotap_active(ic)) {
2785 struct mwl_rx_radiotap_header *tap = &sc->sc_rx_th;
2788 tap->wr_rate = ds->Rate;
2789 tap->wr_antsignal = rssi + nf;
2790 tap->wr_antnoise = nf;
2792 if (IFF_DUMPPKTS_RECV(sc, wh)) {
2793 ieee80211_dump_pkt(ic, mtod(m, caddr_t),
2794 len, ds->Rate, rssi);
2797 ni = ieee80211_find_rxnode(ic,
2798 (const struct ieee80211_frame_min *) wh);
2801 #ifdef MWL_ANT_INFO_SUPPORT
2802 mn->mn_ai.rssi_a = ds->ai.rssi_a;
2803 mn->mn_ai.rssi_b = ds->ai.rssi_b;
2804 mn->mn_ai.rssi_c = ds->ai.rssi_c;
2805 mn->mn_ai.rsvd1 = rssi;
2807 /* tag AMPDU aggregates for reorder processing */
2808 if (ni->ni_flags & IEEE80211_NODE_HT)
2809 m->m_flags |= M_AMPDU;
2810 (void) ieee80211_input(ni, m, rssi, nf);
2811 ieee80211_free_node(ni);
2813 (void) ieee80211_input_all(ic, m, rssi, nf);
2815 /* NB: ignore ENOMEM so we process more descriptors */
2816 (void) mwl_rxbuf_init(sc, bf);
2817 bf = STAILQ_NEXT(bf, bf_list);
2822 if (mbufq_first(&sc->sc_snd) != NULL) {
2823 /* NB: kick fw; the tx thread may have been preempted */
2824 mwl_hal_txstart(sc->sc_mh, 0);
2830 mwl_txq_init(struct mwl_softc *sc, struct mwl_txq *txq, int qnum)
2832 struct mwl_txbuf *bf, *bn;
2833 struct mwl_txdesc *ds;
2835 MWL_TXQ_LOCK_INIT(sc, txq);
2837 txq->txpri = 0; /* XXX */
2839 /* NB: q setup by mwl_txdma_setup XXX */
2840 STAILQ_INIT(&txq->free);
2842 STAILQ_FOREACH(bf, &txq->free, bf_list) {
2846 bn = STAILQ_NEXT(bf, bf_list);
2848 bn = STAILQ_FIRST(&txq->free);
2849 ds->pPhysNext = htole32(bn->bf_daddr);
2851 STAILQ_INIT(&txq->active);
2855 * Setup a hardware data transmit queue for the specified
2856 * access control. We record the mapping from ac's
2857 * to h/w queues for use by mwl_tx_start.
2860 mwl_tx_setup(struct mwl_softc *sc, int ac, int mvtype)
2862 struct mwl_txq *txq;
2864 if (ac >= nitems(sc->sc_ac2q)) {
2865 device_printf(sc->sc_dev, "AC %u out of range, max %zu!\n",
2866 ac, nitems(sc->sc_ac2q));
2869 if (mvtype >= MWL_NUM_TX_QUEUES) {
2870 device_printf(sc->sc_dev, "mvtype %u out of range, max %u!\n",
2871 mvtype, MWL_NUM_TX_QUEUES);
2874 txq = &sc->sc_txq[mvtype];
2875 mwl_txq_init(sc, txq, mvtype);
2876 sc->sc_ac2q[ac] = txq;
2881 * Update WME parameters for a transmit queue.
2884 mwl_txq_update(struct mwl_softc *sc, int ac)
2886 #define MWL_EXPONENT_TO_VALUE(v) ((1<<v)-1)
2887 struct ieee80211com *ic = &sc->sc_ic;
2888 struct chanAccParams chp;
2889 struct mwl_txq *txq = sc->sc_ac2q[ac];
2890 struct wmeParams *wmep;
2891 struct mwl_hal *mh = sc->sc_mh;
2892 int aifs, cwmin, cwmax, txoplim;
2894 ieee80211_wme_ic_getparams(ic, &chp);
2895 wmep = &chp.cap_wmeParams[ac];
2897 aifs = wmep->wmep_aifsn;
2898 /* XXX in sta mode need to pass log values for cwmin/max */
2899 cwmin = MWL_EXPONENT_TO_VALUE(wmep->wmep_logcwmin);
2900 cwmax = MWL_EXPONENT_TO_VALUE(wmep->wmep_logcwmax);
2901 txoplim = wmep->wmep_txopLimit; /* NB: units of 32us */
2903 if (mwl_hal_setedcaparams(mh, txq->qnum, cwmin, cwmax, aifs, txoplim)) {
2904 device_printf(sc->sc_dev, "unable to update hardware queue "
2905 "parameters for %s traffic!\n",
2906 ieee80211_wme_acnames[ac]);
2910 #undef MWL_EXPONENT_TO_VALUE
2914 * Callback from the 802.11 layer to update WME parameters.
2917 mwl_wme_update(struct ieee80211com *ic)
2919 struct mwl_softc *sc = ic->ic_softc;
2921 return !mwl_txq_update(sc, WME_AC_BE) ||
2922 !mwl_txq_update(sc, WME_AC_BK) ||
2923 !mwl_txq_update(sc, WME_AC_VI) ||
2924 !mwl_txq_update(sc, WME_AC_VO) ? EIO : 0;
2928 * Reclaim resources for a setup queue.
2931 mwl_tx_cleanupq(struct mwl_softc *sc, struct mwl_txq *txq)
2934 MWL_TXQ_LOCK_DESTROY(txq);
2938 * Reclaim all tx queue resources.
2941 mwl_tx_cleanup(struct mwl_softc *sc)
2945 for (i = 0; i < MWL_NUM_TX_QUEUES; i++)
2946 mwl_tx_cleanupq(sc, &sc->sc_txq[i]);
2950 mwl_tx_dmasetup(struct mwl_softc *sc, struct mwl_txbuf *bf, struct mbuf *m0)
2956 * Load the DMA map so any coalescing is done. This
2957 * also calculates the number of descriptors we need.
2959 error = bus_dmamap_load_mbuf_sg(sc->sc_dmat, bf->bf_dmamap, m0,
2960 bf->bf_segs, &bf->bf_nseg,
2962 if (error == EFBIG) {
2963 /* XXX packet requires too many descriptors */
2964 bf->bf_nseg = MWL_TXDESC+1;
2965 } else if (error != 0) {
2966 sc->sc_stats.mst_tx_busdma++;
2971 * Discard null packets and check for packets that
2972 * require too many TX descriptors. We try to convert
2973 * the latter to a cluster.
2975 if (error == EFBIG) { /* too many desc's, linearize */
2976 sc->sc_stats.mst_tx_linear++;
2978 m = m_collapse(m0, M_NOWAIT, MWL_TXDESC);
2980 m = m_defrag(m0, M_NOWAIT);
2984 sc->sc_stats.mst_tx_nombuf++;
2988 error = bus_dmamap_load_mbuf_sg(sc->sc_dmat, bf->bf_dmamap, m0,
2989 bf->bf_segs, &bf->bf_nseg,
2992 sc->sc_stats.mst_tx_busdma++;
2996 KASSERT(bf->bf_nseg <= MWL_TXDESC,
2997 ("too many segments after defrag; nseg %u", bf->bf_nseg));
2998 } else if (bf->bf_nseg == 0) { /* null packet, discard */
2999 sc->sc_stats.mst_tx_nodata++;
3003 DPRINTF(sc, MWL_DEBUG_XMIT, "%s: m %p len %u\n",
3004 __func__, m0, m0->m_pkthdr.len);
3005 bus_dmamap_sync(sc->sc_dmat, bf->bf_dmamap, BUS_DMASYNC_PREWRITE);
3012 mwl_cvtlegacyrate(int rate)
3033 * Calculate fixed tx rate information per client state;
3034 * this value is suitable for writing to the Format field
3035 * of a tx descriptor.
3038 mwl_calcformat(uint8_t rate, const struct ieee80211_node *ni)
3042 fmt = SM(3, EAGLE_TXD_ANTENNA)
3043 | (IEEE80211_IS_CHAN_HT40D(ni->ni_chan) ?
3044 EAGLE_TXD_EXTCHAN_LO : EAGLE_TXD_EXTCHAN_HI);
3045 if (rate & IEEE80211_RATE_MCS) { /* HT MCS */
3046 fmt |= EAGLE_TXD_FORMAT_HT
3047 /* NB: 0x80 implicitly stripped from ucastrate */
3048 | SM(rate, EAGLE_TXD_RATE);
3049 /* XXX short/long GI may be wrong; re-check */
3050 if (IEEE80211_IS_CHAN_HT40(ni->ni_chan)) {
3051 fmt |= EAGLE_TXD_CHW_40
3052 | (ni->ni_htcap & IEEE80211_HTCAP_SHORTGI40 ?
3053 EAGLE_TXD_GI_SHORT : EAGLE_TXD_GI_LONG);
3055 fmt |= EAGLE_TXD_CHW_20
3056 | (ni->ni_htcap & IEEE80211_HTCAP_SHORTGI20 ?
3057 EAGLE_TXD_GI_SHORT : EAGLE_TXD_GI_LONG);
3059 } else { /* legacy rate */
3060 fmt |= EAGLE_TXD_FORMAT_LEGACY
3061 | SM(mwl_cvtlegacyrate(rate), EAGLE_TXD_RATE)
3063 /* XXX iv_flags & IEEE80211_F_SHPREAMBLE? */
3064 | (ni->ni_capinfo & IEEE80211_CAPINFO_SHORT_PREAMBLE ?
3065 EAGLE_TXD_PREAMBLE_SHORT : EAGLE_TXD_PREAMBLE_LONG);
3071 mwl_tx_start(struct mwl_softc *sc, struct ieee80211_node *ni, struct mwl_txbuf *bf,
3074 struct ieee80211com *ic = &sc->sc_ic;
3075 struct ieee80211vap *vap = ni->ni_vap;
3076 int error, iswep, ismcast;
3077 int hdrlen, copyhdrlen, pktlen;
3078 struct mwl_txdesc *ds;
3079 struct mwl_txq *txq;
3080 struct ieee80211_frame *wh;
3081 struct mwltxrec *tr;
3082 struct mwl_node *mn;
3088 wh = mtod(m0, struct ieee80211_frame *);
3089 iswep = wh->i_fc[1] & IEEE80211_FC1_PROTECTED;
3090 ismcast = IEEE80211_IS_MULTICAST(wh->i_addr1);
3091 hdrlen = ieee80211_anyhdrsize(wh);
3092 copyhdrlen = hdrlen;
3093 pktlen = m0->m_pkthdr.len;
3094 if (IEEE80211_QOS_HAS_SEQ(wh)) {
3095 qos = *(uint16_t *)ieee80211_getqos(wh);
3096 if (IEEE80211_IS_DSTODS(wh))
3097 copyhdrlen -= sizeof(qos);
3102 const struct ieee80211_cipher *cip;
3103 struct ieee80211_key *k;
3106 * Construct the 802.11 header+trailer for an encrypted
3107 * frame. The only reason this can fail is because of an
3108 * unknown or unsupported cipher/key type.
3110 * NB: we do this even though the firmware will ignore
3111 * what we've done for WEP and TKIP as we need the
3112 * ExtIV filled in for CCMP and this also adjusts
3113 * the headers which simplifies our work below.
3115 k = ieee80211_crypto_encap(ni, m0);
3118 * This can happen when the key is yanked after the
3119 * frame was queued. Just discard the frame; the
3120 * 802.11 layer counts failures and provides
3121 * debugging/diagnostics.
3127 * Adjust the packet length for the crypto additions
3128 * done during encap and any other bits that the f/w
3129 * will add later on.
3132 pktlen += cip->ic_header + cip->ic_miclen + cip->ic_trailer;
3134 /* packet header may have moved, reset our local pointer */
3135 wh = mtod(m0, struct ieee80211_frame *);
3138 if (ieee80211_radiotap_active_vap(vap)) {
3139 sc->sc_tx_th.wt_flags = 0; /* XXX */
3141 sc->sc_tx_th.wt_flags |= IEEE80211_RADIOTAP_F_WEP;
3143 sc->sc_tx_th.wt_rate = ds->DataRate;
3145 sc->sc_tx_th.wt_txpower = ni->ni_txpower;
3146 sc->sc_tx_th.wt_antenna = sc->sc_txantenna;
3148 ieee80211_radiotap_tx(vap, m0);
3151 * Copy up/down the 802.11 header; the firmware requires
3152 * we present a 2-byte payload length followed by a
3153 * 4-address header (w/o QoS), followed (optionally) by
3154 * any WEP/ExtIV header (but only filled in for CCMP).
3155 * We are assured the mbuf has sufficient headroom to
3156 * prepend in-place by the setup of ic_headroom in
3159 if (hdrlen < sizeof(struct mwltxrec)) {
3160 const int space = sizeof(struct mwltxrec) - hdrlen;
3161 if (M_LEADINGSPACE(m0) < space) {
3162 /* NB: should never happen */
3163 device_printf(sc->sc_dev,
3164 "not enough headroom, need %d found %zd, "
3165 "m_flags 0x%x m_len %d\n",
3166 space, M_LEADINGSPACE(m0), m0->m_flags, m0->m_len);
3167 ieee80211_dump_pkt(ic,
3168 mtod(m0, const uint8_t *), m0->m_len, 0, -1);
3170 sc->sc_stats.mst_tx_noheadroom++;
3173 M_PREPEND(m0, space, M_NOWAIT);
3175 tr = mtod(m0, struct mwltxrec *);
3176 if (wh != (struct ieee80211_frame *) &tr->wh)
3177 ovbcopy(wh, &tr->wh, hdrlen);
3179 * Note: the "firmware length" is actually the length
3180 * of the fully formed "802.11 payload". That is, it's
3181 * everything except for the 802.11 header. In particular
3182 * this includes all crypto material including the MIC!
3184 tr->fwlen = htole16(pktlen - hdrlen);
3187 * Load the DMA map so any coalescing is done. This
3188 * also calculates the number of descriptors we need.
3190 error = mwl_tx_dmasetup(sc, bf, m0);
3192 /* NB: stat collected in mwl_tx_dmasetup */
3193 DPRINTF(sc, MWL_DEBUG_XMIT,
3194 "%s: unable to setup dma\n", __func__);
3197 bf->bf_node = ni; /* NB: held reference */
3198 m0 = bf->bf_m; /* NB: may have changed */
3199 tr = mtod(m0, struct mwltxrec *);
3200 wh = (struct ieee80211_frame *)&tr->wh;
3203 * Formulate tx descriptor.
3208 ds->QosCtrl = qos; /* NB: already little-endian */
3211 * NB: multiframes should be zero because the descriptors
3212 * are initialized to zero. This should handle the case
3213 * where the driver is built with MWL_TXDESC=1 but we are
3214 * using firmware with multi-segment support.
3216 ds->PktPtr = htole32(bf->bf_segs[0].ds_addr);
3217 ds->PktLen = htole16(bf->bf_segs[0].ds_len);
3219 ds->multiframes = htole32(bf->bf_nseg);
3220 ds->PktLen = htole16(m0->m_pkthdr.len);
3221 for (i = 0; i < bf->bf_nseg; i++) {
3222 ds->PktPtrArray[i] = htole32(bf->bf_segs[i].ds_addr);
3223 ds->PktLenArray[i] = htole16(bf->bf_segs[i].ds_len);
3226 /* NB: pPhysNext, DataRate, and SapPktInfo setup once, don't touch */
3229 ds->ack_wcb_addr = 0;
3233 * Select transmit rate.
3235 switch (wh->i_fc[0] & IEEE80211_FC0_TYPE_MASK) {
3236 case IEEE80211_FC0_TYPE_MGT:
3237 sc->sc_stats.mst_tx_mgmt++;
3239 case IEEE80211_FC0_TYPE_CTL:
3240 /* NB: assign to BE q to avoid bursting */
3241 ds->TxPriority = MWL_WME_AC_BE;
3243 case IEEE80211_FC0_TYPE_DATA:
3245 const struct ieee80211_txparam *tp = ni->ni_txparms;
3247 * EAPOL frames get forced to a fixed rate and w/o
3248 * aggregation; otherwise check for any fixed rate
3249 * for the client (may depend on association state).
3251 if (m0->m_flags & M_EAPOL) {
3252 const struct mwl_vap *mvp = MWL_VAP_CONST(vap);
3253 ds->Format = mvp->mv_eapolformat;
3255 EAGLE_TXD_FIXED_RATE | EAGLE_TXD_DONT_AGGR);
3256 } else if (tp->ucastrate != IEEE80211_FIXED_RATE_NONE) {
3257 /* XXX pre-calculate per node */
3258 ds->Format = htole16(
3259 mwl_calcformat(tp->ucastrate, ni));
3260 ds->pad = htole16(EAGLE_TXD_FIXED_RATE);
3262 /* NB: EAPOL frames will never have qos set */
3264 ds->TxPriority = txq->qnum;
3266 else if (mwl_bastream_match(&mn->mn_ba[3], qos))
3267 ds->TxPriority = mn->mn_ba[3].txq;
3270 else if (mwl_bastream_match(&mn->mn_ba[2], qos))
3271 ds->TxPriority = mn->mn_ba[2].txq;
3274 else if (mwl_bastream_match(&mn->mn_ba[1], qos))
3275 ds->TxPriority = mn->mn_ba[1].txq;
3278 else if (mwl_bastream_match(&mn->mn_ba[0], qos))
3279 ds->TxPriority = mn->mn_ba[0].txq;
3282 ds->TxPriority = txq->qnum;
3284 ds->TxPriority = txq->qnum;
3287 device_printf(sc->sc_dev, "bogus frame type 0x%x (%s)\n",
3288 wh->i_fc[0] & IEEE80211_FC0_TYPE_MASK, __func__);
3289 sc->sc_stats.mst_tx_badframetype++;
3294 if (IFF_DUMPPKTS_XMIT(sc))
3295 ieee80211_dump_pkt(ic,
3296 mtod(m0, const uint8_t *)+sizeof(uint16_t),
3297 m0->m_len - sizeof(uint16_t), ds->DataRate, -1);
3300 ds->Status = htole32(EAGLE_TXD_STATUS_FW_OWNED);
3301 STAILQ_INSERT_TAIL(&txq->active, bf, bf_list);
3302 MWL_TXDESC_SYNC(txq, ds, BUS_DMASYNC_PREREAD | BUS_DMASYNC_PREWRITE);
3304 sc->sc_tx_timer = 5;
3305 MWL_TXQ_UNLOCK(txq);
3311 mwl_cvtlegacyrix(int rix)
3313 static const int ieeerates[] =
3314 { 2, 4, 11, 22, 44, 12, 18, 24, 36, 48, 72, 96, 108 };
3315 return (rix < nitems(ieeerates) ? ieeerates[rix] : 0);
3319 * Process completed xmit descriptors from the specified queue.
3322 mwl_tx_processq(struct mwl_softc *sc, struct mwl_txq *txq)
3324 #define EAGLE_TXD_STATUS_MCAST \
3325 (EAGLE_TXD_STATUS_MULTICAST_TX | EAGLE_TXD_STATUS_BROADCAST_TX)
3326 struct ieee80211com *ic = &sc->sc_ic;
3327 struct mwl_txbuf *bf;
3328 struct mwl_txdesc *ds;
3329 struct ieee80211_node *ni;
3330 struct mwl_node *an;
3334 DPRINTF(sc, MWL_DEBUG_TX_PROC, "%s: tx queue %u\n", __func__, txq->qnum);
3335 for (nreaped = 0;; nreaped++) {
3337 bf = STAILQ_FIRST(&txq->active);
3339 MWL_TXQ_UNLOCK(txq);
3343 MWL_TXDESC_SYNC(txq, ds,
3344 BUS_DMASYNC_POSTREAD | BUS_DMASYNC_POSTWRITE);
3345 if (ds->Status & htole32(EAGLE_TXD_STATUS_FW_OWNED)) {
3346 MWL_TXQ_UNLOCK(txq);
3349 STAILQ_REMOVE_HEAD(&txq->active, bf_list);
3350 MWL_TXQ_UNLOCK(txq);
3353 if (sc->sc_debug & MWL_DEBUG_XMIT_DESC)
3354 mwl_printtxbuf(bf, txq->qnum, nreaped);
3359 status = le32toh(ds->Status);
3360 if (status & EAGLE_TXD_STATUS_OK) {
3361 uint16_t Format = le16toh(ds->Format);
3362 uint8_t txant = MS(Format, EAGLE_TXD_ANTENNA);
3364 sc->sc_stats.mst_ant_tx[txant]++;
3365 if (status & EAGLE_TXD_STATUS_OK_RETRY)
3366 sc->sc_stats.mst_tx_retries++;
3367 if (status & EAGLE_TXD_STATUS_OK_MORE_RETRY)
3368 sc->sc_stats.mst_tx_mretries++;
3369 if (txq->qnum >= MWL_WME_AC_VO)
3370 ic->ic_wme.wme_hipri_traffic++;
3371 ni->ni_txrate = MS(Format, EAGLE_TXD_RATE);
3372 if ((Format & EAGLE_TXD_FORMAT_HT) == 0) {
3373 ni->ni_txrate = mwl_cvtlegacyrix(
3376 ni->ni_txrate |= IEEE80211_RATE_MCS;
3377 sc->sc_stats.mst_tx_rate = ni->ni_txrate;
3379 if (status & EAGLE_TXD_STATUS_FAILED_LINK_ERROR)
3380 sc->sc_stats.mst_tx_linkerror++;
3381 if (status & EAGLE_TXD_STATUS_FAILED_XRETRY)
3382 sc->sc_stats.mst_tx_xretries++;
3383 if (status & EAGLE_TXD_STATUS_FAILED_AGING)
3384 sc->sc_stats.mst_tx_aging++;
3385 if (bf->bf_m->m_flags & M_FF)
3386 sc->sc_stats.mst_ff_txerr++;
3388 if (bf->bf_m->m_flags & M_TXCB)
3389 /* XXX strip fw len in case header inspected */
3390 m_adj(bf->bf_m, sizeof(uint16_t));
3391 ieee80211_tx_complete(ni, bf->bf_m,
3392 (status & EAGLE_TXD_STATUS_OK) == 0);
3395 ds->Status = htole32(EAGLE_TXD_STATUS_IDLE);
3397 bus_dmamap_sync(sc->sc_dmat, bf->bf_dmamap,
3398 BUS_DMASYNC_POSTWRITE);
3399 bus_dmamap_unload(sc->sc_dmat, bf->bf_dmamap);
3401 mwl_puttxbuf_tail(txq, bf);
3404 #undef EAGLE_TXD_STATUS_MCAST
3408 * Deferred processing of transmit interrupt; special-cased
3409 * for four hardware queues, 0-3.
3412 mwl_tx_proc(void *arg, int npending)
3414 struct mwl_softc *sc = arg;
3418 * Process each active queue.
3421 if (!STAILQ_EMPTY(&sc->sc_txq[0].active))
3422 nreaped += mwl_tx_processq(sc, &sc->sc_txq[0]);
3423 if (!STAILQ_EMPTY(&sc->sc_txq[1].active))
3424 nreaped += mwl_tx_processq(sc, &sc->sc_txq[1]);
3425 if (!STAILQ_EMPTY(&sc->sc_txq[2].active))
3426 nreaped += mwl_tx_processq(sc, &sc->sc_txq[2]);
3427 if (!STAILQ_EMPTY(&sc->sc_txq[3].active))
3428 nreaped += mwl_tx_processq(sc, &sc->sc_txq[3]);
3431 sc->sc_tx_timer = 0;
3432 if (mbufq_first(&sc->sc_snd) != NULL) {
3433 /* NB: kick fw; the tx thread may have been preempted */
3434 mwl_hal_txstart(sc->sc_mh, 0);
3441 mwl_tx_draintxq(struct mwl_softc *sc, struct mwl_txq *txq)
3443 struct ieee80211_node *ni;
3444 struct mwl_txbuf *bf;
3448 * NB: this assumes output has been stopped and
3449 * we do not need to block mwl_tx_tasklet
3451 for (ix = 0;; ix++) {
3453 bf = STAILQ_FIRST(&txq->active);
3455 MWL_TXQ_UNLOCK(txq);
3458 STAILQ_REMOVE_HEAD(&txq->active, bf_list);
3459 MWL_TXQ_UNLOCK(txq);
3461 if (sc->sc_debug & MWL_DEBUG_RESET) {
3462 struct ieee80211com *ic = &sc->sc_ic;
3463 const struct mwltxrec *tr =
3464 mtod(bf->bf_m, const struct mwltxrec *);
3465 mwl_printtxbuf(bf, txq->qnum, ix);
3466 ieee80211_dump_pkt(ic, (const uint8_t *)&tr->wh,
3467 bf->bf_m->m_len - sizeof(tr->fwlen), 0, -1);
3469 #endif /* MWL_DEBUG */
3470 bus_dmamap_unload(sc->sc_dmat, bf->bf_dmamap);
3474 * Reclaim node reference.
3476 ieee80211_free_node(ni);
3480 mwl_puttxbuf_tail(txq, bf);
3485 * Drain the transmit queues and reclaim resources.
3488 mwl_draintxq(struct mwl_softc *sc)
3492 for (i = 0; i < MWL_NUM_TX_QUEUES; i++)
3493 mwl_tx_draintxq(sc, &sc->sc_txq[i]);
3494 sc->sc_tx_timer = 0;
3499 * Reset the transmit queues to a pristine state after a fw download.
3502 mwl_resettxq(struct mwl_softc *sc)
3506 for (i = 0; i < MWL_NUM_TX_QUEUES; i++)
3507 mwl_txq_reset(sc, &sc->sc_txq[i]);
3509 #endif /* MWL_DIAGAPI */
3512 * Clear the transmit queues of any frames submitted for the
3513 * specified vap. This is done when the vap is deleted so we
3514 * don't potentially reference the vap after it is gone.
3515 * Note we cannot remove the frames; we only reclaim the node
3519 mwl_cleartxq(struct mwl_softc *sc, struct ieee80211vap *vap)
3521 struct mwl_txq *txq;
3522 struct mwl_txbuf *bf;
3525 for (i = 0; i < MWL_NUM_TX_QUEUES; i++) {
3526 txq = &sc->sc_txq[i];
3528 STAILQ_FOREACH(bf, &txq->active, bf_list) {
3529 struct ieee80211_node *ni = bf->bf_node;
3530 if (ni != NULL && ni->ni_vap == vap) {
3532 ieee80211_free_node(ni);
3535 MWL_TXQ_UNLOCK(txq);
3540 mwl_recv_action(struct ieee80211_node *ni, const struct ieee80211_frame *wh,
3541 const uint8_t *frm, const uint8_t *efrm)
3543 struct mwl_softc *sc = ni->ni_ic->ic_softc;
3544 const struct ieee80211_action *ia;
3546 ia = (const struct ieee80211_action *) frm;
3547 if (ia->ia_category == IEEE80211_ACTION_CAT_HT &&
3548 ia->ia_action == IEEE80211_ACTION_HT_MIMOPWRSAVE) {
3549 const struct ieee80211_action_ht_mimopowersave *mps =
3550 (const struct ieee80211_action_ht_mimopowersave *) ia;
3552 mwl_hal_setmimops(sc->sc_mh, ni->ni_macaddr,
3553 mps->am_control & IEEE80211_A_HT_MIMOPWRSAVE_ENA,
3554 MS(mps->am_control, IEEE80211_A_HT_MIMOPWRSAVE_MODE));
3557 return sc->sc_recv_action(ni, wh, frm, efrm);
3561 mwl_addba_request(struct ieee80211_node *ni, struct ieee80211_tx_ampdu *tap,
3562 int dialogtoken, int baparamset, int batimeout)
3564 struct mwl_softc *sc = ni->ni_ic->ic_softc;
3565 struct ieee80211vap *vap = ni->ni_vap;
3566 struct mwl_node *mn = MWL_NODE(ni);
3567 struct mwl_bastate *bas;
3569 bas = tap->txa_private;
3571 const MWL_HAL_BASTREAM *sp;
3573 * Check for a free BA stream slot.
3576 if (mn->mn_ba[3].bastream == NULL)
3577 bas = &mn->mn_ba[3];
3581 if (mn->mn_ba[2].bastream == NULL)
3582 bas = &mn->mn_ba[2];
3586 if (mn->mn_ba[1].bastream == NULL)
3587 bas = &mn->mn_ba[1];
3591 if (mn->mn_ba[0].bastream == NULL)
3592 bas = &mn->mn_ba[0];
3596 /* sta already has max BA streams */
3597 /* XXX assign BA stream to highest priority tid */
3598 DPRINTF(sc, MWL_DEBUG_AMPDU,
3599 "%s: already has max bastreams\n", __func__);
3600 sc->sc_stats.mst_ampdu_reject++;
3603 /* NB: no held reference to ni */
3604 sp = mwl_hal_bastream_alloc(MWL_VAP(vap)->mv_hvap,
3605 (baparamset & IEEE80211_BAPS_POLICY_IMMEDIATE) != 0,
3606 ni->ni_macaddr, tap->txa_tid, ni->ni_htparam,
3610 * No available stream, return 0 so no
3611 * a-mpdu aggregation will be done.
3613 DPRINTF(sc, MWL_DEBUG_AMPDU,
3614 "%s: no bastream available\n", __func__);
3615 sc->sc_stats.mst_ampdu_nostream++;
3618 DPRINTF(sc, MWL_DEBUG_AMPDU, "%s: alloc bastream %p\n",
3620 /* NB: qos is left zero so we won't match in mwl_tx_start */
3622 tap->txa_private = bas;
3624 /* fetch current seq# from the firmware; if available */
3625 if (mwl_hal_bastream_get_seqno(sc->sc_mh, bas->bastream,
3626 vap->iv_opmode == IEEE80211_M_STA ? vap->iv_myaddr : ni->ni_macaddr,
3627 &tap->txa_start) != 0)
3629 return sc->sc_addba_request(ni, tap, dialogtoken, baparamset, batimeout);
3633 mwl_addba_response(struct ieee80211_node *ni, struct ieee80211_tx_ampdu *tap,
3634 int code, int baparamset, int batimeout)
3636 struct mwl_softc *sc = ni->ni_ic->ic_softc;
3637 struct mwl_bastate *bas;
3639 bas = tap->txa_private;
3641 /* XXX should not happen */
3642 DPRINTF(sc, MWL_DEBUG_AMPDU,
3643 "%s: no BA stream allocated, TID %d\n",
3644 __func__, tap->txa_tid);
3645 sc->sc_stats.mst_addba_nostream++;
3648 if (code == IEEE80211_STATUS_SUCCESS) {
3649 struct ieee80211vap *vap = ni->ni_vap;
3653 * Tell the firmware to setup the BA stream;
3654 * we know resources are available because we
3655 * pre-allocated one before forming the request.
3657 bufsiz = MS(baparamset, IEEE80211_BAPS_BUFSIZ);
3659 bufsiz = IEEE80211_AGGR_BAWMAX;
3660 error = mwl_hal_bastream_create(MWL_VAP(vap)->mv_hvap,
3661 bas->bastream, bufsiz, bufsiz, tap->txa_start);
3664 * Setup failed, return immediately so no a-mpdu
3665 * aggregation will be done.
3667 mwl_hal_bastream_destroy(sc->sc_mh, bas->bastream);
3668 mwl_bastream_free(bas);
3669 tap->txa_private = NULL;
3671 DPRINTF(sc, MWL_DEBUG_AMPDU,
3672 "%s: create failed, error %d, bufsiz %d TID %d "
3673 "htparam 0x%x\n", __func__, error, bufsiz,
3674 tap->txa_tid, ni->ni_htparam);
3675 sc->sc_stats.mst_bacreate_failed++;
3678 /* NB: cache txq to avoid ptr indirect */
3679 mwl_bastream_setup(bas, tap->txa_tid, bas->bastream->txq);
3680 DPRINTF(sc, MWL_DEBUG_AMPDU,
3681 "%s: bastream %p assigned to txq %d TID %d bufsiz %d "
3682 "htparam 0x%x\n", __func__, bas->bastream,
3683 bas->txq, tap->txa_tid, bufsiz, ni->ni_htparam);
3686 * Other side NAK'd us; return the resources.
3688 DPRINTF(sc, MWL_DEBUG_AMPDU,
3689 "%s: request failed with code %d, destroy bastream %p\n",
3690 __func__, code, bas->bastream);
3691 mwl_hal_bastream_destroy(sc->sc_mh, bas->bastream);
3692 mwl_bastream_free(bas);
3693 tap->txa_private = NULL;
3695 /* NB: firmware sends BAR so we don't need to */
3696 return sc->sc_addba_response(ni, tap, code, baparamset, batimeout);
3700 mwl_addba_stop(struct ieee80211_node *ni, struct ieee80211_tx_ampdu *tap)
3702 struct mwl_softc *sc = ni->ni_ic->ic_softc;
3703 struct mwl_bastate *bas;
3705 bas = tap->txa_private;
3707 DPRINTF(sc, MWL_DEBUG_AMPDU, "%s: destroy bastream %p\n",
3708 __func__, bas->bastream);
3709 mwl_hal_bastream_destroy(sc->sc_mh, bas->bastream);
3710 mwl_bastream_free(bas);
3711 tap->txa_private = NULL;
3713 sc->sc_addba_stop(ni, tap);
3717 * Setup the rx data structures. This should only be
3718 * done once or we may get out of sync with the firmware.
3721 mwl_startrecv(struct mwl_softc *sc)
3723 if (!sc->sc_recvsetup) {
3724 struct mwl_rxbuf *bf, *prev;
3725 struct mwl_rxdesc *ds;
3728 STAILQ_FOREACH(bf, &sc->sc_rxbuf, bf_list) {
3729 int error = mwl_rxbuf_init(sc, bf);
3731 DPRINTF(sc, MWL_DEBUG_RECV,
3732 "%s: mwl_rxbuf_init failed %d\n",
3738 ds->pPhysNext = htole32(bf->bf_daddr);
3745 htole32(STAILQ_FIRST(&sc->sc_rxbuf)->bf_daddr);
3747 sc->sc_recvsetup = 1;
3749 mwl_mode_init(sc); /* set filters, etc. */
3753 static MWL_HAL_APMODE
3754 mwl_getapmode(const struct ieee80211vap *vap, struct ieee80211_channel *chan)
3756 MWL_HAL_APMODE mode;
3758 if (IEEE80211_IS_CHAN_HT(chan)) {
3759 if (vap->iv_flags_ht & IEEE80211_FHT_PUREN)
3760 mode = AP_MODE_N_ONLY;
3761 else if (IEEE80211_IS_CHAN_5GHZ(chan))
3762 mode = AP_MODE_AandN;
3763 else if (vap->iv_flags & IEEE80211_F_PUREG)
3764 mode = AP_MODE_GandN;
3766 mode = AP_MODE_BandGandN;
3767 } else if (IEEE80211_IS_CHAN_ANYG(chan)) {
3768 if (vap->iv_flags & IEEE80211_F_PUREG)
3769 mode = AP_MODE_G_ONLY;
3771 mode = AP_MODE_MIXED;
3772 } else if (IEEE80211_IS_CHAN_B(chan))
3773 mode = AP_MODE_B_ONLY;
3774 else if (IEEE80211_IS_CHAN_A(chan))
3775 mode = AP_MODE_A_ONLY;
3777 mode = AP_MODE_MIXED; /* XXX should not happen? */
3782 mwl_setapmode(struct ieee80211vap *vap, struct ieee80211_channel *chan)
3784 struct mwl_hal_vap *hvap = MWL_VAP(vap)->mv_hvap;
3785 return mwl_hal_setapmode(hvap, mwl_getapmode(vap, chan));
3789 * Set/change channels.
3792 mwl_chan_set(struct mwl_softc *sc, struct ieee80211_channel *chan)
3794 struct mwl_hal *mh = sc->sc_mh;
3795 struct ieee80211com *ic = &sc->sc_ic;
3796 MWL_HAL_CHANNEL hchan;
3799 DPRINTF(sc, MWL_DEBUG_RESET, "%s: chan %u MHz/flags 0x%x\n",
3800 __func__, chan->ic_freq, chan->ic_flags);
3803 * Convert to a HAL channel description with
3804 * the flags constrained to reflect the current
3807 mwl_mapchan(&hchan, chan);
3808 mwl_hal_intrset(mh, 0); /* disable interrupts */
3810 mwl_draintxq(sc); /* clear pending tx frames */
3812 mwl_hal_setchannel(mh, &hchan);
3814 * Tx power is cap'd by the regulatory setting and
3815 * possibly a user-set limit. We pass the min of
3816 * these to the hal to apply them to the cal data
3820 maxtxpow = 2*chan->ic_maxregpower;
3821 if (maxtxpow > ic->ic_txpowlimit)
3822 maxtxpow = ic->ic_txpowlimit;
3823 mwl_hal_settxpower(mh, &hchan, maxtxpow / 2);
3824 /* NB: potentially change mcast/mgt rates */
3825 mwl_setcurchanrates(sc);
3828 * Update internal state.
3830 sc->sc_tx_th.wt_chan_freq = htole16(chan->ic_freq);
3831 sc->sc_rx_th.wr_chan_freq = htole16(chan->ic_freq);
3832 if (IEEE80211_IS_CHAN_A(chan)) {
3833 sc->sc_tx_th.wt_chan_flags = htole16(IEEE80211_CHAN_A);
3834 sc->sc_rx_th.wr_chan_flags = htole16(IEEE80211_CHAN_A);
3835 } else if (IEEE80211_IS_CHAN_ANYG(chan)) {
3836 sc->sc_tx_th.wt_chan_flags = htole16(IEEE80211_CHAN_G);
3837 sc->sc_rx_th.wr_chan_flags = htole16(IEEE80211_CHAN_G);
3839 sc->sc_tx_th.wt_chan_flags = htole16(IEEE80211_CHAN_B);
3840 sc->sc_rx_th.wr_chan_flags = htole16(IEEE80211_CHAN_B);
3842 sc->sc_curchan = hchan;
3843 mwl_hal_intrset(mh, sc->sc_imask);
3849 mwl_scan_start(struct ieee80211com *ic)
3851 struct mwl_softc *sc = ic->ic_softc;
3853 DPRINTF(sc, MWL_DEBUG_STATE, "%s\n", __func__);
3857 mwl_scan_end(struct ieee80211com *ic)
3859 struct mwl_softc *sc = ic->ic_softc;
3861 DPRINTF(sc, MWL_DEBUG_STATE, "%s\n", __func__);
3865 mwl_set_channel(struct ieee80211com *ic)
3867 struct mwl_softc *sc = ic->ic_softc;
3869 (void) mwl_chan_set(sc, ic->ic_curchan);
3873 * Handle a channel switch request. We inform the firmware
3874 * and mark the global state to suppress various actions.
3875 * NB: we issue only one request to the fw; we may be called
3876 * multiple times if there are multiple vap's.
3879 mwl_startcsa(struct ieee80211vap *vap)
3881 struct ieee80211com *ic = vap->iv_ic;
3882 struct mwl_softc *sc = ic->ic_softc;
3883 MWL_HAL_CHANNEL hchan;
3885 if (sc->sc_csapending)
3888 mwl_mapchan(&hchan, ic->ic_csa_newchan);
3889 /* 1 =>'s quiet channel */
3890 mwl_hal_setchannelswitchie(sc->sc_mh, &hchan, 1, ic->ic_csa_count);
3891 sc->sc_csapending = 1;
3895 * Plumb any static WEP key for the station. This is
3896 * necessary as we must propagate the key from the
3897 * global key table of the vap to each sta db entry.
3900 mwl_setanywepkey(struct ieee80211vap *vap, const uint8_t mac[IEEE80211_ADDR_LEN])
3902 if ((vap->iv_flags & (IEEE80211_F_PRIVACY|IEEE80211_F_WPA)) ==
3903 IEEE80211_F_PRIVACY &&
3904 vap->iv_def_txkey != IEEE80211_KEYIX_NONE &&
3905 vap->iv_nw_keys[vap->iv_def_txkey].wk_keyix != IEEE80211_KEYIX_NONE)
3906 (void) _mwl_key_set(vap, &vap->iv_nw_keys[vap->iv_def_txkey],
3911 mwl_peerstadb(struct ieee80211_node *ni, int aid, int staid, MWL_HAL_PEERINFO *pi)
3913 #define WME(ie) ((const struct ieee80211_wme_info *) ie)
3914 struct ieee80211vap *vap = ni->ni_vap;
3915 struct mwl_hal_vap *hvap;
3918 if (vap->iv_opmode == IEEE80211_M_WDS) {
3920 * WDS vap's do not have a f/w vap; instead they piggyback
3921 * on an AP vap and we must install the sta db entry and
3922 * crypto state using that AP's handle (the WDS vap has none).
3924 hvap = MWL_VAP(vap)->mv_ap_hvap;
3926 hvap = MWL_VAP(vap)->mv_hvap;
3927 error = mwl_hal_newstation(hvap, ni->ni_macaddr,
3929 ni->ni_flags & (IEEE80211_NODE_QOS | IEEE80211_NODE_HT),
3930 ni->ni_ies.wme_ie != NULL ? WME(ni->ni_ies.wme_ie)->wme_info : 0);
3933 * Setup security for this station. For sta mode this is
3934 * needed even though do the same thing on transition to
3935 * AUTH state because the call to mwl_hal_newstation
3936 * clobbers the crypto state we setup.
3938 mwl_setanywepkey(vap, ni->ni_macaddr);
3945 mwl_setglobalkeys(struct ieee80211vap *vap)
3947 struct ieee80211_key *wk;
3949 wk = &vap->iv_nw_keys[0];
3950 for (; wk < &vap->iv_nw_keys[IEEE80211_WEP_NKID]; wk++)
3951 if (wk->wk_keyix != IEEE80211_KEYIX_NONE)
3952 (void) _mwl_key_set(vap, wk, vap->iv_myaddr);
3956 * Convert a legacy rate set to a firmware bitmask.
3959 get_rate_bitmap(const struct ieee80211_rateset *rs)
3965 for (i = 0; i < rs->rs_nrates; i++)
3966 switch (rs->rs_rates[i] & IEEE80211_RATE_VAL) {
3967 case 2: rates |= 0x001; break;
3968 case 4: rates |= 0x002; break;
3969 case 11: rates |= 0x004; break;
3970 case 22: rates |= 0x008; break;
3971 case 44: rates |= 0x010; break;
3972 case 12: rates |= 0x020; break;
3973 case 18: rates |= 0x040; break;
3974 case 24: rates |= 0x080; break;
3975 case 36: rates |= 0x100; break;
3976 case 48: rates |= 0x200; break;
3977 case 72: rates |= 0x400; break;
3978 case 96: rates |= 0x800; break;
3979 case 108: rates |= 0x1000; break;
3985 * Construct an HT firmware bitmask from an HT rate set.
3988 get_htrate_bitmap(const struct ieee80211_htrateset *rs)
3994 for (i = 0; i < rs->rs_nrates; i++) {
3995 if (rs->rs_rates[i] < 16)
3996 rates |= 1<<rs->rs_rates[i];
4002 * Craft station database entry for station.
4003 * NB: use host byte order here, the hal handles byte swapping.
4005 static MWL_HAL_PEERINFO *
4006 mkpeerinfo(MWL_HAL_PEERINFO *pi, const struct ieee80211_node *ni)
4008 const struct ieee80211vap *vap = ni->ni_vap;
4010 memset(pi, 0, sizeof(*pi));
4011 pi->LegacyRateBitMap = get_rate_bitmap(&ni->ni_rates);
4012 pi->CapInfo = ni->ni_capinfo;
4013 if (ni->ni_flags & IEEE80211_NODE_HT) {
4014 /* HT capabilities, etc */
4015 pi->HTCapabilitiesInfo = ni->ni_htcap;
4016 /* XXX pi.HTCapabilitiesInfo */
4017 pi->MacHTParamInfo = ni->ni_htparam;
4018 pi->HTRateBitMap = get_htrate_bitmap(&ni->ni_htrates);
4019 pi->AddHtInfo.ControlChan = ni->ni_htctlchan;
4020 pi->AddHtInfo.AddChan = ni->ni_ht2ndchan;
4021 pi->AddHtInfo.OpMode = ni->ni_htopmode;
4022 pi->AddHtInfo.stbc = ni->ni_htstbc;
4024 /* constrain according to local configuration */
4025 if ((vap->iv_flags_ht & IEEE80211_FHT_SHORTGI40) == 0)
4026 pi->HTCapabilitiesInfo &= ~IEEE80211_HTCAP_SHORTGI40;
4027 if ((vap->iv_flags_ht & IEEE80211_FHT_SHORTGI20) == 0)
4028 pi->HTCapabilitiesInfo &= ~IEEE80211_HTCAP_SHORTGI20;
4029 if (ni->ni_chw != 40)
4030 pi->HTCapabilitiesInfo &= ~IEEE80211_HTCAP_CHWIDTH40;
4036 * Re-create the local sta db entry for a vap to ensure
4037 * up to date WME state is pushed to the firmware. Because
4038 * this resets crypto state this must be followed by a
4039 * reload of any keys in the global key table.
4042 mwl_localstadb(struct ieee80211vap *vap)
4044 #define WME(ie) ((const struct ieee80211_wme_info *) ie)
4045 struct mwl_hal_vap *hvap = MWL_VAP(vap)->mv_hvap;
4046 struct ieee80211_node *bss;
4047 MWL_HAL_PEERINFO pi;
4050 switch (vap->iv_opmode) {
4051 case IEEE80211_M_STA:
4053 error = mwl_hal_newstation(hvap, vap->iv_myaddr, 0, 0,
4054 vap->iv_state == IEEE80211_S_RUN ?
4055 mkpeerinfo(&pi, bss) : NULL,
4056 (bss->ni_flags & (IEEE80211_NODE_QOS | IEEE80211_NODE_HT)),
4057 bss->ni_ies.wme_ie != NULL ?
4058 WME(bss->ni_ies.wme_ie)->wme_info : 0);
4060 mwl_setglobalkeys(vap);
4062 case IEEE80211_M_HOSTAP:
4063 case IEEE80211_M_MBSS:
4064 error = mwl_hal_newstation(hvap, vap->iv_myaddr,
4065 0, 0, NULL, vap->iv_flags & IEEE80211_F_WME, 0);
4067 mwl_setglobalkeys(vap);
4078 mwl_newstate(struct ieee80211vap *vap, enum ieee80211_state nstate, int arg)
4080 struct mwl_vap *mvp = MWL_VAP(vap);
4081 struct mwl_hal_vap *hvap = mvp->mv_hvap;
4082 struct ieee80211com *ic = vap->iv_ic;
4083 struct ieee80211_node *ni = NULL;
4084 struct mwl_softc *sc = ic->ic_softc;
4085 struct mwl_hal *mh = sc->sc_mh;
4086 enum ieee80211_state ostate = vap->iv_state;
4089 DPRINTF(sc, MWL_DEBUG_STATE, "%s: %s: %s -> %s\n",
4090 vap->iv_ifp->if_xname, __func__,
4091 ieee80211_state_name[ostate], ieee80211_state_name[nstate]);
4093 callout_stop(&sc->sc_timer);
4095 * Clear current radar detection state.
4097 if (ostate == IEEE80211_S_CAC) {
4098 /* stop quiet mode radar detection */
4099 mwl_hal_setradardetection(mh, DR_CHK_CHANNEL_AVAILABLE_STOP);
4100 } else if (sc->sc_radarena) {
4101 /* stop in-service radar detection */
4102 mwl_hal_setradardetection(mh, DR_DFS_DISABLE);
4103 sc->sc_radarena = 0;
4106 * Carry out per-state actions before doing net80211 work.
4108 if (nstate == IEEE80211_S_INIT) {
4109 /* NB: only ap+sta vap's have a fw entity */
4112 } else if (nstate == IEEE80211_S_SCAN) {
4113 mwl_hal_start(hvap);
4114 /* NB: this disables beacon frames */
4115 mwl_hal_setinframode(hvap);
4116 } else if (nstate == IEEE80211_S_AUTH) {
4118 * Must create a sta db entry in case a WEP key needs to
4119 * be plumbed. This entry will be overwritten if we
4120 * associate; otherwise it will be reclaimed on node free.
4123 MWL_NODE(ni)->mn_hvap = hvap;
4124 (void) mwl_peerstadb(ni, 0, 0, NULL);
4125 } else if (nstate == IEEE80211_S_CSA) {
4126 /* XXX move to below? */
4127 if (vap->iv_opmode == IEEE80211_M_HOSTAP ||
4128 vap->iv_opmode == IEEE80211_M_MBSS)
4130 } else if (nstate == IEEE80211_S_CAC) {
4131 /* XXX move to below? */
4132 /* stop ap xmit and enable quiet mode radar detection */
4133 mwl_hal_setradardetection(mh, DR_CHK_CHANNEL_AVAILABLE_START);
4137 * Invoke the parent method to do net80211 work.
4139 error = mvp->mv_newstate(vap, nstate, arg);
4142 * Carry out work that must be done after net80211 runs;
4143 * this work requires up to date state (e.g. iv_bss).
4145 if (error == 0 && nstate == IEEE80211_S_RUN) {
4146 /* NB: collect bss node again, it may have changed */
4149 DPRINTF(sc, MWL_DEBUG_STATE,
4150 "%s: %s(RUN): iv_flags 0x%08x bintvl %d bssid %s "
4151 "capinfo 0x%04x chan %d\n",
4152 vap->iv_ifp->if_xname, __func__, vap->iv_flags,
4153 ni->ni_intval, ether_sprintf(ni->ni_bssid), ni->ni_capinfo,
4154 ieee80211_chan2ieee(ic, ic->ic_curchan));
4157 * Recreate local sta db entry to update WME/HT state.
4159 mwl_localstadb(vap);
4160 switch (vap->iv_opmode) {
4161 case IEEE80211_M_HOSTAP:
4162 case IEEE80211_M_MBSS:
4163 if (ostate == IEEE80211_S_CAC) {
4164 /* enable in-service radar detection */
4165 mwl_hal_setradardetection(mh,
4166 DR_IN_SERVICE_MONITOR_START);
4167 sc->sc_radarena = 1;
4170 * Allocate and setup the beacon frame
4171 * (and related state).
4173 error = mwl_reset_vap(vap, IEEE80211_S_RUN);
4175 DPRINTF(sc, MWL_DEBUG_STATE,
4176 "%s: beacon setup failed, error %d\n",
4180 /* NB: must be after setting up beacon */
4181 mwl_hal_start(hvap);
4183 case IEEE80211_M_STA:
4184 DPRINTF(sc, MWL_DEBUG_STATE, "%s: %s: aid 0x%x\n",
4185 vap->iv_ifp->if_xname, __func__, ni->ni_associd);
4187 * Set state now that we're associated.
4189 mwl_hal_setassocid(hvap, ni->ni_bssid, ni->ni_associd);
4191 mwl_hal_setrtsthreshold(hvap, vap->iv_rtsthreshold);
4192 if ((vap->iv_flags & IEEE80211_F_DWDS) &&
4193 sc->sc_ndwdsvaps++ == 0)
4194 mwl_hal_setdwds(mh, 1);
4196 case IEEE80211_M_WDS:
4197 DPRINTF(sc, MWL_DEBUG_STATE, "%s: %s: bssid %s\n",
4198 vap->iv_ifp->if_xname, __func__,
4199 ether_sprintf(ni->ni_bssid));
4200 mwl_seteapolformat(vap);
4206 * Set CS mode according to operating channel;
4207 * this mostly an optimization for 5GHz.
4209 * NB: must follow mwl_hal_start which resets csmode
4211 if (IEEE80211_IS_CHAN_5GHZ(ic->ic_bsschan))
4212 mwl_hal_setcsmode(mh, CSMODE_AGGRESSIVE);
4214 mwl_hal_setcsmode(mh, CSMODE_AUTO_ENA);
4216 * Start timer to prod firmware.
4218 if (sc->sc_ageinterval != 0)
4219 callout_reset(&sc->sc_timer, sc->sc_ageinterval*hz,
4220 mwl_agestations, sc);
4221 } else if (nstate == IEEE80211_S_SLEEP) {
4222 /* XXX set chip in power save */
4223 } else if ((vap->iv_flags & IEEE80211_F_DWDS) &&
4224 --sc->sc_ndwdsvaps == 0)
4225 mwl_hal_setdwds(mh, 0);
4231 * Manage station id's; these are separate from AID's
4232 * as AID's may have values out of the range of possible
4233 * station id's acceptable to the firmware.
4236 allocstaid(struct mwl_softc *sc, int aid)
4240 if (!(0 < aid && aid < MWL_MAXSTAID) || isset(sc->sc_staid, aid)) {
4241 /* NB: don't use 0 */
4242 for (staid = 1; staid < MWL_MAXSTAID; staid++)
4243 if (isclr(sc->sc_staid, staid))
4247 setbit(sc->sc_staid, staid);
4252 delstaid(struct mwl_softc *sc, int staid)
4254 clrbit(sc->sc_staid, staid);
4258 * Setup driver-specific state for a newly associated node.
4259 * Note that we're called also on a re-associate, the isnew
4260 * param tells us if this is the first time or not.
4263 mwl_newassoc(struct ieee80211_node *ni, int isnew)
4265 struct ieee80211vap *vap = ni->ni_vap;
4266 struct mwl_softc *sc = vap->iv_ic->ic_softc;
4267 struct mwl_node *mn = MWL_NODE(ni);
4268 MWL_HAL_PEERINFO pi;
4272 aid = IEEE80211_AID(ni->ni_associd);
4274 mn->mn_staid = allocstaid(sc, aid);
4275 mn->mn_hvap = MWL_VAP(vap)->mv_hvap;
4278 /* XXX reset BA stream? */
4280 DPRINTF(sc, MWL_DEBUG_NODE, "%s: mac %s isnew %d aid %d staid %d\n",
4281 __func__, ether_sprintf(ni->ni_macaddr), isnew, aid, mn->mn_staid);
4282 error = mwl_peerstadb(ni, aid, mn->mn_staid, mkpeerinfo(&pi, ni));
4284 DPRINTF(sc, MWL_DEBUG_NODE,
4285 "%s: error %d creating sta db entry\n",
4287 /* XXX how to deal with error? */
4292 * Periodically poke the firmware to age out station state
4293 * (power save queues, pending tx aggregates).
4296 mwl_agestations(void *arg)
4298 struct mwl_softc *sc = arg;
4300 mwl_hal_setkeepalive(sc->sc_mh);
4301 if (sc->sc_ageinterval != 0) /* NB: catch dynamic changes */
4302 callout_schedule(&sc->sc_timer, sc->sc_ageinterval*hz);
4305 static const struct mwl_hal_channel *
4306 findhalchannel(const MWL_HAL_CHANNELINFO *ci, int ieee)
4310 for (i = 0; i < ci->nchannels; i++) {
4311 const struct mwl_hal_channel *hc = &ci->channels[i];
4312 if (hc->ieee == ieee)
4319 mwl_setregdomain(struct ieee80211com *ic, struct ieee80211_regdomain *rd,
4320 int nchan, struct ieee80211_channel chans[])
4322 struct mwl_softc *sc = ic->ic_softc;
4323 struct mwl_hal *mh = sc->sc_mh;
4324 const MWL_HAL_CHANNELINFO *ci;
4327 for (i = 0; i < nchan; i++) {
4328 struct ieee80211_channel *c = &chans[i];
4329 const struct mwl_hal_channel *hc;
4331 if (IEEE80211_IS_CHAN_2GHZ(c)) {
4332 mwl_hal_getchannelinfo(mh, MWL_FREQ_BAND_2DOT4GHZ,
4333 IEEE80211_IS_CHAN_HT40(c) ?
4334 MWL_CH_40_MHz_WIDTH : MWL_CH_20_MHz_WIDTH, &ci);
4335 } else if (IEEE80211_IS_CHAN_5GHZ(c)) {
4336 mwl_hal_getchannelinfo(mh, MWL_FREQ_BAND_5GHZ,
4337 IEEE80211_IS_CHAN_HT40(c) ?
4338 MWL_CH_40_MHz_WIDTH : MWL_CH_20_MHz_WIDTH, &ci);
4340 device_printf(sc->sc_dev,
4341 "%s: channel %u freq %u/0x%x not 2.4/5GHz\n",
4342 __func__, c->ic_ieee, c->ic_freq, c->ic_flags);
4346 * Verify channel has cal data and cap tx power.
4348 hc = findhalchannel(ci, c->ic_ieee);
4350 if (c->ic_maxpower > 2*hc->maxTxPow)
4351 c->ic_maxpower = 2*hc->maxTxPow;
4354 if (IEEE80211_IS_CHAN_HT40(c)) {
4356 * Look for the extension channel since the
4357 * hal table only has the primary channel.
4359 hc = findhalchannel(ci, c->ic_extieee);
4361 if (c->ic_maxpower > 2*hc->maxTxPow)
4362 c->ic_maxpower = 2*hc->maxTxPow;
4366 device_printf(sc->sc_dev,
4367 "%s: no cal data for channel %u ext %u freq %u/0x%x\n",
4368 __func__, c->ic_ieee, c->ic_extieee,
4369 c->ic_freq, c->ic_flags);
4377 #define IEEE80211_CHAN_HTG (IEEE80211_CHAN_HT|IEEE80211_CHAN_G)
4378 #define IEEE80211_CHAN_HTA (IEEE80211_CHAN_HT|IEEE80211_CHAN_A)
4381 addht40channels(struct ieee80211_channel chans[], int maxchans, int *nchans,
4382 const MWL_HAL_CHANNELINFO *ci, int flags)
4386 for (i = 0; i < ci->nchannels; i++) {
4387 const struct mwl_hal_channel *hc = &ci->channels[i];
4389 error = ieee80211_add_channel_ht40(chans, maxchans, nchans,
4390 hc->ieee, hc->maxTxPow, flags);
4391 if (error != 0 && error != ENOENT)
4397 addchannels(struct ieee80211_channel chans[], int maxchans, int *nchans,
4398 const MWL_HAL_CHANNELINFO *ci, const uint8_t bands[])
4403 for (i = 0; i < ci->nchannels && error == 0; i++) {
4404 const struct mwl_hal_channel *hc = &ci->channels[i];
4406 error = ieee80211_add_channel(chans, maxchans, nchans,
4407 hc->ieee, hc->freq, hc->maxTxPow, 0, bands);
4412 getchannels(struct mwl_softc *sc, int maxchans, int *nchans,
4413 struct ieee80211_channel chans[])
4415 const MWL_HAL_CHANNELINFO *ci;
4416 uint8_t bands[IEEE80211_MODE_BYTES];
4419 * Use the channel info from the hal to craft the
4420 * channel list. Note that we pass back an unsorted
4421 * list; the caller is required to sort it for us
4425 if (mwl_hal_getchannelinfo(sc->sc_mh,
4426 MWL_FREQ_BAND_2DOT4GHZ, MWL_CH_20_MHz_WIDTH, &ci) == 0) {
4427 memset(bands, 0, sizeof(bands));
4428 setbit(bands, IEEE80211_MODE_11B);
4429 setbit(bands, IEEE80211_MODE_11G);
4430 setbit(bands, IEEE80211_MODE_11NG);
4431 addchannels(chans, maxchans, nchans, ci, bands);
4433 if (mwl_hal_getchannelinfo(sc->sc_mh,
4434 MWL_FREQ_BAND_5GHZ, MWL_CH_20_MHz_WIDTH, &ci) == 0) {
4435 memset(bands, 0, sizeof(bands));
4436 setbit(bands, IEEE80211_MODE_11A);
4437 setbit(bands, IEEE80211_MODE_11NA);
4438 addchannels(chans, maxchans, nchans, ci, bands);
4440 if (mwl_hal_getchannelinfo(sc->sc_mh,
4441 MWL_FREQ_BAND_2DOT4GHZ, MWL_CH_40_MHz_WIDTH, &ci) == 0)
4442 addht40channels(chans, maxchans, nchans, ci, IEEE80211_CHAN_HTG);
4443 if (mwl_hal_getchannelinfo(sc->sc_mh,
4444 MWL_FREQ_BAND_5GHZ, MWL_CH_40_MHz_WIDTH, &ci) == 0)
4445 addht40channels(chans, maxchans, nchans, ci, IEEE80211_CHAN_HTA);
4449 mwl_getradiocaps(struct ieee80211com *ic,
4450 int maxchans, int *nchans, struct ieee80211_channel chans[])
4452 struct mwl_softc *sc = ic->ic_softc;
4454 getchannels(sc, maxchans, nchans, chans);
4458 mwl_getchannels(struct mwl_softc *sc)
4460 struct ieee80211com *ic = &sc->sc_ic;
4463 * Use the channel info from the hal to craft the
4464 * channel list for net80211. Note that we pass up
4465 * an unsorted list; net80211 will sort it for us.
4467 memset(ic->ic_channels, 0, sizeof(ic->ic_channels));
4469 getchannels(sc, IEEE80211_CHAN_MAX, &ic->ic_nchans, ic->ic_channels);
4471 ic->ic_regdomain.regdomain = SKU_DEBUG;
4472 ic->ic_regdomain.country = CTRY_DEFAULT;
4473 ic->ic_regdomain.location = 'I';
4474 ic->ic_regdomain.isocc[0] = ' '; /* XXX? */
4475 ic->ic_regdomain.isocc[1] = ' ';
4476 return (ic->ic_nchans == 0 ? EIO : 0);
4478 #undef IEEE80211_CHAN_HTA
4479 #undef IEEE80211_CHAN_HTG
4483 mwl_printrxbuf(const struct mwl_rxbuf *bf, u_int ix)
4485 const struct mwl_rxdesc *ds = bf->bf_desc;
4486 uint32_t status = le32toh(ds->Status);
4488 printf("R[%2u] (DS.V:%p DS.P:0x%jx) NEXT:%08x DATA:%08x RC:%02x%s\n"
4489 " STAT:%02x LEN:%04x RSSI:%02x CHAN:%02x RATE:%02x QOS:%04x HT:%04x\n",
4490 ix, ds, (uintmax_t)bf->bf_daddr, le32toh(ds->pPhysNext),
4491 le32toh(ds->pPhysBuffData), ds->RxControl,
4492 ds->RxControl != EAGLE_RXD_CTRL_DRIVER_OWN ?
4493 "" : (status & EAGLE_RXD_STATUS_OK) ? " *" : " !",
4494 ds->Status, le16toh(ds->PktLen), ds->RSSI, ds->Channel,
4495 ds->Rate, le16toh(ds->QosCtrl), le16toh(ds->HtSig2));
4499 mwl_printtxbuf(const struct mwl_txbuf *bf, u_int qnum, u_int ix)
4501 const struct mwl_txdesc *ds = bf->bf_desc;
4502 uint32_t status = le32toh(ds->Status);
4504 printf("Q%u[%3u]", qnum, ix);
4505 printf(" (DS.V:%p DS.P:0x%jx)\n", ds, (uintmax_t)bf->bf_daddr);
4506 printf(" NEXT:%08x DATA:%08x LEN:%04x STAT:%08x%s\n",
4507 le32toh(ds->pPhysNext),
4508 le32toh(ds->PktPtr), le16toh(ds->PktLen), status,
4509 status & EAGLE_TXD_STATUS_USED ?
4510 "" : (status & 3) != 0 ? " *" : " !");
4511 printf(" RATE:%02x PRI:%x QOS:%04x SAP:%08x FORMAT:%04x\n",
4512 ds->DataRate, ds->TxPriority, le16toh(ds->QosCtrl),
4513 le32toh(ds->SapPktInfo), le16toh(ds->Format));
4515 printf(" MULTIFRAMES:%u LEN:%04x %04x %04x %04x %04x %04x\n"
4516 , le32toh(ds->multiframes)
4517 , le16toh(ds->PktLenArray[0]), le16toh(ds->PktLenArray[1])
4518 , le16toh(ds->PktLenArray[2]), le16toh(ds->PktLenArray[3])
4519 , le16toh(ds->PktLenArray[4]), le16toh(ds->PktLenArray[5])
4521 printf(" DATA:%08x %08x %08x %08x %08x %08x\n"
4522 , le32toh(ds->PktPtrArray[0]), le32toh(ds->PktPtrArray[1])
4523 , le32toh(ds->PktPtrArray[2]), le32toh(ds->PktPtrArray[3])
4524 , le32toh(ds->PktPtrArray[4]), le32toh(ds->PktPtrArray[5])
4528 { const uint8_t *cp = (const uint8_t *) ds;
4530 for (i = 0; i < sizeof(struct mwl_txdesc); i++) {
4531 printf("%02x ", cp[i]);
4532 if (((i+1) % 16) == 0)
4539 #endif /* MWL_DEBUG */
4543 mwl_txq_dump(struct mwl_txq *txq)
4545 struct mwl_txbuf *bf;
4549 STAILQ_FOREACH(bf, &txq->active, bf_list) {
4550 struct mwl_txdesc *ds = bf->bf_desc;
4551 MWL_TXDESC_SYNC(txq, ds,
4552 BUS_DMASYNC_POSTREAD | BUS_DMASYNC_POSTWRITE);
4554 mwl_printtxbuf(bf, txq->qnum, i);
4558 MWL_TXQ_UNLOCK(txq);
4563 mwl_watchdog(void *arg)
4565 struct mwl_softc *sc = arg;
4567 callout_reset(&sc->sc_watchdog, hz, mwl_watchdog, sc);
4568 if (sc->sc_tx_timer == 0 || --sc->sc_tx_timer > 0)
4571 if (sc->sc_running && !sc->sc_invalid) {
4572 if (mwl_hal_setkeepalive(sc->sc_mh))
4573 device_printf(sc->sc_dev,
4574 "transmit timeout (firmware hung?)\n");
4576 device_printf(sc->sc_dev,
4577 "transmit timeout\n");
4580 mwl_txq_dump(&sc->sc_txq[0]);/*XXX*/
4582 counter_u64_add(sc->sc_ic.ic_oerrors, 1);
4583 sc->sc_stats.mst_watchdog++;
4589 * Diagnostic interface to the HAL. This is used by various
4590 * tools to do things like retrieve register contents for
4591 * debugging. The mechanism is intentionally opaque so that
4592 * it can change frequently w/o concern for compatibility.
4595 mwl_ioctl_diag(struct mwl_softc *sc, struct mwl_diag *md)
4597 struct mwl_hal *mh = sc->sc_mh;
4598 u_int id = md->md_id & MWL_DIAG_ID;
4599 void *indata = NULL;
4600 void *outdata = NULL;
4601 u_int32_t insize = md->md_in_size;
4602 u_int32_t outsize = md->md_out_size;
4605 if (md->md_id & MWL_DIAG_IN) {
4609 indata = malloc(insize, M_TEMP, M_NOWAIT);
4610 if (indata == NULL) {
4614 error = copyin(md->md_in_data, indata, insize);
4618 if (md->md_id & MWL_DIAG_DYN) {
4620 * Allocate a buffer for the results (otherwise the HAL
4621 * returns a pointer to a buffer where we can read the
4622 * results). Note that we depend on the HAL leaving this
4623 * pointer for us to use below in reclaiming the buffer;
4624 * may want to be more defensive.
4626 outdata = malloc(outsize, M_TEMP, M_NOWAIT);
4627 if (outdata == NULL) {
4632 if (mwl_hal_getdiagstate(mh, id, indata, insize, &outdata, &outsize)) {
4633 if (outsize < md->md_out_size)
4634 md->md_out_size = outsize;
4635 if (outdata != NULL)
4636 error = copyout(outdata, md->md_out_data,
4642 if ((md->md_id & MWL_DIAG_IN) && indata != NULL)
4643 free(indata, M_TEMP);
4644 if ((md->md_id & MWL_DIAG_DYN) && outdata != NULL)
4645 free(outdata, M_TEMP);
4650 mwl_ioctl_reset(struct mwl_softc *sc, struct mwl_diag *md)
4652 struct mwl_hal *mh = sc->sc_mh;
4655 MWL_LOCK_ASSERT(sc);
4657 if (md->md_id == 0 && mwl_hal_fwload(mh, NULL) != 0) {
4658 device_printf(sc->sc_dev, "unable to load firmware\n");
4661 if (mwl_hal_gethwspecs(mh, &sc->sc_hwspecs) != 0) {
4662 device_printf(sc->sc_dev, "unable to fetch h/w specs\n");
4665 error = mwl_setupdma(sc);
4667 /* NB: mwl_setupdma prints a msg */
4671 * Reset tx/rx data structures; after reload we must
4672 * re-start the driver's notion of the next xmit/recv.
4674 mwl_draintxq(sc); /* clear pending frames */
4675 mwl_resettxq(sc); /* rebuild tx q lists */
4676 sc->sc_rxnext = NULL; /* force rx to start at the list head */
4679 #endif /* MWL_DIAGAPI */
4682 mwl_parent(struct ieee80211com *ic)
4684 struct mwl_softc *sc = ic->ic_softc;
4688 if (ic->ic_nrunning > 0) {
4689 if (sc->sc_running) {
4691 * To avoid rescanning another access point,
4692 * do not call mwl_init() here. Instead,
4693 * only reflect promisc mode settings.
4698 * Beware of being called during attach/detach
4699 * to reset promiscuous mode. In that case we
4700 * will still be marked UP but not RUNNING.
4701 * However trying to re-init the interface
4702 * is the wrong thing to do as we've already
4703 * torn down much of our state. There's
4704 * probably a better way to deal with this.
4706 if (!sc->sc_invalid) {
4707 mwl_init(sc); /* XXX lose error */
4715 ieee80211_start_all(ic);
4719 mwl_ioctl(struct ieee80211com *ic, u_long cmd, void *data)
4721 struct mwl_softc *sc = ic->ic_softc;
4722 struct ifreq *ifr = data;
4727 mwl_hal_gethwstats(sc->sc_mh, &sc->sc_stats.hw_stats);
4729 /* NB: embed these numbers to get a consistent view */
4730 sc->sc_stats.mst_tx_packets =
4731 ifp->if_get_counter(ifp, IFCOUNTER_OPACKETS);
4732 sc->sc_stats.mst_rx_packets =
4733 ifp->if_get_counter(ifp, IFCOUNTER_IPACKETS);
4736 * NB: Drop the softc lock in case of a page fault;
4737 * we'll accept any potential inconsisentcy in the
4738 * statistics. The alternative is to copy the data
4739 * to a local structure.
4741 return (copyout(&sc->sc_stats, ifr_data_get_ptr(ifr),
4742 sizeof (sc->sc_stats)));
4745 /* XXX check privs */
4746 return mwl_ioctl_diag(sc, (struct mwl_diag *) ifr);
4748 /* XXX check privs */
4750 error = mwl_ioctl_reset(sc,(struct mwl_diag *) ifr);
4753 #endif /* MWL_DIAGAPI */
4763 mwl_sysctl_debug(SYSCTL_HANDLER_ARGS)
4765 struct mwl_softc *sc = arg1;
4768 debug = sc->sc_debug | (mwl_hal_getdebug(sc->sc_mh) << 24);
4769 error = sysctl_handle_int(oidp, &debug, 0, req);
4770 if (error || !req->newptr)
4772 mwl_hal_setdebug(sc->sc_mh, debug >> 24);
4773 sc->sc_debug = debug & 0x00ffffff;
4776 #endif /* MWL_DEBUG */
4779 mwl_sysctlattach(struct mwl_softc *sc)
4782 struct sysctl_ctx_list *ctx = device_get_sysctl_ctx(sc->sc_dev);
4783 struct sysctl_oid *tree = device_get_sysctl_tree(sc->sc_dev);
4785 sc->sc_debug = mwl_debug;
4786 SYSCTL_ADD_PROC(ctx, SYSCTL_CHILDREN(tree), OID_AUTO,
4787 "debug", CTLTYPE_INT | CTLFLAG_RW, sc, 0,
4788 mwl_sysctl_debug, "I", "control debugging printfs");
4793 * Announce various information on device/driver attach.
4796 mwl_announce(struct mwl_softc *sc)
4799 device_printf(sc->sc_dev, "Rev A%d hardware, v%d.%d.%d.%d firmware (regioncode %d)\n",
4800 sc->sc_hwspecs.hwVersion,
4801 (sc->sc_hwspecs.fwReleaseNumber>>24) & 0xff,
4802 (sc->sc_hwspecs.fwReleaseNumber>>16) & 0xff,
4803 (sc->sc_hwspecs.fwReleaseNumber>>8) & 0xff,
4804 (sc->sc_hwspecs.fwReleaseNumber>>0) & 0xff,
4805 sc->sc_hwspecs.regionCode);
4806 sc->sc_fwrelease = sc->sc_hwspecs.fwReleaseNumber;
4810 for (i = 0; i <= WME_AC_VO; i++) {
4811 struct mwl_txq *txq = sc->sc_ac2q[i];
4812 device_printf(sc->sc_dev, "Use hw queue %u for %s traffic\n",
4813 txq->qnum, ieee80211_wme_acnames[i]);
4816 if (bootverbose || mwl_rxdesc != MWL_RXDESC)
4817 device_printf(sc->sc_dev, "using %u rx descriptors\n", mwl_rxdesc);
4818 if (bootverbose || mwl_rxbuf != MWL_RXBUF)
4819 device_printf(sc->sc_dev, "using %u rx buffers\n", mwl_rxbuf);
4820 if (bootverbose || mwl_txbuf != MWL_TXBUF)
4821 device_printf(sc->sc_dev, "using %u tx buffers\n", mwl_txbuf);
4822 if (bootverbose && mwl_hal_ismbsscapable(sc->sc_mh))
4823 device_printf(sc->sc_dev, "multi-bss support\n");
4824 #ifdef MWL_TX_NODROP
4826 device_printf(sc->sc_dev, "no tx drop\n");