2 * Copyright (c) 2007-2014 QLogic Corporation. All rights reserved.
4 * Redistribution and use in source and binary forms, with or without
5 * modification, are permitted provided that the following conditions
8 * 1. Redistributions of source code must retain the above copyright
9 * notice, this list of conditions and the following disclaimer.
10 * 2. Redistributions in binary form must reproduce the above copyright
11 * notice, this list of conditions and the following disclaimer in the
12 * documentation and/or other materials provided with the distribution.
14 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS'
15 * AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
16 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
17 * ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS
18 * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
19 * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
20 * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
21 * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
22 * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
23 * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF
24 * THE POSSIBILITY OF SUCH DAMAGE.
27 #include <sys/cdefs.h>
28 __FBSDID("$FreeBSD$");
31 #include "bxe_stats.h"
34 #define BITS_PER_LONG 32
36 #define BITS_PER_LONG 64
40 bxe_hilo(uint32_t *hiref)
42 uint32_t lo = *(hiref + 1);
43 #if (BITS_PER_LONG == 64)
45 return (HILO_U64(hi, lo));
51 static inline uint16_t
52 bxe_get_port_stats_dma_len(struct bxe_softc *sc)
57 /* 'newest' convention - shmem2 contains the size of the port stats */
58 if (SHMEM2_HAS(sc, sizeof_port_stats)) {
59 size = SHMEM2_RD(sc, sizeof_port_stats);
64 /* prevent newer BC from causing buffer overflow */
65 if (res > sizeof(struct host_port_stats)) {
66 res = sizeof(struct host_port_stats);
71 * Older convention - all BCs support the port stats fields up until
72 * the 'not_used' field
75 res = (offsetof(struct host_port_stats, not_used) + 4);
77 /* if PFC stats are supported by the MFW, DMA them as well */
78 if (sc->devinfo.bc_ver >= REQ_BC_VER_4_PFC_STATS_SUPPORTED) {
79 res += (offsetof(struct host_port_stats, pfc_frames_rx_lo) -
80 offsetof(struct host_port_stats, pfc_frames_tx_hi) + 4);
86 DBASSERT(sc, !(res > 2 * DMAE_LEN32_RD_MAX), ("big stats dmae length\n"));
91 * Init service functions
95 bxe_dp_stats(struct bxe_softc *sc)
105 " drv_stats_counter %d\n"
107 " stats_counters_addrs %x %x\n",
108 sc->fw_stats_req->hdr.cmd_num,
109 sc->fw_stats_req->hdr.reserved0,
110 sc->fw_stats_req->hdr.drv_stats_counter,
111 sc->fw_stats_req->hdr.reserved1,
112 sc->fw_stats_req->hdr.stats_counters_addrs.hi,
113 sc->fw_stats_req->hdr.stats_counters_addrs.lo);
115 for (i = 0; i < sc->fw_stats_req->hdr.cmd_num; i++) {
124 sc->fw_stats_req->query[i].kind,
125 sc->fw_stats_req->query[i].index,
126 sc->fw_stats_req->query[i].funcID,
127 sc->fw_stats_req->query[i].reserved,
128 sc->fw_stats_req->query[i].address.hi,
129 sc->fw_stats_req->query[i].address.lo);
134 * Post the next statistics ramrod. Protect it with the lock in
135 * order to ensure the strict order between statistics ramrods
136 * (each ramrod has a sequence number passed in a
137 * sc->fw_stats_req->hdr.drv_stats_counter and ramrods must be
141 bxe_storm_stats_post(struct bxe_softc *sc)
145 if (!sc->stats_pending) {
148 if (sc->stats_pending) {
149 BXE_STATS_UNLOCK(sc);
153 sc->fw_stats_req->hdr.drv_stats_counter =
154 htole16(sc->stats_counter++);
157 "sending statistics ramrod %d\n",
158 le16toh(sc->fw_stats_req->hdr.drv_stats_counter));
160 /* adjust the ramrod to include VF queues statistics */
161 // XXX bxe_iov_adjust_stats_req(sc);
165 /* send FW stats ramrod */
166 rc = bxe_sp_post(sc, RAMROD_CMD_ID_COMMON_STAT_QUERY, 0,
167 U64_HI(sc->fw_stats_req_mapping),
168 U64_LO(sc->fw_stats_req_mapping),
169 NONE_CONNECTION_TYPE);
171 sc->stats_pending = 1;
174 BXE_STATS_UNLOCK(sc);
179 bxe_hw_stats_post(struct bxe_softc *sc)
181 struct dmae_cmd *dmae = &sc->stats_dmae;
182 uint32_t *stats_comp = BXE_SP(sc, stats_comp);
186 *stats_comp = DMAE_COMP_VAL;
187 if (CHIP_REV_IS_SLOW(sc)) {
191 /* Update MCP's statistics if possible */
193 memcpy(BXE_SP(sc, func_stats), &sc->func_stats,
194 sizeof(sc->func_stats));
198 if (sc->executer_idx) {
199 loader_idx = PMF_DMAE_C(sc);
200 opcode = bxe_dmae_opcode(sc, DMAE_SRC_PCI, DMAE_DST_GRC,
201 TRUE, DMAE_COMP_GRC);
202 opcode = bxe_dmae_opcode_clr_src_reset(opcode);
204 memset(dmae, 0, sizeof(struct dmae_cmd));
205 dmae->opcode = opcode;
206 dmae->src_addr_lo = U64_LO(BXE_SP_MAPPING(sc, dmae[0]));
207 dmae->src_addr_hi = U64_HI(BXE_SP_MAPPING(sc, dmae[0]));
208 dmae->dst_addr_lo = ((DMAE_REG_CMD_MEM +
209 sizeof(struct dmae_cmd) *
210 (loader_idx + 1)) >> 2);
211 dmae->dst_addr_hi = 0;
212 dmae->len = sizeof(struct dmae_cmd) >> 2;
213 if (CHIP_IS_E1(sc)) {
216 dmae->comp_addr_lo = (dmae_reg_go_c[loader_idx + 1] >> 2);
217 dmae->comp_addr_hi = 0;
221 bxe_post_dmae(sc, dmae, loader_idx);
222 } else if (sc->func_stx) {
224 bxe_post_dmae(sc, dmae, INIT_DMAE_C(sc));
229 bxe_stats_comp(struct bxe_softc *sc)
231 uint32_t *stats_comp = BXE_SP(sc, stats_comp);
234 while (*stats_comp != DMAE_COMP_VAL) {
236 BLOGE(sc, "Timeout waiting for stats finished\n");
237 if(sc->trigger_grcdump) {
252 * Statistics service functions
256 bxe_stats_pmf_update(struct bxe_softc *sc)
258 struct dmae_cmd *dmae;
260 int loader_idx = PMF_DMAE_C(sc);
261 uint32_t *stats_comp = BXE_SP(sc, stats_comp);
263 if (sc->devinfo.bc_ver <= 0x06001400) {
265 * Bootcode v6.0.21 fixed a GRC timeout that occurs when accessing
266 * BRB registers while the BRB block is in reset. The DMA transfer
267 * below triggers this issue resulting in the DMAE to stop
268 * functioning. Skip this initial stats transfer for old bootcode
269 * versions <= 6.0.20.
275 if (!sc->port.pmf || !sc->port.port_stx) {
280 sc->executer_idx = 0;
282 opcode = bxe_dmae_opcode(sc, DMAE_SRC_GRC, DMAE_DST_PCI, FALSE, 0);
284 dmae = BXE_SP(sc, dmae[sc->executer_idx++]);
285 dmae->opcode = bxe_dmae_opcode_add_comp(opcode, DMAE_COMP_GRC);
286 dmae->src_addr_lo = (sc->port.port_stx >> 2);
287 dmae->src_addr_hi = 0;
288 dmae->dst_addr_lo = U64_LO(BXE_SP_MAPPING(sc, port_stats));
289 dmae->dst_addr_hi = U64_HI(BXE_SP_MAPPING(sc, port_stats));
290 dmae->len = DMAE_LEN32_RD_MAX;
291 dmae->comp_addr_lo = (dmae_reg_go_c[loader_idx] >> 2);
292 dmae->comp_addr_hi = 0;
295 dmae = BXE_SP(sc, dmae[sc->executer_idx++]);
296 dmae->opcode = bxe_dmae_opcode_add_comp(opcode, DMAE_COMP_PCI);
297 dmae->src_addr_lo = ((sc->port.port_stx >> 2) + DMAE_LEN32_RD_MAX);
298 dmae->src_addr_hi = 0;
299 dmae->dst_addr_lo = U64_LO(BXE_SP_MAPPING(sc, port_stats) +
300 DMAE_LEN32_RD_MAX * 4);
301 dmae->dst_addr_hi = U64_HI(BXE_SP_MAPPING(sc, port_stats) +
302 DMAE_LEN32_RD_MAX * 4);
303 dmae->len = (bxe_get_port_stats_dma_len(sc) - DMAE_LEN32_RD_MAX);
305 dmae->comp_addr_lo = U64_LO(BXE_SP_MAPPING(sc, stats_comp));
306 dmae->comp_addr_hi = U64_HI(BXE_SP_MAPPING(sc, stats_comp));
307 dmae->comp_val = DMAE_COMP_VAL;
310 bxe_hw_stats_post(sc);
315 bxe_port_stats_init(struct bxe_softc *sc)
317 struct dmae_cmd *dmae;
318 int port = SC_PORT(sc);
320 int loader_idx = PMF_DMAE_C(sc);
322 uint32_t *stats_comp = BXE_SP(sc, stats_comp);
325 if (!sc->link_vars.link_up || !sc->port.pmf) {
330 sc->executer_idx = 0;
333 opcode = bxe_dmae_opcode(sc, DMAE_SRC_PCI, DMAE_DST_GRC,
334 TRUE, DMAE_COMP_GRC);
336 if (sc->port.port_stx) {
337 dmae = BXE_SP(sc, dmae[sc->executer_idx++]);
338 dmae->opcode = opcode;
339 dmae->src_addr_lo = U64_LO(BXE_SP_MAPPING(sc, port_stats));
340 dmae->src_addr_hi = U64_HI(BXE_SP_MAPPING(sc, port_stats));
341 dmae->dst_addr_lo = sc->port.port_stx >> 2;
342 dmae->dst_addr_hi = 0;
343 dmae->len = bxe_get_port_stats_dma_len(sc);
344 dmae->comp_addr_lo = dmae_reg_go_c[loader_idx] >> 2;
345 dmae->comp_addr_hi = 0;
350 dmae = BXE_SP(sc, dmae[sc->executer_idx++]);
351 dmae->opcode = opcode;
352 dmae->src_addr_lo = U64_LO(BXE_SP_MAPPING(sc, func_stats));
353 dmae->src_addr_hi = U64_HI(BXE_SP_MAPPING(sc, func_stats));
354 dmae->dst_addr_lo = (sc->func_stx >> 2);
355 dmae->dst_addr_hi = 0;
356 dmae->len = (sizeof(struct host_func_stats) >> 2);
357 dmae->comp_addr_lo = (dmae_reg_go_c[loader_idx] >> 2);
358 dmae->comp_addr_hi = 0;
363 opcode = bxe_dmae_opcode(sc, DMAE_SRC_GRC, DMAE_DST_PCI,
364 TRUE, DMAE_COMP_GRC);
366 /* EMAC is special */
367 if (sc->link_vars.mac_type == ELINK_MAC_TYPE_EMAC) {
368 mac_addr = (port ? GRCBASE_EMAC1 : GRCBASE_EMAC0);
370 /* EMAC_REG_EMAC_RX_STAT_AC (EMAC_REG_EMAC_RX_STAT_AC_COUNT)*/
371 dmae = BXE_SP(sc, dmae[sc->executer_idx++]);
372 dmae->opcode = opcode;
373 dmae->src_addr_lo = (mac_addr + EMAC_REG_EMAC_RX_STAT_AC) >> 2;
374 dmae->src_addr_hi = 0;
375 dmae->dst_addr_lo = U64_LO(BXE_SP_MAPPING(sc, mac_stats));
376 dmae->dst_addr_hi = U64_HI(BXE_SP_MAPPING(sc, mac_stats));
377 dmae->len = EMAC_REG_EMAC_RX_STAT_AC_COUNT;
378 dmae->comp_addr_lo = (dmae_reg_go_c[loader_idx] >> 2);
379 dmae->comp_addr_hi = 0;
382 /* EMAC_REG_EMAC_RX_STAT_AC_28 */
383 dmae = BXE_SP(sc, dmae[sc->executer_idx++]);
384 dmae->opcode = opcode;
385 dmae->src_addr_lo = ((mac_addr + EMAC_REG_EMAC_RX_STAT_AC_28) >> 2);
386 dmae->src_addr_hi = 0;
387 dmae->dst_addr_lo = U64_LO(BXE_SP_MAPPING(sc, mac_stats) +
388 offsetof(struct emac_stats,
389 rx_stat_falsecarriererrors));
390 dmae->dst_addr_hi = U64_HI(BXE_SP_MAPPING(sc, mac_stats) +
391 offsetof(struct emac_stats,
392 rx_stat_falsecarriererrors));
394 dmae->comp_addr_lo = (dmae_reg_go_c[loader_idx] >> 2);
395 dmae->comp_addr_hi = 0;
398 /* EMAC_REG_EMAC_TX_STAT_AC (EMAC_REG_EMAC_TX_STAT_AC_COUNT)*/
399 dmae = BXE_SP(sc, dmae[sc->executer_idx++]);
400 dmae->opcode = opcode;
401 dmae->src_addr_lo = ((mac_addr + EMAC_REG_EMAC_TX_STAT_AC) >> 2);
402 dmae->src_addr_hi = 0;
403 dmae->dst_addr_lo = U64_LO(BXE_SP_MAPPING(sc, mac_stats) +
404 offsetof(struct emac_stats,
405 tx_stat_ifhcoutoctets));
406 dmae->dst_addr_hi = U64_HI(BXE_SP_MAPPING(sc, mac_stats) +
407 offsetof(struct emac_stats,
408 tx_stat_ifhcoutoctets));
409 dmae->len = EMAC_REG_EMAC_TX_STAT_AC_COUNT;
410 dmae->comp_addr_lo = (dmae_reg_go_c[loader_idx] >> 2);
411 dmae->comp_addr_hi = 0;
414 uint32_t tx_src_addr_lo, rx_src_addr_lo;
415 uint16_t rx_len, tx_len;
417 /* configure the params according to MAC type */
418 switch (sc->link_vars.mac_type) {
419 case ELINK_MAC_TYPE_BMAC:
420 mac_addr = (port) ? NIG_REG_INGRESS_BMAC1_MEM :
421 NIG_REG_INGRESS_BMAC0_MEM;
423 /* BIGMAC_REGISTER_TX_STAT_GTPKT ..
424 BIGMAC_REGISTER_TX_STAT_GTBYT */
425 if (CHIP_IS_E1x(sc)) {
427 ((mac_addr + BIGMAC_REGISTER_TX_STAT_GTPKT) >> 2);
428 tx_len = ((8 + BIGMAC_REGISTER_TX_STAT_GTBYT -
429 BIGMAC_REGISTER_TX_STAT_GTPKT) >> 2);
431 ((mac_addr + BIGMAC_REGISTER_RX_STAT_GR64) >> 2);
432 rx_len = ((8 + BIGMAC_REGISTER_RX_STAT_GRIPJ -
433 BIGMAC_REGISTER_RX_STAT_GR64) >> 2);
436 ((mac_addr + BIGMAC2_REGISTER_TX_STAT_GTPOK) >> 2);
437 tx_len = ((8 + BIGMAC2_REGISTER_TX_STAT_GTBYT -
438 BIGMAC2_REGISTER_TX_STAT_GTPOK) >> 2);
440 ((mac_addr + BIGMAC2_REGISTER_RX_STAT_GR64) >> 2);
441 rx_len = ((8 + BIGMAC2_REGISTER_RX_STAT_GRIPJ -
442 BIGMAC2_REGISTER_RX_STAT_GR64) >> 2);
447 case ELINK_MAC_TYPE_UMAC: /* handled by MSTAT */
448 case ELINK_MAC_TYPE_XMAC: /* handled by MSTAT */
450 mac_addr = (port) ? GRCBASE_MSTAT1 : GRCBASE_MSTAT0;
451 tx_src_addr_lo = ((mac_addr + MSTAT_REG_TX_STAT_GTXPOK_LO) >> 2);
452 rx_src_addr_lo = ((mac_addr + MSTAT_REG_RX_STAT_GR64_LO) >> 2);
454 (sizeof(sc->sp->mac_stats.mstat_stats.stats_tx) >> 2);
456 (sizeof(sc->sp->mac_stats.mstat_stats.stats_rx) >> 2);
461 dmae = BXE_SP(sc, dmae[sc->executer_idx++]);
462 dmae->opcode = opcode;
463 dmae->src_addr_lo = tx_src_addr_lo;
464 dmae->src_addr_hi = 0;
466 dmae->dst_addr_lo = U64_LO(BXE_SP_MAPPING(sc, mac_stats));
467 dmae->dst_addr_hi = U64_HI(BXE_SP_MAPPING(sc, mac_stats));
468 dmae->comp_addr_lo = dmae_reg_go_c[loader_idx] >> 2;
469 dmae->comp_addr_hi = 0;
473 dmae = BXE_SP(sc, dmae[sc->executer_idx++]);
474 dmae->opcode = opcode;
475 dmae->src_addr_hi = 0;
476 dmae->src_addr_lo = rx_src_addr_lo;
478 U64_LO(BXE_SP_MAPPING(sc, mac_stats) + (tx_len << 2));
480 U64_HI(BXE_SP_MAPPING(sc, mac_stats) + (tx_len << 2));
482 dmae->comp_addr_lo = dmae_reg_go_c[loader_idx] >> 2;
483 dmae->comp_addr_hi = 0;
488 if (!CHIP_IS_E3(sc)) {
489 dmae = BXE_SP(sc, dmae[sc->executer_idx++]);
490 dmae->opcode = opcode;
492 (port ? NIG_REG_STAT1_EGRESS_MAC_PKT0 :
493 NIG_REG_STAT0_EGRESS_MAC_PKT0) >> 2;
494 dmae->src_addr_hi = 0;
495 dmae->dst_addr_lo = U64_LO(BXE_SP_MAPPING(sc, nig_stats) +
496 offsetof(struct nig_stats,
497 egress_mac_pkt0_lo));
498 dmae->dst_addr_hi = U64_HI(BXE_SP_MAPPING(sc, nig_stats) +
499 offsetof(struct nig_stats,
500 egress_mac_pkt0_lo));
501 dmae->len = ((2 * sizeof(uint32_t)) >> 2);
502 dmae->comp_addr_lo = (dmae_reg_go_c[loader_idx] >> 2);
503 dmae->comp_addr_hi = 0;
506 dmae = BXE_SP(sc, dmae[sc->executer_idx++]);
507 dmae->opcode = opcode;
509 (port ? NIG_REG_STAT1_EGRESS_MAC_PKT1 :
510 NIG_REG_STAT0_EGRESS_MAC_PKT1) >> 2;
511 dmae->src_addr_hi = 0;
512 dmae->dst_addr_lo = U64_LO(BXE_SP_MAPPING(sc, nig_stats) +
513 offsetof(struct nig_stats,
514 egress_mac_pkt1_lo));
515 dmae->dst_addr_hi = U64_HI(BXE_SP_MAPPING(sc, nig_stats) +
516 offsetof(struct nig_stats,
517 egress_mac_pkt1_lo));
518 dmae->len = ((2 * sizeof(uint32_t)) >> 2);
519 dmae->comp_addr_lo = (dmae_reg_go_c[loader_idx] >> 2);
520 dmae->comp_addr_hi = 0;
524 dmae = BXE_SP(sc, dmae[sc->executer_idx++]);
525 dmae->opcode = bxe_dmae_opcode(sc, DMAE_SRC_GRC, DMAE_DST_PCI,
526 TRUE, DMAE_COMP_PCI);
528 (port ? NIG_REG_STAT1_BRB_DISCARD :
529 NIG_REG_STAT0_BRB_DISCARD) >> 2;
530 dmae->src_addr_hi = 0;
531 dmae->dst_addr_lo = U64_LO(BXE_SP_MAPPING(sc, nig_stats));
532 dmae->dst_addr_hi = U64_HI(BXE_SP_MAPPING(sc, nig_stats));
533 dmae->len = (sizeof(struct nig_stats) - 4*sizeof(uint32_t)) >> 2;
535 dmae->comp_addr_lo = U64_LO(BXE_SP_MAPPING(sc, stats_comp));
536 dmae->comp_addr_hi = U64_HI(BXE_SP_MAPPING(sc, stats_comp));
537 dmae->comp_val = DMAE_COMP_VAL;
543 bxe_func_stats_init(struct bxe_softc *sc)
545 struct dmae_cmd *dmae = &sc->stats_dmae;
546 uint32_t *stats_comp = BXE_SP(sc, stats_comp);
554 sc->executer_idx = 0;
555 memset(dmae, 0, sizeof(struct dmae_cmd));
557 dmae->opcode = bxe_dmae_opcode(sc, DMAE_SRC_PCI, DMAE_DST_GRC,
558 TRUE, DMAE_COMP_PCI);
559 dmae->src_addr_lo = U64_LO(BXE_SP_MAPPING(sc, func_stats));
560 dmae->src_addr_hi = U64_HI(BXE_SP_MAPPING(sc, func_stats));
561 dmae->dst_addr_lo = (sc->func_stx >> 2);
562 dmae->dst_addr_hi = 0;
563 dmae->len = (sizeof(struct host_func_stats) >> 2);
564 dmae->comp_addr_lo = U64_LO(BXE_SP_MAPPING(sc, stats_comp));
565 dmae->comp_addr_hi = U64_HI(BXE_SP_MAPPING(sc, stats_comp));
566 dmae->comp_val = DMAE_COMP_VAL;
572 bxe_stats_start(struct bxe_softc *sc)
575 * VFs travel through here as part of the statistics FSM, but no action
583 bxe_port_stats_init(sc);
586 else if (sc->func_stx) {
587 bxe_func_stats_init(sc);
590 bxe_hw_stats_post(sc);
591 bxe_storm_stats_post(sc);
595 bxe_stats_pmf_start(struct bxe_softc *sc)
598 bxe_stats_pmf_update(sc);
603 bxe_stats_restart(struct bxe_softc *sc)
606 * VFs travel through here as part of the statistics FSM, but no action
618 bxe_bmac_stats_update(struct bxe_softc *sc)
620 struct host_port_stats *pstats = BXE_SP(sc, port_stats);
621 struct bxe_eth_stats *estats = &sc->eth_stats;
627 if (CHIP_IS_E1x(sc)) {
628 struct bmac1_stats *new = BXE_SP(sc, mac_stats.bmac1_stats);
630 /* the macros below will use "bmac1_stats" type */
631 UPDATE_STAT64(rx_stat_grerb, rx_stat_ifhcinbadoctets);
632 UPDATE_STAT64(rx_stat_grfcs, rx_stat_dot3statsfcserrors);
633 UPDATE_STAT64(rx_stat_grund, rx_stat_etherstatsundersizepkts);
634 UPDATE_STAT64(rx_stat_grovr, rx_stat_dot3statsframestoolong);
635 UPDATE_STAT64(rx_stat_grfrg, rx_stat_etherstatsfragments);
636 UPDATE_STAT64(rx_stat_grjbr, rx_stat_etherstatsjabbers);
637 UPDATE_STAT64(rx_stat_grxcf, rx_stat_maccontrolframesreceived);
638 UPDATE_STAT64(rx_stat_grxpf, rx_stat_xoffstateentered);
639 UPDATE_STAT64(rx_stat_grxpf, rx_stat_mac_xpf);
641 UPDATE_STAT64(tx_stat_gtxpf, tx_stat_outxoffsent);
642 UPDATE_STAT64(tx_stat_gtxpf, tx_stat_flowcontroldone);
643 UPDATE_STAT64(tx_stat_gt64, tx_stat_etherstatspkts64octets);
644 UPDATE_STAT64(tx_stat_gt127,
645 tx_stat_etherstatspkts65octetsto127octets);
646 UPDATE_STAT64(tx_stat_gt255,
647 tx_stat_etherstatspkts128octetsto255octets);
648 UPDATE_STAT64(tx_stat_gt511,
649 tx_stat_etherstatspkts256octetsto511octets);
650 UPDATE_STAT64(tx_stat_gt1023,
651 tx_stat_etherstatspkts512octetsto1023octets);
652 UPDATE_STAT64(tx_stat_gt1518,
653 tx_stat_etherstatspkts1024octetsto1522octets);
654 UPDATE_STAT64(tx_stat_gt2047, tx_stat_mac_2047);
655 UPDATE_STAT64(tx_stat_gt4095, tx_stat_mac_4095);
656 UPDATE_STAT64(tx_stat_gt9216, tx_stat_mac_9216);
657 UPDATE_STAT64(tx_stat_gt16383, tx_stat_mac_16383);
658 UPDATE_STAT64(tx_stat_gterr,
659 tx_stat_dot3statsinternalmactransmiterrors);
660 UPDATE_STAT64(tx_stat_gtufl, tx_stat_mac_ufl);
662 struct bmac2_stats *new = BXE_SP(sc, mac_stats.bmac2_stats);
663 struct bxe_fw_port_stats_old *fwstats = &sc->fw_stats_old;
665 /* the macros below will use "bmac2_stats" type */
666 UPDATE_STAT64(rx_stat_grerb, rx_stat_ifhcinbadoctets);
667 UPDATE_STAT64(rx_stat_grfcs, rx_stat_dot3statsfcserrors);
668 UPDATE_STAT64(rx_stat_grund, rx_stat_etherstatsundersizepkts);
669 UPDATE_STAT64(rx_stat_grovr, rx_stat_dot3statsframestoolong);
670 UPDATE_STAT64(rx_stat_grfrg, rx_stat_etherstatsfragments);
671 UPDATE_STAT64(rx_stat_grjbr, rx_stat_etherstatsjabbers);
672 UPDATE_STAT64(rx_stat_grxcf, rx_stat_maccontrolframesreceived);
673 UPDATE_STAT64(rx_stat_grxpf, rx_stat_xoffstateentered);
674 UPDATE_STAT64(rx_stat_grxpf, rx_stat_mac_xpf);
675 UPDATE_STAT64(tx_stat_gtxpf, tx_stat_outxoffsent);
676 UPDATE_STAT64(tx_stat_gtxpf, tx_stat_flowcontroldone);
677 UPDATE_STAT64(tx_stat_gt64, tx_stat_etherstatspkts64octets);
678 UPDATE_STAT64(tx_stat_gt127,
679 tx_stat_etherstatspkts65octetsto127octets);
680 UPDATE_STAT64(tx_stat_gt255,
681 tx_stat_etherstatspkts128octetsto255octets);
682 UPDATE_STAT64(tx_stat_gt511,
683 tx_stat_etherstatspkts256octetsto511octets);
684 UPDATE_STAT64(tx_stat_gt1023,
685 tx_stat_etherstatspkts512octetsto1023octets);
686 UPDATE_STAT64(tx_stat_gt1518,
687 tx_stat_etherstatspkts1024octetsto1522octets);
688 UPDATE_STAT64(tx_stat_gt2047, tx_stat_mac_2047);
689 UPDATE_STAT64(tx_stat_gt4095, tx_stat_mac_4095);
690 UPDATE_STAT64(tx_stat_gt9216, tx_stat_mac_9216);
691 UPDATE_STAT64(tx_stat_gt16383, tx_stat_mac_16383);
692 UPDATE_STAT64(tx_stat_gterr,
693 tx_stat_dot3statsinternalmactransmiterrors);
694 UPDATE_STAT64(tx_stat_gtufl, tx_stat_mac_ufl);
696 /* collect PFC stats */
697 pstats->pfc_frames_tx_hi = new->tx_stat_gtpp_hi;
698 pstats->pfc_frames_tx_lo = new->tx_stat_gtpp_lo;
699 ADD_64(pstats->pfc_frames_tx_hi, fwstats->pfc_frames_tx_hi,
700 pstats->pfc_frames_tx_lo, fwstats->pfc_frames_tx_lo);
702 pstats->pfc_frames_rx_hi = new->rx_stat_grpp_hi;
703 pstats->pfc_frames_rx_lo = new->rx_stat_grpp_lo;
704 ADD_64(pstats->pfc_frames_rx_hi, fwstats->pfc_frames_rx_hi,
705 pstats->pfc_frames_rx_lo, fwstats->pfc_frames_rx_lo);
708 estats->pause_frames_received_hi = pstats->mac_stx[1].rx_stat_mac_xpf_hi;
709 estats->pause_frames_received_lo = pstats->mac_stx[1].rx_stat_mac_xpf_lo;
711 estats->pause_frames_sent_hi = pstats->mac_stx[1].tx_stat_outxoffsent_hi;
712 estats->pause_frames_sent_lo = pstats->mac_stx[1].tx_stat_outxoffsent_lo;
714 estats->pfc_frames_received_hi = pstats->pfc_frames_rx_hi;
715 estats->pfc_frames_received_lo = pstats->pfc_frames_rx_lo;
716 estats->pfc_frames_sent_hi = pstats->pfc_frames_tx_hi;
717 estats->pfc_frames_sent_lo = pstats->pfc_frames_tx_lo;
721 bxe_mstat_stats_update(struct bxe_softc *sc)
723 struct host_port_stats *pstats = BXE_SP(sc, port_stats);
724 struct bxe_eth_stats *estats = &sc->eth_stats;
725 struct mstat_stats *new = BXE_SP(sc, mac_stats.mstat_stats);
727 ADD_STAT64(stats_rx.rx_grerb, rx_stat_ifhcinbadoctets);
728 ADD_STAT64(stats_rx.rx_grfcs, rx_stat_dot3statsfcserrors);
729 ADD_STAT64(stats_rx.rx_grund, rx_stat_etherstatsundersizepkts);
730 ADD_STAT64(stats_rx.rx_grovr, rx_stat_dot3statsframestoolong);
731 ADD_STAT64(stats_rx.rx_grfrg, rx_stat_etherstatsfragments);
732 ADD_STAT64(stats_rx.rx_grxcf, rx_stat_maccontrolframesreceived);
733 ADD_STAT64(stats_rx.rx_grxpf, rx_stat_xoffstateentered);
734 ADD_STAT64(stats_rx.rx_grxpf, rx_stat_mac_xpf);
735 ADD_STAT64(stats_tx.tx_gtxpf, tx_stat_outxoffsent);
736 ADD_STAT64(stats_tx.tx_gtxpf, tx_stat_flowcontroldone);
738 /* collect pfc stats */
739 ADD_64(pstats->pfc_frames_tx_hi, new->stats_tx.tx_gtxpp_hi,
740 pstats->pfc_frames_tx_lo, new->stats_tx.tx_gtxpp_lo);
741 ADD_64(pstats->pfc_frames_rx_hi, new->stats_rx.rx_grxpp_hi,
742 pstats->pfc_frames_rx_lo, new->stats_rx.rx_grxpp_lo);
744 ADD_STAT64(stats_tx.tx_gt64, tx_stat_etherstatspkts64octets);
745 ADD_STAT64(stats_tx.tx_gt127, tx_stat_etherstatspkts65octetsto127octets);
746 ADD_STAT64(stats_tx.tx_gt255, tx_stat_etherstatspkts128octetsto255octets);
747 ADD_STAT64(stats_tx.tx_gt511, tx_stat_etherstatspkts256octetsto511octets);
748 ADD_STAT64(stats_tx.tx_gt1023,
749 tx_stat_etherstatspkts512octetsto1023octets);
750 ADD_STAT64(stats_tx.tx_gt1518,
751 tx_stat_etherstatspkts1024octetsto1522octets);
752 ADD_STAT64(stats_tx.tx_gt2047, tx_stat_mac_2047);
754 ADD_STAT64(stats_tx.tx_gt4095, tx_stat_mac_4095);
755 ADD_STAT64(stats_tx.tx_gt9216, tx_stat_mac_9216);
756 ADD_STAT64(stats_tx.tx_gt16383, tx_stat_mac_16383);
758 ADD_STAT64(stats_tx.tx_gterr, tx_stat_dot3statsinternalmactransmiterrors);
759 ADD_STAT64(stats_tx.tx_gtufl, tx_stat_mac_ufl);
761 estats->etherstatspkts1024octetsto1522octets_hi =
762 pstats->mac_stx[1].tx_stat_etherstatspkts1024octetsto1522octets_hi;
763 estats->etherstatspkts1024octetsto1522octets_lo =
764 pstats->mac_stx[1].tx_stat_etherstatspkts1024octetsto1522octets_lo;
766 estats->etherstatspktsover1522octets_hi =
767 pstats->mac_stx[1].tx_stat_mac_2047_hi;
768 estats->etherstatspktsover1522octets_lo =
769 pstats->mac_stx[1].tx_stat_mac_2047_lo;
771 ADD_64(estats->etherstatspktsover1522octets_hi,
772 pstats->mac_stx[1].tx_stat_mac_4095_hi,
773 estats->etherstatspktsover1522octets_lo,
774 pstats->mac_stx[1].tx_stat_mac_4095_lo);
776 ADD_64(estats->etherstatspktsover1522octets_hi,
777 pstats->mac_stx[1].tx_stat_mac_9216_hi,
778 estats->etherstatspktsover1522octets_lo,
779 pstats->mac_stx[1].tx_stat_mac_9216_lo);
781 ADD_64(estats->etherstatspktsover1522octets_hi,
782 pstats->mac_stx[1].tx_stat_mac_16383_hi,
783 estats->etherstatspktsover1522octets_lo,
784 pstats->mac_stx[1].tx_stat_mac_16383_lo);
786 estats->pause_frames_received_hi = pstats->mac_stx[1].rx_stat_mac_xpf_hi;
787 estats->pause_frames_received_lo = pstats->mac_stx[1].rx_stat_mac_xpf_lo;
789 estats->pause_frames_sent_hi = pstats->mac_stx[1].tx_stat_outxoffsent_hi;
790 estats->pause_frames_sent_lo = pstats->mac_stx[1].tx_stat_outxoffsent_lo;
792 estats->pfc_frames_received_hi = pstats->pfc_frames_rx_hi;
793 estats->pfc_frames_received_lo = pstats->pfc_frames_rx_lo;
794 estats->pfc_frames_sent_hi = pstats->pfc_frames_tx_hi;
795 estats->pfc_frames_sent_lo = pstats->pfc_frames_tx_lo;
799 bxe_emac_stats_update(struct bxe_softc *sc)
801 struct emac_stats *new = BXE_SP(sc, mac_stats.emac_stats);
802 struct host_port_stats *pstats = BXE_SP(sc, port_stats);
803 struct bxe_eth_stats *estats = &sc->eth_stats;
805 UPDATE_EXTEND_STAT(rx_stat_ifhcinbadoctets);
806 UPDATE_EXTEND_STAT(tx_stat_ifhcoutbadoctets);
807 UPDATE_EXTEND_STAT(rx_stat_dot3statsfcserrors);
808 UPDATE_EXTEND_STAT(rx_stat_dot3statsalignmenterrors);
809 UPDATE_EXTEND_STAT(rx_stat_dot3statscarriersenseerrors);
810 UPDATE_EXTEND_STAT(rx_stat_falsecarriererrors);
811 UPDATE_EXTEND_STAT(rx_stat_etherstatsundersizepkts);
812 UPDATE_EXTEND_STAT(rx_stat_dot3statsframestoolong);
813 UPDATE_EXTEND_STAT(rx_stat_etherstatsfragments);
814 UPDATE_EXTEND_STAT(rx_stat_etherstatsjabbers);
815 UPDATE_EXTEND_STAT(rx_stat_maccontrolframesreceived);
816 UPDATE_EXTEND_STAT(rx_stat_xoffstateentered);
817 UPDATE_EXTEND_STAT(rx_stat_xonpauseframesreceived);
818 UPDATE_EXTEND_STAT(rx_stat_xoffpauseframesreceived);
819 UPDATE_EXTEND_STAT(tx_stat_outxonsent);
820 UPDATE_EXTEND_STAT(tx_stat_outxoffsent);
821 UPDATE_EXTEND_STAT(tx_stat_flowcontroldone);
822 UPDATE_EXTEND_STAT(tx_stat_etherstatscollisions);
823 UPDATE_EXTEND_STAT(tx_stat_dot3statssinglecollisionframes);
824 UPDATE_EXTEND_STAT(tx_stat_dot3statsmultiplecollisionframes);
825 UPDATE_EXTEND_STAT(tx_stat_dot3statsdeferredtransmissions);
826 UPDATE_EXTEND_STAT(tx_stat_dot3statsexcessivecollisions);
827 UPDATE_EXTEND_STAT(tx_stat_dot3statslatecollisions);
828 UPDATE_EXTEND_STAT(tx_stat_etherstatspkts64octets);
829 UPDATE_EXTEND_STAT(tx_stat_etherstatspkts65octetsto127octets);
830 UPDATE_EXTEND_STAT(tx_stat_etherstatspkts128octetsto255octets);
831 UPDATE_EXTEND_STAT(tx_stat_etherstatspkts256octetsto511octets);
832 UPDATE_EXTEND_STAT(tx_stat_etherstatspkts512octetsto1023octets);
833 UPDATE_EXTEND_STAT(tx_stat_etherstatspkts1024octetsto1522octets);
834 UPDATE_EXTEND_STAT(tx_stat_etherstatspktsover1522octets);
835 UPDATE_EXTEND_STAT(tx_stat_dot3statsinternalmactransmiterrors);
837 estats->pause_frames_received_hi =
838 pstats->mac_stx[1].rx_stat_xonpauseframesreceived_hi;
839 estats->pause_frames_received_lo =
840 pstats->mac_stx[1].rx_stat_xonpauseframesreceived_lo;
841 ADD_64(estats->pause_frames_received_hi,
842 pstats->mac_stx[1].rx_stat_xoffpauseframesreceived_hi,
843 estats->pause_frames_received_lo,
844 pstats->mac_stx[1].rx_stat_xoffpauseframesreceived_lo);
846 estats->pause_frames_sent_hi =
847 pstats->mac_stx[1].tx_stat_outxonsent_hi;
848 estats->pause_frames_sent_lo =
849 pstats->mac_stx[1].tx_stat_outxonsent_lo;
850 ADD_64(estats->pause_frames_sent_hi,
851 pstats->mac_stx[1].tx_stat_outxoffsent_hi,
852 estats->pause_frames_sent_lo,
853 pstats->mac_stx[1].tx_stat_outxoffsent_lo);
857 bxe_hw_stats_update(struct bxe_softc *sc)
859 struct nig_stats *new = BXE_SP(sc, nig_stats);
860 struct nig_stats *old = &(sc->port.old_nig_stats);
861 struct host_port_stats *pstats = BXE_SP(sc, port_stats);
862 struct bxe_eth_stats *estats = &sc->eth_stats;
863 uint32_t lpi_reg, nig_timer_max;
869 switch (sc->link_vars.mac_type) {
870 case ELINK_MAC_TYPE_BMAC:
871 bxe_bmac_stats_update(sc);
874 case ELINK_MAC_TYPE_EMAC:
875 bxe_emac_stats_update(sc);
878 case ELINK_MAC_TYPE_UMAC:
879 case ELINK_MAC_TYPE_XMAC:
880 bxe_mstat_stats_update(sc);
883 case ELINK_MAC_TYPE_NONE: /* unreached */
885 "stats updated by DMAE but no MAC active\n");
888 default: /* unreached */
889 BLOGE(sc, "stats update failed, unknown MAC type\n");
892 ADD_EXTEND_64(pstats->brb_drop_hi, pstats->brb_drop_lo,
893 new->brb_discard - old->brb_discard);
894 ADD_EXTEND_64(estats->brb_truncate_hi, estats->brb_truncate_lo,
895 new->brb_truncate - old->brb_truncate);
897 if (!CHIP_IS_E3(sc)) {
898 UPDATE_STAT64_NIG(egress_mac_pkt0,
899 etherstatspkts1024octetsto1522octets);
900 UPDATE_STAT64_NIG(egress_mac_pkt1,
901 etherstatspktsover1522octets);
904 memcpy(old, new, sizeof(struct nig_stats));
906 memcpy(&(estats->rx_stat_ifhcinbadoctets_hi), &(pstats->mac_stx[1]),
907 sizeof(struct mac_stx));
908 estats->brb_drop_hi = pstats->brb_drop_hi;
909 estats->brb_drop_lo = pstats->brb_drop_lo;
911 pstats->host_port_stats_counter++;
913 if (CHIP_IS_E3(sc)) {
914 lpi_reg = (SC_PORT(sc)) ?
915 MISC_REG_CPMU_LP_SM_ENT_CNT_P1 :
916 MISC_REG_CPMU_LP_SM_ENT_CNT_P0;
917 estats->eee_tx_lpi += REG_RD(sc, lpi_reg);
920 if (!BXE_NOMCP(sc)) {
921 nig_timer_max = SHMEM_RD(sc, port_mb[SC_PORT(sc)].stat_nig_timer);
922 if (nig_timer_max != estats->nig_timer_max) {
923 estats->nig_timer_max = nig_timer_max;
924 BLOGE(sc, "invalid NIG timer max (%u)\n",
925 estats->nig_timer_max);
933 bxe_storm_stats_validate_counters(struct bxe_softc *sc)
935 struct stats_counter *counters = &sc->fw_stats_data->storm_counters;
936 uint16_t cur_stats_counter;
939 * Make sure we use the value of the counter
940 * used for sending the last stats ramrod.
943 cur_stats_counter = (sc->stats_counter - 1);
944 BXE_STATS_UNLOCK(sc);
946 /* are storm stats valid? */
947 if (le16toh(counters->xstats_counter) != cur_stats_counter) {
949 "stats not updated by xstorm, "
950 "counter 0x%x != stats_counter 0x%x\n",
951 le16toh(counters->xstats_counter), sc->stats_counter);
955 if (le16toh(counters->ustats_counter) != cur_stats_counter) {
957 "stats not updated by ustorm, "
958 "counter 0x%x != stats_counter 0x%x\n",
959 le16toh(counters->ustats_counter), sc->stats_counter);
963 if (le16toh(counters->cstats_counter) != cur_stats_counter) {
965 "stats not updated by cstorm, "
966 "counter 0x%x != stats_counter 0x%x\n",
967 le16toh(counters->cstats_counter), sc->stats_counter);
971 if (le16toh(counters->tstats_counter) != cur_stats_counter) {
973 "stats not updated by tstorm, "
974 "counter 0x%x != stats_counter 0x%x\n",
975 le16toh(counters->tstats_counter), sc->stats_counter);
983 bxe_storm_stats_update(struct bxe_softc *sc)
985 struct tstorm_per_port_stats *tport =
986 &sc->fw_stats_data->port.tstorm_port_statistics;
987 struct tstorm_per_pf_stats *tfunc =
988 &sc->fw_stats_data->pf.tstorm_pf_statistics;
989 struct host_func_stats *fstats = &sc->func_stats;
990 struct bxe_eth_stats *estats = &sc->eth_stats;
991 struct bxe_eth_stats_old *estats_old = &sc->eth_stats_old;
994 /* vfs stat counter is managed by pf */
995 if (IS_PF(sc) && bxe_storm_stats_validate_counters(sc)) {
999 estats->error_bytes_received_hi = 0;
1000 estats->error_bytes_received_lo = 0;
1002 for (i = 0; i < sc->num_queues; i++) {
1003 struct bxe_fastpath *fp = &sc->fp[i];
1004 struct tstorm_per_queue_stats *tclient =
1005 &sc->fw_stats_data->queue_stats[i].tstorm_queue_statistics;
1006 struct tstorm_per_queue_stats *old_tclient = &fp->old_tclient;
1007 struct ustorm_per_queue_stats *uclient =
1008 &sc->fw_stats_data->queue_stats[i].ustorm_queue_statistics;
1009 struct ustorm_per_queue_stats *old_uclient = &fp->old_uclient;
1010 struct xstorm_per_queue_stats *xclient =
1011 &sc->fw_stats_data->queue_stats[i].xstorm_queue_statistics;
1012 struct xstorm_per_queue_stats *old_xclient = &fp->old_xclient;
1013 struct bxe_eth_q_stats *qstats = &fp->eth_q_stats;
1014 struct bxe_eth_q_stats_old *qstats_old = &fp->eth_q_stats_old;
1018 BLOGD(sc, DBG_STATS,
1019 "queue[%d]: ucast_sent 0x%x bcast_sent 0x%x mcast_sent 0x%x\n",
1020 i, xclient->ucast_pkts_sent, xclient->bcast_pkts_sent,
1021 xclient->mcast_pkts_sent);
1023 BLOGD(sc, DBG_STATS, "---------------\n");
1025 UPDATE_QSTAT(tclient->rcv_bcast_bytes,
1026 total_broadcast_bytes_received);
1027 UPDATE_QSTAT(tclient->rcv_mcast_bytes,
1028 total_multicast_bytes_received);
1029 UPDATE_QSTAT(tclient->rcv_ucast_bytes,
1030 total_unicast_bytes_received);
1033 * sum to total_bytes_received all
1034 * unicast/multicast/broadcast
1036 qstats->total_bytes_received_hi =
1037 qstats->total_broadcast_bytes_received_hi;
1038 qstats->total_bytes_received_lo =
1039 qstats->total_broadcast_bytes_received_lo;
1041 ADD_64(qstats->total_bytes_received_hi,
1042 qstats->total_multicast_bytes_received_hi,
1043 qstats->total_bytes_received_lo,
1044 qstats->total_multicast_bytes_received_lo);
1046 ADD_64(qstats->total_bytes_received_hi,
1047 qstats->total_unicast_bytes_received_hi,
1048 qstats->total_bytes_received_lo,
1049 qstats->total_unicast_bytes_received_lo);
1051 qstats->valid_bytes_received_hi = qstats->total_bytes_received_hi;
1052 qstats->valid_bytes_received_lo = qstats->total_bytes_received_lo;
1054 UPDATE_EXTEND_TSTAT(rcv_ucast_pkts, total_unicast_packets_received);
1055 UPDATE_EXTEND_TSTAT(rcv_mcast_pkts, total_multicast_packets_received);
1056 UPDATE_EXTEND_TSTAT(rcv_bcast_pkts, total_broadcast_packets_received);
1057 UPDATE_EXTEND_E_TSTAT(pkts_too_big_discard,
1058 etherstatsoverrsizepkts, 32);
1059 UPDATE_EXTEND_E_TSTAT(no_buff_discard, no_buff_discard, 16);
1061 SUB_EXTEND_USTAT(ucast_no_buff_pkts, total_unicast_packets_received);
1062 SUB_EXTEND_USTAT(mcast_no_buff_pkts,
1063 total_multicast_packets_received);
1064 SUB_EXTEND_USTAT(bcast_no_buff_pkts,
1065 total_broadcast_packets_received);
1066 UPDATE_EXTEND_E_USTAT(ucast_no_buff_pkts, no_buff_discard);
1067 UPDATE_EXTEND_E_USTAT(mcast_no_buff_pkts, no_buff_discard);
1068 UPDATE_EXTEND_E_USTAT(bcast_no_buff_pkts, no_buff_discard);
1070 UPDATE_QSTAT(xclient->bcast_bytes_sent,
1071 total_broadcast_bytes_transmitted);
1072 UPDATE_QSTAT(xclient->mcast_bytes_sent,
1073 total_multicast_bytes_transmitted);
1074 UPDATE_QSTAT(xclient->ucast_bytes_sent,
1075 total_unicast_bytes_transmitted);
1078 * sum to total_bytes_transmitted all
1079 * unicast/multicast/broadcast
1081 qstats->total_bytes_transmitted_hi =
1082 qstats->total_unicast_bytes_transmitted_hi;
1083 qstats->total_bytes_transmitted_lo =
1084 qstats->total_unicast_bytes_transmitted_lo;
1086 ADD_64(qstats->total_bytes_transmitted_hi,
1087 qstats->total_broadcast_bytes_transmitted_hi,
1088 qstats->total_bytes_transmitted_lo,
1089 qstats->total_broadcast_bytes_transmitted_lo);
1091 ADD_64(qstats->total_bytes_transmitted_hi,
1092 qstats->total_multicast_bytes_transmitted_hi,
1093 qstats->total_bytes_transmitted_lo,
1094 qstats->total_multicast_bytes_transmitted_lo);
1096 UPDATE_EXTEND_XSTAT(ucast_pkts_sent,
1097 total_unicast_packets_transmitted);
1098 UPDATE_EXTEND_XSTAT(mcast_pkts_sent,
1099 total_multicast_packets_transmitted);
1100 UPDATE_EXTEND_XSTAT(bcast_pkts_sent,
1101 total_broadcast_packets_transmitted);
1103 UPDATE_EXTEND_TSTAT(checksum_discard,
1104 total_packets_received_checksum_discarded);
1105 UPDATE_EXTEND_TSTAT(ttl0_discard,
1106 total_packets_received_ttl0_discarded);
1108 UPDATE_EXTEND_XSTAT(error_drop_pkts,
1109 total_transmitted_dropped_packets_error);
1111 /* TPA aggregations completed */
1112 UPDATE_EXTEND_E_USTAT(coalesced_events, total_tpa_aggregations);
1113 /* Number of network frames aggregated by TPA */
1114 UPDATE_EXTEND_E_USTAT(coalesced_pkts, total_tpa_aggregated_frames);
1115 /* Total number of bytes in completed TPA aggregations */
1116 UPDATE_QSTAT(uclient->coalesced_bytes, total_tpa_bytes);
1118 UPDATE_ESTAT_QSTAT_64(total_tpa_bytes);
1120 UPDATE_FSTAT_QSTAT(total_bytes_received);
1121 UPDATE_FSTAT_QSTAT(total_bytes_transmitted);
1122 UPDATE_FSTAT_QSTAT(total_unicast_packets_received);
1123 UPDATE_FSTAT_QSTAT(total_multicast_packets_received);
1124 UPDATE_FSTAT_QSTAT(total_broadcast_packets_received);
1125 UPDATE_FSTAT_QSTAT(total_unicast_packets_transmitted);
1126 UPDATE_FSTAT_QSTAT(total_multicast_packets_transmitted);
1127 UPDATE_FSTAT_QSTAT(total_broadcast_packets_transmitted);
1128 UPDATE_FSTAT_QSTAT(valid_bytes_received);
1131 ADD_64(estats->total_bytes_received_hi,
1132 estats->rx_stat_ifhcinbadoctets_hi,
1133 estats->total_bytes_received_lo,
1134 estats->rx_stat_ifhcinbadoctets_lo);
1136 ADD_64_LE(estats->total_bytes_received_hi,
1137 tfunc->rcv_error_bytes.hi,
1138 estats->total_bytes_received_lo,
1139 tfunc->rcv_error_bytes.lo);
1141 ADD_64_LE(estats->error_bytes_received_hi,
1142 tfunc->rcv_error_bytes.hi,
1143 estats->error_bytes_received_lo,
1144 tfunc->rcv_error_bytes.lo);
1146 UPDATE_ESTAT(etherstatsoverrsizepkts, rx_stat_dot3statsframestoolong);
1148 ADD_64(estats->error_bytes_received_hi,
1149 estats->rx_stat_ifhcinbadoctets_hi,
1150 estats->error_bytes_received_lo,
1151 estats->rx_stat_ifhcinbadoctets_lo);
1154 struct bxe_fw_port_stats_old *fwstats = &sc->fw_stats_old;
1155 UPDATE_FW_STAT(mac_filter_discard);
1156 UPDATE_FW_STAT(mf_tag_discard);
1157 UPDATE_FW_STAT(brb_truncate_discard);
1158 UPDATE_FW_STAT(mac_discard);
1161 fstats->host_func_stats_start = ++fstats->host_func_stats_end;
1163 sc->stats_pending = 0;
1169 bxe_net_stats_update(struct bxe_softc *sc)
1171 struct bxe_eth_stats *estats = &sc->eth_stats;
1172 struct ifnet *ifnet = sc->ifnet;
1176 ifnet->if_data.ifi_ipackets =
1177 bxe_hilo(&estats->total_unicast_packets_received_hi) +
1178 bxe_hilo(&estats->total_multicast_packets_received_hi) +
1179 bxe_hilo(&estats->total_broadcast_packets_received_hi);
1181 ifnet->if_data.ifi_opackets =
1182 bxe_hilo(&estats->total_unicast_packets_transmitted_hi) +
1183 bxe_hilo(&estats->total_multicast_packets_transmitted_hi) +
1184 bxe_hilo(&estats->total_broadcast_packets_transmitted_hi);
1186 ifnet->if_data.ifi_ibytes = bxe_hilo(&estats->total_bytes_received_hi);
1188 ifnet->if_data.ifi_obytes = bxe_hilo(&estats->total_bytes_transmitted_hi);
1191 for (i = 0; i < sc->num_queues; i++) {
1192 struct tstorm_per_queue_stats *old_tclient =
1193 &sc->fp[i].old_tclient;
1194 tmp += le32toh(old_tclient->checksum_discard);
1197 ifnet->if_data.ifi_iqdrops = tmp;
1199 ifnet->if_data.ifi_ierrors =
1200 bxe_hilo(&estats->rx_stat_etherstatsundersizepkts_hi) +
1201 bxe_hilo(&estats->etherstatsoverrsizepkts_hi) +
1202 bxe_hilo(&estats->brb_drop_hi) +
1203 bxe_hilo(&estats->brb_truncate_hi) +
1204 bxe_hilo(&estats->rx_stat_dot3statsfcserrors_hi) +
1205 bxe_hilo(&estats->rx_stat_dot3statsalignmenterrors_hi) +
1206 bxe_hilo(&estats->no_buff_discard_hi);
1208 ifnet->if_data.ifi_oerrors =
1209 bxe_hilo(&estats->rx_stat_dot3statscarriersenseerrors_hi) +
1210 bxe_hilo(&estats->tx_stat_dot3statsinternalmactransmiterrors_hi);
1212 ifnet->if_data.ifi_imcasts =
1213 bxe_hilo(&estats->total_multicast_packets_received_hi);
1215 ifnet->if_data.ifi_collisions =
1216 bxe_hilo(&estats->tx_stat_etherstatscollisions_hi) +
1217 bxe_hilo(&estats->tx_stat_dot3statslatecollisions_hi) +
1218 bxe_hilo(&estats->tx_stat_dot3statsexcessivecollisions_hi);
1222 bxe_drv_stats_update(struct bxe_softc *sc)
1224 struct bxe_eth_stats *estats = &sc->eth_stats;
1227 for (i = 0; i < sc->num_queues; i++) {
1228 struct bxe_eth_q_stats *qstats = &sc->fp[i].eth_q_stats;
1229 struct bxe_eth_q_stats_old *qstats_old = &sc->fp[i].eth_q_stats_old;
1231 UPDATE_ESTAT_QSTAT(rx_calls);
1232 UPDATE_ESTAT_QSTAT(rx_pkts);
1233 UPDATE_ESTAT_QSTAT(rx_tpa_pkts);
1234 UPDATE_ESTAT_QSTAT(rx_erroneous_jumbo_sge_pkts);
1235 UPDATE_ESTAT_QSTAT(rx_bxe_service_rxsgl);
1236 UPDATE_ESTAT_QSTAT(rx_jumbo_sge_pkts);
1237 UPDATE_ESTAT_QSTAT(rx_soft_errors);
1238 UPDATE_ESTAT_QSTAT(rx_hw_csum_errors);
1239 UPDATE_ESTAT_QSTAT(rx_ofld_frames_csum_ip);
1240 UPDATE_ESTAT_QSTAT(rx_ofld_frames_csum_tcp_udp);
1241 UPDATE_ESTAT_QSTAT(rx_budget_reached);
1242 UPDATE_ESTAT_QSTAT(tx_pkts);
1243 UPDATE_ESTAT_QSTAT(tx_soft_errors);
1244 UPDATE_ESTAT_QSTAT(tx_ofld_frames_csum_ip);
1245 UPDATE_ESTAT_QSTAT(tx_ofld_frames_csum_tcp);
1246 UPDATE_ESTAT_QSTAT(tx_ofld_frames_csum_udp);
1247 UPDATE_ESTAT_QSTAT(tx_ofld_frames_lso);
1248 UPDATE_ESTAT_QSTAT(tx_ofld_frames_lso_hdr_splits);
1249 UPDATE_ESTAT_QSTAT(tx_encap_failures);
1250 UPDATE_ESTAT_QSTAT(tx_hw_queue_full);
1251 UPDATE_ESTAT_QSTAT(tx_hw_max_queue_depth);
1252 UPDATE_ESTAT_QSTAT(tx_dma_mapping_failure);
1253 UPDATE_ESTAT_QSTAT(tx_max_drbr_queue_depth);
1254 UPDATE_ESTAT_QSTAT(tx_window_violation_std);
1255 UPDATE_ESTAT_QSTAT(tx_window_violation_tso);
1256 //UPDATE_ESTAT_QSTAT(tx_unsupported_tso_request_ipv6);
1257 //UPDATE_ESTAT_QSTAT(tx_unsupported_tso_request_not_tcp);
1258 UPDATE_ESTAT_QSTAT(tx_chain_lost_mbuf);
1259 UPDATE_ESTAT_QSTAT(tx_frames_deferred);
1260 UPDATE_ESTAT_QSTAT(tx_queue_xoff);
1262 /* mbuf driver statistics */
1263 UPDATE_ESTAT_QSTAT(mbuf_defrag_attempts);
1264 UPDATE_ESTAT_QSTAT(mbuf_defrag_failures);
1265 UPDATE_ESTAT_QSTAT(mbuf_rx_bd_alloc_failed);
1266 UPDATE_ESTAT_QSTAT(mbuf_rx_bd_mapping_failed);
1267 UPDATE_ESTAT_QSTAT(mbuf_rx_tpa_alloc_failed);
1268 UPDATE_ESTAT_QSTAT(mbuf_rx_tpa_mapping_failed);
1269 UPDATE_ESTAT_QSTAT(mbuf_rx_sge_alloc_failed);
1270 UPDATE_ESTAT_QSTAT(mbuf_rx_sge_mapping_failed);
1272 /* track the number of allocated mbufs */
1273 UPDATE_ESTAT_QSTAT(mbuf_alloc_tx);
1274 UPDATE_ESTAT_QSTAT(mbuf_alloc_rx);
1275 UPDATE_ESTAT_QSTAT(mbuf_alloc_sge);
1276 UPDATE_ESTAT_QSTAT(mbuf_alloc_tpa);
1281 bxe_edebug_stats_stopped(struct bxe_softc *sc)
1285 if (SHMEM2_HAS(sc, edebug_driver_if[1])) {
1286 val = SHMEM2_RD(sc, edebug_driver_if[1]);
1288 if (val == EDEBUG_DRIVER_IF_OP_CODE_DISABLE_STAT) {
1297 bxe_stats_update(struct bxe_softc *sc)
1299 uint32_t *stats_comp = BXE_SP(sc, stats_comp);
1301 if (bxe_edebug_stats_stopped(sc)) {
1306 if (*stats_comp != DMAE_COMP_VAL) {
1311 bxe_hw_stats_update(sc);
1314 if (bxe_storm_stats_update(sc)) {
1315 if (sc->stats_pending++ == 3) {
1316 if (sc->ifnet->if_drv_flags & IFF_DRV_RUNNING) {
1317 if(sc->trigger_grcdump) {
1318 /* taking grcdump */
1321 atomic_store_rel_long(&sc->chip_tq_flags, CHIP_TQ_REINIT);
1322 taskqueue_enqueue(sc->chip_tq, &sc->chip_tq_task);
1329 * VF doesn't collect HW statistics, and doesn't get completions,
1330 * performs only update.
1332 bxe_storm_stats_update(sc);
1335 bxe_net_stats_update(sc);
1336 bxe_drv_stats_update(sc);
1343 bxe_hw_stats_post(sc);
1344 bxe_storm_stats_post(sc);
1348 bxe_port_stats_stop(struct bxe_softc *sc)
1350 struct dmae_cmd *dmae;
1352 int loader_idx = PMF_DMAE_C(sc);
1353 uint32_t *stats_comp = BXE_SP(sc, stats_comp);
1355 sc->executer_idx = 0;
1357 opcode = bxe_dmae_opcode(sc, DMAE_SRC_PCI, DMAE_DST_GRC, FALSE, 0);
1359 if (sc->port.port_stx) {
1360 dmae = BXE_SP(sc, dmae[sc->executer_idx++]);
1363 dmae->opcode = bxe_dmae_opcode_add_comp(opcode, DMAE_COMP_GRC);
1365 dmae->opcode = bxe_dmae_opcode_add_comp(opcode, DMAE_COMP_PCI);
1368 dmae->src_addr_lo = U64_LO(BXE_SP_MAPPING(sc, port_stats));
1369 dmae->src_addr_hi = U64_HI(BXE_SP_MAPPING(sc, port_stats));
1370 dmae->dst_addr_lo = sc->port.port_stx >> 2;
1371 dmae->dst_addr_hi = 0;
1372 dmae->len = bxe_get_port_stats_dma_len(sc);
1374 dmae->comp_addr_lo = (dmae_reg_go_c[loader_idx] >> 2);
1375 dmae->comp_addr_hi = 0;
1378 dmae->comp_addr_lo = U64_LO(BXE_SP_MAPPING(sc, stats_comp));
1379 dmae->comp_addr_hi = U64_HI(BXE_SP_MAPPING(sc, stats_comp));
1380 dmae->comp_val = DMAE_COMP_VAL;
1387 dmae = BXE_SP(sc, dmae[sc->executer_idx++]);
1388 dmae->opcode = bxe_dmae_opcode_add_comp(opcode, DMAE_COMP_PCI);
1389 dmae->src_addr_lo = U64_LO(BXE_SP_MAPPING(sc, func_stats));
1390 dmae->src_addr_hi = U64_HI(BXE_SP_MAPPING(sc, func_stats));
1391 dmae->dst_addr_lo = (sc->func_stx >> 2);
1392 dmae->dst_addr_hi = 0;
1393 dmae->len = (sizeof(struct host_func_stats) >> 2);
1394 dmae->comp_addr_lo = U64_LO(BXE_SP_MAPPING(sc, stats_comp));
1395 dmae->comp_addr_hi = U64_HI(BXE_SP_MAPPING(sc, stats_comp));
1396 dmae->comp_val = DMAE_COMP_VAL;
1403 bxe_stats_stop(struct bxe_softc *sc)
1405 uint8_t update = FALSE;
1410 update = bxe_hw_stats_update(sc) == 0;
1413 update |= bxe_storm_stats_update(sc) == 0;
1416 bxe_net_stats_update(sc);
1419 bxe_port_stats_stop(sc);
1422 bxe_hw_stats_post(sc);
1428 bxe_stats_do_nothing(struct bxe_softc *sc)
1433 static const struct {
1434 void (*action)(struct bxe_softc *sc);
1435 enum bxe_stats_state next_state;
1436 } bxe_stats_stm[STATS_STATE_MAX][STATS_EVENT_MAX] = {
1438 /* DISABLED PMF */ { bxe_stats_pmf_update, STATS_STATE_DISABLED },
1439 /* LINK_UP */ { bxe_stats_start, STATS_STATE_ENABLED },
1440 /* UPDATE */ { bxe_stats_do_nothing, STATS_STATE_DISABLED },
1441 /* STOP */ { bxe_stats_do_nothing, STATS_STATE_DISABLED }
1444 /* ENABLED PMF */ { bxe_stats_pmf_start, STATS_STATE_ENABLED },
1445 /* LINK_UP */ { bxe_stats_restart, STATS_STATE_ENABLED },
1446 /* UPDATE */ { bxe_stats_update, STATS_STATE_ENABLED },
1447 /* STOP */ { bxe_stats_stop, STATS_STATE_DISABLED }
1451 void bxe_stats_handle(struct bxe_softc *sc,
1452 enum bxe_stats_event event)
1454 enum bxe_stats_state state;
1456 if (__predict_false(sc->panic)) {
1461 state = sc->stats_state;
1462 sc->stats_state = bxe_stats_stm[state][event].next_state;
1463 BXE_STATS_UNLOCK(sc);
1465 bxe_stats_stm[state][event].action(sc);
1467 if (event != STATS_EVENT_UPDATE) {
1468 BLOGD(sc, DBG_STATS,
1469 "state %d -> event %d -> state %d\n",
1470 state, event, sc->stats_state);
1475 bxe_port_stats_base_init(struct bxe_softc *sc)
1477 struct dmae_cmd *dmae;
1478 uint32_t *stats_comp = BXE_SP(sc, stats_comp);
1481 if (!sc->port.pmf || !sc->port.port_stx) {
1482 BLOGE(sc, "BUG!\n");
1486 sc->executer_idx = 0;
1488 dmae = BXE_SP(sc, dmae[sc->executer_idx++]);
1489 dmae->opcode = bxe_dmae_opcode(sc, DMAE_SRC_PCI, DMAE_DST_GRC,
1490 TRUE, DMAE_COMP_PCI);
1491 dmae->src_addr_lo = U64_LO(BXE_SP_MAPPING(sc, port_stats));
1492 dmae->src_addr_hi = U64_HI(BXE_SP_MAPPING(sc, port_stats));
1493 dmae->dst_addr_lo = (sc->port.port_stx >> 2);
1494 dmae->dst_addr_hi = 0;
1495 dmae->len = bxe_get_port_stats_dma_len(sc);
1496 dmae->comp_addr_lo = U64_LO(BXE_SP_MAPPING(sc, stats_comp));
1497 dmae->comp_addr_hi = U64_HI(BXE_SP_MAPPING(sc, stats_comp));
1498 dmae->comp_val = DMAE_COMP_VAL;
1501 bxe_hw_stats_post(sc);
1506 * This function will prepare the statistics ramrod data the way
1507 * we will only have to increment the statistics counter and
1508 * send the ramrod each time we have to.
1511 bxe_prep_fw_stats_req(struct bxe_softc *sc)
1514 int first_queue_query_index;
1515 struct stats_query_header *stats_hdr = &sc->fw_stats_req->hdr;
1516 bus_addr_t cur_data_offset;
1517 struct stats_query_entry *cur_query_entry;
1519 stats_hdr->cmd_num = sc->fw_stats_num;
1520 stats_hdr->drv_stats_counter = 0;
1523 * The storm_counters struct contains the counters of completed
1524 * statistics requests per storm which are incremented by FW
1525 * each time it completes hadning a statistics ramrod. We will
1526 * check these counters in the timer handler and discard a
1527 * (statistics) ramrod completion.
1529 cur_data_offset = (sc->fw_stats_data_mapping +
1530 offsetof(struct bxe_fw_stats_data, storm_counters));
1532 stats_hdr->stats_counters_addrs.hi = htole32(U64_HI(cur_data_offset));
1533 stats_hdr->stats_counters_addrs.lo = htole32(U64_LO(cur_data_offset));
1536 * Prepare the first stats ramrod (will be completed with
1537 * the counters equal to zero) - init counters to somethig different.
1539 memset(&sc->fw_stats_data->storm_counters, 0xff,
1540 sizeof(struct stats_counter));
1542 /**** Port FW statistics data ****/
1543 cur_data_offset = (sc->fw_stats_data_mapping +
1544 offsetof(struct bxe_fw_stats_data, port));
1546 cur_query_entry = &sc->fw_stats_req->query[BXE_PORT_QUERY_IDX];
1548 cur_query_entry->kind = STATS_TYPE_PORT;
1549 /* For port query index is a DONT CARE */
1550 cur_query_entry->index = SC_PORT(sc);
1551 /* For port query funcID is a DONT CARE */
1552 cur_query_entry->funcID = htole16(SC_FUNC(sc));
1553 cur_query_entry->address.hi = htole32(U64_HI(cur_data_offset));
1554 cur_query_entry->address.lo = htole32(U64_LO(cur_data_offset));
1556 /**** PF FW statistics data ****/
1557 cur_data_offset = (sc->fw_stats_data_mapping +
1558 offsetof(struct bxe_fw_stats_data, pf));
1560 cur_query_entry = &sc->fw_stats_req->query[BXE_PF_QUERY_IDX];
1562 cur_query_entry->kind = STATS_TYPE_PF;
1563 /* For PF query index is a DONT CARE */
1564 cur_query_entry->index = SC_PORT(sc);
1565 cur_query_entry->funcID = htole16(SC_FUNC(sc));
1566 cur_query_entry->address.hi = htole32(U64_HI(cur_data_offset));
1567 cur_query_entry->address.lo = htole32(U64_LO(cur_data_offset));
1569 /**** Clients' queries ****/
1570 cur_data_offset = (sc->fw_stats_data_mapping +
1571 offsetof(struct bxe_fw_stats_data, queue_stats));
1574 * First queue query index depends whether FCoE offloaded request will
1575 * be included in the ramrod
1577 first_queue_query_index = (BXE_FIRST_QUEUE_QUERY_IDX - 1);
1579 for (i = 0; i < sc->num_queues; i++) {
1581 &sc->fw_stats_req->query[first_queue_query_index + i];
1583 cur_query_entry->kind = STATS_TYPE_QUEUE;
1584 cur_query_entry->index = bxe_stats_id(&sc->fp[i]);
1585 cur_query_entry->funcID = htole16(SC_FUNC(sc));
1586 cur_query_entry->address.hi = htole32(U64_HI(cur_data_offset));
1587 cur_query_entry->address.lo = htole32(U64_LO(cur_data_offset));
1589 cur_data_offset += sizeof(struct per_queue_stats);
1594 bxe_stats_init(struct bxe_softc *sc)
1596 int /*abs*/port = SC_PORT(sc);
1597 int mb_idx = SC_FW_MB_IDX(sc);
1600 sc->stats_pending = 0;
1601 sc->executer_idx = 0;
1602 sc->stats_counter = 0;
1604 /* port and func stats for management */
1605 if (!BXE_NOMCP(sc)) {
1606 sc->port.port_stx = SHMEM_RD(sc, port_mb[port].port_stx);
1607 sc->func_stx = SHMEM_RD(sc, func_mb[mb_idx].fw_mb_param);
1609 sc->port.port_stx = 0;
1613 BLOGD(sc, DBG_STATS, "port_stx 0x%x func_stx 0x%x\n",
1614 sc->port.port_stx, sc->func_stx);
1616 /* pmf should retrieve port statistics from SP on a non-init*/
1617 if (!sc->stats_init && sc->port.pmf && sc->port.port_stx) {
1618 bxe_stats_handle(sc, STATS_EVENT_PMF);
1623 memset(&(sc->port.old_nig_stats), 0, sizeof(struct nig_stats));
1624 sc->port.old_nig_stats.brb_discard =
1625 REG_RD(sc, NIG_REG_STAT0_BRB_DISCARD + port*0x38);
1626 sc->port.old_nig_stats.brb_truncate =
1627 REG_RD(sc, NIG_REG_STAT0_BRB_TRUNCATE + port*0x38);
1628 if (!CHIP_IS_E3(sc)) {
1629 REG_RD_DMAE(sc, NIG_REG_STAT0_EGRESS_MAC_PKT0 + port*0x50,
1630 &(sc->port.old_nig_stats.egress_mac_pkt0_lo), 2);
1631 REG_RD_DMAE(sc, NIG_REG_STAT0_EGRESS_MAC_PKT1 + port*0x50,
1632 &(sc->port.old_nig_stats.egress_mac_pkt1_lo), 2);
1635 /* function stats */
1636 for (i = 0; i < sc->num_queues; i++) {
1637 memset(&sc->fp[i].old_tclient, 0, sizeof(sc->fp[i].old_tclient));
1638 memset(&sc->fp[i].old_uclient, 0, sizeof(sc->fp[i].old_uclient));
1639 memset(&sc->fp[i].old_xclient, 0, sizeof(sc->fp[i].old_xclient));
1640 if (sc->stats_init) {
1641 memset(&sc->fp[i].eth_q_stats, 0,
1642 sizeof(sc->fp[i].eth_q_stats));
1643 memset(&sc->fp[i].eth_q_stats_old, 0,
1644 sizeof(sc->fp[i].eth_q_stats_old));
1648 /* prepare statistics ramrod data */
1649 bxe_prep_fw_stats_req(sc);
1651 sc->ifnet->if_data.ifi_ipackets = 0;
1652 sc->ifnet->if_data.ifi_opackets = 0;
1653 sc->ifnet->if_data.ifi_ibytes = 0;
1654 sc->ifnet->if_data.ifi_obytes = 0;
1655 sc->ifnet->if_data.ifi_ierrors = 0;
1656 sc->ifnet->if_data.ifi_oerrors = 0;
1657 sc->ifnet->if_data.ifi_imcasts = 0;
1658 sc->ifnet->if_data.ifi_collisions = 0;
1660 if (sc->stats_init) {
1661 memset(&sc->net_stats_old, 0, sizeof(sc->net_stats_old));
1662 memset(&sc->fw_stats_old, 0, sizeof(sc->fw_stats_old));
1663 memset(&sc->eth_stats_old, 0, sizeof(sc->eth_stats_old));
1664 memset(&sc->eth_stats, 0, sizeof(sc->eth_stats));
1665 memset(&sc->func_stats, 0, sizeof(sc->func_stats));
1667 /* Clean SP from previous statistics */
1669 memset(BXE_SP(sc, func_stats), 0, sizeof(struct host_func_stats));
1670 bxe_func_stats_init(sc);
1671 bxe_hw_stats_post(sc);
1676 sc->stats_state = STATS_STATE_DISABLED;
1678 if (sc->port.pmf && sc->port.port_stx) {
1679 bxe_port_stats_base_init(sc);
1682 /* mark the end of statistics initializiation */
1683 sc->stats_init = FALSE;
1687 bxe_save_statistics(struct bxe_softc *sc)
1691 /* save queue statistics */
1692 for (i = 0; i < sc->num_queues; i++) {
1693 struct bxe_fastpath *fp = &sc->fp[i];
1694 struct bxe_eth_q_stats *qstats = &fp->eth_q_stats;
1695 struct bxe_eth_q_stats_old *qstats_old = &fp->eth_q_stats_old;
1697 UPDATE_QSTAT_OLD(total_unicast_bytes_received_hi);
1698 UPDATE_QSTAT_OLD(total_unicast_bytes_received_lo);
1699 UPDATE_QSTAT_OLD(total_broadcast_bytes_received_hi);
1700 UPDATE_QSTAT_OLD(total_broadcast_bytes_received_lo);
1701 UPDATE_QSTAT_OLD(total_multicast_bytes_received_hi);
1702 UPDATE_QSTAT_OLD(total_multicast_bytes_received_lo);
1703 UPDATE_QSTAT_OLD(total_unicast_bytes_transmitted_hi);
1704 UPDATE_QSTAT_OLD(total_unicast_bytes_transmitted_lo);
1705 UPDATE_QSTAT_OLD(total_broadcast_bytes_transmitted_hi);
1706 UPDATE_QSTAT_OLD(total_broadcast_bytes_transmitted_lo);
1707 UPDATE_QSTAT_OLD(total_multicast_bytes_transmitted_hi);
1708 UPDATE_QSTAT_OLD(total_multicast_bytes_transmitted_lo);
1709 UPDATE_QSTAT_OLD(total_tpa_bytes_hi);
1710 UPDATE_QSTAT_OLD(total_tpa_bytes_lo);
1713 /* save net_device_stats statistics */
1714 sc->net_stats_old.rx_dropped = sc->ifnet->if_data.ifi_iqdrops;
1716 /* store port firmware statistics */
1718 struct bxe_eth_stats *estats = &sc->eth_stats;
1719 struct bxe_fw_port_stats_old *fwstats = &sc->fw_stats_old;
1720 struct host_port_stats *pstats = BXE_SP(sc, port_stats);
1722 fwstats->pfc_frames_rx_hi = pstats->pfc_frames_rx_hi;
1723 fwstats->pfc_frames_rx_lo = pstats->pfc_frames_rx_lo;
1724 fwstats->pfc_frames_tx_hi = pstats->pfc_frames_tx_hi;
1725 fwstats->pfc_frames_tx_lo = pstats->pfc_frames_tx_lo;
1728 UPDATE_FW_STAT_OLD(mac_filter_discard);
1729 UPDATE_FW_STAT_OLD(mf_tag_discard);
1730 UPDATE_FW_STAT_OLD(brb_truncate_discard);
1731 UPDATE_FW_STAT_OLD(mac_discard);
1737 bxe_afex_collect_stats(struct bxe_softc *sc,
1738 void *void_afex_stats,
1739 uint32_t stats_type)
1742 struct afex_stats *afex_stats = (struct afex_stats *)void_afex_stats;
1743 struct bxe_eth_stats *estats = &sc->eth_stats;
1745 memset(afex_stats, 0, sizeof(struct afex_stats));
1747 for (i = 0; i < sc->num_queues; i++) {
1748 struct bxe_eth_q_stats *qstats = &sc->fp[i].eth_q_stats;
1750 ADD_64(afex_stats->rx_unicast_bytes_hi,
1751 qstats->total_unicast_bytes_received_hi,
1752 afex_stats->rx_unicast_bytes_lo,
1753 qstats->total_unicast_bytes_received_lo);
1755 ADD_64(afex_stats->rx_broadcast_bytes_hi,
1756 qstats->total_broadcast_bytes_received_hi,
1757 afex_stats->rx_broadcast_bytes_lo,
1758 qstats->total_broadcast_bytes_received_lo);
1760 ADD_64(afex_stats->rx_multicast_bytes_hi,
1761 qstats->total_multicast_bytes_received_hi,
1762 afex_stats->rx_multicast_bytes_lo,
1763 qstats->total_multicast_bytes_received_lo);
1765 ADD_64(afex_stats->rx_unicast_frames_hi,
1766 qstats->total_unicast_packets_received_hi,
1767 afex_stats->rx_unicast_frames_lo,
1768 qstats->total_unicast_packets_received_lo);
1770 ADD_64(afex_stats->rx_broadcast_frames_hi,
1771 qstats->total_broadcast_packets_received_hi,
1772 afex_stats->rx_broadcast_frames_lo,
1773 qstats->total_broadcast_packets_received_lo);
1775 ADD_64(afex_stats->rx_multicast_frames_hi,
1776 qstats->total_multicast_packets_received_hi,
1777 afex_stats->rx_multicast_frames_lo,
1778 qstats->total_multicast_packets_received_lo);
1781 * sum to rx_frames_discarded all discarded
1782 * packets due to size, ttl0 and checksum
1784 ADD_64(afex_stats->rx_frames_discarded_hi,
1785 qstats->total_packets_received_checksum_discarded_hi,
1786 afex_stats->rx_frames_discarded_lo,
1787 qstats->total_packets_received_checksum_discarded_lo);
1789 ADD_64(afex_stats->rx_frames_discarded_hi,
1790 qstats->total_packets_received_ttl0_discarded_hi,
1791 afex_stats->rx_frames_discarded_lo,
1792 qstats->total_packets_received_ttl0_discarded_lo);
1794 ADD_64(afex_stats->rx_frames_discarded_hi,
1795 qstats->etherstatsoverrsizepkts_hi,
1796 afex_stats->rx_frames_discarded_lo,
1797 qstats->etherstatsoverrsizepkts_lo);
1799 ADD_64(afex_stats->rx_frames_dropped_hi,
1800 qstats->no_buff_discard_hi,
1801 afex_stats->rx_frames_dropped_lo,
1802 qstats->no_buff_discard_lo);
1804 ADD_64(afex_stats->tx_unicast_bytes_hi,
1805 qstats->total_unicast_bytes_transmitted_hi,
1806 afex_stats->tx_unicast_bytes_lo,
1807 qstats->total_unicast_bytes_transmitted_lo);
1809 ADD_64(afex_stats->tx_broadcast_bytes_hi,
1810 qstats->total_broadcast_bytes_transmitted_hi,
1811 afex_stats->tx_broadcast_bytes_lo,
1812 qstats->total_broadcast_bytes_transmitted_lo);
1814 ADD_64(afex_stats->tx_multicast_bytes_hi,
1815 qstats->total_multicast_bytes_transmitted_hi,
1816 afex_stats->tx_multicast_bytes_lo,
1817 qstats->total_multicast_bytes_transmitted_lo);
1819 ADD_64(afex_stats->tx_unicast_frames_hi,
1820 qstats->total_unicast_packets_transmitted_hi,
1821 afex_stats->tx_unicast_frames_lo,
1822 qstats->total_unicast_packets_transmitted_lo);
1824 ADD_64(afex_stats->tx_broadcast_frames_hi,
1825 qstats->total_broadcast_packets_transmitted_hi,
1826 afex_stats->tx_broadcast_frames_lo,
1827 qstats->total_broadcast_packets_transmitted_lo);
1829 ADD_64(afex_stats->tx_multicast_frames_hi,
1830 qstats->total_multicast_packets_transmitted_hi,
1831 afex_stats->tx_multicast_frames_lo,
1832 qstats->total_multicast_packets_transmitted_lo);
1834 ADD_64(afex_stats->tx_frames_dropped_hi,
1835 qstats->total_transmitted_dropped_packets_error_hi,
1836 afex_stats->tx_frames_dropped_lo,
1837 qstats->total_transmitted_dropped_packets_error_lo);
1841 * If port stats are requested, add them to the PMF
1842 * stats, as anyway they will be accumulated by the
1843 * MCP before sent to the switch
1845 if ((sc->port.pmf) && (stats_type == VICSTATST_UIF_INDEX)) {
1846 ADD_64(afex_stats->rx_frames_dropped_hi,
1848 afex_stats->rx_frames_dropped_lo,
1849 estats->mac_filter_discard);
1850 ADD_64(afex_stats->rx_frames_dropped_hi,
1852 afex_stats->rx_frames_dropped_lo,
1853 estats->brb_truncate_discard);
1854 ADD_64(afex_stats->rx_frames_discarded_hi,
1856 afex_stats->rx_frames_discarded_lo,
1857 estats->mac_discard);