]> CyberLeo.Net >> Repos - FreeBSD/FreeBSD.git/blob - sys/dev/ixgbe/if_ix.c
ix: Remove cpu_id and related code from ixgbe_if_msix_intr_assign.
[FreeBSD/FreeBSD.git] / sys / dev / ixgbe / if_ix.c
1 /******************************************************************************
2
3   Copyright (c) 2001-2017, Intel Corporation
4   All rights reserved.
5
6   Redistribution and use in source and binary forms, with or without
7   modification, are permitted provided that the following conditions are met:
8
9    1. Redistributions of source code must retain the above copyright notice,
10       this list of conditions and the following disclaimer.
11
12    2. Redistributions in binary form must reproduce the above copyright
13       notice, this list of conditions and the following disclaimer in the
14       documentation and/or other materials provided with the distribution.
15
16    3. Neither the name of the Intel Corporation nor the names of its
17       contributors may be used to endorse or promote products derived from
18       this software without specific prior written permission.
19
20   THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
21   AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
22   IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
23   ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE
24   LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
25   CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
26   SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
27   INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
28   CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
29   ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
30   POSSIBILITY OF SUCH DAMAGE.
31
32 ******************************************************************************/
33 /*$FreeBSD$*/
34
35 #include "opt_inet.h"
36 #include "opt_inet6.h"
37 #include "opt_rss.h"
38
39 #include "ixgbe.h"
40 #include "ixgbe_sriov.h"
41 #include "ifdi_if.h"
42
43 #include <net/netmap.h>
44 #include <dev/netmap/netmap_kern.h>
45
46 /************************************************************************
47  * Driver version
48  ************************************************************************/
49 char ixgbe_driver_version[] = "4.0.1-k";
50
51 /************************************************************************
52  * PCI Device ID Table
53  *
54  *   Used by probe to select devices to load on
55  *   Last field stores an index into ixgbe_strings
56  *   Last entry must be all 0s
57  *
58  *   { Vendor ID, Device ID, SubVendor ID, SubDevice ID, String Index }
59  ************************************************************************/
60 static pci_vendor_info_t ixgbe_vendor_info_array[] =
61 {
62   PVID(IXGBE_INTEL_VENDOR_ID, IXGBE_DEV_ID_82598AF_DUAL_PORT,  "Intel(R) 82598EB AF (Dual Fiber)"),
63   PVID(IXGBE_INTEL_VENDOR_ID, IXGBE_DEV_ID_82598AF_SINGLE_PORT,  "Intel(R) 82598EB AF (Fiber)"),
64   PVID(IXGBE_INTEL_VENDOR_ID, IXGBE_DEV_ID_82598EB_CX4,  "Intel(R) 82598EB AT (CX4)"),
65   PVID(IXGBE_INTEL_VENDOR_ID, IXGBE_DEV_ID_82598AT,  "Intel(R) 82598EB AT"),
66   PVID(IXGBE_INTEL_VENDOR_ID, IXGBE_DEV_ID_82598AT2,  "Intel(R) 82598EB AT2"),
67   PVID(IXGBE_INTEL_VENDOR_ID, IXGBE_DEV_ID_82598,  "Intel(R) 82598"),
68   PVID(IXGBE_INTEL_VENDOR_ID, IXGBE_DEV_ID_82598_DA_DUAL_PORT,  "Intel(R) 82598EB AF DA (Dual Fiber)"),
69   PVID(IXGBE_INTEL_VENDOR_ID, IXGBE_DEV_ID_82598_CX4_DUAL_PORT,  "Intel(R) 82598EB AT (Dual CX4)"),
70   PVID(IXGBE_INTEL_VENDOR_ID, IXGBE_DEV_ID_82598EB_XF_LR,  "Intel(R) 82598EB AF (Dual Fiber LR)"),
71   PVID(IXGBE_INTEL_VENDOR_ID, IXGBE_DEV_ID_82598_SR_DUAL_PORT_EM,  "Intel(R) 82598EB AF (Dual Fiber SR)"),
72   PVID(IXGBE_INTEL_VENDOR_ID, IXGBE_DEV_ID_82598EB_SFP_LOM,  "Intel(R) 82598EB LOM"),
73   PVID(IXGBE_INTEL_VENDOR_ID, IXGBE_DEV_ID_82599_KX4,  "Intel(R) X520 82599 (KX4)"),
74   PVID(IXGBE_INTEL_VENDOR_ID, IXGBE_DEV_ID_82599_KX4_MEZZ,  "Intel(R) X520 82599 (KX4 Mezzanine)"),
75   PVID(IXGBE_INTEL_VENDOR_ID, IXGBE_DEV_ID_82599_SFP,  "Intel(R) X520 82599ES (SFI/SFP+)"),
76   PVID(IXGBE_INTEL_VENDOR_ID, IXGBE_DEV_ID_82599_XAUI_LOM,  "Intel(R) X520 82599 (XAUI/BX4)"),
77   PVID(IXGBE_INTEL_VENDOR_ID, IXGBE_DEV_ID_82599_CX4,  "Intel(R) X520 82599 (Dual CX4)"),
78   PVID(IXGBE_INTEL_VENDOR_ID, IXGBE_DEV_ID_82599_T3_LOM,  "Intel(R) X520-T 82599 LOM"),
79   PVID(IXGBE_INTEL_VENDOR_ID, IXGBE_DEV_ID_82599_COMBO_BACKPLANE,  "Intel(R) X520 82599 (Combined Backplane)"),
80   PVID(IXGBE_INTEL_VENDOR_ID, IXGBE_DEV_ID_82599_BACKPLANE_FCOE,  "Intel(R) X520 82599 (Backplane w/FCoE)"),
81   PVID(IXGBE_INTEL_VENDOR_ID, IXGBE_DEV_ID_82599_SFP_SF2,  "Intel(R) X520 82599 (Dual SFP+)"),
82   PVID(IXGBE_INTEL_VENDOR_ID, IXGBE_DEV_ID_82599_SFP_FCOE,  "Intel(R) X520 82599 (Dual SFP+ w/FCoE)"),
83   PVID(IXGBE_INTEL_VENDOR_ID, IXGBE_DEV_ID_82599EN_SFP,  "Intel(R) X520-1 82599EN (SFP+)"),
84   PVID(IXGBE_INTEL_VENDOR_ID, IXGBE_DEV_ID_82599_SFP_SF_QP,  "Intel(R) X520-4 82599 (Quad SFP+)"),
85   PVID(IXGBE_INTEL_VENDOR_ID, IXGBE_DEV_ID_82599_QSFP_SF_QP,  "Intel(R) X520-Q1 82599 (QSFP+)"),
86   PVID(IXGBE_INTEL_VENDOR_ID, IXGBE_DEV_ID_X540T,  "Intel(R) X540-AT2"),
87   PVID(IXGBE_INTEL_VENDOR_ID, IXGBE_DEV_ID_X540T1,  "Intel(R) X540-T1"),
88   PVID(IXGBE_INTEL_VENDOR_ID, IXGBE_DEV_ID_X550T,  "Intel(R) X550-T2"),
89   PVID(IXGBE_INTEL_VENDOR_ID, IXGBE_DEV_ID_X550T1, "Intel(R) X550-T1"),
90   PVID(IXGBE_INTEL_VENDOR_ID, IXGBE_DEV_ID_X550EM_X_KR,  "Intel(R) X552 (KR Backplane)"),
91   PVID(IXGBE_INTEL_VENDOR_ID, IXGBE_DEV_ID_X550EM_X_KX4,  "Intel(R) X552 (KX4 Backplane)"),
92   PVID(IXGBE_INTEL_VENDOR_ID, IXGBE_DEV_ID_X550EM_X_10G_T,  "Intel(R) X552/X557-AT (10GBASE-T)"),
93   PVID(IXGBE_INTEL_VENDOR_ID, IXGBE_DEV_ID_X550EM_X_1G_T,  "Intel(R) X552 (1000BASE-T)"),
94   PVID(IXGBE_INTEL_VENDOR_ID, IXGBE_DEV_ID_X550EM_X_SFP, "Intel(R) X552 (SFP+)"),
95   PVID(IXGBE_INTEL_VENDOR_ID, IXGBE_DEV_ID_X550EM_A_KR, "Intel(R) X553 (KR Backplane)"),
96   PVID(IXGBE_INTEL_VENDOR_ID, IXGBE_DEV_ID_X550EM_A_KR_L, "Intel(R) X553 L (KR Backplane)"),
97   PVID(IXGBE_INTEL_VENDOR_ID, IXGBE_DEV_ID_X550EM_A_SFP, "Intel(R) X553 (SFP+)"),
98   PVID(IXGBE_INTEL_VENDOR_ID, IXGBE_DEV_ID_X550EM_A_SFP_N, "Intel(R) X553 N (SFP+)"),
99   PVID(IXGBE_INTEL_VENDOR_ID, IXGBE_DEV_ID_X550EM_A_SGMII, "Intel(R) X553 (1GbE SGMII)"),
100   PVID(IXGBE_INTEL_VENDOR_ID, IXGBE_DEV_ID_X550EM_A_SGMII_L, "Intel(R) X553 L (1GbE SGMII)"),
101   PVID(IXGBE_INTEL_VENDOR_ID, IXGBE_DEV_ID_X550EM_A_10G_T, "Intel(R) X553/X557-AT (10GBASE-T)"),
102   PVID(IXGBE_INTEL_VENDOR_ID, IXGBE_DEV_ID_X550EM_A_1G_T, "Intel(R) X553 (1GbE)"),
103   PVID(IXGBE_INTEL_VENDOR_ID, IXGBE_DEV_ID_X550EM_A_1G_T_L, "Intel(R) X553 L (1GbE)"),
104   PVID(IXGBE_INTEL_VENDOR_ID, IXGBE_DEV_ID_X540_BYPASS, "Intel(R) X540-T2 (Bypass)"),
105   PVID(IXGBE_INTEL_VENDOR_ID, IXGBE_DEV_ID_82599_BYPASS, "Intel(R) X520 82599 (Bypass)"),
106         /* required last entry */
107   PVID_END
108 };
109
110 static void *ixgbe_register(device_t);
111 static int  ixgbe_if_attach_pre(if_ctx_t);
112 static int  ixgbe_if_attach_post(if_ctx_t);
113 static int  ixgbe_if_detach(if_ctx_t);
114 static int  ixgbe_if_shutdown(if_ctx_t);
115 static int  ixgbe_if_suspend(if_ctx_t);
116 static int  ixgbe_if_resume(if_ctx_t);
117
118 static void ixgbe_if_stop(if_ctx_t);
119 void ixgbe_if_enable_intr(if_ctx_t);
120 static void ixgbe_if_disable_intr(if_ctx_t);
121 static void ixgbe_link_intr_enable(if_ctx_t);
122 static int  ixgbe_if_rx_queue_intr_enable(if_ctx_t, uint16_t);
123 static void ixgbe_if_media_status(if_ctx_t, struct ifmediareq *);
124 static int  ixgbe_if_media_change(if_ctx_t);
125 static int  ixgbe_if_msix_intr_assign(if_ctx_t, int);
126 static int  ixgbe_if_mtu_set(if_ctx_t, uint32_t);
127 static void ixgbe_if_crcstrip_set(if_ctx_t, int, int);
128 static void ixgbe_if_multi_set(if_ctx_t);
129 static int  ixgbe_if_promisc_set(if_ctx_t, int);
130 static int  ixgbe_if_tx_queues_alloc(if_ctx_t, caddr_t *, uint64_t *, int, int);
131 static int  ixgbe_if_rx_queues_alloc(if_ctx_t, caddr_t *, uint64_t *, int, int);
132 static void ixgbe_if_queues_free(if_ctx_t);
133 static void ixgbe_if_timer(if_ctx_t, uint16_t);
134 static void ixgbe_if_update_admin_status(if_ctx_t);
135 static void ixgbe_if_vlan_register(if_ctx_t, u16);
136 static void ixgbe_if_vlan_unregister(if_ctx_t, u16);
137 static int  ixgbe_if_i2c_req(if_ctx_t, struct ifi2creq *);
138 static bool ixgbe_if_needs_restart(if_ctx_t, enum iflib_restart_event);
139 int ixgbe_intr(void *);
140
141 /************************************************************************
142  * Function prototypes
143  ************************************************************************/
144 static uint64_t ixgbe_if_get_counter(if_ctx_t, ift_counter);
145
146 static void ixgbe_enable_queue(struct ixgbe_softc *, u32);
147 static void ixgbe_disable_queue(struct ixgbe_softc *, u32);
148 static void ixgbe_add_device_sysctls(if_ctx_t);
149 static int  ixgbe_allocate_pci_resources(if_ctx_t);
150 static int  ixgbe_setup_low_power_mode(if_ctx_t);
151
152 static void ixgbe_config_dmac(struct ixgbe_softc *);
153 static void ixgbe_configure_ivars(struct ixgbe_softc *);
154 static void ixgbe_set_ivar(struct ixgbe_softc *, u8, u8, s8);
155 static u8   *ixgbe_mc_array_itr(struct ixgbe_hw *, u8 **, u32 *);
156 static bool ixgbe_sfp_probe(if_ctx_t);
157
158 static void ixgbe_free_pci_resources(if_ctx_t);
159
160 static int  ixgbe_msix_link(void *);
161 static int  ixgbe_msix_que(void *);
162 static void ixgbe_initialize_rss_mapping(struct ixgbe_softc *);
163 static void ixgbe_initialize_receive_units(if_ctx_t);
164 static void ixgbe_initialize_transmit_units(if_ctx_t);
165
166 static int  ixgbe_setup_interface(if_ctx_t);
167 static void ixgbe_init_device_features(struct ixgbe_softc *);
168 static void ixgbe_check_fan_failure(struct ixgbe_softc *, u32, bool);
169 static void ixgbe_sbuf_fw_version(struct ixgbe_hw *, struct sbuf *);
170 static void ixgbe_print_fw_version(if_ctx_t);
171 static void ixgbe_add_media_types(if_ctx_t);
172 static void ixgbe_update_stats_counters(struct ixgbe_softc *);
173 static void ixgbe_config_link(if_ctx_t);
174 static void ixgbe_get_slot_info(struct ixgbe_softc *);
175 static void ixgbe_check_wol_support(struct ixgbe_softc *);
176 static void ixgbe_enable_rx_drop(struct ixgbe_softc *);
177 static void ixgbe_disable_rx_drop(struct ixgbe_softc *);
178
179 static void ixgbe_add_hw_stats(struct ixgbe_softc *);
180 static int  ixgbe_set_flowcntl(struct ixgbe_softc *, int);
181 static int  ixgbe_set_advertise(struct ixgbe_softc *, int);
182 static int  ixgbe_get_default_advertise(struct ixgbe_softc *);
183 static void ixgbe_setup_vlan_hw_support(if_ctx_t);
184 static void ixgbe_config_gpie(struct ixgbe_softc *);
185 static void ixgbe_config_delay_values(struct ixgbe_softc *);
186
187 /* Sysctl handlers */
188 static int  ixgbe_sysctl_flowcntl(SYSCTL_HANDLER_ARGS);
189 static int  ixgbe_sysctl_advertise(SYSCTL_HANDLER_ARGS);
190 static int  ixgbe_sysctl_interrupt_rate_handler(SYSCTL_HANDLER_ARGS);
191 static int  ixgbe_sysctl_dmac(SYSCTL_HANDLER_ARGS);
192 static int  ixgbe_sysctl_phy_temp(SYSCTL_HANDLER_ARGS);
193 static int  ixgbe_sysctl_phy_overtemp_occurred(SYSCTL_HANDLER_ARGS);
194 static int  ixgbe_sysctl_print_fw_version(SYSCTL_HANDLER_ARGS);
195 #ifdef IXGBE_DEBUG
196 static int  ixgbe_sysctl_power_state(SYSCTL_HANDLER_ARGS);
197 static int  ixgbe_sysctl_print_rss_config(SYSCTL_HANDLER_ARGS);
198 #endif
199 static int  ixgbe_sysctl_rdh_handler(SYSCTL_HANDLER_ARGS);
200 static int  ixgbe_sysctl_rdt_handler(SYSCTL_HANDLER_ARGS);
201 static int  ixgbe_sysctl_tdt_handler(SYSCTL_HANDLER_ARGS);
202 static int  ixgbe_sysctl_tdh_handler(SYSCTL_HANDLER_ARGS);
203 static int  ixgbe_sysctl_eee_state(SYSCTL_HANDLER_ARGS);
204 static int  ixgbe_sysctl_wol_enable(SYSCTL_HANDLER_ARGS);
205 static int  ixgbe_sysctl_wufc(SYSCTL_HANDLER_ARGS);
206
207 /* Deferred interrupt tasklets */
208 static void ixgbe_handle_msf(void *);
209 static void ixgbe_handle_mod(void *);
210 static void ixgbe_handle_phy(void *);
211
212 /************************************************************************
213  *  FreeBSD Device Interface Entry Points
214  ************************************************************************/
215 static device_method_t ix_methods[] = {
216         /* Device interface */
217         DEVMETHOD(device_register, ixgbe_register),
218         DEVMETHOD(device_probe, iflib_device_probe),
219         DEVMETHOD(device_attach, iflib_device_attach),
220         DEVMETHOD(device_detach, iflib_device_detach),
221         DEVMETHOD(device_shutdown, iflib_device_shutdown),
222         DEVMETHOD(device_suspend, iflib_device_suspend),
223         DEVMETHOD(device_resume, iflib_device_resume),
224 #ifdef PCI_IOV
225         DEVMETHOD(pci_iov_init, iflib_device_iov_init),
226         DEVMETHOD(pci_iov_uninit, iflib_device_iov_uninit),
227         DEVMETHOD(pci_iov_add_vf, iflib_device_iov_add_vf),
228 #endif /* PCI_IOV */
229         DEVMETHOD_END
230 };
231
232 static driver_t ix_driver = {
233         "ix", ix_methods, sizeof(struct ixgbe_softc),
234 };
235
236 devclass_t ix_devclass;
237 DRIVER_MODULE(ix, pci, ix_driver, ix_devclass, 0, 0);
238 IFLIB_PNP_INFO(pci, ix_driver, ixgbe_vendor_info_array);
239 MODULE_DEPEND(ix, pci, 1, 1, 1);
240 MODULE_DEPEND(ix, ether, 1, 1, 1);
241 MODULE_DEPEND(ix, iflib, 1, 1, 1);
242
243 static device_method_t ixgbe_if_methods[] = {
244         DEVMETHOD(ifdi_attach_pre, ixgbe_if_attach_pre),
245         DEVMETHOD(ifdi_attach_post, ixgbe_if_attach_post),
246         DEVMETHOD(ifdi_detach, ixgbe_if_detach),
247         DEVMETHOD(ifdi_shutdown, ixgbe_if_shutdown),
248         DEVMETHOD(ifdi_suspend, ixgbe_if_suspend),
249         DEVMETHOD(ifdi_resume, ixgbe_if_resume),
250         DEVMETHOD(ifdi_init, ixgbe_if_init),
251         DEVMETHOD(ifdi_stop, ixgbe_if_stop),
252         DEVMETHOD(ifdi_msix_intr_assign, ixgbe_if_msix_intr_assign),
253         DEVMETHOD(ifdi_intr_enable, ixgbe_if_enable_intr),
254         DEVMETHOD(ifdi_intr_disable, ixgbe_if_disable_intr),
255         DEVMETHOD(ifdi_link_intr_enable, ixgbe_link_intr_enable),
256         DEVMETHOD(ifdi_tx_queue_intr_enable, ixgbe_if_rx_queue_intr_enable),
257         DEVMETHOD(ifdi_rx_queue_intr_enable, ixgbe_if_rx_queue_intr_enable),
258         DEVMETHOD(ifdi_tx_queues_alloc, ixgbe_if_tx_queues_alloc),
259         DEVMETHOD(ifdi_rx_queues_alloc, ixgbe_if_rx_queues_alloc),
260         DEVMETHOD(ifdi_queues_free, ixgbe_if_queues_free),
261         DEVMETHOD(ifdi_update_admin_status, ixgbe_if_update_admin_status),
262         DEVMETHOD(ifdi_multi_set, ixgbe_if_multi_set),
263         DEVMETHOD(ifdi_mtu_set, ixgbe_if_mtu_set),
264         DEVMETHOD(ifdi_crcstrip_set, ixgbe_if_crcstrip_set),
265         DEVMETHOD(ifdi_media_status, ixgbe_if_media_status),
266         DEVMETHOD(ifdi_media_change, ixgbe_if_media_change),
267         DEVMETHOD(ifdi_promisc_set, ixgbe_if_promisc_set),
268         DEVMETHOD(ifdi_timer, ixgbe_if_timer),
269         DEVMETHOD(ifdi_vlan_register, ixgbe_if_vlan_register),
270         DEVMETHOD(ifdi_vlan_unregister, ixgbe_if_vlan_unregister),
271         DEVMETHOD(ifdi_get_counter, ixgbe_if_get_counter),
272         DEVMETHOD(ifdi_i2c_req, ixgbe_if_i2c_req),
273         DEVMETHOD(ifdi_needs_restart, ixgbe_if_needs_restart),
274 #ifdef PCI_IOV
275         DEVMETHOD(ifdi_iov_init, ixgbe_if_iov_init),
276         DEVMETHOD(ifdi_iov_uninit, ixgbe_if_iov_uninit),
277         DEVMETHOD(ifdi_iov_vf_add, ixgbe_if_iov_vf_add),
278 #endif /* PCI_IOV */
279         DEVMETHOD_END
280 };
281
282 /*
283  * TUNEABLE PARAMETERS:
284  */
285
286 static SYSCTL_NODE(_hw, OID_AUTO, ix, CTLFLAG_RD | CTLFLAG_MPSAFE, 0,
287     "IXGBE driver parameters");
288 static driver_t ixgbe_if_driver = {
289   "ixgbe_if", ixgbe_if_methods, sizeof(struct ixgbe_softc)
290 };
291
292 static int ixgbe_max_interrupt_rate = (4000000 / IXGBE_LOW_LATENCY);
293 SYSCTL_INT(_hw_ix, OID_AUTO, max_interrupt_rate, CTLFLAG_RDTUN,
294     &ixgbe_max_interrupt_rate, 0, "Maximum interrupts per second");
295
296 /* Flow control setting, default to full */
297 static int ixgbe_flow_control = ixgbe_fc_full;
298 SYSCTL_INT(_hw_ix, OID_AUTO, flow_control, CTLFLAG_RDTUN,
299     &ixgbe_flow_control, 0, "Default flow control used for all adapters");
300
301 /* Advertise Speed, default to 0 (auto) */
302 static int ixgbe_advertise_speed = 0;
303 SYSCTL_INT(_hw_ix, OID_AUTO, advertise_speed, CTLFLAG_RDTUN,
304     &ixgbe_advertise_speed, 0, "Default advertised speed for all adapters");
305
306 /*
307  * Smart speed setting, default to on
308  * this only works as a compile option
309  * right now as its during attach, set
310  * this to 'ixgbe_smart_speed_off' to
311  * disable.
312  */
313 static int ixgbe_smart_speed = ixgbe_smart_speed_on;
314
315 /*
316  * MSI-X should be the default for best performance,
317  * but this allows it to be forced off for testing.
318  */
319 static int ixgbe_enable_msix = 1;
320 SYSCTL_INT(_hw_ix, OID_AUTO, enable_msix, CTLFLAG_RDTUN, &ixgbe_enable_msix, 0,
321     "Enable MSI-X interrupts");
322
323 /*
324  * Defining this on will allow the use
325  * of unsupported SFP+ modules, note that
326  * doing so you are on your own :)
327  */
328 static int allow_unsupported_sfp = false;
329 SYSCTL_INT(_hw_ix, OID_AUTO, unsupported_sfp, CTLFLAG_RDTUN,
330     &allow_unsupported_sfp, 0,
331     "Allow unsupported SFP modules...use at your own risk");
332
333 /*
334  * Not sure if Flow Director is fully baked,
335  * so we'll default to turning it off.
336  */
337 static int ixgbe_enable_fdir = 0;
338 SYSCTL_INT(_hw_ix, OID_AUTO, enable_fdir, CTLFLAG_RDTUN, &ixgbe_enable_fdir, 0,
339     "Enable Flow Director");
340
341 /* Receive-Side Scaling */
342 static int ixgbe_enable_rss = 1;
343 SYSCTL_INT(_hw_ix, OID_AUTO, enable_rss, CTLFLAG_RDTUN, &ixgbe_enable_rss, 0,
344     "Enable Receive-Side Scaling (RSS)");
345
346 /*
347  * AIM: Adaptive Interrupt Moderation
348  * which means that the interrupt rate
349  * is varied over time based on the
350  * traffic for that interrupt vector
351  */
352 static int ixgbe_enable_aim = false;
353 SYSCTL_INT(_hw_ix, OID_AUTO, enable_aim, CTLFLAG_RWTUN, &ixgbe_enable_aim, 0,
354     "Enable adaptive interrupt moderation");
355
356 #if 0
357 /* Keep running tab on them for sanity check */
358 static int ixgbe_total_ports;
359 #endif
360
361 MALLOC_DEFINE(M_IXGBE, "ix", "ix driver allocations");
362
363 /*
364  * For Flow Director: this is the number of TX packets we sample
365  * for the filter pool, this means every 20th packet will be probed.
366  *
367  * This feature can be disabled by setting this to 0.
368  */
369 static int atr_sample_rate = 20;
370
371 extern struct if_txrx ixgbe_txrx;
372
373 static struct if_shared_ctx ixgbe_sctx_init = {
374         .isc_magic = IFLIB_MAGIC,
375         .isc_q_align = PAGE_SIZE,/* max(DBA_ALIGN, PAGE_SIZE) */
376         .isc_tx_maxsize = IXGBE_TSO_SIZE + sizeof(struct ether_vlan_header),
377         .isc_tx_maxsegsize = PAGE_SIZE,
378         .isc_tso_maxsize = IXGBE_TSO_SIZE + sizeof(struct ether_vlan_header),
379         .isc_tso_maxsegsize = PAGE_SIZE,
380         .isc_rx_maxsize = PAGE_SIZE*4,
381         .isc_rx_nsegments = 1,
382         .isc_rx_maxsegsize = PAGE_SIZE*4,
383         .isc_nfl = 1,
384         .isc_ntxqs = 1,
385         .isc_nrxqs = 1,
386
387         .isc_admin_intrcnt = 1,
388         .isc_vendor_info = ixgbe_vendor_info_array,
389         .isc_driver_version = ixgbe_driver_version,
390         .isc_driver = &ixgbe_if_driver,
391         .isc_flags = IFLIB_TSO_INIT_IP,
392
393         .isc_nrxd_min = {MIN_RXD},
394         .isc_ntxd_min = {MIN_TXD},
395         .isc_nrxd_max = {MAX_RXD},
396         .isc_ntxd_max = {MAX_TXD},
397         .isc_nrxd_default = {DEFAULT_RXD},
398         .isc_ntxd_default = {DEFAULT_TXD},
399 };
400
401 /************************************************************************
402  * ixgbe_if_tx_queues_alloc
403  ************************************************************************/
404 static int
405 ixgbe_if_tx_queues_alloc(if_ctx_t ctx, caddr_t *vaddrs, uint64_t *paddrs,
406     int ntxqs, int ntxqsets)
407 {
408         struct ixgbe_softc *sc = iflib_get_softc(ctx);
409         if_softc_ctx_t     scctx = sc->shared;
410         struct ix_tx_queue *que;
411         int                i, j, error;
412
413         MPASS(sc->num_tx_queues > 0);
414         MPASS(sc->num_tx_queues == ntxqsets);
415         MPASS(ntxqs == 1);
416
417         /* Allocate queue structure memory */
418         sc->tx_queues =
419             (struct ix_tx_queue *)malloc(sizeof(struct ix_tx_queue) * ntxqsets,
420                                          M_IXGBE, M_NOWAIT | M_ZERO);
421         if (!sc->tx_queues) {
422                 device_printf(iflib_get_dev(ctx),
423                     "Unable to allocate TX ring memory\n");
424                 return (ENOMEM);
425         }
426
427         for (i = 0, que = sc->tx_queues; i < ntxqsets; i++, que++) {
428                 struct tx_ring *txr = &que->txr;
429
430                 /* In case SR-IOV is enabled, align the index properly */
431                 txr->me = ixgbe_vf_que_index(sc->iov_mode, sc->pool,
432                     i);
433
434                 txr->sc = que->sc = sc;
435
436                 /* Allocate report status array */
437                 txr->tx_rsq = (qidx_t *)malloc(sizeof(qidx_t) * scctx->isc_ntxd[0], M_IXGBE, M_NOWAIT | M_ZERO);
438                 if (txr->tx_rsq == NULL) {
439                         error = ENOMEM;
440                         goto fail;
441                 }
442                 for (j = 0; j < scctx->isc_ntxd[0]; j++)
443                         txr->tx_rsq[j] = QIDX_INVALID;
444                 /* get the virtual and physical address of the hardware queues */
445                 txr->tail = IXGBE_TDT(txr->me);
446                 txr->tx_base = (union ixgbe_adv_tx_desc *)vaddrs[i];
447                 txr->tx_paddr = paddrs[i];
448
449                 txr->bytes = 0;
450                 txr->total_packets = 0;
451
452                 /* Set the rate at which we sample packets */
453                 if (sc->feat_en & IXGBE_FEATURE_FDIR)
454                         txr->atr_sample = atr_sample_rate;
455
456         }
457
458         device_printf(iflib_get_dev(ctx), "allocated for %d queues\n",
459             sc->num_tx_queues);
460
461         return (0);
462
463 fail:
464         ixgbe_if_queues_free(ctx);
465
466         return (error);
467 } /* ixgbe_if_tx_queues_alloc */
468
469 /************************************************************************
470  * ixgbe_if_rx_queues_alloc
471  ************************************************************************/
472 static int
473 ixgbe_if_rx_queues_alloc(if_ctx_t ctx, caddr_t *vaddrs, uint64_t *paddrs,
474     int nrxqs, int nrxqsets)
475 {
476         struct ixgbe_softc     *sc = iflib_get_softc(ctx);
477         struct ix_rx_queue *que;
478         int                i;
479
480         MPASS(sc->num_rx_queues > 0);
481         MPASS(sc->num_rx_queues == nrxqsets);
482         MPASS(nrxqs == 1);
483
484         /* Allocate queue structure memory */
485         sc->rx_queues =
486             (struct ix_rx_queue *)malloc(sizeof(struct ix_rx_queue)*nrxqsets,
487                                          M_IXGBE, M_NOWAIT | M_ZERO);
488         if (!sc->rx_queues) {
489                 device_printf(iflib_get_dev(ctx),
490                     "Unable to allocate TX ring memory\n");
491                 return (ENOMEM);
492         }
493
494         for (i = 0, que = sc->rx_queues; i < nrxqsets; i++, que++) {
495                 struct rx_ring *rxr = &que->rxr;
496
497                 /* In case SR-IOV is enabled, align the index properly */
498                 rxr->me = ixgbe_vf_que_index(sc->iov_mode, sc->pool,
499                     i);
500
501                 rxr->sc = que->sc = sc;
502
503                 /* get the virtual and physical address of the hw queues */
504                 rxr->tail = IXGBE_RDT(rxr->me);
505                 rxr->rx_base = (union ixgbe_adv_rx_desc *)vaddrs[i];
506                 rxr->rx_paddr = paddrs[i];
507                 rxr->bytes = 0;
508                 rxr->que = que;
509         }
510
511         device_printf(iflib_get_dev(ctx), "allocated for %d rx queues\n",
512             sc->num_rx_queues);
513
514         return (0);
515 } /* ixgbe_if_rx_queues_alloc */
516
517 /************************************************************************
518  * ixgbe_if_queues_free
519  ************************************************************************/
520 static void
521 ixgbe_if_queues_free(if_ctx_t ctx)
522 {
523         struct ixgbe_softc     *sc = iflib_get_softc(ctx);
524         struct ix_tx_queue *tx_que = sc->tx_queues;
525         struct ix_rx_queue *rx_que = sc->rx_queues;
526         int                i;
527
528         if (tx_que != NULL) {
529                 for (i = 0; i < sc->num_tx_queues; i++, tx_que++) {
530                         struct tx_ring *txr = &tx_que->txr;
531                         if (txr->tx_rsq == NULL)
532                                 break;
533
534                         free(txr->tx_rsq, M_IXGBE);
535                         txr->tx_rsq = NULL;
536                 }
537
538                 free(sc->tx_queues, M_IXGBE);
539                 sc->tx_queues = NULL;
540         }
541         if (rx_que != NULL) {
542                 free(sc->rx_queues, M_IXGBE);
543                 sc->rx_queues = NULL;
544         }
545 } /* ixgbe_if_queues_free */
546
547 /************************************************************************
548  * ixgbe_initialize_rss_mapping
549  ************************************************************************/
550 static void
551 ixgbe_initialize_rss_mapping(struct ixgbe_softc *sc)
552 {
553         struct ixgbe_hw *hw = &sc->hw;
554         u32             reta = 0, mrqc, rss_key[10];
555         int             queue_id, table_size, index_mult;
556         int             i, j;
557         u32             rss_hash_config;
558
559         if (sc->feat_en & IXGBE_FEATURE_RSS) {
560                 /* Fetch the configured RSS key */
561                 rss_getkey((uint8_t *)&rss_key);
562         } else {
563                 /* set up random bits */
564                 arc4rand(&rss_key, sizeof(rss_key), 0);
565         }
566
567         /* Set multiplier for RETA setup and table size based on MAC */
568         index_mult = 0x1;
569         table_size = 128;
570         switch (sc->hw.mac.type) {
571         case ixgbe_mac_82598EB:
572                 index_mult = 0x11;
573                 break;
574         case ixgbe_mac_X550:
575         case ixgbe_mac_X550EM_x:
576         case ixgbe_mac_X550EM_a:
577                 table_size = 512;
578                 break;
579         default:
580                 break;
581         }
582
583         /* Set up the redirection table */
584         for (i = 0, j = 0; i < table_size; i++, j++) {
585                 if (j == sc->num_rx_queues)
586                         j = 0;
587
588                 if (sc->feat_en & IXGBE_FEATURE_RSS) {
589                         /*
590                          * Fetch the RSS bucket id for the given indirection
591                          * entry. Cap it at the number of configured buckets
592                          * (which is num_rx_queues.)
593                          */
594                         queue_id = rss_get_indirection_to_bucket(i);
595                         queue_id = queue_id % sc->num_rx_queues;
596                 } else
597                         queue_id = (j * index_mult);
598
599                 /*
600                  * The low 8 bits are for hash value (n+0);
601                  * The next 8 bits are for hash value (n+1), etc.
602                  */
603                 reta = reta >> 8;
604                 reta = reta | (((uint32_t)queue_id) << 24);
605                 if ((i & 3) == 3) {
606                         if (i < 128)
607                                 IXGBE_WRITE_REG(hw, IXGBE_RETA(i >> 2), reta);
608                         else
609                                 IXGBE_WRITE_REG(hw, IXGBE_ERETA((i >> 2) - 32),
610                                     reta);
611                         reta = 0;
612                 }
613         }
614
615         /* Now fill our hash function seeds */
616         for (i = 0; i < 10; i++)
617                 IXGBE_WRITE_REG(hw, IXGBE_RSSRK(i), rss_key[i]);
618
619         /* Perform hash on these packet types */
620         if (sc->feat_en & IXGBE_FEATURE_RSS)
621                 rss_hash_config = rss_gethashconfig();
622         else {
623                 /*
624                  * Disable UDP - IP fragments aren't currently being handled
625                  * and so we end up with a mix of 2-tuple and 4-tuple
626                  * traffic.
627                  */
628                 rss_hash_config = RSS_HASHTYPE_RSS_IPV4
629                                 | RSS_HASHTYPE_RSS_TCP_IPV4
630                                 | RSS_HASHTYPE_RSS_IPV6
631                                 | RSS_HASHTYPE_RSS_TCP_IPV6
632                                 | RSS_HASHTYPE_RSS_IPV6_EX
633                                 | RSS_HASHTYPE_RSS_TCP_IPV6_EX;
634         }
635
636         mrqc = IXGBE_MRQC_RSSEN;
637         if (rss_hash_config & RSS_HASHTYPE_RSS_IPV4)
638                 mrqc |= IXGBE_MRQC_RSS_FIELD_IPV4;
639         if (rss_hash_config & RSS_HASHTYPE_RSS_TCP_IPV4)
640                 mrqc |= IXGBE_MRQC_RSS_FIELD_IPV4_TCP;
641         if (rss_hash_config & RSS_HASHTYPE_RSS_IPV6)
642                 mrqc |= IXGBE_MRQC_RSS_FIELD_IPV6;
643         if (rss_hash_config & RSS_HASHTYPE_RSS_TCP_IPV6)
644                 mrqc |= IXGBE_MRQC_RSS_FIELD_IPV6_TCP;
645         if (rss_hash_config & RSS_HASHTYPE_RSS_IPV6_EX)
646                 mrqc |= IXGBE_MRQC_RSS_FIELD_IPV6_EX;
647         if (rss_hash_config & RSS_HASHTYPE_RSS_TCP_IPV6_EX)
648                 mrqc |= IXGBE_MRQC_RSS_FIELD_IPV6_EX_TCP;
649         if (rss_hash_config & RSS_HASHTYPE_RSS_UDP_IPV4)
650                 mrqc |= IXGBE_MRQC_RSS_FIELD_IPV4_UDP;
651         if (rss_hash_config & RSS_HASHTYPE_RSS_UDP_IPV6)
652                 mrqc |= IXGBE_MRQC_RSS_FIELD_IPV6_UDP;
653         if (rss_hash_config & RSS_HASHTYPE_RSS_UDP_IPV6_EX)
654                 mrqc |= IXGBE_MRQC_RSS_FIELD_IPV6_EX_UDP;
655         mrqc |= ixgbe_get_mrqc(sc->iov_mode);
656         IXGBE_WRITE_REG(hw, IXGBE_MRQC, mrqc);
657 } /* ixgbe_initialize_rss_mapping */
658
659 /************************************************************************
660  * ixgbe_initialize_receive_units - Setup receive registers and features.
661  ************************************************************************/
662 #define BSIZEPKT_ROUNDUP ((1<<IXGBE_SRRCTL_BSIZEPKT_SHIFT)-1)
663
664 static void
665 ixgbe_initialize_receive_units(if_ctx_t ctx)
666 {
667         struct ixgbe_softc     *sc = iflib_get_softc(ctx);
668         if_softc_ctx_t     scctx = sc->shared;
669         struct ixgbe_hw    *hw = &sc->hw;
670         struct ifnet       *ifp = iflib_get_ifp(ctx);
671         struct ix_rx_queue *que;
672         int                i, j;
673         u32                bufsz, fctrl, srrctl, rxcsum;
674         u32                hlreg;
675
676         /*
677          * Make sure receives are disabled while
678          * setting up the descriptor ring
679          */
680         ixgbe_disable_rx(hw);
681
682         /* Enable broadcasts */
683         fctrl = IXGBE_READ_REG(hw, IXGBE_FCTRL);
684         fctrl |= IXGBE_FCTRL_BAM;
685         if (sc->hw.mac.type == ixgbe_mac_82598EB) {
686                 fctrl |= IXGBE_FCTRL_DPF;
687                 fctrl |= IXGBE_FCTRL_PMCF;
688         }
689         IXGBE_WRITE_REG(hw, IXGBE_FCTRL, fctrl);
690
691         /* Set for Jumbo Frames? */
692         hlreg = IXGBE_READ_REG(hw, IXGBE_HLREG0);
693         if (ifp->if_mtu > ETHERMTU)
694                 hlreg |= IXGBE_HLREG0_JUMBOEN;
695         else
696                 hlreg &= ~IXGBE_HLREG0_JUMBOEN;
697         IXGBE_WRITE_REG(hw, IXGBE_HLREG0, hlreg);
698
699         bufsz = (sc->rx_mbuf_sz + BSIZEPKT_ROUNDUP) >>
700             IXGBE_SRRCTL_BSIZEPKT_SHIFT;
701
702         /* Setup the Base and Length of the Rx Descriptor Ring */
703         for (i = 0, que = sc->rx_queues; i < sc->num_rx_queues; i++, que++) {
704                 struct rx_ring *rxr = &que->rxr;
705                 u64            rdba = rxr->rx_paddr;
706
707                 j = rxr->me;
708
709                 /* Setup the Base and Length of the Rx Descriptor Ring */
710                 IXGBE_WRITE_REG(hw, IXGBE_RDBAL(j),
711                     (rdba & 0x00000000ffffffffULL));
712                 IXGBE_WRITE_REG(hw, IXGBE_RDBAH(j), (rdba >> 32));
713                 IXGBE_WRITE_REG(hw, IXGBE_RDLEN(j),
714                      scctx->isc_nrxd[0] * sizeof(union ixgbe_adv_rx_desc));
715
716                 /* Set up the SRRCTL register */
717                 srrctl = IXGBE_READ_REG(hw, IXGBE_SRRCTL(j));
718                 srrctl &= ~IXGBE_SRRCTL_BSIZEHDR_MASK;
719                 srrctl &= ~IXGBE_SRRCTL_BSIZEPKT_MASK;
720                 srrctl |= bufsz;
721                 srrctl |= IXGBE_SRRCTL_DESCTYPE_ADV_ONEBUF;
722
723                 /*
724                  * Set DROP_EN iff we have no flow control and >1 queue.
725                  * Note that srrctl was cleared shortly before during reset,
726                  * so we do not need to clear the bit, but do it just in case
727                  * this code is moved elsewhere.
728                  */
729                 if (sc->num_rx_queues > 1 &&
730                     sc->hw.fc.requested_mode == ixgbe_fc_none) {
731                         srrctl |= IXGBE_SRRCTL_DROP_EN;
732                 } else {
733                         srrctl &= ~IXGBE_SRRCTL_DROP_EN;
734                 }
735
736                 IXGBE_WRITE_REG(hw, IXGBE_SRRCTL(j), srrctl);
737
738                 /* Setup the HW Rx Head and Tail Descriptor Pointers */
739                 IXGBE_WRITE_REG(hw, IXGBE_RDH(j), 0);
740                 IXGBE_WRITE_REG(hw, IXGBE_RDT(j), 0);
741
742                 /* Set the driver rx tail address */
743                 rxr->tail =  IXGBE_RDT(rxr->me);
744         }
745
746         if (sc->hw.mac.type != ixgbe_mac_82598EB) {
747                 u32 psrtype = IXGBE_PSRTYPE_TCPHDR
748                             | IXGBE_PSRTYPE_UDPHDR
749                             | IXGBE_PSRTYPE_IPV4HDR
750                             | IXGBE_PSRTYPE_IPV6HDR;
751                 IXGBE_WRITE_REG(hw, IXGBE_PSRTYPE(0), psrtype);
752         }
753
754         rxcsum = IXGBE_READ_REG(hw, IXGBE_RXCSUM);
755
756         ixgbe_initialize_rss_mapping(sc);
757
758         if (sc->num_rx_queues > 1) {
759                 /* RSS and RX IPP Checksum are mutually exclusive */
760                 rxcsum |= IXGBE_RXCSUM_PCSD;
761         }
762
763         if (ifp->if_capenable & IFCAP_RXCSUM)
764                 rxcsum |= IXGBE_RXCSUM_PCSD;
765
766         /* This is useful for calculating UDP/IP fragment checksums */
767         if (!(rxcsum & IXGBE_RXCSUM_PCSD))
768                 rxcsum |= IXGBE_RXCSUM_IPPCSE;
769
770         IXGBE_WRITE_REG(hw, IXGBE_RXCSUM, rxcsum);
771
772 } /* ixgbe_initialize_receive_units */
773
774 /************************************************************************
775  * ixgbe_initialize_transmit_units - Enable transmit units.
776  ************************************************************************/
777 static void
778 ixgbe_initialize_transmit_units(if_ctx_t ctx)
779 {
780         struct ixgbe_softc     *sc = iflib_get_softc(ctx);
781         struct ixgbe_hw    *hw = &sc->hw;
782         if_softc_ctx_t     scctx = sc->shared;
783         struct ix_tx_queue *que;
784         int i;
785
786         /* Setup the Base and Length of the Tx Descriptor Ring */
787         for (i = 0, que = sc->tx_queues; i < sc->num_tx_queues;
788             i++, que++) {
789                 struct tx_ring     *txr = &que->txr;
790                 u64 tdba = txr->tx_paddr;
791                 u32 txctrl = 0;
792                 int j = txr->me;
793
794                 IXGBE_WRITE_REG(hw, IXGBE_TDBAL(j),
795                     (tdba & 0x00000000ffffffffULL));
796                 IXGBE_WRITE_REG(hw, IXGBE_TDBAH(j), (tdba >> 32));
797                 IXGBE_WRITE_REG(hw, IXGBE_TDLEN(j),
798                     scctx->isc_ntxd[0] * sizeof(union ixgbe_adv_tx_desc));
799
800                 /* Setup the HW Tx Head and Tail descriptor pointers */
801                 IXGBE_WRITE_REG(hw, IXGBE_TDH(j), 0);
802                 IXGBE_WRITE_REG(hw, IXGBE_TDT(j), 0);
803
804                 /* Cache the tail address */
805                 txr->tail = IXGBE_TDT(txr->me);
806
807                 txr->tx_rs_cidx = txr->tx_rs_pidx;
808                 txr->tx_cidx_processed = scctx->isc_ntxd[0] - 1;
809                 for (int k = 0; k < scctx->isc_ntxd[0]; k++)
810                         txr->tx_rsq[k] = QIDX_INVALID;
811
812                 /* Disable Head Writeback */
813                 /*
814                  * Note: for X550 series devices, these registers are actually
815                  * prefixed with TPH_ isntead of DCA_, but the addresses and
816                  * fields remain the same.
817                  */
818                 switch (hw->mac.type) {
819                 case ixgbe_mac_82598EB:
820                         txctrl = IXGBE_READ_REG(hw, IXGBE_DCA_TXCTRL(j));
821                         break;
822                 default:
823                         txctrl = IXGBE_READ_REG(hw, IXGBE_DCA_TXCTRL_82599(j));
824                         break;
825                 }
826                 txctrl &= ~IXGBE_DCA_TXCTRL_DESC_WRO_EN;
827                 switch (hw->mac.type) {
828                 case ixgbe_mac_82598EB:
829                         IXGBE_WRITE_REG(hw, IXGBE_DCA_TXCTRL(j), txctrl);
830                         break;
831                 default:
832                         IXGBE_WRITE_REG(hw, IXGBE_DCA_TXCTRL_82599(j), txctrl);
833                         break;
834                 }
835
836         }
837
838         if (hw->mac.type != ixgbe_mac_82598EB) {
839                 u32 dmatxctl, rttdcs;
840
841                 dmatxctl = IXGBE_READ_REG(hw, IXGBE_DMATXCTL);
842                 dmatxctl |= IXGBE_DMATXCTL_TE;
843                 IXGBE_WRITE_REG(hw, IXGBE_DMATXCTL, dmatxctl);
844                 /* Disable arbiter to set MTQC */
845                 rttdcs = IXGBE_READ_REG(hw, IXGBE_RTTDCS);
846                 rttdcs |= IXGBE_RTTDCS_ARBDIS;
847                 IXGBE_WRITE_REG(hw, IXGBE_RTTDCS, rttdcs);
848                 IXGBE_WRITE_REG(hw, IXGBE_MTQC,
849                     ixgbe_get_mtqc(sc->iov_mode));
850                 rttdcs &= ~IXGBE_RTTDCS_ARBDIS;
851                 IXGBE_WRITE_REG(hw, IXGBE_RTTDCS, rttdcs);
852         }
853
854 } /* ixgbe_initialize_transmit_units */
855
856 /************************************************************************
857  * ixgbe_register
858  ************************************************************************/
859 static void *
860 ixgbe_register(device_t dev)
861 {
862         return (&ixgbe_sctx_init);
863 } /* ixgbe_register */
864
865 /************************************************************************
866  * ixgbe_if_attach_pre - Device initialization routine, part 1
867  *
868  *   Called when the driver is being loaded.
869  *   Identifies the type of hardware, initializes the hardware,
870  *   and initializes iflib structures.
871  *
872  *   return 0 on success, positive on failure
873  ************************************************************************/
874 static int
875 ixgbe_if_attach_pre(if_ctx_t ctx)
876 {
877         struct ixgbe_softc  *sc;
878         device_t        dev;
879         if_softc_ctx_t  scctx;
880         struct ixgbe_hw *hw;
881         int             error = 0;
882         u32             ctrl_ext;
883
884         INIT_DEBUGOUT("ixgbe_attach: begin");
885
886         /* Allocate, clear, and link in our adapter structure */
887         dev = iflib_get_dev(ctx);
888         sc = iflib_get_softc(ctx);
889         sc->hw.back = sc;
890         sc->ctx = ctx;
891         sc->dev = dev;
892         scctx = sc->shared = iflib_get_softc_ctx(ctx);
893         sc->media = iflib_get_media(ctx);
894         hw = &sc->hw;
895
896         /* Determine hardware revision */
897         hw->vendor_id = pci_get_vendor(dev);
898         hw->device_id = pci_get_device(dev);
899         hw->revision_id = pci_get_revid(dev);
900         hw->subsystem_vendor_id = pci_get_subvendor(dev);
901         hw->subsystem_device_id = pci_get_subdevice(dev);
902
903         /* Do base PCI setup - map BAR0 */
904         if (ixgbe_allocate_pci_resources(ctx)) {
905                 device_printf(dev, "Allocation of PCI resources failed\n");
906                 return (ENXIO);
907         }
908
909         /* let hardware know driver is loaded */
910         ctrl_ext = IXGBE_READ_REG(hw, IXGBE_CTRL_EXT);
911         ctrl_ext |= IXGBE_CTRL_EXT_DRV_LOAD;
912         IXGBE_WRITE_REG(hw, IXGBE_CTRL_EXT, ctrl_ext);
913
914         /*
915          * Initialize the shared code
916          */
917         if (ixgbe_init_shared_code(hw) != 0) {
918                 device_printf(dev, "Unable to initialize the shared code\n");
919                 error = ENXIO;
920                 goto err_pci;
921         }
922
923         if (hw->mbx.ops.init_params)
924                 hw->mbx.ops.init_params(hw);
925
926         hw->allow_unsupported_sfp = allow_unsupported_sfp;
927
928         if (hw->mac.type != ixgbe_mac_82598EB)
929                 hw->phy.smart_speed = ixgbe_smart_speed;
930
931         ixgbe_init_device_features(sc);
932
933         /* Enable WoL (if supported) */
934         ixgbe_check_wol_support(sc);
935
936         /* Verify adapter fan is still functional (if applicable) */
937         if (sc->feat_en & IXGBE_FEATURE_FAN_FAIL) {
938                 u32 esdp = IXGBE_READ_REG(hw, IXGBE_ESDP);
939                 ixgbe_check_fan_failure(sc, esdp, false);
940         }
941
942         /* Ensure SW/FW semaphore is free */
943         ixgbe_init_swfw_semaphore(hw);
944
945         /* Set an initial default flow control value */
946         hw->fc.requested_mode = ixgbe_flow_control;
947
948         hw->phy.reset_if_overtemp = true;
949         error = ixgbe_reset_hw(hw);
950         hw->phy.reset_if_overtemp = false;
951         if (error == IXGBE_ERR_SFP_NOT_PRESENT) {
952                 /*
953                  * No optics in this port, set up
954                  * so the timer routine will probe
955                  * for later insertion.
956                  */
957                 sc->sfp_probe = true;
958                 error = 0;
959         } else if (error == IXGBE_ERR_SFP_NOT_SUPPORTED) {
960                 device_printf(dev, "Unsupported SFP+ module detected!\n");
961                 error = EIO;
962                 goto err_pci;
963         } else if (error) {
964                 device_printf(dev, "Hardware initialization failed\n");
965                 error = EIO;
966                 goto err_pci;
967         }
968
969         /* Make sure we have a good EEPROM before we read from it */
970         if (ixgbe_validate_eeprom_checksum(&sc->hw, NULL) < 0) {
971                 device_printf(dev, "The EEPROM Checksum Is Not Valid\n");
972                 error = EIO;
973                 goto err_pci;
974         }
975
976         error = ixgbe_start_hw(hw);
977         switch (error) {
978         case IXGBE_ERR_EEPROM_VERSION:
979                 device_printf(dev, "This device is a pre-production adapter/LOM.  Please be aware there may be issues associated with your hardware.\nIf you are experiencing problems please contact your Intel or hardware representative who provided you with this hardware.\n");
980                 break;
981         case IXGBE_ERR_SFP_NOT_SUPPORTED:
982                 device_printf(dev, "Unsupported SFP+ Module\n");
983                 error = EIO;
984                 goto err_pci;
985         case IXGBE_ERR_SFP_NOT_PRESENT:
986                 device_printf(dev, "No SFP+ Module found\n");
987                 /* falls thru */
988         default:
989                 break;
990         }
991
992         /* Most of the iflib initialization... */
993
994         iflib_set_mac(ctx, hw->mac.addr);
995         switch (sc->hw.mac.type) {
996         case ixgbe_mac_X550:
997         case ixgbe_mac_X550EM_x:
998         case ixgbe_mac_X550EM_a:
999                 scctx->isc_rss_table_size = 512;
1000                 scctx->isc_ntxqsets_max = scctx->isc_nrxqsets_max = 64;
1001                 break;
1002         default:
1003                 scctx->isc_rss_table_size = 128;
1004                 scctx->isc_ntxqsets_max = scctx->isc_nrxqsets_max = 16;
1005         }
1006
1007         /* Allow legacy interrupts */
1008         ixgbe_txrx.ift_legacy_intr = ixgbe_intr;
1009
1010         scctx->isc_txqsizes[0] =
1011             roundup2(scctx->isc_ntxd[0] * sizeof(union ixgbe_adv_tx_desc) +
1012             sizeof(u32), DBA_ALIGN),
1013         scctx->isc_rxqsizes[0] =
1014             roundup2(scctx->isc_nrxd[0] * sizeof(union ixgbe_adv_rx_desc),
1015             DBA_ALIGN);
1016
1017         /* XXX */
1018         scctx->isc_tx_csum_flags = CSUM_IP | CSUM_TCP | CSUM_UDP | CSUM_TSO |
1019             CSUM_IP6_TCP | CSUM_IP6_UDP | CSUM_IP6_TSO;
1020         if (sc->hw.mac.type == ixgbe_mac_82598EB) {
1021                 scctx->isc_tx_nsegments = IXGBE_82598_SCATTER;
1022         } else {
1023                 scctx->isc_tx_csum_flags |= CSUM_SCTP |CSUM_IP6_SCTP;
1024                 scctx->isc_tx_nsegments = IXGBE_82599_SCATTER;
1025         }
1026
1027         scctx->isc_msix_bar = pci_msix_table_bar(dev);
1028
1029         scctx->isc_tx_tso_segments_max = scctx->isc_tx_nsegments;
1030         scctx->isc_tx_tso_size_max = IXGBE_TSO_SIZE;
1031         scctx->isc_tx_tso_segsize_max = PAGE_SIZE;
1032
1033         scctx->isc_txrx = &ixgbe_txrx;
1034
1035         scctx->isc_capabilities = scctx->isc_capenable = IXGBE_CAPS;
1036
1037         return (0);
1038
1039 err_pci:
1040         ctrl_ext = IXGBE_READ_REG(&sc->hw, IXGBE_CTRL_EXT);
1041         ctrl_ext &= ~IXGBE_CTRL_EXT_DRV_LOAD;
1042         IXGBE_WRITE_REG(&sc->hw, IXGBE_CTRL_EXT, ctrl_ext);
1043         ixgbe_free_pci_resources(ctx);
1044
1045         return (error);
1046 } /* ixgbe_if_attach_pre */
1047
1048  /*********************************************************************
1049  * ixgbe_if_attach_post - Device initialization routine, part 2
1050  *
1051  *   Called during driver load, but after interrupts and
1052  *   resources have been allocated and configured.
1053  *   Sets up some data structures not relevant to iflib.
1054  *
1055  *   return 0 on success, positive on failure
1056  *********************************************************************/
1057 static int
1058 ixgbe_if_attach_post(if_ctx_t ctx)
1059 {
1060         device_t dev;
1061         struct ixgbe_softc  *sc;
1062         struct ixgbe_hw *hw;
1063         int             error = 0;
1064
1065         dev = iflib_get_dev(ctx);
1066         sc = iflib_get_softc(ctx);
1067         hw = &sc->hw;
1068
1069
1070         if (sc->intr_type == IFLIB_INTR_LEGACY &&
1071                 (sc->feat_cap & IXGBE_FEATURE_LEGACY_IRQ) == 0) {
1072                 device_printf(dev, "Device does not support legacy interrupts");
1073                 error = ENXIO;
1074                 goto err;
1075         }
1076
1077         /* Allocate multicast array memory. */
1078         sc->mta = malloc(sizeof(*sc->mta) *
1079                               MAX_NUM_MULTICAST_ADDRESSES, M_IXGBE, M_NOWAIT);
1080         if (sc->mta == NULL) {
1081                 device_printf(dev, "Can not allocate multicast setup array\n");
1082                 error = ENOMEM;
1083                 goto err;
1084         }
1085
1086         /* hw.ix defaults init */
1087         ixgbe_set_advertise(sc, ixgbe_advertise_speed);
1088
1089         /* Enable the optics for 82599 SFP+ fiber */
1090         ixgbe_enable_tx_laser(hw);
1091
1092         /* Enable power to the phy. */
1093         ixgbe_set_phy_power(hw, true);
1094
1095         ixgbe_initialize_iov(sc);
1096
1097         error = ixgbe_setup_interface(ctx);
1098         if (error) {
1099                 device_printf(dev, "Interface setup failed: %d\n", error);
1100                 goto err;
1101         }
1102
1103         ixgbe_if_update_admin_status(ctx);
1104
1105         /* Initialize statistics */
1106         ixgbe_update_stats_counters(sc);
1107         ixgbe_add_hw_stats(sc);
1108
1109         /* Check PCIE slot type/speed/width */
1110         ixgbe_get_slot_info(sc);
1111
1112         /*
1113          * Do time init and sysctl init here, but
1114          * only on the first port of a bypass sc.
1115          */
1116         ixgbe_bypass_init(sc);
1117
1118         /* Display NVM and Option ROM versions */
1119         ixgbe_print_fw_version(ctx);
1120
1121         /* Set an initial dmac value */
1122         sc->dmac = 0;
1123         /* Set initial advertised speeds (if applicable) */
1124         sc->advertise = ixgbe_get_default_advertise(sc);
1125
1126         if (sc->feat_cap & IXGBE_FEATURE_SRIOV)
1127                 ixgbe_define_iov_schemas(dev, &error);
1128
1129         /* Add sysctls */
1130         ixgbe_add_device_sysctls(ctx);
1131
1132         return (0);
1133 err:
1134         return (error);
1135 } /* ixgbe_if_attach_post */
1136
1137 /************************************************************************
1138  * ixgbe_check_wol_support
1139  *
1140  *   Checks whether the adapter's ports are capable of
1141  *   Wake On LAN by reading the adapter's NVM.
1142  *
1143  *   Sets each port's hw->wol_enabled value depending
1144  *   on the value read here.
1145  ************************************************************************/
1146 static void
1147 ixgbe_check_wol_support(struct ixgbe_softc *sc)
1148 {
1149         struct ixgbe_hw *hw = &sc->hw;
1150         u16             dev_caps = 0;
1151
1152         /* Find out WoL support for port */
1153         sc->wol_support = hw->wol_enabled = 0;
1154         ixgbe_get_device_caps(hw, &dev_caps);
1155         if ((dev_caps & IXGBE_DEVICE_CAPS_WOL_PORT0_1) ||
1156             ((dev_caps & IXGBE_DEVICE_CAPS_WOL_PORT0) &&
1157              hw->bus.func == 0))
1158                 sc->wol_support = hw->wol_enabled = 1;
1159
1160         /* Save initial wake up filter configuration */
1161         sc->wufc = IXGBE_READ_REG(hw, IXGBE_WUFC);
1162
1163         return;
1164 } /* ixgbe_check_wol_support */
1165
1166 /************************************************************************
1167  * ixgbe_setup_interface
1168  *
1169  *   Setup networking device structure and register an interface.
1170  ************************************************************************/
1171 static int
1172 ixgbe_setup_interface(if_ctx_t ctx)
1173 {
1174         struct ifnet   *ifp = iflib_get_ifp(ctx);
1175         struct ixgbe_softc *sc = iflib_get_softc(ctx);
1176
1177         INIT_DEBUGOUT("ixgbe_setup_interface: begin");
1178
1179         if_setbaudrate(ifp, IF_Gbps(10));
1180
1181         sc->max_frame_size = ifp->if_mtu + ETHER_HDR_LEN + ETHER_CRC_LEN;
1182
1183         sc->phy_layer = ixgbe_get_supported_physical_layer(&sc->hw);
1184
1185         ixgbe_add_media_types(ctx);
1186
1187         /* Autoselect media by default */
1188         ifmedia_set(sc->media, IFM_ETHER | IFM_AUTO);
1189
1190         return (0);
1191 } /* ixgbe_setup_interface */
1192
1193 /************************************************************************
1194  * ixgbe_if_get_counter
1195  ************************************************************************/
1196 static uint64_t
1197 ixgbe_if_get_counter(if_ctx_t ctx, ift_counter cnt)
1198 {
1199         struct ixgbe_softc *sc = iflib_get_softc(ctx);
1200         if_t           ifp = iflib_get_ifp(ctx);
1201
1202         switch (cnt) {
1203         case IFCOUNTER_IPACKETS:
1204                 return (sc->ipackets);
1205         case IFCOUNTER_OPACKETS:
1206                 return (sc->opackets);
1207         case IFCOUNTER_IBYTES:
1208                 return (sc->ibytes);
1209         case IFCOUNTER_OBYTES:
1210                 return (sc->obytes);
1211         case IFCOUNTER_IMCASTS:
1212                 return (sc->imcasts);
1213         case IFCOUNTER_OMCASTS:
1214                 return (sc->omcasts);
1215         case IFCOUNTER_COLLISIONS:
1216                 return (0);
1217         case IFCOUNTER_IQDROPS:
1218                 return (sc->iqdrops);
1219         case IFCOUNTER_OQDROPS:
1220                 return (0);
1221         case IFCOUNTER_IERRORS:
1222                 return (sc->ierrors);
1223         default:
1224                 return (if_get_counter_default(ifp, cnt));
1225         }
1226 } /* ixgbe_if_get_counter */
1227
1228 /************************************************************************
1229  * ixgbe_if_i2c_req
1230  ************************************************************************/
1231 static int
1232 ixgbe_if_i2c_req(if_ctx_t ctx, struct ifi2creq *req)
1233 {
1234         struct ixgbe_softc              *sc = iflib_get_softc(ctx);
1235         struct ixgbe_hw         *hw = &sc->hw;
1236         int                     i;
1237
1238
1239         if (hw->phy.ops.read_i2c_byte == NULL)
1240                 return (ENXIO);
1241         for (i = 0; i < req->len; i++)
1242                 hw->phy.ops.read_i2c_byte(hw, req->offset + i,
1243                     req->dev_addr, &req->data[i]);
1244         return (0);
1245 } /* ixgbe_if_i2c_req */
1246
1247 /* ixgbe_if_needs_restart - Tell iflib when the driver needs to be reinitialized
1248  * @ctx: iflib context
1249  * @event: event code to check
1250  *
1251  * Defaults to returning true for unknown events.
1252  *
1253  * @returns true if iflib needs to reinit the interface
1254  */
1255 static bool
1256 ixgbe_if_needs_restart(if_ctx_t ctx __unused, enum iflib_restart_event event)
1257 {
1258         switch (event) {
1259         case IFLIB_RESTART_VLAN_CONFIG:
1260                 return (false);
1261         default:
1262                 return (true);
1263         }
1264 }
1265
1266 /************************************************************************
1267  * ixgbe_add_media_types
1268  ************************************************************************/
1269 static void
1270 ixgbe_add_media_types(if_ctx_t ctx)
1271 {
1272         struct ixgbe_softc  *sc = iflib_get_softc(ctx);
1273         struct ixgbe_hw *hw = &sc->hw;
1274         device_t        dev = iflib_get_dev(ctx);
1275         u64             layer;
1276
1277         layer = sc->phy_layer = ixgbe_get_supported_physical_layer(hw);
1278
1279         /* Media types with matching FreeBSD media defines */
1280         if (layer & IXGBE_PHYSICAL_LAYER_10GBASE_T)
1281                 ifmedia_add(sc->media, IFM_ETHER | IFM_10G_T, 0, NULL);
1282         if (layer & IXGBE_PHYSICAL_LAYER_1000BASE_T)
1283                 ifmedia_add(sc->media, IFM_ETHER | IFM_1000_T, 0, NULL);
1284         if (layer & IXGBE_PHYSICAL_LAYER_100BASE_TX)
1285                 ifmedia_add(sc->media, IFM_ETHER | IFM_100_TX, 0, NULL);
1286         if (layer & IXGBE_PHYSICAL_LAYER_10BASE_T)
1287                 ifmedia_add(sc->media, IFM_ETHER | IFM_10_T, 0, NULL);
1288
1289         if (hw->mac.type == ixgbe_mac_X550) {
1290                 ifmedia_add(sc->media, IFM_ETHER | IFM_2500_T, 0, NULL);
1291                 ifmedia_add(sc->media, IFM_ETHER | IFM_5000_T, 0, NULL);
1292         }
1293
1294         if (layer & IXGBE_PHYSICAL_LAYER_SFP_PLUS_CU ||
1295             layer & IXGBE_PHYSICAL_LAYER_SFP_ACTIVE_DA)
1296                 ifmedia_add(sc->media, IFM_ETHER | IFM_10G_TWINAX, 0,
1297                     NULL);
1298
1299         if (layer & IXGBE_PHYSICAL_LAYER_10GBASE_LR) {
1300                 ifmedia_add(sc->media, IFM_ETHER | IFM_10G_LR, 0, NULL);
1301                 if (hw->phy.multispeed_fiber)
1302                         ifmedia_add(sc->media, IFM_ETHER | IFM_1000_LX, 0,
1303                             NULL);
1304         }
1305         if (layer & IXGBE_PHYSICAL_LAYER_10GBASE_SR) {
1306                 ifmedia_add(sc->media, IFM_ETHER | IFM_10G_SR, 0, NULL);
1307                 if (hw->phy.multispeed_fiber)
1308                         ifmedia_add(sc->media, IFM_ETHER | IFM_1000_SX, 0,
1309                             NULL);
1310         } else if (layer & IXGBE_PHYSICAL_LAYER_1000BASE_SX)
1311                 ifmedia_add(sc->media, IFM_ETHER | IFM_1000_SX, 0, NULL);
1312         if (layer & IXGBE_PHYSICAL_LAYER_10GBASE_CX4)
1313                 ifmedia_add(sc->media, IFM_ETHER | IFM_10G_CX4, 0, NULL);
1314
1315 #ifdef IFM_ETH_XTYPE
1316         if (layer & IXGBE_PHYSICAL_LAYER_10GBASE_KR)
1317                 ifmedia_add(sc->media, IFM_ETHER | IFM_10G_KR, 0, NULL);
1318         if (layer & IXGBE_PHYSICAL_LAYER_10GBASE_KX4)
1319                 ifmedia_add( sc->media, IFM_ETHER | IFM_10G_KX4, 0, NULL);
1320         if (layer & IXGBE_PHYSICAL_LAYER_1000BASE_KX)
1321                 ifmedia_add(sc->media, IFM_ETHER | IFM_1000_KX, 0, NULL);
1322         if (layer & IXGBE_PHYSICAL_LAYER_2500BASE_KX)
1323                 ifmedia_add(sc->media, IFM_ETHER | IFM_2500_KX, 0, NULL);
1324 #else
1325         if (layer & IXGBE_PHYSICAL_LAYER_10GBASE_KR) {
1326                 device_printf(dev, "Media supported: 10GbaseKR\n");
1327                 device_printf(dev, "10GbaseKR mapped to 10GbaseSR\n");
1328                 ifmedia_add(sc->media, IFM_ETHER | IFM_10G_SR, 0, NULL);
1329         }
1330         if (layer & IXGBE_PHYSICAL_LAYER_10GBASE_KX4) {
1331                 device_printf(dev, "Media supported: 10GbaseKX4\n");
1332                 device_printf(dev, "10GbaseKX4 mapped to 10GbaseCX4\n");
1333                 ifmedia_add(sc->media, IFM_ETHER | IFM_10G_CX4, 0, NULL);
1334         }
1335         if (layer & IXGBE_PHYSICAL_LAYER_1000BASE_KX) {
1336                 device_printf(dev, "Media supported: 1000baseKX\n");
1337                 device_printf(dev, "1000baseKX mapped to 1000baseCX\n");
1338                 ifmedia_add(sc->media, IFM_ETHER | IFM_1000_CX, 0, NULL);
1339         }
1340         if (layer & IXGBE_PHYSICAL_LAYER_2500BASE_KX) {
1341                 device_printf(dev, "Media supported: 2500baseKX\n");
1342                 device_printf(dev, "2500baseKX mapped to 2500baseSX\n");
1343                 ifmedia_add(sc->media, IFM_ETHER | IFM_2500_SX, 0, NULL);
1344         }
1345 #endif
1346         if (layer & IXGBE_PHYSICAL_LAYER_1000BASE_BX)
1347                 device_printf(dev, "Media supported: 1000baseBX\n");
1348
1349         if (hw->device_id == IXGBE_DEV_ID_82598AT) {
1350                 ifmedia_add(sc->media, IFM_ETHER | IFM_1000_T | IFM_FDX,
1351                     0, NULL);
1352                 ifmedia_add(sc->media, IFM_ETHER | IFM_1000_T, 0, NULL);
1353         }
1354
1355         ifmedia_add(sc->media, IFM_ETHER | IFM_AUTO, 0, NULL);
1356 } /* ixgbe_add_media_types */
1357
1358 /************************************************************************
1359  * ixgbe_is_sfp
1360  ************************************************************************/
1361 static inline bool
1362 ixgbe_is_sfp(struct ixgbe_hw *hw)
1363 {
1364         switch (hw->mac.type) {
1365         case ixgbe_mac_82598EB:
1366                 if (hw->phy.type == ixgbe_phy_nl)
1367                         return (true);
1368                 return (false);
1369         case ixgbe_mac_82599EB:
1370                 switch (hw->mac.ops.get_media_type(hw)) {
1371                 case ixgbe_media_type_fiber:
1372                 case ixgbe_media_type_fiber_qsfp:
1373                         return (true);
1374                 default:
1375                         return (false);
1376                 }
1377         case ixgbe_mac_X550EM_x:
1378         case ixgbe_mac_X550EM_a:
1379                 if (hw->mac.ops.get_media_type(hw) == ixgbe_media_type_fiber)
1380                         return (true);
1381                 return (false);
1382         default:
1383                 return (false);
1384         }
1385 } /* ixgbe_is_sfp */
1386
1387 /************************************************************************
1388  * ixgbe_config_link
1389  ************************************************************************/
1390 static void
1391 ixgbe_config_link(if_ctx_t ctx)
1392 {
1393         struct ixgbe_softc  *sc = iflib_get_softc(ctx);
1394         struct ixgbe_hw *hw = &sc->hw;
1395         u32             autoneg, err = 0;
1396         bool            sfp, negotiate;
1397
1398         sfp = ixgbe_is_sfp(hw);
1399
1400         if (sfp) {
1401                 sc->task_requests |= IXGBE_REQUEST_TASK_MOD;
1402                 iflib_admin_intr_deferred(ctx);
1403         } else {
1404                 if (hw->mac.ops.check_link)
1405                         err = ixgbe_check_link(hw, &sc->link_speed,
1406                             &sc->link_up, false);
1407                 if (err)
1408                         return;
1409                 autoneg = hw->phy.autoneg_advertised;
1410                 if ((!autoneg) && (hw->mac.ops.get_link_capabilities))
1411                         err = hw->mac.ops.get_link_capabilities(hw, &autoneg,
1412                             &negotiate);
1413                 if (err)
1414                         return;
1415
1416                 if (hw->mac.type == ixgbe_mac_X550 &&
1417                     hw->phy.autoneg_advertised == 0) {
1418                         /*
1419                          * 2.5G and 5G autonegotiation speeds on X550
1420                          * are disabled by default due to reported
1421                          * interoperability issues with some switches.
1422                          *
1423                          * The second condition checks if any operations
1424                          * involving setting autonegotiation speeds have
1425                          * been performed prior to this ixgbe_config_link()
1426                          * call.
1427                          *
1428                          * If hw->phy.autoneg_advertised does not
1429                          * equal 0, this means that the user might have
1430                          * set autonegotiation speeds via the sysctl
1431                          * before bringing the interface up. In this
1432                          * case, we should not disable 2.5G and 5G
1433                          * since that speeds might be selected by the
1434                          * user.
1435                          *
1436                          * Otherwise (i.e. if hw->phy.autoneg_advertised
1437                          * is set to 0), it is the first time we set
1438                          * autonegotiation preferences and the default
1439                          * set of speeds should exclude 2.5G and 5G.
1440                          */
1441                         autoneg &= ~(IXGBE_LINK_SPEED_2_5GB_FULL |
1442                             IXGBE_LINK_SPEED_5GB_FULL);
1443                 }
1444
1445                 if (hw->mac.ops.setup_link)
1446                         err = hw->mac.ops.setup_link(hw, autoneg,
1447                             sc->link_up);
1448         }
1449 } /* ixgbe_config_link */
1450
1451 /************************************************************************
1452  * ixgbe_update_stats_counters - Update board statistics counters.
1453  ************************************************************************/
1454 static void
1455 ixgbe_update_stats_counters(struct ixgbe_softc *sc)
1456 {
1457         struct ixgbe_hw       *hw = &sc->hw;
1458         struct ixgbe_hw_stats *stats = &sc->stats.pf;
1459         u32                   missed_rx = 0, bprc, lxon, lxoff, total;
1460         u32                   lxoffrxc;
1461         u64                   total_missed_rx = 0;
1462
1463         stats->crcerrs += IXGBE_READ_REG(hw, IXGBE_CRCERRS);
1464         stats->illerrc += IXGBE_READ_REG(hw, IXGBE_ILLERRC);
1465         stats->errbc += IXGBE_READ_REG(hw, IXGBE_ERRBC);
1466         stats->mspdc += IXGBE_READ_REG(hw, IXGBE_MSPDC);
1467         stats->mpc[0] += IXGBE_READ_REG(hw, IXGBE_MPC(0));
1468
1469         for (int i = 0; i < 16; i++) {
1470                 stats->qprc[i] += IXGBE_READ_REG(hw, IXGBE_QPRC(i));
1471                 stats->qptc[i] += IXGBE_READ_REG(hw, IXGBE_QPTC(i));
1472                 stats->qprdc[i] += IXGBE_READ_REG(hw, IXGBE_QPRDC(i));
1473         }
1474         stats->mlfc += IXGBE_READ_REG(hw, IXGBE_MLFC);
1475         stats->mrfc += IXGBE_READ_REG(hw, IXGBE_MRFC);
1476         stats->rlec += IXGBE_READ_REG(hw, IXGBE_RLEC);
1477
1478         /* Hardware workaround, gprc counts missed packets */
1479         stats->gprc += IXGBE_READ_REG(hw, IXGBE_GPRC);
1480         stats->gprc -= missed_rx;
1481
1482         if (hw->mac.type != ixgbe_mac_82598EB) {
1483                 stats->gorc += IXGBE_READ_REG(hw, IXGBE_GORCL) +
1484                     ((u64)IXGBE_READ_REG(hw, IXGBE_GORCH) << 32);
1485                 stats->gotc += IXGBE_READ_REG(hw, IXGBE_GOTCL) +
1486                     ((u64)IXGBE_READ_REG(hw, IXGBE_GOTCH) << 32);
1487                 stats->tor += IXGBE_READ_REG(hw, IXGBE_TORL) +
1488                     ((u64)IXGBE_READ_REG(hw, IXGBE_TORH) << 32);
1489                 stats->lxonrxc += IXGBE_READ_REG(hw, IXGBE_LXONRXCNT);
1490                 lxoffrxc = IXGBE_READ_REG(hw, IXGBE_LXOFFRXCNT);
1491                 stats->lxoffrxc += lxoffrxc;
1492         } else {
1493                 stats->lxonrxc += IXGBE_READ_REG(hw, IXGBE_LXONRXC);
1494                 lxoffrxc = IXGBE_READ_REG(hw, IXGBE_LXOFFRXC);
1495                 stats->lxoffrxc += lxoffrxc;
1496                 /* 82598 only has a counter in the high register */
1497                 stats->gorc += IXGBE_READ_REG(hw, IXGBE_GORCH);
1498                 stats->gotc += IXGBE_READ_REG(hw, IXGBE_GOTCH);
1499                 stats->tor += IXGBE_READ_REG(hw, IXGBE_TORH);
1500         }
1501
1502         /*
1503          * For watchdog management we need to know if we have been paused
1504          * during the last interval, so capture that here.
1505         */
1506         if (lxoffrxc)
1507                 sc->shared->isc_pause_frames = 1;
1508
1509         /*
1510          * Workaround: mprc hardware is incorrectly counting
1511          * broadcasts, so for now we subtract those.
1512          */
1513         bprc = IXGBE_READ_REG(hw, IXGBE_BPRC);
1514         stats->bprc += bprc;
1515         stats->mprc += IXGBE_READ_REG(hw, IXGBE_MPRC);
1516         if (hw->mac.type == ixgbe_mac_82598EB)
1517                 stats->mprc -= bprc;
1518
1519         stats->prc64 += IXGBE_READ_REG(hw, IXGBE_PRC64);
1520         stats->prc127 += IXGBE_READ_REG(hw, IXGBE_PRC127);
1521         stats->prc255 += IXGBE_READ_REG(hw, IXGBE_PRC255);
1522         stats->prc511 += IXGBE_READ_REG(hw, IXGBE_PRC511);
1523         stats->prc1023 += IXGBE_READ_REG(hw, IXGBE_PRC1023);
1524         stats->prc1522 += IXGBE_READ_REG(hw, IXGBE_PRC1522);
1525
1526         lxon = IXGBE_READ_REG(hw, IXGBE_LXONTXC);
1527         stats->lxontxc += lxon;
1528         lxoff = IXGBE_READ_REG(hw, IXGBE_LXOFFTXC);
1529         stats->lxofftxc += lxoff;
1530         total = lxon + lxoff;
1531
1532         stats->gptc += IXGBE_READ_REG(hw, IXGBE_GPTC);
1533         stats->mptc += IXGBE_READ_REG(hw, IXGBE_MPTC);
1534         stats->ptc64 += IXGBE_READ_REG(hw, IXGBE_PTC64);
1535         stats->gptc -= total;
1536         stats->mptc -= total;
1537         stats->ptc64 -= total;
1538         stats->gotc -= total * ETHER_MIN_LEN;
1539
1540         stats->ruc += IXGBE_READ_REG(hw, IXGBE_RUC);
1541         stats->rfc += IXGBE_READ_REG(hw, IXGBE_RFC);
1542         stats->roc += IXGBE_READ_REG(hw, IXGBE_ROC);
1543         stats->rjc += IXGBE_READ_REG(hw, IXGBE_RJC);
1544         stats->mngprc += IXGBE_READ_REG(hw, IXGBE_MNGPRC);
1545         stats->mngpdc += IXGBE_READ_REG(hw, IXGBE_MNGPDC);
1546         stats->mngptc += IXGBE_READ_REG(hw, IXGBE_MNGPTC);
1547         stats->tpr += IXGBE_READ_REG(hw, IXGBE_TPR);
1548         stats->tpt += IXGBE_READ_REG(hw, IXGBE_TPT);
1549         stats->ptc127 += IXGBE_READ_REG(hw, IXGBE_PTC127);
1550         stats->ptc255 += IXGBE_READ_REG(hw, IXGBE_PTC255);
1551         stats->ptc511 += IXGBE_READ_REG(hw, IXGBE_PTC511);
1552         stats->ptc1023 += IXGBE_READ_REG(hw, IXGBE_PTC1023);
1553         stats->ptc1522 += IXGBE_READ_REG(hw, IXGBE_PTC1522);
1554         stats->bptc += IXGBE_READ_REG(hw, IXGBE_BPTC);
1555         stats->xec += IXGBE_READ_REG(hw, IXGBE_XEC);
1556         stats->fccrc += IXGBE_READ_REG(hw, IXGBE_FCCRC);
1557         stats->fclast += IXGBE_READ_REG(hw, IXGBE_FCLAST);
1558         /* Only read FCOE on 82599 */
1559         if (hw->mac.type != ixgbe_mac_82598EB) {
1560                 stats->fcoerpdc += IXGBE_READ_REG(hw, IXGBE_FCOERPDC);
1561                 stats->fcoeprc += IXGBE_READ_REG(hw, IXGBE_FCOEPRC);
1562                 stats->fcoeptc += IXGBE_READ_REG(hw, IXGBE_FCOEPTC);
1563                 stats->fcoedwrc += IXGBE_READ_REG(hw, IXGBE_FCOEDWRC);
1564                 stats->fcoedwtc += IXGBE_READ_REG(hw, IXGBE_FCOEDWTC);
1565         }
1566
1567         /* Fill out the OS statistics structure */
1568         IXGBE_SET_IPACKETS(sc, stats->gprc);
1569         IXGBE_SET_OPACKETS(sc, stats->gptc);
1570         IXGBE_SET_IBYTES(sc, stats->gorc);
1571         IXGBE_SET_OBYTES(sc, stats->gotc);
1572         IXGBE_SET_IMCASTS(sc, stats->mprc);
1573         IXGBE_SET_OMCASTS(sc, stats->mptc);
1574         IXGBE_SET_COLLISIONS(sc, 0);
1575         IXGBE_SET_IQDROPS(sc, total_missed_rx);
1576
1577         /*
1578          * Aggregate following types of errors as RX errors:
1579          * - CRC error count,
1580          * - illegal byte error count,
1581          * - checksum error count,
1582          * - missed packets count,
1583          * - length error count,
1584          * - undersized packets count,
1585          * - fragmented packets count,
1586          * - oversized packets count,
1587          * - jabber count.
1588          */
1589         IXGBE_SET_IERRORS(sc, stats->crcerrs + stats->illerrc + stats->xec +
1590             stats->mpc[0] + stats->rlec + stats->ruc + stats->rfc + stats->roc +
1591             stats->rjc);
1592 } /* ixgbe_update_stats_counters */
1593
1594 /************************************************************************
1595  * ixgbe_add_hw_stats
1596  *
1597  *   Add sysctl variables, one per statistic, to the system.
1598  ************************************************************************/
1599 static void
1600 ixgbe_add_hw_stats(struct ixgbe_softc *sc)
1601 {
1602         device_t               dev = iflib_get_dev(sc->ctx);
1603         struct ix_rx_queue     *rx_que;
1604         struct ix_tx_queue     *tx_que;
1605         struct sysctl_ctx_list *ctx = device_get_sysctl_ctx(dev);
1606         struct sysctl_oid      *tree = device_get_sysctl_tree(dev);
1607         struct sysctl_oid_list *child = SYSCTL_CHILDREN(tree);
1608         struct ixgbe_hw_stats  *stats = &sc->stats.pf;
1609         struct sysctl_oid      *stat_node, *queue_node;
1610         struct sysctl_oid_list *stat_list, *queue_list;
1611         int                    i;
1612
1613 #define QUEUE_NAME_LEN 32
1614         char                   namebuf[QUEUE_NAME_LEN];
1615
1616         /* Driver Statistics */
1617         SYSCTL_ADD_ULONG(ctx, child, OID_AUTO, "dropped",
1618             CTLFLAG_RD, &sc->dropped_pkts, "Driver dropped packets");
1619         SYSCTL_ADD_ULONG(ctx, child, OID_AUTO, "watchdog_events",
1620             CTLFLAG_RD, &sc->watchdog_events, "Watchdog timeouts");
1621         SYSCTL_ADD_ULONG(ctx, child, OID_AUTO, "link_irq",
1622             CTLFLAG_RD, &sc->link_irq, "Link MSI-X IRQ Handled");
1623
1624         for (i = 0, tx_que = sc->tx_queues; i < sc->num_tx_queues; i++, tx_que++) {
1625                 struct tx_ring *txr = &tx_que->txr;
1626                 snprintf(namebuf, QUEUE_NAME_LEN, "queue%d", i);
1627                 queue_node = SYSCTL_ADD_NODE(ctx, child, OID_AUTO, namebuf,
1628                     CTLFLAG_RD | CTLFLAG_MPSAFE, NULL, "Queue Name");
1629                 queue_list = SYSCTL_CHILDREN(queue_node);
1630
1631                 SYSCTL_ADD_PROC(ctx, queue_list, OID_AUTO, "txd_head",
1632                     CTLTYPE_UINT | CTLFLAG_RD | CTLFLAG_NEEDGIANT, txr, 0,
1633                     ixgbe_sysctl_tdh_handler, "IU", "Transmit Descriptor Head");
1634                 SYSCTL_ADD_PROC(ctx, queue_list, OID_AUTO, "txd_tail",
1635                     CTLTYPE_UINT | CTLFLAG_RD | CTLFLAG_NEEDGIANT, txr, 0,
1636                     ixgbe_sysctl_tdt_handler, "IU", "Transmit Descriptor Tail");
1637                 SYSCTL_ADD_UQUAD(ctx, queue_list, OID_AUTO, "tso_tx",
1638                     CTLFLAG_RD, &txr->tso_tx, "TSO");
1639                 SYSCTL_ADD_UQUAD(ctx, queue_list, OID_AUTO, "tx_packets",
1640                     CTLFLAG_RD, &txr->total_packets,
1641                     "Queue Packets Transmitted");
1642         }
1643
1644         for (i = 0, rx_que = sc->rx_queues; i < sc->num_rx_queues; i++, rx_que++) {
1645                 struct rx_ring *rxr = &rx_que->rxr;
1646                 snprintf(namebuf, QUEUE_NAME_LEN, "queue%d", i);
1647                 queue_node = SYSCTL_ADD_NODE(ctx, child, OID_AUTO, namebuf,
1648                     CTLFLAG_RD | CTLFLAG_MPSAFE, NULL, "Queue Name");
1649                 queue_list = SYSCTL_CHILDREN(queue_node);
1650
1651                 SYSCTL_ADD_PROC(ctx, queue_list, OID_AUTO, "interrupt_rate",
1652                     CTLTYPE_UINT | CTLFLAG_RW | CTLFLAG_NEEDGIANT,
1653                     &sc->rx_queues[i], 0,
1654                     ixgbe_sysctl_interrupt_rate_handler, "IU",
1655                     "Interrupt Rate");
1656                 SYSCTL_ADD_UQUAD(ctx, queue_list, OID_AUTO, "irqs",
1657                     CTLFLAG_RD, &(sc->rx_queues[i].irqs),
1658                     "irqs on this queue");
1659                 SYSCTL_ADD_PROC(ctx, queue_list, OID_AUTO, "rxd_head",
1660                     CTLTYPE_UINT | CTLFLAG_RD | CTLFLAG_NEEDGIANT, rxr, 0,
1661                     ixgbe_sysctl_rdh_handler, "IU", "Receive Descriptor Head");
1662                 SYSCTL_ADD_PROC(ctx, queue_list, OID_AUTO, "rxd_tail",
1663                     CTLTYPE_UINT | CTLFLAG_RD | CTLFLAG_NEEDGIANT, rxr, 0,
1664                     ixgbe_sysctl_rdt_handler, "IU", "Receive Descriptor Tail");
1665                 SYSCTL_ADD_UQUAD(ctx, queue_list, OID_AUTO, "rx_packets",
1666                     CTLFLAG_RD, &rxr->rx_packets, "Queue Packets Received");
1667                 SYSCTL_ADD_UQUAD(ctx, queue_list, OID_AUTO, "rx_bytes",
1668                     CTLFLAG_RD, &rxr->rx_bytes, "Queue Bytes Received");
1669                 SYSCTL_ADD_UQUAD(ctx, queue_list, OID_AUTO, "rx_copies",
1670                     CTLFLAG_RD, &rxr->rx_copies, "Copied RX Frames");
1671                 SYSCTL_ADD_UQUAD(ctx, queue_list, OID_AUTO, "rx_discarded",
1672                     CTLFLAG_RD, &rxr->rx_discarded, "Discarded RX packets");
1673         }
1674
1675         /* MAC stats get their own sub node */
1676
1677         stat_node = SYSCTL_ADD_NODE(ctx, child, OID_AUTO, "mac_stats",
1678             CTLFLAG_RD | CTLFLAG_MPSAFE, NULL, "MAC Statistics");
1679         stat_list = SYSCTL_CHILDREN(stat_node);
1680
1681         SYSCTL_ADD_UQUAD(ctx, stat_list, OID_AUTO, "rx_errs",
1682             CTLFLAG_RD, &sc->ierrors, IXGBE_SYSCTL_DESC_RX_ERRS);
1683         SYSCTL_ADD_UQUAD(ctx, stat_list, OID_AUTO, "crc_errs",
1684             CTLFLAG_RD, &stats->crcerrs, "CRC Errors");
1685         SYSCTL_ADD_UQUAD(ctx, stat_list, OID_AUTO, "ill_errs",
1686             CTLFLAG_RD, &stats->illerrc, "Illegal Byte Errors");
1687         SYSCTL_ADD_UQUAD(ctx, stat_list, OID_AUTO, "byte_errs",
1688             CTLFLAG_RD, &stats->errbc, "Byte Errors");
1689         SYSCTL_ADD_UQUAD(ctx, stat_list, OID_AUTO, "short_discards",
1690             CTLFLAG_RD, &stats->mspdc, "MAC Short Packets Discarded");
1691         SYSCTL_ADD_UQUAD(ctx, stat_list, OID_AUTO, "local_faults",
1692             CTLFLAG_RD, &stats->mlfc, "MAC Local Faults");
1693         SYSCTL_ADD_UQUAD(ctx, stat_list, OID_AUTO, "remote_faults",
1694             CTLFLAG_RD, &stats->mrfc, "MAC Remote Faults");
1695         SYSCTL_ADD_UQUAD(ctx, stat_list, OID_AUTO, "rec_len_errs",
1696             CTLFLAG_RD, &stats->rlec, "Receive Length Errors");
1697         SYSCTL_ADD_UQUAD(ctx, stat_list, OID_AUTO, "rx_missed_packets",
1698             CTLFLAG_RD, &stats->mpc[0], "RX Missed Packet Count");
1699
1700         /* Flow Control stats */
1701         SYSCTL_ADD_UQUAD(ctx, stat_list, OID_AUTO, "xon_txd",
1702             CTLFLAG_RD, &stats->lxontxc, "Link XON Transmitted");
1703         SYSCTL_ADD_UQUAD(ctx, stat_list, OID_AUTO, "xon_recvd",
1704             CTLFLAG_RD, &stats->lxonrxc, "Link XON Received");
1705         SYSCTL_ADD_UQUAD(ctx, stat_list, OID_AUTO, "xoff_txd",
1706             CTLFLAG_RD, &stats->lxofftxc, "Link XOFF Transmitted");
1707         SYSCTL_ADD_UQUAD(ctx, stat_list, OID_AUTO, "xoff_recvd",
1708             CTLFLAG_RD, &stats->lxoffrxc, "Link XOFF Received");
1709
1710         /* Packet Reception Stats */
1711         SYSCTL_ADD_UQUAD(ctx, stat_list, OID_AUTO, "total_octets_rcvd",
1712             CTLFLAG_RD, &stats->tor, "Total Octets Received");
1713         SYSCTL_ADD_UQUAD(ctx, stat_list, OID_AUTO, "good_octets_rcvd",
1714             CTLFLAG_RD, &stats->gorc, "Good Octets Received");
1715         SYSCTL_ADD_UQUAD(ctx, stat_list, OID_AUTO, "total_pkts_rcvd",
1716             CTLFLAG_RD, &stats->tpr, "Total Packets Received");
1717         SYSCTL_ADD_UQUAD(ctx, stat_list, OID_AUTO, "good_pkts_rcvd",
1718             CTLFLAG_RD, &stats->gprc, "Good Packets Received");
1719         SYSCTL_ADD_UQUAD(ctx, stat_list, OID_AUTO, "mcast_pkts_rcvd",
1720             CTLFLAG_RD, &stats->mprc, "Multicast Packets Received");
1721         SYSCTL_ADD_UQUAD(ctx, stat_list, OID_AUTO, "bcast_pkts_rcvd",
1722             CTLFLAG_RD, &stats->bprc, "Broadcast Packets Received");
1723         SYSCTL_ADD_UQUAD(ctx, stat_list, OID_AUTO, "rx_frames_64",
1724             CTLFLAG_RD, &stats->prc64, "64 byte frames received ");
1725         SYSCTL_ADD_UQUAD(ctx, stat_list, OID_AUTO, "rx_frames_65_127",
1726             CTLFLAG_RD, &stats->prc127, "65-127 byte frames received");
1727         SYSCTL_ADD_UQUAD(ctx, stat_list, OID_AUTO, "rx_frames_128_255",
1728             CTLFLAG_RD, &stats->prc255, "128-255 byte frames received");
1729         SYSCTL_ADD_UQUAD(ctx, stat_list, OID_AUTO, "rx_frames_256_511",
1730             CTLFLAG_RD, &stats->prc511, "256-511 byte frames received");
1731         SYSCTL_ADD_UQUAD(ctx, stat_list, OID_AUTO, "rx_frames_512_1023",
1732             CTLFLAG_RD, &stats->prc1023, "512-1023 byte frames received");
1733         SYSCTL_ADD_UQUAD(ctx, stat_list, OID_AUTO, "rx_frames_1024_1522",
1734             CTLFLAG_RD, &stats->prc1522, "1023-1522 byte frames received");
1735         SYSCTL_ADD_UQUAD(ctx, stat_list, OID_AUTO, "recv_undersized",
1736             CTLFLAG_RD, &stats->ruc, "Receive Undersized");
1737         SYSCTL_ADD_UQUAD(ctx, stat_list, OID_AUTO, "recv_fragmented",
1738             CTLFLAG_RD, &stats->rfc, "Fragmented Packets Received ");
1739         SYSCTL_ADD_UQUAD(ctx, stat_list, OID_AUTO, "recv_oversized",
1740             CTLFLAG_RD, &stats->roc, "Oversized Packets Received");
1741         SYSCTL_ADD_UQUAD(ctx, stat_list, OID_AUTO, "recv_jabberd",
1742             CTLFLAG_RD, &stats->rjc, "Received Jabber");
1743         SYSCTL_ADD_UQUAD(ctx, stat_list, OID_AUTO, "management_pkts_rcvd",
1744             CTLFLAG_RD, &stats->mngprc, "Management Packets Received");
1745         SYSCTL_ADD_UQUAD(ctx, stat_list, OID_AUTO, "management_pkts_drpd",
1746             CTLFLAG_RD, &stats->mngptc, "Management Packets Dropped");
1747         SYSCTL_ADD_UQUAD(ctx, stat_list, OID_AUTO, "checksum_errs",
1748             CTLFLAG_RD, &stats->xec, "Checksum Errors");
1749
1750         /* Packet Transmission Stats */
1751         SYSCTL_ADD_UQUAD(ctx, stat_list, OID_AUTO, "good_octets_txd",
1752             CTLFLAG_RD, &stats->gotc, "Good Octets Transmitted");
1753         SYSCTL_ADD_UQUAD(ctx, stat_list, OID_AUTO, "total_pkts_txd",
1754             CTLFLAG_RD, &stats->tpt, "Total Packets Transmitted");
1755         SYSCTL_ADD_UQUAD(ctx, stat_list, OID_AUTO, "good_pkts_txd",
1756             CTLFLAG_RD, &stats->gptc, "Good Packets Transmitted");
1757         SYSCTL_ADD_UQUAD(ctx, stat_list, OID_AUTO, "bcast_pkts_txd",
1758             CTLFLAG_RD, &stats->bptc, "Broadcast Packets Transmitted");
1759         SYSCTL_ADD_UQUAD(ctx, stat_list, OID_AUTO, "mcast_pkts_txd",
1760             CTLFLAG_RD, &stats->mptc, "Multicast Packets Transmitted");
1761         SYSCTL_ADD_UQUAD(ctx, stat_list, OID_AUTO, "management_pkts_txd",
1762             CTLFLAG_RD, &stats->mngptc, "Management Packets Transmitted");
1763         SYSCTL_ADD_UQUAD(ctx, stat_list, OID_AUTO, "tx_frames_64",
1764             CTLFLAG_RD, &stats->ptc64, "64 byte frames transmitted ");
1765         SYSCTL_ADD_UQUAD(ctx, stat_list, OID_AUTO, "tx_frames_65_127",
1766             CTLFLAG_RD, &stats->ptc127, "65-127 byte frames transmitted");
1767         SYSCTL_ADD_UQUAD(ctx, stat_list, OID_AUTO, "tx_frames_128_255",
1768             CTLFLAG_RD, &stats->ptc255, "128-255 byte frames transmitted");
1769         SYSCTL_ADD_UQUAD(ctx, stat_list, OID_AUTO, "tx_frames_256_511",
1770             CTLFLAG_RD, &stats->ptc511, "256-511 byte frames transmitted");
1771         SYSCTL_ADD_UQUAD(ctx, stat_list, OID_AUTO, "tx_frames_512_1023",
1772             CTLFLAG_RD, &stats->ptc1023, "512-1023 byte frames transmitted");
1773         SYSCTL_ADD_UQUAD(ctx, stat_list, OID_AUTO, "tx_frames_1024_1522",
1774             CTLFLAG_RD, &stats->ptc1522, "1024-1522 byte frames transmitted");
1775 } /* ixgbe_add_hw_stats */
1776
1777 /************************************************************************
1778  * ixgbe_sysctl_tdh_handler - Transmit Descriptor Head handler function
1779  *
1780  *   Retrieves the TDH value from the hardware
1781  ************************************************************************/
1782 static int
1783 ixgbe_sysctl_tdh_handler(SYSCTL_HANDLER_ARGS)
1784 {
1785         struct tx_ring *txr = ((struct tx_ring *)oidp->oid_arg1);
1786         int            error;
1787         unsigned int   val;
1788
1789         if (!txr)
1790                 return (0);
1791
1792         val = IXGBE_READ_REG(&txr->sc->hw, IXGBE_TDH(txr->me));
1793         error = sysctl_handle_int(oidp, &val, 0, req);
1794         if (error || !req->newptr)
1795                 return error;
1796
1797         return (0);
1798 } /* ixgbe_sysctl_tdh_handler */
1799
1800 /************************************************************************
1801  * ixgbe_sysctl_tdt_handler - Transmit Descriptor Tail handler function
1802  *
1803  *   Retrieves the TDT value from the hardware
1804  ************************************************************************/
1805 static int
1806 ixgbe_sysctl_tdt_handler(SYSCTL_HANDLER_ARGS)
1807 {
1808         struct tx_ring *txr = ((struct tx_ring *)oidp->oid_arg1);
1809         int            error;
1810         unsigned int   val;
1811
1812         if (!txr)
1813                 return (0);
1814
1815         val = IXGBE_READ_REG(&txr->sc->hw, IXGBE_TDT(txr->me));
1816         error = sysctl_handle_int(oidp, &val, 0, req);
1817         if (error || !req->newptr)
1818                 return error;
1819
1820         return (0);
1821 } /* ixgbe_sysctl_tdt_handler */
1822
1823 /************************************************************************
1824  * ixgbe_sysctl_rdh_handler - Receive Descriptor Head handler function
1825  *
1826  *   Retrieves the RDH value from the hardware
1827  ************************************************************************/
1828 static int
1829 ixgbe_sysctl_rdh_handler(SYSCTL_HANDLER_ARGS)
1830 {
1831         struct rx_ring *rxr = ((struct rx_ring *)oidp->oid_arg1);
1832         int            error;
1833         unsigned int   val;
1834
1835         if (!rxr)
1836                 return (0);
1837
1838         val = IXGBE_READ_REG(&rxr->sc->hw, IXGBE_RDH(rxr->me));
1839         error = sysctl_handle_int(oidp, &val, 0, req);
1840         if (error || !req->newptr)
1841                 return error;
1842
1843         return (0);
1844 } /* ixgbe_sysctl_rdh_handler */
1845
1846 /************************************************************************
1847  * ixgbe_sysctl_rdt_handler - Receive Descriptor Tail handler function
1848  *
1849  *   Retrieves the RDT value from the hardware
1850  ************************************************************************/
1851 static int
1852 ixgbe_sysctl_rdt_handler(SYSCTL_HANDLER_ARGS)
1853 {
1854         struct rx_ring *rxr = ((struct rx_ring *)oidp->oid_arg1);
1855         int            error;
1856         unsigned int   val;
1857
1858         if (!rxr)
1859                 return (0);
1860
1861         val = IXGBE_READ_REG(&rxr->sc->hw, IXGBE_RDT(rxr->me));
1862         error = sysctl_handle_int(oidp, &val, 0, req);
1863         if (error || !req->newptr)
1864                 return error;
1865
1866         return (0);
1867 } /* ixgbe_sysctl_rdt_handler */
1868
1869 /************************************************************************
1870  * ixgbe_if_vlan_register
1871  *
1872  *   Run via vlan config EVENT, it enables us to use the
1873  *   HW Filter table since we can get the vlan id. This
1874  *   just creates the entry in the soft version of the
1875  *   VFTA, init will repopulate the real table.
1876  ************************************************************************/
1877 static void
1878 ixgbe_if_vlan_register(if_ctx_t ctx, u16 vtag)
1879 {
1880         struct ixgbe_softc *sc = iflib_get_softc(ctx);
1881         u16            index, bit;
1882
1883         index = (vtag >> 5) & 0x7F;
1884         bit = vtag & 0x1F;
1885         sc->shadow_vfta[index] |= (1 << bit);
1886         ++sc->num_vlans;
1887         ixgbe_setup_vlan_hw_support(ctx);
1888 } /* ixgbe_if_vlan_register */
1889
1890 /************************************************************************
1891  * ixgbe_if_vlan_unregister
1892  *
1893  *   Run via vlan unconfig EVENT, remove our entry in the soft vfta.
1894  ************************************************************************/
1895 static void
1896 ixgbe_if_vlan_unregister(if_ctx_t ctx, u16 vtag)
1897 {
1898         struct ixgbe_softc *sc = iflib_get_softc(ctx);
1899         u16            index, bit;
1900
1901         index = (vtag >> 5) & 0x7F;
1902         bit = vtag & 0x1F;
1903         sc->shadow_vfta[index] &= ~(1 << bit);
1904         --sc->num_vlans;
1905         /* Re-init to load the changes */
1906         ixgbe_setup_vlan_hw_support(ctx);
1907 } /* ixgbe_if_vlan_unregister */
1908
1909 /************************************************************************
1910  * ixgbe_setup_vlan_hw_support
1911  ************************************************************************/
1912 static void
1913 ixgbe_setup_vlan_hw_support(if_ctx_t ctx)
1914 {
1915         struct ifnet    *ifp = iflib_get_ifp(ctx);
1916         struct ixgbe_softc  *sc = iflib_get_softc(ctx);
1917         struct ixgbe_hw *hw = &sc->hw;
1918         struct rx_ring  *rxr;
1919         int             i;
1920         u32             ctrl;
1921
1922
1923         /*
1924          * We get here thru init_locked, meaning
1925          * a soft reset, this has already cleared
1926          * the VFTA and other state, so if there
1927          * have been no vlan's registered do nothing.
1928          */
1929         if (sc->num_vlans == 0)
1930                 return;
1931
1932         /* Setup the queues for vlans */
1933         if (ifp->if_capenable & IFCAP_VLAN_HWTAGGING) {
1934                 for (i = 0; i < sc->num_rx_queues; i++) {
1935                         rxr = &sc->rx_queues[i].rxr;
1936                         /* On 82599 the VLAN enable is per/queue in RXDCTL */
1937                         if (hw->mac.type != ixgbe_mac_82598EB) {
1938                                 ctrl = IXGBE_READ_REG(hw, IXGBE_RXDCTL(rxr->me));
1939                                 ctrl |= IXGBE_RXDCTL_VME;
1940                                 IXGBE_WRITE_REG(hw, IXGBE_RXDCTL(rxr->me), ctrl);
1941                         }
1942                         rxr->vtag_strip = true;
1943                 }
1944         }
1945
1946         if ((ifp->if_capenable & IFCAP_VLAN_HWFILTER) == 0)
1947                 return;
1948         /*
1949          * A soft reset zero's out the VFTA, so
1950          * we need to repopulate it now.
1951          */
1952         for (i = 0; i < IXGBE_VFTA_SIZE; i++)
1953                 if (sc->shadow_vfta[i] != 0)
1954                         IXGBE_WRITE_REG(hw, IXGBE_VFTA(i),
1955                             sc->shadow_vfta[i]);
1956
1957         ctrl = IXGBE_READ_REG(hw, IXGBE_VLNCTRL);
1958         /* Enable the Filter Table if enabled */
1959         if (ifp->if_capenable & IFCAP_VLAN_HWFILTER) {
1960                 ctrl &= ~IXGBE_VLNCTRL_CFIEN;
1961                 ctrl |= IXGBE_VLNCTRL_VFE;
1962         }
1963         if (hw->mac.type == ixgbe_mac_82598EB)
1964                 ctrl |= IXGBE_VLNCTRL_VME;
1965         IXGBE_WRITE_REG(hw, IXGBE_VLNCTRL, ctrl);
1966 } /* ixgbe_setup_vlan_hw_support */
1967
1968 /************************************************************************
1969  * ixgbe_get_slot_info
1970  *
1971  *   Get the width and transaction speed of
1972  *   the slot this adapter is plugged into.
1973  ************************************************************************/
1974 static void
1975 ixgbe_get_slot_info(struct ixgbe_softc *sc)
1976 {
1977         device_t        dev = iflib_get_dev(sc->ctx);
1978         struct ixgbe_hw *hw = &sc->hw;
1979         int             bus_info_valid = true;
1980         u32             offset;
1981         u16             link;
1982
1983         /* Some devices are behind an internal bridge */
1984         switch (hw->device_id) {
1985         case IXGBE_DEV_ID_82599_SFP_SF_QP:
1986         case IXGBE_DEV_ID_82599_QSFP_SF_QP:
1987                 goto get_parent_info;
1988         default:
1989                 break;
1990         }
1991
1992         ixgbe_get_bus_info(hw);
1993
1994         /*
1995          * Some devices don't use PCI-E, but there is no need
1996          * to display "Unknown" for bus speed and width.
1997          */
1998         switch (hw->mac.type) {
1999         case ixgbe_mac_X550EM_x:
2000         case ixgbe_mac_X550EM_a:
2001                 return;
2002         default:
2003                 goto display;
2004         }
2005
2006 get_parent_info:
2007         /*
2008          * For the Quad port adapter we need to parse back
2009          * up the PCI tree to find the speed of the expansion
2010          * slot into which this adapter is plugged. A bit more work.
2011          */
2012         dev = device_get_parent(device_get_parent(dev));
2013 #ifdef IXGBE_DEBUG
2014         device_printf(dev, "parent pcib = %x,%x,%x\n", pci_get_bus(dev),
2015             pci_get_slot(dev), pci_get_function(dev));
2016 #endif
2017         dev = device_get_parent(device_get_parent(dev));
2018 #ifdef IXGBE_DEBUG
2019         device_printf(dev, "slot pcib = %x,%x,%x\n", pci_get_bus(dev),
2020             pci_get_slot(dev), pci_get_function(dev));
2021 #endif
2022         /* Now get the PCI Express Capabilities offset */
2023         if (pci_find_cap(dev, PCIY_EXPRESS, &offset)) {
2024                 /*
2025                  * Hmm...can't get PCI-Express capabilities.
2026                  * Falling back to default method.
2027                  */
2028                 bus_info_valid = false;
2029                 ixgbe_get_bus_info(hw);
2030                 goto display;
2031         }
2032         /* ...and read the Link Status Register */
2033         link = pci_read_config(dev, offset + PCIER_LINK_STA, 2);
2034         ixgbe_set_pci_config_data_generic(hw, link);
2035
2036 display:
2037         device_printf(dev, "PCI Express Bus: Speed %s %s\n",
2038             ((hw->bus.speed == ixgbe_bus_speed_8000)    ? "8.0GT/s"  :
2039              (hw->bus.speed == ixgbe_bus_speed_5000)    ? "5.0GT/s"  :
2040              (hw->bus.speed == ixgbe_bus_speed_2500)    ? "2.5GT/s"  :
2041              "Unknown"),
2042             ((hw->bus.width == ixgbe_bus_width_pcie_x8) ? "Width x8" :
2043              (hw->bus.width == ixgbe_bus_width_pcie_x4) ? "Width x4" :
2044              (hw->bus.width == ixgbe_bus_width_pcie_x1) ? "Width x1" :
2045              "Unknown"));
2046
2047         if (bus_info_valid) {
2048                 if ((hw->device_id != IXGBE_DEV_ID_82599_SFP_SF_QP) &&
2049                     ((hw->bus.width <= ixgbe_bus_width_pcie_x4) &&
2050                     (hw->bus.speed == ixgbe_bus_speed_2500))) {
2051                         device_printf(dev, "PCI-Express bandwidth available for this card\n     is not sufficient for optimal performance.\n");
2052                         device_printf(dev, "For optimal performance a x8 PCIE, or x4 PCIE Gen2 slot is required.\n");
2053                 }
2054                 if ((hw->device_id == IXGBE_DEV_ID_82599_SFP_SF_QP) &&
2055                     ((hw->bus.width <= ixgbe_bus_width_pcie_x8) &&
2056                     (hw->bus.speed < ixgbe_bus_speed_8000))) {
2057                         device_printf(dev, "PCI-Express bandwidth available for this card\n     is not sufficient for optimal performance.\n");
2058                         device_printf(dev, "For optimal performance a x8 PCIE Gen3 slot is required.\n");
2059                 }
2060         } else
2061                 device_printf(dev, "Unable to determine slot speed/width. The speed/width reported are that of the internal switch.\n");
2062
2063         return;
2064 } /* ixgbe_get_slot_info */
2065
2066 /************************************************************************
2067  * ixgbe_if_msix_intr_assign
2068  *
2069  *   Setup MSI-X Interrupt resources and handlers
2070  ************************************************************************/
2071 static int
2072 ixgbe_if_msix_intr_assign(if_ctx_t ctx, int msix)
2073 {
2074         struct ixgbe_softc     *sc = iflib_get_softc(ctx);
2075         struct ix_rx_queue *rx_que = sc->rx_queues;
2076         struct ix_tx_queue *tx_que;
2077         int                error, rid, vector = 0;
2078         char               buf[16];
2079
2080         /* Admin Que is vector 0*/
2081         rid = vector + 1;
2082         for (int i = 0; i < sc->num_rx_queues; i++, vector++, rx_que++) {
2083                 rid = vector + 1;
2084
2085                 snprintf(buf, sizeof(buf), "rxq%d", i);
2086                 error = iflib_irq_alloc_generic(ctx, &rx_que->que_irq, rid,
2087                     IFLIB_INTR_RXTX, ixgbe_msix_que, rx_que, rx_que->rxr.me, buf);
2088
2089                 if (error) {
2090                         device_printf(iflib_get_dev(ctx),
2091                             "Failed to allocate que int %d err: %d", i, error);
2092                         sc->num_rx_queues = i + 1;
2093                         goto fail;
2094                 }
2095
2096                 rx_que->msix = vector;
2097         }
2098         for (int i = 0; i < sc->num_tx_queues; i++) {
2099                 snprintf(buf, sizeof(buf), "txq%d", i);
2100                 tx_que = &sc->tx_queues[i];
2101                 tx_que->msix = i % sc->num_rx_queues;
2102                 iflib_softirq_alloc_generic(ctx,
2103                     &sc->rx_queues[tx_que->msix].que_irq,
2104                     IFLIB_INTR_TX, tx_que, tx_que->txr.me, buf);
2105         }
2106         rid = vector + 1;
2107         error = iflib_irq_alloc_generic(ctx, &sc->irq, rid,
2108             IFLIB_INTR_ADMIN, ixgbe_msix_link, sc, 0, "aq");
2109         if (error) {
2110                 device_printf(iflib_get_dev(ctx),
2111                     "Failed to register admin handler");
2112                 return (error);
2113         }
2114
2115         sc->vector = vector;
2116
2117         return (0);
2118 fail:
2119         iflib_irq_free(ctx, &sc->irq);
2120         rx_que = sc->rx_queues;
2121         for (int i = 0; i < sc->num_rx_queues; i++, rx_que++)
2122                 iflib_irq_free(ctx, &rx_que->que_irq);
2123
2124         return (error);
2125 } /* ixgbe_if_msix_intr_assign */
2126
2127 static inline void
2128 ixgbe_perform_aim(struct ixgbe_softc *sc, struct ix_rx_queue *que)
2129 {
2130         uint32_t newitr = 0;
2131         struct rx_ring *rxr = &que->rxr;
2132
2133         /*
2134          * Do Adaptive Interrupt Moderation:
2135          *  - Write out last calculated setting
2136          *  - Calculate based on average size over
2137          *    the last interval.
2138          */
2139         if (que->eitr_setting) {
2140                 IXGBE_WRITE_REG(&sc->hw, IXGBE_EITR(que->msix),
2141                     que->eitr_setting);
2142         }
2143
2144         que->eitr_setting = 0;
2145         /* Idle, do nothing */
2146         if (rxr->bytes == 0) {
2147                 return;
2148         }
2149
2150         if ((rxr->bytes) && (rxr->packets)) {
2151                 newitr = (rxr->bytes / rxr->packets);
2152         }
2153
2154         newitr += 24; /* account for hardware frame, crc */
2155         /* set an upper boundary */
2156         newitr = min(newitr, 3000);
2157
2158         /* Be nice to the mid range */
2159         if ((newitr > 300) && (newitr < 1200)) {
2160                 newitr = (newitr / 3);
2161         } else {
2162                 newitr = (newitr / 2);
2163         }
2164
2165         if (sc->hw.mac.type == ixgbe_mac_82598EB) {
2166                 newitr |= newitr << 16;
2167         } else {
2168                 newitr |= IXGBE_EITR_CNT_WDIS;
2169         }
2170
2171         /* save for next interrupt */
2172         que->eitr_setting = newitr;
2173
2174         /* Reset state */
2175         rxr->bytes = 0;
2176         rxr->packets = 0;
2177
2178         return;
2179 }
2180
2181 /*********************************************************************
2182  * ixgbe_msix_que - MSI-X Queue Interrupt Service routine
2183  **********************************************************************/
2184 static int
2185 ixgbe_msix_que(void *arg)
2186 {
2187         struct ix_rx_queue *que = arg;
2188         struct ixgbe_softc     *sc = que->sc;
2189         struct ifnet       *ifp = iflib_get_ifp(que->sc->ctx);
2190
2191         /* Protect against spurious interrupts */
2192         if ((ifp->if_drv_flags & IFF_DRV_RUNNING) == 0)
2193                 return (FILTER_HANDLED);
2194
2195         ixgbe_disable_queue(sc, que->msix);
2196         ++que->irqs;
2197
2198         /* Check for AIM */
2199         if (sc->enable_aim) {
2200                 ixgbe_perform_aim(sc, que);
2201         }
2202
2203         return (FILTER_SCHEDULE_THREAD);
2204 } /* ixgbe_msix_que */
2205
2206 /************************************************************************
2207  * ixgbe_media_status - Media Ioctl callback
2208  *
2209  *   Called whenever the user queries the status of
2210  *   the interface using ifconfig.
2211  ************************************************************************/
2212 static void
2213 ixgbe_if_media_status(if_ctx_t ctx, struct ifmediareq * ifmr)
2214 {
2215         struct ixgbe_softc  *sc = iflib_get_softc(ctx);
2216         struct ixgbe_hw *hw = &sc->hw;
2217         int             layer;
2218
2219         INIT_DEBUGOUT("ixgbe_if_media_status: begin");
2220
2221         ifmr->ifm_status = IFM_AVALID;
2222         ifmr->ifm_active = IFM_ETHER;
2223
2224         if (!sc->link_active)
2225                 return;
2226
2227         ifmr->ifm_status |= IFM_ACTIVE;
2228         layer = sc->phy_layer;
2229
2230         if (layer & IXGBE_PHYSICAL_LAYER_10GBASE_T ||
2231             layer & IXGBE_PHYSICAL_LAYER_1000BASE_T ||
2232             layer & IXGBE_PHYSICAL_LAYER_100BASE_TX ||
2233             layer & IXGBE_PHYSICAL_LAYER_10BASE_T)
2234                 switch (sc->link_speed) {
2235                 case IXGBE_LINK_SPEED_10GB_FULL:
2236                         ifmr->ifm_active |= IFM_10G_T | IFM_FDX;
2237                         break;
2238                 case IXGBE_LINK_SPEED_1GB_FULL:
2239                         ifmr->ifm_active |= IFM_1000_T | IFM_FDX;
2240                         break;
2241                 case IXGBE_LINK_SPEED_100_FULL:
2242                         ifmr->ifm_active |= IFM_100_TX | IFM_FDX;
2243                         break;
2244                 case IXGBE_LINK_SPEED_10_FULL:
2245                         ifmr->ifm_active |= IFM_10_T | IFM_FDX;
2246                         break;
2247                 }
2248         if (hw->mac.type == ixgbe_mac_X550)
2249                 switch (sc->link_speed) {
2250                 case IXGBE_LINK_SPEED_5GB_FULL:
2251                         ifmr->ifm_active |= IFM_5000_T | IFM_FDX;
2252                         break;
2253                 case IXGBE_LINK_SPEED_2_5GB_FULL:
2254                         ifmr->ifm_active |= IFM_2500_T | IFM_FDX;
2255                         break;
2256                 }
2257         if (layer & IXGBE_PHYSICAL_LAYER_SFP_PLUS_CU ||
2258             layer & IXGBE_PHYSICAL_LAYER_SFP_ACTIVE_DA)
2259                 switch (sc->link_speed) {
2260                 case IXGBE_LINK_SPEED_10GB_FULL:
2261                         ifmr->ifm_active |= IFM_10G_TWINAX | IFM_FDX;
2262                         break;
2263                 }
2264         if (layer & IXGBE_PHYSICAL_LAYER_10GBASE_LR)
2265                 switch (sc->link_speed) {
2266                 case IXGBE_LINK_SPEED_10GB_FULL:
2267                         ifmr->ifm_active |= IFM_10G_LR | IFM_FDX;
2268                         break;
2269                 case IXGBE_LINK_SPEED_1GB_FULL:
2270                         ifmr->ifm_active |= IFM_1000_LX | IFM_FDX;
2271                         break;
2272                 }
2273         if (layer & IXGBE_PHYSICAL_LAYER_10GBASE_LRM)
2274                 switch (sc->link_speed) {
2275                 case IXGBE_LINK_SPEED_10GB_FULL:
2276                         ifmr->ifm_active |= IFM_10G_LRM | IFM_FDX;
2277                         break;
2278                 case IXGBE_LINK_SPEED_1GB_FULL:
2279                         ifmr->ifm_active |= IFM_1000_LX | IFM_FDX;
2280                         break;
2281                 }
2282         if (layer & IXGBE_PHYSICAL_LAYER_10GBASE_SR ||
2283             layer & IXGBE_PHYSICAL_LAYER_1000BASE_SX)
2284                 switch (sc->link_speed) {
2285                 case IXGBE_LINK_SPEED_10GB_FULL:
2286                         ifmr->ifm_active |= IFM_10G_SR | IFM_FDX;
2287                         break;
2288                 case IXGBE_LINK_SPEED_1GB_FULL:
2289                         ifmr->ifm_active |= IFM_1000_SX | IFM_FDX;
2290                         break;
2291                 }
2292         if (layer & IXGBE_PHYSICAL_LAYER_10GBASE_CX4)
2293                 switch (sc->link_speed) {
2294                 case IXGBE_LINK_SPEED_10GB_FULL:
2295                         ifmr->ifm_active |= IFM_10G_CX4 | IFM_FDX;
2296                         break;
2297                 }
2298         /*
2299          * XXX: These need to use the proper media types once
2300          * they're added.
2301          */
2302 #ifndef IFM_ETH_XTYPE
2303         if (layer & IXGBE_PHYSICAL_LAYER_10GBASE_KR)
2304                 switch (sc->link_speed) {
2305                 case IXGBE_LINK_SPEED_10GB_FULL:
2306                         ifmr->ifm_active |= IFM_10G_SR | IFM_FDX;
2307                         break;
2308                 case IXGBE_LINK_SPEED_2_5GB_FULL:
2309                         ifmr->ifm_active |= IFM_2500_SX | IFM_FDX;
2310                         break;
2311                 case IXGBE_LINK_SPEED_1GB_FULL:
2312                         ifmr->ifm_active |= IFM_1000_CX | IFM_FDX;
2313                         break;
2314                 }
2315         else if (layer & IXGBE_PHYSICAL_LAYER_10GBASE_KX4 ||
2316             layer & IXGBE_PHYSICAL_LAYER_2500BASE_KX ||
2317             layer & IXGBE_PHYSICAL_LAYER_1000BASE_KX)
2318                 switch (sc->link_speed) {
2319                 case IXGBE_LINK_SPEED_10GB_FULL:
2320                         ifmr->ifm_active |= IFM_10G_CX4 | IFM_FDX;
2321                         break;
2322                 case IXGBE_LINK_SPEED_2_5GB_FULL:
2323                         ifmr->ifm_active |= IFM_2500_SX | IFM_FDX;
2324                         break;
2325                 case IXGBE_LINK_SPEED_1GB_FULL:
2326                         ifmr->ifm_active |= IFM_1000_CX | IFM_FDX;
2327                         break;
2328                 }
2329 #else
2330         if (layer & IXGBE_PHYSICAL_LAYER_10GBASE_KR)
2331                 switch (sc->link_speed) {
2332                 case IXGBE_LINK_SPEED_10GB_FULL:
2333                         ifmr->ifm_active |= IFM_10G_KR | IFM_FDX;
2334                         break;
2335                 case IXGBE_LINK_SPEED_2_5GB_FULL:
2336                         ifmr->ifm_active |= IFM_2500_KX | IFM_FDX;
2337                         break;
2338                 case IXGBE_LINK_SPEED_1GB_FULL:
2339                         ifmr->ifm_active |= IFM_1000_KX | IFM_FDX;
2340                         break;
2341                 }
2342         else if (layer & IXGBE_PHYSICAL_LAYER_10GBASE_KX4 ||
2343             layer & IXGBE_PHYSICAL_LAYER_2500BASE_KX ||
2344             layer & IXGBE_PHYSICAL_LAYER_1000BASE_KX)
2345                 switch (sc->link_speed) {
2346                 case IXGBE_LINK_SPEED_10GB_FULL:
2347                         ifmr->ifm_active |= IFM_10G_KX4 | IFM_FDX;
2348                         break;
2349                 case IXGBE_LINK_SPEED_2_5GB_FULL:
2350                         ifmr->ifm_active |= IFM_2500_KX | IFM_FDX;
2351                         break;
2352                 case IXGBE_LINK_SPEED_1GB_FULL:
2353                         ifmr->ifm_active |= IFM_1000_KX | IFM_FDX;
2354                         break;
2355                 }
2356 #endif
2357
2358         /* If nothing is recognized... */
2359         if (IFM_SUBTYPE(ifmr->ifm_active) == 0)
2360                 ifmr->ifm_active |= IFM_UNKNOWN;
2361
2362         /* Display current flow control setting used on link */
2363         if (hw->fc.current_mode == ixgbe_fc_rx_pause ||
2364             hw->fc.current_mode == ixgbe_fc_full)
2365                 ifmr->ifm_active |= IFM_ETH_RXPAUSE;
2366         if (hw->fc.current_mode == ixgbe_fc_tx_pause ||
2367             hw->fc.current_mode == ixgbe_fc_full)
2368                 ifmr->ifm_active |= IFM_ETH_TXPAUSE;
2369 } /* ixgbe_media_status */
2370
2371 /************************************************************************
2372  * ixgbe_media_change - Media Ioctl callback
2373  *
2374  *   Called when the user changes speed/duplex using
2375  *   media/mediopt option with ifconfig.
2376  ************************************************************************/
2377 static int
2378 ixgbe_if_media_change(if_ctx_t ctx)
2379 {
2380         struct ixgbe_softc   *sc = iflib_get_softc(ctx);
2381         struct ifmedia   *ifm = iflib_get_media(ctx);
2382         struct ixgbe_hw  *hw = &sc->hw;
2383         ixgbe_link_speed speed = 0;
2384
2385         INIT_DEBUGOUT("ixgbe_if_media_change: begin");
2386
2387         if (IFM_TYPE(ifm->ifm_media) != IFM_ETHER)
2388                 return (EINVAL);
2389
2390         if (hw->phy.media_type == ixgbe_media_type_backplane)
2391                 return (EPERM);
2392
2393         /*
2394          * We don't actually need to check against the supported
2395          * media types of the adapter; ifmedia will take care of
2396          * that for us.
2397          */
2398         switch (IFM_SUBTYPE(ifm->ifm_media)) {
2399         case IFM_AUTO:
2400         case IFM_10G_T:
2401                 speed |= IXGBE_LINK_SPEED_100_FULL;
2402                 speed |= IXGBE_LINK_SPEED_1GB_FULL;
2403                 speed |= IXGBE_LINK_SPEED_10GB_FULL;
2404                 break;
2405         case IFM_10G_LRM:
2406         case IFM_10G_LR:
2407 #ifndef IFM_ETH_XTYPE
2408         case IFM_10G_SR: /* KR, too */
2409         case IFM_10G_CX4: /* KX4 */
2410 #else
2411         case IFM_10G_KR:
2412         case IFM_10G_KX4:
2413 #endif
2414                 speed |= IXGBE_LINK_SPEED_1GB_FULL;
2415                 speed |= IXGBE_LINK_SPEED_10GB_FULL;
2416                 break;
2417 #ifndef IFM_ETH_XTYPE
2418         case IFM_1000_CX: /* KX */
2419 #else
2420         case IFM_1000_KX:
2421 #endif
2422         case IFM_1000_LX:
2423         case IFM_1000_SX:
2424                 speed |= IXGBE_LINK_SPEED_1GB_FULL;
2425                 break;
2426         case IFM_1000_T:
2427                 speed |= IXGBE_LINK_SPEED_100_FULL;
2428                 speed |= IXGBE_LINK_SPEED_1GB_FULL;
2429                 break;
2430         case IFM_10G_TWINAX:
2431                 speed |= IXGBE_LINK_SPEED_10GB_FULL;
2432                 break;
2433         case IFM_5000_T:
2434                 speed |= IXGBE_LINK_SPEED_5GB_FULL;
2435                 break;
2436         case IFM_2500_T:
2437                 speed |= IXGBE_LINK_SPEED_2_5GB_FULL;
2438                 break;
2439         case IFM_100_TX:
2440                 speed |= IXGBE_LINK_SPEED_100_FULL;
2441                 break;
2442         case IFM_10_T:
2443                 speed |= IXGBE_LINK_SPEED_10_FULL;
2444                 break;
2445         default:
2446                 goto invalid;
2447         }
2448
2449         hw->mac.autotry_restart = true;
2450         hw->mac.ops.setup_link(hw, speed, true);
2451         sc->advertise =
2452             ((speed & IXGBE_LINK_SPEED_10GB_FULL)  ? 0x4  : 0) |
2453             ((speed & IXGBE_LINK_SPEED_5GB_FULL)   ? 0x20 : 0) |
2454             ((speed & IXGBE_LINK_SPEED_2_5GB_FULL) ? 0x10 : 0) |
2455             ((speed & IXGBE_LINK_SPEED_1GB_FULL)   ? 0x2  : 0) |
2456             ((speed & IXGBE_LINK_SPEED_100_FULL)   ? 0x1  : 0) |
2457             ((speed & IXGBE_LINK_SPEED_10_FULL)    ? 0x8  : 0);
2458
2459         return (0);
2460
2461 invalid:
2462         device_printf(iflib_get_dev(ctx), "Invalid media type!\n");
2463
2464         return (EINVAL);
2465 } /* ixgbe_if_media_change */
2466
2467 /************************************************************************
2468  * ixgbe_set_promisc
2469  ************************************************************************/
2470 static int
2471 ixgbe_if_promisc_set(if_ctx_t ctx, int flags)
2472 {
2473         struct ixgbe_softc *sc = iflib_get_softc(ctx);
2474         struct ifnet   *ifp = iflib_get_ifp(ctx);
2475         u32            rctl;
2476         int            mcnt = 0;
2477
2478         rctl = IXGBE_READ_REG(&sc->hw, IXGBE_FCTRL);
2479         rctl &= (~IXGBE_FCTRL_UPE);
2480         if (ifp->if_flags & IFF_ALLMULTI)
2481                 mcnt = MAX_NUM_MULTICAST_ADDRESSES;
2482         else {
2483                 mcnt = min(if_llmaddr_count(ifp), MAX_NUM_MULTICAST_ADDRESSES);
2484         }
2485         if (mcnt < MAX_NUM_MULTICAST_ADDRESSES)
2486                 rctl &= (~IXGBE_FCTRL_MPE);
2487         IXGBE_WRITE_REG(&sc->hw, IXGBE_FCTRL, rctl);
2488
2489         if (ifp->if_flags & IFF_PROMISC) {
2490                 rctl |= (IXGBE_FCTRL_UPE | IXGBE_FCTRL_MPE);
2491                 IXGBE_WRITE_REG(&sc->hw, IXGBE_FCTRL, rctl);
2492         } else if (ifp->if_flags & IFF_ALLMULTI) {
2493                 rctl |= IXGBE_FCTRL_MPE;
2494                 rctl &= ~IXGBE_FCTRL_UPE;
2495                 IXGBE_WRITE_REG(&sc->hw, IXGBE_FCTRL, rctl);
2496         }
2497         return (0);
2498 } /* ixgbe_if_promisc_set */
2499
2500 /************************************************************************
2501  * ixgbe_msix_link - Link status change ISR (MSI/MSI-X)
2502  ************************************************************************/
2503 static int
2504 ixgbe_msix_link(void *arg)
2505 {
2506         struct ixgbe_softc  *sc = arg;
2507         struct ixgbe_hw *hw = &sc->hw;
2508         u32             eicr, eicr_mask;
2509         s32             retval;
2510
2511         ++sc->link_irq;
2512
2513         /* Pause other interrupts */
2514         IXGBE_WRITE_REG(hw, IXGBE_EIMC, IXGBE_EIMC_OTHER);
2515
2516         /* First get the cause */
2517         eicr = IXGBE_READ_REG(hw, IXGBE_EICS);
2518         /* Be sure the queue bits are not cleared */
2519         eicr &= ~IXGBE_EICR_RTX_QUEUE;
2520         /* Clear interrupt with write */
2521         IXGBE_WRITE_REG(hw, IXGBE_EICR, eicr);
2522
2523         /* Link status change */
2524         if (eicr & IXGBE_EICR_LSC) {
2525                 IXGBE_WRITE_REG(hw, IXGBE_EIMC, IXGBE_EIMC_LSC);
2526                 sc->task_requests |= IXGBE_REQUEST_TASK_LSC;
2527         }
2528
2529         if (sc->hw.mac.type != ixgbe_mac_82598EB) {
2530                 if ((sc->feat_en & IXGBE_FEATURE_FDIR) &&
2531                     (eicr & IXGBE_EICR_FLOW_DIR)) {
2532                         /* This is probably overkill :) */
2533                         if (!atomic_cmpset_int(&sc->fdir_reinit, 0, 1))
2534                                 return (FILTER_HANDLED);
2535                         /* Disable the interrupt */
2536                         IXGBE_WRITE_REG(hw, IXGBE_EIMC, IXGBE_EICR_FLOW_DIR);
2537                         sc->task_requests |= IXGBE_REQUEST_TASK_FDIR;
2538                 } else
2539                         if (eicr & IXGBE_EICR_ECC) {
2540                                 device_printf(iflib_get_dev(sc->ctx),
2541                                    "\nCRITICAL: ECC ERROR!! Please Reboot!!\n");
2542                                 IXGBE_WRITE_REG(hw, IXGBE_EICR, IXGBE_EICR_ECC);
2543                         }
2544
2545                 /* Check for over temp condition */
2546                 if (sc->feat_en & IXGBE_FEATURE_TEMP_SENSOR) {
2547                         switch (sc->hw.mac.type) {
2548                         case ixgbe_mac_X550EM_a:
2549                                 if (!(eicr & IXGBE_EICR_GPI_SDP0_X550EM_a))
2550                                         break;
2551                                 IXGBE_WRITE_REG(hw, IXGBE_EIMC,
2552                                     IXGBE_EICR_GPI_SDP0_X550EM_a);
2553                                 IXGBE_WRITE_REG(hw, IXGBE_EICR,
2554                                     IXGBE_EICR_GPI_SDP0_X550EM_a);
2555                                 retval = hw->phy.ops.check_overtemp(hw);
2556                                 if (retval != IXGBE_ERR_OVERTEMP)
2557                                         break;
2558                                 device_printf(iflib_get_dev(sc->ctx),
2559                                     "\nCRITICAL: OVER TEMP!! PHY IS SHUT DOWN!!\n");
2560                                 device_printf(iflib_get_dev(sc->ctx),
2561                                     "System shutdown required!\n");
2562                                 break;
2563                         default:
2564                                 if (!(eicr & IXGBE_EICR_TS))
2565                                         break;
2566                                 retval = hw->phy.ops.check_overtemp(hw);
2567                                 if (retval != IXGBE_ERR_OVERTEMP)
2568                                         break;
2569                                 device_printf(iflib_get_dev(sc->ctx),
2570                                     "\nCRITICAL: OVER TEMP!! PHY IS SHUT DOWN!!\n");
2571                                 device_printf(iflib_get_dev(sc->ctx),
2572                                     "System shutdown required!\n");
2573                                 IXGBE_WRITE_REG(hw, IXGBE_EICR, IXGBE_EICR_TS);
2574                                 break;
2575                         }
2576                 }
2577
2578                 /* Check for VF message */
2579                 if ((sc->feat_en & IXGBE_FEATURE_SRIOV) &&
2580                     (eicr & IXGBE_EICR_MAILBOX))
2581                         sc->task_requests |= IXGBE_REQUEST_TASK_MBX;
2582         }
2583
2584         if (ixgbe_is_sfp(hw)) {
2585                 /* Pluggable optics-related interrupt */
2586                 if (hw->mac.type >= ixgbe_mac_X540)
2587                         eicr_mask = IXGBE_EICR_GPI_SDP0_X540;
2588                 else
2589                         eicr_mask = IXGBE_EICR_GPI_SDP2_BY_MAC(hw);
2590
2591                 if (eicr & eicr_mask) {
2592                         IXGBE_WRITE_REG(hw, IXGBE_EICR, eicr_mask);
2593                         sc->task_requests |= IXGBE_REQUEST_TASK_MOD;
2594                 }
2595
2596                 if ((hw->mac.type == ixgbe_mac_82599EB) &&
2597                     (eicr & IXGBE_EICR_GPI_SDP1_BY_MAC(hw))) {
2598                         IXGBE_WRITE_REG(hw, IXGBE_EICR,
2599                             IXGBE_EICR_GPI_SDP1_BY_MAC(hw));
2600                         sc->task_requests |= IXGBE_REQUEST_TASK_MSF;
2601                 }
2602         }
2603
2604         /* Check for fan failure */
2605         if (sc->feat_en & IXGBE_FEATURE_FAN_FAIL) {
2606                 ixgbe_check_fan_failure(sc, eicr, true);
2607                 IXGBE_WRITE_REG(hw, IXGBE_EICR, IXGBE_EICR_GPI_SDP1_BY_MAC(hw));
2608         }
2609
2610         /* External PHY interrupt */
2611         if ((hw->phy.type == ixgbe_phy_x550em_ext_t) &&
2612             (eicr & IXGBE_EICR_GPI_SDP0_X540)) {
2613                 IXGBE_WRITE_REG(hw, IXGBE_EICR, IXGBE_EICR_GPI_SDP0_X540);
2614                 sc->task_requests |= IXGBE_REQUEST_TASK_PHY;
2615         }
2616
2617         return (sc->task_requests != 0) ? FILTER_SCHEDULE_THREAD : FILTER_HANDLED;
2618 } /* ixgbe_msix_link */
2619
2620 /************************************************************************
2621  * ixgbe_sysctl_interrupt_rate_handler
2622  ************************************************************************/
2623 static int
2624 ixgbe_sysctl_interrupt_rate_handler(SYSCTL_HANDLER_ARGS)
2625 {
2626         struct ix_rx_queue *que = ((struct ix_rx_queue *)oidp->oid_arg1);
2627         int                error;
2628         unsigned int       reg, usec, rate;
2629
2630         reg = IXGBE_READ_REG(&que->sc->hw, IXGBE_EITR(que->msix));
2631         usec = ((reg & 0x0FF8) >> 3);
2632         if (usec > 0)
2633                 rate = 500000 / usec;
2634         else
2635                 rate = 0;
2636         error = sysctl_handle_int(oidp, &rate, 0, req);
2637         if (error || !req->newptr)
2638                 return error;
2639         reg &= ~0xfff; /* default, no limitation */
2640         ixgbe_max_interrupt_rate = 0;
2641         if (rate > 0 && rate < 500000) {
2642                 if (rate < 1000)
2643                         rate = 1000;
2644                 ixgbe_max_interrupt_rate = rate;
2645                 reg |= ((4000000/rate) & 0xff8);
2646         }
2647         IXGBE_WRITE_REG(&que->sc->hw, IXGBE_EITR(que->msix), reg);
2648
2649         return (0);
2650 } /* ixgbe_sysctl_interrupt_rate_handler */
2651
2652 /************************************************************************
2653  * ixgbe_add_device_sysctls
2654  ************************************************************************/
2655 static void
2656 ixgbe_add_device_sysctls(if_ctx_t ctx)
2657 {
2658         struct ixgbe_softc         *sc = iflib_get_softc(ctx);
2659         device_t               dev = iflib_get_dev(ctx);
2660         struct ixgbe_hw        *hw = &sc->hw;
2661         struct sysctl_oid_list *child;
2662         struct sysctl_ctx_list *ctx_list;
2663
2664         ctx_list = device_get_sysctl_ctx(dev);
2665         child = SYSCTL_CHILDREN(device_get_sysctl_tree(dev));
2666
2667         /* Sysctls for all devices */
2668         SYSCTL_ADD_PROC(ctx_list, child, OID_AUTO, "fc",
2669             CTLTYPE_INT | CTLFLAG_RW | CTLFLAG_NEEDGIANT,
2670             sc, 0, ixgbe_sysctl_flowcntl, "I",
2671             IXGBE_SYSCTL_DESC_SET_FC);
2672
2673         SYSCTL_ADD_PROC(ctx_list, child, OID_AUTO, "advertise_speed",
2674             CTLTYPE_INT | CTLFLAG_RW | CTLFLAG_NEEDGIANT,
2675             sc, 0, ixgbe_sysctl_advertise, "I",
2676             IXGBE_SYSCTL_DESC_ADV_SPEED);
2677
2678         sc->enable_aim = ixgbe_enable_aim;
2679         SYSCTL_ADD_INT(ctx_list, child, OID_AUTO, "enable_aim", CTLFLAG_RW,
2680             &sc->enable_aim, 0, "Interrupt Moderation");
2681
2682         SYSCTL_ADD_PROC(ctx_list, child, OID_AUTO, "fw_version",
2683             CTLTYPE_STRING | CTLFLAG_RD | CTLFLAG_NEEDGIANT, sc, 0,
2684             ixgbe_sysctl_print_fw_version, "A", "Prints FW/NVM Versions");
2685
2686 #ifdef IXGBE_DEBUG
2687         /* testing sysctls (for all devices) */
2688         SYSCTL_ADD_PROC(ctx_list, child, OID_AUTO, "power_state",
2689             CTLTYPE_INT | CTLFLAG_RW | CTLFLAG_NEEDGIANT,
2690             sc, 0, ixgbe_sysctl_power_state,
2691             "I", "PCI Power State");
2692
2693         SYSCTL_ADD_PROC(ctx_list, child, OID_AUTO, "print_rss_config",
2694             CTLTYPE_STRING | CTLFLAG_RD | CTLFLAG_NEEDGIANT, sc, 0,
2695             ixgbe_sysctl_print_rss_config, "A", "Prints RSS Configuration");
2696 #endif
2697         /* for X550 series devices */
2698         if (hw->mac.type >= ixgbe_mac_X550)
2699                 SYSCTL_ADD_PROC(ctx_list, child, OID_AUTO, "dmac",
2700                     CTLTYPE_U16 | CTLFLAG_RW | CTLFLAG_NEEDGIANT,
2701                     sc, 0, ixgbe_sysctl_dmac,
2702                     "I", "DMA Coalesce");
2703
2704         /* for WoL-capable devices */
2705         if (hw->device_id == IXGBE_DEV_ID_X550EM_X_10G_T) {
2706                 SYSCTL_ADD_PROC(ctx_list, child, OID_AUTO, "wol_enable",
2707                     CTLTYPE_INT | CTLFLAG_RW | CTLFLAG_NEEDGIANT, sc, 0,
2708                     ixgbe_sysctl_wol_enable, "I", "Enable/Disable Wake on LAN");
2709
2710                 SYSCTL_ADD_PROC(ctx_list, child, OID_AUTO, "wufc",
2711                     CTLTYPE_U32 | CTLFLAG_RW | CTLFLAG_NEEDGIANT,
2712                     sc, 0, ixgbe_sysctl_wufc,
2713                     "I", "Enable/Disable Wake Up Filters");
2714         }
2715
2716         /* for X552/X557-AT devices */
2717         if (hw->device_id == IXGBE_DEV_ID_X550EM_X_10G_T) {
2718                 struct sysctl_oid *phy_node;
2719                 struct sysctl_oid_list *phy_list;
2720
2721                 phy_node = SYSCTL_ADD_NODE(ctx_list, child, OID_AUTO, "phy",
2722                     CTLFLAG_RD | CTLFLAG_MPSAFE, NULL, "External PHY sysctls");
2723                 phy_list = SYSCTL_CHILDREN(phy_node);
2724
2725                 SYSCTL_ADD_PROC(ctx_list, phy_list, OID_AUTO, "temp",
2726                     CTLTYPE_U16 | CTLFLAG_RD | CTLFLAG_NEEDGIANT,
2727                     sc, 0, ixgbe_sysctl_phy_temp,
2728                     "I", "Current External PHY Temperature (Celsius)");
2729
2730                 SYSCTL_ADD_PROC(ctx_list, phy_list, OID_AUTO,
2731                     "overtemp_occurred",
2732                     CTLTYPE_U16 | CTLFLAG_RD | CTLFLAG_NEEDGIANT, sc, 0,
2733                     ixgbe_sysctl_phy_overtemp_occurred, "I",
2734                     "External PHY High Temperature Event Occurred");
2735         }
2736
2737         if (sc->feat_cap & IXGBE_FEATURE_EEE) {
2738                 SYSCTL_ADD_PROC(ctx_list, child, OID_AUTO, "eee_state",
2739                     CTLTYPE_INT | CTLFLAG_RW | CTLFLAG_NEEDGIANT, sc, 0,
2740                     ixgbe_sysctl_eee_state, "I", "EEE Power Save State");
2741         }
2742 } /* ixgbe_add_device_sysctls */
2743
2744 /************************************************************************
2745  * ixgbe_allocate_pci_resources
2746  ************************************************************************/
2747 static int
2748 ixgbe_allocate_pci_resources(if_ctx_t ctx)
2749 {
2750         struct ixgbe_softc *sc = iflib_get_softc(ctx);
2751         device_t        dev = iflib_get_dev(ctx);
2752         int             rid;
2753
2754         rid = PCIR_BAR(0);
2755         sc->pci_mem = bus_alloc_resource_any(dev, SYS_RES_MEMORY, &rid,
2756             RF_ACTIVE);
2757
2758         if (!(sc->pci_mem)) {
2759                 device_printf(dev, "Unable to allocate bus resource: memory\n");
2760                 return (ENXIO);
2761         }
2762
2763         /* Save bus_space values for READ/WRITE_REG macros */
2764         sc->osdep.mem_bus_space_tag = rman_get_bustag(sc->pci_mem);
2765         sc->osdep.mem_bus_space_handle =
2766             rman_get_bushandle(sc->pci_mem);
2767         /* Set hw values for shared code */
2768         sc->hw.hw_addr = (u8 *)&sc->osdep.mem_bus_space_handle;
2769
2770         return (0);
2771 } /* ixgbe_allocate_pci_resources */
2772
2773 /************************************************************************
2774  * ixgbe_detach - Device removal routine
2775  *
2776  *   Called when the driver is being removed.
2777  *   Stops the adapter and deallocates all the resources
2778  *   that were allocated for driver operation.
2779  *
2780  *   return 0 on success, positive on failure
2781  ************************************************************************/
2782 static int
2783 ixgbe_if_detach(if_ctx_t ctx)
2784 {
2785         struct ixgbe_softc *sc = iflib_get_softc(ctx);
2786         device_t       dev = iflib_get_dev(ctx);
2787         u32            ctrl_ext;
2788
2789         INIT_DEBUGOUT("ixgbe_detach: begin");
2790
2791         if (ixgbe_pci_iov_detach(dev) != 0) {
2792                 device_printf(dev, "SR-IOV in use; detach first.\n");
2793                 return (EBUSY);
2794         }
2795
2796         ixgbe_setup_low_power_mode(ctx);
2797
2798         /* let hardware know driver is unloading */
2799         ctrl_ext = IXGBE_READ_REG(&sc->hw, IXGBE_CTRL_EXT);
2800         ctrl_ext &= ~IXGBE_CTRL_EXT_DRV_LOAD;
2801         IXGBE_WRITE_REG(&sc->hw, IXGBE_CTRL_EXT, ctrl_ext);
2802
2803         ixgbe_free_pci_resources(ctx);
2804         free(sc->mta, M_IXGBE);
2805
2806         return (0);
2807 } /* ixgbe_if_detach */
2808
2809 /************************************************************************
2810  * ixgbe_setup_low_power_mode - LPLU/WoL preparation
2811  *
2812  *   Prepare the adapter/port for LPLU and/or WoL
2813  ************************************************************************/
2814 static int
2815 ixgbe_setup_low_power_mode(if_ctx_t ctx)
2816 {
2817         struct ixgbe_softc  *sc = iflib_get_softc(ctx);
2818         struct ixgbe_hw *hw = &sc->hw;
2819         device_t        dev = iflib_get_dev(ctx);
2820         s32             error = 0;
2821
2822         if (!hw->wol_enabled)
2823                 ixgbe_set_phy_power(hw, false);
2824
2825         /* Limit power management flow to X550EM baseT */
2826         if (hw->device_id == IXGBE_DEV_ID_X550EM_X_10G_T &&
2827             hw->phy.ops.enter_lplu) {
2828                 /* Turn off support for APM wakeup. (Using ACPI instead) */
2829                 IXGBE_WRITE_REG(hw, IXGBE_GRC,
2830                     IXGBE_READ_REG(hw, IXGBE_GRC) & ~(u32)2);
2831
2832                 /*
2833                  * Clear Wake Up Status register to prevent any previous wakeup
2834                  * events from waking us up immediately after we suspend.
2835                  */
2836                 IXGBE_WRITE_REG(hw, IXGBE_WUS, 0xffffffff);
2837
2838                 /*
2839                  * Program the Wakeup Filter Control register with user filter
2840                  * settings
2841                  */
2842                 IXGBE_WRITE_REG(hw, IXGBE_WUFC, sc->wufc);
2843
2844                 /* Enable wakeups and power management in Wakeup Control */
2845                 IXGBE_WRITE_REG(hw, IXGBE_WUC,
2846                     IXGBE_WUC_WKEN | IXGBE_WUC_PME_EN);
2847
2848                 /* X550EM baseT adapters need a special LPLU flow */
2849                 hw->phy.reset_disable = true;
2850                 ixgbe_if_stop(ctx);
2851                 error = hw->phy.ops.enter_lplu(hw);
2852                 if (error)
2853                         device_printf(dev, "Error entering LPLU: %d\n", error);
2854                 hw->phy.reset_disable = false;
2855         } else {
2856                 /* Just stop for other adapters */
2857                 ixgbe_if_stop(ctx);
2858         }
2859
2860         return error;
2861 } /* ixgbe_setup_low_power_mode */
2862
2863 /************************************************************************
2864  * ixgbe_shutdown - Shutdown entry point
2865  ************************************************************************/
2866 static int
2867 ixgbe_if_shutdown(if_ctx_t ctx)
2868 {
2869         int error = 0;
2870
2871         INIT_DEBUGOUT("ixgbe_shutdown: begin");
2872
2873         error = ixgbe_setup_low_power_mode(ctx);
2874
2875         return (error);
2876 } /* ixgbe_if_shutdown */
2877
2878 /************************************************************************
2879  * ixgbe_suspend
2880  *
2881  *   From D0 to D3
2882  ************************************************************************/
2883 static int
2884 ixgbe_if_suspend(if_ctx_t ctx)
2885 {
2886         int error = 0;
2887
2888         INIT_DEBUGOUT("ixgbe_suspend: begin");
2889
2890         error = ixgbe_setup_low_power_mode(ctx);
2891
2892         return (error);
2893 } /* ixgbe_if_suspend */
2894
2895 /************************************************************************
2896  * ixgbe_resume
2897  *
2898  *   From D3 to D0
2899  ************************************************************************/
2900 static int
2901 ixgbe_if_resume(if_ctx_t ctx)
2902 {
2903         struct ixgbe_softc  *sc = iflib_get_softc(ctx);
2904         device_t        dev = iflib_get_dev(ctx);
2905         struct ifnet    *ifp = iflib_get_ifp(ctx);
2906         struct ixgbe_hw *hw = &sc->hw;
2907         u32             wus;
2908
2909         INIT_DEBUGOUT("ixgbe_resume: begin");
2910
2911         /* Read & clear WUS register */
2912         wus = IXGBE_READ_REG(hw, IXGBE_WUS);
2913         if (wus)
2914                 device_printf(dev, "Woken up by (WUS): %#010x\n",
2915                     IXGBE_READ_REG(hw, IXGBE_WUS));
2916         IXGBE_WRITE_REG(hw, IXGBE_WUS, 0xffffffff);
2917         /* And clear WUFC until next low-power transition */
2918         IXGBE_WRITE_REG(hw, IXGBE_WUFC, 0);
2919
2920         /*
2921          * Required after D3->D0 transition;
2922          * will re-advertise all previous advertised speeds
2923          */
2924         if (ifp->if_flags & IFF_UP)
2925                 ixgbe_if_init(ctx);
2926
2927         return (0);
2928 } /* ixgbe_if_resume */
2929
2930 /************************************************************************
2931  * ixgbe_if_mtu_set - Ioctl mtu entry point
2932  *
2933  *   Return 0 on success, EINVAL on failure
2934  ************************************************************************/
2935 static int
2936 ixgbe_if_mtu_set(if_ctx_t ctx, uint32_t mtu)
2937 {
2938         struct ixgbe_softc *sc = iflib_get_softc(ctx);
2939         int error = 0;
2940
2941         IOCTL_DEBUGOUT("ioctl: SIOCIFMTU (Set Interface MTU)");
2942
2943         if (mtu > IXGBE_MAX_MTU) {
2944                 error = EINVAL;
2945         } else {
2946                 sc->max_frame_size = mtu + IXGBE_MTU_HDR;
2947         }
2948
2949         return error;
2950 } /* ixgbe_if_mtu_set */
2951
2952 /************************************************************************
2953  * ixgbe_if_crcstrip_set
2954  ************************************************************************/
2955 static void
2956 ixgbe_if_crcstrip_set(if_ctx_t ctx, int onoff, int crcstrip)
2957 {
2958         struct ixgbe_softc *sc = iflib_get_softc(ctx);
2959         struct ixgbe_hw *hw = &sc->hw;
2960         /* crc stripping is set in two places:
2961          * IXGBE_HLREG0 (modified on init_locked and hw reset)
2962          * IXGBE_RDRXCTL (set by the original driver in
2963          *      ixgbe_setup_hw_rsc() called in init_locked.
2964          *      We disable the setting when netmap is compiled in).
2965          * We update the values here, but also in ixgbe.c because
2966          * init_locked sometimes is called outside our control.
2967          */
2968         uint32_t hl, rxc;
2969
2970         hl = IXGBE_READ_REG(hw, IXGBE_HLREG0);
2971         rxc = IXGBE_READ_REG(hw, IXGBE_RDRXCTL);
2972 #ifdef NETMAP
2973         if (netmap_verbose)
2974                 D("%s read  HLREG 0x%x rxc 0x%x",
2975                         onoff ? "enter" : "exit", hl, rxc);
2976 #endif
2977         /* hw requirements ... */
2978         rxc &= ~IXGBE_RDRXCTL_RSCFRSTSIZE;
2979         rxc |= IXGBE_RDRXCTL_RSCACKC;
2980         if (onoff && !crcstrip) {
2981                 /* keep the crc. Fast rx */
2982                 hl &= ~IXGBE_HLREG0_RXCRCSTRP;
2983                 rxc &= ~IXGBE_RDRXCTL_CRCSTRIP;
2984         } else {
2985                 /* reset default mode */
2986                 hl |= IXGBE_HLREG0_RXCRCSTRP;
2987                 rxc |= IXGBE_RDRXCTL_CRCSTRIP;
2988         }
2989 #ifdef NETMAP
2990         if (netmap_verbose)
2991                 D("%s write HLREG 0x%x rxc 0x%x",
2992                         onoff ? "enter" : "exit", hl, rxc);
2993 #endif
2994         IXGBE_WRITE_REG(hw, IXGBE_HLREG0, hl);
2995         IXGBE_WRITE_REG(hw, IXGBE_RDRXCTL, rxc);
2996 } /* ixgbe_if_crcstrip_set */
2997
2998 /*********************************************************************
2999  * ixgbe_if_init - Init entry point
3000  *
3001  *   Used in two ways: It is used by the stack as an init
3002  *   entry point in network interface structure. It is also
3003  *   used by the driver as a hw/sw initialization routine to
3004  *   get to a consistent state.
3005  *
3006  *   Return 0 on success, positive on failure
3007  **********************************************************************/
3008 void
3009 ixgbe_if_init(if_ctx_t ctx)
3010 {
3011         struct ixgbe_softc     *sc = iflib_get_softc(ctx);
3012         struct ifnet       *ifp = iflib_get_ifp(ctx);
3013         device_t           dev = iflib_get_dev(ctx);
3014         struct ixgbe_hw *hw = &sc->hw;
3015         struct ix_rx_queue *rx_que;
3016         struct ix_tx_queue *tx_que;
3017         u32             txdctl, mhadd;
3018         u32             rxdctl, rxctrl;
3019         u32             ctrl_ext;
3020
3021         int             i, j, err;
3022
3023         INIT_DEBUGOUT("ixgbe_if_init: begin");
3024
3025         /* Queue indices may change with IOV mode */
3026         ixgbe_align_all_queue_indices(sc);
3027
3028         /* reprogram the RAR[0] in case user changed it. */
3029         ixgbe_set_rar(hw, 0, hw->mac.addr, sc->pool, IXGBE_RAH_AV);
3030
3031         /* Get the latest mac address, User can use a LAA */
3032         bcopy(IF_LLADDR(ifp), hw->mac.addr, IXGBE_ETH_LENGTH_OF_ADDRESS);
3033         ixgbe_set_rar(hw, 0, hw->mac.addr, sc->pool, 1);
3034         hw->addr_ctrl.rar_used_count = 1;
3035
3036         ixgbe_init_hw(hw);
3037
3038         ixgbe_initialize_iov(sc);
3039
3040         ixgbe_initialize_transmit_units(ctx);
3041
3042         /* Setup Multicast table */
3043         ixgbe_if_multi_set(ctx);
3044
3045         /* Determine the correct mbuf pool, based on frame size */
3046         sc->rx_mbuf_sz = iflib_get_rx_mbuf_sz(ctx);
3047
3048         /* Configure RX settings */
3049         ixgbe_initialize_receive_units(ctx);
3050
3051         /*
3052          * Initialize variable holding task enqueue requests
3053          * from MSI-X interrupts
3054          */
3055         sc->task_requests = 0;
3056
3057         /* Enable SDP & MSI-X interrupts based on adapter */
3058         ixgbe_config_gpie(sc);
3059
3060         /* Set MTU size */
3061         if (ifp->if_mtu > ETHERMTU) {
3062                 /* aka IXGBE_MAXFRS on 82599 and newer */
3063                 mhadd = IXGBE_READ_REG(hw, IXGBE_MHADD);
3064                 mhadd &= ~IXGBE_MHADD_MFS_MASK;
3065                 mhadd |= sc->max_frame_size << IXGBE_MHADD_MFS_SHIFT;
3066                 IXGBE_WRITE_REG(hw, IXGBE_MHADD, mhadd);
3067         }
3068
3069         /* Now enable all the queues */
3070         for (i = 0, tx_que = sc->tx_queues; i < sc->num_tx_queues; i++, tx_que++) {
3071                 struct tx_ring *txr = &tx_que->txr;
3072
3073                 txdctl = IXGBE_READ_REG(hw, IXGBE_TXDCTL(txr->me));
3074                 txdctl |= IXGBE_TXDCTL_ENABLE;
3075                 /* Set WTHRESH to 8, burst writeback */
3076                 txdctl |= (8 << 16);
3077                 /*
3078                  * When the internal queue falls below PTHRESH (32),
3079                  * start prefetching as long as there are at least
3080                  * HTHRESH (1) buffers ready. The values are taken
3081                  * from the Intel linux driver 3.8.21.
3082                  * Prefetching enables tx line rate even with 1 queue.
3083                  */
3084                 txdctl |= (32 << 0) | (1 << 8);
3085                 IXGBE_WRITE_REG(hw, IXGBE_TXDCTL(txr->me), txdctl);
3086         }
3087
3088         for (i = 0, rx_que = sc->rx_queues; i < sc->num_rx_queues; i++, rx_que++) {
3089                 struct rx_ring *rxr = &rx_que->rxr;
3090
3091                 rxdctl = IXGBE_READ_REG(hw, IXGBE_RXDCTL(rxr->me));
3092                 if (hw->mac.type == ixgbe_mac_82598EB) {
3093                         /*
3094                          * PTHRESH = 21
3095                          * HTHRESH = 4
3096                          * WTHRESH = 8
3097                          */
3098                         rxdctl &= ~0x3FFFFF;
3099                         rxdctl |= 0x080420;
3100                 }
3101                 rxdctl |= IXGBE_RXDCTL_ENABLE;
3102                 IXGBE_WRITE_REG(hw, IXGBE_RXDCTL(rxr->me), rxdctl);
3103                 for (j = 0; j < 10; j++) {
3104                         if (IXGBE_READ_REG(hw, IXGBE_RXDCTL(rxr->me)) &
3105                             IXGBE_RXDCTL_ENABLE)
3106                                 break;
3107                         else
3108                                 msec_delay(1);
3109                 }
3110                 wmb();
3111         }
3112
3113         /* Enable Receive engine */
3114         rxctrl = IXGBE_READ_REG(hw, IXGBE_RXCTRL);
3115         if (hw->mac.type == ixgbe_mac_82598EB)
3116                 rxctrl |= IXGBE_RXCTRL_DMBYPS;
3117         rxctrl |= IXGBE_RXCTRL_RXEN;
3118         ixgbe_enable_rx_dma(hw, rxctrl);
3119
3120         /* Set up MSI/MSI-X routing */
3121         if (ixgbe_enable_msix)  {
3122                 ixgbe_configure_ivars(sc);
3123                 /* Set up auto-mask */
3124                 if (hw->mac.type == ixgbe_mac_82598EB)
3125                         IXGBE_WRITE_REG(hw, IXGBE_EIAM, IXGBE_EICS_RTX_QUEUE);
3126                 else {
3127                         IXGBE_WRITE_REG(hw, IXGBE_EIAM_EX(0), 0xFFFFFFFF);
3128                         IXGBE_WRITE_REG(hw, IXGBE_EIAM_EX(1), 0xFFFFFFFF);
3129                 }
3130         } else {  /* Simple settings for Legacy/MSI */
3131                 ixgbe_set_ivar(sc, 0, 0, 0);
3132                 ixgbe_set_ivar(sc, 0, 0, 1);
3133                 IXGBE_WRITE_REG(hw, IXGBE_EIAM, IXGBE_EICS_RTX_QUEUE);
3134         }
3135
3136         ixgbe_init_fdir(sc);
3137
3138         /*
3139          * Check on any SFP devices that
3140          * need to be kick-started
3141          */
3142         if (hw->phy.type == ixgbe_phy_none) {
3143                 err = hw->phy.ops.identify(hw);
3144                 if (err == IXGBE_ERR_SFP_NOT_SUPPORTED) {
3145                         device_printf(dev,
3146                             "Unsupported SFP+ module type was detected.\n");
3147                         return;
3148                 }
3149         }
3150
3151         /* Set moderation on the Link interrupt */
3152         IXGBE_WRITE_REG(hw, IXGBE_EITR(sc->vector), IXGBE_LINK_ITR);
3153
3154         /* Enable power to the phy. */
3155         ixgbe_set_phy_power(hw, true);
3156
3157         /* Config/Enable Link */
3158         ixgbe_config_link(ctx);
3159
3160         /* Hardware Packet Buffer & Flow Control setup */
3161         ixgbe_config_delay_values(sc);
3162
3163         /* Initialize the FC settings */
3164         ixgbe_start_hw(hw);
3165
3166         /* Set up VLAN support and filter */
3167         ixgbe_setup_vlan_hw_support(ctx);
3168
3169         /* Setup DMA Coalescing */
3170         ixgbe_config_dmac(sc);
3171
3172         /* And now turn on interrupts */
3173         ixgbe_if_enable_intr(ctx);
3174
3175         /* Enable the use of the MBX by the VF's */
3176         if (sc->feat_en & IXGBE_FEATURE_SRIOV) {
3177                 ctrl_ext = IXGBE_READ_REG(hw, IXGBE_CTRL_EXT);
3178                 ctrl_ext |= IXGBE_CTRL_EXT_PFRSTD;
3179                 IXGBE_WRITE_REG(hw, IXGBE_CTRL_EXT, ctrl_ext);
3180         }
3181
3182 } /* ixgbe_init_locked */
3183
3184 /************************************************************************
3185  * ixgbe_set_ivar
3186  *
3187  *   Setup the correct IVAR register for a particular MSI-X interrupt
3188  *     (yes this is all very magic and confusing :)
3189  *    - entry is the register array entry
3190  *    - vector is the MSI-X vector for this queue
3191  *    - type is RX/TX/MISC
3192  ************************************************************************/
3193 static void
3194 ixgbe_set_ivar(struct ixgbe_softc *sc, u8 entry, u8 vector, s8 type)
3195 {
3196         struct ixgbe_hw *hw = &sc->hw;
3197         u32 ivar, index;
3198
3199         vector |= IXGBE_IVAR_ALLOC_VAL;
3200
3201         switch (hw->mac.type) {
3202         case ixgbe_mac_82598EB:
3203                 if (type == -1)
3204                         entry = IXGBE_IVAR_OTHER_CAUSES_INDEX;
3205                 else
3206                         entry += (type * 64);
3207                 index = (entry >> 2) & 0x1F;
3208                 ivar = IXGBE_READ_REG(hw, IXGBE_IVAR(index));
3209                 ivar &= ~(0xFF << (8 * (entry & 0x3)));
3210                 ivar |= (vector << (8 * (entry & 0x3)));
3211                 IXGBE_WRITE_REG(&sc->hw, IXGBE_IVAR(index), ivar);
3212                 break;
3213         case ixgbe_mac_82599EB:
3214         case ixgbe_mac_X540:
3215         case ixgbe_mac_X550:
3216         case ixgbe_mac_X550EM_x:
3217         case ixgbe_mac_X550EM_a:
3218                 if (type == -1) { /* MISC IVAR */
3219                         index = (entry & 1) * 8;
3220                         ivar = IXGBE_READ_REG(hw, IXGBE_IVAR_MISC);
3221                         ivar &= ~(0xFF << index);
3222                         ivar |= (vector << index);
3223                         IXGBE_WRITE_REG(hw, IXGBE_IVAR_MISC, ivar);
3224                 } else {          /* RX/TX IVARS */
3225                         index = (16 * (entry & 1)) + (8 * type);
3226                         ivar = IXGBE_READ_REG(hw, IXGBE_IVAR(entry >> 1));
3227                         ivar &= ~(0xFF << index);
3228                         ivar |= (vector << index);
3229                         IXGBE_WRITE_REG(hw, IXGBE_IVAR(entry >> 1), ivar);
3230                 }
3231         default:
3232                 break;
3233         }
3234 } /* ixgbe_set_ivar */
3235
3236 /************************************************************************
3237  * ixgbe_configure_ivars
3238  ************************************************************************/
3239 static void
3240 ixgbe_configure_ivars(struct ixgbe_softc *sc)
3241 {
3242         struct ix_rx_queue *rx_que = sc->rx_queues;
3243         struct ix_tx_queue *tx_que = sc->tx_queues;
3244         u32                newitr;
3245
3246         if (ixgbe_max_interrupt_rate > 0)
3247                 newitr = (4000000 / ixgbe_max_interrupt_rate) & 0x0FF8;
3248         else {
3249                 /*
3250                  * Disable DMA coalescing if interrupt moderation is
3251                  * disabled.
3252                  */
3253                 sc->dmac = 0;
3254                 newitr = 0;
3255         }
3256
3257         for (int i = 0; i < sc->num_rx_queues; i++, rx_que++) {
3258                 struct rx_ring *rxr = &rx_que->rxr;
3259
3260                 /* First the RX queue entry */
3261                 ixgbe_set_ivar(sc, rxr->me, rx_que->msix, 0);
3262
3263                 /* Set an Initial EITR value */
3264                 IXGBE_WRITE_REG(&sc->hw, IXGBE_EITR(rx_que->msix), newitr);
3265         }
3266         for (int i = 0; i < sc->num_tx_queues; i++, tx_que++) {
3267                 struct tx_ring *txr = &tx_que->txr;
3268
3269                 /* ... and the TX */
3270                 ixgbe_set_ivar(sc, txr->me, tx_que->msix, 1);
3271         }
3272         /* For the Link interrupt */
3273         ixgbe_set_ivar(sc, 1, sc->vector, -1);
3274 } /* ixgbe_configure_ivars */
3275
3276 /************************************************************************
3277  * ixgbe_config_gpie
3278  ************************************************************************/
3279 static void
3280 ixgbe_config_gpie(struct ixgbe_softc *sc)
3281 {
3282         struct ixgbe_hw *hw = &sc->hw;
3283         u32             gpie;
3284
3285         gpie = IXGBE_READ_REG(hw, IXGBE_GPIE);
3286
3287         if (sc->intr_type == IFLIB_INTR_MSIX) {
3288                 /* Enable Enhanced MSI-X mode */
3289                 gpie |= IXGBE_GPIE_MSIX_MODE
3290                      |  IXGBE_GPIE_EIAME
3291                      |  IXGBE_GPIE_PBA_SUPPORT
3292                      |  IXGBE_GPIE_OCD;
3293         }
3294
3295         /* Fan Failure Interrupt */
3296         if (sc->feat_en & IXGBE_FEATURE_FAN_FAIL)
3297                 gpie |= IXGBE_SDP1_GPIEN;
3298
3299         /* Thermal Sensor Interrupt */
3300         if (sc->feat_en & IXGBE_FEATURE_TEMP_SENSOR)
3301                 gpie |= IXGBE_SDP0_GPIEN_X540;
3302
3303         /* Link detection */
3304         switch (hw->mac.type) {
3305         case ixgbe_mac_82599EB:
3306                 gpie |= IXGBE_SDP1_GPIEN | IXGBE_SDP2_GPIEN;
3307                 break;
3308         case ixgbe_mac_X550EM_x:
3309         case ixgbe_mac_X550EM_a:
3310                 gpie |= IXGBE_SDP0_GPIEN_X540;
3311                 break;
3312         default:
3313                 break;
3314         }
3315
3316         IXGBE_WRITE_REG(hw, IXGBE_GPIE, gpie);
3317
3318 } /* ixgbe_config_gpie */
3319
3320 /************************************************************************
3321  * ixgbe_config_delay_values
3322  *
3323  *   Requires sc->max_frame_size to be set.
3324  ************************************************************************/
3325 static void
3326 ixgbe_config_delay_values(struct ixgbe_softc *sc)
3327 {
3328         struct ixgbe_hw *hw = &sc->hw;
3329         u32             rxpb, frame, size, tmp;
3330
3331         frame = sc->max_frame_size;
3332
3333         /* Calculate High Water */
3334         switch (hw->mac.type) {
3335         case ixgbe_mac_X540:
3336         case ixgbe_mac_X550:
3337         case ixgbe_mac_X550EM_x:
3338         case ixgbe_mac_X550EM_a:
3339                 tmp = IXGBE_DV_X540(frame, frame);
3340                 break;
3341         default:
3342                 tmp = IXGBE_DV(frame, frame);
3343                 break;
3344         }
3345         size = IXGBE_BT2KB(tmp);
3346         rxpb = IXGBE_READ_REG(hw, IXGBE_RXPBSIZE(0)) >> 10;
3347         hw->fc.high_water[0] = rxpb - size;
3348
3349         /* Now calculate Low Water */
3350         switch (hw->mac.type) {
3351         case ixgbe_mac_X540:
3352         case ixgbe_mac_X550:
3353         case ixgbe_mac_X550EM_x:
3354         case ixgbe_mac_X550EM_a:
3355                 tmp = IXGBE_LOW_DV_X540(frame);
3356                 break;
3357         default:
3358                 tmp = IXGBE_LOW_DV(frame);
3359                 break;
3360         }
3361         hw->fc.low_water[0] = IXGBE_BT2KB(tmp);
3362
3363         hw->fc.pause_time = IXGBE_FC_PAUSE;
3364         hw->fc.send_xon = true;
3365 } /* ixgbe_config_delay_values */
3366
3367 /************************************************************************
3368  * ixgbe_set_multi - Multicast Update
3369  *
3370  *   Called whenever multicast address list is updated.
3371  ************************************************************************/
3372 static u_int
3373 ixgbe_mc_filter_apply(void *arg, struct sockaddr_dl *sdl, u_int idx)
3374 {
3375         struct ixgbe_softc *sc = arg;
3376         struct ixgbe_mc_addr *mta = sc->mta;
3377
3378         if (idx == MAX_NUM_MULTICAST_ADDRESSES)
3379                 return (0);
3380         bcopy(LLADDR(sdl), mta[idx].addr, IXGBE_ETH_LENGTH_OF_ADDRESS);
3381         mta[idx].vmdq = sc->pool;
3382
3383         return (1);
3384 } /* ixgbe_mc_filter_apply */
3385
3386 static void
3387 ixgbe_if_multi_set(if_ctx_t ctx)
3388 {
3389         struct ixgbe_softc       *sc = iflib_get_softc(ctx);
3390         struct ixgbe_mc_addr *mta;
3391         struct ifnet         *ifp = iflib_get_ifp(ctx);
3392         u8                   *update_ptr;
3393         u32                  fctrl;
3394         u_int                mcnt;
3395
3396         IOCTL_DEBUGOUT("ixgbe_if_multi_set: begin");
3397
3398         mta = sc->mta;
3399         bzero(mta, sizeof(*mta) * MAX_NUM_MULTICAST_ADDRESSES);
3400
3401         mcnt = if_foreach_llmaddr(iflib_get_ifp(ctx), ixgbe_mc_filter_apply, sc);
3402
3403         fctrl = IXGBE_READ_REG(&sc->hw, IXGBE_FCTRL);
3404
3405         if (ifp->if_flags & IFF_PROMISC)
3406                 fctrl |= (IXGBE_FCTRL_UPE | IXGBE_FCTRL_MPE);
3407         else if (mcnt >= MAX_NUM_MULTICAST_ADDRESSES ||
3408             ifp->if_flags & IFF_ALLMULTI) {
3409                 fctrl |= IXGBE_FCTRL_MPE;
3410                 fctrl &= ~IXGBE_FCTRL_UPE;
3411         } else
3412                 fctrl &= ~(IXGBE_FCTRL_UPE | IXGBE_FCTRL_MPE);
3413
3414         IXGBE_WRITE_REG(&sc->hw, IXGBE_FCTRL, fctrl);
3415
3416         if (mcnt < MAX_NUM_MULTICAST_ADDRESSES) {
3417                 update_ptr = (u8 *)mta;
3418                 ixgbe_update_mc_addr_list(&sc->hw, update_ptr, mcnt,
3419                     ixgbe_mc_array_itr, true);
3420         }
3421
3422 } /* ixgbe_if_multi_set */
3423
3424 /************************************************************************
3425  * ixgbe_mc_array_itr
3426  *
3427  *   An iterator function needed by the multicast shared code.
3428  *   It feeds the shared code routine the addresses in the
3429  *   array of ixgbe_set_multi() one by one.
3430  ************************************************************************/
3431 static u8 *
3432 ixgbe_mc_array_itr(struct ixgbe_hw *hw, u8 **update_ptr, u32 *vmdq)
3433 {
3434         struct ixgbe_mc_addr *mta;
3435
3436         mta = (struct ixgbe_mc_addr *)*update_ptr;
3437         *vmdq = mta->vmdq;
3438
3439         *update_ptr = (u8*)(mta + 1);
3440
3441         return (mta->addr);
3442 } /* ixgbe_mc_array_itr */
3443
3444 /************************************************************************
3445  * ixgbe_local_timer - Timer routine
3446  *
3447  *   Checks for link status, updates statistics,
3448  *   and runs the watchdog check.
3449  ************************************************************************/
3450 static void
3451 ixgbe_if_timer(if_ctx_t ctx, uint16_t qid)
3452 {
3453         struct ixgbe_softc *sc = iflib_get_softc(ctx);
3454
3455         if (qid != 0)
3456                 return;
3457
3458         /* Check for pluggable optics */
3459         if (sc->sfp_probe)
3460                 if (!ixgbe_sfp_probe(ctx))
3461                         return; /* Nothing to do */
3462
3463         ixgbe_check_link(&sc->hw, &sc->link_speed, &sc->link_up, 0);
3464
3465         /* Fire off the adminq task */
3466         iflib_admin_intr_deferred(ctx);
3467
3468 } /* ixgbe_if_timer */
3469
3470 /************************************************************************
3471  * ixgbe_sfp_probe
3472  *
3473  *   Determine if a port had optics inserted.
3474  ************************************************************************/
3475 static bool
3476 ixgbe_sfp_probe(if_ctx_t ctx)
3477 {
3478         struct ixgbe_softc  *sc = iflib_get_softc(ctx);
3479         struct ixgbe_hw *hw = &sc->hw;
3480         device_t        dev = iflib_get_dev(ctx);
3481         bool            result = false;
3482
3483         if ((hw->phy.type == ixgbe_phy_nl) &&
3484             (hw->phy.sfp_type == ixgbe_sfp_type_not_present)) {
3485                 s32 ret = hw->phy.ops.identify_sfp(hw);
3486                 if (ret)
3487                         goto out;
3488                 ret = hw->phy.ops.reset(hw);
3489                 sc->sfp_probe = false;
3490                 if (ret == IXGBE_ERR_SFP_NOT_SUPPORTED) {
3491                         device_printf(dev, "Unsupported SFP+ module detected!");
3492                         device_printf(dev,
3493                             "Reload driver with supported module.\n");
3494                         goto out;
3495                 } else
3496                         device_printf(dev, "SFP+ module detected!\n");
3497                 /* We now have supported optics */
3498                 result = true;
3499         }
3500 out:
3501
3502         return (result);
3503 } /* ixgbe_sfp_probe */
3504
3505 /************************************************************************
3506  * ixgbe_handle_mod - Tasklet for SFP module interrupts
3507  ************************************************************************/
3508 static void
3509 ixgbe_handle_mod(void *context)
3510 {
3511         if_ctx_t        ctx = context;
3512         struct ixgbe_softc  *sc = iflib_get_softc(ctx);
3513         struct ixgbe_hw *hw = &sc->hw;
3514         device_t        dev = iflib_get_dev(ctx);
3515         u32             err, cage_full = 0;
3516
3517         if (sc->hw.need_crosstalk_fix) {
3518                 switch (hw->mac.type) {
3519                 case ixgbe_mac_82599EB:
3520                         cage_full = IXGBE_READ_REG(hw, IXGBE_ESDP) &
3521                             IXGBE_ESDP_SDP2;
3522                         break;
3523                 case ixgbe_mac_X550EM_x:
3524                 case ixgbe_mac_X550EM_a:
3525                         cage_full = IXGBE_READ_REG(hw, IXGBE_ESDP) &
3526                             IXGBE_ESDP_SDP0;
3527                         break;
3528                 default:
3529                         break;
3530                 }
3531
3532                 if (!cage_full)
3533                         goto handle_mod_out;
3534         }
3535
3536         err = hw->phy.ops.identify_sfp(hw);
3537         if (err == IXGBE_ERR_SFP_NOT_SUPPORTED) {
3538                 device_printf(dev,
3539                     "Unsupported SFP+ module type was detected.\n");
3540                 goto handle_mod_out;
3541         }
3542
3543         if (hw->mac.type == ixgbe_mac_82598EB)
3544                 err = hw->phy.ops.reset(hw);
3545         else
3546                 err = hw->mac.ops.setup_sfp(hw);
3547
3548         if (err == IXGBE_ERR_SFP_NOT_SUPPORTED) {
3549                 device_printf(dev,
3550                     "Setup failure - unsupported SFP+ module type.\n");
3551                 goto handle_mod_out;
3552         }
3553         sc->task_requests |= IXGBE_REQUEST_TASK_MSF;
3554         return;
3555
3556 handle_mod_out:
3557         sc->task_requests &= ~(IXGBE_REQUEST_TASK_MSF);
3558 } /* ixgbe_handle_mod */
3559
3560
3561 /************************************************************************
3562  * ixgbe_handle_msf - Tasklet for MSF (multispeed fiber) interrupts
3563  ************************************************************************/
3564 static void
3565 ixgbe_handle_msf(void *context)
3566 {
3567         if_ctx_t        ctx = context;
3568         struct ixgbe_softc  *sc = iflib_get_softc(ctx);
3569         struct ixgbe_hw *hw = &sc->hw;
3570         u32             autoneg;
3571         bool            negotiate;
3572
3573         /* get_supported_phy_layer will call hw->phy.ops.identify_sfp() */
3574         sc->phy_layer = ixgbe_get_supported_physical_layer(hw);
3575
3576         autoneg = hw->phy.autoneg_advertised;
3577         if ((!autoneg) && (hw->mac.ops.get_link_capabilities))
3578                 hw->mac.ops.get_link_capabilities(hw, &autoneg, &negotiate);
3579         if (hw->mac.ops.setup_link)
3580                 hw->mac.ops.setup_link(hw, autoneg, true);
3581
3582         /* Adjust media types shown in ifconfig */
3583         ifmedia_removeall(sc->media);
3584         ixgbe_add_media_types(sc->ctx);
3585         ifmedia_set(sc->media, IFM_ETHER | IFM_AUTO);
3586 } /* ixgbe_handle_msf */
3587
3588 /************************************************************************
3589  * ixgbe_handle_phy - Tasklet for external PHY interrupts
3590  ************************************************************************/
3591 static void
3592 ixgbe_handle_phy(void *context)
3593 {
3594         if_ctx_t        ctx = context;
3595         struct ixgbe_softc  *sc = iflib_get_softc(ctx);
3596         struct ixgbe_hw *hw = &sc->hw;
3597         int             error;
3598
3599         error = hw->phy.ops.handle_lasi(hw);
3600         if (error == IXGBE_ERR_OVERTEMP)
3601                 device_printf(sc->dev, "CRITICAL: EXTERNAL PHY OVER TEMP!!  PHY will downshift to lower power state!\n");
3602         else if (error)
3603                 device_printf(sc->dev,
3604                     "Error handling LASI interrupt: %d\n", error);
3605 } /* ixgbe_handle_phy */
3606
3607 /************************************************************************
3608  * ixgbe_if_stop - Stop the hardware
3609  *
3610  *   Disables all traffic on the adapter by issuing a
3611  *   global reset on the MAC and deallocates TX/RX buffers.
3612  ************************************************************************/
3613 static void
3614 ixgbe_if_stop(if_ctx_t ctx)
3615 {
3616         struct ixgbe_softc  *sc = iflib_get_softc(ctx);
3617         struct ixgbe_hw *hw = &sc->hw;
3618
3619         INIT_DEBUGOUT("ixgbe_if_stop: begin\n");
3620
3621         ixgbe_reset_hw(hw);
3622         hw->adapter_stopped = false;
3623         ixgbe_stop_adapter(hw);
3624         if (hw->mac.type == ixgbe_mac_82599EB)
3625                 ixgbe_stop_mac_link_on_d3_82599(hw);
3626         /* Turn off the laser - noop with no optics */
3627         ixgbe_disable_tx_laser(hw);
3628
3629         /* Update the stack */
3630         sc->link_up = false;
3631         ixgbe_if_update_admin_status(ctx);
3632
3633         /* reprogram the RAR[0] in case user changed it. */
3634         ixgbe_set_rar(&sc->hw, 0, sc->hw.mac.addr, 0, IXGBE_RAH_AV);
3635
3636         return;
3637 } /* ixgbe_if_stop */
3638
3639 /************************************************************************
3640  * ixgbe_update_link_status - Update OS on link state
3641  *
3642  * Note: Only updates the OS on the cached link state.
3643  *       The real check of the hardware only happens with
3644  *       a link interrupt.
3645  ************************************************************************/
3646 static void
3647 ixgbe_if_update_admin_status(if_ctx_t ctx)
3648 {
3649         struct ixgbe_softc *sc = iflib_get_softc(ctx);
3650         device_t       dev = iflib_get_dev(ctx);
3651
3652         if (sc->link_up) {
3653                 if (sc->link_active == false) {
3654                         if (bootverbose)
3655                                 device_printf(dev, "Link is up %d Gbps %s \n",
3656                                     ((sc->link_speed == 128) ? 10 : 1),
3657                                     "Full Duplex");
3658                         sc->link_active = true;
3659                         /* Update any Flow Control changes */
3660                         ixgbe_fc_enable(&sc->hw);
3661                         /* Update DMA coalescing config */
3662                         ixgbe_config_dmac(sc);
3663                         /* should actually be negotiated value */
3664                         iflib_link_state_change(ctx, LINK_STATE_UP, IF_Gbps(10));
3665
3666                         if (sc->feat_en & IXGBE_FEATURE_SRIOV)
3667                                 ixgbe_ping_all_vfs(sc);
3668                 }
3669         } else { /* Link down */
3670                 if (sc->link_active == true) {
3671                         if (bootverbose)
3672                                 device_printf(dev, "Link is Down\n");
3673                         iflib_link_state_change(ctx, LINK_STATE_DOWN, 0);
3674                         sc->link_active = false;
3675                         if (sc->feat_en & IXGBE_FEATURE_SRIOV)
3676                                 ixgbe_ping_all_vfs(sc);
3677                 }
3678         }
3679
3680         /* Handle task requests from msix_link() */
3681         if (sc->task_requests & IXGBE_REQUEST_TASK_MOD)
3682                 ixgbe_handle_mod(ctx);
3683         if (sc->task_requests & IXGBE_REQUEST_TASK_MSF)
3684                 ixgbe_handle_msf(ctx);
3685         if (sc->task_requests & IXGBE_REQUEST_TASK_MBX)
3686                 ixgbe_handle_mbx(ctx);
3687         if (sc->task_requests & IXGBE_REQUEST_TASK_FDIR)
3688                 ixgbe_reinit_fdir(ctx);
3689         if (sc->task_requests & IXGBE_REQUEST_TASK_PHY)
3690                 ixgbe_handle_phy(ctx);
3691         sc->task_requests = 0;
3692
3693         ixgbe_update_stats_counters(sc);
3694 } /* ixgbe_if_update_admin_status */
3695
3696 /************************************************************************
3697  * ixgbe_config_dmac - Configure DMA Coalescing
3698  ************************************************************************/
3699 static void
3700 ixgbe_config_dmac(struct ixgbe_softc *sc)
3701 {
3702         struct ixgbe_hw          *hw = &sc->hw;
3703         struct ixgbe_dmac_config *dcfg = &hw->mac.dmac_config;
3704
3705         if (hw->mac.type < ixgbe_mac_X550 || !hw->mac.ops.dmac_config)
3706                 return;
3707
3708         if (dcfg->watchdog_timer ^ sc->dmac ||
3709             dcfg->link_speed ^ sc->link_speed) {
3710                 dcfg->watchdog_timer = sc->dmac;
3711                 dcfg->fcoe_en = false;
3712                 dcfg->link_speed = sc->link_speed;
3713                 dcfg->num_tcs = 1;
3714
3715                 INIT_DEBUGOUT2("dmac settings: watchdog %d, link speed %d\n",
3716                     dcfg->watchdog_timer, dcfg->link_speed);
3717
3718                 hw->mac.ops.dmac_config(hw);
3719         }
3720 } /* ixgbe_config_dmac */
3721
3722 /************************************************************************
3723  * ixgbe_if_enable_intr
3724  ************************************************************************/
3725 void
3726 ixgbe_if_enable_intr(if_ctx_t ctx)
3727 {
3728         struct ixgbe_softc     *sc = iflib_get_softc(ctx);
3729         struct ixgbe_hw    *hw = &sc->hw;
3730         struct ix_rx_queue *que = sc->rx_queues;
3731         u32                mask, fwsm;
3732
3733         mask = (IXGBE_EIMS_ENABLE_MASK & ~IXGBE_EIMS_RTX_QUEUE);
3734
3735         switch (sc->hw.mac.type) {
3736         case ixgbe_mac_82599EB:
3737                 mask |= IXGBE_EIMS_ECC;
3738                 /* Temperature sensor on some scs */
3739                 mask |= IXGBE_EIMS_GPI_SDP0;
3740                 /* SFP+ (RX_LOS_N & MOD_ABS_N) */
3741                 mask |= IXGBE_EIMS_GPI_SDP1;
3742                 mask |= IXGBE_EIMS_GPI_SDP2;
3743                 break;
3744         case ixgbe_mac_X540:
3745                 /* Detect if Thermal Sensor is enabled */
3746                 fwsm = IXGBE_READ_REG(hw, IXGBE_FWSM);
3747                 if (fwsm & IXGBE_FWSM_TS_ENABLED)
3748                         mask |= IXGBE_EIMS_TS;
3749                 mask |= IXGBE_EIMS_ECC;
3750                 break;
3751         case ixgbe_mac_X550:
3752                 /* MAC thermal sensor is automatically enabled */
3753                 mask |= IXGBE_EIMS_TS;
3754                 mask |= IXGBE_EIMS_ECC;
3755                 break;
3756         case ixgbe_mac_X550EM_x:
3757         case ixgbe_mac_X550EM_a:
3758                 /* Some devices use SDP0 for important information */
3759                 if (hw->device_id == IXGBE_DEV_ID_X550EM_X_SFP ||
3760                     hw->device_id == IXGBE_DEV_ID_X550EM_A_SFP ||
3761                     hw->device_id == IXGBE_DEV_ID_X550EM_A_SFP_N ||
3762                     hw->device_id == IXGBE_DEV_ID_X550EM_X_10G_T)
3763                         mask |= IXGBE_EIMS_GPI_SDP0_BY_MAC(hw);
3764                 if (hw->phy.type == ixgbe_phy_x550em_ext_t)
3765                         mask |= IXGBE_EICR_GPI_SDP0_X540;
3766                 mask |= IXGBE_EIMS_ECC;
3767                 break;
3768         default:
3769                 break;
3770         }
3771
3772         /* Enable Fan Failure detection */
3773         if (sc->feat_en & IXGBE_FEATURE_FAN_FAIL)
3774                 mask |= IXGBE_EIMS_GPI_SDP1;
3775         /* Enable SR-IOV */
3776         if (sc->feat_en & IXGBE_FEATURE_SRIOV)
3777                 mask |= IXGBE_EIMS_MAILBOX;
3778         /* Enable Flow Director */
3779         if (sc->feat_en & IXGBE_FEATURE_FDIR)
3780                 mask |= IXGBE_EIMS_FLOW_DIR;
3781
3782         IXGBE_WRITE_REG(hw, IXGBE_EIMS, mask);
3783
3784         /* With MSI-X we use auto clear */
3785         if (sc->intr_type == IFLIB_INTR_MSIX) {
3786                 mask = IXGBE_EIMS_ENABLE_MASK;
3787                 /* Don't autoclear Link */
3788                 mask &= ~IXGBE_EIMS_OTHER;
3789                 mask &= ~IXGBE_EIMS_LSC;
3790                 if (sc->feat_cap & IXGBE_FEATURE_SRIOV)
3791                         mask &= ~IXGBE_EIMS_MAILBOX;
3792                 IXGBE_WRITE_REG(hw, IXGBE_EIAC, mask);
3793         }
3794
3795         /*
3796          * Now enable all queues, this is done separately to
3797          * allow for handling the extended (beyond 32) MSI-X
3798          * vectors that can be used by 82599
3799          */
3800         for (int i = 0; i < sc->num_rx_queues; i++, que++)
3801                 ixgbe_enable_queue(sc, que->msix);
3802
3803         IXGBE_WRITE_FLUSH(hw);
3804
3805 } /* ixgbe_if_enable_intr */
3806
3807 /************************************************************************
3808  * ixgbe_disable_intr
3809  ************************************************************************/
3810 static void
3811 ixgbe_if_disable_intr(if_ctx_t ctx)
3812 {
3813         struct ixgbe_softc *sc = iflib_get_softc(ctx);
3814
3815         if (sc->intr_type == IFLIB_INTR_MSIX)
3816                 IXGBE_WRITE_REG(&sc->hw, IXGBE_EIAC, 0);
3817         if (sc->hw.mac.type == ixgbe_mac_82598EB) {
3818                 IXGBE_WRITE_REG(&sc->hw, IXGBE_EIMC, ~0);
3819         } else {
3820                 IXGBE_WRITE_REG(&sc->hw, IXGBE_EIMC, 0xFFFF0000);
3821                 IXGBE_WRITE_REG(&sc->hw, IXGBE_EIMC_EX(0), ~0);
3822                 IXGBE_WRITE_REG(&sc->hw, IXGBE_EIMC_EX(1), ~0);
3823         }
3824         IXGBE_WRITE_FLUSH(&sc->hw);
3825
3826 } /* ixgbe_if_disable_intr */
3827
3828 /************************************************************************
3829  * ixgbe_link_intr_enable
3830  ************************************************************************/
3831 static void
3832 ixgbe_link_intr_enable(if_ctx_t ctx)
3833 {
3834         struct ixgbe_hw *hw = &((struct ixgbe_softc *)iflib_get_softc(ctx))->hw;
3835
3836         /* Re-enable other interrupts */
3837         IXGBE_WRITE_REG(hw, IXGBE_EIMS, IXGBE_EIMS_OTHER | IXGBE_EIMS_LSC);
3838 } /* ixgbe_link_intr_enable */
3839
3840 /************************************************************************
3841  * ixgbe_if_rx_queue_intr_enable
3842  ************************************************************************/
3843 static int
3844 ixgbe_if_rx_queue_intr_enable(if_ctx_t ctx, uint16_t rxqid)
3845 {
3846         struct ixgbe_softc     *sc = iflib_get_softc(ctx);
3847         struct ix_rx_queue *que = &sc->rx_queues[rxqid];
3848
3849         ixgbe_enable_queue(sc, que->msix);
3850
3851         return (0);
3852 } /* ixgbe_if_rx_queue_intr_enable */
3853
3854 /************************************************************************
3855  * ixgbe_enable_queue
3856  ************************************************************************/
3857 static void
3858 ixgbe_enable_queue(struct ixgbe_softc *sc, u32 vector)
3859 {
3860         struct ixgbe_hw *hw = &sc->hw;
3861         u64             queue = 1ULL << vector;
3862         u32             mask;
3863
3864         if (hw->mac.type == ixgbe_mac_82598EB) {
3865                 mask = (IXGBE_EIMS_RTX_QUEUE & queue);
3866                 IXGBE_WRITE_REG(hw, IXGBE_EIMS, mask);
3867         } else {
3868                 mask = (queue & 0xFFFFFFFF);
3869                 if (mask)
3870                         IXGBE_WRITE_REG(hw, IXGBE_EIMS_EX(0), mask);
3871                 mask = (queue >> 32);
3872                 if (mask)
3873                         IXGBE_WRITE_REG(hw, IXGBE_EIMS_EX(1), mask);
3874         }
3875 } /* ixgbe_enable_queue */
3876
3877 /************************************************************************
3878  * ixgbe_disable_queue
3879  ************************************************************************/
3880 static void
3881 ixgbe_disable_queue(struct ixgbe_softc *sc, u32 vector)
3882 {
3883         struct ixgbe_hw *hw = &sc->hw;
3884         u64             queue = 1ULL << vector;
3885         u32             mask;
3886
3887         if (hw->mac.type == ixgbe_mac_82598EB) {
3888                 mask = (IXGBE_EIMS_RTX_QUEUE & queue);
3889                 IXGBE_WRITE_REG(hw, IXGBE_EIMC, mask);
3890         } else {
3891                 mask = (queue & 0xFFFFFFFF);
3892                 if (mask)
3893                         IXGBE_WRITE_REG(hw, IXGBE_EIMC_EX(0), mask);
3894                 mask = (queue >> 32);
3895                 if (mask)
3896                         IXGBE_WRITE_REG(hw, IXGBE_EIMC_EX(1), mask);
3897         }
3898 } /* ixgbe_disable_queue */
3899
3900 /************************************************************************
3901  * ixgbe_intr - Legacy Interrupt Service Routine
3902  ************************************************************************/
3903 int
3904 ixgbe_intr(void *arg)
3905 {
3906         struct ixgbe_softc     *sc = arg;
3907         struct ix_rx_queue *que = sc->rx_queues;
3908         struct ixgbe_hw    *hw = &sc->hw;
3909         if_ctx_t           ctx = sc->ctx;
3910         u32                eicr, eicr_mask;
3911
3912         eicr = IXGBE_READ_REG(hw, IXGBE_EICR);
3913
3914         ++que->irqs;
3915         if (eicr == 0) {
3916                 ixgbe_if_enable_intr(ctx);
3917                 return (FILTER_HANDLED);
3918         }
3919
3920         /* Check for fan failure */
3921         if ((hw->device_id == IXGBE_DEV_ID_82598AT) &&
3922             (eicr & IXGBE_EICR_GPI_SDP1)) {
3923                 device_printf(sc->dev,
3924                     "\nCRITICAL: FAN FAILURE!! REPLACE IMMEDIATELY!!\n");
3925                 IXGBE_WRITE_REG(hw, IXGBE_EIMS, IXGBE_EICR_GPI_SDP1_BY_MAC(hw));
3926         }
3927
3928         /* Link status change */
3929         if (eicr & IXGBE_EICR_LSC) {
3930                 IXGBE_WRITE_REG(hw, IXGBE_EIMC, IXGBE_EIMC_LSC);
3931                 iflib_admin_intr_deferred(ctx);
3932         }
3933
3934         if (ixgbe_is_sfp(hw)) {
3935                 /* Pluggable optics-related interrupt */
3936                 if (hw->mac.type >= ixgbe_mac_X540)
3937                         eicr_mask = IXGBE_EICR_GPI_SDP0_X540;
3938                 else
3939                         eicr_mask = IXGBE_EICR_GPI_SDP2_BY_MAC(hw);
3940
3941                 if (eicr & eicr_mask) {
3942                         IXGBE_WRITE_REG(hw, IXGBE_EICR, eicr_mask);
3943                         sc->task_requests |= IXGBE_REQUEST_TASK_MOD;
3944                 }
3945
3946                 if ((hw->mac.type == ixgbe_mac_82599EB) &&
3947                     (eicr & IXGBE_EICR_GPI_SDP1_BY_MAC(hw))) {
3948                         IXGBE_WRITE_REG(hw, IXGBE_EICR,
3949                             IXGBE_EICR_GPI_SDP1_BY_MAC(hw));
3950                         sc->task_requests |= IXGBE_REQUEST_TASK_MSF;
3951                 }
3952         }
3953
3954         /* External PHY interrupt */
3955         if ((hw->phy.type == ixgbe_phy_x550em_ext_t) &&
3956             (eicr & IXGBE_EICR_GPI_SDP0_X540))
3957                 sc->task_requests |= IXGBE_REQUEST_TASK_PHY;
3958
3959         return (FILTER_SCHEDULE_THREAD);
3960 } /* ixgbe_intr */
3961
3962 /************************************************************************
3963  * ixgbe_free_pci_resources
3964  ************************************************************************/
3965 static void
3966 ixgbe_free_pci_resources(if_ctx_t ctx)
3967 {
3968         struct ixgbe_softc *sc = iflib_get_softc(ctx);
3969         struct         ix_rx_queue *que = sc->rx_queues;
3970         device_t       dev = iflib_get_dev(ctx);
3971
3972         /* Release all MSI-X queue resources */
3973         if (sc->intr_type == IFLIB_INTR_MSIX)
3974                 iflib_irq_free(ctx, &sc->irq);
3975
3976         if (que != NULL) {
3977                 for (int i = 0; i < sc->num_rx_queues; i++, que++) {
3978                         iflib_irq_free(ctx, &que->que_irq);
3979                 }
3980         }
3981
3982         if (sc->pci_mem != NULL)
3983                 bus_release_resource(dev, SYS_RES_MEMORY,
3984                     rman_get_rid(sc->pci_mem), sc->pci_mem);
3985 } /* ixgbe_free_pci_resources */
3986
3987 /************************************************************************
3988  * ixgbe_sysctl_flowcntl
3989  *
3990  *   SYSCTL wrapper around setting Flow Control
3991  ************************************************************************/
3992 static int
3993 ixgbe_sysctl_flowcntl(SYSCTL_HANDLER_ARGS)
3994 {
3995         struct ixgbe_softc *sc;
3996         int            error, fc;
3997
3998         sc = (struct ixgbe_softc *)arg1;
3999         fc = sc->hw.fc.current_mode;
4000
4001         error = sysctl_handle_int(oidp, &fc, 0, req);
4002         if ((error) || (req->newptr == NULL))
4003                 return (error);
4004
4005         /* Don't bother if it's not changed */
4006         if (fc == sc->hw.fc.current_mode)
4007                 return (0);
4008
4009         return ixgbe_set_flowcntl(sc, fc);
4010 } /* ixgbe_sysctl_flowcntl */
4011
4012 /************************************************************************
4013  * ixgbe_set_flowcntl - Set flow control
4014  *
4015  *   Flow control values:
4016  *     0 - off
4017  *     1 - rx pause
4018  *     2 - tx pause
4019  *     3 - full
4020  ************************************************************************/
4021 static int
4022 ixgbe_set_flowcntl(struct ixgbe_softc *sc, int fc)
4023 {
4024         switch (fc) {
4025         case ixgbe_fc_rx_pause:
4026         case ixgbe_fc_tx_pause:
4027         case ixgbe_fc_full:
4028                 sc->hw.fc.requested_mode = fc;
4029                 if (sc->num_rx_queues > 1)
4030                         ixgbe_disable_rx_drop(sc);
4031                 break;
4032         case ixgbe_fc_none:
4033                 sc->hw.fc.requested_mode = ixgbe_fc_none;
4034                 if (sc->num_rx_queues > 1)
4035                         ixgbe_enable_rx_drop(sc);
4036                 break;
4037         default:
4038                 return (EINVAL);
4039         }
4040
4041         /* Don't autoneg if forcing a value */
4042         sc->hw.fc.disable_fc_autoneg = true;
4043         ixgbe_fc_enable(&sc->hw);
4044
4045         return (0);
4046 } /* ixgbe_set_flowcntl */
4047
4048 /************************************************************************
4049  * ixgbe_enable_rx_drop
4050  *
4051  *   Enable the hardware to drop packets when the buffer is
4052  *   full. This is useful with multiqueue, so that no single
4053  *   queue being full stalls the entire RX engine. We only
4054  *   enable this when Multiqueue is enabled AND Flow Control
4055  *   is disabled.
4056  ************************************************************************/
4057 static void
4058 ixgbe_enable_rx_drop(struct ixgbe_softc *sc)
4059 {
4060         struct ixgbe_hw *hw = &sc->hw;
4061         struct rx_ring  *rxr;
4062         u32             srrctl;
4063
4064         for (int i = 0; i < sc->num_rx_queues; i++) {
4065                 rxr = &sc->rx_queues[i].rxr;
4066                 srrctl = IXGBE_READ_REG(hw, IXGBE_SRRCTL(rxr->me));
4067                 srrctl |= IXGBE_SRRCTL_DROP_EN;
4068                 IXGBE_WRITE_REG(hw, IXGBE_SRRCTL(rxr->me), srrctl);
4069         }
4070
4071         /* enable drop for each vf */
4072         for (int i = 0; i < sc->num_vfs; i++) {
4073                 IXGBE_WRITE_REG(hw, IXGBE_QDE,
4074                                 (IXGBE_QDE_WRITE | (i << IXGBE_QDE_IDX_SHIFT) |
4075                                 IXGBE_QDE_ENABLE));
4076         }
4077 } /* ixgbe_enable_rx_drop */
4078
4079 /************************************************************************
4080  * ixgbe_disable_rx_drop
4081  ************************************************************************/
4082 static void
4083 ixgbe_disable_rx_drop(struct ixgbe_softc *sc)
4084 {
4085         struct ixgbe_hw *hw = &sc->hw;
4086         struct rx_ring  *rxr;
4087         u32             srrctl;
4088
4089         for (int i = 0; i < sc->num_rx_queues; i++) {
4090                 rxr = &sc->rx_queues[i].rxr;
4091                 srrctl = IXGBE_READ_REG(hw, IXGBE_SRRCTL(rxr->me));
4092                 srrctl &= ~IXGBE_SRRCTL_DROP_EN;
4093                 IXGBE_WRITE_REG(hw, IXGBE_SRRCTL(rxr->me), srrctl);
4094         }
4095
4096         /* disable drop for each vf */
4097         for (int i = 0; i < sc->num_vfs; i++) {
4098                 IXGBE_WRITE_REG(hw, IXGBE_QDE,
4099                     (IXGBE_QDE_WRITE | (i << IXGBE_QDE_IDX_SHIFT)));
4100         }
4101 } /* ixgbe_disable_rx_drop */
4102
4103 /************************************************************************
4104  * ixgbe_sysctl_advertise
4105  *
4106  *   SYSCTL wrapper around setting advertised speed
4107  ************************************************************************/
4108 static int
4109 ixgbe_sysctl_advertise(SYSCTL_HANDLER_ARGS)
4110 {
4111         struct ixgbe_softc *sc;
4112         int            error, advertise;
4113
4114         sc = (struct ixgbe_softc *)arg1;
4115         advertise = sc->advertise;
4116
4117         error = sysctl_handle_int(oidp, &advertise, 0, req);
4118         if ((error) || (req->newptr == NULL))
4119                 return (error);
4120
4121         return ixgbe_set_advertise(sc, advertise);
4122 } /* ixgbe_sysctl_advertise */
4123
4124 /************************************************************************
4125  * ixgbe_set_advertise - Control advertised link speed
4126  *
4127  *   Flags:
4128  *     0x1  - advertise 100 Mb
4129  *     0x2  - advertise 1G
4130  *     0x4  - advertise 10G
4131  *     0x8  - advertise 10 Mb (yes, Mb)
4132  *     0x10 - advertise 2.5G (disabled by default)
4133  *     0x20 - advertise 5G (disabled by default)
4134  *
4135  ************************************************************************/
4136 static int
4137 ixgbe_set_advertise(struct ixgbe_softc *sc, int advertise)
4138 {
4139         device_t         dev = iflib_get_dev(sc->ctx);
4140         struct ixgbe_hw  *hw;
4141         ixgbe_link_speed speed = 0;
4142         ixgbe_link_speed link_caps = 0;
4143         s32              err = IXGBE_NOT_IMPLEMENTED;
4144         bool             negotiate = false;
4145
4146         /* Checks to validate new value */
4147         if (sc->advertise == advertise) /* no change */
4148                 return (0);
4149
4150         hw = &sc->hw;
4151
4152         /* No speed changes for backplane media */
4153         if (hw->phy.media_type == ixgbe_media_type_backplane)
4154                 return (ENODEV);
4155
4156         if (!((hw->phy.media_type == ixgbe_media_type_copper) ||
4157               (hw->phy.multispeed_fiber))) {
4158                 device_printf(dev, "Advertised speed can only be set on copper or multispeed fiber media types.\n");
4159                 return (EINVAL);
4160         }
4161
4162         if (advertise < 0x1 || advertise > 0x3F) {
4163                 device_printf(dev, "Invalid advertised speed; valid modes are 0x1 through 0x3F\n");
4164                 return (EINVAL);
4165         }
4166
4167         if (hw->mac.ops.get_link_capabilities) {
4168                 err = hw->mac.ops.get_link_capabilities(hw, &link_caps,
4169                     &negotiate);
4170                 if (err != IXGBE_SUCCESS) {
4171                         device_printf(dev, "Unable to determine supported advertise speeds\n");
4172                         return (ENODEV);
4173                 }
4174         }
4175
4176         /* Set new value and report new advertised mode */
4177         if (advertise & 0x1) {
4178                 if (!(link_caps & IXGBE_LINK_SPEED_100_FULL)) {
4179                         device_printf(dev, "Interface does not support 100Mb advertised speed\n");
4180                         return (EINVAL);
4181                 }
4182                 speed |= IXGBE_LINK_SPEED_100_FULL;
4183         }
4184         if (advertise & 0x2) {
4185                 if (!(link_caps & IXGBE_LINK_SPEED_1GB_FULL)) {
4186                         device_printf(dev, "Interface does not support 1Gb advertised speed\n");
4187                         return (EINVAL);
4188                 }
4189                 speed |= IXGBE_LINK_SPEED_1GB_FULL;
4190         }
4191         if (advertise & 0x4) {
4192                 if (!(link_caps & IXGBE_LINK_SPEED_10GB_FULL)) {
4193                         device_printf(dev, "Interface does not support 10Gb advertised speed\n");
4194                         return (EINVAL);
4195                 }
4196                 speed |= IXGBE_LINK_SPEED_10GB_FULL;
4197         }
4198         if (advertise & 0x8) {
4199                 if (!(link_caps & IXGBE_LINK_SPEED_10_FULL)) {
4200                         device_printf(dev, "Interface does not support 10Mb advertised speed\n");
4201                         return (EINVAL);
4202                 }
4203                 speed |= IXGBE_LINK_SPEED_10_FULL;
4204         }
4205         if (advertise & 0x10) {
4206                 if (!(link_caps & IXGBE_LINK_SPEED_2_5GB_FULL)) {
4207                         device_printf(dev, "Interface does not support 2.5G advertised speed\n");
4208                         return (EINVAL);
4209                 }
4210                 speed |= IXGBE_LINK_SPEED_2_5GB_FULL;
4211         }
4212         if (advertise & 0x20) {
4213                 if (!(link_caps & IXGBE_LINK_SPEED_5GB_FULL)) {
4214                         device_printf(dev, "Interface does not support 5G advertised speed\n");
4215                         return (EINVAL);
4216                 }
4217                 speed |= IXGBE_LINK_SPEED_5GB_FULL;
4218         }
4219
4220         hw->mac.autotry_restart = true;
4221         hw->mac.ops.setup_link(hw, speed, true);
4222         sc->advertise = advertise;
4223
4224         return (0);
4225 } /* ixgbe_set_advertise */
4226
4227 /************************************************************************
4228  * ixgbe_get_default_advertise - Get default advertised speed settings
4229  *
4230  *   Formatted for sysctl usage.
4231  *   Flags:
4232  *     0x1 - advertise 100 Mb
4233  *     0x2 - advertise 1G
4234  *     0x4 - advertise 10G
4235  *     0x8 - advertise 10 Mb (yes, Mb)
4236  *     0x10 - advertise 2.5G (disabled by default)
4237  *     0x20 - advertise 5G (disabled by default)
4238  ************************************************************************/
4239 static int
4240 ixgbe_get_default_advertise(struct ixgbe_softc *sc)
4241 {
4242         struct ixgbe_hw  *hw = &sc->hw;
4243         int              speed;
4244         ixgbe_link_speed link_caps = 0;
4245         s32              err;
4246         bool             negotiate = false;
4247
4248         /*
4249          * Advertised speed means nothing unless it's copper or
4250          * multi-speed fiber
4251          */
4252         if (!(hw->phy.media_type == ixgbe_media_type_copper) &&
4253             !(hw->phy.multispeed_fiber))
4254                 return (0);
4255
4256         err = hw->mac.ops.get_link_capabilities(hw, &link_caps, &negotiate);
4257         if (err != IXGBE_SUCCESS)
4258                 return (0);
4259
4260         if (hw->mac.type == ixgbe_mac_X550) {
4261                 /*
4262                  * 2.5G and 5G autonegotiation speeds on X550
4263                  * are disabled by default due to reported
4264                  * interoperability issues with some switches.
4265                  */
4266                 link_caps &= ~(IXGBE_LINK_SPEED_2_5GB_FULL |
4267                     IXGBE_LINK_SPEED_5GB_FULL);
4268         }
4269
4270         speed =
4271             ((link_caps & IXGBE_LINK_SPEED_10GB_FULL)  ? 0x4  : 0) |
4272             ((link_caps & IXGBE_LINK_SPEED_5GB_FULL)   ? 0x20 : 0) |
4273             ((link_caps & IXGBE_LINK_SPEED_2_5GB_FULL) ? 0x10 : 0) |
4274             ((link_caps & IXGBE_LINK_SPEED_1GB_FULL)   ? 0x2  : 0) |
4275             ((link_caps & IXGBE_LINK_SPEED_100_FULL)   ? 0x1  : 0) |
4276             ((link_caps & IXGBE_LINK_SPEED_10_FULL)    ? 0x8  : 0);
4277
4278         return speed;
4279 } /* ixgbe_get_default_advertise */
4280
4281 /************************************************************************
4282  * ixgbe_sysctl_dmac - Manage DMA Coalescing
4283  *
4284  *   Control values:
4285  *     0/1 - off / on (use default value of 1000)
4286  *
4287  *     Legal timer values are:
4288  *     50,100,250,500,1000,2000,5000,10000
4289  *
4290  *     Turning off interrupt moderation will also turn this off.
4291  ************************************************************************/
4292 static int
4293 ixgbe_sysctl_dmac(SYSCTL_HANDLER_ARGS)
4294 {
4295         struct ixgbe_softc *sc = (struct ixgbe_softc *)arg1;
4296         struct ifnet   *ifp = iflib_get_ifp(sc->ctx);
4297         int            error;
4298         u16            newval;
4299
4300         newval = sc->dmac;
4301         error = sysctl_handle_16(oidp, &newval, 0, req);
4302         if ((error) || (req->newptr == NULL))
4303                 return (error);
4304
4305         switch (newval) {
4306         case 0:
4307                 /* Disabled */
4308                 sc->dmac = 0;
4309                 break;
4310         case 1:
4311                 /* Enable and use default */
4312                 sc->dmac = 1000;
4313                 break;
4314         case 50:
4315         case 100:
4316         case 250:
4317         case 500:
4318         case 1000:
4319         case 2000:
4320         case 5000:
4321         case 10000:
4322                 /* Legal values - allow */
4323                 sc->dmac = newval;
4324                 break;
4325         default:
4326                 /* Do nothing, illegal value */
4327                 return (EINVAL);
4328         }
4329
4330         /* Re-initialize hardware if it's already running */
4331         if (ifp->if_drv_flags & IFF_DRV_RUNNING)
4332                 ifp->if_init(ifp);
4333
4334         return (0);
4335 } /* ixgbe_sysctl_dmac */
4336
4337 #ifdef IXGBE_DEBUG
4338 /************************************************************************
4339  * ixgbe_sysctl_power_state
4340  *
4341  *   Sysctl to test power states
4342  *   Values:
4343  *     0      - set device to D0
4344  *     3      - set device to D3
4345  *     (none) - get current device power state
4346  ************************************************************************/
4347 static int
4348 ixgbe_sysctl_power_state(SYSCTL_HANDLER_ARGS)
4349 {
4350         struct ixgbe_softc *sc = (struct ixgbe_softc *)arg1;
4351         device_t       dev = sc->dev;
4352         int            curr_ps, new_ps, error = 0;
4353
4354         curr_ps = new_ps = pci_get_powerstate(dev);
4355
4356         error = sysctl_handle_int(oidp, &new_ps, 0, req);
4357         if ((error) || (req->newptr == NULL))
4358                 return (error);
4359
4360         if (new_ps == curr_ps)
4361                 return (0);
4362
4363         if (new_ps == 3 && curr_ps == 0)
4364                 error = DEVICE_SUSPEND(dev);
4365         else if (new_ps == 0 && curr_ps == 3)
4366                 error = DEVICE_RESUME(dev);
4367         else
4368                 return (EINVAL);
4369
4370         device_printf(dev, "New state: %d\n", pci_get_powerstate(dev));
4371
4372         return (error);
4373 } /* ixgbe_sysctl_power_state */
4374 #endif
4375
4376 /************************************************************************
4377  * ixgbe_sysctl_wol_enable
4378  *
4379  *   Sysctl to enable/disable the WoL capability,
4380  *   if supported by the adapter.
4381  *
4382  *   Values:
4383  *     0 - disabled
4384  *     1 - enabled
4385  ************************************************************************/
4386 static int
4387 ixgbe_sysctl_wol_enable(SYSCTL_HANDLER_ARGS)
4388 {
4389         struct ixgbe_softc  *sc = (struct ixgbe_softc *)arg1;
4390         struct ixgbe_hw *hw = &sc->hw;
4391         int             new_wol_enabled;
4392         int             error = 0;
4393
4394         new_wol_enabled = hw->wol_enabled;
4395         error = sysctl_handle_int(oidp, &new_wol_enabled, 0, req);
4396         if ((error) || (req->newptr == NULL))
4397                 return (error);
4398         new_wol_enabled = !!(new_wol_enabled);
4399         if (new_wol_enabled == hw->wol_enabled)
4400                 return (0);
4401
4402         if (new_wol_enabled > 0 && !sc->wol_support)
4403                 return (ENODEV);
4404         else
4405                 hw->wol_enabled = new_wol_enabled;
4406
4407         return (0);
4408 } /* ixgbe_sysctl_wol_enable */
4409
4410 /************************************************************************
4411  * ixgbe_sysctl_wufc - Wake Up Filter Control
4412  *
4413  *   Sysctl to enable/disable the types of packets that the
4414  *   adapter will wake up on upon receipt.
4415  *   Flags:
4416  *     0x1  - Link Status Change
4417  *     0x2  - Magic Packet
4418  *     0x4  - Direct Exact
4419  *     0x8  - Directed Multicast
4420  *     0x10 - Broadcast
4421  *     0x20 - ARP/IPv4 Request Packet
4422  *     0x40 - Direct IPv4 Packet
4423  *     0x80 - Direct IPv6 Packet
4424  *
4425  *   Settings not listed above will cause the sysctl to return an error.
4426  ************************************************************************/
4427 static int
4428 ixgbe_sysctl_wufc(SYSCTL_HANDLER_ARGS)
4429 {
4430         struct ixgbe_softc *sc = (struct ixgbe_softc *)arg1;
4431         int            error = 0;
4432         u32            new_wufc;
4433
4434         new_wufc = sc->wufc;
4435
4436         error = sysctl_handle_32(oidp, &new_wufc, 0, req);
4437         if ((error) || (req->newptr == NULL))
4438                 return (error);
4439         if (new_wufc == sc->wufc)
4440                 return (0);
4441
4442         if (new_wufc & 0xffffff00)
4443                 return (EINVAL);
4444
4445         new_wufc &= 0xff;
4446         new_wufc |= (0xffffff & sc->wufc);
4447         sc->wufc = new_wufc;
4448
4449         return (0);
4450 } /* ixgbe_sysctl_wufc */
4451
4452 #ifdef IXGBE_DEBUG
4453 /************************************************************************
4454  * ixgbe_sysctl_print_rss_config
4455  ************************************************************************/
4456 static int
4457 ixgbe_sysctl_print_rss_config(SYSCTL_HANDLER_ARGS)
4458 {
4459         struct ixgbe_softc  *sc = (struct ixgbe_softc *)arg1;
4460         struct ixgbe_hw *hw = &sc->hw;
4461         device_t        dev = sc->dev;
4462         struct sbuf     *buf;
4463         int             error = 0, reta_size;
4464         u32             reg;
4465
4466         buf = sbuf_new_for_sysctl(NULL, NULL, 128, req);
4467         if (!buf) {
4468                 device_printf(dev, "Could not allocate sbuf for output.\n");
4469                 return (ENOMEM);
4470         }
4471
4472         // TODO: use sbufs to make a string to print out
4473         /* Set multiplier for RETA setup and table size based on MAC */
4474         switch (sc->hw.mac.type) {
4475         case ixgbe_mac_X550:
4476         case ixgbe_mac_X550EM_x:
4477         case ixgbe_mac_X550EM_a:
4478                 reta_size = 128;
4479                 break;
4480         default:
4481                 reta_size = 32;
4482                 break;
4483         }
4484
4485         /* Print out the redirection table */
4486         sbuf_cat(buf, "\n");
4487         for (int i = 0; i < reta_size; i++) {
4488                 if (i < 32) {
4489                         reg = IXGBE_READ_REG(hw, IXGBE_RETA(i));
4490                         sbuf_printf(buf, "RETA(%2d): 0x%08x\n", i, reg);
4491                 } else {
4492                         reg = IXGBE_READ_REG(hw, IXGBE_ERETA(i - 32));
4493                         sbuf_printf(buf, "ERETA(%2d): 0x%08x\n", i - 32, reg);
4494                 }
4495         }
4496
4497         // TODO: print more config
4498
4499         error = sbuf_finish(buf);
4500         if (error)
4501                 device_printf(dev, "Error finishing sbuf: %d\n", error);
4502
4503         sbuf_delete(buf);
4504
4505         return (0);
4506 } /* ixgbe_sysctl_print_rss_config */
4507 #endif /* IXGBE_DEBUG */
4508
4509 /************************************************************************
4510  * ixgbe_sysctl_phy_temp - Retrieve temperature of PHY
4511  *
4512  *   For X552/X557-AT devices using an external PHY
4513  ************************************************************************/
4514 static int
4515 ixgbe_sysctl_phy_temp(SYSCTL_HANDLER_ARGS)
4516 {
4517         struct ixgbe_softc  *sc = (struct ixgbe_softc *)arg1;
4518         struct ixgbe_hw *hw = &sc->hw;
4519         u16             reg;
4520
4521         if (hw->device_id != IXGBE_DEV_ID_X550EM_X_10G_T) {
4522                 device_printf(iflib_get_dev(sc->ctx),
4523                     "Device has no supported external thermal sensor.\n");
4524                 return (ENODEV);
4525         }
4526
4527         if (hw->phy.ops.read_reg(hw, IXGBE_PHY_CURRENT_TEMP,
4528             IXGBE_MDIO_VENDOR_SPECIFIC_1_DEV_TYPE, &reg)) {
4529                 device_printf(iflib_get_dev(sc->ctx),
4530                     "Error reading from PHY's current temperature register\n");
4531                 return (EAGAIN);
4532         }
4533
4534         /* Shift temp for output */
4535         reg = reg >> 8;
4536
4537         return (sysctl_handle_16(oidp, NULL, reg, req));
4538 } /* ixgbe_sysctl_phy_temp */
4539
4540 /************************************************************************
4541  * ixgbe_sysctl_phy_overtemp_occurred
4542  *
4543  *   Reports (directly from the PHY) whether the current PHY
4544  *   temperature is over the overtemp threshold.
4545  ************************************************************************/
4546 static int
4547 ixgbe_sysctl_phy_overtemp_occurred(SYSCTL_HANDLER_ARGS)
4548 {
4549         struct ixgbe_softc  *sc = (struct ixgbe_softc *)arg1;
4550         struct ixgbe_hw *hw = &sc->hw;
4551         u16             reg;
4552
4553         if (hw->device_id != IXGBE_DEV_ID_X550EM_X_10G_T) {
4554                 device_printf(iflib_get_dev(sc->ctx),
4555                     "Device has no supported external thermal sensor.\n");
4556                 return (ENODEV);
4557         }
4558
4559         if (hw->phy.ops.read_reg(hw, IXGBE_PHY_OVERTEMP_STATUS,
4560             IXGBE_MDIO_VENDOR_SPECIFIC_1_DEV_TYPE, &reg)) {
4561                 device_printf(iflib_get_dev(sc->ctx),
4562                     "Error reading from PHY's temperature status register\n");
4563                 return (EAGAIN);
4564         }
4565
4566         /* Get occurrence bit */
4567         reg = !!(reg & 0x4000);
4568
4569         return (sysctl_handle_16(oidp, 0, reg, req));
4570 } /* ixgbe_sysctl_phy_overtemp_occurred */
4571
4572 /************************************************************************
4573  * ixgbe_sysctl_eee_state
4574  *
4575  *   Sysctl to set EEE power saving feature
4576  *   Values:
4577  *     0      - disable EEE
4578  *     1      - enable EEE
4579  *     (none) - get current device EEE state
4580  ************************************************************************/
4581 static int
4582 ixgbe_sysctl_eee_state(SYSCTL_HANDLER_ARGS)
4583 {
4584         struct ixgbe_softc *sc = (struct ixgbe_softc *)arg1;
4585         device_t       dev = sc->dev;
4586         struct ifnet   *ifp = iflib_get_ifp(sc->ctx);
4587         int            curr_eee, new_eee, error = 0;
4588         s32            retval;
4589
4590         curr_eee = new_eee = !!(sc->feat_en & IXGBE_FEATURE_EEE);
4591
4592         error = sysctl_handle_int(oidp, &new_eee, 0, req);
4593         if ((error) || (req->newptr == NULL))
4594                 return (error);
4595
4596         /* Nothing to do */
4597         if (new_eee == curr_eee)
4598                 return (0);
4599
4600         /* Not supported */
4601         if (!(sc->feat_cap & IXGBE_FEATURE_EEE))
4602                 return (EINVAL);
4603
4604         /* Bounds checking */
4605         if ((new_eee < 0) || (new_eee > 1))
4606                 return (EINVAL);
4607
4608         retval = ixgbe_setup_eee(&sc->hw, new_eee);
4609         if (retval) {
4610                 device_printf(dev, "Error in EEE setup: 0x%08X\n", retval);
4611                 return (EINVAL);
4612         }
4613
4614         /* Restart auto-neg */
4615         ifp->if_init(ifp);
4616
4617         device_printf(dev, "New EEE state: %d\n", new_eee);
4618
4619         /* Cache new value */
4620         if (new_eee)
4621                 sc->feat_en |= IXGBE_FEATURE_EEE;
4622         else
4623                 sc->feat_en &= ~IXGBE_FEATURE_EEE;
4624
4625         return (error);
4626 } /* ixgbe_sysctl_eee_state */
4627
4628 /************************************************************************
4629  * ixgbe_init_device_features
4630  ************************************************************************/
4631 static void
4632 ixgbe_init_device_features(struct ixgbe_softc *sc)
4633 {
4634         sc->feat_cap = IXGBE_FEATURE_NETMAP
4635                           | IXGBE_FEATURE_RSS
4636                           | IXGBE_FEATURE_MSI
4637                           | IXGBE_FEATURE_MSIX
4638                           | IXGBE_FEATURE_LEGACY_IRQ;
4639
4640         /* Set capabilities first... */
4641         switch (sc->hw.mac.type) {
4642         case ixgbe_mac_82598EB:
4643                 if (sc->hw.device_id == IXGBE_DEV_ID_82598AT)
4644                         sc->feat_cap |= IXGBE_FEATURE_FAN_FAIL;
4645                 break;
4646         case ixgbe_mac_X540:
4647                 sc->feat_cap |= IXGBE_FEATURE_SRIOV;
4648                 sc->feat_cap |= IXGBE_FEATURE_FDIR;
4649                 if ((sc->hw.device_id == IXGBE_DEV_ID_X540_BYPASS) &&
4650                     (sc->hw.bus.func == 0))
4651                         sc->feat_cap |= IXGBE_FEATURE_BYPASS;
4652                 break;
4653         case ixgbe_mac_X550:
4654                 sc->feat_cap |= IXGBE_FEATURE_TEMP_SENSOR;
4655                 sc->feat_cap |= IXGBE_FEATURE_SRIOV;
4656                 sc->feat_cap |= IXGBE_FEATURE_FDIR;
4657                 break;
4658         case ixgbe_mac_X550EM_x:
4659                 sc->feat_cap |= IXGBE_FEATURE_SRIOV;
4660                 sc->feat_cap |= IXGBE_FEATURE_FDIR;
4661                 break;
4662         case ixgbe_mac_X550EM_a:
4663                 sc->feat_cap |= IXGBE_FEATURE_SRIOV;
4664                 sc->feat_cap |= IXGBE_FEATURE_FDIR;
4665                 sc->feat_cap &= ~IXGBE_FEATURE_LEGACY_IRQ;
4666                 if ((sc->hw.device_id == IXGBE_DEV_ID_X550EM_A_1G_T) ||
4667                     (sc->hw.device_id == IXGBE_DEV_ID_X550EM_A_1G_T_L)) {
4668                         sc->feat_cap |= IXGBE_FEATURE_TEMP_SENSOR;
4669                         sc->feat_cap |= IXGBE_FEATURE_EEE;
4670                 }
4671                 break;
4672         case ixgbe_mac_82599EB:
4673                 sc->feat_cap |= IXGBE_FEATURE_SRIOV;
4674                 sc->feat_cap |= IXGBE_FEATURE_FDIR;
4675                 if ((sc->hw.device_id == IXGBE_DEV_ID_82599_BYPASS) &&
4676                     (sc->hw.bus.func == 0))
4677                         sc->feat_cap |= IXGBE_FEATURE_BYPASS;
4678                 if (sc->hw.device_id == IXGBE_DEV_ID_82599_QSFP_SF_QP)
4679                         sc->feat_cap &= ~IXGBE_FEATURE_LEGACY_IRQ;
4680                 break;
4681         default:
4682                 break;
4683         }
4684
4685         /* Enabled by default... */
4686         /* Fan failure detection */
4687         if (sc->feat_cap & IXGBE_FEATURE_FAN_FAIL)
4688                 sc->feat_en |= IXGBE_FEATURE_FAN_FAIL;
4689         /* Netmap */
4690         if (sc->feat_cap & IXGBE_FEATURE_NETMAP)
4691                 sc->feat_en |= IXGBE_FEATURE_NETMAP;
4692         /* EEE */
4693         if (sc->feat_cap & IXGBE_FEATURE_EEE)
4694                 sc->feat_en |= IXGBE_FEATURE_EEE;
4695         /* Thermal Sensor */
4696         if (sc->feat_cap & IXGBE_FEATURE_TEMP_SENSOR)
4697                 sc->feat_en |= IXGBE_FEATURE_TEMP_SENSOR;
4698
4699         /* Enabled via global sysctl... */
4700         /* Flow Director */
4701         if (ixgbe_enable_fdir) {
4702                 if (sc->feat_cap & IXGBE_FEATURE_FDIR)
4703                         sc->feat_en |= IXGBE_FEATURE_FDIR;
4704                 else
4705                         device_printf(sc->dev, "Device does not support Flow Director. Leaving disabled.");
4706         }
4707         /*
4708          * Message Signal Interrupts - Extended (MSI-X)
4709          * Normal MSI is only enabled if MSI-X calls fail.
4710          */
4711         if (!ixgbe_enable_msix)
4712                 sc->feat_cap &= ~IXGBE_FEATURE_MSIX;
4713         /* Receive-Side Scaling (RSS) */
4714         if ((sc->feat_cap & IXGBE_FEATURE_RSS) && ixgbe_enable_rss)
4715                 sc->feat_en |= IXGBE_FEATURE_RSS;
4716
4717         /* Disable features with unmet dependencies... */
4718         /* No MSI-X */
4719         if (!(sc->feat_cap & IXGBE_FEATURE_MSIX)) {
4720                 sc->feat_cap &= ~IXGBE_FEATURE_RSS;
4721                 sc->feat_cap &= ~IXGBE_FEATURE_SRIOV;
4722                 sc->feat_en &= ~IXGBE_FEATURE_RSS;
4723                 sc->feat_en &= ~IXGBE_FEATURE_SRIOV;
4724         }
4725 } /* ixgbe_init_device_features */
4726
4727 /************************************************************************
4728  * ixgbe_check_fan_failure
4729  ************************************************************************/
4730 static void
4731 ixgbe_check_fan_failure(struct ixgbe_softc *sc, u32 reg, bool in_interrupt)
4732 {
4733         u32 mask;
4734
4735         mask = (in_interrupt) ? IXGBE_EICR_GPI_SDP1_BY_MAC(&sc->hw) :
4736             IXGBE_ESDP_SDP1;
4737
4738         if (reg & mask)
4739                 device_printf(sc->dev, "\nCRITICAL: FAN FAILURE!! REPLACE IMMEDIATELY!!\n");
4740 } /* ixgbe_check_fan_failure */
4741
4742 /************************************************************************
4743  * ixgbe_sbuf_fw_version
4744  ************************************************************************/
4745 static void
4746 ixgbe_sbuf_fw_version(struct ixgbe_hw *hw, struct sbuf *buf)
4747 {
4748         struct ixgbe_nvm_version nvm_ver = {0};
4749         uint16_t phyfw = 0;
4750         int status;
4751         const char *space = "";
4752
4753         ixgbe_get_oem_prod_version(hw, &nvm_ver); /* OEM's NVM version */
4754         ixgbe_get_orom_version(hw, &nvm_ver); /* Option ROM */
4755         ixgbe_get_etk_id(hw, &nvm_ver); /* eTrack identifies a build in Intel's SCM */
4756         status = ixgbe_get_phy_firmware_version(hw, &phyfw);
4757
4758         if (nvm_ver.oem_valid) {
4759                 sbuf_printf(buf, "NVM OEM V%d.%d R%d", nvm_ver.oem_major,
4760                     nvm_ver.oem_minor, nvm_ver.oem_release);
4761                 space = " ";
4762         }
4763
4764         if (nvm_ver.or_valid) {
4765                 sbuf_printf(buf, "%sOption ROM V%d-b%d-p%d",
4766                     space, nvm_ver.or_major, nvm_ver.or_build, nvm_ver.or_patch);
4767                 space = " ";
4768         }
4769
4770         if (nvm_ver.etk_id != ((NVM_VER_INVALID << NVM_ETK_SHIFT) |
4771             NVM_VER_INVALID)) {
4772                 sbuf_printf(buf, "%seTrack 0x%08x", space, nvm_ver.etk_id);
4773                 space = " ";
4774         }
4775
4776         if (phyfw != 0 && status == IXGBE_SUCCESS)
4777                 sbuf_printf(buf, "%sPHY FW V%d", space, phyfw);
4778 } /* ixgbe_sbuf_fw_version */
4779
4780 /************************************************************************
4781  * ixgbe_print_fw_version
4782  ************************************************************************/
4783 static void
4784 ixgbe_print_fw_version(if_ctx_t ctx)
4785 {
4786         struct ixgbe_softc *sc = iflib_get_softc(ctx);
4787         struct ixgbe_hw *hw = &sc->hw;
4788         device_t dev = sc->dev;
4789         struct sbuf *buf;
4790         int error = 0;
4791
4792         buf = sbuf_new_auto();
4793         if (!buf) {
4794                 device_printf(dev, "Could not allocate sbuf for output.\n");
4795                 return;
4796         }
4797
4798         ixgbe_sbuf_fw_version(hw, buf);
4799
4800         error = sbuf_finish(buf);
4801         if (error)
4802                 device_printf(dev, "Error finishing sbuf: %d\n", error);
4803         else if (sbuf_len(buf))
4804                 device_printf(dev, "%s\n", sbuf_data(buf));
4805
4806         sbuf_delete(buf);
4807 } /* ixgbe_print_fw_version */
4808
4809 /************************************************************************
4810  * ixgbe_sysctl_print_fw_version
4811  ************************************************************************/
4812 static int
4813 ixgbe_sysctl_print_fw_version(SYSCTL_HANDLER_ARGS)
4814 {
4815         struct ixgbe_softc  *sc = (struct ixgbe_softc *)arg1;
4816         struct ixgbe_hw *hw = &sc->hw;
4817         device_t dev = sc->dev;
4818         struct sbuf *buf;
4819         int error = 0;
4820
4821         buf = sbuf_new_for_sysctl(NULL, NULL, 128, req);
4822         if (!buf) {
4823                 device_printf(dev, "Could not allocate sbuf for output.\n");
4824                 return (ENOMEM);
4825         }
4826
4827         ixgbe_sbuf_fw_version(hw, buf);
4828
4829         error = sbuf_finish(buf);
4830         if (error)
4831                 device_printf(dev, "Error finishing sbuf: %d\n", error);
4832
4833         sbuf_delete(buf);
4834
4835         return (0);
4836 } /* ixgbe_sysctl_print_fw_version */