]> CyberLeo.Net >> Repos - FreeBSD/releng/10.0.git/blob - sys/dev/qlxgb/qla_isr.c
- Copy stable/10 (r259064) to releng/10.0 as part of the
[FreeBSD/releng/10.0.git] / sys / dev / qlxgb / qla_isr.c
1 /*
2  * Copyright (c) 2011-2013 Qlogic Corporation
3  * All rights reserved.
4  *
5  *  Redistribution and use in source and binary forms, with or without
6  *  modification, are permitted provided that the following conditions
7  *  are met:
8  *
9  *  1. Redistributions of source code must retain the above copyright
10  *     notice, this list of conditions and the following disclaimer.
11  *  2. Redistributions in binary form must reproduce the above copyright
12  *     notice, this list of conditions and the following disclaimer in the
13  *     documentation and/or other materials provided with the distribution.
14  *
15  *  THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
16  *  AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
17  *  IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
18  *  ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE
19  *  LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
20  *  CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
21  *  SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS
22  *  INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN
23  *  CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE)
24  *  ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
25  *  POSSIBILITY OF SUCH DAMAGE.
26  */
27
28 /*
29  * File: qla_isr.c
30  * Author : David C Somayajulu, Qlogic Corporation, Aliso Viejo, CA 92656.
31  */
32
33 #include <sys/cdefs.h>
34 __FBSDID("$FreeBSD$");
35
36 #include "qla_os.h"
37 #include "qla_reg.h"
38 #include "qla_hw.h"
39 #include "qla_def.h"
40 #include "qla_inline.h"
41 #include "qla_ver.h"
42 #include "qla_glbl.h"
43 #include "qla_dbg.h"
44
45 static void qla_replenish_normal_rx(qla_host_t *ha, qla_sds_t *sdsp);
46 static void qla_replenish_jumbo_rx(qla_host_t *ha, qla_sds_t *sdsp);
47
48 /*
49  * Name: qla_rx_intr
50  * Function: Handles normal ethernet frames received
51  */
52 static void
53 qla_rx_intr(qla_host_t *ha, uint64_t data, uint32_t sds_idx,
54         struct lro_ctrl *lro)
55 {
56         uint32_t idx, length, status, ring;
57         qla_rx_buf_t *rxb;
58         struct mbuf *mp;
59         struct ifnet *ifp = ha->ifp;
60         qla_sds_t *sdsp;
61         struct ether_vlan_header *eh;
62         
63         sdsp = &ha->hw.sds[sds_idx];
64         
65         ring = (uint32_t)Q8_STAT_DESC_TYPE(data);
66         idx = (uint32_t)Q8_STAT_DESC_HANDLE(data);
67         length = (uint32_t)Q8_STAT_DESC_TOTAL_LENGTH(data);
68         status = (uint32_t)Q8_STAT_DESC_STATUS(data);
69
70         if (ring == 0) {
71                 if ((idx >= NUM_RX_DESCRIPTORS) || (length > MCLBYTES)) {
72                         device_printf(ha->pci_dev, "%s: ring[%d] index[0x%08x]"
73                                 " len[0x%08x] invalid\n",
74                                 __func__, ring, idx, length);
75                         return;
76                 }
77         } else {
78                 if ((idx >= NUM_RX_JUMBO_DESCRIPTORS)||(length > MJUM9BYTES)) {
79                         device_printf(ha->pci_dev, "%s: ring[%d] index[0x%08x]"
80                                 " len[0x%08x] invalid\n",
81                                 __func__, ring, idx, length);
82                         return;
83                 }
84         }
85
86         if (ring == 0)
87                 rxb = &ha->rx_buf[idx];
88         else 
89                 rxb = &ha->rx_jbuf[idx];
90
91         QL_ASSERT((rxb != NULL),\
92                 ("%s: [r, i, sds_idx]=[%d, 0x%x, %d] rxb != NULL\n",\
93                  __func__, ring, idx, sds_idx));
94
95         mp = rxb->m_head;
96
97         QL_ASSERT((mp != NULL),\
98                 ("%s: [r,i,rxb, sds_idx]=[%d, 0x%x, %p, %d] mp != NULL\n",\
99                  __func__, ring, idx, rxb, sds_idx));
100
101         bus_dmamap_sync(ha->rx_tag, rxb->map, BUS_DMASYNC_POSTREAD);
102
103         if (ring == 0) {
104                 rxb->m_head = NULL;
105                 rxb->next = sdsp->rxb_free;
106                 sdsp->rxb_free = rxb;
107                 sdsp->rx_free++;
108         } else {
109                 rxb->m_head = NULL;
110                 rxb->next = sdsp->rxjb_free;
111                 sdsp->rxjb_free = rxb;
112                 sdsp->rxj_free++;
113         }
114         
115         mp->m_len = length;
116         mp->m_pkthdr.len = length;
117         mp->m_pkthdr.rcvif = ifp;
118         
119         eh = mtod(mp, struct ether_vlan_header *);
120
121         if (eh->evl_encap_proto == htons(ETHERTYPE_VLAN)) {
122                 uint32_t *data = (uint32_t *)eh;
123
124                 mp->m_pkthdr.ether_vtag = ntohs(eh->evl_tag);
125                 mp->m_flags |= M_VLANTAG;
126
127                 *(data + 3) = *(data + 2);
128                 *(data + 2) = *(data + 1);
129                 *(data + 1) = *data;
130
131                 m_adj(mp, ETHER_VLAN_ENCAP_LEN);
132         }
133
134         if (status == Q8_STAT_DESC_STATUS_CHKSUM_OK) {
135                 mp->m_pkthdr.csum_flags = (CSUM_IP_CHECKED | CSUM_IP_VALID);
136         } else {
137                 mp->m_pkthdr.csum_flags = 0;
138         }
139
140         if (lro->lro_cnt && (tcp_lro_rx(lro, mp, 0) == 0)) {
141                 /* LRO packet has been successfuly queued */
142         } else {
143                 (*ifp->if_input)(ifp, mp);
144         }
145
146         if (sdsp->rx_free > std_replenish)
147                 qla_replenish_normal_rx(ha, sdsp);
148
149         if (sdsp->rxj_free > jumbo_replenish)
150                 qla_replenish_jumbo_rx(ha, sdsp);
151
152         return;
153 }
154
155 static void
156 qla_replenish_jumbo_rx(qla_host_t *ha, qla_sds_t *sdsp)
157 {
158         qla_rx_buf_t *rxb;
159         int count = jumbo_replenish;
160         uint32_t rxj_next;
161
162         if (!mtx_trylock(&ha->rxj_lock))
163                 return;
164
165         rxj_next = ha->hw.rxj_next;
166
167         while (count--) {
168                 rxb = sdsp->rxjb_free;
169
170                 if (rxb == NULL)
171                         break;
172
173                 sdsp->rxjb_free = rxb->next;
174                 sdsp->rxj_free--;
175
176
177                 if (qla_get_mbuf(ha, rxb, NULL, RDS_RING_INDEX_JUMBO) == 0) {
178                         qla_set_hw_rcv_desc(ha, RDS_RING_INDEX_JUMBO,
179                                 ha->hw.rxj_in, rxb->handle, rxb->paddr,
180                                 (rxb->m_head)->m_pkthdr.len);
181                         ha->hw.rxj_in++;
182                         if (ha->hw.rxj_in == NUM_RX_JUMBO_DESCRIPTORS)
183                                 ha->hw.rxj_in = 0;
184                         ha->hw.rxj_next++;
185                         if (ha->hw.rxj_next == NUM_RX_JUMBO_DESCRIPTORS)
186                                 ha->hw.rxj_next = 0;
187                 } else {
188                         device_printf(ha->pci_dev,
189                                 "%s: qla_get_mbuf [1,(%d),(%d)] failed\n",
190                                 __func__, ha->hw.rxj_in, rxb->handle);
191
192                         rxb->m_head = NULL;
193                         rxb->next = sdsp->rxjb_free;
194                         sdsp->rxjb_free = rxb;
195                         sdsp->rxj_free++;
196
197                         break;
198                 }
199         }
200
201         if (rxj_next != ha->hw.rxj_next) {
202                 QL_UPDATE_RDS_PRODUCER_INDEX(ha, 1, ha->hw.rxj_next);
203         }
204         mtx_unlock(&ha->rxj_lock);
205 }
206
207 static void
208 qla_replenish_normal_rx(qla_host_t *ha, qla_sds_t *sdsp)
209 {
210         qla_rx_buf_t *rxb;
211         int count = std_replenish;
212         uint32_t rx_next;
213
214         if (!mtx_trylock(&ha->rx_lock))
215                 return;
216
217         rx_next = ha->hw.rx_next;
218
219         while (count--) {
220                 rxb = sdsp->rxb_free;
221
222                 if (rxb == NULL)
223                         break;
224
225                 sdsp->rxb_free = rxb->next;
226                 sdsp->rx_free--;
227
228                 if (qla_get_mbuf(ha, rxb, NULL, RDS_RING_INDEX_NORMAL) == 0) {
229                         qla_set_hw_rcv_desc(ha, RDS_RING_INDEX_NORMAL,
230                                 ha->hw.rx_in, rxb->handle, rxb->paddr,
231                                 (rxb->m_head)->m_pkthdr.len);
232                         ha->hw.rx_in++;
233                         if (ha->hw.rx_in == NUM_RX_DESCRIPTORS)
234                                 ha->hw.rx_in = 0;
235                         ha->hw.rx_next++;
236                         if (ha->hw.rx_next == NUM_RX_DESCRIPTORS)
237                                 ha->hw.rx_next = 0;
238                 } else {
239                         device_printf(ha->pci_dev,
240                                 "%s: qla_get_mbuf [0,(%d),(%d)] failed\n",
241                                 __func__, ha->hw.rx_in, rxb->handle);
242
243                         rxb->m_head = NULL;
244                         rxb->next = sdsp->rxb_free;
245                         sdsp->rxb_free = rxb;
246                         sdsp->rx_free++;
247
248                         break;
249                 }
250         }
251
252         if (rx_next != ha->hw.rx_next) {
253                 QL_UPDATE_RDS_PRODUCER_INDEX(ha, 0, ha->hw.rx_next);
254         }
255         mtx_unlock(&ha->rx_lock);
256 }
257
258 /*
259  * Name: qla_isr
260  * Function: Main Interrupt Service Routine
261  */
262 static uint32_t
263 qla_rcv_isr(qla_host_t *ha, uint32_t sds_idx, uint32_t count)
264 {
265         device_t dev;
266         qla_hw_t *hw;
267         uint32_t comp_idx, desc_count;
268         q80_stat_desc_t *sdesc;
269         struct lro_ctrl *lro;
270         struct lro_entry *queued;
271         uint32_t ret = 0;
272
273         dev = ha->pci_dev;
274         hw = &ha->hw;
275
276         hw->sds[sds_idx].rcv_active = 1;
277         if (ha->flags.stop_rcv) {
278                 hw->sds[sds_idx].rcv_active = 0;
279                 return 0;
280         }
281
282         QL_DPRINT2((dev, "%s: [%d]enter\n", __func__, sds_idx));
283
284         /*
285          * receive interrupts
286          */
287         comp_idx = hw->sds[sds_idx].sdsr_next;
288         lro = &hw->sds[sds_idx].lro;
289
290         while (count--) {
291
292                 sdesc = (q80_stat_desc_t *)
293                                 &hw->sds[sds_idx].sds_ring_base[comp_idx];
294
295                 if (Q8_STAT_DESC_OWNER((sdesc->data[0])) !=
296                         Q8_STAT_DESC_OWNER_HOST) {
297                         QL_DPRINT2((dev, "%s:  data %p sdsr_next 0x%08x\n",
298                                 __func__, (void *)sdesc->data[0], comp_idx));
299                         break;
300                 }
301
302                 desc_count = Q8_STAT_DESC_COUNT((sdesc->data[0]));
303
304                 switch (Q8_STAT_DESC_OPCODE((sdesc->data[0]))) {
305
306                 case Q8_STAT_DESC_OPCODE_RCV_PKT:
307                 case Q8_STAT_DESC_OPCODE_SYN_OFFLOAD:
308                         qla_rx_intr(ha, (sdesc->data[0]), sds_idx, lro);
309                         
310                         break;
311
312                 default:
313                         device_printf(dev, "%s: default 0x%llx!\n", __func__,
314                                         (long long unsigned int)sdesc->data[0]);
315                         break;
316                 }
317
318                 while (desc_count--) {
319                         sdesc->data[0] =
320                                 Q8_STAT_DESC_SET_OWNER(Q8_STAT_DESC_OWNER_FW);
321                         comp_idx = (comp_idx + 1) & (NUM_STATUS_DESCRIPTORS-1);
322                         sdesc = (q80_stat_desc_t *)
323                                 &hw->sds[sds_idx].sds_ring_base[comp_idx];
324                 }
325         }
326
327         while((!SLIST_EMPTY(&lro->lro_active))) {
328                 queued = SLIST_FIRST(&lro->lro_active);
329                 SLIST_REMOVE_HEAD(&lro->lro_active, next);
330                 tcp_lro_flush(lro, queued);
331         }
332
333         if (hw->sds[sds_idx].sdsr_next != comp_idx) {
334                 QL_UPDATE_SDS_CONSUMER_INDEX(ha, sds_idx, comp_idx);
335         }
336         hw->sds[sds_idx].sdsr_next = comp_idx;
337
338         sdesc = (q80_stat_desc_t *)&hw->sds[sds_idx].sds_ring_base[comp_idx];
339         if ((sds_idx == 0) && (Q8_STAT_DESC_OWNER((sdesc->data[0])) ==
340                                         Q8_STAT_DESC_OWNER_HOST)) {
341                 ret = -1;
342         }
343
344         hw->sds[sds_idx].rcv_active = 0;
345         return (ret);
346 }
347
348 void
349 qla_isr(void *arg)
350 {
351         qla_ivec_t *ivec = arg;
352         qla_host_t *ha;
353         uint32_t sds_idx;
354         uint32_t ret;
355
356         ha = ivec->ha;
357         sds_idx = ivec->irq_rid - 1;
358
359         if (sds_idx >= ha->hw.num_sds_rings) {
360                 device_printf(ha->pci_dev, "%s: bogus sds_idx 0x%x\n", __func__,
361                         sds_idx);
362                 
363                 return;
364         }
365
366         if (sds_idx == 0)
367                 taskqueue_enqueue(ha->tx_tq, &ha->tx_task);
368
369         ret = qla_rcv_isr(ha, sds_idx, rcv_pkt_thres);
370
371         if (sds_idx == 0)
372                 taskqueue_enqueue(ha->tx_tq, &ha->tx_task);
373
374         if (ret) {
375                 taskqueue_enqueue(ha->irq_vec[sds_idx].rcv_tq,
376                         &ha->irq_vec[sds_idx].rcv_task);
377         } else {
378                 QL_ENABLE_INTERRUPTS(ha, sds_idx);
379         }
380 }
381
382 void
383 qla_rcv(void *context, int pending)
384 {
385         qla_ivec_t *ivec = context;
386         qla_host_t *ha;
387         device_t dev;
388         qla_hw_t *hw;
389         uint32_t sds_idx;
390         uint32_t ret;
391         struct ifnet *ifp;
392
393         ha = ivec->ha;
394         dev = ha->pci_dev;
395         hw = &ha->hw;
396         sds_idx = ivec->irq_rid - 1;
397         ifp = ha->ifp;
398
399         do {
400                 if (sds_idx == 0) {
401                         if (qla_le32_to_host(*(hw->tx_cons)) != hw->txr_comp) {
402                                 taskqueue_enqueue(ha->tx_tq, &ha->tx_task);
403                         } else if ((ifp->if_snd.ifq_head != NULL) &&
404                                         QL_RUNNING(ifp)) {
405                                 taskqueue_enqueue(ha->tx_tq, &ha->tx_task);
406                         }
407                 }
408                 ret = qla_rcv_isr(ha, sds_idx, rcv_pkt_thres_d);
409         } while (ret);
410
411         if (sds_idx == 0)
412                 taskqueue_enqueue(ha->tx_tq, &ha->tx_task);
413
414         QL_ENABLE_INTERRUPTS(ha, sds_idx);
415 }
416