]> CyberLeo.Net >> Repos - FreeBSD/releng/10.3.git/blob - sys/dev/ixl/i40e_adminq.c
- Copy stable/10@296371 to releng/10.3 in preparation for 10.3-RC1
[FreeBSD/releng/10.3.git] / sys / dev / ixl / i40e_adminq.c
1 /******************************************************************************
2
3   Copyright (c) 2013-2015, Intel Corporation 
4   All rights reserved.
5   
6   Redistribution and use in source and binary forms, with or without 
7   modification, are permitted provided that the following conditions are met:
8   
9    1. Redistributions of source code must retain the above copyright notice, 
10       this list of conditions and the following disclaimer.
11   
12    2. Redistributions in binary form must reproduce the above copyright 
13       notice, this list of conditions and the following disclaimer in the 
14       documentation and/or other materials provided with the distribution.
15   
16    3. Neither the name of the Intel Corporation nor the names of its 
17       contributors may be used to endorse or promote products derived from 
18       this software without specific prior written permission.
19   
20   THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS"
21   AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE 
22   IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE 
23   ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE 
24   LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR 
25   CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF 
26   SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS 
27   INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN 
28   CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) 
29   ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
30   POSSIBILITY OF SUCH DAMAGE.
31
32 ******************************************************************************/
33 /*$FreeBSD$*/
34
35 #include "i40e_status.h"
36 #include "i40e_type.h"
37 #include "i40e_register.h"
38 #include "i40e_adminq.h"
39 #include "i40e_prototype.h"
40
41 /**
42  * i40e_is_nvm_update_op - return TRUE if this is an NVM update operation
43  * @desc: API request descriptor
44  **/
45 static INLINE bool i40e_is_nvm_update_op(struct i40e_aq_desc *desc)
46 {
47         return (desc->opcode == CPU_TO_LE16(i40e_aqc_opc_nvm_erase) ||
48                 desc->opcode == CPU_TO_LE16(i40e_aqc_opc_nvm_update));
49 }
50
51 /**
52  *  i40e_adminq_init_regs - Initialize AdminQ registers
53  *  @hw: pointer to the hardware structure
54  *
55  *  This assumes the alloc_asq and alloc_arq functions have already been called
56  **/
57 static void i40e_adminq_init_regs(struct i40e_hw *hw)
58 {
59         /* set head and tail registers in our local struct */
60         if (i40e_is_vf(hw)) {
61                 hw->aq.asq.tail = I40E_VF_ATQT1;
62                 hw->aq.asq.head = I40E_VF_ATQH1;
63                 hw->aq.asq.len  = I40E_VF_ATQLEN1;
64                 hw->aq.asq.bal  = I40E_VF_ATQBAL1;
65                 hw->aq.asq.bah  = I40E_VF_ATQBAH1;
66                 hw->aq.arq.tail = I40E_VF_ARQT1;
67                 hw->aq.arq.head = I40E_VF_ARQH1;
68                 hw->aq.arq.len  = I40E_VF_ARQLEN1;
69                 hw->aq.arq.bal  = I40E_VF_ARQBAL1;
70                 hw->aq.arq.bah  = I40E_VF_ARQBAH1;
71         } else {
72                 hw->aq.asq.tail = I40E_PF_ATQT;
73                 hw->aq.asq.head = I40E_PF_ATQH;
74                 hw->aq.asq.len  = I40E_PF_ATQLEN;
75                 hw->aq.asq.bal  = I40E_PF_ATQBAL;
76                 hw->aq.asq.bah  = I40E_PF_ATQBAH;
77                 hw->aq.arq.tail = I40E_PF_ARQT;
78                 hw->aq.arq.head = I40E_PF_ARQH;
79                 hw->aq.arq.len  = I40E_PF_ARQLEN;
80                 hw->aq.arq.bal  = I40E_PF_ARQBAL;
81                 hw->aq.arq.bah  = I40E_PF_ARQBAH;
82         }
83 }
84
85 /**
86  *  i40e_alloc_adminq_asq_ring - Allocate Admin Queue send rings
87  *  @hw: pointer to the hardware structure
88  **/
89 enum i40e_status_code i40e_alloc_adminq_asq_ring(struct i40e_hw *hw)
90 {
91         enum i40e_status_code ret_code;
92
93         ret_code = i40e_allocate_dma_mem(hw, &hw->aq.asq.desc_buf,
94                                          i40e_mem_atq_ring,
95                                          (hw->aq.num_asq_entries *
96                                          sizeof(struct i40e_aq_desc)),
97                                          I40E_ADMINQ_DESC_ALIGNMENT);
98         if (ret_code)
99                 return ret_code;
100
101         ret_code = i40e_allocate_virt_mem(hw, &hw->aq.asq.cmd_buf,
102                                           (hw->aq.num_asq_entries *
103                                           sizeof(struct i40e_asq_cmd_details)));
104         if (ret_code) {
105                 i40e_free_dma_mem(hw, &hw->aq.asq.desc_buf);
106                 return ret_code;
107         }
108
109         return ret_code;
110 }
111
112 /**
113  *  i40e_alloc_adminq_arq_ring - Allocate Admin Queue receive rings
114  *  @hw: pointer to the hardware structure
115  **/
116 enum i40e_status_code i40e_alloc_adminq_arq_ring(struct i40e_hw *hw)
117 {
118         enum i40e_status_code ret_code;
119
120         ret_code = i40e_allocate_dma_mem(hw, &hw->aq.arq.desc_buf,
121                                          i40e_mem_arq_ring,
122                                          (hw->aq.num_arq_entries *
123                                          sizeof(struct i40e_aq_desc)),
124                                          I40E_ADMINQ_DESC_ALIGNMENT);
125
126         return ret_code;
127 }
128
129 /**
130  *  i40e_free_adminq_asq - Free Admin Queue send rings
131  *  @hw: pointer to the hardware structure
132  *
133  *  This assumes the posted send buffers have already been cleaned
134  *  and de-allocated
135  **/
136 void i40e_free_adminq_asq(struct i40e_hw *hw)
137 {
138         i40e_free_dma_mem(hw, &hw->aq.asq.desc_buf);
139 }
140
141 /**
142  *  i40e_free_adminq_arq - Free Admin Queue receive rings
143  *  @hw: pointer to the hardware structure
144  *
145  *  This assumes the posted receive buffers have already been cleaned
146  *  and de-allocated
147  **/
148 void i40e_free_adminq_arq(struct i40e_hw *hw)
149 {
150         i40e_free_dma_mem(hw, &hw->aq.arq.desc_buf);
151 }
152
153 /**
154  *  i40e_alloc_arq_bufs - Allocate pre-posted buffers for the receive queue
155  *  @hw: pointer to the hardware structure
156  **/
157 static enum i40e_status_code i40e_alloc_arq_bufs(struct i40e_hw *hw)
158 {
159         enum i40e_status_code ret_code;
160         struct i40e_aq_desc *desc;
161         struct i40e_dma_mem *bi;
162         int i;
163
164         /* We'll be allocating the buffer info memory first, then we can
165          * allocate the mapped buffers for the event processing
166          */
167
168         /* buffer_info structures do not need alignment */
169         ret_code = i40e_allocate_virt_mem(hw, &hw->aq.arq.dma_head,
170                 (hw->aq.num_arq_entries * sizeof(struct i40e_dma_mem)));
171         if (ret_code)
172                 goto alloc_arq_bufs;
173         hw->aq.arq.r.arq_bi = (struct i40e_dma_mem *)hw->aq.arq.dma_head.va;
174
175         /* allocate the mapped buffers */
176         for (i = 0; i < hw->aq.num_arq_entries; i++) {
177                 bi = &hw->aq.arq.r.arq_bi[i];
178                 ret_code = i40e_allocate_dma_mem(hw, bi,
179                                                  i40e_mem_arq_buf,
180                                                  hw->aq.arq_buf_size,
181                                                  I40E_ADMINQ_DESC_ALIGNMENT);
182                 if (ret_code)
183                         goto unwind_alloc_arq_bufs;
184
185                 /* now configure the descriptors for use */
186                 desc = I40E_ADMINQ_DESC(hw->aq.arq, i);
187
188                 desc->flags = CPU_TO_LE16(I40E_AQ_FLAG_BUF);
189                 if (hw->aq.arq_buf_size > I40E_AQ_LARGE_BUF)
190                         desc->flags |= CPU_TO_LE16(I40E_AQ_FLAG_LB);
191                 desc->opcode = 0;
192                 /* This is in accordance with Admin queue design, there is no
193                  * register for buffer size configuration
194                  */
195                 desc->datalen = CPU_TO_LE16((u16)bi->size);
196                 desc->retval = 0;
197                 desc->cookie_high = 0;
198                 desc->cookie_low = 0;
199                 desc->params.external.addr_high =
200                         CPU_TO_LE32(I40E_HI_DWORD(bi->pa));
201                 desc->params.external.addr_low =
202                         CPU_TO_LE32(I40E_LO_DWORD(bi->pa));
203                 desc->params.external.param0 = 0;
204                 desc->params.external.param1 = 0;
205         }
206
207 alloc_arq_bufs:
208         return ret_code;
209
210 unwind_alloc_arq_bufs:
211         /* don't try to free the one that failed... */
212         i--;
213         for (; i >= 0; i--)
214                 i40e_free_dma_mem(hw, &hw->aq.arq.r.arq_bi[i]);
215         i40e_free_virt_mem(hw, &hw->aq.arq.dma_head);
216
217         return ret_code;
218 }
219
220 /**
221  *  i40e_alloc_asq_bufs - Allocate empty buffer structs for the send queue
222  *  @hw: pointer to the hardware structure
223  **/
224 static enum i40e_status_code i40e_alloc_asq_bufs(struct i40e_hw *hw)
225 {
226         enum i40e_status_code ret_code;
227         struct i40e_dma_mem *bi;
228         int i;
229
230         /* No mapped memory needed yet, just the buffer info structures */
231         ret_code = i40e_allocate_virt_mem(hw, &hw->aq.asq.dma_head,
232                 (hw->aq.num_asq_entries * sizeof(struct i40e_dma_mem)));
233         if (ret_code)
234                 goto alloc_asq_bufs;
235         hw->aq.asq.r.asq_bi = (struct i40e_dma_mem *)hw->aq.asq.dma_head.va;
236
237         /* allocate the mapped buffers */
238         for (i = 0; i < hw->aq.num_asq_entries; i++) {
239                 bi = &hw->aq.asq.r.asq_bi[i];
240                 ret_code = i40e_allocate_dma_mem(hw, bi,
241                                                  i40e_mem_asq_buf,
242                                                  hw->aq.asq_buf_size,
243                                                  I40E_ADMINQ_DESC_ALIGNMENT);
244                 if (ret_code)
245                         goto unwind_alloc_asq_bufs;
246         }
247 alloc_asq_bufs:
248         return ret_code;
249
250 unwind_alloc_asq_bufs:
251         /* don't try to free the one that failed... */
252         i--;
253         for (; i >= 0; i--)
254                 i40e_free_dma_mem(hw, &hw->aq.asq.r.asq_bi[i]);
255         i40e_free_virt_mem(hw, &hw->aq.asq.dma_head);
256
257         return ret_code;
258 }
259
260 /**
261  *  i40e_free_arq_bufs - Free receive queue buffer info elements
262  *  @hw: pointer to the hardware structure
263  **/
264 static void i40e_free_arq_bufs(struct i40e_hw *hw)
265 {
266         int i;
267
268         /* free descriptors */
269         for (i = 0; i < hw->aq.num_arq_entries; i++)
270                 i40e_free_dma_mem(hw, &hw->aq.arq.r.arq_bi[i]);
271
272         /* free the descriptor memory */
273         i40e_free_dma_mem(hw, &hw->aq.arq.desc_buf);
274
275         /* free the dma header */
276         i40e_free_virt_mem(hw, &hw->aq.arq.dma_head);
277 }
278
279 /**
280  *  i40e_free_asq_bufs - Free send queue buffer info elements
281  *  @hw: pointer to the hardware structure
282  **/
283 static void i40e_free_asq_bufs(struct i40e_hw *hw)
284 {
285         int i;
286
287         /* only unmap if the address is non-NULL */
288         for (i = 0; i < hw->aq.num_asq_entries; i++)
289                 if (hw->aq.asq.r.asq_bi[i].pa)
290                         i40e_free_dma_mem(hw, &hw->aq.asq.r.asq_bi[i]);
291
292         /* free the buffer info list */
293         i40e_free_virt_mem(hw, &hw->aq.asq.cmd_buf);
294
295         /* free the descriptor memory */
296         i40e_free_dma_mem(hw, &hw->aq.asq.desc_buf);
297
298         /* free the dma header */
299         i40e_free_virt_mem(hw, &hw->aq.asq.dma_head);
300 }
301
302 /**
303  *  i40e_config_asq_regs - configure ASQ registers
304  *  @hw: pointer to the hardware structure
305  *
306  *  Configure base address and length registers for the transmit queue
307  **/
308 static enum i40e_status_code i40e_config_asq_regs(struct i40e_hw *hw)
309 {
310         enum i40e_status_code ret_code = I40E_SUCCESS;
311         u32 reg = 0;
312
313         /* Clear Head and Tail */
314         wr32(hw, hw->aq.asq.head, 0);
315         wr32(hw, hw->aq.asq.tail, 0);
316
317         /* set starting point */
318         if (!i40e_is_vf(hw))
319                 wr32(hw, hw->aq.asq.len, (hw->aq.num_asq_entries |
320                                           I40E_PF_ATQLEN_ATQENABLE_MASK));
321         if (i40e_is_vf(hw))
322                 wr32(hw, hw->aq.asq.len, (hw->aq.num_asq_entries |
323                                           I40E_VF_ATQLEN1_ATQENABLE_MASK));
324         wr32(hw, hw->aq.asq.bal, I40E_LO_DWORD(hw->aq.asq.desc_buf.pa));
325         wr32(hw, hw->aq.asq.bah, I40E_HI_DWORD(hw->aq.asq.desc_buf.pa));
326
327         /* Check one register to verify that config was applied */
328         reg = rd32(hw, hw->aq.asq.bal);
329         if (reg != I40E_LO_DWORD(hw->aq.asq.desc_buf.pa))
330                 ret_code = I40E_ERR_ADMIN_QUEUE_ERROR;
331
332         return ret_code;
333 }
334
335 /**
336  *  i40e_config_arq_regs - ARQ register configuration
337  *  @hw: pointer to the hardware structure
338  *
339  * Configure base address and length registers for the receive (event queue)
340  **/
341 static enum i40e_status_code i40e_config_arq_regs(struct i40e_hw *hw)
342 {
343         enum i40e_status_code ret_code = I40E_SUCCESS;
344         u32 reg = 0;
345
346         /* Clear Head and Tail */
347         wr32(hw, hw->aq.arq.head, 0);
348         wr32(hw, hw->aq.arq.tail, 0);
349
350         /* set starting point */
351         if (!i40e_is_vf(hw))
352                 wr32(hw, hw->aq.arq.len, (hw->aq.num_arq_entries |
353                                           I40E_PF_ARQLEN_ARQENABLE_MASK));
354         if (i40e_is_vf(hw))
355                 wr32(hw, hw->aq.arq.len, (hw->aq.num_arq_entries |
356                                           I40E_VF_ARQLEN1_ARQENABLE_MASK));
357         wr32(hw, hw->aq.arq.bal, I40E_LO_DWORD(hw->aq.arq.desc_buf.pa));
358         wr32(hw, hw->aq.arq.bah, I40E_HI_DWORD(hw->aq.arq.desc_buf.pa));
359
360         /* Update tail in the HW to post pre-allocated buffers */
361         wr32(hw, hw->aq.arq.tail, hw->aq.num_arq_entries - 1);
362
363         /* Check one register to verify that config was applied */
364         reg = rd32(hw, hw->aq.arq.bal);
365         if (reg != I40E_LO_DWORD(hw->aq.arq.desc_buf.pa))
366                 ret_code = I40E_ERR_ADMIN_QUEUE_ERROR;
367
368         return ret_code;
369 }
370
371 /**
372  *  i40e_init_asq - main initialization routine for ASQ
373  *  @hw: pointer to the hardware structure
374  *
375  *  This is the main initialization routine for the Admin Send Queue
376  *  Prior to calling this function, drivers *MUST* set the following fields
377  *  in the hw->aq structure:
378  *     - hw->aq.num_asq_entries
379  *     - hw->aq.arq_buf_size
380  *
381  *  Do *NOT* hold the lock when calling this as the memory allocation routines
382  *  called are not going to be atomic context safe
383  **/
384 enum i40e_status_code i40e_init_asq(struct i40e_hw *hw)
385 {
386         enum i40e_status_code ret_code = I40E_SUCCESS;
387
388         if (hw->aq.asq.count > 0) {
389                 /* queue already initialized */
390                 ret_code = I40E_ERR_NOT_READY;
391                 goto init_adminq_exit;
392         }
393
394         /* verify input for valid configuration */
395         if ((hw->aq.num_asq_entries == 0) ||
396             (hw->aq.asq_buf_size == 0)) {
397                 ret_code = I40E_ERR_CONFIG;
398                 goto init_adminq_exit;
399         }
400
401         hw->aq.asq.next_to_use = 0;
402         hw->aq.asq.next_to_clean = 0;
403         hw->aq.asq.count = hw->aq.num_asq_entries;
404
405         /* allocate the ring memory */
406         ret_code = i40e_alloc_adminq_asq_ring(hw);
407         if (ret_code != I40E_SUCCESS)
408                 goto init_adminq_exit;
409
410         /* allocate buffers in the rings */
411         ret_code = i40e_alloc_asq_bufs(hw);
412         if (ret_code != I40E_SUCCESS)
413                 goto init_adminq_free_rings;
414
415         /* initialize base registers */
416         ret_code = i40e_config_asq_regs(hw);
417         if (ret_code != I40E_SUCCESS)
418                 goto init_adminq_free_rings;
419
420         /* success! */
421         goto init_adminq_exit;
422
423 init_adminq_free_rings:
424         i40e_free_adminq_asq(hw);
425
426 init_adminq_exit:
427         return ret_code;
428 }
429
430 /**
431  *  i40e_init_arq - initialize ARQ
432  *  @hw: pointer to the hardware structure
433  *
434  *  The main initialization routine for the Admin Receive (Event) Queue.
435  *  Prior to calling this function, drivers *MUST* set the following fields
436  *  in the hw->aq structure:
437  *     - hw->aq.num_asq_entries
438  *     - hw->aq.arq_buf_size
439  *
440  *  Do *NOT* hold the lock when calling this as the memory allocation routines
441  *  called are not going to be atomic context safe
442  **/
443 enum i40e_status_code i40e_init_arq(struct i40e_hw *hw)
444 {
445         enum i40e_status_code ret_code = I40E_SUCCESS;
446
447         if (hw->aq.arq.count > 0) {
448                 /* queue already initialized */
449                 ret_code = I40E_ERR_NOT_READY;
450                 goto init_adminq_exit;
451         }
452
453         /* verify input for valid configuration */
454         if ((hw->aq.num_arq_entries == 0) ||
455             (hw->aq.arq_buf_size == 0)) {
456                 ret_code = I40E_ERR_CONFIG;
457                 goto init_adminq_exit;
458         }
459
460         hw->aq.arq.next_to_use = 0;
461         hw->aq.arq.next_to_clean = 0;
462         hw->aq.arq.count = hw->aq.num_arq_entries;
463
464         /* allocate the ring memory */
465         ret_code = i40e_alloc_adminq_arq_ring(hw);
466         if (ret_code != I40E_SUCCESS)
467                 goto init_adminq_exit;
468
469         /* allocate buffers in the rings */
470         ret_code = i40e_alloc_arq_bufs(hw);
471         if (ret_code != I40E_SUCCESS)
472                 goto init_adminq_free_rings;
473
474         /* initialize base registers */
475         ret_code = i40e_config_arq_regs(hw);
476         if (ret_code != I40E_SUCCESS)
477                 goto init_adminq_free_rings;
478
479         /* success! */
480         goto init_adminq_exit;
481
482 init_adminq_free_rings:
483         i40e_free_adminq_arq(hw);
484
485 init_adminq_exit:
486         return ret_code;
487 }
488
489 /**
490  *  i40e_shutdown_asq - shutdown the ASQ
491  *  @hw: pointer to the hardware structure
492  *
493  *  The main shutdown routine for the Admin Send Queue
494  **/
495 enum i40e_status_code i40e_shutdown_asq(struct i40e_hw *hw)
496 {
497         enum i40e_status_code ret_code = I40E_SUCCESS;
498
499         if (hw->aq.asq.count == 0)
500                 return I40E_ERR_NOT_READY;
501
502         /* Stop firmware AdminQ processing */
503         wr32(hw, hw->aq.asq.head, 0);
504         wr32(hw, hw->aq.asq.tail, 0);
505         wr32(hw, hw->aq.asq.len, 0);
506         wr32(hw, hw->aq.asq.bal, 0);
507         wr32(hw, hw->aq.asq.bah, 0);
508
509         /* make sure spinlock is available */
510         i40e_acquire_spinlock(&hw->aq.asq_spinlock);
511
512         hw->aq.asq.count = 0; /* to indicate uninitialized queue */
513
514         /* free ring buffers */
515         i40e_free_asq_bufs(hw);
516
517         i40e_release_spinlock(&hw->aq.asq_spinlock);
518
519         return ret_code;
520 }
521
522 /**
523  *  i40e_shutdown_arq - shutdown ARQ
524  *  @hw: pointer to the hardware structure
525  *
526  *  The main shutdown routine for the Admin Receive Queue
527  **/
528 enum i40e_status_code i40e_shutdown_arq(struct i40e_hw *hw)
529 {
530         enum i40e_status_code ret_code = I40E_SUCCESS;
531
532         if (hw->aq.arq.count == 0)
533                 return I40E_ERR_NOT_READY;
534
535         /* Stop firmware AdminQ processing */
536         wr32(hw, hw->aq.arq.head, 0);
537         wr32(hw, hw->aq.arq.tail, 0);
538         wr32(hw, hw->aq.arq.len, 0);
539         wr32(hw, hw->aq.arq.bal, 0);
540         wr32(hw, hw->aq.arq.bah, 0);
541
542         /* make sure spinlock is available */
543         i40e_acquire_spinlock(&hw->aq.arq_spinlock);
544
545         hw->aq.arq.count = 0; /* to indicate uninitialized queue */
546
547         /* free ring buffers */
548         i40e_free_arq_bufs(hw);
549
550         i40e_release_spinlock(&hw->aq.arq_spinlock);
551
552         return ret_code;
553 }
554
555 /**
556  *  i40e_init_adminq - main initialization routine for Admin Queue
557  *  @hw: pointer to the hardware structure
558  *
559  *  Prior to calling this function, drivers *MUST* set the following fields
560  *  in the hw->aq structure:
561  *     - hw->aq.num_asq_entries
562  *     - hw->aq.num_arq_entries
563  *     - hw->aq.arq_buf_size
564  *     - hw->aq.asq_buf_size
565  **/
566 enum i40e_status_code i40e_init_adminq(struct i40e_hw *hw)
567 {
568         enum i40e_status_code ret_code;
569         u16 eetrack_lo, eetrack_hi;
570         u16 cfg_ptr, oem_hi, oem_lo;
571         int retry = 0;
572         /* verify input for valid configuration */
573         if ((hw->aq.num_arq_entries == 0) ||
574             (hw->aq.num_asq_entries == 0) ||
575             (hw->aq.arq_buf_size == 0) ||
576             (hw->aq.asq_buf_size == 0)) {
577                 ret_code = I40E_ERR_CONFIG;
578                 goto init_adminq_exit;
579         }
580
581         /* initialize spin locks */
582         i40e_init_spinlock(&hw->aq.asq_spinlock);
583         i40e_init_spinlock(&hw->aq.arq_spinlock);
584
585         /* Set up register offsets */
586         i40e_adminq_init_regs(hw);
587
588         /* setup ASQ command write back timeout */
589         hw->aq.asq_cmd_timeout = I40E_ASQ_CMD_TIMEOUT;
590
591         /* allocate the ASQ */
592         ret_code = i40e_init_asq(hw);
593         if (ret_code != I40E_SUCCESS)
594                 goto init_adminq_destroy_spinlocks;
595
596         /* allocate the ARQ */
597         ret_code = i40e_init_arq(hw);
598         if (ret_code != I40E_SUCCESS)
599                 goto init_adminq_free_asq;
600
601         /* VF has no need of firmware */
602         if (i40e_is_vf(hw))
603                 goto init_adminq_exit;
604         /* There are some cases where the firmware may not be quite ready
605          * for AdminQ operations, so we retry the AdminQ setup a few times
606          * if we see timeouts in this first AQ call.
607          */
608         do {
609                 ret_code = i40e_aq_get_firmware_version(hw,
610                                                         &hw->aq.fw_maj_ver,
611                                                         &hw->aq.fw_min_ver,
612                                                         &hw->aq.fw_build,
613                                                         &hw->aq.api_maj_ver,
614                                                         &hw->aq.api_min_ver,
615                                                         NULL);
616                 if (ret_code != I40E_ERR_ADMIN_QUEUE_TIMEOUT)
617                         break;
618                 retry++;
619                 i40e_msec_delay(100);
620                 i40e_resume_aq(hw);
621         } while (retry < 10);
622         if (ret_code != I40E_SUCCESS)
623                 goto init_adminq_free_arq;
624
625         /* get the NVM version info */
626         i40e_read_nvm_word(hw, I40E_SR_NVM_DEV_STARTER_VERSION,
627                            &hw->nvm.version);
628         i40e_read_nvm_word(hw, I40E_SR_NVM_EETRACK_LO, &eetrack_lo);
629         i40e_read_nvm_word(hw, I40E_SR_NVM_EETRACK_HI, &eetrack_hi);
630         hw->nvm.eetrack = (eetrack_hi << 16) | eetrack_lo;
631         i40e_read_nvm_word(hw, I40E_SR_BOOT_CONFIG_PTR, &cfg_ptr);
632         i40e_read_nvm_word(hw, (cfg_ptr + I40E_NVM_OEM_VER_OFF),
633                            &oem_hi);
634         i40e_read_nvm_word(hw, (cfg_ptr + (I40E_NVM_OEM_VER_OFF + 1)),
635                            &oem_lo);
636         hw->nvm.oem_ver = ((u32)oem_hi << 16) | oem_lo;
637
638         if (hw->aq.api_maj_ver > I40E_FW_API_VERSION_MAJOR) {
639                 ret_code = I40E_ERR_FIRMWARE_API_VERSION;
640                 goto init_adminq_free_arq;
641         }
642
643         /* pre-emptive resource lock release */
644         i40e_aq_release_resource(hw, I40E_NVM_RESOURCE_ID, 0, NULL);
645         hw->aq.nvm_release_on_done = FALSE;
646         hw->nvmupd_state = I40E_NVMUPD_STATE_INIT;
647
648         ret_code = i40e_aq_set_hmc_resource_profile(hw,
649                                                     I40E_HMC_PROFILE_DEFAULT,
650                                                     0,
651                                                     NULL);
652         ret_code = I40E_SUCCESS;
653
654         /* success! */
655         goto init_adminq_exit;
656
657 init_adminq_free_arq:
658         i40e_shutdown_arq(hw);
659 init_adminq_free_asq:
660         i40e_shutdown_asq(hw);
661 init_adminq_destroy_spinlocks:
662         i40e_destroy_spinlock(&hw->aq.asq_spinlock);
663         i40e_destroy_spinlock(&hw->aq.arq_spinlock);
664
665 init_adminq_exit:
666         return ret_code;
667 }
668
669 /**
670  *  i40e_shutdown_adminq - shutdown routine for the Admin Queue
671  *  @hw: pointer to the hardware structure
672  **/
673 enum i40e_status_code i40e_shutdown_adminq(struct i40e_hw *hw)
674 {
675         enum i40e_status_code ret_code = I40E_SUCCESS;
676
677         if (i40e_check_asq_alive(hw))
678                 i40e_aq_queue_shutdown(hw, TRUE);
679
680         i40e_shutdown_asq(hw);
681         i40e_shutdown_arq(hw);
682
683         /* destroy the spinlocks */
684         i40e_destroy_spinlock(&hw->aq.asq_spinlock);
685         i40e_destroy_spinlock(&hw->aq.arq_spinlock);
686
687         if (hw->nvm_buff.va)
688                 i40e_free_virt_mem(hw, &hw->nvm_buff);
689
690         return ret_code;
691 }
692
693 /**
694  *  i40e_clean_asq - cleans Admin send queue
695  *  @hw: pointer to the hardware structure
696  *
697  *  returns the number of free desc
698  **/
699 u16 i40e_clean_asq(struct i40e_hw *hw)
700 {
701         struct i40e_adminq_ring *asq = &(hw->aq.asq);
702         struct i40e_asq_cmd_details *details;
703         u16 ntc = asq->next_to_clean;
704         struct i40e_aq_desc desc_cb;
705         struct i40e_aq_desc *desc;
706
707         desc = I40E_ADMINQ_DESC(*asq, ntc);
708         details = I40E_ADMINQ_DETAILS(*asq, ntc);
709
710         while (rd32(hw, hw->aq.asq.head) != ntc) {
711                 i40e_debug(hw, I40E_DEBUG_AQ_MESSAGE,
712                            "ntc %d head %d.\n", ntc, rd32(hw, hw->aq.asq.head));
713
714                 if (details->callback) {
715                         I40E_ADMINQ_CALLBACK cb_func =
716                                         (I40E_ADMINQ_CALLBACK)details->callback;
717                         i40e_memcpy(&desc_cb, desc, sizeof(struct i40e_aq_desc),
718                                     I40E_DMA_TO_DMA);
719                         cb_func(hw, &desc_cb);
720                 }
721                 i40e_memset(desc, 0, sizeof(*desc), I40E_DMA_MEM);
722                 i40e_memset(details, 0, sizeof(*details), I40E_NONDMA_MEM);
723                 ntc++;
724                 if (ntc == asq->count)
725                         ntc = 0;
726                 desc = I40E_ADMINQ_DESC(*asq, ntc);
727                 details = I40E_ADMINQ_DETAILS(*asq, ntc);
728         }
729
730         asq->next_to_clean = ntc;
731
732         return I40E_DESC_UNUSED(asq);
733 }
734
735 /**
736  *  i40e_asq_done - check if FW has processed the Admin Send Queue
737  *  @hw: pointer to the hw struct
738  *
739  *  Returns TRUE if the firmware has processed all descriptors on the
740  *  admin send queue. Returns FALSE if there are still requests pending.
741  **/
742 bool i40e_asq_done(struct i40e_hw *hw)
743 {
744         /* AQ designers suggest use of head for better
745          * timing reliability than DD bit
746          */
747         return rd32(hw, hw->aq.asq.head) == hw->aq.asq.next_to_use;
748
749 }
750
751 /**
752  *  i40e_asq_send_command - send command to Admin Queue
753  *  @hw: pointer to the hw struct
754  *  @desc: prefilled descriptor describing the command (non DMA mem)
755  *  @buff: buffer to use for indirect commands
756  *  @buff_size: size of buffer for indirect commands
757  *  @cmd_details: pointer to command details structure
758  *
759  *  This is the main send command driver routine for the Admin Queue send
760  *  queue.  It runs the queue, cleans the queue, etc
761  **/
762 enum i40e_status_code i40e_asq_send_command(struct i40e_hw *hw,
763                                 struct i40e_aq_desc *desc,
764                                 void *buff, /* can be NULL */
765                                 u16  buff_size,
766                                 struct i40e_asq_cmd_details *cmd_details)
767 {
768         enum i40e_status_code status = I40E_SUCCESS;
769         struct i40e_dma_mem *dma_buff = NULL;
770         struct i40e_asq_cmd_details *details;
771         struct i40e_aq_desc *desc_on_ring;
772         bool cmd_completed = FALSE;
773         u16  retval = 0;
774         u32  val = 0;
775
776         hw->aq.asq_last_status = I40E_AQ_RC_OK;
777
778         val = rd32(hw, hw->aq.asq.head);
779         if (val >= hw->aq.num_asq_entries) {
780                 i40e_debug(hw, I40E_DEBUG_AQ_MESSAGE,
781                            "AQTX: head overrun at %d\n", val);
782                 status = I40E_ERR_QUEUE_EMPTY;
783                 goto asq_send_command_exit;
784         }
785
786         if (hw->aq.asq.count == 0) {
787                 i40e_debug(hw, I40E_DEBUG_AQ_MESSAGE,
788                            "AQTX: Admin queue not initialized.\n");
789                 status = I40E_ERR_QUEUE_EMPTY;
790                 goto asq_send_command_exit;
791         }
792
793         details = I40E_ADMINQ_DETAILS(hw->aq.asq, hw->aq.asq.next_to_use);
794         if (cmd_details) {
795                 i40e_memcpy(details,
796                             cmd_details,
797                             sizeof(struct i40e_asq_cmd_details),
798                             I40E_NONDMA_TO_NONDMA);
799
800                 /* If the cmd_details are defined copy the cookie.  The
801                  * CPU_TO_LE32 is not needed here because the data is ignored
802                  * by the FW, only used by the driver
803                  */
804                 if (details->cookie) {
805                         desc->cookie_high =
806                                 CPU_TO_LE32(I40E_HI_DWORD(details->cookie));
807                         desc->cookie_low =
808                                 CPU_TO_LE32(I40E_LO_DWORD(details->cookie));
809                 }
810         } else {
811                 i40e_memset(details, 0,
812                             sizeof(struct i40e_asq_cmd_details),
813                             I40E_NONDMA_MEM);
814         }
815
816         /* clear requested flags and then set additional flags if defined */
817         desc->flags &= ~CPU_TO_LE16(details->flags_dis);
818         desc->flags |= CPU_TO_LE16(details->flags_ena);
819
820         i40e_acquire_spinlock(&hw->aq.asq_spinlock);
821
822         if (buff_size > hw->aq.asq_buf_size) {
823                 i40e_debug(hw,
824                            I40E_DEBUG_AQ_MESSAGE,
825                            "AQTX: Invalid buffer size: %d.\n",
826                            buff_size);
827                 status = I40E_ERR_INVALID_SIZE;
828                 goto asq_send_command_error;
829         }
830
831         if (details->postpone && !details->async) {
832                 i40e_debug(hw,
833                            I40E_DEBUG_AQ_MESSAGE,
834                            "AQTX: Async flag not set along with postpone flag");
835                 status = I40E_ERR_PARAM;
836                 goto asq_send_command_error;
837         }
838
839         /* call clean and check queue available function to reclaim the
840          * descriptors that were processed by FW, the function returns the
841          * number of desc available
842          */
843         /* the clean function called here could be called in a separate thread
844          * in case of asynchronous completions
845          */
846         if (i40e_clean_asq(hw) == 0) {
847                 i40e_debug(hw,
848                            I40E_DEBUG_AQ_MESSAGE,
849                            "AQTX: Error queue is full.\n");
850                 status = I40E_ERR_ADMIN_QUEUE_FULL;
851                 goto asq_send_command_error;
852         }
853
854         /* initialize the temp desc pointer with the right desc */
855         desc_on_ring = I40E_ADMINQ_DESC(hw->aq.asq, hw->aq.asq.next_to_use);
856
857         /* if the desc is available copy the temp desc to the right place */
858         i40e_memcpy(desc_on_ring, desc, sizeof(struct i40e_aq_desc),
859                     I40E_NONDMA_TO_DMA);
860
861         /* if buff is not NULL assume indirect command */
862         if (buff != NULL) {
863                 dma_buff = &(hw->aq.asq.r.asq_bi[hw->aq.asq.next_to_use]);
864                 /* copy the user buff into the respective DMA buff */
865                 i40e_memcpy(dma_buff->va, buff, buff_size,
866                             I40E_NONDMA_TO_DMA);
867                 desc_on_ring->datalen = CPU_TO_LE16(buff_size);
868
869                 /* Update the address values in the desc with the pa value
870                  * for respective buffer
871                  */
872                 desc_on_ring->params.external.addr_high =
873                                 CPU_TO_LE32(I40E_HI_DWORD(dma_buff->pa));
874                 desc_on_ring->params.external.addr_low =
875                                 CPU_TO_LE32(I40E_LO_DWORD(dma_buff->pa));
876         }
877
878         /* bump the tail */
879         i40e_debug(hw, I40E_DEBUG_AQ_MESSAGE, "AQTX: desc and buffer:\n");
880         i40e_debug_aq(hw, I40E_DEBUG_AQ_COMMAND, (void *)desc_on_ring,
881                       buff, buff_size);
882         (hw->aq.asq.next_to_use)++;
883         if (hw->aq.asq.next_to_use == hw->aq.asq.count)
884                 hw->aq.asq.next_to_use = 0;
885         if (!details->postpone)
886                 wr32(hw, hw->aq.asq.tail, hw->aq.asq.next_to_use);
887
888         /* if cmd_details are not defined or async flag is not set,
889          * we need to wait for desc write back
890          */
891         if (!details->async && !details->postpone) {
892                 u32 total_delay = 0;
893
894                 do {
895                         /* AQ designers suggest use of head for better
896                          * timing reliability than DD bit
897                          */
898                         if (i40e_asq_done(hw))
899                                 break;
900                         /* ugh! delay while spin_lock */
901                         i40e_msec_delay(1);
902                         total_delay++;
903                 } while (total_delay < hw->aq.asq_cmd_timeout);
904         }
905
906         /* if ready, copy the desc back to temp */
907         if (i40e_asq_done(hw)) {
908                 i40e_memcpy(desc, desc_on_ring, sizeof(struct i40e_aq_desc),
909                             I40E_DMA_TO_NONDMA);
910                 if (buff != NULL)
911                         i40e_memcpy(buff, dma_buff->va, buff_size,
912                                     I40E_DMA_TO_NONDMA);
913                 retval = LE16_TO_CPU(desc->retval);
914                 if (retval != 0) {
915                         i40e_debug(hw,
916                                    I40E_DEBUG_AQ_MESSAGE,
917                                    "AQTX: Command completed with error 0x%X.\n",
918                                    retval);
919
920                         /* strip off FW internal code */
921                         retval &= 0xff;
922                 }
923                 cmd_completed = TRUE;
924                 if ((enum i40e_admin_queue_err)retval == I40E_AQ_RC_OK)
925                         status = I40E_SUCCESS;
926                 else
927                         status = I40E_ERR_ADMIN_QUEUE_ERROR;
928                 hw->aq.asq_last_status = (enum i40e_admin_queue_err)retval;
929         }
930
931         i40e_debug(hw, I40E_DEBUG_AQ_MESSAGE,
932                    "AQTX: desc and buffer writeback:\n");
933         i40e_debug_aq(hw, I40E_DEBUG_AQ_COMMAND, (void *)desc, buff, buff_size);
934
935         /* save writeback aq if requested */
936         if (details->wb_desc)
937                 i40e_memcpy(details->wb_desc, desc_on_ring,
938                             sizeof(struct i40e_aq_desc), I40E_DMA_TO_NONDMA);
939
940         /* update the error if time out occurred */
941         if ((!cmd_completed) &&
942             (!details->async && !details->postpone)) {
943                 i40e_debug(hw,
944                            I40E_DEBUG_AQ_MESSAGE,
945                            "AQTX: Writeback timeout.\n");
946                 status = I40E_ERR_ADMIN_QUEUE_TIMEOUT;
947         }
948
949 asq_send_command_error:
950         i40e_release_spinlock(&hw->aq.asq_spinlock);
951 asq_send_command_exit:
952         return status;
953 }
954
955 /**
956  *  i40e_fill_default_direct_cmd_desc - AQ descriptor helper function
957  *  @desc:     pointer to the temp descriptor (non DMA mem)
958  *  @opcode:   the opcode can be used to decide which flags to turn off or on
959  *
960  *  Fill the desc with default values
961  **/
962 void i40e_fill_default_direct_cmd_desc(struct i40e_aq_desc *desc,
963                                        u16 opcode)
964 {
965         /* zero out the desc */
966         i40e_memset((void *)desc, 0, sizeof(struct i40e_aq_desc),
967                     I40E_NONDMA_MEM);
968         desc->opcode = CPU_TO_LE16(opcode);
969         desc->flags = CPU_TO_LE16(I40E_AQ_FLAG_SI);
970 }
971
972 /**
973  *  i40e_clean_arq_element
974  *  @hw: pointer to the hw struct
975  *  @e: event info from the receive descriptor, includes any buffers
976  *  @pending: number of events that could be left to process
977  *
978  *  This function cleans one Admin Receive Queue element and returns
979  *  the contents through e.  It can also return how many events are
980  *  left to process through 'pending'
981  **/
982 enum i40e_status_code i40e_clean_arq_element(struct i40e_hw *hw,
983                                              struct i40e_arq_event_info *e,
984                                              u16 *pending)
985 {
986         enum i40e_status_code ret_code = I40E_SUCCESS;
987         u16 ntc = hw->aq.arq.next_to_clean;
988         struct i40e_aq_desc *desc;
989         struct i40e_dma_mem *bi;
990         u16 desc_idx;
991         u16 datalen;
992         u16 flags;
993         u16 ntu;
994
995         /* take the lock before we start messing with the ring */
996         i40e_acquire_spinlock(&hw->aq.arq_spinlock);
997
998         /* set next_to_use to head */
999         if (!i40e_is_vf(hw))
1000                 ntu = (rd32(hw, hw->aq.arq.head) & I40E_PF_ARQH_ARQH_MASK);
1001         if (i40e_is_vf(hw))
1002                 ntu = (rd32(hw, hw->aq.arq.head) & I40E_VF_ARQH1_ARQH_MASK);
1003         if (ntu == ntc) {
1004                 /* nothing to do - shouldn't need to update ring's values */
1005                 ret_code = I40E_ERR_ADMIN_QUEUE_NO_WORK;
1006                 goto clean_arq_element_out;
1007         }
1008
1009         /* now clean the next descriptor */
1010         desc = I40E_ADMINQ_DESC(hw->aq.arq, ntc);
1011         desc_idx = ntc;
1012
1013         flags = LE16_TO_CPU(desc->flags);
1014         if (flags & I40E_AQ_FLAG_ERR) {
1015                 ret_code = I40E_ERR_ADMIN_QUEUE_ERROR;
1016                 hw->aq.arq_last_status =
1017                         (enum i40e_admin_queue_err)LE16_TO_CPU(desc->retval);
1018                 i40e_debug(hw,
1019                            I40E_DEBUG_AQ_MESSAGE,
1020                            "AQRX: Event received with error 0x%X.\n",
1021                            hw->aq.arq_last_status);
1022         }
1023
1024         i40e_memcpy(&e->desc, desc, sizeof(struct i40e_aq_desc),
1025                     I40E_DMA_TO_NONDMA);
1026         datalen = LE16_TO_CPU(desc->datalen);
1027         e->msg_len = min(datalen, e->buf_len);
1028         if (e->msg_buf != NULL && (e->msg_len != 0))
1029                 i40e_memcpy(e->msg_buf,
1030                             hw->aq.arq.r.arq_bi[desc_idx].va,
1031                             e->msg_len, I40E_DMA_TO_NONDMA);
1032
1033         i40e_debug(hw, I40E_DEBUG_AQ_MESSAGE, "AQRX: desc and buffer:\n");
1034         i40e_debug_aq(hw, I40E_DEBUG_AQ_COMMAND, (void *)desc, e->msg_buf,
1035                       hw->aq.arq_buf_size);
1036
1037         /* Restore the original datalen and buffer address in the desc,
1038          * FW updates datalen to indicate the event message
1039          * size
1040          */
1041         bi = &hw->aq.arq.r.arq_bi[ntc];
1042         i40e_memset((void *)desc, 0, sizeof(struct i40e_aq_desc), I40E_DMA_MEM);
1043
1044         desc->flags = CPU_TO_LE16(I40E_AQ_FLAG_BUF);
1045         if (hw->aq.arq_buf_size > I40E_AQ_LARGE_BUF)
1046                 desc->flags |= CPU_TO_LE16(I40E_AQ_FLAG_LB);
1047         desc->datalen = CPU_TO_LE16((u16)bi->size);
1048         desc->params.external.addr_high = CPU_TO_LE32(I40E_HI_DWORD(bi->pa));
1049         desc->params.external.addr_low = CPU_TO_LE32(I40E_LO_DWORD(bi->pa));
1050
1051         /* set tail = the last cleaned desc index. */
1052         wr32(hw, hw->aq.arq.tail, ntc);
1053         /* ntc is updated to tail + 1 */
1054         ntc++;
1055         if (ntc == hw->aq.num_arq_entries)
1056                 ntc = 0;
1057         hw->aq.arq.next_to_clean = ntc;
1058         hw->aq.arq.next_to_use = ntu;
1059
1060 clean_arq_element_out:
1061         /* Set pending if needed, unlock and return */
1062         if (pending != NULL)
1063                 *pending = (ntc > ntu ? hw->aq.arq.count : 0) + (ntu - ntc);
1064         i40e_release_spinlock(&hw->aq.arq_spinlock);
1065
1066         if (i40e_is_nvm_update_op(&e->desc)) {
1067                 if (hw->aq.nvm_release_on_done) {
1068                         i40e_release_nvm(hw);
1069                         hw->aq.nvm_release_on_done = FALSE;
1070                 }
1071
1072                 switch (hw->nvmupd_state) {
1073                 case I40E_NVMUPD_STATE_INIT_WAIT:
1074                         hw->nvmupd_state = I40E_NVMUPD_STATE_INIT;
1075                         break;
1076
1077                 case I40E_NVMUPD_STATE_WRITE_WAIT:
1078                         hw->nvmupd_state = I40E_NVMUPD_STATE_WRITING;
1079                         break;
1080
1081                 default:
1082                         break;
1083                 }
1084         }
1085
1086         return ret_code;
1087 }
1088
1089 void i40e_resume_aq(struct i40e_hw *hw)
1090 {
1091         /* Registers are reset after PF reset */
1092         hw->aq.asq.next_to_use = 0;
1093         hw->aq.asq.next_to_clean = 0;
1094
1095         i40e_config_asq_regs(hw);
1096
1097         hw->aq.arq.next_to_use = 0;
1098         hw->aq.arq.next_to_clean = 0;
1099
1100         i40e_config_arq_regs(hw);
1101 }