]> CyberLeo.Net >> Repos - FreeBSD/releng/10.3.git/blob - sys/dev/nvme/nvme_sysctl.c
- Copy stable/10@296371 to releng/10.3 in preparation for 10.3-RC1
[FreeBSD/releng/10.3.git] / sys / dev / nvme / nvme_sysctl.c
1 /*-
2  * Copyright (C) 2012-2016 Intel Corporation
3  * All rights reserved.
4  *
5  * Redistribution and use in source and binary forms, with or without
6  * modification, are permitted provided that the following conditions
7  * are met:
8  * 1. Redistributions of source code must retain the above copyright
9  *    notice, this list of conditions and the following disclaimer.
10  * 2. Redistributions in binary form must reproduce the above copyright
11  *    notice, this list of conditions and the following disclaimer in the
12  *    documentation and/or other materials provided with the distribution.
13  *
14  * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
15  * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
16  * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
17  * ARE DISCLAIMED.  IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
18  * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
19  * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
20  * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
21  * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
22  * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
23  * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
24  * SUCH DAMAGE.
25  */
26
27 #include <sys/cdefs.h>
28 __FBSDID("$FreeBSD$");
29
30 #include <sys/param.h>
31 #include <sys/bus.h>
32 #include <sys/sysctl.h>
33
34 #include "nvme_private.h"
35
36 SYSCTL_NODE(_kern, OID_AUTO, nvme, CTLFLAG_RD, 0, "NVM Express");
37 /*
38  * Intel NVMe controllers have a slow path for I/Os that span a 128KB
39  * stripe boundary but ZFS limits ashift, which is derived from
40  * d_stripesize, to 13 (8KB) so we limit the stripesize reported to
41  * geom(8) to 4KB by default.
42  *
43  * This may result in a small number of additional I/Os to require
44  * splitting in nvme(4), however the NVMe I/O path is very efficient
45  * so these additional I/Os will cause very minimal (if any) difference
46  * in performance or CPU utilisation.
47  */
48 int nvme_max_optimal_sectorsize = 1<<12;
49 SYSCTL_INT(_kern_nvme, OID_AUTO, max_optimal_sectorsize, CTLFLAG_RWTUN,
50     &nvme_max_optimal_sectorsize, 0, "The maximum optimal sectorsize reported");
51
52 /*
53  * CTLTYPE_S64 and sysctl_handle_64 were added in r217616.  Define these
54  *  explicitly here for older kernels that don't include the r217616
55  *  changeset.
56  */
57 #ifndef CTLTYPE_S64
58 #define CTLTYPE_S64             CTLTYPE_QUAD
59 #define sysctl_handle_64        sysctl_handle_quad
60 #endif
61
62 static void
63 nvme_dump_queue(struct nvme_qpair *qpair)
64 {
65         struct nvme_completion *cpl;
66         struct nvme_command *cmd;
67         int i;
68
69         printf("id:%04Xh phase:%d\n", qpair->id, qpair->phase);
70
71         printf("Completion queue:\n");
72         for (i = 0; i < qpair->num_entries; i++) {
73                 cpl = &qpair->cpl[i];
74                 printf("%05d: ", i);
75                 nvme_dump_completion(cpl);
76         }
77
78         printf("Submission queue:\n");
79         for (i = 0; i < qpair->num_entries; i++) {
80                 cmd = &qpair->cmd[i];
81                 printf("%05d: ", i);
82                 nvme_dump_command(cmd);
83         }
84 }
85
86
87 static int
88 nvme_sysctl_dump_debug(SYSCTL_HANDLER_ARGS)
89 {
90         struct nvme_qpair       *qpair = arg1;
91         uint32_t                val = 0;
92
93         int error = sysctl_handle_int(oidp, &val, 0, req);
94
95         if (error)
96                 return (error);
97
98         if (val != 0)
99                 nvme_dump_queue(qpair);
100
101         return (0);
102 }
103
104 static int
105 nvme_sysctl_int_coal_time(SYSCTL_HANDLER_ARGS)
106 {
107         struct nvme_controller *ctrlr = arg1;
108         uint32_t oldval = ctrlr->int_coal_time;
109         int error = sysctl_handle_int(oidp, &ctrlr->int_coal_time, 0,
110             req);
111
112         if (error)
113                 return (error);
114
115         if (oldval != ctrlr->int_coal_time)
116                 nvme_ctrlr_cmd_set_interrupt_coalescing(ctrlr,
117                     ctrlr->int_coal_time, ctrlr->int_coal_threshold, NULL,
118                     NULL);
119
120         return (0);
121 }
122
123 static int
124 nvme_sysctl_int_coal_threshold(SYSCTL_HANDLER_ARGS)
125 {
126         struct nvme_controller *ctrlr = arg1;
127         uint32_t oldval = ctrlr->int_coal_threshold;
128         int error = sysctl_handle_int(oidp, &ctrlr->int_coal_threshold, 0,
129             req);
130
131         if (error)
132                 return (error);
133
134         if (oldval != ctrlr->int_coal_threshold)
135                 nvme_ctrlr_cmd_set_interrupt_coalescing(ctrlr,
136                     ctrlr->int_coal_time, ctrlr->int_coal_threshold, NULL,
137                     NULL);
138
139         return (0);
140 }
141
142 static int
143 nvme_sysctl_timeout_period(SYSCTL_HANDLER_ARGS)
144 {
145         struct nvme_controller *ctrlr = arg1;
146         uint32_t oldval = ctrlr->timeout_period;
147         int error = sysctl_handle_int(oidp, &ctrlr->timeout_period, 0, req);
148
149         if (error)
150                 return (error);
151
152         if (ctrlr->timeout_period > NVME_MAX_TIMEOUT_PERIOD ||
153             ctrlr->timeout_period < NVME_MIN_TIMEOUT_PERIOD) {
154                 ctrlr->timeout_period = oldval;
155                 return (EINVAL);
156         }
157
158         return (0);
159 }
160
161 static void
162 nvme_qpair_reset_stats(struct nvme_qpair *qpair)
163 {
164
165         qpair->num_cmds = 0;
166         qpair->num_intr_handler_calls = 0;
167 }
168
169 static int
170 nvme_sysctl_num_cmds(SYSCTL_HANDLER_ARGS)
171 {
172         struct nvme_controller  *ctrlr = arg1;
173         int64_t                 num_cmds = 0;
174         int                     i;
175
176         num_cmds = ctrlr->adminq.num_cmds;
177
178         for (i = 0; i < ctrlr->num_io_queues; i++)
179                 num_cmds += ctrlr->ioq[i].num_cmds;
180
181         return (sysctl_handle_64(oidp, &num_cmds, 0, req));
182 }
183
184 static int
185 nvme_sysctl_num_intr_handler_calls(SYSCTL_HANDLER_ARGS)
186 {
187         struct nvme_controller  *ctrlr = arg1;
188         int64_t                 num_intr_handler_calls = 0;
189         int                     i;
190
191         num_intr_handler_calls = ctrlr->adminq.num_intr_handler_calls;
192
193         for (i = 0; i < ctrlr->num_io_queues; i++)
194                 num_intr_handler_calls += ctrlr->ioq[i].num_intr_handler_calls;
195
196         return (sysctl_handle_64(oidp, &num_intr_handler_calls, 0, req));
197 }
198
199 static int
200 nvme_sysctl_reset_stats(SYSCTL_HANDLER_ARGS)
201 {
202         struct nvme_controller  *ctrlr = arg1;
203         uint32_t                i, val = 0;
204
205         int error = sysctl_handle_int(oidp, &val, 0, req);
206
207         if (error)
208                 return (error);
209
210         if (val != 0) {
211                 nvme_qpair_reset_stats(&ctrlr->adminq);
212
213                 for (i = 0; i < ctrlr->num_io_queues; i++)
214                         nvme_qpair_reset_stats(&ctrlr->ioq[i]);
215         }
216
217         return (0);
218 }
219
220
221 static void
222 nvme_sysctl_initialize_queue(struct nvme_qpair *qpair,
223     struct sysctl_ctx_list *ctrlr_ctx, struct sysctl_oid *que_tree)
224 {
225         struct sysctl_oid_list  *que_list = SYSCTL_CHILDREN(que_tree);
226
227         SYSCTL_ADD_UINT(ctrlr_ctx, que_list, OID_AUTO, "num_entries",
228             CTLFLAG_RD, &qpair->num_entries, 0,
229             "Number of entries in hardware queue");
230         SYSCTL_ADD_UINT(ctrlr_ctx, que_list, OID_AUTO, "num_trackers",
231             CTLFLAG_RD, &qpair->num_trackers, 0,
232             "Number of trackers pre-allocated for this queue pair");
233         SYSCTL_ADD_UINT(ctrlr_ctx, que_list, OID_AUTO, "sq_head",
234             CTLFLAG_RD, &qpair->sq_head, 0,
235             "Current head of submission queue (as observed by driver)");
236         SYSCTL_ADD_UINT(ctrlr_ctx, que_list, OID_AUTO, "sq_tail",
237             CTLFLAG_RD, &qpair->sq_tail, 0,
238             "Current tail of submission queue (as observed by driver)");
239         SYSCTL_ADD_UINT(ctrlr_ctx, que_list, OID_AUTO, "cq_head",
240             CTLFLAG_RD, &qpair->cq_head, 0,
241             "Current head of completion queue (as observed by driver)");
242
243         SYSCTL_ADD_QUAD(ctrlr_ctx, que_list, OID_AUTO, "num_cmds",
244             CTLFLAG_RD, &qpair->num_cmds, "Number of commands submitted");
245         SYSCTL_ADD_QUAD(ctrlr_ctx, que_list, OID_AUTO, "num_intr_handler_calls",
246             CTLFLAG_RD, &qpair->num_intr_handler_calls,
247             "Number of times interrupt handler was invoked (will typically be "
248             "less than number of actual interrupts generated due to "
249             "coalescing)");
250
251         SYSCTL_ADD_PROC(ctrlr_ctx, que_list, OID_AUTO,
252             "dump_debug", CTLTYPE_UINT | CTLFLAG_RW, qpair, 0,
253             nvme_sysctl_dump_debug, "IU", "Dump debug data");
254 }
255
256 void
257 nvme_sysctl_initialize_ctrlr(struct nvme_controller *ctrlr)
258 {
259         struct sysctl_ctx_list  *ctrlr_ctx;
260         struct sysctl_oid       *ctrlr_tree, *que_tree;
261         struct sysctl_oid_list  *ctrlr_list;
262 #define QUEUE_NAME_LENGTH       16
263         char                    queue_name[QUEUE_NAME_LENGTH];
264         int                     i;
265
266         ctrlr_ctx = device_get_sysctl_ctx(ctrlr->dev);
267         ctrlr_tree = device_get_sysctl_tree(ctrlr->dev);
268         ctrlr_list = SYSCTL_CHILDREN(ctrlr_tree);
269
270         SYSCTL_ADD_UINT(ctrlr_ctx, ctrlr_list, OID_AUTO, "num_cpus_per_ioq",
271             CTLFLAG_RD, &ctrlr->num_cpus_per_ioq, 0,
272             "Number of CPUs assigned per I/O queue pair");
273
274         SYSCTL_ADD_PROC(ctrlr_ctx, ctrlr_list, OID_AUTO,
275             "int_coal_time", CTLTYPE_UINT | CTLFLAG_RW, ctrlr, 0,
276             nvme_sysctl_int_coal_time, "IU",
277             "Interrupt coalescing timeout (in microseconds)");
278
279         SYSCTL_ADD_PROC(ctrlr_ctx, ctrlr_list, OID_AUTO,
280             "int_coal_threshold", CTLTYPE_UINT | CTLFLAG_RW, ctrlr, 0,
281             nvme_sysctl_int_coal_threshold, "IU",
282             "Interrupt coalescing threshold");
283
284         SYSCTL_ADD_PROC(ctrlr_ctx, ctrlr_list, OID_AUTO,
285             "timeout_period", CTLTYPE_UINT | CTLFLAG_RW, ctrlr, 0,
286             nvme_sysctl_timeout_period, "IU",
287             "Timeout period (in seconds)");
288
289         SYSCTL_ADD_PROC(ctrlr_ctx, ctrlr_list, OID_AUTO,
290             "num_cmds", CTLTYPE_S64 | CTLFLAG_RD,
291             ctrlr, 0, nvme_sysctl_num_cmds, "IU",
292             "Number of commands submitted");
293
294         SYSCTL_ADD_PROC(ctrlr_ctx, ctrlr_list, OID_AUTO,
295             "num_intr_handler_calls", CTLTYPE_S64 | CTLFLAG_RD,
296             ctrlr, 0, nvme_sysctl_num_intr_handler_calls, "IU",
297             "Number of times interrupt handler was invoked (will "
298             "typically be less than number of actual interrupts "
299             "generated due to coalescing)");
300
301         SYSCTL_ADD_PROC(ctrlr_ctx, ctrlr_list, OID_AUTO,
302             "reset_stats", CTLTYPE_UINT | CTLFLAG_RW, ctrlr, 0,
303             nvme_sysctl_reset_stats, "IU", "Reset statistics to zero");
304
305         que_tree = SYSCTL_ADD_NODE(ctrlr_ctx, ctrlr_list, OID_AUTO, "adminq",
306             CTLFLAG_RD, NULL, "Admin Queue");
307
308         nvme_sysctl_initialize_queue(&ctrlr->adminq, ctrlr_ctx, que_tree);
309
310         for (i = 0; i < ctrlr->num_io_queues; i++) {
311                 snprintf(queue_name, QUEUE_NAME_LENGTH, "ioq%d", i);
312                 que_tree = SYSCTL_ADD_NODE(ctrlr_ctx, ctrlr_list, OID_AUTO,
313                     queue_name, CTLFLAG_RD, NULL, "IO Queue");
314                 nvme_sysctl_initialize_queue(&ctrlr->ioq[i], ctrlr_ctx,
315                     que_tree);
316         }
317 }