]> CyberLeo.Net >> Repos - FreeBSD/FreeBSD.git/blob - sys/dev/nvme/nvme.c
nvme: start qpair in state RECOVERY_WAITING
[FreeBSD/FreeBSD.git] / sys / dev / nvme / nvme.c
1 /*-
2  * SPDX-License-Identifier: BSD-2-Clause-FreeBSD
3  *
4  * Copyright (C) 2012-2014 Intel Corporation
5  * All rights reserved.
6  *
7  * Redistribution and use in source and binary forms, with or without
8  * modification, are permitted provided that the following conditions
9  * are met:
10  * 1. Redistributions of source code must retain the above copyright
11  *    notice, this list of conditions and the following disclaimer.
12  * 2. Redistributions in binary form must reproduce the above copyright
13  *    notice, this list of conditions and the following disclaimer in the
14  *    documentation and/or other materials provided with the distribution.
15  *
16  * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
17  * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
18  * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
19  * ARE DISCLAIMED.  IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
20  * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
21  * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
22  * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
23  * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
24  * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
25  * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
26  * SUCH DAMAGE.
27  */
28
29 #include <sys/cdefs.h>
30 __FBSDID("$FreeBSD$");
31
32 #include <sys/param.h>
33 #include <sys/bus.h>
34 #include <sys/conf.h>
35 #include <sys/module.h>
36
37 #include <vm/uma.h>
38
39 #include "nvme_private.h"
40
41 struct nvme_consumer {
42         uint32_t                id;
43         nvme_cons_ns_fn_t       ns_fn;
44         nvme_cons_ctrlr_fn_t    ctrlr_fn;
45         nvme_cons_async_fn_t    async_fn;
46         nvme_cons_fail_fn_t     fail_fn;
47 };
48
49 struct nvme_consumer nvme_consumer[NVME_MAX_CONSUMERS];
50 #define INVALID_CONSUMER_ID     0xFFFF
51
52 int32_t         nvme_retry_count;
53
54 MALLOC_DEFINE(M_NVME, "nvme", "nvme(4) memory allocations");
55
56 devclass_t nvme_devclass;
57
58 static void
59 nvme_init(void)
60 {
61         uint32_t        i;
62
63         for (i = 0; i < NVME_MAX_CONSUMERS; i++)
64                 nvme_consumer[i].id = INVALID_CONSUMER_ID;
65 }
66
67 SYSINIT(nvme_register, SI_SUB_DRIVERS, SI_ORDER_SECOND, nvme_init, NULL);
68
69 static void
70 nvme_uninit(void)
71 {
72 }
73
74 SYSUNINIT(nvme_unregister, SI_SUB_DRIVERS, SI_ORDER_SECOND, nvme_uninit, NULL);
75
76 int
77 nvme_shutdown(device_t dev)
78 {
79         struct nvme_controller  *ctrlr;
80
81         ctrlr = DEVICE2SOFTC(dev);
82         nvme_ctrlr_shutdown(ctrlr);
83
84         return (0);
85 }
86
87 void
88 nvme_dump_command(struct nvme_command *cmd)
89 {
90
91         printf(
92 "opc:%x f:%x cid:%x nsid:%x r2:%x r3:%x mptr:%jx prp1:%jx prp2:%jx cdw:%x %x %x %x %x %x\n",
93             cmd->opc, cmd->fuse, cmd->cid, le32toh(cmd->nsid),
94             cmd->rsvd2, cmd->rsvd3,
95             (uintmax_t)le64toh(cmd->mptr), (uintmax_t)le64toh(cmd->prp1), (uintmax_t)le64toh(cmd->prp2),
96             le32toh(cmd->cdw10), le32toh(cmd->cdw11), le32toh(cmd->cdw12),
97             le32toh(cmd->cdw13), le32toh(cmd->cdw14), le32toh(cmd->cdw15));
98 }
99
100 void
101 nvme_dump_completion(struct nvme_completion *cpl)
102 {
103         uint8_t p, sc, sct, m, dnr;
104         uint16_t status;
105
106         status = le16toh(cpl->status);
107
108         p = NVME_STATUS_GET_P(status);
109         sc = NVME_STATUS_GET_SC(status);
110         sct = NVME_STATUS_GET_SCT(status);
111         m = NVME_STATUS_GET_M(status);
112         dnr = NVME_STATUS_GET_DNR(status);
113
114         printf("cdw0:%08x sqhd:%04x sqid:%04x "
115             "cid:%04x p:%x sc:%02x sct:%x m:%x dnr:%x\n",
116             le32toh(cpl->cdw0), le16toh(cpl->sqhd), le16toh(cpl->sqid),
117             cpl->cid, p, sc, sct, m, dnr);
118 }
119
120 int
121 nvme_attach(device_t dev)
122 {
123         struct nvme_controller  *ctrlr = DEVICE2SOFTC(dev);
124         int                     status;
125
126         status = nvme_ctrlr_construct(ctrlr, dev);
127         if (status != 0) {
128                 nvme_ctrlr_destruct(ctrlr, dev);
129                 return (status);
130         }
131
132         ctrlr->config_hook.ich_func = nvme_ctrlr_start_config_hook;
133         ctrlr->config_hook.ich_arg = ctrlr;
134
135         if (config_intrhook_establish(&ctrlr->config_hook) != 0)
136                 return (ENOMEM);
137
138         return (0);
139 }
140
141 int
142 nvme_detach(device_t dev)
143 {
144         struct nvme_controller  *ctrlr = DEVICE2SOFTC(dev);
145
146         config_intrhook_drain(&ctrlr->config_hook);
147
148         nvme_ctrlr_destruct(ctrlr, dev);
149         return (0);
150 }
151
152 static void
153 nvme_notify(struct nvme_consumer *cons,
154             struct nvme_controller *ctrlr)
155 {
156         struct nvme_namespace   *ns;
157         void                    *ctrlr_cookie;
158         int                     cmpset, ns_idx;
159
160         /*
161          * The consumer may register itself after the nvme devices
162          *  have registered with the kernel, but before the
163          *  driver has completed initialization.  In that case,
164          *  return here, and when initialization completes, the
165          *  controller will make sure the consumer gets notified.
166          */
167         if (!ctrlr->is_initialized)
168                 return;
169
170         cmpset = atomic_cmpset_32(&ctrlr->notification_sent, 0, 1);
171         if (cmpset == 0)
172                 return;
173
174         if (cons->ctrlr_fn != NULL)
175                 ctrlr_cookie = (*cons->ctrlr_fn)(ctrlr);
176         else
177                 ctrlr_cookie = (void *)(uintptr_t)0xdeadc0dedeadc0de;
178         ctrlr->cons_cookie[cons->id] = ctrlr_cookie;
179
180         /* ctrlr_fn has failed.  Nothing to notify here any more. */
181         if (ctrlr_cookie == NULL)
182                 return;
183
184         if (ctrlr->is_failed) {
185                 ctrlr->cons_cookie[cons->id] = NULL;
186                 if (cons->fail_fn != NULL)
187                         (*cons->fail_fn)(ctrlr_cookie);
188                 /*
189                  * Do not notify consumers about the namespaces of a
190                  *  failed controller.
191                  */
192                 return;
193         }
194         for (ns_idx = 0; ns_idx < min(ctrlr->cdata.nn, NVME_MAX_NAMESPACES); ns_idx++) {
195                 ns = &ctrlr->ns[ns_idx];
196                 if (ns->data.nsze == 0)
197                         continue;
198                 if (cons->ns_fn != NULL)
199                         ns->cons_cookie[cons->id] =
200                             (*cons->ns_fn)(ns, ctrlr_cookie);
201         }
202 }
203
204 void
205 nvme_notify_new_controller(struct nvme_controller *ctrlr)
206 {
207         int i;
208
209         for (i = 0; i < NVME_MAX_CONSUMERS; i++) {
210                 if (nvme_consumer[i].id != INVALID_CONSUMER_ID) {
211                         nvme_notify(&nvme_consumer[i], ctrlr);
212                 }
213         }
214 }
215
216 static void
217 nvme_notify_new_consumer(struct nvme_consumer *cons)
218 {
219         device_t                *devlist;
220         struct nvme_controller  *ctrlr;
221         int                     dev_idx, devcount;
222
223         if (devclass_get_devices(nvme_devclass, &devlist, &devcount))
224                 return;
225
226         for (dev_idx = 0; dev_idx < devcount; dev_idx++) {
227                 ctrlr = DEVICE2SOFTC(devlist[dev_idx]);
228                 nvme_notify(cons, ctrlr);
229         }
230
231         free(devlist, M_TEMP);
232 }
233
234 void
235 nvme_notify_async_consumers(struct nvme_controller *ctrlr,
236                             const struct nvme_completion *async_cpl,
237                             uint32_t log_page_id, void *log_page_buffer,
238                             uint32_t log_page_size)
239 {
240         struct nvme_consumer    *cons;
241         void                    *ctrlr_cookie;
242         uint32_t                i;
243
244         for (i = 0; i < NVME_MAX_CONSUMERS; i++) {
245                 cons = &nvme_consumer[i];
246                 if (cons->id != INVALID_CONSUMER_ID && cons->async_fn != NULL &&
247                     (ctrlr_cookie = ctrlr->cons_cookie[i]) != NULL) {
248                         (*cons->async_fn)(ctrlr_cookie, async_cpl,
249                             log_page_id, log_page_buffer, log_page_size);
250                 }
251         }
252 }
253
254 void
255 nvme_notify_fail_consumers(struct nvme_controller *ctrlr)
256 {
257         struct nvme_consumer    *cons;
258         void                    *ctrlr_cookie;
259         uint32_t                i;
260
261         /*
262          * This controller failed during initialization (i.e. IDENTIFY
263          *  command failed or timed out).  Do not notify any nvme
264          *  consumers of the failure here, since the consumer does not
265          *  even know about the controller yet.
266          */
267         if (!ctrlr->is_initialized)
268                 return;
269
270         for (i = 0; i < NVME_MAX_CONSUMERS; i++) {
271                 cons = &nvme_consumer[i];
272                 if (cons->id != INVALID_CONSUMER_ID &&
273                     (ctrlr_cookie = ctrlr->cons_cookie[i]) != NULL) {
274                         ctrlr->cons_cookie[i] = NULL;
275                         if (cons->fail_fn != NULL)
276                                 cons->fail_fn(ctrlr_cookie);
277                 }
278         }
279 }
280
281 void
282 nvme_notify_ns(struct nvme_controller *ctrlr, int nsid)
283 {
284         struct nvme_consumer    *cons;
285         struct nvme_namespace   *ns;
286         void                    *ctrlr_cookie;
287         uint32_t                i;
288
289         KASSERT(nsid <= NVME_MAX_NAMESPACES,
290             ("%s: Namespace notification to nsid %d exceeds range\n",
291                 device_get_nameunit(ctrlr->dev), nsid));
292
293         if (!ctrlr->is_initialized)
294                 return;
295
296         ns = &ctrlr->ns[nsid - 1];
297         for (i = 0; i < NVME_MAX_CONSUMERS; i++) {
298                 cons = &nvme_consumer[i];
299                 if (cons->id != INVALID_CONSUMER_ID && cons->ns_fn != NULL &&
300                     (ctrlr_cookie = ctrlr->cons_cookie[i]) != NULL)
301                         ns->cons_cookie[i] = (*cons->ns_fn)(ns, ctrlr_cookie);
302         }
303 }
304
305 struct nvme_consumer *
306 nvme_register_consumer(nvme_cons_ns_fn_t ns_fn, nvme_cons_ctrlr_fn_t ctrlr_fn,
307                        nvme_cons_async_fn_t async_fn,
308                        nvme_cons_fail_fn_t fail_fn)
309 {
310         int i;
311
312         /*
313          * TODO: add locking around consumer registration.
314          */
315         for (i = 0; i < NVME_MAX_CONSUMERS; i++)
316                 if (nvme_consumer[i].id == INVALID_CONSUMER_ID) {
317                         nvme_consumer[i].id = i;
318                         nvme_consumer[i].ns_fn = ns_fn;
319                         nvme_consumer[i].ctrlr_fn = ctrlr_fn;
320                         nvme_consumer[i].async_fn = async_fn;
321                         nvme_consumer[i].fail_fn = fail_fn;
322
323                         nvme_notify_new_consumer(&nvme_consumer[i]);
324                         return (&nvme_consumer[i]);
325                 }
326
327         printf("nvme(4): consumer not registered - no slots available\n");
328         return (NULL);
329 }
330
331 void
332 nvme_unregister_consumer(struct nvme_consumer *consumer)
333 {
334
335         consumer->id = INVALID_CONSUMER_ID;
336 }
337
338 void
339 nvme_completion_poll_cb(void *arg, const struct nvme_completion *cpl)
340 {
341         struct nvme_completion_poll_status      *status = arg;
342
343         /*
344          * Copy status into the argument passed by the caller, so that
345          *  the caller can check the status to determine if the
346          *  the request passed or failed.
347          */
348         memcpy(&status->cpl, cpl, sizeof(*cpl));
349         atomic_store_rel_int(&status->done, 1);
350 }
351
352 static int
353 nvme_modevent(module_t mod __unused, int type __unused, void *argp __unused)
354 {
355        return (0);
356 }
357
358 static moduledata_t nvme_mod = {
359        "nvme",
360        nvme_modevent,
361        0
362 };
363
364 DECLARE_MODULE(nvme, nvme_mod, SI_SUB_DRIVERS, SI_ORDER_FIRST);
365 MODULE_VERSION(nvme, 1);
366 MODULE_DEPEND(nvme, cam, 1, 1, 1);