2 * Copyright (C) 2012-2016 Intel Corporation
5 * Redistribution and use in source and binary forms, with or without
6 * modification, are permitted provided that the following conditions
8 * 1. Redistributions of source code must retain the above copyright
9 * notice, this list of conditions and the following disclaimer.
10 * 2. Redistributions in binary form must reproduce the above copyright
11 * notice, this list of conditions and the following disclaimer in the
12 * documentation and/or other materials provided with the distribution.
14 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
15 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
16 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
17 * ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
18 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
19 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
20 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
21 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
22 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
23 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
27 #include <sys/cdefs.h>
28 __FBSDID("$FreeBSD$");
30 #include <sys/param.h>
32 #include <sys/kernel.h>
33 #include <sys/malloc.h>
34 #include <sys/module.h>
35 #include <sys/sysctl.h>
36 #include <sys/systm.h>
37 #include <sys/taskqueue.h>
39 #include <geom/geom.h>
40 #include <geom/geom_disk.h>
42 #include <dev/nvme/nvme.h>
48 static disk_ioctl_t nvd_ioctl;
49 static disk_strategy_t nvd_strategy;
51 static void nvd_done(void *arg, const struct nvme_completion *cpl);
53 static void *nvd_new_disk(struct nvme_namespace *ns, void *ctrlr);
54 static void destroy_geom_disk(struct nvd_disk *ndisk);
56 static void *nvd_new_controller(struct nvme_controller *ctrlr);
57 static void nvd_controller_fail(void *ctrlr);
59 static int nvd_load(void);
60 static void nvd_unload(void);
62 MALLOC_DEFINE(M_NVD, "nvd", "nvd(4) allocations");
64 struct nvme_consumer *consumer_handle;
68 struct bio_queue_head bioq;
74 struct nvme_namespace *ns;
77 uint32_t ordered_in_flight;
79 TAILQ_ENTRY(nvd_disk) global_tailq;
80 TAILQ_ENTRY(nvd_disk) ctrlr_tailq;
83 struct nvd_controller {
85 TAILQ_ENTRY(nvd_controller) tailq;
86 TAILQ_HEAD(, nvd_disk) disk_head;
89 static TAILQ_HEAD(, nvd_controller) ctrlr_head;
90 static TAILQ_HEAD(disk_list, nvd_disk) disk_head;
92 static SYSCTL_NODE(_hw, OID_AUTO, nvd, CTLFLAG_RD, 0, "nvd driver parameters");
94 * The NVMe specification does not define a maximum or optimal delete size, so
95 * technically max delete size is min(full size of the namespace, 2^32 - 1
96 * LBAs). A single delete for a multi-TB NVMe namespace though may take much
97 * longer to complete than the nvme(4) I/O timeout period. So choose a sensible
98 * default here that is still suitably large to minimize the number of overall
101 static uint64_t nvd_delete_max = (1024 * 1024 * 1024); /* 1GB */
102 SYSCTL_UQUAD(_hw_nvd, OID_AUTO, delete_max, CTLFLAG_RDTUN, &nvd_delete_max, 0,
103 "nvd maximum BIO_DELETE size in bytes");
105 static int nvd_modevent(module_t mod, int type, void *arg)
123 moduledata_t nvd_mod = {
125 (modeventhand_t)nvd_modevent,
129 DECLARE_MODULE(nvd, nvd_mod, SI_SUB_DRIVERS, SI_ORDER_ANY);
130 MODULE_VERSION(nvd, 1);
131 MODULE_DEPEND(nvd, nvme, 1, 1, 1);
137 TAILQ_INIT(&ctrlr_head);
138 TAILQ_INIT(&disk_head);
140 consumer_handle = nvme_register_consumer(nvd_new_disk,
141 nvd_new_controller, NULL, nvd_controller_fail);
143 return (consumer_handle != NULL ? 0 : -1);
149 struct nvd_controller *ctrlr;
150 struct nvd_disk *disk;
152 while (!TAILQ_EMPTY(&ctrlr_head)) {
153 ctrlr = TAILQ_FIRST(&ctrlr_head);
154 TAILQ_REMOVE(&ctrlr_head, ctrlr, tailq);
158 while (!TAILQ_EMPTY(&disk_head)) {
159 disk = TAILQ_FIRST(&disk_head);
160 TAILQ_REMOVE(&disk_head, disk, global_tailq);
161 destroy_geom_disk(disk);
165 nvme_unregister_consumer(consumer_handle);
169 nvd_bio_submit(struct nvd_disk *ndisk, struct bio *bp)
173 bp->bio_driver1 = NULL;
174 atomic_add_int(&ndisk->cur_depth, 1);
175 err = nvme_ns_bio_process(ndisk->ns, bp, nvd_done);
177 atomic_add_int(&ndisk->cur_depth, -1);
178 if (__predict_false(bp->bio_flags & BIO_ORDERED))
179 atomic_add_int(&ndisk->ordered_in_flight, -1);
181 bp->bio_flags |= BIO_ERROR;
182 bp->bio_resid = bp->bio_bcount;
191 nvd_strategy(struct bio *bp)
193 struct nvd_disk *ndisk;
195 ndisk = (struct nvd_disk *)bp->bio_disk->d_drv1;
197 if (__predict_false(bp->bio_flags & BIO_ORDERED))
198 atomic_add_int(&ndisk->ordered_in_flight, 1);
200 if (__predict_true(ndisk->ordered_in_flight == 0)) {
201 nvd_bio_submit(ndisk, bp);
206 * There are ordered bios in flight, so we need to submit
207 * bios through the task queue to enforce ordering.
209 mtx_lock(&ndisk->bioqlock);
210 bioq_insert_tail(&ndisk->bioq, bp);
211 mtx_unlock(&ndisk->bioqlock);
212 taskqueue_enqueue(ndisk->tq, &ndisk->bioqtask);
216 nvd_ioctl(struct disk *ndisk, u_long cmd, void *data, int fflag,
230 nvd_done(void *arg, const struct nvme_completion *cpl)
233 struct nvd_disk *ndisk;
235 bp = (struct bio *)arg;
237 ndisk = bp->bio_disk->d_drv1;
239 atomic_add_int(&ndisk->cur_depth, -1);
240 if (__predict_false(bp->bio_flags & BIO_ORDERED))
241 atomic_add_int(&ndisk->ordered_in_flight, -1);
247 nvd_bioq_process(void *arg, int pending)
249 struct nvd_disk *ndisk = arg;
253 mtx_lock(&ndisk->bioqlock);
254 bp = bioq_takefirst(&ndisk->bioq);
255 mtx_unlock(&ndisk->bioqlock);
259 if (nvd_bio_submit(ndisk, bp) != 0) {
265 * BIO_ORDERED flag dictates that the bio with BIO_ORDERED
266 * flag set must be completed before proceeding with
269 if (bp->bio_flags & BIO_ORDERED) {
270 while (ndisk->cur_depth > 0) {
271 pause("nvd flush", 1);
279 nvd_new_controller(struct nvme_controller *ctrlr)
281 struct nvd_controller *nvd_ctrlr;
283 nvd_ctrlr = malloc(sizeof(struct nvd_controller), M_NVD,
286 TAILQ_INIT(&nvd_ctrlr->disk_head);
287 TAILQ_INSERT_TAIL(&ctrlr_head, nvd_ctrlr, tailq);
293 nvd_new_disk(struct nvme_namespace *ns, void *ctrlr_arg)
295 uint8_t descr[NVME_MODEL_NUMBER_LENGTH+1];
296 struct nvd_disk *ndisk;
298 struct nvd_controller *ctrlr = ctrlr_arg;
300 ndisk = malloc(sizeof(struct nvd_disk), M_NVD, M_ZERO | M_WAITOK);
303 disk->d_strategy = nvd_strategy;
304 disk->d_ioctl = nvd_ioctl;
305 disk->d_name = NVD_STR;
306 disk->d_drv1 = ndisk;
308 disk->d_maxsize = nvme_ns_get_max_io_xfer_size(ns);
309 disk->d_sectorsize = nvme_ns_get_sector_size(ns);
310 disk->d_mediasize = (off_t)nvme_ns_get_size(ns);
311 disk->d_delmaxsize = (off_t)nvme_ns_get_size(ns);
312 if (disk->d_delmaxsize > nvd_delete_max)
313 disk->d_delmaxsize = nvd_delete_max;
314 disk->d_stripesize = nvme_ns_get_optimal_sector_size(ns);
316 if (TAILQ_EMPTY(&disk_head))
320 TAILQ_LAST(&disk_head, disk_list)->disk->d_unit + 1;
322 disk->d_flags = DISKFLAG_DIRECT_COMPLETION;
324 if (nvme_ns_get_flags(ns) & NVME_NS_DEALLOCATE_SUPPORTED)
325 disk->d_flags |= DISKFLAG_CANDELETE;
327 if (nvme_ns_get_flags(ns) & NVME_NS_FLUSH_SUPPORTED)
328 disk->d_flags |= DISKFLAG_CANFLUSHCACHE;
330 /* ifdef used here to ease porting to stable branches at a later point. */
331 #ifdef DISKFLAG_UNMAPPED_BIO
332 disk->d_flags |= DISKFLAG_UNMAPPED_BIO;
336 * d_ident and d_descr are both far bigger than the length of either
337 * the serial or model number strings.
339 nvme_strvis(disk->d_ident, nvme_ns_get_serial_number(ns),
340 sizeof(disk->d_ident), NVME_SERIAL_NUMBER_LENGTH);
342 nvme_strvis(descr, nvme_ns_get_model_number(ns), sizeof(descr),
343 NVME_MODEL_NUMBER_LENGTH);
345 #if __FreeBSD_version >= 900034
346 strlcpy(disk->d_descr, descr, sizeof(descr));
351 ndisk->cur_depth = 0;
352 ndisk->ordered_in_flight = 0;
354 mtx_init(&ndisk->bioqlock, "NVD bioq lock", NULL, MTX_DEF);
355 bioq_init(&ndisk->bioq);
357 TASK_INIT(&ndisk->bioqtask, 0, nvd_bioq_process, ndisk);
358 ndisk->tq = taskqueue_create("nvd_taskq", M_WAITOK,
359 taskqueue_thread_enqueue, &ndisk->tq);
360 taskqueue_start_threads(&ndisk->tq, 1, PI_DISK, "nvd taskq");
362 TAILQ_INSERT_TAIL(&disk_head, ndisk, global_tailq);
363 TAILQ_INSERT_TAIL(&ctrlr->disk_head, ndisk, ctrlr_tailq);
365 disk_create(disk, DISK_VERSION);
367 printf(NVD_STR"%u: <%s> NVMe namespace\n", disk->d_unit, descr);
368 printf(NVD_STR"%u: %juMB (%ju %u byte sectors)\n", disk->d_unit,
369 (uintmax_t)disk->d_mediasize / (1024*1024),
370 (uintmax_t)disk->d_mediasize / disk->d_sectorsize,
377 destroy_geom_disk(struct nvd_disk *ndisk)
386 taskqueue_free(ndisk->tq);
388 disk_destroy(ndisk->disk);
390 mtx_lock(&ndisk->bioqlock);
392 bp = bioq_takefirst(&ndisk->bioq);
396 bp->bio_flags |= BIO_ERROR;
397 bp->bio_resid = bp->bio_bcount;
402 printf(NVD_STR"%u: lost device - %d outstanding\n", unit, cnt);
403 printf(NVD_STR"%u: removing device entry\n", unit);
405 mtx_unlock(&ndisk->bioqlock);
407 mtx_destroy(&ndisk->bioqlock);
411 nvd_controller_fail(void *ctrlr_arg)
413 struct nvd_controller *ctrlr = ctrlr_arg;
414 struct nvd_disk *disk;
416 while (!TAILQ_EMPTY(&ctrlr->disk_head)) {
417 disk = TAILQ_FIRST(&ctrlr->disk_head);
418 TAILQ_REMOVE(&disk_head, disk, global_tailq);
419 TAILQ_REMOVE(&ctrlr->disk_head, disk, ctrlr_tailq);
420 destroy_geom_disk(disk);
424 TAILQ_REMOVE(&ctrlr_head, ctrlr, tailq);