2 * Copyright (c) 2003-2009 Silicon Graphics International Corp.
3 * Copyright (c) 2012 The FreeBSD Foundation
4 * Copyright (c) 2015 Alexander Motin <mav@FreeBSD.org>
7 * Redistribution and use in source and binary forms, with or without
8 * modification, are permitted provided that the following conditions
10 * 1. Redistributions of source code must retain the above copyright
11 * notice, this list of conditions and the following disclaimer,
12 * without modification, immediately at the beginning of the file.
13 * 2. Redistributions in binary form must reproduce the above copyright
14 * notice, this list of conditions and the following disclaimer in the
15 * documentation and/or other materials provided with the distribution.
17 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR ``AS IS'' AND ANY EXPRESS OR
18 * IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES
19 * OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED.
20 * IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR ANY DIRECT, INDIRECT,
21 * INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT
22 * NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE,
23 * DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY
24 * THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT
25 * (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF
26 * THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
29 #include <sys/cdefs.h>
30 __FBSDID("$FreeBSD$");
32 #include <sys/param.h>
33 #include <sys/systm.h>
34 #include <sys/kernel.h>
35 #include <sys/types.h>
37 #include <sys/module.h>
38 #include <sys/mutex.h>
39 #include <sys/condvar.h>
40 #include <sys/malloc.h>
42 #include <sys/queue.h>
43 #include <sys/sysctl.h>
46 #include <cam/scsi/scsi_all.h>
47 #include <cam/scsi/scsi_da.h>
48 #include <cam/ctl/ctl_io.h>
49 #include <cam/ctl/ctl.h>
50 #include <cam/ctl/ctl_frontend.h>
51 #include <cam/ctl/ctl_util.h>
52 #include <cam/ctl/ctl_backend.h>
53 #include <cam/ctl/ctl_ioctl.h>
54 #include <cam/ctl/ctl_ha.h>
55 #include <cam/ctl/ctl_private.h>
56 #include <cam/ctl/ctl_debug.h>
57 #include <cam/ctl/ctl_error.h>
65 struct ctl_fe_ioctl_params {
68 ctl_fe_ioctl_state state;
76 static struct cfi_softc cfi_softc;
78 static int cfi_init(void);
79 static void cfi_shutdown(void);
80 static void cfi_datamove(union ctl_io *io);
81 static void cfi_done(union ctl_io *io);
83 static struct ctl_frontend cfi_frontend =
87 .shutdown = cfi_shutdown,
89 CTL_FRONTEND_DECLARE(ctlioctl, cfi_frontend);
94 struct cfi_softc *isoftc = &cfi_softc;
95 struct ctl_port *port;
97 memset(isoftc, 0, sizeof(*isoftc));
100 port->frontend = &cfi_frontend;
101 port->port_type = CTL_PORT_IOCTL;
102 port->num_requested_ctl_io = 100;
103 port->port_name = "ioctl";
104 port->fe_datamove = cfi_datamove;
105 port->fe_done = cfi_done;
106 port->max_targets = 1;
107 port->max_target_id = 0;
108 port->targ_port = -1;
109 port->max_initiators = 1;
111 if (ctl_port_register(port) != 0) {
112 printf("%s: ioctl port registration failed\n", __func__);
115 ctl_port_online(port);
122 struct cfi_softc *isoftc = &cfi_softc;
123 struct ctl_port *port;
125 port = &isoftc->port;
126 ctl_port_offline(port);
127 if (ctl_port_deregister(&isoftc->port) != 0)
128 printf("%s: ctl_frontend_deregister() failed\n", __func__);
132 * Data movement routine for the CTL ioctl frontend port.
135 ctl_ioctl_do_datamove(struct ctl_scsiio *ctsio)
137 struct ctl_sg_entry *ext_sglist, *kern_sglist;
138 struct ctl_sg_entry ext_entry, kern_entry;
139 int ext_sglen, ext_sg_entries, kern_sg_entries;
140 int ext_sg_start, ext_offset;
141 int len_to_copy, len_copied;
142 int kern_watermark, ext_watermark;
143 int ext_sglist_malloced;
146 ext_sglist_malloced = 0;
150 CTL_DEBUG_PRINT(("ctl_ioctl_do_datamove\n"));
153 * If this flag is set, fake the data transfer.
155 if (ctsio->io_hdr.flags & CTL_FLAG_NO_DATAMOVE) {
156 ctsio->ext_data_filled = ctsio->ext_data_len;
161 * To simplify things here, if we have a single buffer, stick it in
162 * a S/G entry and just make it a single entry S/G list.
164 if (ctsio->ext_sg_entries > 0) {
167 ext_sglen = ctsio->ext_sg_entries * sizeof(*ext_sglist);
169 ext_sglist = (struct ctl_sg_entry *)malloc(ext_sglen, M_CTL,
171 ext_sglist_malloced = 1;
172 if (copyin(ctsio->ext_data_ptr, ext_sglist, ext_sglen) != 0) {
173 ctsio->io_hdr.port_status = 31343;
176 ext_sg_entries = ctsio->ext_sg_entries;
178 for (i = 0; i < ext_sg_entries; i++) {
179 if ((len_seen + ext_sglist[i].len) >=
180 ctsio->ext_data_filled) {
182 ext_offset = ctsio->ext_data_filled - len_seen;
185 len_seen += ext_sglist[i].len;
188 ext_sglist = &ext_entry;
189 ext_sglist->addr = ctsio->ext_data_ptr;
190 ext_sglist->len = ctsio->ext_data_len;
193 ext_offset = ctsio->ext_data_filled;
196 if (ctsio->kern_sg_entries > 0) {
197 kern_sglist = (struct ctl_sg_entry *)ctsio->kern_data_ptr;
198 kern_sg_entries = ctsio->kern_sg_entries;
200 kern_sglist = &kern_entry;
201 kern_sglist->addr = ctsio->kern_data_ptr;
202 kern_sglist->len = ctsio->kern_data_len;
208 ext_watermark = ext_offset;
210 for (i = ext_sg_start, j = 0;
211 i < ext_sg_entries && j < kern_sg_entries;) {
212 uint8_t *ext_ptr, *kern_ptr;
214 len_to_copy = MIN(ext_sglist[i].len - ext_watermark,
215 kern_sglist[j].len - kern_watermark);
217 ext_ptr = (uint8_t *)ext_sglist[i].addr;
218 ext_ptr = ext_ptr + ext_watermark;
219 if (ctsio->io_hdr.flags & CTL_FLAG_BUS_ADDR) {
223 panic("need to implement bus address support");
225 kern_ptr = bus_to_virt(kern_sglist[j].addr);
228 kern_ptr = (uint8_t *)kern_sglist[j].addr;
229 kern_ptr = kern_ptr + kern_watermark;
231 kern_watermark += len_to_copy;
232 ext_watermark += len_to_copy;
234 if ((ctsio->io_hdr.flags & CTL_FLAG_DATA_MASK) ==
236 CTL_DEBUG_PRINT(("ctl_ioctl_do_datamove: copying %d "
237 "bytes to user\n", len_to_copy));
238 CTL_DEBUG_PRINT(("ctl_ioctl_do_datamove: from %p "
239 "to %p\n", kern_ptr, ext_ptr));
240 if (copyout(kern_ptr, ext_ptr, len_to_copy) != 0) {
241 ctsio->io_hdr.port_status = 31344;
245 CTL_DEBUG_PRINT(("ctl_ioctl_do_datamove: copying %d "
246 "bytes from user\n", len_to_copy));
247 CTL_DEBUG_PRINT(("ctl_ioctl_do_datamove: from %p "
248 "to %p\n", ext_ptr, kern_ptr));
249 if (copyin(ext_ptr, kern_ptr, len_to_copy)!= 0){
250 ctsio->io_hdr.port_status = 31345;
255 len_copied += len_to_copy;
257 if (ext_sglist[i].len == ext_watermark) {
262 if (kern_sglist[j].len == kern_watermark) {
268 ctsio->ext_data_filled += len_copied;
270 CTL_DEBUG_PRINT(("ctl_ioctl_do_datamove: ext_sg_entries: %d, "
271 "kern_sg_entries: %d\n", ext_sg_entries,
273 CTL_DEBUG_PRINT(("ctl_ioctl_do_datamove: ext_data_len = %d, "
274 "kern_data_len = %d\n", ctsio->ext_data_len,
275 ctsio->kern_data_len));
278 /* XXX KDM set residual?? */
281 if (ext_sglist_malloced != 0)
282 free(ext_sglist, M_CTL);
284 return (CTL_RETVAL_COMPLETE);
288 cfi_datamove(union ctl_io *io)
290 struct ctl_fe_ioctl_params *params;
292 params = (struct ctl_fe_ioctl_params *)
293 io->io_hdr.ctl_private[CTL_PRIV_FRONTEND].ptr;
295 mtx_lock(¶ms->ioctl_mtx);
296 params->state = CTL_IOCTL_DATAMOVE;
297 cv_broadcast(¶ms->sem);
298 mtx_unlock(¶ms->ioctl_mtx);
302 cfi_done(union ctl_io *io)
304 struct ctl_fe_ioctl_params *params;
306 params = (struct ctl_fe_ioctl_params *)
307 io->io_hdr.ctl_private[CTL_PRIV_FRONTEND].ptr;
309 mtx_lock(¶ms->ioctl_mtx);
310 params->state = CTL_IOCTL_DONE;
311 cv_broadcast(¶ms->sem);
312 mtx_unlock(¶ms->ioctl_mtx);
316 cfi_submit_wait(union ctl_io *io)
318 struct ctl_fe_ioctl_params params;
319 ctl_fe_ioctl_state last_state;
322 bzero(¶ms, sizeof(params));
323 mtx_init(¶ms.ioctl_mtx, "ctliocmtx", NULL, MTX_DEF);
324 cv_init(¶ms.sem, "ctlioccv");
325 params.state = CTL_IOCTL_INPROG;
326 last_state = params.state;
328 io->io_hdr.ctl_private[CTL_PRIV_FRONTEND].ptr = ¶ms;
330 CTL_DEBUG_PRINT(("cfi_submit_wait\n"));
332 /* This shouldn't happen */
333 if ((retval = ctl_queue(io)) != CTL_RETVAL_COMPLETE)
339 mtx_lock(¶ms.ioctl_mtx);
341 * Check the state here, and don't sleep if the state has
342 * already changed (i.e. wakeup has already occurred, but we
343 * weren't waiting yet).
345 if (params.state == last_state) {
346 /* XXX KDM cv_wait_sig instead? */
347 cv_wait(¶ms.sem, ¶ms.ioctl_mtx);
349 last_state = params.state;
351 switch (params.state) {
352 case CTL_IOCTL_INPROG:
353 /* Why did we wake up? */
354 /* XXX KDM error here? */
355 mtx_unlock(¶ms.ioctl_mtx);
357 case CTL_IOCTL_DATAMOVE:
358 CTL_DEBUG_PRINT(("got CTL_IOCTL_DATAMOVE\n"));
361 * change last_state back to INPROG to avoid
362 * deadlock on subsequent data moves.
364 params.state = last_state = CTL_IOCTL_INPROG;
366 mtx_unlock(¶ms.ioctl_mtx);
367 ctl_ioctl_do_datamove(&io->scsiio);
369 * Note that in some cases, most notably writes,
370 * this will queue the I/O and call us back later.
371 * In other cases, generally reads, this routine
372 * will immediately call back and wake us up,
373 * probably using our own context.
375 io->scsiio.be_move_done(io);
378 mtx_unlock(¶ms.ioctl_mtx);
379 CTL_DEBUG_PRINT(("got CTL_IOCTL_DONE\n"));
383 mtx_unlock(¶ms.ioctl_mtx);
384 /* XXX KDM error here? */
389 mtx_destroy(¶ms.ioctl_mtx);
390 cv_destroy(¶ms.sem);
392 return (CTL_RETVAL_COMPLETE);
396 ctl_ioctl_io(struct cdev *dev, u_long cmd, caddr_t addr, int flag,
404 * If we haven't been "enabled", don't allow any SCSI I/O
407 if ((cfi_softc.port.status & CTL_PORT_STATUS_ONLINE) == 0)
410 io = ctl_alloc_io(cfi_softc.port.ctl_pool_ref);
413 * Need to save the pool reference so it doesn't get
414 * spammed by the user's ctl_io.
416 pool_tmp = io->io_hdr.pool;
417 memcpy(io, (void *)addr, sizeof(*io));
418 io->io_hdr.pool = pool_tmp;
421 * No status yet, so make sure the status is set properly.
423 io->io_hdr.status = CTL_STATUS_NONE;
426 * The user sets the initiator ID, target and LUN IDs.
428 io->io_hdr.nexus.targ_port = cfi_softc.port.targ_port;
429 io->io_hdr.flags |= CTL_FLAG_USER_REQ;
430 if ((io->io_hdr.io_type == CTL_IO_SCSI) &&
431 (io->scsiio.tag_type != CTL_TAG_UNTAGGED))
432 io->scsiio.tag_num = cfi_softc.cur_tag_num++;
434 retval = cfi_submit_wait(io);
436 memcpy((void *)addr, io, sizeof(*io));