3 * Copyright (c) 2008 Hans Petter Selasky. All rights reserved.
5 * Redistribution and use in source and binary forms, with or without
6 * modification, are permitted provided that the following conditions
8 * 1. Redistributions of source code must retain the above copyright
9 * notice, this list of conditions and the following disclaimer.
10 * 2. Redistributions in binary form must reproduce the above copyright
11 * notice, this list of conditions and the following disclaimer in the
12 * documentation and/or other materials provided with the distribution.
14 * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND
15 * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE
16 * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE
17 * ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE
18 * FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
19 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
20 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
21 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT
22 * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY
23 * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF
27 #ifdef USB_GLOBAL_INCLUDE_FILE
28 #include USB_GLOBAL_INCLUDE_FILE
30 #include <sys/stdint.h>
31 #include <sys/stddef.h>
32 #include <sys/param.h>
33 #include <sys/queue.h>
34 #include <sys/types.h>
35 #include <sys/systm.h>
36 #include <sys/kernel.h>
38 #include <sys/module.h>
40 #include <sys/mutex.h>
41 #include <sys/condvar.h>
42 #include <sys/sysctl.h>
44 #include <sys/unistd.h>
45 #include <sys/callout.h>
46 #include <sys/malloc.h>
49 #include <dev/usb/usb.h>
50 #include <dev/usb/usbdi.h>
51 #include <dev/usb/usbdi_util.h>
53 #define USB_DEBUG_VAR usb_debug
55 #include <dev/usb/usb_core.h>
56 #include <dev/usb/usb_busdma.h>
57 #include <dev/usb/usb_process.h>
58 #include <dev/usb/usb_transfer.h>
59 #include <dev/usb/usb_device.h>
60 #include <dev/usb/usb_util.h>
61 #include <dev/usb/usb_debug.h>
63 #include <dev/usb/usb_controller.h>
64 #include <dev/usb/usb_bus.h>
65 #endif /* USB_GLOBAL_INCLUDE_FILE */
68 static void usb_dma_tag_create(struct usb_dma_tag *, usb_size_t, usb_size_t);
69 static void usb_dma_tag_destroy(struct usb_dma_tag *);
70 static void usb_dma_lock_cb(void *, bus_dma_lock_op_t);
71 static void usb_pc_alloc_mem_cb(void *, bus_dma_segment_t *, int, int);
72 static void usb_pc_load_mem_cb(void *, bus_dma_segment_t *, int, int);
73 static void usb_pc_common_mem_cb(void *, bus_dma_segment_t *, int, int,
77 /*------------------------------------------------------------------------*
78 * usbd_get_page - lookup DMA-able memory for the given offset
80 * NOTE: Only call this function when the "page_cache" structure has
81 * been properly initialized !
82 *------------------------------------------------------------------------*/
84 usbd_get_page(struct usb_page_cache *pc, usb_frlength_t offset,
85 struct usb_page_search *res)
88 struct usb_page *page;
92 /* Case 1 - something has been loaded into DMA */
96 /* Case 1a - Kernel Virtual Address */
98 res->buffer = USB_ADD_BYTES(pc->buffer, offset);
100 offset += pc->page_offset_buf;
102 /* compute destination page */
104 page = pc->page_start;
106 if (pc->ismultiseg) {
108 page += (offset / USB_PAGE_SIZE);
110 offset %= USB_PAGE_SIZE;
112 res->length = USB_PAGE_SIZE - offset;
113 res->physaddr = page->physaddr + offset;
115 res->length = (usb_size_t)-1;
116 res->physaddr = page->physaddr + offset;
120 /* Case 1b - Non Kernel Virtual Address */
122 res->buffer = USB_ADD_BYTES(page->buffer, offset);
127 /* Case 2 - Plain PIO */
129 res->buffer = USB_ADD_BYTES(pc->buffer, offset);
130 res->length = (usb_size_t)-1;
136 /*------------------------------------------------------------------------*
137 * usb_pc_buffer_is_aligned - verify alignment
139 * This function is used to check if a page cache buffer is properly
140 * aligned to reduce the use of bounce buffers in PIO mode.
141 *------------------------------------------------------------------------*/
143 usb_pc_buffer_is_aligned(struct usb_page_cache *pc, usb_frlength_t offset,
144 usb_frlength_t len, usb_frlength_t mask)
146 struct usb_page_search buf_res;
150 usbd_get_page(pc, offset, &buf_res);
152 if (buf_res.length > len)
153 buf_res.length = len;
154 if (USB_P2U(buf_res.buffer) & mask)
156 if (buf_res.length & mask)
159 offset += buf_res.length;
160 len -= buf_res.length;
165 /*------------------------------------------------------------------------*
166 * usbd_copy_in - copy directly to DMA-able memory
167 *------------------------------------------------------------------------*/
169 usbd_copy_in(struct usb_page_cache *cache, usb_frlength_t offset,
170 const void *ptr, usb_frlength_t len)
172 struct usb_page_search buf_res;
176 usbd_get_page(cache, offset, &buf_res);
178 if (buf_res.length > len) {
179 buf_res.length = len;
181 memcpy(buf_res.buffer, ptr, buf_res.length);
183 offset += buf_res.length;
184 len -= buf_res.length;
185 ptr = USB_ADD_BYTES(ptr, buf_res.length);
189 /*------------------------------------------------------------------------*
190 * usbd_copy_in_user - copy directly to DMA-able memory from userland
195 *------------------------------------------------------------------------*/
198 usbd_copy_in_user(struct usb_page_cache *cache, usb_frlength_t offset,
199 const void *ptr, usb_frlength_t len)
201 struct usb_page_search buf_res;
206 usbd_get_page(cache, offset, &buf_res);
208 if (buf_res.length > len) {
209 buf_res.length = len;
211 error = copyin(ptr, buf_res.buffer, buf_res.length);
215 offset += buf_res.length;
216 len -= buf_res.length;
217 ptr = USB_ADD_BYTES(ptr, buf_res.length);
219 return (0); /* success */
223 /*------------------------------------------------------------------------*
224 * usbd_m_copy_in - copy a mbuf chain directly into DMA-able memory
225 *------------------------------------------------------------------------*/
227 struct usb_m_copy_in_arg {
228 struct usb_page_cache *cache;
229 usb_frlength_t dst_offset;
233 usbd_m_copy_in_cb(void *arg, void *src, uint32_t count)
235 struct usb_m_copy_in_arg *ua = arg;
237 usbd_copy_in(ua->cache, ua->dst_offset, src, count);
238 ua->dst_offset += count;
243 usbd_m_copy_in(struct usb_page_cache *cache, usb_frlength_t dst_offset,
244 struct mbuf *m, usb_size_t src_offset, usb_frlength_t src_len)
246 struct usb_m_copy_in_arg arg = {cache, dst_offset};
247 (void) m_apply(m, src_offset, src_len, &usbd_m_copy_in_cb, &arg);
251 /*------------------------------------------------------------------------*
252 * usb_uiomove - factored out code
253 *------------------------------------------------------------------------*/
256 usb_uiomove(struct usb_page_cache *pc, struct uio *uio,
257 usb_frlength_t pc_offset, usb_frlength_t len)
259 struct usb_page_search res;
264 usbd_get_page(pc, pc_offset, &res);
266 if (res.length > len) {
270 * "uiomove()" can sleep so one needs to make a wrapper,
271 * exiting the mutex and checking things
273 error = uiomove(res.buffer, res.length, uio);
278 pc_offset += res.length;
285 /*------------------------------------------------------------------------*
286 * usbd_copy_out - copy directly from DMA-able memory
287 *------------------------------------------------------------------------*/
289 usbd_copy_out(struct usb_page_cache *cache, usb_frlength_t offset,
290 void *ptr, usb_frlength_t len)
292 struct usb_page_search res;
296 usbd_get_page(cache, offset, &res);
298 if (res.length > len) {
301 memcpy(ptr, res.buffer, res.length);
303 offset += res.length;
305 ptr = USB_ADD_BYTES(ptr, res.length);
309 /*------------------------------------------------------------------------*
310 * usbd_copy_out_user - copy directly from DMA-able memory to userland
315 *------------------------------------------------------------------------*/
318 usbd_copy_out_user(struct usb_page_cache *cache, usb_frlength_t offset,
319 void *ptr, usb_frlength_t len)
321 struct usb_page_search res;
326 usbd_get_page(cache, offset, &res);
328 if (res.length > len) {
331 error = copyout(res.buffer, ptr, res.length);
335 offset += res.length;
337 ptr = USB_ADD_BYTES(ptr, res.length);
339 return (0); /* success */
343 /*------------------------------------------------------------------------*
344 * usbd_frame_zero - zero DMA-able memory
345 *------------------------------------------------------------------------*/
347 usbd_frame_zero(struct usb_page_cache *cache, usb_frlength_t offset,
350 struct usb_page_search res;
354 usbd_get_page(cache, offset, &res);
356 if (res.length > len) {
359 memset(res.buffer, 0, res.length);
361 offset += res.length;
368 /*------------------------------------------------------------------------*
369 * usb_dma_lock_cb - dummy callback
370 *------------------------------------------------------------------------*/
372 usb_dma_lock_cb(void *arg, bus_dma_lock_op_t op)
374 /* we use "mtx_owned()" instead of this function */
377 /*------------------------------------------------------------------------*
378 * usb_dma_tag_create - allocate a DMA tag
380 * NOTE: If the "align" parameter has a value of 1 the DMA-tag will
381 * allow multi-segment mappings. Else all mappings are single-segment.
382 *------------------------------------------------------------------------*/
384 usb_dma_tag_create(struct usb_dma_tag *udt,
385 usb_size_t size, usb_size_t align)
389 if (bus_dma_tag_create
390 ( /* parent */ udt->tag_parent->tag,
391 /* alignment */ align,
393 /* lowaddr */ (2ULL << (udt->tag_parent->dma_bits - 1)) - 1,
394 /* highaddr */ BUS_SPACE_MAXADDR,
396 /* filterarg */ NULL,
398 /* nsegments */ (align == 1 && size > 1) ?
399 (2 + (size / USB_PAGE_SIZE)) : 1,
400 /* maxsegsz */ (align == 1 && size > USB_PAGE_SIZE) ?
401 USB_PAGE_SIZE : size,
402 /* flags */ BUS_DMA_KEEP_PG_OFFSET,
403 /* lockfn */ &usb_dma_lock_cb,
411 /*------------------------------------------------------------------------*
412 * usb_dma_tag_free - free a DMA tag
413 *------------------------------------------------------------------------*/
415 usb_dma_tag_destroy(struct usb_dma_tag *udt)
417 bus_dma_tag_destroy(udt->tag);
420 /*------------------------------------------------------------------------*
421 * usb_pc_alloc_mem_cb - BUS-DMA callback function
422 *------------------------------------------------------------------------*/
424 usb_pc_alloc_mem_cb(void *arg, bus_dma_segment_t *segs,
427 usb_pc_common_mem_cb(arg, segs, nseg, error, 0);
430 /*------------------------------------------------------------------------*
431 * usb_pc_load_mem_cb - BUS-DMA callback function
432 *------------------------------------------------------------------------*/
434 usb_pc_load_mem_cb(void *arg, bus_dma_segment_t *segs,
437 usb_pc_common_mem_cb(arg, segs, nseg, error, 1);
440 /*------------------------------------------------------------------------*
441 * usb_pc_common_mem_cb - BUS-DMA callback function
442 *------------------------------------------------------------------------*/
444 usb_pc_common_mem_cb(void *arg, bus_dma_segment_t *segs,
445 int nseg, int error, uint8_t isload)
447 struct usb_dma_parent_tag *uptag;
448 struct usb_page_cache *pc;
455 uptag = pc->tag_parent;
458 * XXX There is sometimes recursive locking here.
459 * XXX We should try to find a better solution.
460 * XXX Until further the "owned" variable does
470 pg->physaddr = rounddown2(segs->ds_addr, USB_PAGE_SIZE);
471 rem = segs->ds_addr & (USB_PAGE_SIZE - 1);
472 pc->page_offset_buf = rem;
473 pc->page_offset_end += rem;
478 for (x = 0; x != nseg - 1; x++) {
479 if (((segs[x].ds_addr + segs[x].ds_len) & (USB_PAGE_SIZE - 1)) ==
480 ((segs[x + 1].ds_addr & (USB_PAGE_SIZE - 1))))
483 * This check verifies there is no page offset
484 * hole between any of the segments. See the
485 * BUS_DMA_KEEP_PG_OFFSET flag.
487 DPRINTFN(0, "Page offset was not preserved\n");
493 while (pc->ismultiseg) {
494 off += USB_PAGE_SIZE;
495 if (off >= (segs->ds_len + rem)) {
505 pg->physaddr = rounddown2(segs->ds_addr + off, USB_PAGE_SIZE);
509 owned = mtx_owned(uptag->mtx);
511 USB_MTX_LOCK(uptag->mtx);
513 uptag->dma_error = (error ? 1 : 0);
515 (uptag->func) (uptag);
517 cv_broadcast(uptag->cv);
520 USB_MTX_UNLOCK(uptag->mtx);
523 /*------------------------------------------------------------------------*
524 * usb_pc_alloc_mem - allocate DMA'able memory
529 *------------------------------------------------------------------------*/
531 usb_pc_alloc_mem(struct usb_page_cache *pc, struct usb_page *pg,
532 usb_size_t size, usb_size_t align)
534 struct usb_dma_parent_tag *uptag;
535 struct usb_dma_tag *utag;
540 uptag = pc->tag_parent;
544 * The alignment must be greater or equal to the
545 * "size" else the object can be split between two
546 * memory pages and we get a problem!
548 while (align < size) {
556 * XXX BUS-DMA workaround - FIXME later:
558 * We assume that that the aligment at this point of
559 * the code is greater than or equal to the size and
560 * less than two times the size, so that if we double
561 * the size, the size will be greater than the
564 * The bus-dma system has a check for "alignment"
565 * being less than "size". If that check fails we end
566 * up using contigmalloc which is page based even for
567 * small allocations. Try to avoid that to save
568 * memory, hence we sometimes to a large number of
571 if (size <= (USB_PAGE_SIZE / 2)) {
576 /* get the correct DMA tag */
577 utag = usb_dma_tag_find(uptag, size, align);
581 /* allocate memory */
582 if (bus_dmamem_alloc(
583 utag->tag, &ptr, (BUS_DMA_WAITOK | BUS_DMA_COHERENT), &map)) {
586 /* setup page cache */
589 pc->page_offset_buf = 0;
590 pc->page_offset_end = size;
593 pc->ismultiseg = (align == 1);
595 USB_MTX_LOCK(uptag->mtx);
597 /* load memory into DMA */
598 err = bus_dmamap_load(
599 utag->tag, map, ptr, size, &usb_pc_alloc_mem_cb,
600 pc, (BUS_DMA_WAITOK | BUS_DMA_COHERENT));
602 if (err == EINPROGRESS) {
603 cv_wait(uptag->cv, uptag->mtx);
606 USB_MTX_UNLOCK(uptag->mtx);
608 if (err || uptag->dma_error) {
609 bus_dmamem_free(utag->tag, ptr, map);
612 memset(ptr, 0, size);
614 usb_pc_cpu_flush(pc);
619 /* reset most of the page cache */
621 pc->page_start = NULL;
622 pc->page_offset_buf = 0;
623 pc->page_offset_end = 0;
629 /*------------------------------------------------------------------------*
630 * usb_pc_free_mem - free DMA memory
632 * This function is NULL safe.
633 *------------------------------------------------------------------------*/
635 usb_pc_free_mem(struct usb_page_cache *pc)
637 if (pc && pc->buffer) {
639 bus_dmamap_unload(pc->tag, pc->map);
641 bus_dmamem_free(pc->tag, pc->buffer, pc->map);
647 /*------------------------------------------------------------------------*
648 * usb_pc_load_mem - load virtual memory into DMA
653 *------------------------------------------------------------------------*/
655 usb_pc_load_mem(struct usb_page_cache *pc, usb_size_t size, uint8_t sync)
657 /* setup page cache */
658 pc->page_offset_buf = 0;
659 pc->page_offset_end = size;
662 USB_MTX_ASSERT(pc->tag_parent->mtx, MA_OWNED);
666 struct usb_dma_parent_tag *uptag;
669 uptag = pc->tag_parent;
672 * We have to unload the previous loaded DMA
673 * pages before trying to load a new one!
675 bus_dmamap_unload(pc->tag, pc->map);
678 * Try to load memory into DMA.
680 err = bus_dmamap_load(
681 pc->tag, pc->map, pc->buffer, size,
682 &usb_pc_alloc_mem_cb, pc, BUS_DMA_WAITOK);
683 if (err == EINPROGRESS) {
684 cv_wait(uptag->cv, uptag->mtx);
687 if (err || uptag->dma_error) {
693 * We have to unload the previous loaded DMA
694 * pages before trying to load a new one!
696 bus_dmamap_unload(pc->tag, pc->map);
699 * Try to load memory into DMA. The callback
700 * will be called in all cases:
703 pc->tag, pc->map, pc->buffer, size,
704 &usb_pc_load_mem_cb, pc, BUS_DMA_WAITOK)) {
710 * Call callback so that refcount is decremented
713 pc->tag_parent->dma_error = 0;
714 (pc->tag_parent->func) (pc->tag_parent);
720 /*------------------------------------------------------------------------*
721 * usb_pc_cpu_invalidate - invalidate CPU cache
722 *------------------------------------------------------------------------*/
724 usb_pc_cpu_invalidate(struct usb_page_cache *pc)
726 if (pc->page_offset_end == pc->page_offset_buf) {
727 /* nothing has been loaded into this page cache! */
732 * TODO: We currently do XXX_POSTREAD and XXX_PREREAD at the
733 * same time, but in the future we should try to isolate the
734 * different cases to optimise the code. --HPS
736 bus_dmamap_sync(pc->tag, pc->map, BUS_DMASYNC_POSTREAD);
737 bus_dmamap_sync(pc->tag, pc->map, BUS_DMASYNC_PREREAD);
740 /*------------------------------------------------------------------------*
741 * usb_pc_cpu_flush - flush CPU cache
742 *------------------------------------------------------------------------*/
744 usb_pc_cpu_flush(struct usb_page_cache *pc)
746 if (pc->page_offset_end == pc->page_offset_buf) {
747 /* nothing has been loaded into this page cache! */
750 bus_dmamap_sync(pc->tag, pc->map, BUS_DMASYNC_PREWRITE);
753 /*------------------------------------------------------------------------*
754 * usb_pc_dmamap_create - create a DMA map
759 *------------------------------------------------------------------------*/
761 usb_pc_dmamap_create(struct usb_page_cache *pc, usb_size_t size)
763 struct usb_xfer_root *info;
764 struct usb_dma_tag *utag;
767 info = USB_DMATAG_TO_XROOT(pc->tag_parent);
773 utag = usb_dma_tag_find(pc->tag_parent, size, 1);
778 if (bus_dmamap_create(utag->tag, 0, &pc->map)) {
782 return 0; /* success */
787 return 1; /* failure */
790 /*------------------------------------------------------------------------*
791 * usb_pc_dmamap_destroy
793 * This function is NULL safe.
794 *------------------------------------------------------------------------*/
796 usb_pc_dmamap_destroy(struct usb_page_cache *pc)
799 bus_dmamap_destroy(pc->tag, pc->map);
805 /*------------------------------------------------------------------------*
806 * usb_dma_tag_find - factored out code
807 *------------------------------------------------------------------------*/
809 usb_dma_tag_find(struct usb_dma_parent_tag *udpt,
810 usb_size_t size, usb_size_t align)
812 struct usb_dma_tag *udt;
815 USB_ASSERT(align > 0, ("Invalid parameter align = 0\n"));
816 USB_ASSERT(size > 0, ("Invalid parameter size = 0\n"));
818 udt = udpt->utag_first;
819 nudt = udpt->utag_max;
823 if (udt->align == 0) {
824 usb_dma_tag_create(udt, size, align);
825 if (udt->tag == NULL) {
832 if ((udt->align == align) && (udt->size == size)) {
840 /*------------------------------------------------------------------------*
841 * usb_dma_tag_setup - initialise USB DMA tags
842 *------------------------------------------------------------------------*/
844 usb_dma_tag_setup(struct usb_dma_parent_tag *udpt,
845 struct usb_dma_tag *udt, bus_dma_tag_t dmat,
846 struct mtx *mtx, usb_dma_callback_t *func,
847 uint8_t ndmabits, uint8_t nudt)
849 memset(udpt, 0, sizeof(*udpt));
851 /* sanity checking */
855 /* something is corrupt */
858 /* initialise condition variable */
859 cv_init(udpt->cv, "USB DMA CV");
861 /* store some information */
865 udpt->utag_first = udt;
866 udpt->utag_max = nudt;
867 udpt->dma_bits = ndmabits;
870 memset(udt, 0, sizeof(*udt));
871 udt->tag_parent = udpt;
876 /*------------------------------------------------------------------------*
877 * usb_bus_tag_unsetup - factored out code
878 *------------------------------------------------------------------------*/
880 usb_dma_tag_unsetup(struct usb_dma_parent_tag *udpt)
882 struct usb_dma_tag *udt;
885 udt = udpt->utag_first;
886 nudt = udpt->utag_max;
891 /* destroy the USB DMA tag */
892 usb_dma_tag_destroy(udt);
898 if (udpt->utag_max) {
899 /* destroy the condition variable */
900 cv_destroy(udpt->cv);
904 /*------------------------------------------------------------------------*
907 * This function handles loading of virtual buffers into DMA and is
908 * only called when "dma_refcount" is zero.
909 *------------------------------------------------------------------------*/
911 usb_bdma_work_loop(struct usb_xfer_queue *pq)
913 struct usb_xfer_root *info;
914 struct usb_xfer *xfer;
915 usb_frcount_t nframes;
920 USB_MTX_ASSERT(info->xfer_mtx, MA_OWNED);
923 /* some error happened */
924 USB_BUS_LOCK(info->bus);
925 usbd_transfer_done(xfer, 0);
926 USB_BUS_UNLOCK(info->bus);
929 if (!xfer->flags_int.bdma_setup) {
931 usb_frlength_t frlength_0;
934 xfer->flags_int.bdma_setup = 1;
936 /* reset BUS-DMA load state */
940 if (xfer->flags_int.isochronous_xfr) {
941 /* only one frame buffer */
943 frlength_0 = xfer->sumlen;
945 /* can be multiple frame buffers */
946 nframes = xfer->nframes;
947 frlength_0 = xfer->frlengths[0];
951 * Set DMA direction first. This is needed to
952 * select the correct cache invalidate and cache
955 isread = USB_GET_DATA_ISREAD(xfer);
956 pg = xfer->dma_page_ptr;
958 if (xfer->flags_int.control_xfr &&
959 xfer->flags_int.control_hdr) {
961 if (xfer->flags_int.usb_mode == USB_MODE_DEVICE) {
962 /* The device controller writes to memory */
963 xfer->frbuffers[0].isread = 1;
965 /* The host controller reads from memory */
966 xfer->frbuffers[0].isread = 0;
970 xfer->frbuffers[0].isread = isread;
974 * Setup the "page_start" pointer which points to an array of
975 * USB pages where information about the physical address of a
976 * page will be stored. Also initialise the "isread" field of
977 * the USB page caches.
979 xfer->frbuffers[0].page_start = pg;
981 info->dma_nframes = nframes;
982 info->dma_currframe = 0;
983 info->dma_frlength_0 = frlength_0;
985 pg += (frlength_0 / USB_PAGE_SIZE);
988 while (--nframes > 0) {
989 xfer->frbuffers[nframes].isread = isread;
990 xfer->frbuffers[nframes].page_start = pg;
992 pg += (xfer->frlengths[nframes] / USB_PAGE_SIZE);
997 if (info->dma_error) {
998 USB_BUS_LOCK(info->bus);
999 usbd_transfer_done(xfer, USB_ERR_DMA_LOAD_FAILED);
1000 USB_BUS_UNLOCK(info->bus);
1003 if (info->dma_currframe != info->dma_nframes) {
1005 if (info->dma_currframe == 0) {
1007 usb_pc_load_mem(xfer->frbuffers,
1008 info->dma_frlength_0, 0);
1011 nframes = info->dma_currframe;
1012 usb_pc_load_mem(xfer->frbuffers + nframes,
1013 xfer->frlengths[nframes], 0);
1016 /* advance frame index */
1017 info->dma_currframe++;
1022 usb_bdma_pre_sync(xfer);
1024 /* start loading next USB transfer, if any */
1025 usb_command_wrapper(pq, NULL);
1027 /* finally start the hardware */
1028 usbd_pipe_enter(xfer);
1031 /*------------------------------------------------------------------------*
1032 * usb_bdma_done_event
1034 * This function is called when the BUS-DMA has loaded virtual memory
1036 *------------------------------------------------------------------------*/
1038 usb_bdma_done_event(struct usb_dma_parent_tag *udpt)
1040 struct usb_xfer_root *info;
1042 info = USB_DMATAG_TO_XROOT(udpt);
1044 USB_MTX_ASSERT(info->xfer_mtx, MA_OWNED);
1047 info->dma_error = udpt->dma_error;
1049 /* enter workloop again */
1050 usb_command_wrapper(&info->dma_q,
1054 /*------------------------------------------------------------------------*
1057 * This function handles DMA synchronisation that must be done before
1058 * an USB transfer is started.
1059 *------------------------------------------------------------------------*/
1061 usb_bdma_pre_sync(struct usb_xfer *xfer)
1063 struct usb_page_cache *pc;
1064 usb_frcount_t nframes;
1066 if (xfer->flags_int.isochronous_xfr) {
1067 /* only one frame buffer */
1070 /* can be multiple frame buffers */
1071 nframes = xfer->nframes;
1074 pc = xfer->frbuffers;
1079 usb_pc_cpu_invalidate(pc);
1081 usb_pc_cpu_flush(pc);
1087 /*------------------------------------------------------------------------*
1088 * usb_bdma_post_sync
1090 * This function handles DMA synchronisation that must be done after
1091 * an USB transfer is complete.
1092 *------------------------------------------------------------------------*/
1094 usb_bdma_post_sync(struct usb_xfer *xfer)
1096 struct usb_page_cache *pc;
1097 usb_frcount_t nframes;
1099 if (xfer->flags_int.isochronous_xfr) {
1100 /* only one frame buffer */
1103 /* can be multiple frame buffers */
1104 nframes = xfer->nframes;
1107 pc = xfer->frbuffers;
1111 usb_pc_cpu_invalidate(pc);