]> CyberLeo.Net >> Repos - FreeBSD/releng/10.0.git/blob - sys/dev/drm2/ttm/ttm_tt.c
- Copy stable/10 (r259064) to releng/10.0 as part of the
[FreeBSD/releng/10.0.git] / sys / dev / drm2 / ttm / ttm_tt.c
1 /**************************************************************************
2  *
3  * Copyright (c) 2006-2009 VMware, Inc., Palo Alto, CA., USA
4  * All Rights Reserved.
5  *
6  * Permission is hereby granted, free of charge, to any person obtaining a
7  * copy of this software and associated documentation files (the
8  * "Software"), to deal in the Software without restriction, including
9  * without limitation the rights to use, copy, modify, merge, publish,
10  * distribute, sub license, and/or sell copies of the Software, and to
11  * permit persons to whom the Software is furnished to do so, subject to
12  * the following conditions:
13  *
14  * The above copyright notice and this permission notice (including the
15  * next paragraph) shall be included in all copies or substantial portions
16  * of the Software.
17  *
18  * THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
19  * IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
20  * FITNESS FOR A PARTICULAR PURPOSE AND NON-INFRINGEMENT. IN NO EVENT SHALL
21  * THE COPYRIGHT HOLDERS, AUTHORS AND/OR ITS SUPPLIERS BE LIABLE FOR ANY CLAIM,
22  * DAMAGES OR OTHER LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR
23  * OTHERWISE, ARISING FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE
24  * USE OR OTHER DEALINGS IN THE SOFTWARE.
25  *
26  **************************************************************************/
27 /*
28  * Authors: Thomas Hellstrom <thellstrom-at-vmware-dot-com>
29  */
30 /*
31  * Copyright (c) 2013 The FreeBSD Foundation
32  * All rights reserved.
33  *
34  * Portions of this software were developed by Konstantin Belousov
35  * <kib@FreeBSD.org> under sponsorship from the FreeBSD Foundation.
36  */
37
38 #include <sys/cdefs.h>
39 __FBSDID("$FreeBSD$");
40
41 #include <dev/drm2/drmP.h>
42 #include <dev/drm2/ttm/ttm_module.h>
43 #include <dev/drm2/ttm/ttm_bo_driver.h>
44 #include <dev/drm2/ttm/ttm_placement.h>
45 #include <dev/drm2/ttm/ttm_page_alloc.h>
46
47 MALLOC_DEFINE(M_TTM_PD, "ttm_pd", "TTM Page Directories");
48
49 /**
50  * Allocates storage for pointers to the pages that back the ttm.
51  */
52 static void ttm_tt_alloc_page_directory(struct ttm_tt *ttm)
53 {
54         ttm->pages = malloc(ttm->num_pages * sizeof(void *),
55             M_TTM_PD, M_WAITOK | M_ZERO);
56 }
57
58 static void ttm_dma_tt_alloc_page_directory(struct ttm_dma_tt *ttm)
59 {
60         ttm->ttm.pages = malloc(ttm->ttm.num_pages * sizeof(void *),
61             M_TTM_PD, M_WAITOK | M_ZERO);
62         ttm->dma_address = malloc(ttm->ttm.num_pages *
63             sizeof(*ttm->dma_address), M_TTM_PD, M_WAITOK);
64 }
65
66 #if defined(__i386__) || defined(__amd64__)
67 static inline int ttm_tt_set_page_caching(vm_page_t p,
68                                           enum ttm_caching_state c_old,
69                                           enum ttm_caching_state c_new)
70 {
71
72         /* XXXKIB our VM does not need this. */
73 #if 0
74         if (c_old != tt_cached) {
75                 /* p isn't in the default caching state, set it to
76                  * writeback first to free its current memtype. */
77                 pmap_page_set_memattr(p, VM_MEMATTR_WRITE_BACK);
78         }
79 #endif
80
81         if (c_new == tt_wc)
82                 pmap_page_set_memattr(p, VM_MEMATTR_WRITE_COMBINING);
83         else if (c_new == tt_uncached)
84                 pmap_page_set_memattr(p, VM_MEMATTR_UNCACHEABLE);
85
86         return (0);
87 }
88 #else
89 static inline int ttm_tt_set_page_caching(vm_page_t p,
90                                           enum ttm_caching_state c_old,
91                                           enum ttm_caching_state c_new)
92 {
93         return 0;
94 }
95 #endif
96
97 /*
98  * Change caching policy for the linear kernel map
99  * for range of pages in a ttm.
100  */
101
102 static int ttm_tt_set_caching(struct ttm_tt *ttm,
103                               enum ttm_caching_state c_state)
104 {
105         int i, j;
106         vm_page_t cur_page;
107         int ret;
108
109         if (ttm->caching_state == c_state)
110                 return 0;
111
112         if (ttm->state == tt_unpopulated) {
113                 /* Change caching but don't populate */
114                 ttm->caching_state = c_state;
115                 return 0;
116         }
117
118         if (ttm->caching_state == tt_cached)
119                 drm_clflush_pages(ttm->pages, ttm->num_pages);
120
121         for (i = 0; i < ttm->num_pages; ++i) {
122                 cur_page = ttm->pages[i];
123                 if (likely(cur_page != NULL)) {
124                         ret = ttm_tt_set_page_caching(cur_page,
125                                                       ttm->caching_state,
126                                                       c_state);
127                         if (unlikely(ret != 0))
128                                 goto out_err;
129                 }
130         }
131
132         ttm->caching_state = c_state;
133
134         return 0;
135
136 out_err:
137         for (j = 0; j < i; ++j) {
138                 cur_page = ttm->pages[j];
139                 if (cur_page != NULL) {
140                         (void)ttm_tt_set_page_caching(cur_page, c_state,
141                                                       ttm->caching_state);
142                 }
143         }
144
145         return ret;
146 }
147
148 int ttm_tt_set_placement_caching(struct ttm_tt *ttm, uint32_t placement)
149 {
150         enum ttm_caching_state state;
151
152         if (placement & TTM_PL_FLAG_WC)
153                 state = tt_wc;
154         else if (placement & TTM_PL_FLAG_UNCACHED)
155                 state = tt_uncached;
156         else
157                 state = tt_cached;
158
159         return ttm_tt_set_caching(ttm, state);
160 }
161
162 void ttm_tt_destroy(struct ttm_tt *ttm)
163 {
164         if (unlikely(ttm == NULL))
165                 return;
166
167         if (ttm->state == tt_bound) {
168                 ttm_tt_unbind(ttm);
169         }
170
171         if (likely(ttm->pages != NULL)) {
172                 ttm->bdev->driver->ttm_tt_unpopulate(ttm);
173         }
174
175         if (!(ttm->page_flags & TTM_PAGE_FLAG_PERSISTENT_SWAP) &&
176             ttm->swap_storage)
177                 vm_object_deallocate(ttm->swap_storage);
178
179         ttm->swap_storage = NULL;
180         ttm->func->destroy(ttm);
181 }
182
183 int ttm_tt_init(struct ttm_tt *ttm, struct ttm_bo_device *bdev,
184                 unsigned long size, uint32_t page_flags,
185                 vm_page_t dummy_read_page)
186 {
187         ttm->bdev = bdev;
188         ttm->glob = bdev->glob;
189         ttm->num_pages = (size + PAGE_SIZE - 1) >> PAGE_SHIFT;
190         ttm->caching_state = tt_cached;
191         ttm->page_flags = page_flags;
192         ttm->dummy_read_page = dummy_read_page;
193         ttm->state = tt_unpopulated;
194         ttm->swap_storage = NULL;
195
196         ttm_tt_alloc_page_directory(ttm);
197         if (!ttm->pages) {
198                 ttm_tt_destroy(ttm);
199                 printf("Failed allocating page table\n");
200                 return -ENOMEM;
201         }
202         return 0;
203 }
204
205 void ttm_tt_fini(struct ttm_tt *ttm)
206 {
207         free(ttm->pages, M_TTM_PD);
208         ttm->pages = NULL;
209 }
210
211 int ttm_dma_tt_init(struct ttm_dma_tt *ttm_dma, struct ttm_bo_device *bdev,
212                 unsigned long size, uint32_t page_flags,
213                 vm_page_t dummy_read_page)
214 {
215         struct ttm_tt *ttm = &ttm_dma->ttm;
216
217         ttm->bdev = bdev;
218         ttm->glob = bdev->glob;
219         ttm->num_pages = (size + PAGE_SIZE - 1) >> PAGE_SHIFT;
220         ttm->caching_state = tt_cached;
221         ttm->page_flags = page_flags;
222         ttm->dummy_read_page = dummy_read_page;
223         ttm->state = tt_unpopulated;
224         ttm->swap_storage = NULL;
225
226         INIT_LIST_HEAD(&ttm_dma->pages_list);
227         ttm_dma_tt_alloc_page_directory(ttm_dma);
228         if (!ttm->pages || !ttm_dma->dma_address) {
229                 ttm_tt_destroy(ttm);
230                 printf("Failed allocating page table\n");
231                 return -ENOMEM;
232         }
233         return 0;
234 }
235
236 void ttm_dma_tt_fini(struct ttm_dma_tt *ttm_dma)
237 {
238         struct ttm_tt *ttm = &ttm_dma->ttm;
239
240         free(ttm->pages, M_TTM_PD);
241         ttm->pages = NULL;
242         free(ttm_dma->dma_address, M_TTM_PD);
243         ttm_dma->dma_address = NULL;
244 }
245
246 void ttm_tt_unbind(struct ttm_tt *ttm)
247 {
248         int ret;
249
250         if (ttm->state == tt_bound) {
251                 ret = ttm->func->unbind(ttm);
252                 MPASS(ret == 0);
253                 ttm->state = tt_unbound;
254         }
255 }
256
257 int ttm_tt_bind(struct ttm_tt *ttm, struct ttm_mem_reg *bo_mem)
258 {
259         int ret = 0;
260
261         if (!ttm)
262                 return -EINVAL;
263
264         if (ttm->state == tt_bound)
265                 return 0;
266
267         ret = ttm->bdev->driver->ttm_tt_populate(ttm);
268         if (ret)
269                 return ret;
270
271         ret = ttm->func->bind(ttm, bo_mem);
272         if (unlikely(ret != 0))
273                 return ret;
274
275         ttm->state = tt_bound;
276
277         return 0;
278 }
279
280 int ttm_tt_swapin(struct ttm_tt *ttm)
281 {
282         vm_object_t obj;
283         vm_page_t from_page, to_page;
284         int i, ret, rv;
285
286         obj = ttm->swap_storage;
287
288         VM_OBJECT_WLOCK(obj);
289         vm_object_pip_add(obj, 1);
290         for (i = 0; i < ttm->num_pages; ++i) {
291                 from_page = vm_page_grab(obj, i, VM_ALLOC_NORMAL);
292                 if (from_page->valid != VM_PAGE_BITS_ALL) {
293                         if (vm_pager_has_page(obj, i, NULL, NULL)) {
294                                 rv = vm_pager_get_pages(obj, &from_page, 1, 0);
295                                 if (rv != VM_PAGER_OK) {
296                                         vm_page_lock(from_page);
297                                         vm_page_free(from_page);
298                                         vm_page_unlock(from_page);
299                                         ret = -EIO;
300                                         goto err_ret;
301                                 }
302                         } else
303                                 vm_page_zero_invalid(from_page, TRUE);
304                 }
305                 vm_page_xunbusy(from_page);
306                 to_page = ttm->pages[i];
307                 if (unlikely(to_page == NULL)) {
308                         ret = -ENOMEM;
309                         goto err_ret;
310                 }
311                 pmap_copy_page(from_page, to_page);
312         }
313         vm_object_pip_wakeup(obj);
314         VM_OBJECT_WUNLOCK(obj);
315
316         if (!(ttm->page_flags & TTM_PAGE_FLAG_PERSISTENT_SWAP))
317                 vm_object_deallocate(obj);
318         ttm->swap_storage = NULL;
319         ttm->page_flags &= ~TTM_PAGE_FLAG_SWAPPED;
320         return (0);
321
322 err_ret:
323         vm_object_pip_wakeup(obj);
324         VM_OBJECT_WUNLOCK(obj);
325         return (ret);
326 }
327
328 int ttm_tt_swapout(struct ttm_tt *ttm, vm_object_t persistent_swap_storage)
329 {
330         vm_object_t obj;
331         vm_page_t from_page, to_page;
332         int i;
333
334         MPASS(ttm->state == tt_unbound || ttm->state == tt_unpopulated);
335         MPASS(ttm->caching_state == tt_cached);
336
337         if (persistent_swap_storage == NULL) {
338                 obj = vm_pager_allocate(OBJT_SWAP, NULL,
339                     IDX_TO_OFF(ttm->num_pages), VM_PROT_DEFAULT, 0,
340                     curthread->td_ucred);
341                 if (obj == NULL) {
342                         printf("[TTM] Failed allocating swap storage\n");
343                         return (-ENOMEM);
344                 }
345         } else
346                 obj = persistent_swap_storage;
347
348         VM_OBJECT_WLOCK(obj);
349         vm_object_pip_add(obj, 1);
350         for (i = 0; i < ttm->num_pages; ++i) {
351                 from_page = ttm->pages[i];
352                 if (unlikely(from_page == NULL))
353                         continue;
354                 to_page = vm_page_grab(obj, i, VM_ALLOC_NORMAL);
355                 pmap_copy_page(from_page, to_page);
356                 to_page->valid = VM_PAGE_BITS_ALL;
357                 vm_page_dirty(to_page);
358                 vm_page_xunbusy(to_page);
359         }
360         vm_object_pip_wakeup(obj);
361         VM_OBJECT_WUNLOCK(obj);
362
363         ttm->bdev->driver->ttm_tt_unpopulate(ttm);
364         ttm->swap_storage = obj;
365         ttm->page_flags |= TTM_PAGE_FLAG_SWAPPED;
366         if (persistent_swap_storage != NULL)
367                 ttm->page_flags |= TTM_PAGE_FLAG_PERSISTENT_SWAP;
368         return (0);
369 }