.. | .. |
---|
25 | 25 | * Alex Deucher |
---|
26 | 26 | * Jerome Glisse |
---|
27 | 27 | */ |
---|
28 | | -#include <drm/drmP.h> |
---|
| 28 | + |
---|
| 29 | +#include <linux/pci.h> |
---|
| 30 | +#include <linux/vmalloc.h> |
---|
| 31 | + |
---|
29 | 32 | #include <drm/amdgpu_drm.h> |
---|
30 | 33 | #ifdef CONFIG_X86 |
---|
31 | 34 | #include <asm/set_memory.h> |
---|
.. | .. |
---|
68 | 71 | */ |
---|
69 | 72 | static int amdgpu_gart_dummy_page_init(struct amdgpu_device *adev) |
---|
70 | 73 | { |
---|
71 | | - struct page *dummy_page = adev->mman.bdev.glob->dummy_read_page; |
---|
| 74 | + struct page *dummy_page = ttm_bo_glob.dummy_read_page; |
---|
72 | 75 | |
---|
73 | 76 | if (adev->dummy_page_addr) |
---|
74 | 77 | return 0; |
---|
.. | .. |
---|
112 | 115 | { |
---|
113 | 116 | int r; |
---|
114 | 117 | |
---|
115 | | - if (adev->gart.robj == NULL) { |
---|
| 118 | + if (adev->gart.bo == NULL) { |
---|
116 | 119 | struct amdgpu_bo_param bp; |
---|
117 | 120 | |
---|
118 | 121 | memset(&bp, 0, sizeof(bp)); |
---|
.. | .. |
---|
123 | 126 | AMDGPU_GEM_CREATE_VRAM_CONTIGUOUS; |
---|
124 | 127 | bp.type = ttm_bo_type_kernel; |
---|
125 | 128 | bp.resv = NULL; |
---|
126 | | - r = amdgpu_bo_create(adev, &bp, &adev->gart.robj); |
---|
| 129 | + r = amdgpu_bo_create(adev, &bp, &adev->gart.bo); |
---|
127 | 130 | if (r) { |
---|
128 | 131 | return r; |
---|
129 | 132 | } |
---|
.. | .. |
---|
145 | 148 | { |
---|
146 | 149 | int r; |
---|
147 | 150 | |
---|
148 | | - r = amdgpu_bo_reserve(adev->gart.robj, false); |
---|
| 151 | + r = amdgpu_bo_reserve(adev->gart.bo, false); |
---|
149 | 152 | if (unlikely(r != 0)) |
---|
150 | 153 | return r; |
---|
151 | | - r = amdgpu_bo_pin(adev->gart.robj, AMDGPU_GEM_DOMAIN_VRAM); |
---|
| 154 | + r = amdgpu_bo_pin(adev->gart.bo, AMDGPU_GEM_DOMAIN_VRAM); |
---|
152 | 155 | if (r) { |
---|
153 | | - amdgpu_bo_unreserve(adev->gart.robj); |
---|
| 156 | + amdgpu_bo_unreserve(adev->gart.bo); |
---|
154 | 157 | return r; |
---|
155 | 158 | } |
---|
156 | | - r = amdgpu_bo_kmap(adev->gart.robj, &adev->gart.ptr); |
---|
| 159 | + r = amdgpu_bo_kmap(adev->gart.bo, &adev->gart.ptr); |
---|
157 | 160 | if (r) |
---|
158 | | - amdgpu_bo_unpin(adev->gart.robj); |
---|
159 | | - amdgpu_bo_unreserve(adev->gart.robj); |
---|
160 | | - adev->gart.table_addr = amdgpu_bo_gpu_offset(adev->gart.robj); |
---|
| 161 | + amdgpu_bo_unpin(adev->gart.bo); |
---|
| 162 | + amdgpu_bo_unreserve(adev->gart.bo); |
---|
161 | 163 | return r; |
---|
162 | 164 | } |
---|
163 | 165 | |
---|
.. | .. |
---|
173 | 175 | { |
---|
174 | 176 | int r; |
---|
175 | 177 | |
---|
176 | | - if (adev->gart.robj == NULL) { |
---|
| 178 | + if (adev->gart.bo == NULL) { |
---|
177 | 179 | return; |
---|
178 | 180 | } |
---|
179 | | - r = amdgpu_bo_reserve(adev->gart.robj, true); |
---|
| 181 | + r = amdgpu_bo_reserve(adev->gart.bo, true); |
---|
180 | 182 | if (likely(r == 0)) { |
---|
181 | | - amdgpu_bo_kunmap(adev->gart.robj); |
---|
182 | | - amdgpu_bo_unpin(adev->gart.robj); |
---|
183 | | - amdgpu_bo_unreserve(adev->gart.robj); |
---|
| 183 | + amdgpu_bo_kunmap(adev->gart.bo); |
---|
| 184 | + amdgpu_bo_unpin(adev->gart.bo); |
---|
| 185 | + amdgpu_bo_unreserve(adev->gart.bo); |
---|
184 | 186 | adev->gart.ptr = NULL; |
---|
185 | 187 | } |
---|
186 | 188 | } |
---|
.. | .. |
---|
196 | 198 | */ |
---|
197 | 199 | void amdgpu_gart_table_vram_free(struct amdgpu_device *adev) |
---|
198 | 200 | { |
---|
199 | | - if (adev->gart.robj == NULL) { |
---|
| 201 | + if (adev->gart.bo == NULL) { |
---|
200 | 202 | return; |
---|
201 | 203 | } |
---|
202 | | - amdgpu_bo_unref(&adev->gart.robj); |
---|
| 204 | + amdgpu_bo_unref(&adev->gart.bo); |
---|
203 | 205 | } |
---|
204 | 206 | |
---|
205 | 207 | /* |
---|
.. | .. |
---|
249 | 251 | } |
---|
250 | 252 | mb(); |
---|
251 | 253 | amdgpu_asic_flush_hdp(adev, NULL); |
---|
252 | | - amdgpu_gmc_flush_gpu_tlb(adev, 0); |
---|
| 254 | + for (i = 0; i < adev->num_vmhubs; i++) |
---|
| 255 | + amdgpu_gmc_flush_gpu_tlb(adev, 0, i, 0); |
---|
| 256 | + |
---|
253 | 257 | return 0; |
---|
254 | 258 | } |
---|
255 | 259 | |
---|
.. | .. |
---|
260 | 264 | * @offset: offset into the GPU's gart aperture |
---|
261 | 265 | * @pages: number of pages to bind |
---|
262 | 266 | * @dma_addr: DMA addresses of pages |
---|
| 267 | + * @flags: page table entry flags |
---|
| 268 | + * @dst: CPU address of the gart table |
---|
263 | 269 | * |
---|
264 | 270 | * Map the dma_addresses into GART entries (all asics). |
---|
265 | 271 | * Returns 0 for success, -EINVAL for failure. |
---|
.. | .. |
---|
296 | 302 | * @pages: number of pages to bind |
---|
297 | 303 | * @pagelist: pages to bind |
---|
298 | 304 | * @dma_addr: DMA addresses of pages |
---|
| 305 | + * @flags: page table entry flags |
---|
299 | 306 | * |
---|
300 | 307 | * Binds the requested pages to the gart page table |
---|
301 | 308 | * (all asics). |
---|
.. | .. |
---|
306 | 313 | uint64_t flags) |
---|
307 | 314 | { |
---|
308 | 315 | #ifdef CONFIG_DRM_AMDGPU_GART_DEBUGFS |
---|
309 | | - unsigned i,t,p; |
---|
| 316 | + unsigned t,p; |
---|
310 | 317 | #endif |
---|
311 | | - int r; |
---|
| 318 | + int r, i; |
---|
312 | 319 | |
---|
313 | 320 | if (!adev->gart.ready) { |
---|
314 | 321 | WARN(1, "trying to bind memory to uninitialized GART !\n"); |
---|
.. | .. |
---|
332 | 339 | |
---|
333 | 340 | mb(); |
---|
334 | 341 | amdgpu_asic_flush_hdp(adev, NULL); |
---|
335 | | - amdgpu_gmc_flush_gpu_tlb(adev, 0); |
---|
| 342 | + for (i = 0; i < adev->num_vmhubs; i++) |
---|
| 343 | + amdgpu_gmc_flush_gpu_tlb(adev, 0, i, 0); |
---|
336 | 344 | return 0; |
---|
337 | 345 | } |
---|
338 | 346 | |
---|