| .. | .. |
|---|
| 25 | 25 | * Alex Deucher |
|---|
| 26 | 26 | * Jerome Glisse |
|---|
| 27 | 27 | */ |
|---|
| 28 | | -#include <drm/drmP.h> |
|---|
| 28 | + |
|---|
| 29 | +#include <linux/pci.h> |
|---|
| 30 | +#include <linux/vmalloc.h> |
|---|
| 31 | + |
|---|
| 29 | 32 | #include <drm/amdgpu_drm.h> |
|---|
| 30 | 33 | #ifdef CONFIG_X86 |
|---|
| 31 | 34 | #include <asm/set_memory.h> |
|---|
| .. | .. |
|---|
| 68 | 71 | */ |
|---|
| 69 | 72 | static int amdgpu_gart_dummy_page_init(struct amdgpu_device *adev) |
|---|
| 70 | 73 | { |
|---|
| 71 | | - struct page *dummy_page = adev->mman.bdev.glob->dummy_read_page; |
|---|
| 74 | + struct page *dummy_page = ttm_bo_glob.dummy_read_page; |
|---|
| 72 | 75 | |
|---|
| 73 | 76 | if (adev->dummy_page_addr) |
|---|
| 74 | 77 | return 0; |
|---|
| .. | .. |
|---|
| 112 | 115 | { |
|---|
| 113 | 116 | int r; |
|---|
| 114 | 117 | |
|---|
| 115 | | - if (adev->gart.robj == NULL) { |
|---|
| 118 | + if (adev->gart.bo == NULL) { |
|---|
| 116 | 119 | struct amdgpu_bo_param bp; |
|---|
| 117 | 120 | |
|---|
| 118 | 121 | memset(&bp, 0, sizeof(bp)); |
|---|
| .. | .. |
|---|
| 123 | 126 | AMDGPU_GEM_CREATE_VRAM_CONTIGUOUS; |
|---|
| 124 | 127 | bp.type = ttm_bo_type_kernel; |
|---|
| 125 | 128 | bp.resv = NULL; |
|---|
| 126 | | - r = amdgpu_bo_create(adev, &bp, &adev->gart.robj); |
|---|
| 129 | + r = amdgpu_bo_create(adev, &bp, &adev->gart.bo); |
|---|
| 127 | 130 | if (r) { |
|---|
| 128 | 131 | return r; |
|---|
| 129 | 132 | } |
|---|
| .. | .. |
|---|
| 145 | 148 | { |
|---|
| 146 | 149 | int r; |
|---|
| 147 | 150 | |
|---|
| 148 | | - r = amdgpu_bo_reserve(adev->gart.robj, false); |
|---|
| 151 | + r = amdgpu_bo_reserve(adev->gart.bo, false); |
|---|
| 149 | 152 | if (unlikely(r != 0)) |
|---|
| 150 | 153 | return r; |
|---|
| 151 | | - r = amdgpu_bo_pin(adev->gart.robj, AMDGPU_GEM_DOMAIN_VRAM); |
|---|
| 154 | + r = amdgpu_bo_pin(adev->gart.bo, AMDGPU_GEM_DOMAIN_VRAM); |
|---|
| 152 | 155 | if (r) { |
|---|
| 153 | | - amdgpu_bo_unreserve(adev->gart.robj); |
|---|
| 156 | + amdgpu_bo_unreserve(adev->gart.bo); |
|---|
| 154 | 157 | return r; |
|---|
| 155 | 158 | } |
|---|
| 156 | | - r = amdgpu_bo_kmap(adev->gart.robj, &adev->gart.ptr); |
|---|
| 159 | + r = amdgpu_bo_kmap(adev->gart.bo, &adev->gart.ptr); |
|---|
| 157 | 160 | if (r) |
|---|
| 158 | | - amdgpu_bo_unpin(adev->gart.robj); |
|---|
| 159 | | - amdgpu_bo_unreserve(adev->gart.robj); |
|---|
| 160 | | - adev->gart.table_addr = amdgpu_bo_gpu_offset(adev->gart.robj); |
|---|
| 161 | + amdgpu_bo_unpin(adev->gart.bo); |
|---|
| 162 | + amdgpu_bo_unreserve(adev->gart.bo); |
|---|
| 161 | 163 | return r; |
|---|
| 162 | 164 | } |
|---|
| 163 | 165 | |
|---|
| .. | .. |
|---|
| 173 | 175 | { |
|---|
| 174 | 176 | int r; |
|---|
| 175 | 177 | |
|---|
| 176 | | - if (adev->gart.robj == NULL) { |
|---|
| 178 | + if (adev->gart.bo == NULL) { |
|---|
| 177 | 179 | return; |
|---|
| 178 | 180 | } |
|---|
| 179 | | - r = amdgpu_bo_reserve(adev->gart.robj, true); |
|---|
| 181 | + r = amdgpu_bo_reserve(adev->gart.bo, true); |
|---|
| 180 | 182 | if (likely(r == 0)) { |
|---|
| 181 | | - amdgpu_bo_kunmap(adev->gart.robj); |
|---|
| 182 | | - amdgpu_bo_unpin(adev->gart.robj); |
|---|
| 183 | | - amdgpu_bo_unreserve(adev->gart.robj); |
|---|
| 183 | + amdgpu_bo_kunmap(adev->gart.bo); |
|---|
| 184 | + amdgpu_bo_unpin(adev->gart.bo); |
|---|
| 185 | + amdgpu_bo_unreserve(adev->gart.bo); |
|---|
| 184 | 186 | adev->gart.ptr = NULL; |
|---|
| 185 | 187 | } |
|---|
| 186 | 188 | } |
|---|
| .. | .. |
|---|
| 196 | 198 | */ |
|---|
| 197 | 199 | void amdgpu_gart_table_vram_free(struct amdgpu_device *adev) |
|---|
| 198 | 200 | { |
|---|
| 199 | | - if (adev->gart.robj == NULL) { |
|---|
| 201 | + if (adev->gart.bo == NULL) { |
|---|
| 200 | 202 | return; |
|---|
| 201 | 203 | } |
|---|
| 202 | | - amdgpu_bo_unref(&adev->gart.robj); |
|---|
| 204 | + amdgpu_bo_unref(&adev->gart.bo); |
|---|
| 203 | 205 | } |
|---|
| 204 | 206 | |
|---|
| 205 | 207 | /* |
|---|
| .. | .. |
|---|
| 249 | 251 | } |
|---|
| 250 | 252 | mb(); |
|---|
| 251 | 253 | amdgpu_asic_flush_hdp(adev, NULL); |
|---|
| 252 | | - amdgpu_gmc_flush_gpu_tlb(adev, 0); |
|---|
| 254 | + for (i = 0; i < adev->num_vmhubs; i++) |
|---|
| 255 | + amdgpu_gmc_flush_gpu_tlb(adev, 0, i, 0); |
|---|
| 256 | + |
|---|
| 253 | 257 | return 0; |
|---|
| 254 | 258 | } |
|---|
| 255 | 259 | |
|---|
| .. | .. |
|---|
| 260 | 264 | * @offset: offset into the GPU's gart aperture |
|---|
| 261 | 265 | * @pages: number of pages to bind |
|---|
| 262 | 266 | * @dma_addr: DMA addresses of pages |
|---|
| 267 | + * @flags: page table entry flags |
|---|
| 268 | + * @dst: CPU address of the gart table |
|---|
| 263 | 269 | * |
|---|
| 264 | 270 | * Map the dma_addresses into GART entries (all asics). |
|---|
| 265 | 271 | * Returns 0 for success, -EINVAL for failure. |
|---|
| .. | .. |
|---|
| 296 | 302 | * @pages: number of pages to bind |
|---|
| 297 | 303 | * @pagelist: pages to bind |
|---|
| 298 | 304 | * @dma_addr: DMA addresses of pages |
|---|
| 305 | + * @flags: page table entry flags |
|---|
| 299 | 306 | * |
|---|
| 300 | 307 | * Binds the requested pages to the gart page table |
|---|
| 301 | 308 | * (all asics). |
|---|
| .. | .. |
|---|
| 306 | 313 | uint64_t flags) |
|---|
| 307 | 314 | { |
|---|
| 308 | 315 | #ifdef CONFIG_DRM_AMDGPU_GART_DEBUGFS |
|---|
| 309 | | - unsigned i,t,p; |
|---|
| 316 | + unsigned t,p; |
|---|
| 310 | 317 | #endif |
|---|
| 311 | | - int r; |
|---|
| 318 | + int r, i; |
|---|
| 312 | 319 | |
|---|
| 313 | 320 | if (!adev->gart.ready) { |
|---|
| 314 | 321 | WARN(1, "trying to bind memory to uninitialized GART !\n"); |
|---|
| .. | .. |
|---|
| 332 | 339 | |
|---|
| 333 | 340 | mb(); |
|---|
| 334 | 341 | amdgpu_asic_flush_hdp(adev, NULL); |
|---|
| 335 | | - amdgpu_gmc_flush_gpu_tlb(adev, 0); |
|---|
| 342 | + for (i = 0; i < adev->num_vmhubs; i++) |
|---|
| 343 | + amdgpu_gmc_flush_gpu_tlb(adev, 0, i, 0); |
|---|
| 336 | 344 | return 0; |
|---|
| 337 | 345 | } |
|---|
| 338 | 346 | |
|---|