| .. | .. |
|---|
| 1 | 1 | // SPDX-License-Identifier: GPL-2.0 WITH Linux-syscall-note |
|---|
| 2 | 2 | /* |
|---|
| 3 | 3 | * |
|---|
| 4 | | - * (C) COPYRIGHT 2010-2021 ARM Limited. All rights reserved. |
|---|
| 4 | + * (C) COPYRIGHT 2010-2023 ARM Limited. All rights reserved. |
|---|
| 5 | 5 | * |
|---|
| 6 | 6 | * This program is free software and is provided to you under the terms of the |
|---|
| 7 | 7 | * GNU General Public License version 2 as published by the Free Software |
|---|
| .. | .. |
|---|
| 31 | 31 | #include <linux/fs.h> |
|---|
| 32 | 32 | #include <linux/version.h> |
|---|
| 33 | 33 | #include <linux/dma-mapping.h> |
|---|
| 34 | | -#if (KERNEL_VERSION(4, 8, 0) > LINUX_VERSION_CODE) |
|---|
| 35 | | -#include <linux/dma-attrs.h> |
|---|
| 36 | | -#endif /* LINUX_VERSION_CODE < 4.8.0 */ |
|---|
| 37 | 34 | #include <linux/dma-buf.h> |
|---|
| 38 | 35 | #include <linux/shrinker.h> |
|---|
| 39 | 36 | #include <linux/cache.h> |
|---|
| 40 | 37 | #include <linux/memory_group_manager.h> |
|---|
| 41 | | - |
|---|
| 38 | +#include <linux/math64.h> |
|---|
| 39 | +#include <linux/migrate.h> |
|---|
| 40 | +#include <linux/version.h> |
|---|
| 42 | 41 | #include <mali_kbase.h> |
|---|
| 43 | 42 | #include <mali_kbase_mem_linux.h> |
|---|
| 44 | 43 | #include <tl/mali_kbase_tracepoints.h> |
|---|
| .. | .. |
|---|
| 86 | 85 | #define IR_THRESHOLD_STEPS (256u) |
|---|
| 87 | 86 | |
|---|
| 88 | 87 | #if MALI_USE_CSF |
|---|
| 89 | | -static int kbase_csf_cpu_mmap_user_reg_page(struct kbase_context *kctx, |
|---|
| 90 | | - struct vm_area_struct *vma); |
|---|
| 91 | | -static int kbase_csf_cpu_mmap_user_io_pages(struct kbase_context *kctx, |
|---|
| 92 | | - struct vm_area_struct *vma); |
|---|
| 88 | +static int kbase_csf_cpu_mmap_user_reg_page(struct kbase_context *kctx, struct vm_area_struct *vma); |
|---|
| 89 | +static int kbase_csf_cpu_mmap_user_io_pages(struct kbase_context *kctx, struct vm_area_struct *vma); |
|---|
| 93 | 90 | #endif |
|---|
| 94 | 91 | |
|---|
| 95 | | -static int kbase_vmap_phy_pages(struct kbase_context *kctx, |
|---|
| 96 | | - struct kbase_va_region *reg, u64 offset_bytes, size_t size, |
|---|
| 97 | | - struct kbase_vmap_struct *map); |
|---|
| 92 | +static int kbase_vmap_phy_pages(struct kbase_context *kctx, struct kbase_va_region *reg, |
|---|
| 93 | + u64 offset_bytes, size_t size, struct kbase_vmap_struct *map, |
|---|
| 94 | + kbase_vmap_flag vmap_flags); |
|---|
| 98 | 95 | static void kbase_vunmap_phy_pages(struct kbase_context *kctx, |
|---|
| 99 | 96 | struct kbase_vmap_struct *map); |
|---|
| 100 | 97 | |
|---|
| 101 | 98 | static int kbase_tracking_page_setup(struct kbase_context *kctx, struct vm_area_struct *vma); |
|---|
| 102 | 99 | |
|---|
| 103 | | -static int kbase_mem_shrink_gpu_mapping(struct kbase_context *kctx, |
|---|
| 104 | | - struct kbase_va_region *reg, |
|---|
| 105 | | - u64 new_pages, u64 old_pages); |
|---|
| 100 | +static bool is_process_exiting(struct vm_area_struct *vma) |
|---|
| 101 | +{ |
|---|
| 102 | + /* PF_EXITING flag can't be reliably used here for the detection |
|---|
| 103 | + * of process exit, as 'mm_users' counter could still be non-zero |
|---|
| 104 | + * when all threads of the process have exited. Later when the |
|---|
| 105 | + * thread (which took a reference on the 'mm' of process that |
|---|
| 106 | + * exited) drops it reference, the vm_ops->close method would be |
|---|
| 107 | + * called for all the vmas (owned by 'mm' of process that exited) |
|---|
| 108 | + * but the PF_EXITING flag may not be neccessarily set for the |
|---|
| 109 | + * thread at that time. |
|---|
| 110 | + */ |
|---|
| 111 | + if (atomic_read(&vma->vm_mm->mm_users)) |
|---|
| 112 | + return false; |
|---|
| 113 | + |
|---|
| 114 | + return true; |
|---|
| 115 | +} |
|---|
| 106 | 116 | |
|---|
| 107 | 117 | /* Retrieve the associated region pointer if the GPU address corresponds to |
|---|
| 108 | 118 | * one of the event memory pages. The enclosing region, if found, shouldn't |
|---|
| .. | .. |
|---|
| 184 | 194 | reg->cpu_alloc->type != KBASE_MEM_TYPE_NATIVE) |
|---|
| 185 | 195 | return -EINVAL; |
|---|
| 186 | 196 | |
|---|
| 187 | | - if (size > (KBASE_PERMANENTLY_MAPPED_MEM_LIMIT_PAGES - |
|---|
| 188 | | - atomic_read(&kctx->permanent_mapped_pages))) { |
|---|
| 189 | | - dev_warn(kctx->kbdev->dev, "Request for %llu more pages mem needing a permanent mapping would breach limit %lu, currently at %d pages", |
|---|
| 190 | | - (u64)size, |
|---|
| 191 | | - KBASE_PERMANENTLY_MAPPED_MEM_LIMIT_PAGES, |
|---|
| 192 | | - atomic_read(&kctx->permanent_mapped_pages)); |
|---|
| 193 | | - return -ENOMEM; |
|---|
| 194 | | - } |
|---|
| 195 | | - |
|---|
| 196 | 197 | kern_mapping = kzalloc(sizeof(*kern_mapping), GFP_KERNEL); |
|---|
| 197 | 198 | if (!kern_mapping) |
|---|
| 198 | 199 | return -ENOMEM; |
|---|
| 199 | 200 | |
|---|
| 200 | | - err = kbase_vmap_phy_pages(kctx, reg, 0u, size_bytes, kern_mapping); |
|---|
| 201 | + err = kbase_vmap_phy_pages(kctx, reg, 0u, size_bytes, kern_mapping, |
|---|
| 202 | + KBASE_VMAP_FLAG_PERMANENT_MAP_ACCOUNTING); |
|---|
| 201 | 203 | if (err < 0) |
|---|
| 202 | 204 | goto vmap_fail; |
|---|
| 203 | 205 | |
|---|
| .. | .. |
|---|
| 205 | 207 | reg->flags &= ~KBASE_REG_GROWABLE; |
|---|
| 206 | 208 | |
|---|
| 207 | 209 | reg->cpu_alloc->permanent_map = kern_mapping; |
|---|
| 208 | | - atomic_add(size, &kctx->permanent_mapped_pages); |
|---|
| 209 | 210 | |
|---|
| 210 | 211 | return 0; |
|---|
| 211 | 212 | vmap_fail: |
|---|
| .. | .. |
|---|
| 221 | 222 | kfree(alloc->permanent_map); |
|---|
| 222 | 223 | |
|---|
| 223 | 224 | alloc->permanent_map = NULL; |
|---|
| 224 | | - |
|---|
| 225 | | - /* Mappings are only done on cpu_alloc, so don't need to worry about |
|---|
| 226 | | - * this being reduced a second time if a separate gpu_alloc is |
|---|
| 227 | | - * freed |
|---|
| 228 | | - */ |
|---|
| 229 | | - WARN_ON(alloc->nents > atomic_read(&kctx->permanent_mapped_pages)); |
|---|
| 230 | | - atomic_sub(alloc->nents, &kctx->permanent_mapped_pages); |
|---|
| 231 | 225 | } |
|---|
| 232 | 226 | |
|---|
| 233 | 227 | void *kbase_phy_alloc_mapping_get(struct kbase_context *kctx, |
|---|
| .. | .. |
|---|
| 291 | 285 | */ |
|---|
| 292 | 286 | } |
|---|
| 293 | 287 | |
|---|
| 294 | | -struct kbase_va_region *kbase_mem_alloc(struct kbase_context *kctx, |
|---|
| 295 | | - u64 va_pages, u64 commit_pages, |
|---|
| 296 | | - u64 extension, u64 *flags, u64 *gpu_va) |
|---|
| 288 | +struct kbase_va_region *kbase_mem_alloc(struct kbase_context *kctx, u64 va_pages, u64 commit_pages, |
|---|
| 289 | + u64 extension, u64 *flags, u64 *gpu_va, |
|---|
| 290 | + enum kbase_caller_mmu_sync_info mmu_sync_info) |
|---|
| 297 | 291 | { |
|---|
| 298 | 292 | int zone; |
|---|
| 299 | 293 | struct kbase_va_region *reg; |
|---|
| .. | .. |
|---|
| 310 | 304 | va_pages, commit_pages, extension, *flags); |
|---|
| 311 | 305 | |
|---|
| 312 | 306 | #if MALI_USE_CSF |
|---|
| 313 | | - *gpu_va = 0; /* return 0 on failure */ |
|---|
| 307 | + if (!(*flags & BASE_MEM_FIXED)) |
|---|
| 308 | + *gpu_va = 0; /* return 0 on failure */ |
|---|
| 314 | 309 | #else |
|---|
| 315 | 310 | if (!(*flags & BASE_MEM_FLAG_MAP_FIXED)) |
|---|
| 316 | 311 | *gpu_va = 0; /* return 0 on failure */ |
|---|
| 312 | +#endif |
|---|
| 317 | 313 | else |
|---|
| 318 | | - dev_err(dev, |
|---|
| 314 | + dev_dbg(dev, |
|---|
| 319 | 315 | "Keeping requested GPU VA of 0x%llx\n", |
|---|
| 320 | 316 | (unsigned long long)*gpu_va); |
|---|
| 321 | | -#endif |
|---|
| 322 | 317 | |
|---|
| 323 | 318 | if (!kbase_check_alloc_flags(*flags)) { |
|---|
| 324 | 319 | dev_warn(dev, |
|---|
| 325 | | - "kbase_mem_alloc called with bad flags (%llx)", |
|---|
| 320 | + "%s called with bad flags (%llx)", |
|---|
| 321 | + __func__, |
|---|
| 326 | 322 | (unsigned long long)*flags); |
|---|
| 327 | 323 | goto bad_flags; |
|---|
| 328 | 324 | } |
|---|
| .. | .. |
|---|
| 344 | 340 | } |
|---|
| 345 | 341 | if ((*flags & BASE_MEM_COHERENT_SYSTEM_REQUIRED) != 0 && |
|---|
| 346 | 342 | !kbase_device_is_cpu_coherent(kctx->kbdev)) { |
|---|
| 347 | | - dev_warn(dev, "kbase_mem_alloc call required coherent mem when unavailable"); |
|---|
| 343 | + dev_warn(dev, "%s call required coherent mem when unavailable", |
|---|
| 344 | + __func__); |
|---|
| 348 | 345 | goto bad_flags; |
|---|
| 349 | 346 | } |
|---|
| 350 | 347 | if ((*flags & BASE_MEM_COHERENT_SYSTEM) != 0 && |
|---|
| .. | .. |
|---|
| 367 | 364 | if (*flags & BASE_MEM_SAME_VA) { |
|---|
| 368 | 365 | rbtree = &kctx->reg_rbtree_same; |
|---|
| 369 | 366 | zone = KBASE_REG_ZONE_SAME_VA; |
|---|
| 370 | | - } else if ((*flags & BASE_MEM_PROT_GPU_EX) && kbase_has_exec_va_zone(kctx)) { |
|---|
| 367 | + } |
|---|
| 368 | +#if MALI_USE_CSF |
|---|
| 369 | + /* fixed va_zone always exists */ |
|---|
| 370 | + else if (*flags & (BASE_MEM_FIXED | BASE_MEM_FIXABLE)) { |
|---|
| 371 | + if (*flags & BASE_MEM_PROT_GPU_EX) { |
|---|
| 372 | + rbtree = &kctx->reg_rbtree_exec_fixed; |
|---|
| 373 | + zone = KBASE_REG_ZONE_EXEC_FIXED_VA; |
|---|
| 374 | + } else { |
|---|
| 375 | + rbtree = &kctx->reg_rbtree_fixed; |
|---|
| 376 | + zone = KBASE_REG_ZONE_FIXED_VA; |
|---|
| 377 | + } |
|---|
| 378 | + } |
|---|
| 379 | +#endif |
|---|
| 380 | + else if ((*flags & BASE_MEM_PROT_GPU_EX) && kbase_has_exec_va_zone(kctx)) { |
|---|
| 371 | 381 | rbtree = &kctx->reg_rbtree_exec; |
|---|
| 372 | 382 | zone = KBASE_REG_ZONE_EXEC_VA; |
|---|
| 373 | 383 | } else { |
|---|
| .. | .. |
|---|
| 375 | 385 | zone = KBASE_REG_ZONE_CUSTOM_VA; |
|---|
| 376 | 386 | } |
|---|
| 377 | 387 | |
|---|
| 378 | | - reg = kbase_alloc_free_region(rbtree, PFN_DOWN(*gpu_va), |
|---|
| 379 | | - va_pages, zone); |
|---|
| 388 | + reg = kbase_alloc_free_region(kctx->kbdev, rbtree, PFN_DOWN(*gpu_va), va_pages, zone); |
|---|
| 380 | 389 | |
|---|
| 381 | 390 | if (!reg) { |
|---|
| 382 | 391 | dev_err(dev, "Failed to allocate free region"); |
|---|
| .. | .. |
|---|
| 387 | 396 | goto invalid_flags; |
|---|
| 388 | 397 | |
|---|
| 389 | 398 | if (kbase_reg_prepare_native(reg, kctx, |
|---|
| 390 | | - base_mem_group_id_get(*flags)) != 0) { |
|---|
| 399 | + kbase_mem_group_id_get(*flags)) != 0) { |
|---|
| 391 | 400 | dev_err(dev, "Failed to prepare region"); |
|---|
| 392 | 401 | goto prepare_failed; |
|---|
| 393 | 402 | } |
|---|
| .. | .. |
|---|
| 469 | 478 | |
|---|
| 470 | 479 | *gpu_va = (u64) cookie; |
|---|
| 471 | 480 | } else /* we control the VA */ { |
|---|
| 472 | | - if (kbase_gpu_mmap(kctx, reg, *gpu_va, va_pages, 1) != 0) { |
|---|
| 481 | + size_t align = 1; |
|---|
| 482 | + |
|---|
| 483 | + if (kctx->kbdev->pagesize_2mb) { |
|---|
| 484 | + /* If there's enough (> 33 bits) of GPU VA space, align to 2MB |
|---|
| 485 | + * boundaries. The similar condition is used for mapping from |
|---|
| 486 | + * the SAME_VA zone inside kbase_context_get_unmapped_area(). |
|---|
| 487 | + */ |
|---|
| 488 | + if (kctx->kbdev->gpu_props.mmu.va_bits > 33) { |
|---|
| 489 | + if (va_pages >= (SZ_2M / SZ_4K)) |
|---|
| 490 | + align = (SZ_2M / SZ_4K); |
|---|
| 491 | + } |
|---|
| 492 | + if (*gpu_va) |
|---|
| 493 | + align = 1; |
|---|
| 494 | +#if !MALI_USE_CSF |
|---|
| 495 | + if (reg->flags & KBASE_REG_TILER_ALIGN_TOP) |
|---|
| 496 | + align = 1; |
|---|
| 497 | +#endif /* !MALI_USE_CSF */ |
|---|
| 498 | + } |
|---|
| 499 | + if (kbase_gpu_mmap(kctx, reg, *gpu_va, va_pages, align, |
|---|
| 500 | + mmu_sync_info) != 0) { |
|---|
| 473 | 501 | dev_warn(dev, "Failed to map memory on GPU"); |
|---|
| 474 | 502 | kbase_gpu_vm_unlock(kctx); |
|---|
| 475 | 503 | goto no_mmap; |
|---|
| .. | .. |
|---|
| 490 | 518 | #endif /* MALI_JIT_PRESSURE_LIMIT_BASE */ |
|---|
| 491 | 519 | |
|---|
| 492 | 520 | kbase_gpu_vm_unlock(kctx); |
|---|
| 521 | + |
|---|
| 522 | +#if MALI_USE_CSF |
|---|
| 523 | + if (*flags & BASE_MEM_FIXABLE) |
|---|
| 524 | + atomic64_inc(&kctx->num_fixable_allocs); |
|---|
| 525 | + else if (*flags & BASE_MEM_FIXED) |
|---|
| 526 | + atomic64_inc(&kctx->num_fixed_allocs); |
|---|
| 527 | +#endif |
|---|
| 528 | + |
|---|
| 493 | 529 | return reg; |
|---|
| 494 | 530 | |
|---|
| 495 | 531 | no_mmap: |
|---|
| .. | .. |
|---|
| 600 | 636 | #if MALI_USE_CSF |
|---|
| 601 | 637 | if (KBASE_REG_CSF_EVENT & reg->flags) |
|---|
| 602 | 638 | *out |= BASE_MEM_CSF_EVENT; |
|---|
| 639 | + if (((KBASE_REG_ZONE_MASK & reg->flags) == KBASE_REG_ZONE_FIXED_VA) || |
|---|
| 640 | + ((KBASE_REG_ZONE_MASK & reg->flags) == KBASE_REG_ZONE_EXEC_FIXED_VA)) { |
|---|
| 641 | + if (KBASE_REG_FIXED_ADDRESS & reg->flags) |
|---|
| 642 | + *out |= BASE_MEM_FIXED; |
|---|
| 643 | + else |
|---|
| 644 | + *out |= BASE_MEM_FIXABLE; |
|---|
| 645 | + } |
|---|
| 603 | 646 | #endif |
|---|
| 604 | 647 | if (KBASE_REG_GPU_VA_SAME_4GB_PAGE & reg->flags) |
|---|
| 605 | 648 | *out |= BASE_MEM_GPU_VA_SAME_4GB_PAGE; |
|---|
| 606 | 649 | |
|---|
| 607 | | - *out |= base_mem_group_id_set(reg->cpu_alloc->group_id); |
|---|
| 650 | + *out |= kbase_mem_group_id_set(reg->cpu_alloc->group_id); |
|---|
| 608 | 651 | |
|---|
| 609 | 652 | WARN(*out & ~BASE_MEM_FLAGS_QUERYABLE, |
|---|
| 610 | 653 | "BASE_MEM_FLAGS_QUERYABLE needs updating\n"); |
|---|
| .. | .. |
|---|
| 629 | 672 | * @s: Shrinker |
|---|
| 630 | 673 | * @sc: Shrinker control |
|---|
| 631 | 674 | * |
|---|
| 632 | | - * Return: Number of pages which can be freed. |
|---|
| 675 | + * Return: Number of pages which can be freed or SHRINK_EMPTY if no page remains. |
|---|
| 633 | 676 | */ |
|---|
| 634 | 677 | static |
|---|
| 635 | 678 | unsigned long kbase_mem_evictable_reclaim_count_objects(struct shrinker *s, |
|---|
| 636 | 679 | struct shrink_control *sc) |
|---|
| 637 | 680 | { |
|---|
| 638 | | - struct kbase_context *kctx; |
|---|
| 639 | | - |
|---|
| 640 | | - kctx = container_of(s, struct kbase_context, reclaim); |
|---|
| 681 | + struct kbase_context *kctx = container_of(s, struct kbase_context, reclaim); |
|---|
| 682 | + int evict_nents = atomic_read(&kctx->evict_nents); |
|---|
| 683 | + unsigned long nr_freeable_items; |
|---|
| 641 | 684 | |
|---|
| 642 | 685 | WARN((sc->gfp_mask & __GFP_ATOMIC), |
|---|
| 643 | 686 | "Shrinkers cannot be called for GFP_ATOMIC allocations. Check kernel mm for problems. gfp_mask==%x\n", |
|---|
| 644 | 687 | sc->gfp_mask); |
|---|
| 645 | 688 | WARN(in_atomic(), |
|---|
| 646 | | - "Shrinker called whilst in atomic context. The caller must switch to using GFP_ATOMIC or similar. gfp_mask==%x\n", |
|---|
| 689 | + "Shrinker called in atomic context. The caller must use GFP_ATOMIC or similar, then Shrinkers must not be called. gfp_mask==%x\n", |
|---|
| 647 | 690 | sc->gfp_mask); |
|---|
| 648 | 691 | |
|---|
| 649 | | - return atomic_read(&kctx->evict_nents); |
|---|
| 692 | + if (unlikely(evict_nents < 0)) { |
|---|
| 693 | + dev_err(kctx->kbdev->dev, "invalid evict_nents(%d)", evict_nents); |
|---|
| 694 | + nr_freeable_items = 0; |
|---|
| 695 | + } else { |
|---|
| 696 | + nr_freeable_items = evict_nents; |
|---|
| 697 | + } |
|---|
| 698 | + |
|---|
| 699 | +#if KERNEL_VERSION(4, 19, 0) <= LINUX_VERSION_CODE |
|---|
| 700 | + if (nr_freeable_items == 0) |
|---|
| 701 | + nr_freeable_items = SHRINK_EMPTY; |
|---|
| 702 | +#endif |
|---|
| 703 | + |
|---|
| 704 | + return nr_freeable_items; |
|---|
| 650 | 705 | } |
|---|
| 651 | 706 | |
|---|
| 652 | 707 | /** |
|---|
| .. | .. |
|---|
| 655 | 710 | * @s: Shrinker |
|---|
| 656 | 711 | * @sc: Shrinker control |
|---|
| 657 | 712 | * |
|---|
| 658 | | - * Return: Number of pages freed (can be less then requested) or -1 if the |
|---|
| 659 | | - * shrinker failed to free pages in its pool. |
|---|
| 713 | + * Return: Number of pages freed (can be less then requested) or |
|---|
| 714 | + * SHRINK_STOP if reclaim isn't possible. |
|---|
| 660 | 715 | * |
|---|
| 661 | 716 | * Note: |
|---|
| 662 | 717 | * This function accesses region structures without taking the region lock, |
|---|
| .. | .. |
|---|
| 684 | 739 | list_for_each_entry_safe(alloc, tmp, &kctx->evict_list, evict_node) { |
|---|
| 685 | 740 | int err; |
|---|
| 686 | 741 | |
|---|
| 742 | + if (!alloc->reg) |
|---|
| 743 | + continue; |
|---|
| 744 | + |
|---|
| 687 | 745 | err = kbase_mem_shrink_gpu_mapping(kctx, alloc->reg, |
|---|
| 688 | 746 | 0, alloc->nents); |
|---|
| 689 | | - if (err != 0) { |
|---|
| 690 | | - /* |
|---|
| 691 | | - * Failed to remove GPU mapping, tell the shrinker |
|---|
| 692 | | - * to stop trying to shrink our slab even though we |
|---|
| 693 | | - * have pages in it. |
|---|
| 694 | | - */ |
|---|
| 695 | | - freed = -1; |
|---|
| 696 | | - goto out_unlock; |
|---|
| 697 | | - } |
|---|
| 747 | + |
|---|
| 748 | + /* Failed to remove GPU mapping, proceed to next one. */ |
|---|
| 749 | + if (err != 0) |
|---|
| 750 | + continue; |
|---|
| 698 | 751 | |
|---|
| 699 | 752 | /* |
|---|
| 700 | 753 | * Update alloc->evicted before freeing the backing so the |
|---|
| .. | .. |
|---|
| 718 | 771 | if (freed > sc->nr_to_scan) |
|---|
| 719 | 772 | break; |
|---|
| 720 | 773 | } |
|---|
| 721 | | -out_unlock: |
|---|
| 774 | + |
|---|
| 722 | 775 | mutex_unlock(&kctx->jit_evict_lock); |
|---|
| 723 | 776 | |
|---|
| 724 | 777 | return freed; |
|---|
| .. | .. |
|---|
| 738 | 791 | * struct shrinker does not define batch |
|---|
| 739 | 792 | */ |
|---|
| 740 | 793 | kctx->reclaim.batch = 0; |
|---|
| 794 | +#if KERNEL_VERSION(6, 0, 0) > LINUX_VERSION_CODE |
|---|
| 741 | 795 | register_shrinker(&kctx->reclaim); |
|---|
| 796 | +#else |
|---|
| 797 | + register_shrinker(&kctx->reclaim, "mali-mem"); |
|---|
| 798 | +#endif |
|---|
| 742 | 799 | return 0; |
|---|
| 743 | 800 | } |
|---|
| 744 | 801 | |
|---|
| .. | .. |
|---|
| 802 | 859 | |
|---|
| 803 | 860 | lockdep_assert_held(&kctx->reg_lock); |
|---|
| 804 | 861 | |
|---|
| 862 | + /* Memory is in the process of transitioning to the shrinker, and |
|---|
| 863 | + * should ignore migration attempts |
|---|
| 864 | + */ |
|---|
| 805 | 865 | kbase_mem_shrink_cpu_mapping(kctx, gpu_alloc->reg, |
|---|
| 806 | 866 | 0, gpu_alloc->nents); |
|---|
| 807 | 867 | |
|---|
| .. | .. |
|---|
| 809 | 869 | /* This allocation can't already be on a list. */ |
|---|
| 810 | 870 | WARN_ON(!list_empty(&gpu_alloc->evict_node)); |
|---|
| 811 | 871 | |
|---|
| 812 | | - /* |
|---|
| 813 | | - * Add the allocation to the eviction list, after this point the shrink |
|---|
| 872 | + /* Add the allocation to the eviction list, after this point the shrink |
|---|
| 814 | 873 | * can reclaim it. |
|---|
| 815 | 874 | */ |
|---|
| 816 | 875 | list_add(&gpu_alloc->evict_node, &kctx->evict_list); |
|---|
| 817 | 876 | atomic_add(gpu_alloc->nents, &kctx->evict_nents); |
|---|
| 877 | + |
|---|
| 878 | + /* Indicate to page migration that the memory can be reclaimed by the shrinker. |
|---|
| 879 | + */ |
|---|
| 880 | + if (kbase_page_migration_enabled) |
|---|
| 881 | + kbase_set_phy_alloc_page_status(gpu_alloc, NOT_MOVABLE); |
|---|
| 882 | + |
|---|
| 818 | 883 | mutex_unlock(&kctx->jit_evict_lock); |
|---|
| 819 | 884 | kbase_mem_evictable_mark_reclaim(gpu_alloc); |
|---|
| 820 | 885 | |
|---|
| .. | .. |
|---|
| 826 | 891 | { |
|---|
| 827 | 892 | struct kbase_context *kctx = gpu_alloc->imported.native.kctx; |
|---|
| 828 | 893 | int err = 0; |
|---|
| 894 | + |
|---|
| 895 | + /* Calls to this function are inherently asynchronous, with respect to |
|---|
| 896 | + * MMU operations. |
|---|
| 897 | + */ |
|---|
| 898 | + const enum kbase_caller_mmu_sync_info mmu_sync_info = CALLER_MMU_ASYNC; |
|---|
| 829 | 899 | |
|---|
| 830 | 900 | lockdep_assert_held(&kctx->reg_lock); |
|---|
| 831 | 901 | |
|---|
| .. | .. |
|---|
| 856 | 926 | * pre-eviction size. |
|---|
| 857 | 927 | */ |
|---|
| 858 | 928 | if (!err) |
|---|
| 859 | | - err = kbase_mem_grow_gpu_mapping(kctx, |
|---|
| 860 | | - gpu_alloc->reg, |
|---|
| 861 | | - gpu_alloc->evicted, 0); |
|---|
| 929 | + err = kbase_mem_grow_gpu_mapping( |
|---|
| 930 | + kctx, gpu_alloc->reg, |
|---|
| 931 | + gpu_alloc->evicted, 0, mmu_sync_info); |
|---|
| 862 | 932 | |
|---|
| 863 | 933 | gpu_alloc->evicted = 0; |
|---|
| 934 | + |
|---|
| 935 | + /* Since the allocation is no longer evictable, and we ensure that |
|---|
| 936 | + * it grows back to its pre-eviction size, we will consider the |
|---|
| 937 | + * state of it to be ALLOCATED_MAPPED, as that is the only state |
|---|
| 938 | + * in which a physical allocation could transition to NOT_MOVABLE |
|---|
| 939 | + * from. |
|---|
| 940 | + */ |
|---|
| 941 | + if (kbase_page_migration_enabled) |
|---|
| 942 | + kbase_set_phy_alloc_page_status(gpu_alloc, ALLOCATED_MAPPED); |
|---|
| 864 | 943 | } |
|---|
| 865 | 944 | } |
|---|
| 866 | 945 | |
|---|
| .. | .. |
|---|
| 911 | 990 | /* Validate the region */ |
|---|
| 912 | 991 | reg = kbase_region_tracker_find_region_base_address(kctx, gpu_addr); |
|---|
| 913 | 992 | if (kbase_is_region_invalid_or_free(reg)) |
|---|
| 993 | + goto out_unlock; |
|---|
| 994 | + |
|---|
| 995 | + /* There is no use case to support MEM_FLAGS_CHANGE ioctl for allocations |
|---|
| 996 | + * that have NO_USER_FREE flag set, to mark them as evictable/reclaimable. |
|---|
| 997 | + * This would usually include JIT allocations, Tiler heap related allocations |
|---|
| 998 | + * & GPU queue ringbuffer and none of them needs to be explicitly marked |
|---|
| 999 | + * as evictable by Userspace. |
|---|
| 1000 | + */ |
|---|
| 1001 | + if (kbase_va_region_is_no_user_free(reg)) |
|---|
| 914 | 1002 | goto out_unlock; |
|---|
| 915 | 1003 | |
|---|
| 916 | 1004 | /* Is the region being transitioning between not needed and needed? */ |
|---|
| .. | .. |
|---|
| 1022 | 1110 | struct kbase_va_region *reg, enum kbase_sync_type sync_fn) |
|---|
| 1023 | 1111 | { |
|---|
| 1024 | 1112 | int ret = -EINVAL; |
|---|
| 1025 | | - struct dma_buf *dma_buf; |
|---|
| 1113 | + struct dma_buf __maybe_unused *dma_buf; |
|---|
| 1026 | 1114 | enum dma_data_direction dir = DMA_BIDIRECTIONAL; |
|---|
| 1027 | 1115 | |
|---|
| 1028 | 1116 | lockdep_assert_held(&kctx->reg_lock); |
|---|
| .. | .. |
|---|
| 1066 | 1154 | ret = 0; |
|---|
| 1067 | 1155 | } |
|---|
| 1068 | 1156 | #else |
|---|
| 1069 | | - /* Though the below version check could be superfluous depending upon the version condition |
|---|
| 1070 | | - * used for enabling KBASE_MEM_ION_SYNC_WORKAROUND, we still keep this check here to allow |
|---|
| 1071 | | - * ease of modification for non-ION systems or systems where ION has been patched. |
|---|
| 1072 | | - */ |
|---|
| 1073 | | -#if KERNEL_VERSION(4, 6, 0) > LINUX_VERSION_CODE && !defined(CONFIG_CHROMEOS) |
|---|
| 1074 | | - dma_buf_end_cpu_access(dma_buf, |
|---|
| 1075 | | - 0, dma_buf->size, |
|---|
| 1076 | | - dir); |
|---|
| 1077 | | - ret = 0; |
|---|
| 1078 | | -#else |
|---|
| 1079 | | - ret = dma_buf_end_cpu_access(dma_buf, |
|---|
| 1080 | | - dir); |
|---|
| 1081 | | -#endif |
|---|
| 1157 | + ret = dma_buf_end_cpu_access(dma_buf, dir); |
|---|
| 1082 | 1158 | #endif /* KBASE_MEM_ION_SYNC_WORKAROUND */ |
|---|
| 1083 | 1159 | break; |
|---|
| 1084 | 1160 | case KBASE_SYNC_TO_CPU: |
|---|
| .. | .. |
|---|
| 1095 | 1171 | ret = 0; |
|---|
| 1096 | 1172 | } |
|---|
| 1097 | 1173 | #else |
|---|
| 1098 | | - ret = dma_buf_begin_cpu_access(dma_buf, |
|---|
| 1099 | | -#if KERNEL_VERSION(4, 6, 0) > LINUX_VERSION_CODE && !defined(CONFIG_CHROMEOS) |
|---|
| 1100 | | - 0, dma_buf->size, |
|---|
| 1101 | | -#endif |
|---|
| 1102 | | - dir); |
|---|
| 1174 | + ret = dma_buf_begin_cpu_access(dma_buf, dir); |
|---|
| 1103 | 1175 | #endif /* KBASE_MEM_ION_SYNC_WORKAROUND */ |
|---|
| 1104 | 1176 | break; |
|---|
| 1105 | 1177 | } |
|---|
| .. | .. |
|---|
| 1218 | 1290 | struct kbase_mem_phy_alloc *alloc; |
|---|
| 1219 | 1291 | unsigned long gwt_mask = ~0; |
|---|
| 1220 | 1292 | |
|---|
| 1293 | + /* Calls to this function are inherently asynchronous, with respect to |
|---|
| 1294 | + * MMU operations. |
|---|
| 1295 | + */ |
|---|
| 1296 | + const enum kbase_caller_mmu_sync_info mmu_sync_info = CALLER_MMU_ASYNC; |
|---|
| 1297 | + |
|---|
| 1221 | 1298 | lockdep_assert_held(&kctx->reg_lock); |
|---|
| 1222 | 1299 | |
|---|
| 1223 | 1300 | alloc = reg->gpu_alloc; |
|---|
| .. | .. |
|---|
| 1244 | 1321 | gwt_mask = ~KBASE_REG_GPU_WR; |
|---|
| 1245 | 1322 | #endif |
|---|
| 1246 | 1323 | |
|---|
| 1247 | | - err = kbase_mmu_insert_pages(kctx->kbdev, |
|---|
| 1248 | | - &kctx->mmu, |
|---|
| 1249 | | - reg->start_pfn, |
|---|
| 1250 | | - kbase_get_gpu_phy_pages(reg), |
|---|
| 1251 | | - kbase_reg_current_backed_size(reg), |
|---|
| 1252 | | - reg->flags & gwt_mask, |
|---|
| 1253 | | - kctx->as_nr, |
|---|
| 1254 | | - alloc->group_id); |
|---|
| 1324 | + err = kbase_mmu_insert_imported_pages(kctx->kbdev, &kctx->mmu, reg->start_pfn, |
|---|
| 1325 | + kbase_get_gpu_phy_pages(reg), |
|---|
| 1326 | + kbase_reg_current_backed_size(reg), |
|---|
| 1327 | + reg->flags & gwt_mask, kctx->as_nr, alloc->group_id, |
|---|
| 1328 | + mmu_sync_info, NULL); |
|---|
| 1255 | 1329 | if (err) |
|---|
| 1256 | 1330 | goto bad_insert; |
|---|
| 1257 | 1331 | |
|---|
| .. | .. |
|---|
| 1264 | 1338 | * Assume alloc->nents is the number of actual pages in the |
|---|
| 1265 | 1339 | * dma-buf memory. |
|---|
| 1266 | 1340 | */ |
|---|
| 1267 | | - err = kbase_mmu_insert_single_page(kctx, |
|---|
| 1268 | | - reg->start_pfn + alloc->nents, |
|---|
| 1269 | | - kctx->aliasing_sink_page, |
|---|
| 1270 | | - reg->nr_pages - alloc->nents, |
|---|
| 1271 | | - (reg->flags | KBASE_REG_GPU_RD) & |
|---|
| 1272 | | - ~KBASE_REG_GPU_WR, |
|---|
| 1273 | | - KBASE_MEM_GROUP_SINK); |
|---|
| 1341 | + err = kbase_mmu_insert_single_imported_page( |
|---|
| 1342 | + kctx, reg->start_pfn + alloc->nents, kctx->aliasing_sink_page, |
|---|
| 1343 | + reg->nr_pages - alloc->nents, |
|---|
| 1344 | + (reg->flags | KBASE_REG_GPU_RD) & ~KBASE_REG_GPU_WR, KBASE_MEM_GROUP_SINK, |
|---|
| 1345 | + mmu_sync_info); |
|---|
| 1274 | 1346 | if (err) |
|---|
| 1275 | 1347 | goto bad_pad_insert; |
|---|
| 1276 | 1348 | } |
|---|
| .. | .. |
|---|
| 1278 | 1350 | return 0; |
|---|
| 1279 | 1351 | |
|---|
| 1280 | 1352 | bad_pad_insert: |
|---|
| 1281 | | - kbase_mmu_teardown_pages(kctx->kbdev, |
|---|
| 1282 | | - &kctx->mmu, |
|---|
| 1283 | | - reg->start_pfn, |
|---|
| 1284 | | - alloc->nents, |
|---|
| 1285 | | - kctx->as_nr); |
|---|
| 1353 | + kbase_mmu_teardown_pages(kctx->kbdev, &kctx->mmu, reg->start_pfn, alloc->pages, |
|---|
| 1354 | + alloc->nents, alloc->nents, kctx->as_nr, true); |
|---|
| 1286 | 1355 | bad_insert: |
|---|
| 1287 | 1356 | kbase_mem_umm_unmap_attachment(kctx, alloc); |
|---|
| 1288 | 1357 | bad_map_attachment: |
|---|
| .. | .. |
|---|
| 1310 | 1379 | if (!kbase_is_region_invalid_or_free(reg) && reg->gpu_alloc == alloc) { |
|---|
| 1311 | 1380 | int err; |
|---|
| 1312 | 1381 | |
|---|
| 1313 | | - err = kbase_mmu_teardown_pages(kctx->kbdev, |
|---|
| 1314 | | - &kctx->mmu, |
|---|
| 1315 | | - reg->start_pfn, |
|---|
| 1316 | | - reg->nr_pages, |
|---|
| 1317 | | - kctx->as_nr); |
|---|
| 1382 | + err = kbase_mmu_teardown_pages(kctx->kbdev, &kctx->mmu, reg->start_pfn, |
|---|
| 1383 | + alloc->pages, reg->nr_pages, reg->nr_pages, |
|---|
| 1384 | + kctx->as_nr, true); |
|---|
| 1318 | 1385 | WARN_ON(err); |
|---|
| 1319 | 1386 | } |
|---|
| 1320 | 1387 | |
|---|
| .. | .. |
|---|
| 1386 | 1453 | return NULL; |
|---|
| 1387 | 1454 | } |
|---|
| 1388 | 1455 | |
|---|
| 1456 | + if (!kbase_import_size_is_valid(kctx->kbdev, *va_pages)) |
|---|
| 1457 | + return NULL; |
|---|
| 1458 | + |
|---|
| 1389 | 1459 | /* ignore SAME_VA */ |
|---|
| 1390 | 1460 | *flags &= ~BASE_MEM_SAME_VA; |
|---|
| 1391 | 1461 | |
|---|
| .. | .. |
|---|
| 1406 | 1476 | if (*flags & BASE_MEM_IMPORT_SYNC_ON_MAP_UNMAP) |
|---|
| 1407 | 1477 | need_sync = true; |
|---|
| 1408 | 1478 | |
|---|
| 1409 | | -#if IS_ENABLED(CONFIG_64BIT) |
|---|
| 1410 | | - if (!kbase_ctx_flag(kctx, KCTX_COMPAT)) { |
|---|
| 1479 | + if (!kbase_ctx_compat_mode(kctx)) { |
|---|
| 1411 | 1480 | /* |
|---|
| 1412 | 1481 | * 64-bit tasks require us to reserve VA on the CPU that we use |
|---|
| 1413 | 1482 | * on the GPU. |
|---|
| 1414 | 1483 | */ |
|---|
| 1415 | 1484 | shared_zone = true; |
|---|
| 1416 | 1485 | } |
|---|
| 1417 | | -#endif |
|---|
| 1418 | 1486 | |
|---|
| 1419 | 1487 | if (shared_zone) { |
|---|
| 1420 | 1488 | *flags |= BASE_MEM_NEED_MMAP; |
|---|
| 1421 | | - reg = kbase_alloc_free_region(&kctx->reg_rbtree_same, |
|---|
| 1422 | | - 0, *va_pages, KBASE_REG_ZONE_SAME_VA); |
|---|
| 1489 | + reg = kbase_alloc_free_region(kctx->kbdev, &kctx->reg_rbtree_same, 0, *va_pages, |
|---|
| 1490 | + KBASE_REG_ZONE_SAME_VA); |
|---|
| 1423 | 1491 | } else { |
|---|
| 1424 | | - reg = kbase_alloc_free_region(&kctx->reg_rbtree_custom, |
|---|
| 1425 | | - 0, *va_pages, KBASE_REG_ZONE_CUSTOM_VA); |
|---|
| 1492 | + reg = kbase_alloc_free_region(kctx->kbdev, &kctx->reg_rbtree_custom, 0, *va_pages, |
|---|
| 1493 | + KBASE_REG_ZONE_CUSTOM_VA); |
|---|
| 1426 | 1494 | } |
|---|
| 1427 | 1495 | |
|---|
| 1428 | 1496 | if (!reg) { |
|---|
| .. | .. |
|---|
| 1507 | 1575 | struct kbase_context *kctx, unsigned long address, |
|---|
| 1508 | 1576 | unsigned long size, u64 *va_pages, u64 *flags) |
|---|
| 1509 | 1577 | { |
|---|
| 1510 | | - long i; |
|---|
| 1578 | + long i, dma_mapped_pages; |
|---|
| 1511 | 1579 | struct kbase_va_region *reg; |
|---|
| 1512 | 1580 | struct rb_root *rbtree; |
|---|
| 1513 | 1581 | long faulted_pages; |
|---|
| .. | .. |
|---|
| 1516 | 1584 | u32 cache_line_alignment = kbase_get_cache_line_alignment(kctx->kbdev); |
|---|
| 1517 | 1585 | struct kbase_alloc_import_user_buf *user_buf; |
|---|
| 1518 | 1586 | struct page **pages = NULL; |
|---|
| 1587 | + struct tagged_addr *pa; |
|---|
| 1588 | + struct device *dev; |
|---|
| 1519 | 1589 | int write; |
|---|
| 1520 | 1590 | |
|---|
| 1521 | 1591 | /* Flag supported only for dma-buf imported memory */ |
|---|
| .. | .. |
|---|
| 1553 | 1623 | /* 64-bit address range is the max */ |
|---|
| 1554 | 1624 | goto bad_size; |
|---|
| 1555 | 1625 | |
|---|
| 1626 | + if (!kbase_import_size_is_valid(kctx->kbdev, *va_pages)) |
|---|
| 1627 | + goto bad_size; |
|---|
| 1628 | + |
|---|
| 1556 | 1629 | /* SAME_VA generally not supported with imported memory (no known use cases) */ |
|---|
| 1557 | 1630 | *flags &= ~BASE_MEM_SAME_VA; |
|---|
| 1558 | 1631 | |
|---|
| 1559 | 1632 | if (*flags & BASE_MEM_IMPORT_SHARED) |
|---|
| 1560 | 1633 | shared_zone = true; |
|---|
| 1561 | 1634 | |
|---|
| 1562 | | -#if IS_ENABLED(CONFIG_64BIT) |
|---|
| 1563 | | - if (!kbase_ctx_flag(kctx, KCTX_COMPAT)) { |
|---|
| 1635 | + if (!kbase_ctx_compat_mode(kctx)) { |
|---|
| 1564 | 1636 | /* |
|---|
| 1565 | 1637 | * 64-bit tasks require us to reserve VA on the CPU that we use |
|---|
| 1566 | 1638 | * on the GPU. |
|---|
| 1567 | 1639 | */ |
|---|
| 1568 | 1640 | shared_zone = true; |
|---|
| 1569 | 1641 | } |
|---|
| 1570 | | -#endif |
|---|
| 1571 | 1642 | |
|---|
| 1572 | 1643 | if (shared_zone) { |
|---|
| 1573 | 1644 | *flags |= BASE_MEM_NEED_MMAP; |
|---|
| .. | .. |
|---|
| 1576 | 1647 | } else |
|---|
| 1577 | 1648 | rbtree = &kctx->reg_rbtree_custom; |
|---|
| 1578 | 1649 | |
|---|
| 1579 | | - reg = kbase_alloc_free_region(rbtree, 0, *va_pages, zone); |
|---|
| 1650 | + reg = kbase_alloc_free_region(kctx->kbdev, rbtree, 0, *va_pages, zone); |
|---|
| 1580 | 1651 | |
|---|
| 1581 | 1652 | if (!reg) |
|---|
| 1582 | 1653 | goto no_region; |
|---|
| .. | .. |
|---|
| 1602 | 1673 | user_buf->address = address; |
|---|
| 1603 | 1674 | user_buf->nr_pages = *va_pages; |
|---|
| 1604 | 1675 | user_buf->mm = current->mm; |
|---|
| 1605 | | -#if KERNEL_VERSION(4, 11, 0) > LINUX_VERSION_CODE |
|---|
| 1606 | | - atomic_inc(¤t->mm->mm_count); |
|---|
| 1607 | | -#else |
|---|
| 1608 | | - mmgrab(current->mm); |
|---|
| 1609 | | -#endif |
|---|
| 1676 | + kbase_mem_mmgrab(); |
|---|
| 1610 | 1677 | if (reg->gpu_alloc->properties & KBASE_MEM_PHY_ALLOC_LARGE) |
|---|
| 1611 | 1678 | user_buf->pages = vmalloc(*va_pages * sizeof(struct page *)); |
|---|
| 1612 | 1679 | else |
|---|
| .. | .. |
|---|
| 1632 | 1699 | |
|---|
| 1633 | 1700 | write = reg->flags & (KBASE_REG_CPU_WR | KBASE_REG_GPU_WR); |
|---|
| 1634 | 1701 | |
|---|
| 1635 | | -#if KERNEL_VERSION(4, 6, 0) > LINUX_VERSION_CODE |
|---|
| 1636 | | - faulted_pages = get_user_pages(current, current->mm, address, *va_pages, |
|---|
| 1637 | | -#if KERNEL_VERSION(4, 4, 168) <= LINUX_VERSION_CODE && \ |
|---|
| 1638 | | -KERNEL_VERSION(4, 5, 0) > LINUX_VERSION_CODE |
|---|
| 1639 | | - write ? FOLL_WRITE : 0, pages, NULL); |
|---|
| 1640 | | -#else |
|---|
| 1641 | | - write, 0, pages, NULL); |
|---|
| 1642 | | -#endif |
|---|
| 1643 | | -#elif KERNEL_VERSION(4, 9, 0) > LINUX_VERSION_CODE |
|---|
| 1644 | | - faulted_pages = get_user_pages(address, *va_pages, |
|---|
| 1645 | | - write, 0, pages, NULL); |
|---|
| 1646 | | -#else |
|---|
| 1702 | +#if KERNEL_VERSION(5, 9, 0) > LINUX_VERSION_CODE |
|---|
| 1647 | 1703 | faulted_pages = get_user_pages(address, *va_pages, |
|---|
| 1648 | 1704 | write ? FOLL_WRITE : 0, pages, NULL); |
|---|
| 1705 | +#else |
|---|
| 1706 | + /* pin_user_pages function cannot be called with pages param NULL. |
|---|
| 1707 | + * get_user_pages function will be used instead because it is safe to be |
|---|
| 1708 | + * used with NULL pages param as long as it doesn't have FOLL_GET flag. |
|---|
| 1709 | + */ |
|---|
| 1710 | + if (pages != NULL) { |
|---|
| 1711 | + faulted_pages = |
|---|
| 1712 | + pin_user_pages(address, *va_pages, write ? FOLL_WRITE : 0, pages, NULL); |
|---|
| 1713 | + } else { |
|---|
| 1714 | + faulted_pages = |
|---|
| 1715 | + get_user_pages(address, *va_pages, write ? FOLL_WRITE : 0, pages, NULL); |
|---|
| 1716 | + } |
|---|
| 1649 | 1717 | #endif |
|---|
| 1650 | 1718 | |
|---|
| 1651 | 1719 | up_read(kbase_mem_get_process_mmap_lock()); |
|---|
| .. | .. |
|---|
| 1656 | 1724 | reg->gpu_alloc->nents = 0; |
|---|
| 1657 | 1725 | reg->extension = 0; |
|---|
| 1658 | 1726 | |
|---|
| 1659 | | - if (pages) { |
|---|
| 1660 | | - struct device *dev = kctx->kbdev->dev; |
|---|
| 1661 | | - unsigned long local_size = user_buf->size; |
|---|
| 1662 | | - unsigned long offset = user_buf->address & ~PAGE_MASK; |
|---|
| 1663 | | - struct tagged_addr *pa = kbase_get_gpu_phy_pages(reg); |
|---|
| 1727 | + pa = kbase_get_gpu_phy_pages(reg); |
|---|
| 1728 | + dev = kctx->kbdev->dev; |
|---|
| 1664 | 1729 | |
|---|
| 1730 | + if (pages) { |
|---|
| 1665 | 1731 | /* Top bit signifies that this was pinned on import */ |
|---|
| 1666 | 1732 | user_buf->current_mapping_usage_count |= PINNED_ON_IMPORT; |
|---|
| 1667 | 1733 | |
|---|
| 1734 | + /* Manual CPU cache synchronization. |
|---|
| 1735 | + * |
|---|
| 1736 | + * The driver disables automatic CPU cache synchronization because the |
|---|
| 1737 | + * memory pages that enclose the imported region may also contain |
|---|
| 1738 | + * sub-regions which are not imported and that are allocated and used |
|---|
| 1739 | + * by the user process. This may be the case of memory at the beginning |
|---|
| 1740 | + * of the first page and at the end of the last page. Automatic CPU cache |
|---|
| 1741 | + * synchronization would force some operations on those memory allocations, |
|---|
| 1742 | + * unbeknown to the user process: in particular, a CPU cache invalidate |
|---|
| 1743 | + * upon unmapping would destroy the content of dirty CPU caches and cause |
|---|
| 1744 | + * the user process to lose CPU writes to the non-imported sub-regions. |
|---|
| 1745 | + * |
|---|
| 1746 | + * When the GPU claims ownership of the imported memory buffer, it shall |
|---|
| 1747 | + * commit CPU writes for the whole of all pages that enclose the imported |
|---|
| 1748 | + * region, otherwise the initial content of memory would be wrong. |
|---|
| 1749 | + */ |
|---|
| 1668 | 1750 | for (i = 0; i < faulted_pages; i++) { |
|---|
| 1669 | 1751 | dma_addr_t dma_addr; |
|---|
| 1670 | | - unsigned long min; |
|---|
| 1671 | | - |
|---|
| 1672 | | - min = MIN(PAGE_SIZE - offset, local_size); |
|---|
| 1673 | | - dma_addr = dma_map_page(dev, pages[i], |
|---|
| 1674 | | - offset, min, |
|---|
| 1675 | | - DMA_BIDIRECTIONAL); |
|---|
| 1752 | +#if (KERNEL_VERSION(4, 10, 0) > LINUX_VERSION_CODE) |
|---|
| 1753 | + dma_addr = dma_map_page(dev, pages[i], 0, PAGE_SIZE, DMA_BIDIRECTIONAL); |
|---|
| 1754 | +#else |
|---|
| 1755 | + dma_addr = dma_map_page_attrs(dev, pages[i], 0, PAGE_SIZE, |
|---|
| 1756 | + DMA_BIDIRECTIONAL, DMA_ATTR_SKIP_CPU_SYNC); |
|---|
| 1757 | +#endif |
|---|
| 1676 | 1758 | if (dma_mapping_error(dev, dma_addr)) |
|---|
| 1677 | 1759 | goto unwind_dma_map; |
|---|
| 1678 | 1760 | |
|---|
| 1679 | 1761 | user_buf->dma_addrs[i] = dma_addr; |
|---|
| 1680 | 1762 | pa[i] = as_tagged(page_to_phys(pages[i])); |
|---|
| 1681 | 1763 | |
|---|
| 1682 | | - local_size -= min; |
|---|
| 1683 | | - offset = 0; |
|---|
| 1764 | + dma_sync_single_for_device(dev, dma_addr, PAGE_SIZE, DMA_BIDIRECTIONAL); |
|---|
| 1684 | 1765 | } |
|---|
| 1685 | 1766 | |
|---|
| 1686 | 1767 | reg->gpu_alloc->nents = faulted_pages; |
|---|
| .. | .. |
|---|
| 1689 | 1770 | return reg; |
|---|
| 1690 | 1771 | |
|---|
| 1691 | 1772 | unwind_dma_map: |
|---|
| 1692 | | - while (i--) { |
|---|
| 1693 | | - dma_unmap_page(kctx->kbdev->dev, |
|---|
| 1694 | | - user_buf->dma_addrs[i], |
|---|
| 1695 | | - PAGE_SIZE, DMA_BIDIRECTIONAL); |
|---|
| 1773 | + dma_mapped_pages = i; |
|---|
| 1774 | + /* Run the unmap loop in the same order as map loop, and perform again |
|---|
| 1775 | + * CPU cache synchronization to re-write the content of dirty CPU caches |
|---|
| 1776 | + * to memory. This precautionary measure is kept here to keep this code |
|---|
| 1777 | + * aligned with kbase_jd_user_buf_map() to allow for a potential refactor |
|---|
| 1778 | + * in the future. |
|---|
| 1779 | + */ |
|---|
| 1780 | + for (i = 0; i < dma_mapped_pages; i++) { |
|---|
| 1781 | + dma_addr_t dma_addr = user_buf->dma_addrs[i]; |
|---|
| 1782 | + |
|---|
| 1783 | + dma_sync_single_for_device(dev, dma_addr, PAGE_SIZE, DMA_BIDIRECTIONAL); |
|---|
| 1784 | +#if (KERNEL_VERSION(4, 10, 0) > LINUX_VERSION_CODE) |
|---|
| 1785 | + dma_unmap_page(dev, dma_addr, PAGE_SIZE, DMA_BIDIRECTIONAL); |
|---|
| 1786 | +#else |
|---|
| 1787 | + dma_unmap_page_attrs(dev, dma_addr, PAGE_SIZE, DMA_BIDIRECTIONAL, |
|---|
| 1788 | + DMA_ATTR_SKIP_CPU_SYNC); |
|---|
| 1789 | +#endif |
|---|
| 1696 | 1790 | } |
|---|
| 1697 | 1791 | fault_mismatch: |
|---|
| 1698 | 1792 | if (pages) { |
|---|
| 1793 | + /* In this case, the region was not yet in the region tracker, |
|---|
| 1794 | + * and so there are no CPU mappings to remove before we unpin |
|---|
| 1795 | + * the page |
|---|
| 1796 | + */ |
|---|
| 1699 | 1797 | for (i = 0; i < faulted_pages; i++) |
|---|
| 1700 | | - put_page(pages[i]); |
|---|
| 1798 | + kbase_unpin_user_buf_page(pages[i]); |
|---|
| 1701 | 1799 | } |
|---|
| 1702 | 1800 | no_page_array: |
|---|
| 1703 | 1801 | invalid_flags: |
|---|
| .. | .. |
|---|
| 1708 | 1806 | no_region: |
|---|
| 1709 | 1807 | bad_size: |
|---|
| 1710 | 1808 | return NULL; |
|---|
| 1711 | | - |
|---|
| 1712 | 1809 | } |
|---|
| 1713 | 1810 | |
|---|
| 1714 | 1811 | |
|---|
| .. | .. |
|---|
| 1720 | 1817 | u64 gpu_va; |
|---|
| 1721 | 1818 | size_t i; |
|---|
| 1722 | 1819 | bool coherent; |
|---|
| 1820 | + uint64_t max_stride; |
|---|
| 1821 | + |
|---|
| 1822 | + /* Calls to this function are inherently asynchronous, with respect to |
|---|
| 1823 | + * MMU operations. |
|---|
| 1824 | + */ |
|---|
| 1825 | + const enum kbase_caller_mmu_sync_info mmu_sync_info = CALLER_MMU_ASYNC; |
|---|
| 1723 | 1826 | |
|---|
| 1724 | 1827 | KBASE_DEBUG_ASSERT(kctx); |
|---|
| 1725 | 1828 | KBASE_DEBUG_ASSERT(flags); |
|---|
| .. | .. |
|---|
| 1733 | 1836 | |
|---|
| 1734 | 1837 | if (!(*flags & (BASE_MEM_PROT_GPU_RD | BASE_MEM_PROT_GPU_WR))) { |
|---|
| 1735 | 1838 | dev_warn(kctx->kbdev->dev, |
|---|
| 1736 | | - "kbase_mem_alias called with bad flags (%llx)", |
|---|
| 1839 | + "%s called with bad flags (%llx)", |
|---|
| 1840 | + __func__, |
|---|
| 1737 | 1841 | (unsigned long long)*flags); |
|---|
| 1738 | 1842 | goto bad_flags; |
|---|
| 1739 | 1843 | } |
|---|
| .. | .. |
|---|
| 1746 | 1850 | if (!nents) |
|---|
| 1747 | 1851 | goto bad_nents; |
|---|
| 1748 | 1852 | |
|---|
| 1853 | + max_stride = div64_u64(U64_MAX, nents); |
|---|
| 1854 | + |
|---|
| 1855 | + if (stride > max_stride) |
|---|
| 1856 | + goto bad_size; |
|---|
| 1857 | + |
|---|
| 1749 | 1858 | if ((nents * stride) > (U64_MAX / PAGE_SIZE)) |
|---|
| 1750 | 1859 | /* 64-bit address range is the max */ |
|---|
| 1751 | 1860 | goto bad_size; |
|---|
| .. | .. |
|---|
| 1753 | 1862 | /* calculate the number of pages this alias will cover */ |
|---|
| 1754 | 1863 | *num_pages = nents * stride; |
|---|
| 1755 | 1864 | |
|---|
| 1756 | | -#if IS_ENABLED(CONFIG_64BIT) |
|---|
| 1757 | | - if (!kbase_ctx_flag(kctx, KCTX_COMPAT)) { |
|---|
| 1865 | + if (!kbase_alias_size_is_valid(kctx->kbdev, *num_pages)) |
|---|
| 1866 | + goto bad_size; |
|---|
| 1867 | + |
|---|
| 1868 | + if (!kbase_ctx_compat_mode(kctx)) { |
|---|
| 1758 | 1869 | /* 64-bit tasks must MMAP anyway, but not expose this address to |
|---|
| 1759 | 1870 | * clients |
|---|
| 1760 | 1871 | */ |
|---|
| 1761 | 1872 | *flags |= BASE_MEM_NEED_MMAP; |
|---|
| 1762 | | - reg = kbase_alloc_free_region(&kctx->reg_rbtree_same, 0, |
|---|
| 1763 | | - *num_pages, |
|---|
| 1764 | | - KBASE_REG_ZONE_SAME_VA); |
|---|
| 1873 | + reg = kbase_alloc_free_region(kctx->kbdev, &kctx->reg_rbtree_same, 0, *num_pages, |
|---|
| 1874 | + KBASE_REG_ZONE_SAME_VA); |
|---|
| 1765 | 1875 | } else { |
|---|
| 1766 | | -#else |
|---|
| 1767 | | - if (1) { |
|---|
| 1768 | | -#endif |
|---|
| 1769 | | - reg = kbase_alloc_free_region(&kctx->reg_rbtree_custom, |
|---|
| 1770 | | - 0, *num_pages, |
|---|
| 1771 | | - KBASE_REG_ZONE_CUSTOM_VA); |
|---|
| 1876 | + reg = kbase_alloc_free_region(kctx->kbdev, &kctx->reg_rbtree_custom, 0, *num_pages, |
|---|
| 1877 | + KBASE_REG_ZONE_CUSTOM_VA); |
|---|
| 1772 | 1878 | } |
|---|
| 1773 | 1879 | |
|---|
| 1774 | 1880 | if (!reg) |
|---|
| .. | .. |
|---|
| 1817 | 1923 | /* validate found region */ |
|---|
| 1818 | 1924 | if (kbase_is_region_invalid_or_free(aliasing_reg)) |
|---|
| 1819 | 1925 | goto bad_handle; /* Not found/already free */ |
|---|
| 1820 | | - if (aliasing_reg->flags & KBASE_REG_DONT_NEED) |
|---|
| 1926 | + if (kbase_is_region_shrinkable(aliasing_reg)) |
|---|
| 1821 | 1927 | goto bad_handle; /* Ephemeral region */ |
|---|
| 1822 | | - if (aliasing_reg->flags & KBASE_REG_NO_USER_FREE) |
|---|
| 1928 | + if (kbase_va_region_is_no_user_free(aliasing_reg)) |
|---|
| 1823 | 1929 | goto bad_handle; /* JIT regions can't be |
|---|
| 1824 | 1930 | * aliased. NO_USER_FREE flag |
|---|
| 1825 | 1931 | * covers the entire lifetime |
|---|
| .. | .. |
|---|
| 1874 | 1980 | } |
|---|
| 1875 | 1981 | } |
|---|
| 1876 | 1982 | |
|---|
| 1877 | | -#if IS_ENABLED(CONFIG_64BIT) |
|---|
| 1878 | | - if (!kbase_ctx_flag(kctx, KCTX_COMPAT)) { |
|---|
| 1983 | + if (!kbase_ctx_compat_mode(kctx)) { |
|---|
| 1879 | 1984 | /* Bind to a cookie */ |
|---|
| 1880 | 1985 | if (bitmap_empty(kctx->cookies, BITS_PER_LONG)) { |
|---|
| 1881 | 1986 | dev_err(kctx->kbdev->dev, "No cookies available for allocation!"); |
|---|
| .. | .. |
|---|
| 1890 | 1995 | /* relocate to correct base */ |
|---|
| 1891 | 1996 | gpu_va += PFN_DOWN(BASE_MEM_COOKIE_BASE); |
|---|
| 1892 | 1997 | gpu_va <<= PAGE_SHIFT; |
|---|
| 1893 | | - } else /* we control the VA */ { |
|---|
| 1894 | | -#else |
|---|
| 1895 | | - if (1) { |
|---|
| 1896 | | -#endif |
|---|
| 1897 | | - if (kbase_gpu_mmap(kctx, reg, 0, *num_pages, 1) != 0) { |
|---|
| 1998 | + } else { |
|---|
| 1999 | + /* we control the VA */ |
|---|
| 2000 | + if (kbase_gpu_mmap(kctx, reg, 0, *num_pages, 1, |
|---|
| 2001 | + mmu_sync_info) != 0) { |
|---|
| 1898 | 2002 | dev_warn(kctx->kbdev->dev, "Failed to map memory on GPU"); |
|---|
| 1899 | 2003 | goto no_mmap; |
|---|
| 1900 | 2004 | } |
|---|
| .. | .. |
|---|
| 1909 | 2013 | |
|---|
| 1910 | 2014 | return gpu_va; |
|---|
| 1911 | 2015 | |
|---|
| 1912 | | -#if IS_ENABLED(CONFIG_64BIT) |
|---|
| 1913 | 2016 | no_cookie: |
|---|
| 1914 | | -#endif |
|---|
| 1915 | 2017 | no_mmap: |
|---|
| 1916 | 2018 | bad_handle: |
|---|
| 1917 | 2019 | /* Marking the source allocs as not being mapped on the GPU and putting |
|---|
| .. | .. |
|---|
| 1939 | 2041 | { |
|---|
| 1940 | 2042 | struct kbase_va_region *reg; |
|---|
| 1941 | 2043 | |
|---|
| 2044 | + /* Calls to this function are inherently asynchronous, with respect to |
|---|
| 2045 | + * MMU operations. |
|---|
| 2046 | + */ |
|---|
| 2047 | + const enum kbase_caller_mmu_sync_info mmu_sync_info = CALLER_MMU_ASYNC; |
|---|
| 2048 | + |
|---|
| 1942 | 2049 | KBASE_DEBUG_ASSERT(kctx); |
|---|
| 1943 | 2050 | KBASE_DEBUG_ASSERT(gpu_va); |
|---|
| 1944 | 2051 | KBASE_DEBUG_ASSERT(va_pages); |
|---|
| .. | .. |
|---|
| 1950 | 2057 | |
|---|
| 1951 | 2058 | if (!kbase_check_import_flags(*flags)) { |
|---|
| 1952 | 2059 | dev_warn(kctx->kbdev->dev, |
|---|
| 1953 | | - "kbase_mem_import called with bad flags (%llx)", |
|---|
| 2060 | + "%s called with bad flags (%llx)", |
|---|
| 2061 | + __func__, |
|---|
| 1954 | 2062 | (unsigned long long)*flags); |
|---|
| 1955 | 2063 | goto bad_flags; |
|---|
| 1956 | 2064 | } |
|---|
| .. | .. |
|---|
| 1963 | 2071 | if ((*flags & BASE_MEM_COHERENT_SYSTEM_REQUIRED) != 0 && |
|---|
| 1964 | 2072 | !kbase_device_is_cpu_coherent(kctx->kbdev)) { |
|---|
| 1965 | 2073 | dev_warn(kctx->kbdev->dev, |
|---|
| 1966 | | - "kbase_mem_import call required coherent mem when unavailable"); |
|---|
| 2074 | + "%s call required coherent mem when unavailable", |
|---|
| 2075 | + __func__); |
|---|
| 1967 | 2076 | goto bad_flags; |
|---|
| 1968 | 2077 | } |
|---|
| 1969 | 2078 | if ((*flags & BASE_MEM_COHERENT_SYSTEM) != 0 && |
|---|
| .. | .. |
|---|
| 1971 | 2080 | /* Remove COHERENT_SYSTEM flag if coherent mem is unavailable */ |
|---|
| 1972 | 2081 | *flags &= ~BASE_MEM_COHERENT_SYSTEM; |
|---|
| 1973 | 2082 | } |
|---|
| 1974 | | - |
|---|
| 2083 | + if (((*flags & BASE_MEM_CACHED_CPU) == 0) && (type == BASE_MEM_IMPORT_TYPE_USER_BUFFER)) { |
|---|
| 2084 | + dev_warn(kctx->kbdev->dev, "USER_BUFFER must be CPU cached"); |
|---|
| 2085 | + goto bad_flags; |
|---|
| 2086 | + } |
|---|
| 1975 | 2087 | if ((padding != 0) && (type != BASE_MEM_IMPORT_TYPE_UMM)) { |
|---|
| 1976 | 2088 | dev_warn(kctx->kbdev->dev, |
|---|
| 1977 | 2089 | "padding is only supported for UMM"); |
|---|
| .. | .. |
|---|
| 2038 | 2150 | |
|---|
| 2039 | 2151 | } else if (*flags & KBASE_MEM_IMPORT_HAVE_PAGES) { |
|---|
| 2040 | 2152 | /* we control the VA, mmap now to the GPU */ |
|---|
| 2041 | | - if (kbase_gpu_mmap(kctx, reg, 0, *va_pages, 1) != 0) |
|---|
| 2153 | + if (kbase_gpu_mmap(kctx, reg, 0, *va_pages, 1, mmu_sync_info) != |
|---|
| 2154 | + 0) |
|---|
| 2042 | 2155 | goto no_gpu_va; |
|---|
| 2043 | 2156 | /* return real GPU VA */ |
|---|
| 2044 | 2157 | *gpu_va = reg->start_pfn << PAGE_SHIFT; |
|---|
| .. | .. |
|---|
| 2072 | 2185 | } |
|---|
| 2073 | 2186 | |
|---|
| 2074 | 2187 | int kbase_mem_grow_gpu_mapping(struct kbase_context *kctx, |
|---|
| 2075 | | - struct kbase_va_region *reg, |
|---|
| 2076 | | - u64 new_pages, u64 old_pages) |
|---|
| 2188 | + struct kbase_va_region *reg, u64 new_pages, |
|---|
| 2189 | + u64 old_pages, |
|---|
| 2190 | + enum kbase_caller_mmu_sync_info mmu_sync_info) |
|---|
| 2077 | 2191 | { |
|---|
| 2078 | 2192 | struct tagged_addr *phy_pages; |
|---|
| 2079 | 2193 | u64 delta = new_pages - old_pages; |
|---|
| .. | .. |
|---|
| 2083 | 2197 | |
|---|
| 2084 | 2198 | /* Map the new pages into the GPU */ |
|---|
| 2085 | 2199 | phy_pages = kbase_get_gpu_phy_pages(reg); |
|---|
| 2086 | | - ret = kbase_mmu_insert_pages(kctx->kbdev, &kctx->mmu, |
|---|
| 2087 | | - reg->start_pfn + old_pages, phy_pages + old_pages, delta, |
|---|
| 2088 | | - reg->flags, kctx->as_nr, reg->gpu_alloc->group_id); |
|---|
| 2200 | + ret = kbase_mmu_insert_pages(kctx->kbdev, &kctx->mmu, reg->start_pfn + old_pages, |
|---|
| 2201 | + phy_pages + old_pages, delta, reg->flags, kctx->as_nr, |
|---|
| 2202 | + reg->gpu_alloc->group_id, mmu_sync_info, reg, false); |
|---|
| 2089 | 2203 | |
|---|
| 2090 | 2204 | return ret; |
|---|
| 2091 | 2205 | } |
|---|
| .. | .. |
|---|
| 2105 | 2219 | (old_pages - new_pages)<<PAGE_SHIFT, 1); |
|---|
| 2106 | 2220 | } |
|---|
| 2107 | 2221 | |
|---|
| 2108 | | -/** |
|---|
| 2109 | | - * kbase_mem_shrink_gpu_mapping - Shrink the GPU mapping of an allocation |
|---|
| 2110 | | - * @kctx: Context the region belongs to |
|---|
| 2111 | | - * @reg: The GPU region or NULL if there isn't one |
|---|
| 2112 | | - * @new_pages: The number of pages after the shrink |
|---|
| 2113 | | - * @old_pages: The number of pages before the shrink |
|---|
| 2114 | | - * |
|---|
| 2115 | | - * Return: 0 on success, negative -errno on error |
|---|
| 2116 | | - * |
|---|
| 2117 | | - * Unmap the shrunk pages from the GPU mapping. Note that the size of the region |
|---|
| 2118 | | - * itself is unmodified as we still need to reserve the VA, only the page tables |
|---|
| 2119 | | - * will be modified by this function. |
|---|
| 2120 | | - */ |
|---|
| 2121 | | -static int kbase_mem_shrink_gpu_mapping(struct kbase_context *const kctx, |
|---|
| 2122 | | - struct kbase_va_region *const reg, |
|---|
| 2123 | | - u64 const new_pages, u64 const old_pages) |
|---|
| 2222 | +int kbase_mem_shrink_gpu_mapping(struct kbase_context *const kctx, |
|---|
| 2223 | + struct kbase_va_region *const reg, u64 const new_pages, |
|---|
| 2224 | + u64 const old_pages) |
|---|
| 2124 | 2225 | { |
|---|
| 2125 | 2226 | u64 delta = old_pages - new_pages; |
|---|
| 2227 | + struct kbase_mem_phy_alloc *alloc = reg->gpu_alloc; |
|---|
| 2126 | 2228 | int ret = 0; |
|---|
| 2127 | 2229 | |
|---|
| 2128 | | - ret = kbase_mmu_teardown_pages(kctx->kbdev, &kctx->mmu, |
|---|
| 2129 | | - reg->start_pfn + new_pages, delta, kctx->as_nr); |
|---|
| 2230 | + ret = kbase_mmu_teardown_pages(kctx->kbdev, &kctx->mmu, reg->start_pfn + new_pages, |
|---|
| 2231 | + alloc->pages + new_pages, delta, delta, kctx->as_nr, false); |
|---|
| 2130 | 2232 | |
|---|
| 2131 | 2233 | return ret; |
|---|
| 2132 | 2234 | } |
|---|
| .. | .. |
|---|
| 2138 | 2240 | int res = -EINVAL; |
|---|
| 2139 | 2241 | struct kbase_va_region *reg; |
|---|
| 2140 | 2242 | bool read_locked = false; |
|---|
| 2243 | + |
|---|
| 2244 | + /* Calls to this function are inherently asynchronous, with respect to |
|---|
| 2245 | + * MMU operations. |
|---|
| 2246 | + */ |
|---|
| 2247 | + const enum kbase_caller_mmu_sync_info mmu_sync_info = CALLER_MMU_ASYNC; |
|---|
| 2141 | 2248 | |
|---|
| 2142 | 2249 | KBASE_DEBUG_ASSERT(kctx); |
|---|
| 2143 | 2250 | KBASE_DEBUG_ASSERT(gpu_addr != 0); |
|---|
| .. | .. |
|---|
| 2185 | 2292 | |
|---|
| 2186 | 2293 | if (atomic_read(®->cpu_alloc->kernel_mappings) > 0) |
|---|
| 2187 | 2294 | goto out_unlock; |
|---|
| 2188 | | - /* can't grow regions which are ephemeral */ |
|---|
| 2189 | | - if (reg->flags & KBASE_REG_DONT_NEED) |
|---|
| 2295 | + |
|---|
| 2296 | + if (kbase_is_region_shrinkable(reg)) |
|---|
| 2297 | + goto out_unlock; |
|---|
| 2298 | + |
|---|
| 2299 | + if (kbase_va_region_is_no_user_free(reg)) |
|---|
| 2190 | 2300 | goto out_unlock; |
|---|
| 2191 | 2301 | |
|---|
| 2192 | 2302 | #ifdef CONFIG_MALI_MEMORY_FULLY_BACKED |
|---|
| .. | .. |
|---|
| 2230 | 2340 | /* No update required for CPU mappings, that's done on fault. */ |
|---|
| 2231 | 2341 | |
|---|
| 2232 | 2342 | /* Update GPU mapping. */ |
|---|
| 2233 | | - res = kbase_mem_grow_gpu_mapping(kctx, reg, |
|---|
| 2234 | | - new_pages, old_pages); |
|---|
| 2343 | + res = kbase_mem_grow_gpu_mapping(kctx, reg, new_pages, |
|---|
| 2344 | + old_pages, mmu_sync_info); |
|---|
| 2235 | 2345 | |
|---|
| 2236 | 2346 | /* On error free the new pages */ |
|---|
| 2237 | 2347 | if (res) { |
|---|
| .. | .. |
|---|
| 2259 | 2369 | } |
|---|
| 2260 | 2370 | |
|---|
| 2261 | 2371 | int kbase_mem_shrink(struct kbase_context *const kctx, |
|---|
| 2262 | | - struct kbase_va_region *const reg, u64 const new_pages) |
|---|
| 2372 | + struct kbase_va_region *const reg, u64 new_pages) |
|---|
| 2263 | 2373 | { |
|---|
| 2264 | 2374 | u64 delta, old_pages; |
|---|
| 2265 | 2375 | int err; |
|---|
| .. | .. |
|---|
| 2289 | 2399 | kbase_free_phy_pages_helper(reg->cpu_alloc, delta); |
|---|
| 2290 | 2400 | if (reg->cpu_alloc != reg->gpu_alloc) |
|---|
| 2291 | 2401 | kbase_free_phy_pages_helper(reg->gpu_alloc, delta); |
|---|
| 2402 | + |
|---|
| 2403 | + if (kctx->kbdev->pagesize_2mb) { |
|---|
| 2404 | + if (kbase_reg_current_backed_size(reg) > new_pages) { |
|---|
| 2405 | + old_pages = new_pages; |
|---|
| 2406 | + new_pages = kbase_reg_current_backed_size(reg); |
|---|
| 2407 | + |
|---|
| 2408 | + /* Update GPU mapping. */ |
|---|
| 2409 | + err = kbase_mem_grow_gpu_mapping(kctx, reg, new_pages, old_pages, |
|---|
| 2410 | + CALLER_MMU_ASYNC); |
|---|
| 2411 | + } |
|---|
| 2412 | + } else { |
|---|
| 2413 | + WARN_ON(kbase_reg_current_backed_size(reg) != new_pages); |
|---|
| 2414 | + } |
|---|
| 2292 | 2415 | } |
|---|
| 2293 | 2416 | |
|---|
| 2294 | 2417 | return err; |
|---|
| .. | .. |
|---|
| 2327 | 2450 | /* Avoid freeing memory on the process death which results in |
|---|
| 2328 | 2451 | * GPU Page Fault. Memory will be freed in kbase_destroy_context |
|---|
| 2329 | 2452 | */ |
|---|
| 2330 | | - if (!(current->flags & PF_EXITING)) |
|---|
| 2453 | + if (!is_process_exiting(vma)) |
|---|
| 2331 | 2454 | kbase_mem_free_region(map->kctx, map->region); |
|---|
| 2332 | 2455 | } |
|---|
| 2333 | 2456 | |
|---|
| .. | .. |
|---|
| 2559 | 2682 | while (kbase_jit_evict(kctx)) |
|---|
| 2560 | 2683 | ; |
|---|
| 2561 | 2684 | } |
|---|
| 2562 | | -#endif |
|---|
| 2563 | 2685 | |
|---|
| 2564 | 2686 | static int kbase_mmu_dump_mmap(struct kbase_context *kctx, |
|---|
| 2565 | 2687 | struct vm_area_struct *vma, |
|---|
| .. | .. |
|---|
| 2572 | 2694 | size_t size; |
|---|
| 2573 | 2695 | int err = 0; |
|---|
| 2574 | 2696 | |
|---|
| 2575 | | - dev_dbg(kctx->kbdev->dev, "in kbase_mmu_dump_mmap\n"); |
|---|
| 2697 | + lockdep_assert_held(&kctx->reg_lock); |
|---|
| 2698 | + |
|---|
| 2699 | + dev_dbg(kctx->kbdev->dev, "%s\n", __func__); |
|---|
| 2576 | 2700 | size = (vma->vm_end - vma->vm_start); |
|---|
| 2577 | 2701 | nr_pages = size >> PAGE_SHIFT; |
|---|
| 2578 | 2702 | |
|---|
| 2579 | | -#ifdef CONFIG_MALI_VECTOR_DUMP |
|---|
| 2580 | 2703 | kbase_free_unused_jit_allocations(kctx); |
|---|
| 2581 | | -#endif |
|---|
| 2582 | 2704 | |
|---|
| 2583 | 2705 | kaddr = kbase_mmu_dump(kctx, nr_pages); |
|---|
| 2584 | 2706 | |
|---|
| .. | .. |
|---|
| 2587 | 2709 | goto out; |
|---|
| 2588 | 2710 | } |
|---|
| 2589 | 2711 | |
|---|
| 2590 | | - new_reg = kbase_alloc_free_region(&kctx->reg_rbtree_same, 0, nr_pages, |
|---|
| 2591 | | - KBASE_REG_ZONE_SAME_VA); |
|---|
| 2712 | + new_reg = kbase_alloc_free_region(kctx->kbdev, &kctx->reg_rbtree_same, 0, nr_pages, |
|---|
| 2713 | + KBASE_REG_ZONE_SAME_VA); |
|---|
| 2592 | 2714 | if (!new_reg) { |
|---|
| 2593 | 2715 | err = -ENOMEM; |
|---|
| 2594 | 2716 | WARN_ON(1); |
|---|
| .. | .. |
|---|
| 2617 | 2739 | *kmap_addr = kaddr; |
|---|
| 2618 | 2740 | *reg = new_reg; |
|---|
| 2619 | 2741 | |
|---|
| 2620 | | - dev_dbg(kctx->kbdev->dev, "kbase_mmu_dump_mmap done\n"); |
|---|
| 2742 | + dev_dbg(kctx->kbdev->dev, "%s done\n", __func__); |
|---|
| 2621 | 2743 | return 0; |
|---|
| 2622 | 2744 | |
|---|
| 2623 | 2745 | out_no_alloc: |
|---|
| .. | .. |
|---|
| 2626 | 2748 | out: |
|---|
| 2627 | 2749 | return err; |
|---|
| 2628 | 2750 | } |
|---|
| 2629 | | - |
|---|
| 2751 | +#endif |
|---|
| 2630 | 2752 | |
|---|
| 2631 | 2753 | void kbase_os_mem_map_lock(struct kbase_context *kctx) |
|---|
| 2632 | 2754 | { |
|---|
| .. | .. |
|---|
| 2646 | 2768 | size_t *nr_pages, size_t *aligned_offset) |
|---|
| 2647 | 2769 | |
|---|
| 2648 | 2770 | { |
|---|
| 2649 | | - int cookie = vma->vm_pgoff - PFN_DOWN(BASE_MEM_COOKIE_BASE); |
|---|
| 2771 | + unsigned int cookie = vma->vm_pgoff - PFN_DOWN(BASE_MEM_COOKIE_BASE); |
|---|
| 2650 | 2772 | struct kbase_va_region *reg; |
|---|
| 2651 | 2773 | int err = 0; |
|---|
| 2652 | 2774 | |
|---|
| 2775 | + /* Calls to this function are inherently asynchronous, with respect to |
|---|
| 2776 | + * MMU operations. |
|---|
| 2777 | + */ |
|---|
| 2778 | + const enum kbase_caller_mmu_sync_info mmu_sync_info = CALLER_MMU_ASYNC; |
|---|
| 2779 | + |
|---|
| 2653 | 2780 | *aligned_offset = 0; |
|---|
| 2654 | 2781 | |
|---|
| 2655 | | - dev_dbg(kctx->kbdev->dev, "in kbasep_reg_mmap\n"); |
|---|
| 2782 | + dev_dbg(kctx->kbdev->dev, "%s\n", __func__); |
|---|
| 2656 | 2783 | |
|---|
| 2657 | 2784 | /* SAME_VA stuff, fetch the right region */ |
|---|
| 2658 | 2785 | reg = kctx->pending_regions[cookie]; |
|---|
| .. | .. |
|---|
| 2682 | 2809 | |
|---|
| 2683 | 2810 | /* adjust down nr_pages to what we have physically */ |
|---|
| 2684 | 2811 | *nr_pages = kbase_reg_current_backed_size(reg); |
|---|
| 2685 | | - |
|---|
| 2686 | 2812 | if (kbase_gpu_mmap(kctx, reg, vma->vm_start + *aligned_offset, |
|---|
| 2687 | | - reg->nr_pages, 1) != 0) { |
|---|
| 2813 | + reg->nr_pages, 1, mmu_sync_info) != 0) { |
|---|
| 2688 | 2814 | dev_err(kctx->kbdev->dev, "%s:%d\n", __FILE__, __LINE__); |
|---|
| 2689 | 2815 | /* Unable to map in GPU space. */ |
|---|
| 2690 | 2816 | WARN_ON(1); |
|---|
| .. | .. |
|---|
| 2709 | 2835 | vma->vm_pgoff = reg->start_pfn - ((*aligned_offset)>>PAGE_SHIFT); |
|---|
| 2710 | 2836 | out: |
|---|
| 2711 | 2837 | *regm = reg; |
|---|
| 2712 | | - dev_dbg(kctx->kbdev->dev, "kbasep_reg_mmap done\n"); |
|---|
| 2838 | + dev_dbg(kctx->kbdev->dev, "%s done\n", __func__); |
|---|
| 2713 | 2839 | |
|---|
| 2714 | 2840 | return err; |
|---|
| 2715 | 2841 | } |
|---|
| .. | .. |
|---|
| 2750 | 2876 | goto out_unlock; |
|---|
| 2751 | 2877 | } |
|---|
| 2752 | 2878 | |
|---|
| 2753 | | - /* if not the MTP, verify that the MTP has been mapped */ |
|---|
| 2754 | | - rcu_read_lock(); |
|---|
| 2755 | | - /* catches both when the special page isn't present or |
|---|
| 2756 | | - * when we've forked |
|---|
| 2757 | | - */ |
|---|
| 2758 | | - if (rcu_dereference(kctx->process_mm) != current->mm) { |
|---|
| 2879 | + if (!kbase_mem_allow_alloc(kctx)) { |
|---|
| 2759 | 2880 | err = -EINVAL; |
|---|
| 2760 | | - rcu_read_unlock(); |
|---|
| 2761 | 2881 | goto out_unlock; |
|---|
| 2762 | 2882 | } |
|---|
| 2763 | | - rcu_read_unlock(); |
|---|
| 2764 | 2883 | |
|---|
| 2765 | 2884 | switch (vma->vm_pgoff) { |
|---|
| 2766 | 2885 | case PFN_DOWN(BASEP_MEM_INVALID_HANDLE): |
|---|
| .. | .. |
|---|
| 2769 | 2888 | err = -EINVAL; |
|---|
| 2770 | 2889 | goto out_unlock; |
|---|
| 2771 | 2890 | case PFN_DOWN(BASE_MEM_MMU_DUMP_HANDLE): |
|---|
| 2891 | +#if defined(CONFIG_MALI_VECTOR_DUMP) |
|---|
| 2772 | 2892 | /* MMU dump */ |
|---|
| 2773 | 2893 | err = kbase_mmu_dump_mmap(kctx, vma, ®, &kaddr); |
|---|
| 2774 | 2894 | if (err != 0) |
|---|
| .. | .. |
|---|
| 2776 | 2896 | /* free the region on munmap */ |
|---|
| 2777 | 2897 | free_on_close = 1; |
|---|
| 2778 | 2898 | break; |
|---|
| 2899 | +#else |
|---|
| 2900 | + /* Illegal handle for direct map */ |
|---|
| 2901 | + err = -EINVAL; |
|---|
| 2902 | + goto out_unlock; |
|---|
| 2903 | +#endif /* defined(CONFIG_MALI_VECTOR_DUMP) */ |
|---|
| 2779 | 2904 | #if MALI_USE_CSF |
|---|
| 2780 | 2905 | case PFN_DOWN(BASEP_MEM_CSF_USER_REG_PAGE_HANDLE): |
|---|
| 2781 | 2906 | kbase_gpu_vm_unlock(kctx); |
|---|
| .. | .. |
|---|
| 2846 | 2971 | dev_warn(dev, "mmap aliased: invalid params!\n"); |
|---|
| 2847 | 2972 | goto out_unlock; |
|---|
| 2848 | 2973 | } |
|---|
| 2849 | | - } |
|---|
| 2850 | | - else if (reg->cpu_alloc->nents < |
|---|
| 2974 | + } else if (reg->cpu_alloc->nents < |
|---|
| 2851 | 2975 | (vma->vm_pgoff - reg->start_pfn + nr_pages)) { |
|---|
| 2852 | 2976 | /* limit what we map to the amount currently backed */ |
|---|
| 2853 | 2977 | if ((vma->vm_pgoff - reg->start_pfn) >= reg->cpu_alloc->nents) |
|---|
| .. | .. |
|---|
| 2864 | 2988 | |
|---|
| 2865 | 2989 | err = kbase_cpu_mmap(kctx, reg, vma, kaddr, nr_pages, aligned_offset, |
|---|
| 2866 | 2990 | free_on_close); |
|---|
| 2867 | | - |
|---|
| 2991 | +#if defined(CONFIG_MALI_VECTOR_DUMP) |
|---|
| 2868 | 2992 | if (vma->vm_pgoff == PFN_DOWN(BASE_MEM_MMU_DUMP_HANDLE)) { |
|---|
| 2869 | 2993 | /* MMU dump - userspace should now have a reference on |
|---|
| 2870 | 2994 | * the pages, so we can now free the kernel mapping |
|---|
| .. | .. |
|---|
| 2883 | 3007 | */ |
|---|
| 2884 | 3008 | vma->vm_pgoff = PFN_DOWN(vma->vm_start); |
|---|
| 2885 | 3009 | } |
|---|
| 2886 | | - |
|---|
| 3010 | +#endif /* defined(CONFIG_MALI_VECTOR_DUMP) */ |
|---|
| 2887 | 3011 | out_unlock: |
|---|
| 2888 | 3012 | kbase_gpu_vm_unlock(kctx); |
|---|
| 2889 | 3013 | out: |
|---|
| .. | .. |
|---|
| 2925 | 3049 | } |
|---|
| 2926 | 3050 | } |
|---|
| 2927 | 3051 | |
|---|
| 2928 | | -static int kbase_vmap_phy_pages(struct kbase_context *kctx, |
|---|
| 2929 | | - struct kbase_va_region *reg, u64 offset_bytes, size_t size, |
|---|
| 2930 | | - struct kbase_vmap_struct *map) |
|---|
| 3052 | +/** |
|---|
| 3053 | + * kbase_vmap_phy_pages_migrate_count_increment - Increment VMAP count for |
|---|
| 3054 | + * array of physical pages |
|---|
| 3055 | + * |
|---|
| 3056 | + * @pages: Array of pages. |
|---|
| 3057 | + * @page_count: Number of pages. |
|---|
| 3058 | + * @flags: Region flags. |
|---|
| 3059 | + * |
|---|
| 3060 | + * This function is supposed to be called only if page migration support |
|---|
| 3061 | + * is enabled in the driver. |
|---|
| 3062 | + * |
|---|
| 3063 | + * The counter of kernel CPU mappings of the physical pages involved in a |
|---|
| 3064 | + * mapping operation is incremented by 1. Errors are handled by making pages |
|---|
| 3065 | + * not movable. Permanent kernel mappings will be marked as not movable, too. |
|---|
| 3066 | + */ |
|---|
| 3067 | +static void kbase_vmap_phy_pages_migrate_count_increment(struct tagged_addr *pages, |
|---|
| 3068 | + size_t page_count, unsigned long flags) |
|---|
| 3069 | +{ |
|---|
| 3070 | + size_t i; |
|---|
| 3071 | + |
|---|
| 3072 | + for (i = 0; i < page_count; i++) { |
|---|
| 3073 | + struct page *p = as_page(pages[i]); |
|---|
| 3074 | + struct kbase_page_metadata *page_md = kbase_page_private(p); |
|---|
| 3075 | + |
|---|
| 3076 | + /* Skip the 4KB page that is part of a large page, as the large page is |
|---|
| 3077 | + * excluded from the migration process. |
|---|
| 3078 | + */ |
|---|
| 3079 | + if (is_huge(pages[i]) || is_partial(pages[i])) |
|---|
| 3080 | + continue; |
|---|
| 3081 | + |
|---|
| 3082 | + spin_lock(&page_md->migrate_lock); |
|---|
| 3083 | + /* Mark permanent kernel mappings as NOT_MOVABLE because they're likely |
|---|
| 3084 | + * to stay mapped for a long time. However, keep on counting the number |
|---|
| 3085 | + * of mappings even for them: they don't represent an exception for the |
|---|
| 3086 | + * vmap_count. |
|---|
| 3087 | + * |
|---|
| 3088 | + * At the same time, errors need to be handled if a client tries to add |
|---|
| 3089 | + * too many mappings, hence a page may end up in the NOT_MOVABLE state |
|---|
| 3090 | + * anyway even if it's not a permanent kernel mapping. |
|---|
| 3091 | + */ |
|---|
| 3092 | + if (flags & KBASE_REG_PERMANENT_KERNEL_MAPPING) |
|---|
| 3093 | + page_md->status = PAGE_STATUS_SET(page_md->status, (u8)NOT_MOVABLE); |
|---|
| 3094 | + if (page_md->vmap_count < U8_MAX) |
|---|
| 3095 | + page_md->vmap_count++; |
|---|
| 3096 | + else |
|---|
| 3097 | + page_md->status = PAGE_STATUS_SET(page_md->status, (u8)NOT_MOVABLE); |
|---|
| 3098 | + spin_unlock(&page_md->migrate_lock); |
|---|
| 3099 | + } |
|---|
| 3100 | +} |
|---|
| 3101 | + |
|---|
| 3102 | +/** |
|---|
| 3103 | + * kbase_vunmap_phy_pages_migrate_count_decrement - Decrement VMAP count for |
|---|
| 3104 | + * array of physical pages |
|---|
| 3105 | + * |
|---|
| 3106 | + * @pages: Array of pages. |
|---|
| 3107 | + * @page_count: Number of pages. |
|---|
| 3108 | + * |
|---|
| 3109 | + * This function is supposed to be called only if page migration support |
|---|
| 3110 | + * is enabled in the driver. |
|---|
| 3111 | + * |
|---|
| 3112 | + * The counter of kernel CPU mappings of the physical pages involved in a |
|---|
| 3113 | + * mapping operation is decremented by 1. Errors are handled by making pages |
|---|
| 3114 | + * not movable. |
|---|
| 3115 | + */ |
|---|
| 3116 | +static void kbase_vunmap_phy_pages_migrate_count_decrement(struct tagged_addr *pages, |
|---|
| 3117 | + size_t page_count) |
|---|
| 3118 | +{ |
|---|
| 3119 | + size_t i; |
|---|
| 3120 | + |
|---|
| 3121 | + for (i = 0; i < page_count; i++) { |
|---|
| 3122 | + struct page *p = as_page(pages[i]); |
|---|
| 3123 | + struct kbase_page_metadata *page_md = kbase_page_private(p); |
|---|
| 3124 | + |
|---|
| 3125 | + /* Skip the 4KB page that is part of a large page, as the large page is |
|---|
| 3126 | + * excluded from the migration process. |
|---|
| 3127 | + */ |
|---|
| 3128 | + if (is_huge(pages[i]) || is_partial(pages[i])) |
|---|
| 3129 | + continue; |
|---|
| 3130 | + |
|---|
| 3131 | + spin_lock(&page_md->migrate_lock); |
|---|
| 3132 | + /* Decrement the number of mappings for all kinds of pages, including |
|---|
| 3133 | + * pages which are NOT_MOVABLE (e.g. permanent kernel mappings). |
|---|
| 3134 | + * However, errors still need to be handled if a client tries to remove |
|---|
| 3135 | + * more mappings than created. |
|---|
| 3136 | + */ |
|---|
| 3137 | + if (page_md->vmap_count == 0) |
|---|
| 3138 | + page_md->status = PAGE_STATUS_SET(page_md->status, (u8)NOT_MOVABLE); |
|---|
| 3139 | + else |
|---|
| 3140 | + page_md->vmap_count--; |
|---|
| 3141 | + spin_unlock(&page_md->migrate_lock); |
|---|
| 3142 | + } |
|---|
| 3143 | +} |
|---|
| 3144 | + |
|---|
| 3145 | +static int kbase_vmap_phy_pages(struct kbase_context *kctx, struct kbase_va_region *reg, |
|---|
| 3146 | + u64 offset_bytes, size_t size, struct kbase_vmap_struct *map, |
|---|
| 3147 | + kbase_vmap_flag vmap_flags) |
|---|
| 2931 | 3148 | { |
|---|
| 2932 | 3149 | unsigned long page_index; |
|---|
| 2933 | 3150 | unsigned int offset_in_page = offset_bytes & ~PAGE_MASK; |
|---|
| .. | .. |
|---|
| 2937 | 3154 | void *cpu_addr = NULL; |
|---|
| 2938 | 3155 | pgprot_t prot; |
|---|
| 2939 | 3156 | size_t i; |
|---|
| 3157 | + |
|---|
| 3158 | + if (WARN_ON(vmap_flags & ~KBASE_VMAP_INPUT_FLAGS)) |
|---|
| 3159 | + return -EINVAL; |
|---|
| 3160 | + |
|---|
| 3161 | + if (WARN_ON(kbase_is_region_invalid_or_free(reg))) |
|---|
| 3162 | + return -EINVAL; |
|---|
| 2940 | 3163 | |
|---|
| 2941 | 3164 | if (!size || !map || !reg->cpu_alloc || !reg->gpu_alloc) |
|---|
| 2942 | 3165 | return -EINVAL; |
|---|
| .. | .. |
|---|
| 2953 | 3176 | |
|---|
| 2954 | 3177 | if (page_index + page_count > kbase_reg_current_backed_size(reg)) |
|---|
| 2955 | 3178 | return -ENOMEM; |
|---|
| 3179 | + |
|---|
| 3180 | + if ((vmap_flags & KBASE_VMAP_FLAG_PERMANENT_MAP_ACCOUNTING) && |
|---|
| 3181 | + (page_count > (KBASE_PERMANENTLY_MAPPED_MEM_LIMIT_PAGES - |
|---|
| 3182 | + atomic_read(&kctx->permanent_mapped_pages)))) { |
|---|
| 3183 | + dev_warn( |
|---|
| 3184 | + kctx->kbdev->dev, |
|---|
| 3185 | + "Request for %llu more pages mem needing a permanent mapping would breach limit %lu, currently at %d pages", |
|---|
| 3186 | + (u64)page_count, KBASE_PERMANENTLY_MAPPED_MEM_LIMIT_PAGES, |
|---|
| 3187 | + atomic_read(&kctx->permanent_mapped_pages)); |
|---|
| 3188 | + return -ENOMEM; |
|---|
| 3189 | + } |
|---|
| 2956 | 3190 | |
|---|
| 2957 | 3191 | if (reg->flags & KBASE_REG_DONT_NEED) |
|---|
| 2958 | 3192 | return -EINVAL; |
|---|
| .. | .. |
|---|
| 2980 | 3214 | */ |
|---|
| 2981 | 3215 | cpu_addr = vmap(pages, page_count, VM_MAP, prot); |
|---|
| 2982 | 3216 | |
|---|
| 3217 | + /* If page migration is enabled, increment the number of VMA mappings |
|---|
| 3218 | + * of all physical pages. In case of errors, e.g. too many mappings, |
|---|
| 3219 | + * make the page not movable to prevent trouble. |
|---|
| 3220 | + */ |
|---|
| 3221 | + if (kbase_page_migration_enabled && !kbase_mem_is_imported(reg->gpu_alloc->type)) |
|---|
| 3222 | + kbase_vmap_phy_pages_migrate_count_increment(page_array, page_count, reg->flags); |
|---|
| 3223 | + |
|---|
| 2983 | 3224 | kfree(pages); |
|---|
| 2984 | 3225 | |
|---|
| 2985 | 3226 | if (!cpu_addr) |
|---|
| .. | .. |
|---|
| 2992 | 3233 | map->gpu_pages = &kbase_get_gpu_phy_pages(reg)[page_index]; |
|---|
| 2993 | 3234 | map->addr = (void *)((uintptr_t)cpu_addr + offset_in_page); |
|---|
| 2994 | 3235 | map->size = size; |
|---|
| 2995 | | - map->sync_needed = ((reg->flags & KBASE_REG_CPU_CACHED) != 0) && |
|---|
| 2996 | | - !kbase_mem_is_imported(map->gpu_alloc->type); |
|---|
| 3236 | + map->flags = vmap_flags; |
|---|
| 3237 | + if ((reg->flags & KBASE_REG_CPU_CACHED) && !kbase_mem_is_imported(map->gpu_alloc->type)) |
|---|
| 3238 | + map->flags |= KBASE_VMAP_FLAG_SYNC_NEEDED; |
|---|
| 2997 | 3239 | |
|---|
| 2998 | | - if (map->sync_needed) |
|---|
| 3240 | + if (map->flags & KBASE_VMAP_FLAG_SYNC_NEEDED) |
|---|
| 2999 | 3241 | kbase_sync_mem_regions(kctx, map, KBASE_SYNC_TO_CPU); |
|---|
| 3000 | 3242 | |
|---|
| 3243 | + if (vmap_flags & KBASE_VMAP_FLAG_PERMANENT_MAP_ACCOUNTING) |
|---|
| 3244 | + atomic_add(page_count, &kctx->permanent_mapped_pages); |
|---|
| 3245 | + |
|---|
| 3001 | 3246 | kbase_mem_phy_alloc_kernel_mapped(reg->cpu_alloc); |
|---|
| 3247 | + |
|---|
| 3002 | 3248 | return 0; |
|---|
| 3249 | +} |
|---|
| 3250 | + |
|---|
| 3251 | +void *kbase_vmap_reg(struct kbase_context *kctx, struct kbase_va_region *reg, u64 gpu_addr, |
|---|
| 3252 | + size_t size, unsigned long prot_request, struct kbase_vmap_struct *map, |
|---|
| 3253 | + kbase_vmap_flag vmap_flags) |
|---|
| 3254 | +{ |
|---|
| 3255 | + u64 offset_bytes; |
|---|
| 3256 | + struct kbase_mem_phy_alloc *cpu_alloc; |
|---|
| 3257 | + struct kbase_mem_phy_alloc *gpu_alloc; |
|---|
| 3258 | + int err; |
|---|
| 3259 | + |
|---|
| 3260 | + lockdep_assert_held(&kctx->reg_lock); |
|---|
| 3261 | + |
|---|
| 3262 | + if (WARN_ON(kbase_is_region_invalid_or_free(reg))) |
|---|
| 3263 | + return NULL; |
|---|
| 3264 | + |
|---|
| 3265 | + /* check access permissions can be satisfied |
|---|
| 3266 | + * Intended only for checking KBASE_REG_{CPU,GPU}_{RD,WR} |
|---|
| 3267 | + */ |
|---|
| 3268 | + if ((reg->flags & prot_request) != prot_request) |
|---|
| 3269 | + return NULL; |
|---|
| 3270 | + |
|---|
| 3271 | + offset_bytes = gpu_addr - (reg->start_pfn << PAGE_SHIFT); |
|---|
| 3272 | + cpu_alloc = kbase_mem_phy_alloc_get(reg->cpu_alloc); |
|---|
| 3273 | + gpu_alloc = kbase_mem_phy_alloc_get(reg->gpu_alloc); |
|---|
| 3274 | + |
|---|
| 3275 | + err = kbase_vmap_phy_pages(kctx, reg, offset_bytes, size, map, vmap_flags); |
|---|
| 3276 | + if (err < 0) |
|---|
| 3277 | + goto fail_vmap_phy_pages; |
|---|
| 3278 | + |
|---|
| 3279 | + return map->addr; |
|---|
| 3280 | + |
|---|
| 3281 | +fail_vmap_phy_pages: |
|---|
| 3282 | + kbase_mem_phy_alloc_put(cpu_alloc); |
|---|
| 3283 | + kbase_mem_phy_alloc_put(gpu_alloc); |
|---|
| 3284 | + return NULL; |
|---|
| 3003 | 3285 | } |
|---|
| 3004 | 3286 | |
|---|
| 3005 | 3287 | void *kbase_vmap_prot(struct kbase_context *kctx, u64 gpu_addr, size_t size, |
|---|
| .. | .. |
|---|
| 3007 | 3289 | { |
|---|
| 3008 | 3290 | struct kbase_va_region *reg; |
|---|
| 3009 | 3291 | void *addr = NULL; |
|---|
| 3010 | | - u64 offset_bytes; |
|---|
| 3011 | | - struct kbase_mem_phy_alloc *cpu_alloc; |
|---|
| 3012 | | - struct kbase_mem_phy_alloc *gpu_alloc; |
|---|
| 3013 | | - int err; |
|---|
| 3014 | 3292 | |
|---|
| 3015 | 3293 | kbase_gpu_vm_lock(kctx); |
|---|
| 3016 | 3294 | |
|---|
| 3017 | | - reg = kbase_region_tracker_find_region_enclosing_address(kctx, |
|---|
| 3018 | | - gpu_addr); |
|---|
| 3295 | + reg = kbase_region_tracker_find_region_enclosing_address(kctx, gpu_addr); |
|---|
| 3019 | 3296 | if (kbase_is_region_invalid_or_free(reg)) |
|---|
| 3020 | 3297 | goto out_unlock; |
|---|
| 3021 | 3298 | |
|---|
| 3022 | | - /* check access permissions can be satisfied |
|---|
| 3023 | | - * Intended only for checking KBASE_REG_{CPU,GPU}_{RD,WR} |
|---|
| 3024 | | - */ |
|---|
| 3025 | | - if ((reg->flags & prot_request) != prot_request) |
|---|
| 3299 | + if (reg->gpu_alloc->type != KBASE_MEM_TYPE_NATIVE) |
|---|
| 3026 | 3300 | goto out_unlock; |
|---|
| 3027 | 3301 | |
|---|
| 3028 | | - offset_bytes = gpu_addr - (reg->start_pfn << PAGE_SHIFT); |
|---|
| 3029 | | - cpu_alloc = kbase_mem_phy_alloc_get(reg->cpu_alloc); |
|---|
| 3030 | | - gpu_alloc = kbase_mem_phy_alloc_get(reg->gpu_alloc); |
|---|
| 3031 | | - |
|---|
| 3032 | | - err = kbase_vmap_phy_pages(kctx, reg, offset_bytes, size, map); |
|---|
| 3033 | | - if (err < 0) |
|---|
| 3034 | | - goto fail_vmap_phy_pages; |
|---|
| 3035 | | - |
|---|
| 3036 | | - addr = map->addr; |
|---|
| 3302 | + addr = kbase_vmap_reg(kctx, reg, gpu_addr, size, prot_request, map, 0u); |
|---|
| 3037 | 3303 | |
|---|
| 3038 | 3304 | out_unlock: |
|---|
| 3039 | 3305 | kbase_gpu_vm_unlock(kctx); |
|---|
| 3040 | 3306 | return addr; |
|---|
| 3041 | | - |
|---|
| 3042 | | -fail_vmap_phy_pages: |
|---|
| 3043 | | - kbase_gpu_vm_unlock(kctx); |
|---|
| 3044 | | - kbase_mem_phy_alloc_put(cpu_alloc); |
|---|
| 3045 | | - kbase_mem_phy_alloc_put(gpu_alloc); |
|---|
| 3046 | | - |
|---|
| 3047 | | - return NULL; |
|---|
| 3048 | 3307 | } |
|---|
| 3049 | 3308 | |
|---|
| 3050 | 3309 | void *kbase_vmap(struct kbase_context *kctx, u64 gpu_addr, size_t size, |
|---|
| .. | .. |
|---|
| 3064 | 3323 | struct kbase_vmap_struct *map) |
|---|
| 3065 | 3324 | { |
|---|
| 3066 | 3325 | void *addr = (void *)((uintptr_t)map->addr & PAGE_MASK); |
|---|
| 3326 | + |
|---|
| 3067 | 3327 | vunmap(addr); |
|---|
| 3068 | 3328 | |
|---|
| 3069 | | - if (map->sync_needed) |
|---|
| 3329 | + /* If page migration is enabled, decrement the number of VMA mappings |
|---|
| 3330 | + * for all physical pages. Now is a good time to do it because references |
|---|
| 3331 | + * haven't been released yet. |
|---|
| 3332 | + */ |
|---|
| 3333 | + if (kbase_page_migration_enabled && !kbase_mem_is_imported(map->gpu_alloc->type)) { |
|---|
| 3334 | + const size_t page_count = PFN_UP(map->offset_in_page + map->size); |
|---|
| 3335 | + struct tagged_addr *pages_array = map->cpu_pages; |
|---|
| 3336 | + |
|---|
| 3337 | + kbase_vunmap_phy_pages_migrate_count_decrement(pages_array, page_count); |
|---|
| 3338 | + } |
|---|
| 3339 | + |
|---|
| 3340 | + if (map->flags & KBASE_VMAP_FLAG_SYNC_NEEDED) |
|---|
| 3070 | 3341 | kbase_sync_mem_regions(kctx, map, KBASE_SYNC_TO_DEVICE); |
|---|
| 3342 | + if (map->flags & KBASE_VMAP_FLAG_PERMANENT_MAP_ACCOUNTING) { |
|---|
| 3343 | + size_t page_count = PFN_UP(map->offset_in_page + map->size); |
|---|
| 3344 | + |
|---|
| 3345 | + WARN_ON(page_count > atomic_read(&kctx->permanent_mapped_pages)); |
|---|
| 3346 | + atomic_sub(page_count, &kctx->permanent_mapped_pages); |
|---|
| 3347 | + } |
|---|
| 3071 | 3348 | |
|---|
| 3072 | 3349 | kbase_mem_phy_alloc_kernel_unmapped(map->cpu_alloc); |
|---|
| 3350 | + |
|---|
| 3073 | 3351 | map->offset_in_page = 0; |
|---|
| 3074 | 3352 | map->cpu_pages = NULL; |
|---|
| 3075 | 3353 | map->gpu_pages = NULL; |
|---|
| 3076 | 3354 | map->addr = NULL; |
|---|
| 3077 | 3355 | map->size = 0; |
|---|
| 3078 | | - map->sync_needed = false; |
|---|
| 3356 | + map->flags = 0; |
|---|
| 3079 | 3357 | } |
|---|
| 3080 | 3358 | |
|---|
| 3081 | 3359 | void kbase_vunmap(struct kbase_context *kctx, struct kbase_vmap_struct *map) |
|---|
| .. | .. |
|---|
| 3102 | 3380 | |
|---|
| 3103 | 3381 | void kbasep_os_process_page_usage_update(struct kbase_context *kctx, int pages) |
|---|
| 3104 | 3382 | { |
|---|
| 3105 | | - struct mm_struct *mm; |
|---|
| 3383 | + struct mm_struct *mm = kctx->process_mm; |
|---|
| 3106 | 3384 | |
|---|
| 3107 | | - rcu_read_lock(); |
|---|
| 3108 | | - mm = rcu_dereference(kctx->process_mm); |
|---|
| 3109 | | - if (mm) { |
|---|
| 3110 | | - atomic_add(pages, &kctx->nonmapped_pages); |
|---|
| 3111 | | -#ifdef SPLIT_RSS_COUNTING |
|---|
| 3112 | | - kbasep_add_mm_counter(mm, MM_FILEPAGES, pages); |
|---|
| 3113 | | -#else |
|---|
| 3114 | | - spin_lock(&mm->page_table_lock); |
|---|
| 3115 | | - kbasep_add_mm_counter(mm, MM_FILEPAGES, pages); |
|---|
| 3116 | | - spin_unlock(&mm->page_table_lock); |
|---|
| 3117 | | -#endif |
|---|
| 3118 | | - } |
|---|
| 3119 | | - rcu_read_unlock(); |
|---|
| 3120 | | -} |
|---|
| 3121 | | - |
|---|
| 3122 | | -static void kbasep_os_process_page_usage_drain(struct kbase_context *kctx) |
|---|
| 3123 | | -{ |
|---|
| 3124 | | - int pages; |
|---|
| 3125 | | - struct mm_struct *mm; |
|---|
| 3126 | | - |
|---|
| 3127 | | - spin_lock(&kctx->mm_update_lock); |
|---|
| 3128 | | - mm = rcu_dereference_protected(kctx->process_mm, lockdep_is_held(&kctx->mm_update_lock)); |
|---|
| 3129 | | - if (!mm) { |
|---|
| 3130 | | - spin_unlock(&kctx->mm_update_lock); |
|---|
| 3385 | + if (unlikely(!mm)) |
|---|
| 3131 | 3386 | return; |
|---|
| 3132 | | - } |
|---|
| 3133 | 3387 | |
|---|
| 3134 | | - rcu_assign_pointer(kctx->process_mm, NULL); |
|---|
| 3135 | | - spin_unlock(&kctx->mm_update_lock); |
|---|
| 3136 | | - synchronize_rcu(); |
|---|
| 3137 | | - |
|---|
| 3138 | | - pages = atomic_xchg(&kctx->nonmapped_pages, 0); |
|---|
| 3388 | + atomic_add(pages, &kctx->nonmapped_pages); |
|---|
| 3139 | 3389 | #ifdef SPLIT_RSS_COUNTING |
|---|
| 3140 | | - kbasep_add_mm_counter(mm, MM_FILEPAGES, -pages); |
|---|
| 3390 | + kbasep_add_mm_counter(mm, MM_FILEPAGES, pages); |
|---|
| 3141 | 3391 | #else |
|---|
| 3142 | 3392 | spin_lock(&mm->page_table_lock); |
|---|
| 3143 | | - kbasep_add_mm_counter(mm, MM_FILEPAGES, -pages); |
|---|
| 3393 | + kbasep_add_mm_counter(mm, MM_FILEPAGES, pages); |
|---|
| 3144 | 3394 | spin_unlock(&mm->page_table_lock); |
|---|
| 3145 | 3395 | #endif |
|---|
| 3146 | 3396 | } |
|---|
| 3147 | 3397 | |
|---|
| 3148 | | -static void kbase_special_vm_close(struct vm_area_struct *vma) |
|---|
| 3149 | | -{ |
|---|
| 3150 | | - struct kbase_context *kctx; |
|---|
| 3151 | | - |
|---|
| 3152 | | - kctx = vma->vm_private_data; |
|---|
| 3153 | | - kbasep_os_process_page_usage_drain(kctx); |
|---|
| 3154 | | -} |
|---|
| 3155 | | - |
|---|
| 3156 | | -static const struct vm_operations_struct kbase_vm_special_ops = { |
|---|
| 3157 | | - .close = kbase_special_vm_close, |
|---|
| 3158 | | -}; |
|---|
| 3159 | | - |
|---|
| 3160 | 3398 | static int kbase_tracking_page_setup(struct kbase_context *kctx, struct vm_area_struct *vma) |
|---|
| 3161 | 3399 | { |
|---|
| 3162 | | - /* check that this is the only tracking page */ |
|---|
| 3163 | | - spin_lock(&kctx->mm_update_lock); |
|---|
| 3164 | | - if (rcu_dereference_protected(kctx->process_mm, lockdep_is_held(&kctx->mm_update_lock))) { |
|---|
| 3165 | | - spin_unlock(&kctx->mm_update_lock); |
|---|
| 3166 | | - return -EFAULT; |
|---|
| 3167 | | - } |
|---|
| 3168 | | - |
|---|
| 3169 | | - rcu_assign_pointer(kctx->process_mm, current->mm); |
|---|
| 3170 | | - |
|---|
| 3171 | | - spin_unlock(&kctx->mm_update_lock); |
|---|
| 3400 | + if (vma_pages(vma) != 1) |
|---|
| 3401 | + return -EINVAL; |
|---|
| 3172 | 3402 | |
|---|
| 3173 | 3403 | /* no real access */ |
|---|
| 3174 | 3404 | vma->vm_flags &= ~(VM_READ | VM_MAYREAD | VM_WRITE | VM_MAYWRITE | VM_EXEC | VM_MAYEXEC); |
|---|
| 3175 | 3405 | vma->vm_flags |= VM_DONTCOPY | VM_DONTEXPAND | VM_DONTDUMP | VM_IO; |
|---|
| 3176 | | - vma->vm_ops = &kbase_vm_special_ops; |
|---|
| 3177 | | - vma->vm_private_data = kctx; |
|---|
| 3178 | 3406 | |
|---|
| 3179 | 3407 | return 0; |
|---|
| 3180 | 3408 | } |
|---|
| .. | .. |
|---|
| 3189 | 3417 | * assigned one, otherwise a dummy page. Always return the |
|---|
| 3190 | 3418 | * dummy page in no mali builds. |
|---|
| 3191 | 3419 | */ |
|---|
| 3420 | +#if IS_ENABLED(CONFIG_MALI_BIFROST_NO_MALI) |
|---|
| 3421 | + return PFN_DOWN(as_phys_addr_t(kbdev->csf.dummy_db_page)); |
|---|
| 3422 | +#else |
|---|
| 3192 | 3423 | if (queue->doorbell_nr == KBASEP_USER_DB_NR_INVALID) |
|---|
| 3193 | 3424 | return PFN_DOWN(as_phys_addr_t(kbdev->csf.dummy_db_page)); |
|---|
| 3425 | +#endif |
|---|
| 3194 | 3426 | return (PFN_DOWN(kbdev->reg_start + CSF_HW_DOORBELL_PAGE_OFFSET + |
|---|
| 3195 | 3427 | (u64)queue->doorbell_nr * CSF_HW_DOORBELL_PAGE_SIZE)); |
|---|
| 3196 | 3428 | } |
|---|
| 3197 | 3429 | |
|---|
| 3430 | +static int |
|---|
| 3431 | +#if (KERNEL_VERSION(5, 13, 0) <= LINUX_VERSION_CODE || \ |
|---|
| 3432 | + KERNEL_VERSION(5, 11, 0) > LINUX_VERSION_CODE) |
|---|
| 3433 | +kbase_csf_user_io_pages_vm_mremap(struct vm_area_struct *vma) |
|---|
| 3434 | +#else |
|---|
| 3435 | +kbase_csf_user_io_pages_vm_mremap(struct vm_area_struct *vma, unsigned long flags) |
|---|
| 3436 | +#endif |
|---|
| 3437 | +{ |
|---|
| 3438 | + pr_debug("Unexpected call to mremap method for User IO pages mapping vma\n"); |
|---|
| 3439 | + return -EINVAL; |
|---|
| 3440 | +} |
|---|
| 3441 | + |
|---|
| 3442 | +static int kbase_csf_user_io_pages_vm_split(struct vm_area_struct *vma, unsigned long addr) |
|---|
| 3443 | +{ |
|---|
| 3444 | + pr_debug("Unexpected call to split method for User IO pages mapping vma\n"); |
|---|
| 3445 | + return -EINVAL; |
|---|
| 3446 | +} |
|---|
| 3447 | + |
|---|
| 3198 | 3448 | static void kbase_csf_user_io_pages_vm_open(struct vm_area_struct *vma) |
|---|
| 3199 | 3449 | { |
|---|
| 3200 | | - WARN(1, "Unexpected attempt to clone private vma\n"); |
|---|
| 3450 | + pr_debug("Unexpected call to the open method for User IO pages mapping vma\n"); |
|---|
| 3201 | 3451 | vma->vm_private_data = NULL; |
|---|
| 3202 | 3452 | } |
|---|
| 3203 | 3453 | |
|---|
| .. | .. |
|---|
| 3209 | 3459 | int err; |
|---|
| 3210 | 3460 | bool reset_prevented = false; |
|---|
| 3211 | 3461 | |
|---|
| 3212 | | - if (WARN_ON(!queue)) |
|---|
| 3462 | + if (!queue) { |
|---|
| 3463 | + pr_debug("Close method called for the new User IO pages mapping vma\n"); |
|---|
| 3213 | 3464 | return; |
|---|
| 3465 | + } |
|---|
| 3214 | 3466 | |
|---|
| 3215 | 3467 | kctx = queue->kctx; |
|---|
| 3216 | 3468 | kbdev = kctx->kbdev; |
|---|
| .. | .. |
|---|
| 3225 | 3477 | reset_prevented = true; |
|---|
| 3226 | 3478 | |
|---|
| 3227 | 3479 | mutex_lock(&kctx->csf.lock); |
|---|
| 3228 | | - kbase_csf_queue_unbind(queue); |
|---|
| 3480 | + kbase_csf_queue_unbind(queue, is_process_exiting(vma)); |
|---|
| 3229 | 3481 | mutex_unlock(&kctx->csf.lock); |
|---|
| 3230 | 3482 | |
|---|
| 3231 | 3483 | if (reset_prevented) |
|---|
| .. | .. |
|---|
| 3254 | 3506 | struct memory_group_manager_device *mgm_dev; |
|---|
| 3255 | 3507 | |
|---|
| 3256 | 3508 | /* Few sanity checks up front */ |
|---|
| 3257 | | - if ((nr_pages != BASEP_QUEUE_NR_MMAP_USER_PAGES) || |
|---|
| 3258 | | - (vma->vm_pgoff != queue->db_file_offset)) |
|---|
| 3509 | + if (!queue || (nr_pages != BASEP_QUEUE_NR_MMAP_USER_PAGES) || |
|---|
| 3510 | + (vma->vm_pgoff != queue->db_file_offset)) { |
|---|
| 3511 | + pr_warn("Unexpected CPU page fault on User IO pages mapping for process %s tgid %d pid %d\n", |
|---|
| 3512 | + current->comm, current->tgid, current->pid); |
|---|
| 3259 | 3513 | return VM_FAULT_SIGBUS; |
|---|
| 3514 | + } |
|---|
| 3260 | 3515 | |
|---|
| 3261 | | - mutex_lock(&queue->kctx->csf.lock); |
|---|
| 3262 | 3516 | kbdev = queue->kctx->kbdev; |
|---|
| 3263 | 3517 | mgm_dev = kbdev->mgm_dev; |
|---|
| 3518 | + |
|---|
| 3519 | + mutex_lock(&kbdev->csf.reg_lock); |
|---|
| 3264 | 3520 | |
|---|
| 3265 | 3521 | /* Always map the doorbell page as uncached */ |
|---|
| 3266 | 3522 | doorbell_pgprot = pgprot_device(vma->vm_page_prot); |
|---|
| 3267 | 3523 | |
|---|
| 3268 | | -#if ((KERNEL_VERSION(4, 4, 147) >= LINUX_VERSION_CODE) || \ |
|---|
| 3269 | | - ((KERNEL_VERSION(4, 6, 0) > LINUX_VERSION_CODE) && \ |
|---|
| 3270 | | - (KERNEL_VERSION(4, 5, 0) <= LINUX_VERSION_CODE))) |
|---|
| 3271 | | - vma->vm_page_prot = doorbell_pgprot; |
|---|
| 3272 | | - input_page_pgprot = doorbell_pgprot; |
|---|
| 3273 | | - output_page_pgprot = doorbell_pgprot; |
|---|
| 3274 | | -#else |
|---|
| 3275 | 3524 | if (kbdev->system_coherency == COHERENCY_NONE) { |
|---|
| 3276 | 3525 | input_page_pgprot = pgprot_writecombine(vma->vm_page_prot); |
|---|
| 3277 | 3526 | output_page_pgprot = pgprot_writecombine(vma->vm_page_prot); |
|---|
| .. | .. |
|---|
| 3279 | 3528 | input_page_pgprot = vma->vm_page_prot; |
|---|
| 3280 | 3529 | output_page_pgprot = vma->vm_page_prot; |
|---|
| 3281 | 3530 | } |
|---|
| 3282 | | -#endif |
|---|
| 3283 | 3531 | |
|---|
| 3284 | 3532 | doorbell_cpu_addr = vma->vm_start; |
|---|
| 3285 | 3533 | |
|---|
| .. | .. |
|---|
| 3288 | 3536 | #else |
|---|
| 3289 | 3537 | if (vmf->address == doorbell_cpu_addr) { |
|---|
| 3290 | 3538 | #endif |
|---|
| 3291 | | - mutex_lock(&kbdev->csf.reg_lock); |
|---|
| 3292 | 3539 | doorbell_page_pfn = get_queue_doorbell_pfn(kbdev, queue); |
|---|
| 3293 | 3540 | ret = mgm_dev->ops.mgm_vmf_insert_pfn_prot(mgm_dev, |
|---|
| 3294 | 3541 | KBASE_MEM_GROUP_CSF_IO, vma, doorbell_cpu_addr, |
|---|
| 3295 | 3542 | doorbell_page_pfn, doorbell_pgprot); |
|---|
| 3296 | | - mutex_unlock(&kbdev->csf.reg_lock); |
|---|
| 3297 | 3543 | } else { |
|---|
| 3298 | 3544 | /* Map the Input page */ |
|---|
| 3299 | 3545 | input_cpu_addr = doorbell_cpu_addr + PAGE_SIZE; |
|---|
| .. | .. |
|---|
| 3313 | 3559 | } |
|---|
| 3314 | 3560 | |
|---|
| 3315 | 3561 | exit: |
|---|
| 3316 | | - mutex_unlock(&queue->kctx->csf.lock); |
|---|
| 3562 | + mutex_unlock(&kbdev->csf.reg_lock); |
|---|
| 3317 | 3563 | return ret; |
|---|
| 3318 | 3564 | } |
|---|
| 3319 | 3565 | |
|---|
| 3320 | 3566 | static const struct vm_operations_struct kbase_csf_user_io_pages_vm_ops = { |
|---|
| 3321 | 3567 | .open = kbase_csf_user_io_pages_vm_open, |
|---|
| 3322 | 3568 | .close = kbase_csf_user_io_pages_vm_close, |
|---|
| 3569 | +#if KERNEL_VERSION(5, 11, 0) <= LINUX_VERSION_CODE |
|---|
| 3570 | + .may_split = kbase_csf_user_io_pages_vm_split, |
|---|
| 3571 | +#else |
|---|
| 3572 | + .split = kbase_csf_user_io_pages_vm_split, |
|---|
| 3573 | +#endif |
|---|
| 3574 | + .mremap = kbase_csf_user_io_pages_vm_mremap, |
|---|
| 3323 | 3575 | .fault = kbase_csf_user_io_pages_vm_fault |
|---|
| 3324 | 3576 | }; |
|---|
| 3325 | 3577 | |
|---|
| .. | .. |
|---|
| 3399 | 3651 | return err; |
|---|
| 3400 | 3652 | } |
|---|
| 3401 | 3653 | |
|---|
| 3654 | +/** |
|---|
| 3655 | + * kbase_csf_user_reg_vm_open - VMA open function for the USER page |
|---|
| 3656 | + * |
|---|
| 3657 | + * @vma: Pointer to the struct containing information about |
|---|
| 3658 | + * the userspace mapping of USER page. |
|---|
| 3659 | + * Note: |
|---|
| 3660 | + * This function isn't expected to be called. If called (i.e> mremap), |
|---|
| 3661 | + * set private_data as NULL to indicate to close() and fault() functions. |
|---|
| 3662 | + */ |
|---|
| 3663 | +static void kbase_csf_user_reg_vm_open(struct vm_area_struct *vma) |
|---|
| 3664 | +{ |
|---|
| 3665 | + pr_debug("Unexpected call to the open method for USER register mapping"); |
|---|
| 3666 | + vma->vm_private_data = NULL; |
|---|
| 3667 | +} |
|---|
| 3668 | + |
|---|
| 3669 | +/** |
|---|
| 3670 | + * kbase_csf_user_reg_vm_close - VMA close function for the USER page |
|---|
| 3671 | + * |
|---|
| 3672 | + * @vma: Pointer to the struct containing information about |
|---|
| 3673 | + * the userspace mapping of USER page. |
|---|
| 3674 | + */ |
|---|
| 3402 | 3675 | static void kbase_csf_user_reg_vm_close(struct vm_area_struct *vma) |
|---|
| 3403 | 3676 | { |
|---|
| 3404 | 3677 | struct kbase_context *kctx = vma->vm_private_data; |
|---|
| 3678 | + struct kbase_device *kbdev; |
|---|
| 3405 | 3679 | |
|---|
| 3406 | | - WARN_ON(!kctx->csf.user_reg_vma); |
|---|
| 3680 | + if (unlikely(!kctx)) { |
|---|
| 3681 | + pr_debug("Close function called for the unexpected mapping"); |
|---|
| 3682 | + return; |
|---|
| 3683 | + } |
|---|
| 3407 | 3684 | |
|---|
| 3408 | | - kctx->csf.user_reg_vma = NULL; |
|---|
| 3685 | + kbdev = kctx->kbdev; |
|---|
| 3686 | + |
|---|
| 3687 | + if (unlikely(!kctx->csf.user_reg.vma)) |
|---|
| 3688 | + dev_warn(kbdev->dev, "user_reg VMA pointer unexpectedly NULL for ctx %d_%d", |
|---|
| 3689 | + kctx->tgid, kctx->id); |
|---|
| 3690 | + |
|---|
| 3691 | + mutex_lock(&kbdev->csf.reg_lock); |
|---|
| 3692 | + list_del_init(&kctx->csf.user_reg.link); |
|---|
| 3693 | + mutex_unlock(&kbdev->csf.reg_lock); |
|---|
| 3694 | + |
|---|
| 3695 | + kctx->csf.user_reg.vma = NULL; |
|---|
| 3696 | + |
|---|
| 3697 | + /* Now as the VMA is closed, drop the reference on mali device file */ |
|---|
| 3698 | + fput(kctx->filp); |
|---|
| 3699 | +} |
|---|
| 3700 | + |
|---|
| 3701 | +/** |
|---|
| 3702 | + * kbase_csf_user_reg_vm_mremap - VMA mremap function for the USER page |
|---|
| 3703 | + * |
|---|
| 3704 | + * @vma: Pointer to the struct containing information about |
|---|
| 3705 | + * the userspace mapping of USER page. |
|---|
| 3706 | + * |
|---|
| 3707 | + * Return: -EINVAL |
|---|
| 3708 | + * |
|---|
| 3709 | + * Note: |
|---|
| 3710 | + * User space must not attempt mremap on USER page mapping. |
|---|
| 3711 | + * This function will return an error to fail the attempt. |
|---|
| 3712 | + */ |
|---|
| 3713 | +static int |
|---|
| 3714 | +#if ((KERNEL_VERSION(5, 13, 0) <= LINUX_VERSION_CODE) || \ |
|---|
| 3715 | + (KERNEL_VERSION(5, 11, 0) > LINUX_VERSION_CODE)) |
|---|
| 3716 | +kbase_csf_user_reg_vm_mremap(struct vm_area_struct *vma) |
|---|
| 3717 | +#else |
|---|
| 3718 | +kbase_csf_user_reg_vm_mremap(struct vm_area_struct *vma, unsigned long flags) |
|---|
| 3719 | +#endif |
|---|
| 3720 | +{ |
|---|
| 3721 | + pr_debug("Unexpected call to mremap method for USER page mapping vma\n"); |
|---|
| 3722 | + return -EINVAL; |
|---|
| 3409 | 3723 | } |
|---|
| 3410 | 3724 | |
|---|
| 3411 | 3725 | #if (KERNEL_VERSION(4, 11, 0) > LINUX_VERSION_CODE) |
|---|
| .. | .. |
|---|
| 3418 | 3732 | struct vm_area_struct *vma = vmf->vma; |
|---|
| 3419 | 3733 | #endif |
|---|
| 3420 | 3734 | struct kbase_context *kctx = vma->vm_private_data; |
|---|
| 3421 | | - struct kbase_device *kbdev = kctx->kbdev; |
|---|
| 3422 | | - struct memory_group_manager_device *mgm_dev = kbdev->mgm_dev; |
|---|
| 3423 | | - unsigned long pfn = PFN_DOWN(kbdev->reg_start + USER_BASE); |
|---|
| 3735 | + struct kbase_device *kbdev; |
|---|
| 3736 | + struct memory_group_manager_device *mgm_dev; |
|---|
| 3737 | + unsigned long pfn; |
|---|
| 3424 | 3738 | size_t nr_pages = PFN_DOWN(vma->vm_end - vma->vm_start); |
|---|
| 3425 | 3739 | vm_fault_t ret = VM_FAULT_SIGBUS; |
|---|
| 3740 | + unsigned long flags; |
|---|
| 3426 | 3741 | |
|---|
| 3427 | 3742 | /* Few sanity checks up front */ |
|---|
| 3428 | | - if (WARN_ON(nr_pages != 1) || |
|---|
| 3429 | | - WARN_ON(vma != kctx->csf.user_reg_vma) || |
|---|
| 3430 | | - WARN_ON(vma->vm_pgoff != |
|---|
| 3431 | | - PFN_DOWN(BASEP_MEM_CSF_USER_REG_PAGE_HANDLE))) |
|---|
| 3743 | + |
|---|
| 3744 | + if (!kctx || (nr_pages != 1) || (vma != kctx->csf.user_reg.vma) || |
|---|
| 3745 | + (vma->vm_pgoff != kctx->csf.user_reg.file_offset)) { |
|---|
| 3746 | + pr_err("Unexpected CPU page fault on USER page mapping for process %s tgid %d pid %d\n", |
|---|
| 3747 | + current->comm, current->tgid, current->pid); |
|---|
| 3432 | 3748 | return VM_FAULT_SIGBUS; |
|---|
| 3749 | + } |
|---|
| 3433 | 3750 | |
|---|
| 3434 | | - mutex_lock(&kbdev->pm.lock); |
|---|
| 3751 | + kbdev = kctx->kbdev; |
|---|
| 3752 | + mgm_dev = kbdev->mgm_dev; |
|---|
| 3753 | + pfn = PFN_DOWN(kbdev->reg_start + USER_BASE); |
|---|
| 3435 | 3754 | |
|---|
| 3436 | | - /* Don't map in the actual register page if GPU is powered down. |
|---|
| 3437 | | - * Always map in the dummy page in no mali builds. |
|---|
| 3755 | + mutex_lock(&kbdev->csf.reg_lock); |
|---|
| 3756 | + |
|---|
| 3757 | + spin_lock_irqsave(&kbdev->hwaccess_lock, flags); |
|---|
| 3758 | + /* Dummy page will be mapped during GPU off. |
|---|
| 3759 | + * |
|---|
| 3760 | + * In no mail builds, always map in the dummy page. |
|---|
| 3438 | 3761 | */ |
|---|
| 3439 | | - if (!kbdev->pm.backend.gpu_powered) |
|---|
| 3440 | | - pfn = PFN_DOWN(as_phys_addr_t(kbdev->csf.dummy_user_reg_page)); |
|---|
| 3762 | + if (IS_ENABLED(CONFIG_MALI_BIFROST_NO_MALI) || !kbdev->pm.backend.gpu_powered) |
|---|
| 3763 | + pfn = PFN_DOWN(as_phys_addr_t(kbdev->csf.user_reg.dummy_page)); |
|---|
| 3764 | + spin_unlock_irqrestore(&kbdev->hwaccess_lock, flags); |
|---|
| 3441 | 3765 | |
|---|
| 3766 | + list_move_tail(&kctx->csf.user_reg.link, &kbdev->csf.user_reg.list); |
|---|
| 3442 | 3767 | ret = mgm_dev->ops.mgm_vmf_insert_pfn_prot(mgm_dev, |
|---|
| 3443 | 3768 | KBASE_MEM_GROUP_CSF_FW, vma, |
|---|
| 3444 | 3769 | vma->vm_start, pfn, |
|---|
| 3445 | 3770 | vma->vm_page_prot); |
|---|
| 3446 | 3771 | |
|---|
| 3447 | | - mutex_unlock(&kbdev->pm.lock); |
|---|
| 3772 | + mutex_unlock(&kbdev->csf.reg_lock); |
|---|
| 3448 | 3773 | |
|---|
| 3449 | 3774 | return ret; |
|---|
| 3450 | 3775 | } |
|---|
| 3451 | 3776 | |
|---|
| 3452 | 3777 | static const struct vm_operations_struct kbase_csf_user_reg_vm_ops = { |
|---|
| 3778 | + .open = kbase_csf_user_reg_vm_open, |
|---|
| 3453 | 3779 | .close = kbase_csf_user_reg_vm_close, |
|---|
| 3780 | + .mremap = kbase_csf_user_reg_vm_mremap, |
|---|
| 3454 | 3781 | .fault = kbase_csf_user_reg_vm_fault |
|---|
| 3455 | 3782 | }; |
|---|
| 3456 | 3783 | |
|---|
| .. | .. |
|---|
| 3458 | 3785 | struct vm_area_struct *vma) |
|---|
| 3459 | 3786 | { |
|---|
| 3460 | 3787 | size_t nr_pages = PFN_DOWN(vma->vm_end - vma->vm_start); |
|---|
| 3788 | + struct kbase_device *kbdev = kctx->kbdev; |
|---|
| 3461 | 3789 | |
|---|
| 3462 | 3790 | /* Few sanity checks */ |
|---|
| 3463 | | - if (kctx->csf.user_reg_vma) |
|---|
| 3791 | + if (kctx->csf.user_reg.vma) |
|---|
| 3464 | 3792 | return -EBUSY; |
|---|
| 3465 | 3793 | |
|---|
| 3466 | 3794 | if (nr_pages != 1) |
|---|
| .. | .. |
|---|
| 3479 | 3807 | */ |
|---|
| 3480 | 3808 | vma->vm_flags |= VM_PFNMAP; |
|---|
| 3481 | 3809 | |
|---|
| 3482 | | - kctx->csf.user_reg_vma = vma; |
|---|
| 3810 | + kctx->csf.user_reg.vma = vma; |
|---|
| 3483 | 3811 | |
|---|
| 3812 | + mutex_lock(&kbdev->csf.reg_lock); |
|---|
| 3813 | + kctx->csf.user_reg.file_offset = kbdev->csf.user_reg.file_offset++; |
|---|
| 3814 | + mutex_unlock(&kbdev->csf.reg_lock); |
|---|
| 3815 | + |
|---|
| 3816 | + /* Make VMA point to the special internal file, but don't drop the |
|---|
| 3817 | + * reference on mali device file (that would be done later when the |
|---|
| 3818 | + * VMA is closed). |
|---|
| 3819 | + */ |
|---|
| 3820 | + vma->vm_file = kctx->kbdev->csf.user_reg.filp; |
|---|
| 3821 | + get_file(vma->vm_file); |
|---|
| 3822 | + |
|---|
| 3823 | + /* Also adjust the vm_pgoff */ |
|---|
| 3824 | + vma->vm_pgoff = kctx->csf.user_reg.file_offset; |
|---|
| 3484 | 3825 | vma->vm_ops = &kbase_csf_user_reg_vm_ops; |
|---|
| 3485 | 3826 | vma->vm_private_data = kctx; |
|---|
| 3486 | 3827 | |
|---|