/* SPDX-License-Identifier: GPL-2.0 WITH Linux-syscall-note */
|
/*
|
*
|
* (C) COPYRIGHT 2019-2022 ARM Limited. All rights reserved.
|
*
|
* This program is free software and is provided to you under the terms of the
|
* GNU General Public License version 2 as published by the Free Software
|
* Foundation, and any use by you of this program is subject to the terms
|
* of such GNU license.
|
*
|
* This program is distributed in the hope that it will be useful,
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
* GNU General Public License for more details.
|
*
|
* You should have received a copy of the GNU General Public License
|
* along with this program; if not, you can access it online at
|
* http://www.gnu.org/licenses/gpl-2.0.html.
|
*
|
*/
|
|
#ifndef _KBASE_MMU_H_
|
#define _KBASE_MMU_H_
|
|
#include <uapi/gpu/arm/bifrost/mali_base_kernel.h>
|
|
#define KBASE_MMU_PAGE_ENTRIES 512
|
#define KBASE_MMU_INVALID_PGD_ADDRESS (~(phys_addr_t)0)
|
|
struct kbase_context;
|
struct kbase_mmu_table;
|
struct kbase_va_region;
|
|
/**
|
* enum kbase_caller_mmu_sync_info - MMU-synchronous caller info.
|
* A pointer to this type is passed down from the outer-most callers in the kbase
|
* module - where the information resides as to the synchronous / asynchronous
|
* nature of the call flow, with respect to MMU operations. ie - does the call flow relate to
|
* existing GPU work does it come from requests (like ioctl) from user-space, power management,
|
* etc.
|
*
|
* @CALLER_MMU_UNSET_SYNCHRONICITY: default value must be invalid to avoid accidental choice
|
* of a 'valid' value
|
* @CALLER_MMU_SYNC: Arbitrary value for 'synchronous that isn't easy to choose by accident
|
* @CALLER_MMU_ASYNC: Also hard to choose by accident
|
*/
|
enum kbase_caller_mmu_sync_info {
|
CALLER_MMU_UNSET_SYNCHRONICITY,
|
CALLER_MMU_SYNC = 0x02,
|
CALLER_MMU_ASYNC
|
};
|
|
/**
|
* enum kbase_mmu_op_type - enum for MMU operations
|
* @KBASE_MMU_OP_NONE: To help catch uninitialized struct
|
* @KBASE_MMU_OP_FIRST: The lower boundary of enum
|
* @KBASE_MMU_OP_LOCK: Lock memory region
|
* @KBASE_MMU_OP_UNLOCK: Unlock memory region
|
* @KBASE_MMU_OP_FLUSH_PT: Flush page table (CLN+INV L2 only)
|
* @KBASE_MMU_OP_FLUSH_MEM: Flush memory (CLN+INV L2+LSC)
|
* @KBASE_MMU_OP_COUNT: The upper boundary of enum
|
*/
|
enum kbase_mmu_op_type {
|
KBASE_MMU_OP_NONE = 0, /* Must be zero */
|
KBASE_MMU_OP_FIRST, /* Must be the first non-zero op */
|
KBASE_MMU_OP_LOCK = KBASE_MMU_OP_FIRST,
|
KBASE_MMU_OP_UNLOCK,
|
KBASE_MMU_OP_FLUSH_PT,
|
KBASE_MMU_OP_FLUSH_MEM,
|
KBASE_MMU_OP_COUNT /* Must be the last in enum */
|
};
|
|
/**
|
* kbase_mmu_as_init() - Initialising GPU address space object.
|
*
|
* @kbdev: The kbase device structure for the device (must be a valid pointer).
|
* @i: Array index of address space object.
|
*
|
* This is called from device probe to initialise an address space object
|
* of the device.
|
*
|
* Return: 0 on success and non-zero value on failure.
|
*/
|
int kbase_mmu_as_init(struct kbase_device *kbdev, unsigned int i);
|
|
/**
|
* kbase_mmu_as_term() - Terminate address space object.
|
*
|
* @kbdev: The kbase device structure for the device (must be a valid pointer).
|
* @i: Array index of address space object.
|
*
|
* This is called upon device termination to destroy
|
* the address space object of the device.
|
*/
|
void kbase_mmu_as_term(struct kbase_device *kbdev, unsigned int i);
|
|
/**
|
* kbase_mmu_init - Initialise an object representing GPU page tables
|
*
|
* @kbdev: Instance of GPU platform device, allocated from the probe method.
|
* @mmut: GPU page tables to be initialized.
|
* @kctx: Optional kbase context, may be NULL if this set of MMU tables
|
* is not associated with a context.
|
* @group_id: The physical group ID from which to allocate GPU page tables.
|
* Valid range is 0..(MEMORY_GROUP_MANAGER_NR_GROUPS-1).
|
*
|
* The structure should be terminated using kbase_mmu_term()
|
*
|
* Return: 0 if successful, otherwise a negative error code.
|
*/
|
int kbase_mmu_init(struct kbase_device *kbdev, struct kbase_mmu_table *mmut,
|
struct kbase_context *kctx, int group_id);
|
|
/**
|
* kbase_mmu_interrupt - Process an MMU interrupt.
|
*
|
* @kbdev: Pointer to the kbase device for which the interrupt happened.
|
* @irq_stat: Value of the MMU_IRQ_STATUS register.
|
*
|
* Process the MMU interrupt that was reported by the &kbase_device.
|
*/
|
void kbase_mmu_interrupt(struct kbase_device *kbdev, u32 irq_stat);
|
|
/**
|
* kbase_mmu_term - Terminate an object representing GPU page tables
|
*
|
* @kbdev: Instance of GPU platform device, allocated from the probe method.
|
* @mmut: GPU page tables to be destroyed.
|
*
|
* This will free any page tables that have been allocated
|
*/
|
void kbase_mmu_term(struct kbase_device *kbdev, struct kbase_mmu_table *mmut);
|
|
/**
|
* kbase_mmu_create_ate - Create an address translation entry
|
*
|
* @kbdev: Instance of GPU platform device, allocated from the probe method.
|
* @phy: Physical address of the page to be mapped for GPU access.
|
* @flags: Bitmask of attributes of the GPU memory region being mapped.
|
* @level: Page table level for which to build an address translation entry.
|
* @group_id: The physical memory group in which the page was allocated.
|
* Valid range is 0..(MEMORY_GROUP_MANAGER_NR_GROUPS-1).
|
*
|
* This function creates an address translation entry to encode the physical
|
* address of a page to be mapped for access by the GPU, along with any extra
|
* attributes required for the GPU memory region.
|
*
|
* Return: An address translation entry, either in LPAE or AArch64 format
|
* (depending on the driver's configuration).
|
*/
|
u64 kbase_mmu_create_ate(struct kbase_device *kbdev,
|
struct tagged_addr phy, unsigned long flags, int level, int group_id);
|
|
int kbase_mmu_insert_pages_no_flush(struct kbase_device *kbdev, struct kbase_mmu_table *mmut,
|
u64 vpfn, struct tagged_addr *phys, size_t nr,
|
unsigned long flags, int group_id, u64 *dirty_pgds,
|
struct kbase_va_region *reg, bool ignore_page_migration);
|
int kbase_mmu_insert_pages(struct kbase_device *kbdev, struct kbase_mmu_table *mmut, u64 vpfn,
|
struct tagged_addr *phys, size_t nr, unsigned long flags, int as_nr,
|
int group_id, enum kbase_caller_mmu_sync_info mmu_sync_info,
|
struct kbase_va_region *reg, bool ignore_page_migration);
|
int kbase_mmu_insert_imported_pages(struct kbase_device *kbdev, struct kbase_mmu_table *mmut,
|
u64 vpfn, struct tagged_addr *phys, size_t nr,
|
unsigned long flags, int as_nr, int group_id,
|
enum kbase_caller_mmu_sync_info mmu_sync_info,
|
struct kbase_va_region *reg);
|
int kbase_mmu_insert_aliased_pages(struct kbase_device *kbdev, struct kbase_mmu_table *mmut,
|
u64 vpfn, struct tagged_addr *phys, size_t nr,
|
unsigned long flags, int as_nr, int group_id,
|
enum kbase_caller_mmu_sync_info mmu_sync_info,
|
struct kbase_va_region *reg);
|
int kbase_mmu_insert_single_page(struct kbase_context *kctx, u64 vpfn, struct tagged_addr phys,
|
size_t nr, unsigned long flags, int group_id,
|
enum kbase_caller_mmu_sync_info mmu_sync_info,
|
bool ignore_page_migration);
|
int kbase_mmu_insert_single_imported_page(struct kbase_context *kctx, u64 vpfn,
|
struct tagged_addr phys, size_t nr, unsigned long flags,
|
int group_id,
|
enum kbase_caller_mmu_sync_info mmu_sync_info);
|
int kbase_mmu_insert_single_aliased_page(struct kbase_context *kctx, u64 vpfn,
|
struct tagged_addr phys, size_t nr, unsigned long flags,
|
int group_id,
|
enum kbase_caller_mmu_sync_info mmu_sync_info);
|
|
/**
|
* kbase_mmu_teardown_pages - Remove GPU virtual addresses from the MMU page table
|
*
|
* @kbdev: Pointer to kbase device.
|
* @mmut: Pointer to GPU MMU page table.
|
* @vpfn: Start page frame number of the GPU virtual pages to unmap.
|
* @phys: Array of physical pages currently mapped to the virtual
|
* pages to unmap, or NULL. This is used for GPU cache maintenance
|
* and page migration support.
|
* @nr_phys_pages: Number of physical pages to flush.
|
* @nr_virt_pages: Number of virtual pages whose PTEs should be destroyed.
|
* @as_nr: Address space number, for GPU cache maintenance operations
|
* that happen outside a specific kbase context.
|
* @ignore_page_migration: Whether page migration metadata should be ignored.
|
*
|
* We actually discard the ATE and free the page table pages if no valid entries
|
* exist in PGD.
|
*
|
* IMPORTANT: This uses kbasep_js_runpool_release_ctx() when the context is
|
* currently scheduled into the runpool, and so potentially uses a lot of locks.
|
* These locks must be taken in the correct order with respect to others
|
* already held by the caller. Refer to kbasep_js_runpool_release_ctx() for more
|
* information.
|
*
|
* The @p phys pointer to physical pages is not necessary for unmapping virtual memory,
|
* but it is used for fine-grained GPU cache maintenance. If @p phys is NULL,
|
* GPU cache maintenance will be done as usual, that is invalidating the whole GPU caches
|
* instead of specific physical address ranges.
|
*
|
* Return: 0 on success, otherwise an error code.
|
*/
|
int kbase_mmu_teardown_pages(struct kbase_device *kbdev, struct kbase_mmu_table *mmut, u64 vpfn,
|
struct tagged_addr *phys, size_t nr_phys_pages, size_t nr_virt_pages,
|
int as_nr, bool ignore_page_migration);
|
|
int kbase_mmu_update_pages(struct kbase_context *kctx, u64 vpfn,
|
struct tagged_addr *phys, size_t nr,
|
unsigned long flags, int const group_id);
|
#if MALI_USE_CSF
|
/**
|
* kbase_mmu_update_csf_mcu_pages - Update MCU mappings with changes of phys and flags
|
*
|
* @kbdev: Pointer to kbase device.
|
* @vpfn: Virtual PFN (Page Frame Number) of the first page to update
|
* @phys: Pointer to the array of tagged physical addresses of the physical
|
* pages that are pointed to by the page table entries (that need to
|
* be updated).
|
* @nr: Number of pages to update
|
* @flags: Flags
|
* @group_id: The physical memory group in which the page was allocated.
|
* Valid range is 0..(MEMORY_GROUP_MANAGER_NR_GROUPS-1).
|
*
|
* Return: 0 on success, otherwise an error code.
|
*/
|
int kbase_mmu_update_csf_mcu_pages(struct kbase_device *kbdev, u64 vpfn, struct tagged_addr *phys,
|
size_t nr, unsigned long flags, int const group_id);
|
#endif
|
|
/**
|
* kbase_mmu_migrate_page - Migrate GPU mappings and content between memory pages
|
*
|
* @old_phys: Old physical page to be replaced.
|
* @new_phys: New physical page used to replace old physical page.
|
* @old_dma_addr: DMA address of the old page.
|
* @new_dma_addr: DMA address of the new page.
|
* @level: MMU page table level of the provided PGD.
|
*
|
* The page migration process is made of 2 big steps:
|
*
|
* 1) Copy the content of the old page to the new page.
|
* 2) Remap the virtual page, that is: replace either the ATE (if the old page
|
* was a regular page) or the PTE (if the old page was used as a PGD) in the
|
* MMU page table with the new page.
|
*
|
* During the process, the MMU region is locked to prevent GPU access to the
|
* virtual memory page that is being remapped.
|
*
|
* Before copying the content of the old page to the new page and while the
|
* MMU region is locked, a GPU cache flush is performed to make sure that
|
* pending GPU writes are finalized to the old page before copying.
|
* That is necessary because otherwise there's a risk that GPU writes might
|
* be finalized to the old page, and not new page, after migration.
|
* The MMU region is unlocked only at the end of the migration operation.
|
*
|
* Return: 0 on success, otherwise an error code.
|
*/
|
int kbase_mmu_migrate_page(struct tagged_addr old_phys, struct tagged_addr new_phys,
|
dma_addr_t old_dma_addr, dma_addr_t new_dma_addr, int level);
|
|
/**
|
* kbase_mmu_flush_pa_range() - Flush physical address range from the GPU caches
|
*
|
* @kbdev: Instance of GPU platform device, allocated from the probe method.
|
* @kctx: Pointer to kbase context, it can be NULL if the physical address
|
* range is not associated with User created context.
|
* @phys: Starting address of the physical range to start the operation on.
|
* @size: Number of bytes to work on.
|
* @flush_op: Type of cache flush operation to perform.
|
*
|
* Issue a cache flush physical range command. This function won't perform any
|
* flush if the GPU doesn't support FLUSH_PA_RANGE command. The flush would be
|
* performed only if the context has a JASID assigned to it.
|
* This function is basically a wrapper for kbase_gpu_cache_flush_pa_range_and_busy_wait().
|
*/
|
void kbase_mmu_flush_pa_range(struct kbase_device *kbdev, struct kbase_context *kctx,
|
phys_addr_t phys, size_t size,
|
enum kbase_mmu_op_type flush_op);
|
|
/**
|
* kbase_mmu_bus_fault_interrupt - Process a bus fault interrupt.
|
*
|
* @kbdev: Pointer to the kbase device for which bus fault was reported.
|
* @status: Value of the GPU_FAULTSTATUS register.
|
* @as_nr: GPU address space for which the bus fault occurred.
|
*
|
* Process the bus fault interrupt that was reported for a particular GPU
|
* address space.
|
*
|
* Return: zero if the operation was successful, non-zero otherwise.
|
*/
|
int kbase_mmu_bus_fault_interrupt(struct kbase_device *kbdev, u32 status,
|
u32 as_nr);
|
|
/**
|
* kbase_mmu_gpu_fault_interrupt() - Report a GPU fault.
|
*
|
* @kbdev: Kbase device pointer
|
* @status: GPU fault status
|
* @as_nr: Faulty address space
|
* @address: GPU fault address
|
* @as_valid: true if address space is valid
|
*
|
* This function builds GPU fault information to submit a work
|
* for reporting the details of the fault.
|
*/
|
void kbase_mmu_gpu_fault_interrupt(struct kbase_device *kbdev, u32 status,
|
u32 as_nr, u64 address, bool as_valid);
|
|
/**
|
* kbase_context_mmu_group_id_get - Decode a memory group ID from
|
* base_context_create_flags
|
*
|
* @flags: Bitmask of flags to pass to base_context_init.
|
*
|
* Memory allocated for GPU page tables will come from the returned group.
|
*
|
* Return: Physical memory group ID. Valid range is 0..(BASE_MEM_GROUP_COUNT-1).
|
*/
|
static inline int
|
kbase_context_mmu_group_id_get(base_context_create_flags const flags)
|
{
|
KBASE_DEBUG_ASSERT(flags ==
|
(flags & BASEP_CONTEXT_CREATE_ALLOWED_FLAGS));
|
return (int)BASE_CONTEXT_MMU_GROUP_ID_GET(flags);
|
}
|
|
#endif /* _KBASE_MMU_H_ */
|