/* SPDX-License-Identifier: GPL-2.0 WITH Linux-syscall-note */ /* * * (C) COPYRIGHT 2010-2022 ARM Limited. All rights reserved. * * This program is free software and is provided to you under the terms of the * GNU General Public License version 2 as published by the Free Software * Foundation, and any use by you of this program is subject to the terms * of such GNU license. * * This program is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the * GNU General Public License for more details. * * You should have received a copy of the GNU General Public License * along with this program; if not, you can access it online at * http://www.gnu.org/licenses/gpl-2.0.html. * */ #ifndef _KBASE_GPU_REGMAP_H_ #define _KBASE_GPU_REGMAP_H_ #include #include #include #if MALI_USE_CSF #include "backend/mali_kbase_gpu_regmap_csf.h" #else #include "backend/mali_kbase_gpu_regmap_jm.h" #endif /* GPU_U definition */ #ifdef __ASSEMBLER__ #define GPU_U(x) x #define GPU_UL(x) x #define GPU_ULL(x) x #else #define GPU_U(x) x##u #define GPU_UL(x) x##ul #define GPU_ULL(x) x##ull #endif /* __ASSEMBLER__ */ /* Begin Register Offsets */ /* GPU control registers */ #define L2_FEATURES 0x004 /* (RO) Level 2 cache features */ #define TILER_FEATURES 0x00C /* (RO) Tiler Features */ #define MEM_FEATURES 0x010 /* (RO) Memory system features */ #define MMU_FEATURES 0x014 /* (RO) MMU features */ #define AS_PRESENT 0x018 /* (RO) Address space slots present */ #define GPU_IRQ_RAWSTAT 0x020 /* (RW) */ #define GPU_IRQ_MASK 0x028 /* (RW) */ #define GPU_COMMAND 0x030 /* (WO) */ #define GPU_STATUS 0x034 /* (RO) */ #define GPU_DBGEN (1 << 8) /* DBGEN wire status */ #define GPU_FAULTSTATUS 0x03C /* (RO) GPU exception type and fault status */ #define GPU_FAULTADDRESS_LO 0x040 /* (RO) GPU exception fault address, low word */ #define GPU_FAULTADDRESS_HI 0x044 /* (RO) GPU exception fault address, high word */ #define L2_CONFIG 0x048 /* (RW) Level 2 cache configuration */ #define GROUPS_L2_COHERENT (1 << 0) /* Cores groups are l2 coherent */ #define SUPER_L2_COHERENT (1 << 1) /* Shader cores within a core * supergroup are l2 coherent */ #define PWR_KEY 0x050 /* (WO) Power manager key register */ #define PWR_OVERRIDE0 0x054 /* (RW) Power manager override settings */ #define PWR_OVERRIDE1 0x058 /* (RW) Power manager override settings */ #define GPU_FEATURES_LO 0x060 /* (RO) GPU features, low word */ #define GPU_FEATURES_HI 0x064 /* (RO) GPU features, high word */ #define PRFCNT_FEATURES 0x068 /* (RO) Performance counter features */ #define TIMESTAMP_OFFSET_LO 0x088 /* (RW) Global time stamp offset, low word */ #define TIMESTAMP_OFFSET_HI 0x08C /* (RW) Global time stamp offset, high word */ #define CYCLE_COUNT_LO 0x090 /* (RO) Cycle counter, low word */ #define CYCLE_COUNT_HI 0x094 /* (RO) Cycle counter, high word */ #define TIMESTAMP_LO 0x098 /* (RO) Global time stamp counter, low word */ #define TIMESTAMP_HI 0x09C /* (RO) Global time stamp counter, high word */ #define THREAD_MAX_THREADS 0x0A0 /* (RO) Maximum number of threads per core */ #define THREAD_MAX_WORKGROUP_SIZE 0x0A4 /* (RO) Maximum workgroup size */ #define THREAD_MAX_BARRIER_SIZE 0x0A8 /* (RO) Maximum threads waiting at a barrier */ #define THREAD_FEATURES 0x0AC /* (RO) Thread features */ #define THREAD_TLS_ALLOC 0x310 /* (RO) Number of threads per core that TLS must be allocated for */ #define TEXTURE_FEATURES_0 0x0B0 /* (RO) Support flags for indexed texture formats 0..31 */ #define TEXTURE_FEATURES_1 0x0B4 /* (RO) Support flags for indexed texture formats 32..63 */ #define TEXTURE_FEATURES_2 0x0B8 /* (RO) Support flags for indexed texture formats 64..95 */ #define TEXTURE_FEATURES_3 0x0BC /* (RO) Support flags for texture order */ #define TEXTURE_FEATURES_REG(n) GPU_CONTROL_REG(TEXTURE_FEATURES_0 + ((n) << 2)) #define GPU_COMMAND_ARG0_LO 0x0D0 /* (RW) Additional parameter 0 for GPU commands, low word */ #define GPU_COMMAND_ARG0_HI 0x0D4 /* (RW) Additional parameter 0 for GPU commands, high word */ #define GPU_COMMAND_ARG1_LO 0x0D8 /* (RW) Additional parameter 1 for GPU commands, low word */ #define GPU_COMMAND_ARG1_HI 0x0DC /* (RW) Additional parameter 1 for GPU commands, high word */ #define SHADER_PRESENT_LO 0x100 /* (RO) Shader core present bitmap, low word */ #define SHADER_PRESENT_HI 0x104 /* (RO) Shader core present bitmap, high word */ #define TILER_PRESENT_LO 0x110 /* (RO) Tiler core present bitmap, low word */ #define TILER_PRESENT_HI 0x114 /* (RO) Tiler core present bitmap, high word */ #define L2_PRESENT_LO 0x120 /* (RO) Level 2 cache present bitmap, low word */ #define L2_PRESENT_HI 0x124 /* (RO) Level 2 cache present bitmap, high word */ #define STACK_PRESENT_LO 0xE00 /* (RO) Core stack present bitmap, low word */ #define STACK_PRESENT_HI 0xE04 /* (RO) Core stack present bitmap, high word */ #define STACK_READY_LO 0xE10 /* (RO) Core stack ready bitmap, low word */ #define STACK_READY_HI 0xE14 /* (RO) Core stack ready bitmap, high word */ #define SHADER_PWRFEATURES 0x188 /* (RW) Shader core power features */ #define STACK_PWRON_LO 0xE20 /* (RO) Core stack power on bitmap, low word */ #define STACK_PWRON_HI 0xE24 /* (RO) Core stack power on bitmap, high word */ #define SHADER_PWROFF_LO 0x1C0 /* (WO) Shader core power off bitmap, low word */ #define SHADER_PWROFF_HI 0x1C4 /* (WO) Shader core power off bitmap, high word */ #define TILER_PWROFF_LO 0x1D0 /* (WO) Tiler core power off bitmap, low word */ #define TILER_PWROFF_HI 0x1D4 /* (WO) Tiler core power off bitmap, high word */ #define L2_PWROFF_LO 0x1E0 /* (WO) Level 2 cache power off bitmap, low word */ #define L2_PWROFF_HI 0x1E4 /* (WO) Level 2 cache power off bitmap, high word */ #define STACK_PWROFF_LO 0xE30 /* (RO) Core stack power off bitmap, low word */ #define STACK_PWROFF_HI 0xE34 /* (RO) Core stack power off bitmap, high word */ #define SHADER_PWRTRANS_LO 0x200 /* (RO) Shader core power transition bitmap, low word */ #define SHADER_PWRTRANS_HI 0x204 /* (RO) Shader core power transition bitmap, high word */ #define TILER_PWRTRANS_LO 0x210 /* (RO) Tiler core power transition bitmap, low word */ #define TILER_PWRTRANS_HI 0x214 /* (RO) Tiler core power transition bitmap, high word */ #define L2_PWRTRANS_LO 0x220 /* (RO) Level 2 cache power transition bitmap, low word */ #define L2_PWRTRANS_HI 0x224 /* (RO) Level 2 cache power transition bitmap, high word */ #define ASN_HASH_0 0x02C0 /* (RW) ASN hash function argument 0 */ #define ASN_HASH(n) (ASN_HASH_0 + (n)*4) #define ASN_HASH_COUNT 3 #define SYSC_ALLOC0 0x0340 /* (RW) System cache allocation hint from source ID */ #define SYSC_ALLOC(n) (SYSC_ALLOC0 + (n)*4) #define SYSC_ALLOC_COUNT 8 #define STACK_PWRTRANS_LO 0xE40 /* (RO) Core stack power transition bitmap, low word */ #define STACK_PWRTRANS_HI 0xE44 /* (RO) Core stack power transition bitmap, high word */ #define SHADER_PWRACTIVE_LO 0x240 /* (RO) Shader core active bitmap, low word */ #define SHADER_PWRACTIVE_HI 0x244 /* (RO) Shader core active bitmap, high word */ #define TILER_PWRACTIVE_LO 0x250 /* (RO) Tiler core active bitmap, low word */ #define TILER_PWRACTIVE_HI 0x254 /* (RO) Tiler core active bitmap, high word */ #define L2_PWRACTIVE_LO 0x260 /* (RO) Level 2 cache active bitmap, low word */ #define L2_PWRACTIVE_HI 0x264 /* (RO) Level 2 cache active bitmap, high word */ #define COHERENCY_FEATURES 0x300 /* (RO) Coherency features present */ #define COHERENCY_ENABLE 0x304 /* (RW) Coherency enable */ #define AMBA_FEATURES 0x300 /* (RO) AMBA bus supported features */ #define AMBA_ENABLE 0x304 /* (RW) AMBA features enable */ #define SHADER_CONFIG 0xF04 /* (RW) Shader core configuration (implementation-specific) */ #define TILER_CONFIG 0xF08 /* (RW) Tiler core configuration (implementation-specific) */ #define L2_MMU_CONFIG 0xF0C /* (RW) L2 cache and MMU configuration (implementation-specific) */ /* Job control registers */ #define JOB_IRQ_RAWSTAT 0x000 /* Raw interrupt status register */ /* MMU control registers */ #define MMU_AS1 0x440 /* Configuration registers for address space 1 */ #define MMU_AS2 0x480 /* Configuration registers for address space 2 */ #define MMU_AS3 0x4C0 /* Configuration registers for address space 3 */ #define MMU_AS4 0x500 /* Configuration registers for address space 4 */ #define MMU_AS5 0x540 /* Configuration registers for address space 5 */ #define MMU_AS6 0x580 /* Configuration registers for address space 6 */ #define MMU_AS7 0x5C0 /* Configuration registers for address space 7 */ #define MMU_AS8 0x600 /* Configuration registers for address space 8 */ #define MMU_AS9 0x640 /* Configuration registers for address space 9 */ #define MMU_AS10 0x680 /* Configuration registers for address space 10 */ #define MMU_AS11 0x6C0 /* Configuration registers for address space 11 */ #define MMU_AS12 0x700 /* Configuration registers for address space 12 */ #define MMU_AS13 0x740 /* Configuration registers for address space 13 */ #define MMU_AS14 0x780 /* Configuration registers for address space 14 */ #define MMU_AS15 0x7C0 /* Configuration registers for address space 15 */ /* MMU address space control registers */ #define AS_LOCKADDR_LO 0x10 /* (RW) Lock region address for address space n, low word */ #define AS_LOCKADDR_HI 0x14 /* (RW) Lock region address for address space n, high word */ #define AS_FAULTSTATUS 0x1C /* (RO) MMU fault status register for address space n */ #define AS_FAULTADDRESS_LO 0x20 /* (RO) Fault Address for address space n, low word */ #define AS_FAULTADDRESS_HI 0x24 /* (RO) Fault Address for address space n, high word */ #define AS_STATUS 0x28 /* (RO) Status flags for address space n */ /* (RO) Secondary fault address for address space n, low word */ #define AS_FAULTEXTRA_LO 0x38 /* (RO) Secondary fault address for address space n, high word */ #define AS_FAULTEXTRA_HI 0x3C /* End Register Offsets */ #define GPU_IRQ_REG_ALL (GPU_IRQ_REG_COMMON) /* * MMU_IRQ_RAWSTAT register values. Values are valid also for * MMU_IRQ_CLEAR, MMU_IRQ_MASK, MMU_IRQ_STATUS registers. */ #define MMU_PAGE_FAULT_FLAGS 16 /* Macros returning a bitmask to retrieve page fault or bus error flags from * MMU registers */ #define MMU_PAGE_FAULT(n) (1UL << (n)) #define MMU_BUS_ERROR(n) (1UL << ((n) + MMU_PAGE_FAULT_FLAGS)) /* * Begin AARCH64 MMU TRANSTAB register values */ #define MMU_HW_OUTA_BITS 40 #define AS_TRANSTAB_BASE_MASK ((1ULL << MMU_HW_OUTA_BITS) - (1ULL << 4)) /* * Begin MMU STATUS register values */ #define AS_STATUS_AS_ACTIVE 0x01 #define AS_FAULTSTATUS_EXCEPTION_CODE_MASK (0x7<<3) #define AS_FAULTSTATUS_EXCEPTION_CODE_TRANSLATION_FAULT (0x0<<3) #define AS_FAULTSTATUS_EXCEPTION_CODE_PERMISSION_FAULT (0x1<<3) #define AS_FAULTSTATUS_EXCEPTION_CODE_TRANSTAB_BUS_FAULT (0x2<<3) #define AS_FAULTSTATUS_EXCEPTION_CODE_ACCESS_FLAG (0x3<<3) #define AS_FAULTSTATUS_EXCEPTION_CODE_ADDRESS_SIZE_FAULT (0x4<<3) #define AS_FAULTSTATUS_EXCEPTION_CODE_MEMORY_ATTRIBUTES_FAULT (0x5<<3) #define AS_FAULTSTATUS_EXCEPTION_TYPE_SHIFT 0 #define AS_FAULTSTATUS_EXCEPTION_TYPE_MASK (0xFF << AS_FAULTSTATUS_EXCEPTION_TYPE_SHIFT) #define AS_FAULTSTATUS_EXCEPTION_TYPE_GET(reg_val) \ (((reg_val)&AS_FAULTSTATUS_EXCEPTION_TYPE_MASK) >> AS_FAULTSTATUS_EXCEPTION_TYPE_SHIFT) #define AS_FAULTSTATUS_EXCEPTION_TYPE_TRANSLATION_FAULT_0 0xC0 #define AS_FAULTSTATUS_ACCESS_TYPE_SHIFT 8 #define AS_FAULTSTATUS_ACCESS_TYPE_MASK (0x3 << AS_FAULTSTATUS_ACCESS_TYPE_SHIFT) #define AS_FAULTSTATUS_ACCESS_TYPE_GET(reg_val) \ (((reg_val)&AS_FAULTSTATUS_ACCESS_TYPE_MASK) >> AS_FAULTSTATUS_ACCESS_TYPE_SHIFT) #define AS_FAULTSTATUS_ACCESS_TYPE_ATOMIC (0x0) #define AS_FAULTSTATUS_ACCESS_TYPE_EX (0x1) #define AS_FAULTSTATUS_ACCESS_TYPE_READ (0x2) #define AS_FAULTSTATUS_ACCESS_TYPE_WRITE (0x3) #define AS_FAULTSTATUS_SOURCE_ID_SHIFT 16 #define AS_FAULTSTATUS_SOURCE_ID_MASK (0xFFFF << AS_FAULTSTATUS_SOURCE_ID_SHIFT) #define AS_FAULTSTATUS_SOURCE_ID_GET(reg_val) \ (((reg_val)&AS_FAULTSTATUS_SOURCE_ID_MASK) >> AS_FAULTSTATUS_SOURCE_ID_SHIFT) #define PRFCNT_FEATURES_COUNTER_BLOCK_SIZE_SHIFT (0) #define PRFCNT_FEATURES_COUNTER_BLOCK_SIZE_MASK \ ((0xFF) << PRFCNT_FEATURES_COUNTER_BLOCK_SIZE_SHIFT) #define PRFCNT_FEATURES_COUNTER_BLOCK_SIZE_GET(reg_val) \ (((reg_val)&PRFCNT_FEATURES_COUNTER_BLOCK_SIZE_MASK) >> \ PRFCNT_FEATURES_COUNTER_BLOCK_SIZE_SHIFT) /* * Begin MMU TRANSCFG register values */ #define AS_TRANSCFG_ADRMODE_LEGACY 0 #define AS_TRANSCFG_ADRMODE_UNMAPPED 1 #define AS_TRANSCFG_ADRMODE_IDENTITY 2 #define AS_TRANSCFG_ADRMODE_AARCH64_4K 6 #define AS_TRANSCFG_ADRMODE_AARCH64_64K 8 #define AS_TRANSCFG_ADRMODE_MASK 0xF /* * Begin TRANSCFG register values */ #define AS_TRANSCFG_PTW_MEMATTR_MASK (3ull << 24) #define AS_TRANSCFG_PTW_MEMATTR_NON_CACHEABLE (1ull << 24) #define AS_TRANSCFG_PTW_MEMATTR_WRITE_BACK (2ull << 24) #define AS_TRANSCFG_PTW_SH_MASK ((3ull << 28)) #define AS_TRANSCFG_PTW_SH_OS (2ull << 28) #define AS_TRANSCFG_PTW_SH_IS (3ull << 28) #define AS_TRANSCFG_R_ALLOCATE (1ull << 30) /* * Begin Command Values */ /* AS_COMMAND register commands */ #define AS_COMMAND_NOP 0x00 /* NOP Operation */ #define AS_COMMAND_UPDATE 0x01 /* Broadcasts the values in AS_TRANSTAB and ASn_MEMATTR to all MMUs */ #define AS_COMMAND_LOCK 0x02 /* Issue a lock region command to all MMUs */ #define AS_COMMAND_UNLOCK 0x03 /* Issue a flush region command to all MMUs */ /* Flush all L2 caches then issue a flush region command to all MMUs */ #define AS_COMMAND_FLUSH_PT 0x04 /* Wait for memory accesses to complete, flush all the L1s cache then flush all * L2 caches then issue a flush region command to all MMUs */ #define AS_COMMAND_FLUSH_MEM 0x05 /* AS_LOCKADDR register */ #define AS_LOCKADDR_LOCKADDR_SIZE_SHIFT GPU_U(0) #define AS_LOCKADDR_LOCKADDR_SIZE_MASK \ (GPU_U(0x3F) << AS_LOCKADDR_LOCKADDR_SIZE_SHIFT) #define AS_LOCKADDR_LOCKADDR_SIZE_GET(reg_val) \ (((reg_val)&AS_LOCKADDR_LOCKADDR_SIZE_MASK) >> \ AS_LOCKADDR_LOCKADDR_SIZE_SHIFT) #define AS_LOCKADDR_LOCKADDR_SIZE_SET(reg_val, value) \ (((reg_val) & ~AS_LOCKADDR_LOCKADDR_SIZE_MASK) | \ (((value) << AS_LOCKADDR_LOCKADDR_SIZE_SHIFT) & \ AS_LOCKADDR_LOCKADDR_SIZE_MASK)) #define AS_LOCKADDR_LOCKADDR_BASE_SHIFT GPU_U(12) #define AS_LOCKADDR_LOCKADDR_BASE_MASK \ (GPU_ULL(0xFFFFFFFFFFFFF) << AS_LOCKADDR_LOCKADDR_BASE_SHIFT) #define AS_LOCKADDR_LOCKADDR_BASE_GET(reg_val) \ (((reg_val)&AS_LOCKADDR_LOCKADDR_BASE_MASK) >> \ AS_LOCKADDR_LOCKADDR_BASE_SHIFT) #define AS_LOCKADDR_LOCKADDR_BASE_SET(reg_val, value) \ (((reg_val) & ~AS_LOCKADDR_LOCKADDR_BASE_MASK) | \ (((value) << AS_LOCKADDR_LOCKADDR_BASE_SHIFT) & \ AS_LOCKADDR_LOCKADDR_BASE_MASK)) #define AS_LOCKADDR_FLUSH_SKIP_LEVELS_SHIFT (6) #define AS_LOCKADDR_FLUSH_SKIP_LEVELS_MASK ((0xF) << AS_LOCKADDR_FLUSH_SKIP_LEVELS_SHIFT) #define AS_LOCKADDR_FLUSH_SKIP_LEVELS_SET(reg_val, value) \ (((reg_val) & ~AS_LOCKADDR_FLUSH_SKIP_LEVELS_MASK) | \ ((value << AS_LOCKADDR_FLUSH_SKIP_LEVELS_SHIFT) & AS_LOCKADDR_FLUSH_SKIP_LEVELS_MASK)) /* GPU_STATUS values */ #define GPU_STATUS_PRFCNT_ACTIVE (1 << 2) /* Set if the performance counters are active. */ #define GPU_STATUS_CYCLE_COUNT_ACTIVE (1 << 6) /* Set if the cycle counter is active. */ #define GPU_STATUS_PROTECTED_MODE_ACTIVE (1 << 7) /* Set if protected mode is active */ /* PRFCNT_CONFIG register values */ #define PRFCNT_CONFIG_MODE_SHIFT 0 /* Counter mode position. */ #define PRFCNT_CONFIG_AS_SHIFT 4 /* Address space bitmap position. */ #define PRFCNT_CONFIG_SETSELECT_SHIFT 8 /* Set select position. */ /* The performance counters are disabled. */ #define PRFCNT_CONFIG_MODE_OFF 0 /* The performance counters are enabled, but are only written out when a * PRFCNT_SAMPLE command is issued using the GPU_COMMAND register. */ #define PRFCNT_CONFIG_MODE_MANUAL 1 /* The performance counters are enabled, and are written out each time a tile * finishes rendering. */ #define PRFCNT_CONFIG_MODE_TILE 2 /* AS_MEMATTR values from MMU_MEMATTR_STAGE1: */ /* Use GPU implementation-defined caching policy. */ #define AS_MEMATTR_IMPL_DEF_CACHE_POLICY 0x88ull /* The attribute set to force all resources to be cached. */ #define AS_MEMATTR_FORCE_TO_CACHE_ALL 0x8Full /* Inner write-alloc cache setup, no outer caching */ #define AS_MEMATTR_WRITE_ALLOC 0x8Dull /* Use GPU implementation-defined caching policy. */ #define AS_MEMATTR_LPAE_IMPL_DEF_CACHE_POLICY 0x48ull /* The attribute set to force all resources to be cached. */ #define AS_MEMATTR_LPAE_FORCE_TO_CACHE_ALL 0x4Full /* Inner write-alloc cache setup, no outer caching */ #define AS_MEMATTR_LPAE_WRITE_ALLOC 0x4Dull /* Set to implementation defined, outer caching */ #define AS_MEMATTR_LPAE_OUTER_IMPL_DEF 0x88ull /* Set to write back memory, outer caching */ #define AS_MEMATTR_LPAE_OUTER_WA 0x8Dull /* There is no LPAE support for non-cacheable, since the memory type is always * write-back. * Marking this setting as reserved for LPAE */ #define AS_MEMATTR_LPAE_NON_CACHEABLE_RESERVED /* L2_MMU_CONFIG register */ #define L2_MMU_CONFIG_ALLOW_SNOOP_DISPARITY_SHIFT (23) #define L2_MMU_CONFIG_ALLOW_SNOOP_DISPARITY (0x1 << L2_MMU_CONFIG_ALLOW_SNOOP_DISPARITY_SHIFT) /* End L2_MMU_CONFIG register */ /* THREAD_* registers */ /* THREAD_FEATURES IMPLEMENTATION_TECHNOLOGY values */ #define IMPLEMENTATION_UNSPECIFIED 0 #define IMPLEMENTATION_SILICON 1 #define IMPLEMENTATION_FPGA 2 #define IMPLEMENTATION_MODEL 3 /* Default values when registers are not supported by the implemented hardware */ #define THREAD_MT_DEFAULT 256 #define THREAD_MWS_DEFAULT 256 #define THREAD_MBS_DEFAULT 256 #define THREAD_MR_DEFAULT 1024 #define THREAD_MTQ_DEFAULT 4 #define THREAD_MTGS_DEFAULT 10 /* End THREAD_* registers */ /* SHADER_CONFIG register */ #define SC_LS_ALLOW_ATTR_TYPES (1ul << 16) #define SC_TLS_HASH_ENABLE (1ul << 17) #define SC_LS_ATTR_CHECK_DISABLE (1ul << 18) #define SC_VAR_ALGORITHM (1ul << 29) /* End SHADER_CONFIG register */ /* TILER_CONFIG register */ #define TC_CLOCK_GATE_OVERRIDE (1ul << 0) /* End TILER_CONFIG register */ /* L2_CONFIG register */ #define L2_CONFIG_SIZE_SHIFT 16 #define L2_CONFIG_SIZE_MASK (0xFFul << L2_CONFIG_SIZE_SHIFT) #define L2_CONFIG_HASH_SHIFT 24 #define L2_CONFIG_HASH_MASK (0xFFul << L2_CONFIG_HASH_SHIFT) #define L2_CONFIG_ASN_HASH_ENABLE_SHIFT 24 #define L2_CONFIG_ASN_HASH_ENABLE_MASK (1ul << L2_CONFIG_ASN_HASH_ENABLE_SHIFT) /* End L2_CONFIG register */ /* AMBA_FEATURES register */ #define AMBA_FEATURES_ACE_LITE_SHIFT GPU_U(0) #define AMBA_FEATURES_ACE_LITE_MASK (GPU_U(0x1) << AMBA_FEATURES_ACE_LITE_SHIFT) #define AMBA_FEATURES_ACE_LITE_GET(reg_val) \ (((reg_val)&AMBA_FEATURES_ACE_LITE_MASK) >> \ AMBA_FEATURES_ACE_LITE_SHIFT) #define AMBA_FEATURES_ACE_LITE_SET(reg_val, value) \ (((reg_val) & ~AMBA_FEATURES_ACE_LITE_MASK) | \ (((value) << AMBA_FEATURES_ACE_LITE_SHIFT) & \ AMBA_FEATURES_ACE_LITE_MASK)) #define AMBA_FEATURES_ACE_SHIFT GPU_U(1) #define AMBA_FEATURES_ACE_MASK (GPU_U(0x1) << AMBA_FEATURES_ACE_SHIFT) #define AMBA_FEATURES_ACE_GET(reg_val) \ (((reg_val)&AMBA_FEATURES_ACE_MASK) >> AMBA_FEATURES_ACE_SHIFT) #define AMBA_FEATURES_ACE_SET(reg_val, value) \ (((reg_val) & ~AMBA_FEATURES_ACE_MASK) | \ (((value) << AMBA_FEATURES_ACE_SHIFT) & AMBA_FEATURES_ACE_MASK)) #define AMBA_FEATURES_MEMORY_CACHE_SUPPORT_SHIFT GPU_U(5) #define AMBA_FEATURES_MEMORY_CACHE_SUPPORT_MASK \ (GPU_U(0x1) << AMBA_FEATURES_MEMORY_CACHE_SUPPORT_SHIFT) #define AMBA_FEATURES_MEMORY_CACHE_SUPPORT_GET(reg_val) \ (((reg_val)&AMBA_FEATURES_MEMORY_CACHE_SUPPORT_MASK) >> \ AMBA_FEATURES_MEMORY_CACHE_SUPPORT_SHIFT) #define AMBA_FEATURES_MEMORY_CACHE_SUPPORT_SET(reg_val, value) \ (((reg_val) & ~AMBA_FEATURES_MEMORY_CACHE_SUPPORT_MASK) | \ (((value) << AMBA_FEATURES_MEMORY_CACHE_SUPPORT_SHIFT) & \ AMBA_FEATURES_MEMORY_CACHE_SUPPORT_MASK)) #define AMBA_FEATURES_INVALIDATE_HINT_SHIFT GPU_U(6) #define AMBA_FEATURES_INVALIDATE_HINT_MASK \ (GPU_U(0x1) << AMBA_FEATURES_INVALIDATE_HINT_SHIFT) #define AMBA_FEATURES_INVALIDATE_HINT_GET(reg_val) \ (((reg_val)&AMBA_FEATURES_INVALIDATE_HINT_MASK) >> \ AMBA_FEATURES_INVALIDATE_HINT_SHIFT) #define AMBA_FEATURES_INVALIDATE_HINT_SET(reg_val, value) \ (((reg_val) & ~AMBA_FEATURES_INVALIDATE_HINT_MASK) | \ (((value) << AMBA_FEATURES_INVALIDATE_HINT_SHIFT) & \ AMBA_FEATURES_INVALIDATE_HINT_MASK)) /* AMBA_ENABLE register */ #define AMBA_ENABLE_COHERENCY_PROTOCOL_SHIFT GPU_U(0) #define AMBA_ENABLE_COHERENCY_PROTOCOL_MASK \ (GPU_U(0x1F) << AMBA_ENABLE_COHERENCY_PROTOCOL_SHIFT) #define AMBA_ENABLE_COHERENCY_PROTOCOL_GET(reg_val) \ (((reg_val)&AMBA_ENABLE_COHERENCY_PROTOCOL_MASK) >> \ AMBA_ENABLE_COHERENCY_PROTOCOL_SHIFT) #define AMBA_ENABLE_COHERENCY_PROTOCOL_SET(reg_val, value) \ (((reg_val) & ~AMBA_ENABLE_COHERENCY_PROTOCOL_MASK) | \ (((value) << AMBA_ENABLE_COHERENCY_PROTOCOL_SHIFT) & \ AMBA_ENABLE_COHERENCY_PROTOCOL_MASK)) /* AMBA_ENABLE_coherency_protocol values */ #define AMBA_ENABLE_COHERENCY_PROTOCOL_ACE_LITE 0x0 #define AMBA_ENABLE_COHERENCY_PROTOCOL_ACE 0x1 #define AMBA_ENABLE_COHERENCY_PROTOCOL_NO_COHERENCY 0x1F /* End of AMBA_ENABLE_coherency_protocol values */ #define AMBA_ENABLE_MEMORY_CACHE_SUPPORT_SHIFT GPU_U(5) #define AMBA_ENABLE_MEMORY_CACHE_SUPPORT_MASK \ (GPU_U(0x1) << AMBA_ENABLE_MEMORY_CACHE_SUPPORT_SHIFT) #define AMBA_ENABLE_MEMORY_CACHE_SUPPORT_GET(reg_val) \ (((reg_val)&AMBA_ENABLE_MEMORY_CACHE_SUPPORT_MASK) >> \ AMBA_ENABLE_MEMORY_CACHE_SUPPORT_SHIFT) #define AMBA_ENABLE_MEMORY_CACHE_SUPPORT_SET(reg_val, value) \ (((reg_val) & ~AMBA_ENABLE_MEMORY_CACHE_SUPPORT_MASK) | \ (((value) << AMBA_ENABLE_MEMORY_CACHE_SUPPORT_SHIFT) & \ AMBA_ENABLE_MEMORY_CACHE_SUPPORT_MASK)) #define AMBA_ENABLE_INVALIDATE_HINT_SHIFT GPU_U(6) #define AMBA_ENABLE_INVALIDATE_HINT_MASK \ (GPU_U(0x1) << AMBA_ENABLE_INVALIDATE_HINT_SHIFT) #define AMBA_ENABLE_INVALIDATE_HINT_GET(reg_val) \ (((reg_val)&AMBA_ENABLE_INVALIDATE_HINT_MASK) >> \ AMBA_ENABLE_INVALIDATE_HINT_SHIFT) #define AMBA_ENABLE_INVALIDATE_HINT_SET(reg_val, value) \ (((reg_val) & ~AMBA_ENABLE_INVALIDATE_HINT_MASK) | \ (((value) << AMBA_ENABLE_INVALIDATE_HINT_SHIFT) & \ AMBA_ENABLE_INVALIDATE_HINT_MASK)) /* IDVS_GROUP register */ #define IDVS_GROUP_SIZE_SHIFT (16) #define IDVS_GROUP_MAX_SIZE (0x3F) /* SYSC_ALLOC read IDs */ #define SYSC_ALLOC_ID_R_OTHER 0x00 #define SYSC_ALLOC_ID_R_CSF 0x02 #define SYSC_ALLOC_ID_R_MMU 0x04 #define SYSC_ALLOC_ID_R_TILER_VERT 0x08 #define SYSC_ALLOC_ID_R_TILER_PTR 0x09 #define SYSC_ALLOC_ID_R_TILER_INDEX 0x0A #define SYSC_ALLOC_ID_R_TILER_OTHER 0x0B #define SYSC_ALLOC_ID_R_IC 0x10 #define SYSC_ALLOC_ID_R_ATTR 0x11 #define SYSC_ALLOC_ID_R_SCM 0x12 #define SYSC_ALLOC_ID_R_FSDC 0x13 #define SYSC_ALLOC_ID_R_VL 0x14 #define SYSC_ALLOC_ID_R_PLR 0x15 #define SYSC_ALLOC_ID_R_TEX 0x18 #define SYSC_ALLOC_ID_R_LSC 0x1c /* SYSC_ALLOC write IDs */ #define SYSC_ALLOC_ID_W_OTHER 0x00 #define SYSC_ALLOC_ID_W_CSF 0x02 #define SYSC_ALLOC_ID_W_PCB 0x07 #define SYSC_ALLOC_ID_W_TILER_PTR 0x09 #define SYSC_ALLOC_ID_W_TILER_VERT_PLIST 0x0A #define SYSC_ALLOC_ID_W_TILER_OTHER 0x0B #define SYSC_ALLOC_ID_W_L2_EVICT 0x0C #define SYSC_ALLOC_ID_W_L2_FLUSH 0x0D #define SYSC_ALLOC_ID_W_TIB_COLOR 0x10 #define SYSC_ALLOC_ID_W_TIB_COLOR_AFBCH 0x11 #define SYSC_ALLOC_ID_W_TIB_COLOR_AFBCB 0x12 #define SYSC_ALLOC_ID_W_TIB_CRC 0x13 #define SYSC_ALLOC_ID_W_TIB_DS 0x14 #define SYSC_ALLOC_ID_W_TIB_DS_AFBCH 0x15 #define SYSC_ALLOC_ID_W_TIB_DS_AFBCB 0x16 #define SYSC_ALLOC_ID_W_LSC 0x1C /* SYSC_ALLOC values */ #define SYSC_ALLOC_L2_ALLOC 0x0 #define SYSC_ALLOC_NEVER_ALLOC 0x2 #define SYSC_ALLOC_ALWAYS_ALLOC 0x3 #define SYSC_ALLOC_PTL_ALLOC 0x4 #define SYSC_ALLOC_L2_PTL_ALLOC 0x5 /* SYSC_ALLOC register */ #define SYSC_ALLOC_R_SYSC_ALLOC0_SHIFT (0) #define SYSC_ALLOC_R_SYSC_ALLOC0_MASK ((0xF) << SYSC_ALLOC_R_SYSC_ALLOC0_SHIFT) #define SYSC_ALLOC_R_SYSC_ALLOC0_GET(reg_val) \ (((reg_val)&SYSC_ALLOC_R_SYSC_ALLOC0_MASK) >> \ SYSC_ALLOC_R_SYSC_ALLOC0_SHIFT) #define SYSC_ALLOC_R_SYSC_ALLOC0_SET(reg_val, value) \ (((reg_val) & ~SYSC_ALLOC_R_SYSC_ALLOC0_MASK) | \ (((value) << SYSC_ALLOC_R_SYSC_ALLOC0_SHIFT) & \ SYSC_ALLOC_R_SYSC_ALLOC0_MASK)) /* End of SYSC_ALLOC_R_SYSC_ALLOC0 values */ #define SYSC_ALLOC_W_SYSC_ALLOC0_SHIFT (4) #define SYSC_ALLOC_W_SYSC_ALLOC0_MASK ((0xF) << SYSC_ALLOC_W_SYSC_ALLOC0_SHIFT) #define SYSC_ALLOC_W_SYSC_ALLOC0_GET(reg_val) \ (((reg_val)&SYSC_ALLOC_W_SYSC_ALLOC0_MASK) >> \ SYSC_ALLOC_W_SYSC_ALLOC0_SHIFT) #define SYSC_ALLOC_W_SYSC_ALLOC0_SET(reg_val, value) \ (((reg_val) & ~SYSC_ALLOC_W_SYSC_ALLOC0_MASK) | \ (((value) << SYSC_ALLOC_W_SYSC_ALLOC0_SHIFT) & \ SYSC_ALLOC_W_SYSC_ALLOC0_MASK)) /* End of SYSC_ALLOC_W_SYSC_ALLOC0 values */ #define SYSC_ALLOC_R_SYSC_ALLOC1_SHIFT (8) #define SYSC_ALLOC_R_SYSC_ALLOC1_MASK ((0xF) << SYSC_ALLOC_R_SYSC_ALLOC1_SHIFT) #define SYSC_ALLOC_R_SYSC_ALLOC1_GET(reg_val) \ (((reg_val)&SYSC_ALLOC_R_SYSC_ALLOC1_MASK) >> \ SYSC_ALLOC_R_SYSC_ALLOC1_SHIFT) #define SYSC_ALLOC_R_SYSC_ALLOC1_SET(reg_val, value) \ (((reg_val) & ~SYSC_ALLOC_R_SYSC_ALLOC1_MASK) | \ (((value) << SYSC_ALLOC_R_SYSC_ALLOC1_SHIFT) & \ SYSC_ALLOC_R_SYSC_ALLOC1_MASK)) /* End of SYSC_ALLOC_R_SYSC_ALLOC1 values */ #define SYSC_ALLOC_W_SYSC_ALLOC1_SHIFT (12) #define SYSC_ALLOC_W_SYSC_ALLOC1_MASK ((0xF) << SYSC_ALLOC_W_SYSC_ALLOC1_SHIFT) #define SYSC_ALLOC_W_SYSC_ALLOC1_GET(reg_val) \ (((reg_val)&SYSC_ALLOC_W_SYSC_ALLOC1_MASK) >> \ SYSC_ALLOC_W_SYSC_ALLOC1_SHIFT) #define SYSC_ALLOC_W_SYSC_ALLOC1_SET(reg_val, value) \ (((reg_val) & ~SYSC_ALLOC_W_SYSC_ALLOC1_MASK) | \ (((value) << SYSC_ALLOC_W_SYSC_ALLOC1_SHIFT) & \ SYSC_ALLOC_W_SYSC_ALLOC1_MASK)) /* End of SYSC_ALLOC_W_SYSC_ALLOC1 values */ #define SYSC_ALLOC_R_SYSC_ALLOC2_SHIFT (16) #define SYSC_ALLOC_R_SYSC_ALLOC2_MASK ((0xF) << SYSC_ALLOC_R_SYSC_ALLOC2_SHIFT) #define SYSC_ALLOC_R_SYSC_ALLOC2_GET(reg_val) \ (((reg_val)&SYSC_ALLOC_R_SYSC_ALLOC2_MASK) >> \ SYSC_ALLOC_R_SYSC_ALLOC2_SHIFT) #define SYSC_ALLOC_R_SYSC_ALLOC2_SET(reg_val, value) \ (((reg_val) & ~SYSC_ALLOC_R_SYSC_ALLOC2_MASK) | \ (((value) << SYSC_ALLOC_R_SYSC_ALLOC2_SHIFT) & \ SYSC_ALLOC_R_SYSC_ALLOC2_MASK)) /* End of SYSC_ALLOC_R_SYSC_ALLOC2 values */ #define SYSC_ALLOC_W_SYSC_ALLOC2_SHIFT (20) #define SYSC_ALLOC_W_SYSC_ALLOC2_MASK ((0xF) << SYSC_ALLOC_W_SYSC_ALLOC2_SHIFT) #define SYSC_ALLOC_W_SYSC_ALLOC2_GET(reg_val) \ (((reg_val)&SYSC_ALLOC_W_SYSC_ALLOC2_MASK) >> \ SYSC_ALLOC_W_SYSC_ALLOC2_SHIFT) #define SYSC_ALLOC_W_SYSC_ALLOC2_SET(reg_val, value) \ (((reg_val) & ~SYSC_ALLOC_W_SYSC_ALLOC2_MASK) | \ (((value) << SYSC_ALLOC_W_SYSC_ALLOC2_SHIFT) & \ SYSC_ALLOC_W_SYSC_ALLOC2_MASK)) /* End of SYSC_ALLOC_W_SYSC_ALLOC2 values */ #define SYSC_ALLOC_R_SYSC_ALLOC3_SHIFT (24) #define SYSC_ALLOC_R_SYSC_ALLOC3_MASK ((0xF) << SYSC_ALLOC_R_SYSC_ALLOC3_SHIFT) #define SYSC_ALLOC_R_SYSC_ALLOC3_GET(reg_val) \ (((reg_val)&SYSC_ALLOC_R_SYSC_ALLOC3_MASK) >> \ SYSC_ALLOC_R_SYSC_ALLOC3_SHIFT) #define SYSC_ALLOC_R_SYSC_ALLOC3_SET(reg_val, value) \ (((reg_val) & ~SYSC_ALLOC_R_SYSC_ALLOC3_MASK) | \ (((value) << SYSC_ALLOC_R_SYSC_ALLOC3_SHIFT) & \ SYSC_ALLOC_R_SYSC_ALLOC3_MASK)) /* End of SYSC_ALLOC_R_SYSC_ALLOC3 values */ #define SYSC_ALLOC_W_SYSC_ALLOC3_SHIFT (28) #define SYSC_ALLOC_W_SYSC_ALLOC3_MASK ((0xF) << SYSC_ALLOC_W_SYSC_ALLOC3_SHIFT) #define SYSC_ALLOC_W_SYSC_ALLOC3_GET(reg_val) \ (((reg_val)&SYSC_ALLOC_W_SYSC_ALLOC3_MASK) >> \ SYSC_ALLOC_W_SYSC_ALLOC3_SHIFT) #define SYSC_ALLOC_W_SYSC_ALLOC3_SET(reg_val, value) \ (((reg_val) & ~SYSC_ALLOC_W_SYSC_ALLOC3_MASK) | \ (((value) << SYSC_ALLOC_W_SYSC_ALLOC3_SHIFT) & \ SYSC_ALLOC_W_SYSC_ALLOC3_MASK)) /* End of SYSC_ALLOC_W_SYSC_ALLOC3 values */ /* Include POWER_CHANGED_SINGLE in debug builds for use in irq latency test. */ #ifdef CONFIG_MALI_BIFROST_DEBUG #undef GPU_IRQ_REG_ALL #define GPU_IRQ_REG_ALL (GPU_IRQ_REG_COMMON | POWER_CHANGED_SINGLE) #endif /* CONFIG_MALI_BIFROST_DEBUG */ #endif /* _KBASE_GPU_REGMAP_H_ */