// SPDX-License-Identifier: GPL-2.0 WITH Linux-syscall-note /* * * (C) COPYRIGHT 2014-2023 ARM Limited. All rights reserved. * * This program is free software and is provided to you under the terms of the * GNU General Public License version 2 as published by the Free Software * Foundation, and any use by you of this program is subject to the terms * of such GNU license. * * This program is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the * GNU General Public License for more details. * * You should have received a copy of the GNU General Public License * along with this program; if not, you can access it online at * http://www.gnu.org/licenses/gpl-2.0.html. * */ #include #include #if MALI_USE_CSF #include #include #include #endif #include #include #include void kbase_backend_get_gpu_time_norequest(struct kbase_device *kbdev, u64 *cycle_counter, u64 *system_time, struct timespec64 *ts) { u32 hi1, hi2; if (cycle_counter) *cycle_counter = kbase_backend_get_cycle_cnt(kbdev); if (system_time) { /* Read hi, lo, hi to ensure a coherent u64 */ do { hi1 = kbase_reg_read(kbdev, GPU_CONTROL_REG(TIMESTAMP_HI)); *system_time = kbase_reg_read(kbdev, GPU_CONTROL_REG(TIMESTAMP_LO)); hi2 = kbase_reg_read(kbdev, GPU_CONTROL_REG(TIMESTAMP_HI)); } while (hi1 != hi2); *system_time |= (((u64) hi1) << 32); } /* Record the CPU's idea of current time */ if (ts != NULL) #if (KERNEL_VERSION(4, 17, 0) > LINUX_VERSION_CODE) *ts = ktime_to_timespec64(ktime_get_raw()); #else ktime_get_raw_ts64(ts); #endif } #if !MALI_USE_CSF /** * timedwait_cycle_count_active() - Timed wait till CYCLE_COUNT_ACTIVE is active * * @kbdev: Kbase device * * Return: true if CYCLE_COUNT_ACTIVE is active within the timeout. */ static bool timedwait_cycle_count_active(struct kbase_device *kbdev) { #if IS_ENABLED(CONFIG_MALI_BIFROST_NO_MALI) return true; #else bool success = false; const unsigned int timeout = 100; const unsigned long remaining = jiffies + msecs_to_jiffies(timeout); while (time_is_after_jiffies(remaining)) { if ((kbase_reg_read(kbdev, GPU_CONTROL_REG(GPU_STATUS)) & GPU_STATUS_CYCLE_COUNT_ACTIVE)) { success = true; break; } } return success; #endif } #endif void kbase_backend_get_gpu_time(struct kbase_device *kbdev, u64 *cycle_counter, u64 *system_time, struct timespec64 *ts) { #if !MALI_USE_CSF kbase_pm_request_gpu_cycle_counter(kbdev); WARN_ONCE(kbdev->pm.backend.l2_state != KBASE_L2_ON, "L2 not powered up"); WARN_ONCE((!timedwait_cycle_count_active(kbdev)), "Timed out on CYCLE_COUNT_ACTIVE"); #endif kbase_backend_get_gpu_time_norequest(kbdev, cycle_counter, system_time, ts); #if !MALI_USE_CSF kbase_pm_release_gpu_cycle_counter(kbdev); #endif } unsigned int kbase_get_timeout_ms(struct kbase_device *kbdev, enum kbase_timeout_selector selector) { /* Timeout calculation: * dividing number of cycles by freq in KHz automatically gives value * in milliseconds. nr_cycles will have to be multiplied by 1e3 to * get result in microseconds, and 1e6 to get result in nanoseconds. */ u64 timeout, nr_cycles = 0; u64 freq_khz; /* Only for debug messages, safe default in case it's mis-maintained */ const char *selector_str = "(unknown)"; if (!kbdev->lowest_gpu_freq_khz) { dev_dbg(kbdev->dev, "Lowest frequency uninitialized! Using reference frequency for scaling"); freq_khz = DEFAULT_REF_TIMEOUT_FREQ_KHZ; } else { freq_khz = kbdev->lowest_gpu_freq_khz; } switch (selector) { case MMU_AS_INACTIVE_WAIT_TIMEOUT: selector_str = "MMU_AS_INACTIVE_WAIT_TIMEOUT"; nr_cycles = MMU_AS_INACTIVE_WAIT_TIMEOUT_CYCLES; break; case KBASE_TIMEOUT_SELECTOR_COUNT: default: #if !MALI_USE_CSF WARN(1, "Invalid timeout selector used! Using default value"); nr_cycles = JM_DEFAULT_TIMEOUT_CYCLES; break; case JM_DEFAULT_JS_FREE_TIMEOUT: selector_str = "JM_DEFAULT_JS_FREE_TIMEOUT"; nr_cycles = JM_DEFAULT_JS_FREE_TIMEOUT_CYCLES; break; #else /* Use Firmware timeout if invalid selection */ WARN(1, "Invalid timeout selector used! Using CSF Firmware timeout"); fallthrough; case CSF_FIRMWARE_TIMEOUT: selector_str = "CSF_FIRMWARE_TIMEOUT"; /* Any FW timeout cannot be longer than the FW ping interval, after which * the firmware_aliveness_monitor will be triggered and may restart * the GPU if the FW is unresponsive. */ nr_cycles = min(CSF_FIRMWARE_PING_TIMEOUT_CYCLES, CSF_FIRMWARE_TIMEOUT_CYCLES); if (nr_cycles == CSF_FIRMWARE_PING_TIMEOUT_CYCLES) dev_warn(kbdev->dev, "Capping %s to CSF_FIRMWARE_PING_TIMEOUT\n", selector_str); break; case CSF_PM_TIMEOUT: selector_str = "CSF_PM_TIMEOUT"; nr_cycles = CSF_PM_TIMEOUT_CYCLES; break; case CSF_GPU_RESET_TIMEOUT: selector_str = "CSF_GPU_RESET_TIMEOUT"; nr_cycles = CSF_GPU_RESET_TIMEOUT_CYCLES; break; case CSF_CSG_SUSPEND_TIMEOUT: selector_str = "CSF_CSG_SUSPEND_TIMEOUT"; nr_cycles = CSF_CSG_SUSPEND_TIMEOUT_CYCLES; break; case CSF_FIRMWARE_BOOT_TIMEOUT: selector_str = "CSF_FIRMWARE_BOOT_TIMEOUT"; nr_cycles = CSF_FIRMWARE_BOOT_TIMEOUT_CYCLES; break; case CSF_FIRMWARE_PING_TIMEOUT: selector_str = "CSF_FIRMWARE_PING_TIMEOUT"; nr_cycles = CSF_FIRMWARE_PING_TIMEOUT_CYCLES; break; case CSF_SCHED_PROTM_PROGRESS_TIMEOUT: selector_str = "CSF_SCHED_PROTM_PROGRESS_TIMEOUT"; nr_cycles = kbase_csf_timeout_get(kbdev); break; #endif } timeout = div_u64(nr_cycles, freq_khz); if (WARN(timeout > UINT_MAX, "Capping excessive timeout %llums for %s at freq %llukHz to UINT_MAX ms", (unsigned long long)timeout, selector_str, (unsigned long long)freq_khz)) timeout = UINT_MAX; return (unsigned int)timeout; } KBASE_EXPORT_TEST_API(kbase_get_timeout_ms); u64 kbase_backend_get_cycle_cnt(struct kbase_device *kbdev) { u32 hi1, hi2, lo; /* Read hi, lo, hi to ensure a coherent u64 */ do { hi1 = kbase_reg_read(kbdev, GPU_CONTROL_REG(CYCLE_COUNT_HI)); lo = kbase_reg_read(kbdev, GPU_CONTROL_REG(CYCLE_COUNT_LO)); hi2 = kbase_reg_read(kbdev, GPU_CONTROL_REG(CYCLE_COUNT_HI)); } while (hi1 != hi2); return lo | (((u64) hi1) << 32); } #if MALI_USE_CSF u64 __maybe_unused kbase_backend_time_convert_gpu_to_cpu(struct kbase_device *kbdev, u64 gpu_ts) { if (WARN_ON(!kbdev)) return 0; return div64_u64(gpu_ts * kbdev->backend_time.multiplier, kbdev->backend_time.divisor) + kbdev->backend_time.offset; } /** * get_cpu_gpu_time() - Get current CPU and GPU timestamps. * * @kbdev: Kbase device. * @cpu_ts: Output CPU timestamp. * @gpu_ts: Output GPU timestamp. * @gpu_cycle: Output GPU cycle counts. */ static void get_cpu_gpu_time(struct kbase_device *kbdev, u64 *cpu_ts, u64 *gpu_ts, u64 *gpu_cycle) { struct timespec64 ts; kbase_backend_get_gpu_time(kbdev, gpu_cycle, gpu_ts, &ts); if (cpu_ts) *cpu_ts = ts.tv_sec * NSEC_PER_SEC + ts.tv_nsec; } #endif int kbase_backend_time_init(struct kbase_device *kbdev) { #if MALI_USE_CSF u64 cpu_ts = 0; u64 gpu_ts = 0; u64 freq; u64 common_factor; get_cpu_gpu_time(kbdev, &cpu_ts, &gpu_ts, NULL); freq = arch_timer_get_cntfrq(); if (!freq) { dev_warn(kbdev->dev, "arch_timer_get_rate() is zero!"); return -EINVAL; } common_factor = gcd(NSEC_PER_SEC, freq); kbdev->backend_time.multiplier = div64_u64(NSEC_PER_SEC, common_factor); kbdev->backend_time.divisor = div64_u64(freq, common_factor); if (!kbdev->backend_time.divisor) { dev_warn(kbdev->dev, "CPU to GPU divisor is zero!"); return -EINVAL; } kbdev->backend_time.offset = cpu_ts - div64_u64(gpu_ts * kbdev->backend_time.multiplier, kbdev->backend_time.divisor); #endif return 0; }