/* * * (C) COPYRIGHT 2017 ARM Limited. All rights reserved. * * This program is free software and is provided to you under the terms of the * GNU General Public License version 2 as published by the Free Software * Foundation, and any use by you of this program is subject to the terms * of such GNU licence. * * A copy of the licence is included with the program, and can also be obtained * from Free Software Foundation, Inc., 51 Franklin Street, Fifth Floor, * Boston, MA 02110-1301, USA. * */ /* * A core availability policy implementing core mask selection from devfreq OPPs * */ #include #include #include #include void kbase_devfreq_set_core_mask(struct kbase_device *kbdev, u64 core_mask) { struct kbasep_pm_ca_policy_devfreq *data = &kbdev->pm.backend.ca_policy_data.devfreq; unsigned long flags; spin_lock_irqsave(&kbdev->hwaccess_lock, flags); data->cores_desired = core_mask; /* Disable any cores that are now unwanted */ data->cores_enabled &= data->cores_desired; kbdev->pm.backend.ca_in_transition = true; /* If there are no cores to be powered off then power on desired cores */ if (!(data->cores_used & ~data->cores_desired)) { data->cores_enabled = data->cores_desired; kbdev->pm.backend.ca_in_transition = false; } kbase_pm_update_cores_state_nolock(kbdev); spin_unlock_irqrestore(&kbdev->hwaccess_lock, flags); dev_dbg(kbdev->dev, "Devfreq policy : new core mask=%llX %llX\n", data->cores_desired, data->cores_enabled); } static void devfreq_init(struct kbase_device *kbdev) { struct kbasep_pm_ca_policy_devfreq *data = &kbdev->pm.backend.ca_policy_data.devfreq; if (kbdev->current_core_mask) { data->cores_enabled = kbdev->current_core_mask; data->cores_desired = kbdev->current_core_mask; } else { data->cores_enabled = kbdev->gpu_props.props.raw_props.shader_present; data->cores_desired = kbdev->gpu_props.props.raw_props.shader_present; } data->cores_used = 0; kbdev->pm.backend.ca_in_transition = false; } static void devfreq_term(struct kbase_device *kbdev) { } static u64 devfreq_get_core_mask(struct kbase_device *kbdev) { return kbdev->pm.backend.ca_policy_data.devfreq.cores_enabled; } static void devfreq_update_core_status(struct kbase_device *kbdev, u64 cores_ready, u64 cores_transitioning) { struct kbasep_pm_ca_policy_devfreq *data = &kbdev->pm.backend.ca_policy_data.devfreq; lockdep_assert_held(&kbdev->hwaccess_lock); data->cores_used = cores_ready | cores_transitioning; /* If in desired state then clear transition flag */ if (data->cores_enabled == data->cores_desired) kbdev->pm.backend.ca_in_transition = false; /* If all undesired cores are now off then power on desired cores. * The direct comparison against cores_enabled limits potential * recursion to one level */ if (!(data->cores_used & ~data->cores_desired) && data->cores_enabled != data->cores_desired) { data->cores_enabled = data->cores_desired; kbase_pm_update_cores_state_nolock(kbdev); kbdev->pm.backend.ca_in_transition = false; } } /* * The struct kbase_pm_ca_policy structure for the devfreq core availability * policy. * * This is the static structure that defines the devfreq core availability power * policy's callback and name. */ const struct kbase_pm_ca_policy kbase_pm_ca_devfreq_policy_ops = { "devfreq", /* name */ devfreq_init, /* init */ devfreq_term, /* term */ devfreq_get_core_mask, /* get_core_mask */ devfreq_update_core_status, /* update_core_status */ 0u, /* flags */ KBASE_PM_CA_POLICY_ID_DEVFREQ, /* id */ };