// SPDX-License-Identifier: GPL-2.0 WITH Linux-syscall-note /* * * (C) COPYRIGHT 2010, 2012-2015, 2017-2022 ARM Limited. All rights reserved. * * This program is free software and is provided to you under the terms of the * GNU General Public License version 2 as published by the Free Software * Foundation, and any use by you of this program is subject to the terms * of such GNU license. * * This program is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the * GNU General Public License for more details. * * You should have received a copy of the GNU General Public License * along with this program; if not, you can access it online at * http://www.gnu.org/licenses/gpl-2.0.html. * */ /* * Model Linux Framework interfaces. */ #include #include #include "backend/gpu/mali_kbase_model_linux.h" #include "device/mali_kbase_device.h" #include "mali_kbase_irq_internal.h" #include struct model_irq_data { struct kbase_device *kbdev; struct work_struct work; }; static void serve_job_irq(struct work_struct *work) { struct model_irq_data *data = container_of(work, struct model_irq_data, work); struct kbase_device *kbdev = data->kbdev; /* Make sure no worker is already serving this IRQ */ while (atomic_cmpxchg(&kbdev->serving_job_irq, 1, 0) == 1) { u32 val; while ((val = kbase_reg_read(kbdev, JOB_CONTROL_REG(JOB_IRQ_STATUS)))) { unsigned long flags; /* Handle the IRQ */ spin_lock_irqsave(&kbdev->hwaccess_lock, flags); #if MALI_USE_CSF kbase_csf_interrupt(kbdev, val); #else kbase_job_done(kbdev, val); #endif spin_unlock_irqrestore(&kbdev->hwaccess_lock, flags); } } kmem_cache_free(kbdev->irq_slab, data); } static void serve_gpu_irq(struct work_struct *work) { struct model_irq_data *data = container_of(work, struct model_irq_data, work); struct kbase_device *kbdev = data->kbdev; /* Make sure no worker is already serving this IRQ */ while (atomic_cmpxchg(&kbdev->serving_gpu_irq, 1, 0) == 1) { u32 val; while ((val = kbase_reg_read(kbdev, GPU_CONTROL_REG(GPU_IRQ_STATUS)))) { /* Handle the IRQ */ kbase_gpu_interrupt(kbdev, val); } } kmem_cache_free(kbdev->irq_slab, data); } static void serve_mmu_irq(struct work_struct *work) { struct model_irq_data *data = container_of(work, struct model_irq_data, work); struct kbase_device *kbdev = data->kbdev; /* Make sure no worker is already serving this IRQ */ if (atomic_cmpxchg(&kbdev->serving_mmu_irq, 1, 0) == 1) { u32 val; while ((val = kbase_reg_read(kbdev, MMU_REG(MMU_IRQ_STATUS)))) { /* Handle the IRQ */ kbase_mmu_interrupt(kbdev, val); } } kmem_cache_free(kbdev->irq_slab, data); } void gpu_device_raise_irq(void *model, u32 irq) { struct model_irq_data *data; struct kbase_device *kbdev = gpu_device_get_data(model); KBASE_DEBUG_ASSERT(kbdev); data = kmem_cache_alloc(kbdev->irq_slab, GFP_ATOMIC); if (data == NULL) return; data->kbdev = kbdev; switch (irq) { case MODEL_LINUX_JOB_IRQ: INIT_WORK(&data->work, serve_job_irq); atomic_set(&kbdev->serving_job_irq, 1); break; case MODEL_LINUX_GPU_IRQ: INIT_WORK(&data->work, serve_gpu_irq); atomic_set(&kbdev->serving_gpu_irq, 1); break; case MODEL_LINUX_MMU_IRQ: INIT_WORK(&data->work, serve_mmu_irq); atomic_set(&kbdev->serving_mmu_irq, 1); break; default: dev_warn(kbdev->dev, "Unknown IRQ"); kmem_cache_free(kbdev->irq_slab, data); data = NULL; break; } if (data != NULL) queue_work(kbdev->irq_workq, &data->work); } void kbase_reg_write(struct kbase_device *kbdev, u32 offset, u32 value) { unsigned long flags; spin_lock_irqsave(&kbdev->reg_op_lock, flags); midgard_model_write_reg(kbdev->model, offset, value); spin_unlock_irqrestore(&kbdev->reg_op_lock, flags); } KBASE_EXPORT_TEST_API(kbase_reg_write); u32 kbase_reg_read(struct kbase_device *kbdev, u32 offset) { unsigned long flags; u32 val; spin_lock_irqsave(&kbdev->reg_op_lock, flags); midgard_model_read_reg(kbdev->model, offset, &val); spin_unlock_irqrestore(&kbdev->reg_op_lock, flags); return val; } KBASE_EXPORT_TEST_API(kbase_reg_read); int kbase_install_interrupts(struct kbase_device *kbdev) { KBASE_DEBUG_ASSERT(kbdev); atomic_set(&kbdev->serving_job_irq, 0); atomic_set(&kbdev->serving_gpu_irq, 0); atomic_set(&kbdev->serving_mmu_irq, 0); kbdev->irq_workq = alloc_ordered_workqueue("dummy irq queue", 0); if (kbdev->irq_workq == NULL) return -ENOMEM; kbdev->irq_slab = kmem_cache_create("dummy_irq_slab", sizeof(struct model_irq_data), 0, 0, NULL); if (kbdev->irq_slab == NULL) { destroy_workqueue(kbdev->irq_workq); return -ENOMEM; } return 0; } void kbase_release_interrupts(struct kbase_device *kbdev) { KBASE_DEBUG_ASSERT(kbdev); destroy_workqueue(kbdev->irq_workq); kmem_cache_destroy(kbdev->irq_slab); } void kbase_synchronize_irqs(struct kbase_device *kbdev) { KBASE_DEBUG_ASSERT(kbdev); flush_workqueue(kbdev->irq_workq); } KBASE_EXPORT_TEST_API(kbase_synchronize_irqs); int kbase_set_custom_irq_handler(struct kbase_device *kbdev, irq_handler_t custom_handler, int irq_type) { return 0; } KBASE_EXPORT_TEST_API(kbase_set_custom_irq_handler); irqreturn_t kbase_gpu_irq_test_handler(int irq, void *data, u32 val) { if (!val) return IRQ_NONE; return IRQ_HANDLED; } KBASE_EXPORT_TEST_API(kbase_gpu_irq_test_handler); int kbase_gpu_device_create(struct kbase_device *kbdev) { kbdev->model = midgard_model_create(kbdev); if (kbdev->model == NULL) return -ENOMEM; spin_lock_init(&kbdev->reg_op_lock); return 0; } /** * kbase_gpu_device_destroy - Destroy GPU device * * @kbdev: kbase device */ void kbase_gpu_device_destroy(struct kbase_device *kbdev) { midgard_model_destroy(kbdev->model); }