// SPDX-License-Identifier: GPL-2.0 WITH Linux-syscall-note
|
/*
|
*
|
* (C) COPYRIGHT 2019-2021 ARM Limited. All rights reserved.
|
*
|
* This program is free software and is provided to you under the terms of the
|
* GNU General Public License version 2 as published by the Free Software
|
* Foundation, and any use by you of this program is subject to the terms
|
* of such GNU license.
|
*
|
* This program is distributed in the hope that it will be useful,
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
* GNU General Public License for more details.
|
*
|
* You should have received a copy of the GNU General Public License
|
* along with this program; if not, you can access it online at
|
* http://www.gnu.org/licenses/gpl-2.0.html.
|
*
|
*/
|
|
#include "mali_kbase_csf_protected_memory.h"
|
#include <linux/protected_memory_allocator.h>
|
|
#if IS_ENABLED(CONFIG_OF)
|
#include <linux/of_platform.h>
|
#endif
|
|
int kbase_csf_protected_memory_init(struct kbase_device *const kbdev)
|
{
|
int err = 0;
|
|
#if IS_ENABLED(CONFIG_OF)
|
struct device_node *pma_node = of_parse_phandle(kbdev->dev->of_node,
|
"protected-memory-allocator", 0);
|
if (!pma_node) {
|
dev_info(kbdev->dev, "Protected memory allocator not available\n");
|
} else {
|
struct platform_device *const pdev =
|
of_find_device_by_node(pma_node);
|
|
kbdev->csf.pma_dev = NULL;
|
if (!pdev) {
|
dev_err(kbdev->dev, "Platform device for Protected memory allocator not found\n");
|
} else {
|
kbdev->csf.pma_dev = platform_get_drvdata(pdev);
|
if (!kbdev->csf.pma_dev) {
|
dev_info(kbdev->dev, "Protected memory allocator is not ready\n");
|
err = -EPROBE_DEFER;
|
} else if (!try_module_get(kbdev->csf.pma_dev->owner)) {
|
dev_err(kbdev->dev, "Failed to get Protected memory allocator module\n");
|
err = -ENODEV;
|
} else {
|
dev_info(kbdev->dev, "Protected memory allocator successfully loaded\n");
|
}
|
}
|
of_node_put(pma_node);
|
}
|
#endif
|
|
return err;
|
}
|
|
void kbase_csf_protected_memory_term(struct kbase_device *const kbdev)
|
{
|
if (kbdev->csf.pma_dev)
|
module_put(kbdev->csf.pma_dev->owner);
|
}
|
|
struct protected_memory_allocation **
|
kbase_csf_protected_memory_alloc(
|
struct kbase_device *const kbdev,
|
struct tagged_addr *phys,
|
size_t num_pages,
|
bool is_small_page)
|
{
|
size_t i;
|
struct protected_memory_allocator_device *pma_dev =
|
kbdev->csf.pma_dev;
|
struct protected_memory_allocation **pma = NULL;
|
unsigned int order = KBASE_MEM_POOL_2MB_PAGE_TABLE_ORDER;
|
unsigned int num_pages_order;
|
|
if (is_small_page)
|
order = KBASE_MEM_POOL_4KB_PAGE_TABLE_ORDER;
|
|
num_pages_order = (1u << order);
|
|
/* Ensure the requested num_pages is aligned with
|
* the order type passed as argument.
|
*
|
* pma_alloc_page() will then handle the granularity
|
* of the allocation based on order.
|
*/
|
num_pages = div64_u64(num_pages + num_pages_order - 1, num_pages_order);
|
|
pma = kmalloc_array(num_pages, sizeof(*pma), GFP_KERNEL);
|
|
if (WARN_ON(!pma_dev) || WARN_ON(!phys) || !pma)
|
return NULL;
|
|
for (i = 0; i < num_pages; i++) {
|
phys_addr_t phys_addr;
|
|
pma[i] = pma_dev->ops.pma_alloc_page(pma_dev, order);
|
if (!pma[i])
|
break;
|
|
phys_addr = pma_dev->ops.pma_get_phys_addr(pma_dev, pma[i]);
|
|
if (order) {
|
size_t j;
|
|
*phys++ = as_tagged_tag(phys_addr, HUGE_HEAD | HUGE_PAGE);
|
|
for (j = 1; j < num_pages_order; j++) {
|
*phys++ = as_tagged_tag(phys_addr +
|
PAGE_SIZE * j,
|
HUGE_PAGE);
|
}
|
} else {
|
phys[i] = as_tagged(phys_addr);
|
}
|
}
|
|
if (i != num_pages) {
|
kbase_csf_protected_memory_free(kbdev, pma, i * num_pages_order, is_small_page);
|
return NULL;
|
}
|
|
return pma;
|
}
|
|
void kbase_csf_protected_memory_free(
|
struct kbase_device *const kbdev,
|
struct protected_memory_allocation **pma,
|
size_t num_pages,
|
bool is_small_page)
|
{
|
size_t i;
|
struct protected_memory_allocator_device *pma_dev =
|
kbdev->csf.pma_dev;
|
unsigned int num_pages_order = (1u << KBASE_MEM_POOL_2MB_PAGE_TABLE_ORDER);
|
|
if (is_small_page)
|
num_pages_order = (1u << KBASE_MEM_POOL_4KB_PAGE_TABLE_ORDER);
|
|
if (WARN_ON(!pma_dev) || WARN_ON(!pma))
|
return;
|
|
/* Ensure the requested num_pages is aligned with
|
* the order type passed as argument.
|
*
|
* pma_alloc_page() will then handle the granularity
|
* of the allocation based on order.
|
*/
|
num_pages = div64_u64(num_pages + num_pages_order - 1, num_pages_order);
|
|
for (i = 0; i < num_pages; i++)
|
pma_dev->ops.pma_free_page(pma_dev, pma[i]);
|
|
kfree(pma);
|
}
|