forked from ~ljy/RK356X_SDK_RELEASE

hc
2024-05-11 04dd17822334871b23ea2862f7798fb0e0007777
kernel/drivers/gpu/drm/nouveau/include/nvkm/subdev/mmu.h
....@@ -1,4 +1,4 @@
1
-/* SPDX-License-Identifier: GPL-2.0 */
1
+/* SPDX-License-Identifier: MIT */
22 #ifndef __NVKM_MMU_H__
33 #define __NVKM_MMU_H__
44 #include <core/subdev.h>
....@@ -17,6 +17,7 @@
1717 bool part:1; /* Region was split from an allocated region by map(). */
1818 bool user:1; /* Region user-allocated. */
1919 bool busy:1; /* Region busy (for temporarily preventing user access). */
20
+ bool mapped:1; /* Region contains valid pages. */
2021 struct nvkm_memory *memory; /* Memory currently mapped into VMA. */
2122 struct nvkm_tags *tags; /* Compression tag reference. */
2223 };
....@@ -44,6 +45,8 @@
4445
4546 dma_addr_t null;
4647 void *nullp;
48
+
49
+ bool replay;
4750 };
4851
4952 int nvkm_vmm_new(struct nvkm_device *, u64 addr, u64 size, void *argv, u32 argc,
....@@ -63,6 +66,7 @@
6366 struct nvkm_mm_node *mem;
6467 struct scatterlist *sgl;
6568 dma_addr_t *dma;
69
+ u64 *pfn;
6670 u64 off;
6771
6872 const struct nvkm_vmm_page *page;
....@@ -130,4 +134,5 @@
130134 int gp100_mmu_new(struct nvkm_device *, int, struct nvkm_mmu **);
131135 int gp10b_mmu_new(struct nvkm_device *, int, struct nvkm_mmu **);
132136 int gv100_mmu_new(struct nvkm_device *, int, struct nvkm_mmu **);
137
+int tu102_mmu_new(struct nvkm_device *, int, struct nvkm_mmu **);
133138 #endif