forked from ~ljy/RK356X_SDK_RELEASE

hc
2024-05-13 9d77db3c730780c8ef5ccd4b66403ff5675cfe4e
kernel/drivers/gpu/drm/vmwgfx/vmwgfx_ttm_glue.c
....@@ -25,74 +25,68 @@
2525 *
2626 **************************************************************************/
2727
28
-#include <drm/drmP.h>
2928 #include "vmwgfx_drv.h"
3029
3130 int vmw_mmap(struct file *filp, struct vm_area_struct *vma)
3231 {
33
- struct drm_file *file_priv;
34
- struct vmw_private *dev_priv;
32
+ static const struct vm_operations_struct vmw_vm_ops = {
33
+ .pfn_mkwrite = vmw_bo_vm_mkwrite,
34
+ .page_mkwrite = vmw_bo_vm_mkwrite,
35
+ .fault = vmw_bo_vm_fault,
36
+ .open = ttm_bo_vm_open,
37
+ .close = ttm_bo_vm_close,
38
+#ifdef CONFIG_TRANSPARENT_HUGEPAGE
39
+ .huge_fault = vmw_bo_vm_huge_fault,
40
+#endif
41
+ };
42
+ struct drm_file *file_priv = filp->private_data;
43
+ struct vmw_private *dev_priv = vmw_priv(file_priv->minor->dev);
44
+ int ret = ttm_bo_mmap(filp, vma, &dev_priv->bdev);
3545
36
- if (unlikely(vma->vm_pgoff < VMWGFX_FILE_PAGE_OFFSET)) {
37
- DRM_ERROR("Illegal attempt to mmap old fifo space.\n");
38
- return -EINVAL;
39
- }
40
-
41
- file_priv = filp->private_data;
42
- dev_priv = vmw_priv(file_priv->minor->dev);
43
- return ttm_bo_mmap(filp, vma, &dev_priv->bdev);
44
-}
45
-
46
-static int vmw_ttm_mem_global_init(struct drm_global_reference *ref)
47
-{
48
- DRM_INFO("global init.\n");
49
- return ttm_mem_global_init(ref->object);
50
-}
51
-
52
-static void vmw_ttm_mem_global_release(struct drm_global_reference *ref)
53
-{
54
- ttm_mem_global_release(ref->object);
55
-}
56
-
57
-int vmw_ttm_global_init(struct vmw_private *dev_priv)
58
-{
59
- struct drm_global_reference *global_ref;
60
- int ret;
61
-
62
- global_ref = &dev_priv->mem_global_ref;
63
- global_ref->global_type = DRM_GLOBAL_TTM_MEM;
64
- global_ref->size = sizeof(struct ttm_mem_global);
65
- global_ref->init = &vmw_ttm_mem_global_init;
66
- global_ref->release = &vmw_ttm_mem_global_release;
67
-
68
- ret = drm_global_item_ref(global_ref);
69
- if (unlikely(ret != 0)) {
70
- DRM_ERROR("Failed setting up TTM memory accounting.\n");
46
+ if (ret)
7147 return ret;
72
- }
7348
74
- dev_priv->bo_global_ref.mem_glob =
75
- dev_priv->mem_global_ref.object;
76
- global_ref = &dev_priv->bo_global_ref.ref;
77
- global_ref->global_type = DRM_GLOBAL_TTM_BO;
78
- global_ref->size = sizeof(struct ttm_bo_global);
79
- global_ref->init = &ttm_bo_global_init;
80
- global_ref->release = &ttm_bo_global_release;
81
- ret = drm_global_item_ref(global_ref);
49
+ vma->vm_ops = &vmw_vm_ops;
8250
83
- if (unlikely(ret != 0)) {
84
- DRM_ERROR("Failed setting up TTM buffer objects.\n");
85
- goto out_no_bo;
86
- }
51
+ /* Use VM_PFNMAP rather than VM_MIXEDMAP if not a COW mapping */
52
+ if ((vma->vm_flags & (VM_SHARED | VM_MAYWRITE)) != VM_MAYWRITE)
53
+ vma->vm_flags = (vma->vm_flags & ~VM_MIXEDMAP) | VM_PFNMAP;
8754
8855 return 0;
89
-out_no_bo:
90
- drm_global_item_unref(&dev_priv->mem_global_ref);
91
- return ret;
9256 }
9357
94
-void vmw_ttm_global_release(struct vmw_private *dev_priv)
58
+/* struct vmw_validation_mem callback */
59
+static int vmw_vmt_reserve(struct vmw_validation_mem *m, size_t size)
9560 {
96
- drm_global_item_unref(&dev_priv->bo_global_ref.ref);
97
- drm_global_item_unref(&dev_priv->mem_global_ref);
61
+ static struct ttm_operation_ctx ctx = {.interruptible = false,
62
+ .no_wait_gpu = false};
63
+ struct vmw_private *dev_priv = container_of(m, struct vmw_private, vvm);
64
+
65
+ return ttm_mem_global_alloc(vmw_mem_glob(dev_priv), size, &ctx);
66
+}
67
+
68
+/* struct vmw_validation_mem callback */
69
+static void vmw_vmt_unreserve(struct vmw_validation_mem *m, size_t size)
70
+{
71
+ struct vmw_private *dev_priv = container_of(m, struct vmw_private, vvm);
72
+
73
+ return ttm_mem_global_free(vmw_mem_glob(dev_priv), size);
74
+}
75
+
76
+/**
77
+ * vmw_validation_mem_init_ttm - Interface the validation memory tracker
78
+ * to ttm.
79
+ * @dev_priv: Pointer to struct vmw_private. The reason we choose a vmw private
80
+ * rather than a struct vmw_validation_mem is to make sure assumption in the
81
+ * callbacks that struct vmw_private derives from struct vmw_validation_mem
82
+ * holds true.
83
+ * @gran: The recommended allocation granularity
84
+ */
85
+void vmw_validation_mem_init_ttm(struct vmw_private *dev_priv, size_t gran)
86
+{
87
+ struct vmw_validation_mem *vvm = &dev_priv->vvm;
88
+
89
+ vvm->reserve_mem = vmw_vmt_reserve;
90
+ vvm->unreserve_mem = vmw_vmt_unreserve;
91
+ vvm->gran = gran;
9892 }