.. | .. |
---|
37 | 37 | #define __GVT_RENDER_H__ |
---|
38 | 38 | |
---|
39 | 39 | struct engine_mmio { |
---|
40 | | - int ring_id; |
---|
| 40 | + enum intel_engine_id id; |
---|
41 | 41 | i915_reg_t reg; |
---|
42 | 42 | u32 mask; |
---|
43 | 43 | bool in_context; |
---|
.. | .. |
---|
45 | 45 | }; |
---|
46 | 46 | |
---|
47 | 47 | void intel_gvt_switch_mmio(struct intel_vgpu *pre, |
---|
48 | | - struct intel_vgpu *next, int ring_id); |
---|
| 48 | + struct intel_vgpu *next, |
---|
| 49 | + const struct intel_engine_cs *engine); |
---|
49 | 50 | |
---|
50 | 51 | void intel_gvt_init_engine_mmio_context(struct intel_gvt *gvt); |
---|
51 | 52 | |
---|
.. | .. |
---|
54 | 55 | int intel_vgpu_restore_inhibit_context(struct intel_vgpu *vgpu, |
---|
55 | 56 | struct i915_request *req); |
---|
56 | 57 | |
---|
| 58 | +#define IS_RESTORE_INHIBIT(a) \ |
---|
| 59 | + IS_MASKED_BITS_ENABLED(a, CTX_CTRL_ENGINE_CTX_RESTORE_INHIBIT) |
---|
| 60 | + |
---|
57 | 61 | #endif |
---|