mirror of
git://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
synced 2025-09-04 20:19:47 +08:00

Currently intel_dpt_resume() tries to blindly rewrite all the PTEs for currently bound DPT VMAs. That is problematic because the CPU mapping for the DPT is only really guaranteed to exist while the DPT object has been pinned. In the past we worked around this issue by making DPT objects unshrinkable, but that is undesirable as it'll waste physical RAM. Let's instead forcefully evict all the DPT VMAs on suspend, thus guaranteeing that intel_dpt_resume() has nothing to do. To guarantee that all the DPT VMAs are evictable by intel_dpt_suspend() we need to flush the cleanup workqueue after the display output has been shut down. And for good measure throw in a few extra WARNs to catch any mistakes. Cc: Brian Geffon <bgeffon@google.com> Cc: Vidya Srinivas <vidya.srinivas@intel.com> Signed-off-by: Ville Syrjälä <ville.syrjala@linux.intel.com> Link: https://patchwork.freedesktop.org/patch/msgid/20241127061117.25622-4-ville.syrjala@linux.intel.com Reviewed-by: Vidya Srinivas <vidya.srinivas@intel.com> Tested-by: Vidya Srinivas <vidya.srinivas@intel.com>
324 lines
7.9 KiB
C
324 lines
7.9 KiB
C
// SPDX-License-Identifier: MIT
|
|
/*
|
|
* Copyright © 2021 Intel Corporation
|
|
*/
|
|
|
|
#include "gem/i915_gem_domain.h"
|
|
#include "gem/i915_gem_internal.h"
|
|
#include "gem/i915_gem_lmem.h"
|
|
#include "gt/gen8_ppgtt.h"
|
|
|
|
#include "i915_drv.h"
|
|
#include "intel_display_types.h"
|
|
#include "intel_dpt.h"
|
|
#include "intel_fb.h"
|
|
|
|
struct i915_dpt {
|
|
struct i915_address_space vm;
|
|
|
|
struct drm_i915_gem_object *obj;
|
|
struct i915_vma *vma;
|
|
void __iomem *iomem;
|
|
};
|
|
|
|
#define i915_is_dpt(vm) ((vm)->is_dpt)
|
|
|
|
static inline struct i915_dpt *
|
|
i915_vm_to_dpt(struct i915_address_space *vm)
|
|
{
|
|
BUILD_BUG_ON(offsetof(struct i915_dpt, vm));
|
|
drm_WARN_ON(&vm->i915->drm, !i915_is_dpt(vm));
|
|
return container_of(vm, struct i915_dpt, vm);
|
|
}
|
|
|
|
#define dpt_total_entries(dpt) ((dpt)->vm.total >> PAGE_SHIFT)
|
|
|
|
static void gen8_set_pte(void __iomem *addr, gen8_pte_t pte)
|
|
{
|
|
writeq(pte, addr);
|
|
}
|
|
|
|
static void dpt_insert_page(struct i915_address_space *vm,
|
|
dma_addr_t addr,
|
|
u64 offset,
|
|
unsigned int pat_index,
|
|
u32 flags)
|
|
{
|
|
struct i915_dpt *dpt = i915_vm_to_dpt(vm);
|
|
gen8_pte_t __iomem *base = dpt->iomem;
|
|
|
|
gen8_set_pte(base + offset / I915_GTT_PAGE_SIZE,
|
|
vm->pte_encode(addr, pat_index, flags));
|
|
}
|
|
|
|
static void dpt_insert_entries(struct i915_address_space *vm,
|
|
struct i915_vma_resource *vma_res,
|
|
unsigned int pat_index,
|
|
u32 flags)
|
|
{
|
|
struct i915_dpt *dpt = i915_vm_to_dpt(vm);
|
|
gen8_pte_t __iomem *base = dpt->iomem;
|
|
const gen8_pte_t pte_encode = vm->pte_encode(0, pat_index, flags);
|
|
struct sgt_iter sgt_iter;
|
|
dma_addr_t addr;
|
|
int i;
|
|
|
|
/*
|
|
* Note that we ignore PTE_READ_ONLY here. The caller must be careful
|
|
* not to allow the user to override access to a read only page.
|
|
*/
|
|
|
|
i = vma_res->start / I915_GTT_PAGE_SIZE;
|
|
for_each_sgt_daddr(addr, sgt_iter, vma_res->bi.pages)
|
|
gen8_set_pte(&base[i++], pte_encode | addr);
|
|
}
|
|
|
|
static void dpt_clear_range(struct i915_address_space *vm,
|
|
u64 start, u64 length)
|
|
{
|
|
}
|
|
|
|
static void dpt_bind_vma(struct i915_address_space *vm,
|
|
struct i915_vm_pt_stash *stash,
|
|
struct i915_vma_resource *vma_res,
|
|
unsigned int pat_index,
|
|
u32 flags)
|
|
{
|
|
u32 pte_flags;
|
|
|
|
if (vma_res->bound_flags)
|
|
return;
|
|
|
|
/* Applicable to VLV (gen8+ do not support RO in the GGTT) */
|
|
pte_flags = 0;
|
|
if (vm->has_read_only && vma_res->bi.readonly)
|
|
pte_flags |= PTE_READ_ONLY;
|
|
if (vma_res->bi.lmem)
|
|
pte_flags |= PTE_LM;
|
|
|
|
vm->insert_entries(vm, vma_res, pat_index, pte_flags);
|
|
|
|
vma_res->page_sizes_gtt = I915_GTT_PAGE_SIZE;
|
|
|
|
/*
|
|
* Without aliasing PPGTT there's no difference between
|
|
* GLOBAL/LOCAL_BIND, it's all the same ptes. Hence unconditionally
|
|
* upgrade to both bound if we bind either to avoid double-binding.
|
|
*/
|
|
vma_res->bound_flags = I915_VMA_GLOBAL_BIND | I915_VMA_LOCAL_BIND;
|
|
}
|
|
|
|
static void dpt_unbind_vma(struct i915_address_space *vm,
|
|
struct i915_vma_resource *vma_res)
|
|
{
|
|
vm->clear_range(vm, vma_res->start, vma_res->vma_size);
|
|
}
|
|
|
|
static void dpt_cleanup(struct i915_address_space *vm)
|
|
{
|
|
struct i915_dpt *dpt = i915_vm_to_dpt(vm);
|
|
|
|
i915_gem_object_put(dpt->obj);
|
|
}
|
|
|
|
struct i915_vma *intel_dpt_pin_to_ggtt(struct i915_address_space *vm,
|
|
unsigned int alignment)
|
|
{
|
|
struct drm_i915_private *i915 = vm->i915;
|
|
struct i915_dpt *dpt = i915_vm_to_dpt(vm);
|
|
intel_wakeref_t wakeref;
|
|
struct i915_vma *vma;
|
|
void __iomem *iomem;
|
|
struct i915_gem_ww_ctx ww;
|
|
u64 pin_flags = 0;
|
|
int err;
|
|
|
|
if (i915_gem_object_is_stolen(dpt->obj))
|
|
pin_flags |= PIN_MAPPABLE;
|
|
|
|
wakeref = intel_runtime_pm_get(&i915->runtime_pm);
|
|
atomic_inc(&i915->gpu_error.pending_fb_pin);
|
|
|
|
for_i915_gem_ww(&ww, err, true) {
|
|
err = i915_gem_object_lock(dpt->obj, &ww);
|
|
if (err)
|
|
continue;
|
|
|
|
vma = i915_gem_object_ggtt_pin_ww(dpt->obj, &ww, NULL, 0,
|
|
alignment, pin_flags);
|
|
if (IS_ERR(vma)) {
|
|
err = PTR_ERR(vma);
|
|
continue;
|
|
}
|
|
|
|
iomem = i915_vma_pin_iomap(vma);
|
|
i915_vma_unpin(vma);
|
|
|
|
if (IS_ERR(iomem)) {
|
|
err = PTR_ERR(iomem);
|
|
continue;
|
|
}
|
|
|
|
dpt->vma = vma;
|
|
dpt->iomem = iomem;
|
|
|
|
i915_vma_get(vma);
|
|
}
|
|
|
|
dpt->obj->mm.dirty = true;
|
|
|
|
atomic_dec(&i915->gpu_error.pending_fb_pin);
|
|
intel_runtime_pm_put(&i915->runtime_pm, wakeref);
|
|
|
|
return err ? ERR_PTR(err) : vma;
|
|
}
|
|
|
|
void intel_dpt_unpin_from_ggtt(struct i915_address_space *vm)
|
|
{
|
|
struct i915_dpt *dpt = i915_vm_to_dpt(vm);
|
|
|
|
i915_vma_unpin_iomap(dpt->vma);
|
|
i915_vma_put(dpt->vma);
|
|
}
|
|
|
|
/**
|
|
* intel_dpt_resume - restore the memory mapping for all DPT FBs during system resume
|
|
* @i915: device instance
|
|
*
|
|
* Restore the memory mapping during system resume for all framebuffers which
|
|
* are mapped to HW via a GGTT->DPT page table. The content of these page
|
|
* tables are not stored in the hibernation image during S4 and S3RST->S4
|
|
* transitions, so here we reprogram the PTE entries in those tables.
|
|
*
|
|
* This function must be called after the mappings in GGTT have been restored calling
|
|
* i915_ggtt_resume().
|
|
*/
|
|
void intel_dpt_resume(struct drm_i915_private *i915)
|
|
{
|
|
struct drm_framebuffer *drm_fb;
|
|
|
|
if (!HAS_DISPLAY(i915))
|
|
return;
|
|
|
|
mutex_lock(&i915->drm.mode_config.fb_lock);
|
|
drm_for_each_fb(drm_fb, &i915->drm) {
|
|
struct intel_framebuffer *fb = to_intel_framebuffer(drm_fb);
|
|
|
|
if (fb->dpt_vm)
|
|
i915_ggtt_resume_vm(fb->dpt_vm, true);
|
|
}
|
|
mutex_unlock(&i915->drm.mode_config.fb_lock);
|
|
}
|
|
|
|
/**
|
|
* intel_dpt_suspend - suspend the memory mapping for all DPT FBs during system suspend
|
|
* @i915: device instance
|
|
*
|
|
* Suspend the memory mapping during system suspend for all framebuffers which
|
|
* are mapped to HW via a GGTT->DPT page table.
|
|
*
|
|
* This function must be called before the mappings in GGTT are suspended calling
|
|
* i915_ggtt_suspend().
|
|
*/
|
|
void intel_dpt_suspend(struct drm_i915_private *i915)
|
|
{
|
|
struct drm_framebuffer *drm_fb;
|
|
|
|
if (!HAS_DISPLAY(i915))
|
|
return;
|
|
|
|
mutex_lock(&i915->drm.mode_config.fb_lock);
|
|
|
|
drm_for_each_fb(drm_fb, &i915->drm) {
|
|
struct intel_framebuffer *fb = to_intel_framebuffer(drm_fb);
|
|
|
|
if (fb->dpt_vm)
|
|
i915_ggtt_suspend_vm(fb->dpt_vm, true);
|
|
}
|
|
|
|
mutex_unlock(&i915->drm.mode_config.fb_lock);
|
|
}
|
|
|
|
struct i915_address_space *
|
|
intel_dpt_create(struct intel_framebuffer *fb)
|
|
{
|
|
struct drm_gem_object *obj = intel_fb_bo(&fb->base);
|
|
struct drm_i915_private *i915 = to_i915(obj->dev);
|
|
struct drm_i915_gem_object *dpt_obj;
|
|
struct i915_address_space *vm;
|
|
struct i915_dpt *dpt;
|
|
size_t size;
|
|
int ret;
|
|
|
|
if (intel_fb_needs_pot_stride_remap(fb))
|
|
size = intel_remapped_info_size(&fb->remapped_view.gtt.remapped);
|
|
else
|
|
size = DIV_ROUND_UP_ULL(obj->size, I915_GTT_PAGE_SIZE);
|
|
|
|
size = round_up(size * sizeof(gen8_pte_t), I915_GTT_PAGE_SIZE);
|
|
|
|
dpt_obj = i915_gem_object_create_lmem(i915, size, I915_BO_ALLOC_CONTIGUOUS);
|
|
if (IS_ERR(dpt_obj) && i915_ggtt_has_aperture(to_gt(i915)->ggtt))
|
|
dpt_obj = i915_gem_object_create_stolen(i915, size);
|
|
if (IS_ERR(dpt_obj) && !HAS_LMEM(i915)) {
|
|
drm_dbg_kms(&i915->drm, "Allocating dpt from smem\n");
|
|
dpt_obj = i915_gem_object_create_shmem(i915, size);
|
|
}
|
|
if (IS_ERR(dpt_obj))
|
|
return ERR_CAST(dpt_obj);
|
|
|
|
ret = i915_gem_object_lock_interruptible(dpt_obj, NULL);
|
|
if (!ret) {
|
|
ret = i915_gem_object_set_cache_level(dpt_obj, I915_CACHE_NONE);
|
|
i915_gem_object_unlock(dpt_obj);
|
|
}
|
|
if (ret) {
|
|
i915_gem_object_put(dpt_obj);
|
|
return ERR_PTR(ret);
|
|
}
|
|
|
|
dpt = kzalloc(sizeof(*dpt), GFP_KERNEL);
|
|
if (!dpt) {
|
|
i915_gem_object_put(dpt_obj);
|
|
return ERR_PTR(-ENOMEM);
|
|
}
|
|
|
|
vm = &dpt->vm;
|
|
|
|
vm->gt = to_gt(i915);
|
|
vm->i915 = i915;
|
|
vm->dma = i915->drm.dev;
|
|
vm->total = (size / sizeof(gen8_pte_t)) * I915_GTT_PAGE_SIZE;
|
|
vm->is_dpt = true;
|
|
|
|
i915_address_space_init(vm, VM_CLASS_DPT);
|
|
|
|
vm->insert_page = dpt_insert_page;
|
|
vm->clear_range = dpt_clear_range;
|
|
vm->insert_entries = dpt_insert_entries;
|
|
vm->cleanup = dpt_cleanup;
|
|
|
|
vm->vma_ops.bind_vma = dpt_bind_vma;
|
|
vm->vma_ops.unbind_vma = dpt_unbind_vma;
|
|
|
|
vm->pte_encode = vm->gt->ggtt->vm.pte_encode;
|
|
|
|
dpt->obj = dpt_obj;
|
|
dpt->obj->is_dpt = true;
|
|
|
|
return &dpt->vm;
|
|
}
|
|
|
|
void intel_dpt_destroy(struct i915_address_space *vm)
|
|
{
|
|
struct i915_dpt *dpt = i915_vm_to_dpt(vm);
|
|
|
|
dpt->obj->is_dpt = false;
|
|
i915_vm_put(&dpt->vm);
|
|
}
|
|
|
|
u64 intel_dpt_offset(struct i915_vma *dpt_vma)
|
|
{
|
|
return dpt_vma->node.start;
|
|
}
|