mirror of
git://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
synced 2026-04-05 14:09:10 +08:00
Merge tag 'drm-fixes-2025-06-06' of https://gitlab.freedesktop.org/drm/kernel
Pull more drm fixes from Simona Vetter:
"Another small batch of drm fixes, this time with a different baseline
and hence separate.
Drivers:
- ivpu:
- dma_resv locking
- warning fixes
- reset failure handling
- improve logging
- update fw file names
- fix cmdqueue unregister
- panel-simple: add Evervision VGG644804
Core Changes:
- sysfb: screen_info type check
- video: screen_info for relocated pci fb
- drm/sched: signal fence of killed job
- dummycon: deferred takeover fix"
* tag 'drm-fixes-2025-06-06' of https://gitlab.freedesktop.org/drm/kernel:
sysfb: Fix screen_info type check for VGA
video: screen_info: Relocate framebuffers behind PCI bridges
accel/ivpu: Fix warning in ivpu_gem_bo_free()
accel/ivpu: Trigger device recovery on engine reset/resume failure
accel/ivpu: Use dma_resv_lock() instead of a custom mutex
drm/panel-simple: fix the warnings for the Evervision VGG644804
accel/ivpu: Reorder Doorbell Unregister and Command Queue Destruction
accel/ivpu: Use firmware names from upstream repo
accel/ivpu: Improve buffer object logging
dummycon: Trigger redraw when switching consoles with deferred takeover
drm/scheduler: signal scheduled fence when kill job
This commit is contained in:
@@ -55,18 +55,18 @@ static struct {
|
||||
int gen;
|
||||
const char *name;
|
||||
} fw_names[] = {
|
||||
{ IVPU_HW_IP_37XX, "vpu_37xx.bin" },
|
||||
{ IVPU_HW_IP_37XX, "intel/vpu/vpu_37xx_v1.bin" },
|
||||
{ IVPU_HW_IP_37XX, "intel/vpu/vpu_37xx_v0.0.bin" },
|
||||
{ IVPU_HW_IP_40XX, "vpu_40xx.bin" },
|
||||
{ IVPU_HW_IP_40XX, "intel/vpu/vpu_40xx_v1.bin" },
|
||||
{ IVPU_HW_IP_40XX, "intel/vpu/vpu_40xx_v0.0.bin" },
|
||||
{ IVPU_HW_IP_50XX, "vpu_50xx.bin" },
|
||||
{ IVPU_HW_IP_50XX, "intel/vpu/vpu_50xx_v1.bin" },
|
||||
{ IVPU_HW_IP_50XX, "intel/vpu/vpu_50xx_v0.0.bin" },
|
||||
};
|
||||
|
||||
/* Production fw_names from the table above */
|
||||
MODULE_FIRMWARE("intel/vpu/vpu_37xx_v0.0.bin");
|
||||
MODULE_FIRMWARE("intel/vpu/vpu_40xx_v0.0.bin");
|
||||
MODULE_FIRMWARE("intel/vpu/vpu_50xx_v0.0.bin");
|
||||
MODULE_FIRMWARE("intel/vpu/vpu_37xx_v1.bin");
|
||||
MODULE_FIRMWARE("intel/vpu/vpu_40xx_v1.bin");
|
||||
MODULE_FIRMWARE("intel/vpu/vpu_50xx_v1.bin");
|
||||
|
||||
static int ivpu_fw_request(struct ivpu_device *vdev)
|
||||
{
|
||||
|
||||
@@ -28,11 +28,21 @@ static inline void ivpu_dbg_bo(struct ivpu_device *vdev, struct ivpu_bo *bo, con
|
||||
{
|
||||
ivpu_dbg(vdev, BO,
|
||||
"%6s: bo %8p vpu_addr %9llx size %8zu ctx %d has_pages %d dma_mapped %d mmu_mapped %d wc %d imported %d\n",
|
||||
action, bo, bo->vpu_addr, ivpu_bo_size(bo), bo->ctx ? bo->ctx->id : 0,
|
||||
action, bo, bo->vpu_addr, ivpu_bo_size(bo), bo->ctx_id,
|
||||
(bool)bo->base.pages, (bool)bo->base.sgt, bo->mmu_mapped, bo->base.map_wc,
|
||||
(bool)drm_gem_is_imported(&bo->base.base));
|
||||
}
|
||||
|
||||
static inline int ivpu_bo_lock(struct ivpu_bo *bo)
|
||||
{
|
||||
return dma_resv_lock(bo->base.base.resv, NULL);
|
||||
}
|
||||
|
||||
static inline void ivpu_bo_unlock(struct ivpu_bo *bo)
|
||||
{
|
||||
dma_resv_unlock(bo->base.base.resv);
|
||||
}
|
||||
|
||||
/*
|
||||
* ivpu_bo_pin() - pin the backing physical pages and map them to VPU.
|
||||
*
|
||||
@@ -43,22 +53,22 @@ static inline void ivpu_dbg_bo(struct ivpu_device *vdev, struct ivpu_bo *bo, con
|
||||
int __must_check ivpu_bo_pin(struct ivpu_bo *bo)
|
||||
{
|
||||
struct ivpu_device *vdev = ivpu_bo_to_vdev(bo);
|
||||
struct sg_table *sgt;
|
||||
int ret = 0;
|
||||
|
||||
mutex_lock(&bo->lock);
|
||||
|
||||
ivpu_dbg_bo(vdev, bo, "pin");
|
||||
drm_WARN_ON(&vdev->drm, !bo->ctx);
|
||||
|
||||
sgt = drm_gem_shmem_get_pages_sgt(&bo->base);
|
||||
if (IS_ERR(sgt)) {
|
||||
ret = PTR_ERR(sgt);
|
||||
ivpu_err(vdev, "Failed to map BO in IOMMU: %d\n", ret);
|
||||
return ret;
|
||||
}
|
||||
|
||||
ivpu_bo_lock(bo);
|
||||
|
||||
if (!bo->mmu_mapped) {
|
||||
struct sg_table *sgt = drm_gem_shmem_get_pages_sgt(&bo->base);
|
||||
|
||||
if (IS_ERR(sgt)) {
|
||||
ret = PTR_ERR(sgt);
|
||||
ivpu_err(vdev, "Failed to map BO in IOMMU: %d\n", ret);
|
||||
goto unlock;
|
||||
}
|
||||
|
||||
drm_WARN_ON(&vdev->drm, !bo->ctx);
|
||||
ret = ivpu_mmu_context_map_sgt(vdev, bo->ctx, bo->vpu_addr, sgt,
|
||||
ivpu_bo_is_snooped(bo));
|
||||
if (ret) {
|
||||
@@ -69,7 +79,7 @@ int __must_check ivpu_bo_pin(struct ivpu_bo *bo)
|
||||
}
|
||||
|
||||
unlock:
|
||||
mutex_unlock(&bo->lock);
|
||||
ivpu_bo_unlock(bo);
|
||||
|
||||
return ret;
|
||||
}
|
||||
@@ -84,7 +94,7 @@ ivpu_bo_alloc_vpu_addr(struct ivpu_bo *bo, struct ivpu_mmu_context *ctx,
|
||||
if (!drm_dev_enter(&vdev->drm, &idx))
|
||||
return -ENODEV;
|
||||
|
||||
mutex_lock(&bo->lock);
|
||||
ivpu_bo_lock(bo);
|
||||
|
||||
ret = ivpu_mmu_context_insert_node(ctx, range, ivpu_bo_size(bo), &bo->mm_node);
|
||||
if (!ret) {
|
||||
@@ -94,9 +104,7 @@ ivpu_bo_alloc_vpu_addr(struct ivpu_bo *bo, struct ivpu_mmu_context *ctx,
|
||||
ivpu_err(vdev, "Failed to add BO to context %u: %d\n", ctx->id, ret);
|
||||
}
|
||||
|
||||
ivpu_dbg_bo(vdev, bo, "alloc");
|
||||
|
||||
mutex_unlock(&bo->lock);
|
||||
ivpu_bo_unlock(bo);
|
||||
|
||||
drm_dev_exit(idx);
|
||||
|
||||
@@ -107,7 +115,7 @@ static void ivpu_bo_unbind_locked(struct ivpu_bo *bo)
|
||||
{
|
||||
struct ivpu_device *vdev = ivpu_bo_to_vdev(bo);
|
||||
|
||||
lockdep_assert(lockdep_is_held(&bo->lock) || !kref_read(&bo->base.base.refcount));
|
||||
lockdep_assert(dma_resv_held(bo->base.base.resv) || !kref_read(&bo->base.base.refcount));
|
||||
|
||||
if (bo->mmu_mapped) {
|
||||
drm_WARN_ON(&vdev->drm, !bo->ctx);
|
||||
@@ -125,14 +133,12 @@ static void ivpu_bo_unbind_locked(struct ivpu_bo *bo)
|
||||
if (drm_gem_is_imported(&bo->base.base))
|
||||
return;
|
||||
|
||||
dma_resv_lock(bo->base.base.resv, NULL);
|
||||
if (bo->base.sgt) {
|
||||
dma_unmap_sgtable(vdev->drm.dev, bo->base.sgt, DMA_BIDIRECTIONAL, 0);
|
||||
sg_free_table(bo->base.sgt);
|
||||
kfree(bo->base.sgt);
|
||||
bo->base.sgt = NULL;
|
||||
}
|
||||
dma_resv_unlock(bo->base.base.resv);
|
||||
}
|
||||
|
||||
void ivpu_bo_unbind_all_bos_from_context(struct ivpu_device *vdev, struct ivpu_mmu_context *ctx)
|
||||
@@ -144,12 +150,12 @@ void ivpu_bo_unbind_all_bos_from_context(struct ivpu_device *vdev, struct ivpu_m
|
||||
|
||||
mutex_lock(&vdev->bo_list_lock);
|
||||
list_for_each_entry(bo, &vdev->bo_list, bo_list_node) {
|
||||
mutex_lock(&bo->lock);
|
||||
ivpu_bo_lock(bo);
|
||||
if (bo->ctx == ctx) {
|
||||
ivpu_dbg_bo(vdev, bo, "unbind");
|
||||
ivpu_bo_unbind_locked(bo);
|
||||
}
|
||||
mutex_unlock(&bo->lock);
|
||||
ivpu_bo_unlock(bo);
|
||||
}
|
||||
mutex_unlock(&vdev->bo_list_lock);
|
||||
}
|
||||
@@ -169,7 +175,6 @@ struct drm_gem_object *ivpu_gem_create_object(struct drm_device *dev, size_t siz
|
||||
bo->base.pages_mark_dirty_on_put = true; /* VPU can dirty a BO anytime */
|
||||
|
||||
INIT_LIST_HEAD(&bo->bo_list_node);
|
||||
mutex_init(&bo->lock);
|
||||
|
||||
return &bo->base.base;
|
||||
}
|
||||
@@ -215,7 +220,7 @@ fail_detach:
|
||||
return ERR_PTR(ret);
|
||||
}
|
||||
|
||||
static struct ivpu_bo *ivpu_bo_alloc(struct ivpu_device *vdev, u64 size, u32 flags)
|
||||
static struct ivpu_bo *ivpu_bo_alloc(struct ivpu_device *vdev, u64 size, u32 flags, u32 ctx_id)
|
||||
{
|
||||
struct drm_gem_shmem_object *shmem;
|
||||
struct ivpu_bo *bo;
|
||||
@@ -233,6 +238,7 @@ static struct ivpu_bo *ivpu_bo_alloc(struct ivpu_device *vdev, u64 size, u32 fla
|
||||
return ERR_CAST(shmem);
|
||||
|
||||
bo = to_ivpu_bo(&shmem->base);
|
||||
bo->ctx_id = ctx_id;
|
||||
bo->base.map_wc = flags & DRM_IVPU_BO_WC;
|
||||
bo->flags = flags;
|
||||
|
||||
@@ -240,6 +246,8 @@ static struct ivpu_bo *ivpu_bo_alloc(struct ivpu_device *vdev, u64 size, u32 fla
|
||||
list_add_tail(&bo->bo_list_node, &vdev->bo_list);
|
||||
mutex_unlock(&vdev->bo_list_lock);
|
||||
|
||||
ivpu_dbg_bo(vdev, bo, "alloc");
|
||||
|
||||
return bo;
|
||||
}
|
||||
|
||||
@@ -277,10 +285,14 @@ static void ivpu_gem_bo_free(struct drm_gem_object *obj)
|
||||
list_del(&bo->bo_list_node);
|
||||
mutex_unlock(&vdev->bo_list_lock);
|
||||
|
||||
drm_WARN_ON(&vdev->drm, !dma_resv_test_signaled(obj->resv, DMA_RESV_USAGE_READ));
|
||||
drm_WARN_ON(&vdev->drm, !drm_gem_is_imported(&bo->base.base) &&
|
||||
!dma_resv_test_signaled(obj->resv, DMA_RESV_USAGE_READ));
|
||||
drm_WARN_ON(&vdev->drm, ivpu_bo_size(bo) == 0);
|
||||
drm_WARN_ON(&vdev->drm, bo->base.vaddr);
|
||||
|
||||
ivpu_bo_unbind_locked(bo);
|
||||
mutex_destroy(&bo->lock);
|
||||
drm_WARN_ON(&vdev->drm, bo->mmu_mapped);
|
||||
drm_WARN_ON(&vdev->drm, bo->ctx);
|
||||
|
||||
drm_WARN_ON(obj->dev, refcount_read(&bo->base.pages_use_count) > 1);
|
||||
drm_gem_shmem_free(&bo->base);
|
||||
@@ -314,7 +326,7 @@ int ivpu_bo_create_ioctl(struct drm_device *dev, void *data, struct drm_file *fi
|
||||
if (size == 0)
|
||||
return -EINVAL;
|
||||
|
||||
bo = ivpu_bo_alloc(vdev, size, args->flags);
|
||||
bo = ivpu_bo_alloc(vdev, size, args->flags, file_priv->ctx.id);
|
||||
if (IS_ERR(bo)) {
|
||||
ivpu_err(vdev, "Failed to allocate BO: %pe (ctx %u size %llu flags 0x%x)",
|
||||
bo, file_priv->ctx.id, args->size, args->flags);
|
||||
@@ -322,7 +334,10 @@ int ivpu_bo_create_ioctl(struct drm_device *dev, void *data, struct drm_file *fi
|
||||
}
|
||||
|
||||
ret = drm_gem_handle_create(file, &bo->base.base, &args->handle);
|
||||
if (!ret)
|
||||
if (ret)
|
||||
ivpu_err(vdev, "Failed to create handle for BO: %pe (ctx %u size %llu flags 0x%x)",
|
||||
bo, file_priv->ctx.id, args->size, args->flags);
|
||||
else
|
||||
args->vpu_addr = bo->vpu_addr;
|
||||
|
||||
drm_gem_object_put(&bo->base.base);
|
||||
@@ -345,7 +360,7 @@ ivpu_bo_create(struct ivpu_device *vdev, struct ivpu_mmu_context *ctx,
|
||||
drm_WARN_ON(&vdev->drm, !PAGE_ALIGNED(range->end));
|
||||
drm_WARN_ON(&vdev->drm, !PAGE_ALIGNED(size));
|
||||
|
||||
bo = ivpu_bo_alloc(vdev, size, flags);
|
||||
bo = ivpu_bo_alloc(vdev, size, flags, IVPU_GLOBAL_CONTEXT_MMU_SSID);
|
||||
if (IS_ERR(bo)) {
|
||||
ivpu_err(vdev, "Failed to allocate BO: %pe (vpu_addr 0x%llx size %llu flags 0x%x)",
|
||||
bo, range->start, size, flags);
|
||||
@@ -361,9 +376,9 @@ ivpu_bo_create(struct ivpu_device *vdev, struct ivpu_mmu_context *ctx,
|
||||
goto err_put;
|
||||
|
||||
if (flags & DRM_IVPU_BO_MAPPABLE) {
|
||||
dma_resv_lock(bo->base.base.resv, NULL);
|
||||
ivpu_bo_lock(bo);
|
||||
ret = drm_gem_shmem_vmap_locked(&bo->base, &map);
|
||||
dma_resv_unlock(bo->base.base.resv);
|
||||
ivpu_bo_unlock(bo);
|
||||
|
||||
if (ret)
|
||||
goto err_put;
|
||||
@@ -386,9 +401,9 @@ void ivpu_bo_free(struct ivpu_bo *bo)
|
||||
struct iosys_map map = IOSYS_MAP_INIT_VADDR(bo->base.vaddr);
|
||||
|
||||
if (bo->flags & DRM_IVPU_BO_MAPPABLE) {
|
||||
dma_resv_lock(bo->base.base.resv, NULL);
|
||||
ivpu_bo_lock(bo);
|
||||
drm_gem_shmem_vunmap_locked(&bo->base, &map);
|
||||
dma_resv_unlock(bo->base.base.resv);
|
||||
ivpu_bo_unlock(bo);
|
||||
}
|
||||
|
||||
drm_gem_object_put(&bo->base.base);
|
||||
@@ -407,12 +422,12 @@ int ivpu_bo_info_ioctl(struct drm_device *dev, void *data, struct drm_file *file
|
||||
|
||||
bo = to_ivpu_bo(obj);
|
||||
|
||||
mutex_lock(&bo->lock);
|
||||
ivpu_bo_lock(bo);
|
||||
args->flags = bo->flags;
|
||||
args->mmap_offset = drm_vma_node_offset_addr(&obj->vma_node);
|
||||
args->vpu_addr = bo->vpu_addr;
|
||||
args->size = obj->size;
|
||||
mutex_unlock(&bo->lock);
|
||||
ivpu_bo_unlock(bo);
|
||||
|
||||
drm_gem_object_put(obj);
|
||||
return ret;
|
||||
@@ -449,10 +464,10 @@ int ivpu_bo_wait_ioctl(struct drm_device *dev, void *data, struct drm_file *file
|
||||
|
||||
static void ivpu_bo_print_info(struct ivpu_bo *bo, struct drm_printer *p)
|
||||
{
|
||||
mutex_lock(&bo->lock);
|
||||
ivpu_bo_lock(bo);
|
||||
|
||||
drm_printf(p, "%-9p %-3u 0x%-12llx %-10lu 0x%-8x %-4u",
|
||||
bo, bo->ctx ? bo->ctx->id : 0, bo->vpu_addr, bo->base.base.size,
|
||||
bo, bo->ctx_id, bo->vpu_addr, bo->base.base.size,
|
||||
bo->flags, kref_read(&bo->base.base.refcount));
|
||||
|
||||
if (bo->base.pages)
|
||||
@@ -466,7 +481,7 @@ static void ivpu_bo_print_info(struct ivpu_bo *bo, struct drm_printer *p)
|
||||
|
||||
drm_printf(p, "\n");
|
||||
|
||||
mutex_unlock(&bo->lock);
|
||||
ivpu_bo_unlock(bo);
|
||||
}
|
||||
|
||||
void ivpu_bo_list(struct drm_device *dev, struct drm_printer *p)
|
||||
|
||||
@@ -17,10 +17,10 @@ struct ivpu_bo {
|
||||
struct list_head bo_list_node;
|
||||
struct drm_mm_node mm_node;
|
||||
|
||||
struct mutex lock; /* Protects: ctx, mmu_mapped, vpu_addr */
|
||||
u64 vpu_addr;
|
||||
u32 flags;
|
||||
u32 job_status; /* Valid only for command buffer */
|
||||
u32 ctx_id;
|
||||
bool mmu_mapped;
|
||||
};
|
||||
|
||||
|
||||
@@ -247,6 +247,10 @@ static int ivpu_cmdq_unregister(struct ivpu_file_priv *file_priv, struct ivpu_cm
|
||||
if (!cmdq->db_id)
|
||||
return 0;
|
||||
|
||||
ret = ivpu_jsm_unregister_db(vdev, cmdq->db_id);
|
||||
if (!ret)
|
||||
ivpu_dbg(vdev, JOB, "DB %d unregistered\n", cmdq->db_id);
|
||||
|
||||
if (vdev->fw->sched_mode == VPU_SCHEDULING_MODE_HW) {
|
||||
ret = ivpu_jsm_hws_destroy_cmdq(vdev, file_priv->ctx.id, cmdq->id);
|
||||
if (!ret)
|
||||
@@ -254,10 +258,6 @@ static int ivpu_cmdq_unregister(struct ivpu_file_priv *file_priv, struct ivpu_cm
|
||||
cmdq->id, file_priv->ctx.id);
|
||||
}
|
||||
|
||||
ret = ivpu_jsm_unregister_db(vdev, cmdq->db_id);
|
||||
if (!ret)
|
||||
ivpu_dbg(vdev, JOB, "DB %d unregistered\n", cmdq->db_id);
|
||||
|
||||
xa_erase(&file_priv->vdev->db_xa, cmdq->db_id);
|
||||
cmdq->db_id = 0;
|
||||
|
||||
@@ -986,7 +986,8 @@ void ivpu_context_abort_work_fn(struct work_struct *work)
|
||||
return;
|
||||
|
||||
if (vdev->fw->sched_mode == VPU_SCHEDULING_MODE_HW)
|
||||
ivpu_jsm_reset_engine(vdev, 0);
|
||||
if (ivpu_jsm_reset_engine(vdev, 0))
|
||||
return;
|
||||
|
||||
mutex_lock(&vdev->context_list_lock);
|
||||
xa_for_each(&vdev->context_xa, ctx_id, file_priv) {
|
||||
@@ -1009,7 +1010,8 @@ void ivpu_context_abort_work_fn(struct work_struct *work)
|
||||
if (vdev->fw->sched_mode != VPU_SCHEDULING_MODE_HW)
|
||||
goto runtime_put;
|
||||
|
||||
ivpu_jsm_hws_resume_engine(vdev, 0);
|
||||
if (ivpu_jsm_hws_resume_engine(vdev, 0))
|
||||
return;
|
||||
/*
|
||||
* In hardware scheduling mode NPU already has stopped processing jobs
|
||||
* and won't send us any further notifications, thus we have to free job related resources
|
||||
|
||||
@@ -7,6 +7,7 @@
|
||||
#include "ivpu_hw.h"
|
||||
#include "ivpu_ipc.h"
|
||||
#include "ivpu_jsm_msg.h"
|
||||
#include "ivpu_pm.h"
|
||||
#include "vpu_jsm_api.h"
|
||||
|
||||
const char *ivpu_jsm_msg_type_to_str(enum vpu_ipc_msg_type type)
|
||||
@@ -163,8 +164,10 @@ int ivpu_jsm_reset_engine(struct ivpu_device *vdev, u32 engine)
|
||||
|
||||
ret = ivpu_ipc_send_receive(vdev, &req, VPU_JSM_MSG_ENGINE_RESET_DONE, &resp,
|
||||
VPU_IPC_CHAN_ASYNC_CMD, vdev->timeout.jsm);
|
||||
if (ret)
|
||||
if (ret) {
|
||||
ivpu_err_ratelimited(vdev, "Failed to reset engine %d: %d\n", engine, ret);
|
||||
ivpu_pm_trigger_recovery(vdev, "Engine reset failed");
|
||||
}
|
||||
|
||||
return ret;
|
||||
}
|
||||
@@ -354,8 +357,10 @@ int ivpu_jsm_hws_resume_engine(struct ivpu_device *vdev, u32 engine)
|
||||
|
||||
ret = ivpu_ipc_send_receive(vdev, &req, VPU_JSM_MSG_HWS_RESUME_ENGINE_DONE, &resp,
|
||||
VPU_IPC_CHAN_ASYNC_CMD, vdev->timeout.jsm);
|
||||
if (ret)
|
||||
if (ret) {
|
||||
ivpu_err_ratelimited(vdev, "Failed to resume engine %d: %d\n", engine, ret);
|
||||
ivpu_pm_trigger_recovery(vdev, "Engine resume failed");
|
||||
}
|
||||
|
||||
return ret;
|
||||
}
|
||||
|
||||
@@ -143,6 +143,7 @@ static __init int sysfb_init(void)
|
||||
{
|
||||
struct screen_info *si = &screen_info;
|
||||
struct device *parent;
|
||||
unsigned int type;
|
||||
struct simplefb_platform_data mode;
|
||||
const char *name;
|
||||
bool compatible;
|
||||
@@ -170,17 +171,26 @@ static __init int sysfb_init(void)
|
||||
goto put_device;
|
||||
}
|
||||
|
||||
type = screen_info_video_type(si);
|
||||
|
||||
/* if the FB is incompatible, create a legacy framebuffer device */
|
||||
if (si->orig_video_isVGA == VIDEO_TYPE_EFI)
|
||||
name = "efi-framebuffer";
|
||||
else if (si->orig_video_isVGA == VIDEO_TYPE_VLFB)
|
||||
name = "vesa-framebuffer";
|
||||
else if (si->orig_video_isVGA == VIDEO_TYPE_VGAC)
|
||||
name = "vga-framebuffer";
|
||||
else if (si->orig_video_isVGA == VIDEO_TYPE_EGAC)
|
||||
switch (type) {
|
||||
case VIDEO_TYPE_EGAC:
|
||||
name = "ega-framebuffer";
|
||||
else
|
||||
break;
|
||||
case VIDEO_TYPE_VGAC:
|
||||
name = "vga-framebuffer";
|
||||
break;
|
||||
case VIDEO_TYPE_VLFB:
|
||||
name = "vesa-framebuffer";
|
||||
break;
|
||||
case VIDEO_TYPE_EFI:
|
||||
name = "efi-framebuffer";
|
||||
break;
|
||||
default:
|
||||
name = "platform-framebuffer";
|
||||
break;
|
||||
}
|
||||
|
||||
pd = platform_device_alloc(name, 0);
|
||||
if (!pd) {
|
||||
|
||||
@@ -2198,13 +2198,14 @@ static const struct display_timing evervision_vgg644804_timing = {
|
||||
static const struct panel_desc evervision_vgg644804 = {
|
||||
.timings = &evervision_vgg644804_timing,
|
||||
.num_timings = 1,
|
||||
.bpc = 8,
|
||||
.bpc = 6,
|
||||
.size = {
|
||||
.width = 115,
|
||||
.height = 86,
|
||||
},
|
||||
.bus_format = MEDIA_BUS_FMT_RGB666_1X7X3_SPWG,
|
||||
.bus_flags = DRM_BUS_FLAG_DE_HIGH | DRM_BUS_FLAG_PIXDATA_SAMPLE_NEGEDGE,
|
||||
.bus_flags = DRM_BUS_FLAG_DE_HIGH,
|
||||
.connector_type = DRM_MODE_CONNECTOR_LVDS,
|
||||
};
|
||||
|
||||
static const struct display_timing evervision_vgg804821_timing = {
|
||||
|
||||
@@ -176,6 +176,7 @@ static void drm_sched_entity_kill_jobs_work(struct work_struct *wrk)
|
||||
{
|
||||
struct drm_sched_job *job = container_of(wrk, typeof(*job), work);
|
||||
|
||||
drm_sched_fence_scheduled(job->s_fence, NULL);
|
||||
drm_sched_fence_finished(job->s_fence, -ESRCH);
|
||||
WARN_ON(job->s_fence->parent);
|
||||
job->sched->ops->free_job(job);
|
||||
|
||||
@@ -85,6 +85,15 @@ static bool dummycon_blank(struct vc_data *vc, enum vesa_blank_mode blank,
|
||||
/* Redraw, so that we get putc(s) for output done while blanked */
|
||||
return true;
|
||||
}
|
||||
|
||||
static bool dummycon_switch(struct vc_data *vc)
|
||||
{
|
||||
/*
|
||||
* Redraw, so that we get putc(s) for output done while switched
|
||||
* away. Informs deferred consoles to take over the display.
|
||||
*/
|
||||
return true;
|
||||
}
|
||||
#else
|
||||
static void dummycon_putc(struct vc_data *vc, u16 c, unsigned int y,
|
||||
unsigned int x) { }
|
||||
@@ -95,6 +104,10 @@ static bool dummycon_blank(struct vc_data *vc, enum vesa_blank_mode blank,
|
||||
{
|
||||
return false;
|
||||
}
|
||||
static bool dummycon_switch(struct vc_data *vc)
|
||||
{
|
||||
return false;
|
||||
}
|
||||
#endif
|
||||
|
||||
static const char *dummycon_startup(void)
|
||||
@@ -124,11 +137,6 @@ static bool dummycon_scroll(struct vc_data *vc, unsigned int top,
|
||||
return false;
|
||||
}
|
||||
|
||||
static bool dummycon_switch(struct vc_data *vc)
|
||||
{
|
||||
return false;
|
||||
}
|
||||
|
||||
/*
|
||||
* The console `switch' structure for the dummy console
|
||||
*
|
||||
|
||||
@@ -7,8 +7,8 @@
|
||||
|
||||
static struct pci_dev *screen_info_lfb_pdev;
|
||||
static size_t screen_info_lfb_bar;
|
||||
static resource_size_t screen_info_lfb_offset;
|
||||
static struct resource screen_info_lfb_res = DEFINE_RES_MEM(0, 0);
|
||||
static resource_size_t screen_info_lfb_res_start; // original start of resource
|
||||
static resource_size_t screen_info_lfb_offset; // framebuffer offset within resource
|
||||
|
||||
static bool __screen_info_relocation_is_valid(const struct screen_info *si, struct resource *pr)
|
||||
{
|
||||
@@ -31,7 +31,7 @@ void screen_info_apply_fixups(void)
|
||||
if (screen_info_lfb_pdev) {
|
||||
struct resource *pr = &screen_info_lfb_pdev->resource[screen_info_lfb_bar];
|
||||
|
||||
if (pr->start != screen_info_lfb_res.start) {
|
||||
if (pr->start != screen_info_lfb_res_start) {
|
||||
if (__screen_info_relocation_is_valid(si, pr)) {
|
||||
/*
|
||||
* Only update base if we have an actual
|
||||
@@ -47,46 +47,67 @@ void screen_info_apply_fixups(void)
|
||||
}
|
||||
}
|
||||
|
||||
static int __screen_info_lfb_pci_bus_region(const struct screen_info *si, unsigned int type,
|
||||
struct pci_bus_region *r)
|
||||
{
|
||||
u64 base, size;
|
||||
|
||||
base = __screen_info_lfb_base(si);
|
||||
if (!base)
|
||||
return -EINVAL;
|
||||
|
||||
size = __screen_info_lfb_size(si, type);
|
||||
if (!size)
|
||||
return -EINVAL;
|
||||
|
||||
r->start = base;
|
||||
r->end = base + size - 1;
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
static void screen_info_fixup_lfb(struct pci_dev *pdev)
|
||||
{
|
||||
unsigned int type;
|
||||
struct resource res[SCREEN_INFO_MAX_RESOURCES];
|
||||
size_t i, numres;
|
||||
struct pci_bus_region bus_region;
|
||||
int ret;
|
||||
struct resource r = {
|
||||
.flags = IORESOURCE_MEM,
|
||||
};
|
||||
const struct resource *pr;
|
||||
const struct screen_info *si = &screen_info;
|
||||
|
||||
if (screen_info_lfb_pdev)
|
||||
return; // already found
|
||||
|
||||
type = screen_info_video_type(si);
|
||||
if (type != VIDEO_TYPE_EFI)
|
||||
return; // only applies to EFI
|
||||
if (!__screen_info_has_lfb(type))
|
||||
return; // only applies to EFI; maybe VESA
|
||||
|
||||
ret = screen_info_resources(si, res, ARRAY_SIZE(res));
|
||||
ret = __screen_info_lfb_pci_bus_region(si, type, &bus_region);
|
||||
if (ret < 0)
|
||||
return;
|
||||
numres = ret;
|
||||
|
||||
for (i = 0; i < numres; ++i) {
|
||||
struct resource *r = &res[i];
|
||||
const struct resource *pr;
|
||||
/*
|
||||
* Translate the PCI bus address to resource. Account
|
||||
* for an offset if the framebuffer is behind a PCI host
|
||||
* bridge.
|
||||
*/
|
||||
pcibios_bus_to_resource(pdev->bus, &r, &bus_region);
|
||||
|
||||
if (!(r->flags & IORESOURCE_MEM))
|
||||
continue;
|
||||
pr = pci_find_resource(pdev, r);
|
||||
if (!pr)
|
||||
continue;
|
||||
pr = pci_find_resource(pdev, &r);
|
||||
if (!pr)
|
||||
return;
|
||||
|
||||
/*
|
||||
* We've found a PCI device with the framebuffer
|
||||
* resource. Store away the parameters to track
|
||||
* relocation of the framebuffer aperture.
|
||||
*/
|
||||
screen_info_lfb_pdev = pdev;
|
||||
screen_info_lfb_bar = pr - pdev->resource;
|
||||
screen_info_lfb_offset = r->start - pr->start;
|
||||
memcpy(&screen_info_lfb_res, r, sizeof(screen_info_lfb_res));
|
||||
}
|
||||
/*
|
||||
* We've found a PCI device with the framebuffer
|
||||
* resource. Store away the parameters to track
|
||||
* relocation of the framebuffer aperture.
|
||||
*/
|
||||
screen_info_lfb_pdev = pdev;
|
||||
screen_info_lfb_bar = pr - pdev->resource;
|
||||
screen_info_lfb_offset = r.start - pr->start;
|
||||
screen_info_lfb_res_start = bus_region.start;
|
||||
}
|
||||
DECLARE_PCI_FIXUP_CLASS_HEADER(PCI_ANY_ID, PCI_ANY_ID, PCI_BASE_CLASS_DISPLAY, 16,
|
||||
screen_info_fixup_lfb);
|
||||
|
||||
Reference in New Issue
Block a user