Currently the KMD is using enum i915_cache_level to set caching policy for buffer objects. This is flaky because the PAT index which really controls the caching behavior in PTE has far more levels than what's defined in the enum. In addition, the PAT index is platform dependent, having to translate between i915_cache_level and PAT index is not reliable, and makes the code more complicated. From UMD's perspective there is also a necessity to set caching policy for performance fine tuning. It's much easier for the UMD to directly use PAT index because the behavior of each PAT index is clearly defined in Bspec. Having the abstracted i915_cache_level sitting in between would only cause more ambiguity. PAT is expected to work much like MOCS already works today, and by design userspace is expected to select the index that exactly matches the desired behavior described in the hardware specification. For these reasons this patch replaces i915_cache_level with PAT index. Also note, the cache_level is not completely removed yet, because the KMD still has the need of creating buffer objects with simple cache settings such as cached, uncached, or writethrough. For kernel objects, cache_level is used for simplicity and backward compatibility. For Pre-gen12 platforms PAT can have 1:1 mapping to i915_cache_level, so these two are interchangeable. see the use of LEGACY_CACHELEVEL. One consequence of this change is that gen8_pte_encode is no longer working for gen12 platforms due to the fact that gen12 platforms has different PAT definitions. In the meantime the mtl_pte_encode introduced specfically for MTL becomes generic for all gen12 platforms. This patch renames the MTL PTE encode function into gen12_pte_encode and apply it to all gen12. Even though this change looks unrelated, but separating them would temporarily break gen12 PTE encoding, thus squash them in one patch. Special note: this patch changes the way caching behavior is controlled in the sense that some objects are left to be managed by userspace. For such objects we need to be careful not to change the userspace settings.There are kerneldoc and comments added around obj->cache_coherent, cache_dirty, and how to bypass the checkings by i915_gem_object_has_cache_level. For full understanding, these changes need to be looked at together with the two follow-up patches, one disables the {set|get}_caching ioctl's and the other adds set_pat extension to the GEM_CREATE uAPI. Bspec: 63019 Cc: Chris Wilson <chris.p.wilson@linux.intel.com> Signed-off-by: Fei Yang <fei.yang@intel.com> Reviewed-by: Andi Shyti <andi.shyti@linux.intel.com> Reviewed-by: Matt Roper <matthew.d.roper@intel.com> Signed-off-by: Andi Shyti <andi.shyti@linux.intel.com> Link: https://patchwork.freedesktop.org/patch/msgid/20230509165200.1740-3-fei.yang@intel.com
392 lines
8.4 KiB
C
392 lines
8.4 KiB
C
// SPDX-License-Identifier: MIT
|
|
/*
|
|
* Copyright © 2018 Intel Corporation
|
|
*/
|
|
|
|
#include <linux/crc32.h>
|
|
|
|
#include "gem/i915_gem_stolen.h"
|
|
|
|
#include "i915_memcpy.h"
|
|
#include "i915_selftest.h"
|
|
#include "intel_gpu_commands.h"
|
|
#include "selftests/igt_reset.h"
|
|
#include "selftests/igt_atomic.h"
|
|
#include "selftests/igt_spinner.h"
|
|
|
|
static int
|
|
__igt_reset_stolen(struct intel_gt *gt,
|
|
intel_engine_mask_t mask,
|
|
const char *msg)
|
|
{
|
|
struct i915_ggtt *ggtt = gt->ggtt;
|
|
const struct resource *dsm = >->i915->dsm.stolen;
|
|
resource_size_t num_pages, page;
|
|
struct intel_engine_cs *engine;
|
|
intel_wakeref_t wakeref;
|
|
enum intel_engine_id id;
|
|
struct igt_spinner spin;
|
|
long max, count;
|
|
void *tmp;
|
|
u32 *crc;
|
|
int err;
|
|
|
|
if (!drm_mm_node_allocated(&ggtt->error_capture))
|
|
return 0;
|
|
|
|
num_pages = resource_size(dsm) >> PAGE_SHIFT;
|
|
if (!num_pages)
|
|
return 0;
|
|
|
|
crc = kmalloc_array(num_pages, sizeof(u32), GFP_KERNEL);
|
|
if (!crc)
|
|
return -ENOMEM;
|
|
|
|
tmp = kmalloc(PAGE_SIZE, GFP_KERNEL);
|
|
if (!tmp) {
|
|
err = -ENOMEM;
|
|
goto err_crc;
|
|
}
|
|
|
|
igt_global_reset_lock(gt);
|
|
wakeref = intel_runtime_pm_get(gt->uncore->rpm);
|
|
|
|
err = igt_spinner_init(&spin, gt);
|
|
if (err)
|
|
goto err_lock;
|
|
|
|
for_each_engine(engine, gt, id) {
|
|
struct intel_context *ce;
|
|
struct i915_request *rq;
|
|
|
|
if (!(mask & engine->mask))
|
|
continue;
|
|
|
|
if (!intel_engine_can_store_dword(engine))
|
|
continue;
|
|
|
|
ce = intel_context_create(engine);
|
|
if (IS_ERR(ce)) {
|
|
err = PTR_ERR(ce);
|
|
goto err_spin;
|
|
}
|
|
rq = igt_spinner_create_request(&spin, ce, MI_ARB_CHECK);
|
|
intel_context_put(ce);
|
|
if (IS_ERR(rq)) {
|
|
err = PTR_ERR(rq);
|
|
goto err_spin;
|
|
}
|
|
i915_request_add(rq);
|
|
}
|
|
|
|
for (page = 0; page < num_pages; page++) {
|
|
dma_addr_t dma = (dma_addr_t)dsm->start + (page << PAGE_SHIFT);
|
|
void __iomem *s;
|
|
void *in;
|
|
|
|
ggtt->vm.insert_page(&ggtt->vm, dma,
|
|
ggtt->error_capture.start,
|
|
i915_gem_get_pat_index(gt->i915,
|
|
I915_CACHE_NONE),
|
|
0);
|
|
mb();
|
|
|
|
s = io_mapping_map_wc(&ggtt->iomap,
|
|
ggtt->error_capture.start,
|
|
PAGE_SIZE);
|
|
|
|
if (!__drm_mm_interval_first(>->i915->mm.stolen,
|
|
page << PAGE_SHIFT,
|
|
((page + 1) << PAGE_SHIFT) - 1))
|
|
memset_io(s, STACK_MAGIC, PAGE_SIZE);
|
|
|
|
in = (void __force *)s;
|
|
if (i915_memcpy_from_wc(tmp, in, PAGE_SIZE))
|
|
in = tmp;
|
|
crc[page] = crc32_le(0, in, PAGE_SIZE);
|
|
|
|
io_mapping_unmap(s);
|
|
}
|
|
mb();
|
|
ggtt->vm.clear_range(&ggtt->vm, ggtt->error_capture.start, PAGE_SIZE);
|
|
|
|
if (mask == ALL_ENGINES) {
|
|
intel_gt_reset(gt, mask, NULL);
|
|
} else {
|
|
for_each_engine(engine, gt, id) {
|
|
if (mask & engine->mask)
|
|
intel_engine_reset(engine, NULL);
|
|
}
|
|
}
|
|
|
|
max = -1;
|
|
count = 0;
|
|
for (page = 0; page < num_pages; page++) {
|
|
dma_addr_t dma = (dma_addr_t)dsm->start + (page << PAGE_SHIFT);
|
|
void __iomem *s;
|
|
void *in;
|
|
u32 x;
|
|
|
|
ggtt->vm.insert_page(&ggtt->vm, dma,
|
|
ggtt->error_capture.start,
|
|
i915_gem_get_pat_index(gt->i915,
|
|
I915_CACHE_NONE),
|
|
0);
|
|
mb();
|
|
|
|
s = io_mapping_map_wc(&ggtt->iomap,
|
|
ggtt->error_capture.start,
|
|
PAGE_SIZE);
|
|
|
|
in = (void __force *)s;
|
|
if (i915_memcpy_from_wc(tmp, in, PAGE_SIZE))
|
|
in = tmp;
|
|
x = crc32_le(0, in, PAGE_SIZE);
|
|
|
|
if (x != crc[page] &&
|
|
!__drm_mm_interval_first(>->i915->mm.stolen,
|
|
page << PAGE_SHIFT,
|
|
((page + 1) << PAGE_SHIFT) - 1)) {
|
|
pr_debug("unused stolen page %pa modified by GPU reset\n",
|
|
&page);
|
|
if (count++ == 0)
|
|
igt_hexdump(in, PAGE_SIZE);
|
|
max = page;
|
|
}
|
|
|
|
io_mapping_unmap(s);
|
|
}
|
|
mb();
|
|
ggtt->vm.clear_range(&ggtt->vm, ggtt->error_capture.start, PAGE_SIZE);
|
|
|
|
if (count > 0) {
|
|
pr_info("%s reset clobbered %ld pages of stolen, last clobber at page %ld\n",
|
|
msg, count, max);
|
|
}
|
|
if (max >= I915_GEM_STOLEN_BIAS >> PAGE_SHIFT) {
|
|
pr_err("%s reset clobbered unreserved area [above %x] of stolen; may cause severe faults\n",
|
|
msg, I915_GEM_STOLEN_BIAS);
|
|
err = -EINVAL;
|
|
}
|
|
|
|
err_spin:
|
|
igt_spinner_fini(&spin);
|
|
|
|
err_lock:
|
|
intel_runtime_pm_put(gt->uncore->rpm, wakeref);
|
|
igt_global_reset_unlock(gt);
|
|
|
|
kfree(tmp);
|
|
err_crc:
|
|
kfree(crc);
|
|
return err;
|
|
}
|
|
|
|
static int igt_reset_device_stolen(void *arg)
|
|
{
|
|
return __igt_reset_stolen(arg, ALL_ENGINES, "device");
|
|
}
|
|
|
|
static int igt_reset_engines_stolen(void *arg)
|
|
{
|
|
struct intel_gt *gt = arg;
|
|
struct intel_engine_cs *engine;
|
|
enum intel_engine_id id;
|
|
int err;
|
|
|
|
if (!intel_has_reset_engine(gt))
|
|
return 0;
|
|
|
|
for_each_engine(engine, gt, id) {
|
|
err = __igt_reset_stolen(gt, engine->mask, engine->name);
|
|
if (err)
|
|
return err;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int igt_global_reset(void *arg)
|
|
{
|
|
struct intel_gt *gt = arg;
|
|
unsigned int reset_count;
|
|
intel_wakeref_t wakeref;
|
|
int err = 0;
|
|
|
|
/* Check that we can issue a global GPU reset */
|
|
|
|
igt_global_reset_lock(gt);
|
|
wakeref = intel_runtime_pm_get(gt->uncore->rpm);
|
|
|
|
reset_count = i915_reset_count(>->i915->gpu_error);
|
|
|
|
intel_gt_reset(gt, ALL_ENGINES, NULL);
|
|
|
|
if (i915_reset_count(>->i915->gpu_error) == reset_count) {
|
|
pr_err("No GPU reset recorded!\n");
|
|
err = -EINVAL;
|
|
}
|
|
|
|
intel_runtime_pm_put(gt->uncore->rpm, wakeref);
|
|
igt_global_reset_unlock(gt);
|
|
|
|
if (intel_gt_is_wedged(gt))
|
|
err = -EIO;
|
|
|
|
return err;
|
|
}
|
|
|
|
static int igt_wedged_reset(void *arg)
|
|
{
|
|
struct intel_gt *gt = arg;
|
|
intel_wakeref_t wakeref;
|
|
|
|
/* Check that we can recover a wedged device with a GPU reset */
|
|
|
|
igt_global_reset_lock(gt);
|
|
wakeref = intel_runtime_pm_get(gt->uncore->rpm);
|
|
|
|
intel_gt_set_wedged(gt);
|
|
|
|
GEM_BUG_ON(!intel_gt_is_wedged(gt));
|
|
intel_gt_reset(gt, ALL_ENGINES, NULL);
|
|
|
|
intel_runtime_pm_put(gt->uncore->rpm, wakeref);
|
|
igt_global_reset_unlock(gt);
|
|
|
|
return intel_gt_is_wedged(gt) ? -EIO : 0;
|
|
}
|
|
|
|
static int igt_atomic_reset(void *arg)
|
|
{
|
|
struct intel_gt *gt = arg;
|
|
const typeof(*igt_atomic_phases) *p;
|
|
int err = 0;
|
|
|
|
/* Check that the resets are usable from atomic context */
|
|
|
|
intel_gt_pm_get(gt);
|
|
igt_global_reset_lock(gt);
|
|
|
|
/* Flush any requests before we get started and check basics */
|
|
if (!igt_force_reset(gt))
|
|
goto unlock;
|
|
|
|
for (p = igt_atomic_phases; p->name; p++) {
|
|
intel_engine_mask_t awake;
|
|
|
|
GEM_TRACE("__intel_gt_reset under %s\n", p->name);
|
|
|
|
awake = reset_prepare(gt);
|
|
p->critical_section_begin();
|
|
|
|
err = __intel_gt_reset(gt, ALL_ENGINES);
|
|
|
|
p->critical_section_end();
|
|
reset_finish(gt, awake);
|
|
|
|
if (err) {
|
|
pr_err("__intel_gt_reset failed under %s\n", p->name);
|
|
break;
|
|
}
|
|
}
|
|
|
|
/* As we poke around the guts, do a full reset before continuing. */
|
|
igt_force_reset(gt);
|
|
|
|
unlock:
|
|
igt_global_reset_unlock(gt);
|
|
intel_gt_pm_put(gt);
|
|
|
|
return err;
|
|
}
|
|
|
|
static int igt_atomic_engine_reset(void *arg)
|
|
{
|
|
struct intel_gt *gt = arg;
|
|
const typeof(*igt_atomic_phases) *p;
|
|
struct intel_engine_cs *engine;
|
|
enum intel_engine_id id;
|
|
int err = 0;
|
|
|
|
/* Check that the resets are usable from atomic context */
|
|
|
|
if (!intel_has_reset_engine(gt))
|
|
return 0;
|
|
|
|
if (intel_uc_uses_guc_submission(>->uc))
|
|
return 0;
|
|
|
|
intel_gt_pm_get(gt);
|
|
igt_global_reset_lock(gt);
|
|
|
|
/* Flush any requests before we get started and check basics */
|
|
if (!igt_force_reset(gt))
|
|
goto out_unlock;
|
|
|
|
for_each_engine(engine, gt, id) {
|
|
struct tasklet_struct *t = &engine->sched_engine->tasklet;
|
|
|
|
if (t->func)
|
|
tasklet_disable(t);
|
|
intel_engine_pm_get(engine);
|
|
|
|
for (p = igt_atomic_phases; p->name; p++) {
|
|
GEM_TRACE("intel_engine_reset(%s) under %s\n",
|
|
engine->name, p->name);
|
|
if (strcmp(p->name, "softirq"))
|
|
local_bh_disable();
|
|
|
|
p->critical_section_begin();
|
|
err = __intel_engine_reset_bh(engine, NULL);
|
|
p->critical_section_end();
|
|
|
|
if (strcmp(p->name, "softirq"))
|
|
local_bh_enable();
|
|
|
|
if (err) {
|
|
pr_err("intel_engine_reset(%s) failed under %s\n",
|
|
engine->name, p->name);
|
|
break;
|
|
}
|
|
}
|
|
|
|
intel_engine_pm_put(engine);
|
|
if (t->func) {
|
|
tasklet_enable(t);
|
|
tasklet_hi_schedule(t);
|
|
}
|
|
if (err)
|
|
break;
|
|
}
|
|
|
|
/* As we poke around the guts, do a full reset before continuing. */
|
|
igt_force_reset(gt);
|
|
|
|
out_unlock:
|
|
igt_global_reset_unlock(gt);
|
|
intel_gt_pm_put(gt);
|
|
|
|
return err;
|
|
}
|
|
|
|
int intel_reset_live_selftests(struct drm_i915_private *i915)
|
|
{
|
|
static const struct i915_subtest tests[] = {
|
|
SUBTEST(igt_global_reset), /* attempt to recover GPU first */
|
|
SUBTEST(igt_reset_device_stolen),
|
|
SUBTEST(igt_reset_engines_stolen),
|
|
SUBTEST(igt_wedged_reset),
|
|
SUBTEST(igt_atomic_reset),
|
|
SUBTEST(igt_atomic_engine_reset),
|
|
};
|
|
struct intel_gt *gt = to_gt(i915);
|
|
|
|
if (!intel_has_gpu_reset(gt))
|
|
return 0;
|
|
|
|
if (intel_gt_is_wedged(gt))
|
|
return -EIO; /* we're long past hope of a successful reset */
|
|
|
|
return intel_gt_live_subtests(tests, gt);
|
|
}
|