2019-01-16 23:33:04 +08:00
|
|
|
/*
|
|
|
|
* SPDX-License-Identifier: MIT
|
|
|
|
*
|
|
|
|
* Copyright © 2008-2018 Intel Corporation
|
|
|
|
*/
|
|
|
|
|
|
|
|
#include <linux/sched/mm.h>
|
2019-01-25 21:22:28 +08:00
|
|
|
#include <linux/stop_machine.h>
|
2019-01-16 23:33:04 +08:00
|
|
|
|
|
|
|
#include "i915_drv.h"
|
|
|
|
#include "i915_gpu_error.h"
|
|
|
|
#include "i915_reset.h"
|
|
|
|
|
|
|
|
#include "intel_guc.h"
|
|
|
|
|
2019-01-25 21:22:26 +08:00
|
|
|
#define RESET_MAX_RETRIES 3
|
|
|
|
|
2019-01-25 21:22:28 +08:00
|
|
|
/* XXX How to handle concurrent GGTT updates using tiling registers? */
|
|
|
|
#define RESET_UNDER_STOP_MACHINE 0
|
|
|
|
|
2019-01-16 23:33:04 +08:00
|
|
|
static void engine_skip_context(struct i915_request *rq)
|
|
|
|
{
|
|
|
|
struct intel_engine_cs *engine = rq->engine;
|
|
|
|
struct i915_gem_context *hung_ctx = rq->gem_context;
|
|
|
|
|
2019-01-25 21:22:28 +08:00
|
|
|
lockdep_assert_held(&engine->timeline.lock);
|
2019-01-16 23:33:04 +08:00
|
|
|
|
2019-02-14 02:27:37 +08:00
|
|
|
if (!i915_request_is_active(rq))
|
|
|
|
return;
|
2019-01-16 23:33:04 +08:00
|
|
|
|
2019-02-14 02:27:37 +08:00
|
|
|
list_for_each_entry_continue(rq, &engine->timeline.requests, link)
|
|
|
|
if (rq->gem_context == hung_ctx)
|
|
|
|
i915_request_skip(rq, -EIO);
|
2019-01-16 23:33:04 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
static void client_mark_guilty(struct drm_i915_file_private *file_priv,
|
|
|
|
const struct i915_gem_context *ctx)
|
|
|
|
{
|
|
|
|
unsigned int score;
|
|
|
|
unsigned long prev_hang;
|
|
|
|
|
|
|
|
if (i915_gem_context_is_banned(ctx))
|
|
|
|
score = I915_CLIENT_SCORE_CONTEXT_BAN;
|
|
|
|
else
|
|
|
|
score = 0;
|
|
|
|
|
|
|
|
prev_hang = xchg(&file_priv->hang_timestamp, jiffies);
|
|
|
|
if (time_before(jiffies, prev_hang + I915_CLIENT_FAST_HANG_JIFFIES))
|
|
|
|
score += I915_CLIENT_SCORE_HANG_FAST;
|
|
|
|
|
|
|
|
if (score) {
|
|
|
|
atomic_add(score, &file_priv->ban_score);
|
|
|
|
|
|
|
|
DRM_DEBUG_DRIVER("client %s: gained %u ban score, now %u\n",
|
|
|
|
ctx->name, score,
|
|
|
|
atomic_read(&file_priv->ban_score));
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2019-01-25 21:22:28 +08:00
|
|
|
static bool context_mark_guilty(struct i915_gem_context *ctx)
|
2019-01-16 23:33:04 +08:00
|
|
|
{
|
drm/i915: Use time based guilty context banning
Currently, we accumulate each time a context hangs the GPU, offset
against the number of requests it submits, and if that score exceeds a
certain threshold, we ban that context from submitting any more requests
(cancelling any work in flight). In contrast, we use a simple timer on
the file, that if we see more than a 9 hangs faster than 60s apart in
total across all of its contexts, we will ban the client from creating
any more contexts. This leads to a confusing situation where the file
may be banned before the context, so lets use a simple timer scheme for
each.
If the context submits 3 hanging requests within a 120s period, declare
it forbidden to ever send more requests.
This has the advantage of not being easy to repair by simply sending
empty requests, but has the disadvantage that if the context is idle
then it is forgiven. However, if the context is idle, it is not
disrupting the system, but a hog can evade the request counting and
cause much more severe disruption to the system.
Updating ban_score from request retirement is dubious as the retirement
is purposely not in sync with request submission (i.e. we try and batch
retirement to reduce overhead and avoid latency on submission), which
leads to surprising situations where we can forgive a hang immediately
due to a backlog of requests from before the hang being retired
afterwards.
Signed-off-by: Chris Wilson <chris@chris-wilson.co.uk>
Cc: Mika Kuoppala <mika.kuoppala@intel.com>
Reviewed-by: Mika Kuoppala <mika.kuoppala@intel.com>
Link: https://patchwork.freedesktop.org/patch/msgid/20190219122215.8941-2-chris@chris-wilson.co.uk
2019-02-19 20:21:52 +08:00
|
|
|
unsigned long prev_hang;
|
|
|
|
bool banned;
|
|
|
|
int i;
|
2019-01-16 23:33:04 +08:00
|
|
|
|
|
|
|
atomic_inc(&ctx->guilty_count);
|
|
|
|
|
drm/i915: Use time based guilty context banning
Currently, we accumulate each time a context hangs the GPU, offset
against the number of requests it submits, and if that score exceeds a
certain threshold, we ban that context from submitting any more requests
(cancelling any work in flight). In contrast, we use a simple timer on
the file, that if we see more than a 9 hangs faster than 60s apart in
total across all of its contexts, we will ban the client from creating
any more contexts. This leads to a confusing situation where the file
may be banned before the context, so lets use a simple timer scheme for
each.
If the context submits 3 hanging requests within a 120s period, declare
it forbidden to ever send more requests.
This has the advantage of not being easy to repair by simply sending
empty requests, but has the disadvantage that if the context is idle
then it is forgiven. However, if the context is idle, it is not
disrupting the system, but a hog can evade the request counting and
cause much more severe disruption to the system.
Updating ban_score from request retirement is dubious as the retirement
is purposely not in sync with request submission (i.e. we try and batch
retirement to reduce overhead and avoid latency on submission), which
leads to surprising situations where we can forgive a hang immediately
due to a backlog of requests from before the hang being retired
afterwards.
Signed-off-by: Chris Wilson <chris@chris-wilson.co.uk>
Cc: Mika Kuoppala <mika.kuoppala@intel.com>
Reviewed-by: Mika Kuoppala <mika.kuoppala@intel.com>
Link: https://patchwork.freedesktop.org/patch/msgid/20190219122215.8941-2-chris@chris-wilson.co.uk
2019-02-19 20:21:52 +08:00
|
|
|
/* Cool contexts are too cool to be banned! (Used for reset testing.) */
|
|
|
|
if (!i915_gem_context_is_bannable(ctx))
|
2019-01-25 21:22:28 +08:00
|
|
|
return false;
|
2019-01-16 23:33:04 +08:00
|
|
|
|
drm/i915: Use time based guilty context banning
Currently, we accumulate each time a context hangs the GPU, offset
against the number of requests it submits, and if that score exceeds a
certain threshold, we ban that context from submitting any more requests
(cancelling any work in flight). In contrast, we use a simple timer on
the file, that if we see more than a 9 hangs faster than 60s apart in
total across all of its contexts, we will ban the client from creating
any more contexts. This leads to a confusing situation where the file
may be banned before the context, so lets use a simple timer scheme for
each.
If the context submits 3 hanging requests within a 120s period, declare
it forbidden to ever send more requests.
This has the advantage of not being easy to repair by simply sending
empty requests, but has the disadvantage that if the context is idle
then it is forgiven. However, if the context is idle, it is not
disrupting the system, but a hog can evade the request counting and
cause much more severe disruption to the system.
Updating ban_score from request retirement is dubious as the retirement
is purposely not in sync with request submission (i.e. we try and batch
retirement to reduce overhead and avoid latency on submission), which
leads to surprising situations where we can forgive a hang immediately
due to a backlog of requests from before the hang being retired
afterwards.
Signed-off-by: Chris Wilson <chris@chris-wilson.co.uk>
Cc: Mika Kuoppala <mika.kuoppala@intel.com>
Reviewed-by: Mika Kuoppala <mika.kuoppala@intel.com>
Link: https://patchwork.freedesktop.org/patch/msgid/20190219122215.8941-2-chris@chris-wilson.co.uk
2019-02-19 20:21:52 +08:00
|
|
|
/* Record the timestamp for the last N hangs */
|
|
|
|
prev_hang = ctx->hang_timestamp[0];
|
|
|
|
for (i = 0; i < ARRAY_SIZE(ctx->hang_timestamp) - 1; i++)
|
|
|
|
ctx->hang_timestamp[i] = ctx->hang_timestamp[i + 1];
|
|
|
|
ctx->hang_timestamp[i] = jiffies;
|
|
|
|
|
|
|
|
/* If we have hung N+1 times in rapid succession, we ban the context! */
|
|
|
|
banned = !i915_gem_context_is_recoverable(ctx);
|
|
|
|
if (time_before(jiffies, prev_hang + CONTEXT_FAST_HANG_JIFFIES))
|
|
|
|
banned = true;
|
2019-01-16 23:33:04 +08:00
|
|
|
if (banned) {
|
drm/i915: Use time based guilty context banning
Currently, we accumulate each time a context hangs the GPU, offset
against the number of requests it submits, and if that score exceeds a
certain threshold, we ban that context from submitting any more requests
(cancelling any work in flight). In contrast, we use a simple timer on
the file, that if we see more than a 9 hangs faster than 60s apart in
total across all of its contexts, we will ban the client from creating
any more contexts. This leads to a confusing situation where the file
may be banned before the context, so lets use a simple timer scheme for
each.
If the context submits 3 hanging requests within a 120s period, declare
it forbidden to ever send more requests.
This has the advantage of not being easy to repair by simply sending
empty requests, but has the disadvantage that if the context is idle
then it is forgiven. However, if the context is idle, it is not
disrupting the system, but a hog can evade the request counting and
cause much more severe disruption to the system.
Updating ban_score from request retirement is dubious as the retirement
is purposely not in sync with request submission (i.e. we try and batch
retirement to reduce overhead and avoid latency on submission), which
leads to surprising situations where we can forgive a hang immediately
due to a backlog of requests from before the hang being retired
afterwards.
Signed-off-by: Chris Wilson <chris@chris-wilson.co.uk>
Cc: Mika Kuoppala <mika.kuoppala@intel.com>
Reviewed-by: Mika Kuoppala <mika.kuoppala@intel.com>
Link: https://patchwork.freedesktop.org/patch/msgid/20190219122215.8941-2-chris@chris-wilson.co.uk
2019-02-19 20:21:52 +08:00
|
|
|
DRM_DEBUG_DRIVER("context %s: guilty %d, banned\n",
|
|
|
|
ctx->name, atomic_read(&ctx->guilty_count));
|
2019-01-16 23:33:04 +08:00
|
|
|
i915_gem_context_set_banned(ctx);
|
|
|
|
}
|
|
|
|
|
|
|
|
if (!IS_ERR_OR_NULL(ctx->file_priv))
|
|
|
|
client_mark_guilty(ctx->file_priv, ctx);
|
2019-01-25 21:22:28 +08:00
|
|
|
|
|
|
|
return banned;
|
2019-01-16 23:33:04 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
static void context_mark_innocent(struct i915_gem_context *ctx)
|
|
|
|
{
|
|
|
|
atomic_inc(&ctx->active_count);
|
|
|
|
}
|
|
|
|
|
2019-01-25 21:22:28 +08:00
|
|
|
void i915_reset_request(struct i915_request *rq, bool guilty)
|
|
|
|
{
|
2019-03-12 19:11:45 +08:00
|
|
|
GEM_TRACE("%s rq=%llx:%lld, guilty? %s\n",
|
|
|
|
rq->engine->name,
|
|
|
|
rq->fence.context,
|
|
|
|
rq->fence.seqno,
|
|
|
|
yesno(guilty));
|
|
|
|
|
2019-01-25 21:22:28 +08:00
|
|
|
lockdep_assert_held(&rq->engine->timeline.lock);
|
|
|
|
GEM_BUG_ON(i915_request_completed(rq));
|
|
|
|
|
|
|
|
if (guilty) {
|
|
|
|
i915_request_skip(rq, -EIO);
|
|
|
|
if (context_mark_guilty(rq->gem_context))
|
|
|
|
engine_skip_context(rq);
|
|
|
|
} else {
|
|
|
|
dma_fence_set_error(&rq->fence, -EAGAIN);
|
|
|
|
context_mark_innocent(rq->gem_context);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2019-01-16 23:33:04 +08:00
|
|
|
static void gen3_stop_engine(struct intel_engine_cs *engine)
|
|
|
|
{
|
|
|
|
struct drm_i915_private *dev_priv = engine->i915;
|
|
|
|
const u32 base = engine->mmio_base;
|
|
|
|
|
2019-02-14 07:20:47 +08:00
|
|
|
GEM_TRACE("%s\n", engine->name);
|
|
|
|
|
2019-01-16 23:33:04 +08:00
|
|
|
if (intel_engine_stop_cs(engine))
|
2019-02-14 07:20:47 +08:00
|
|
|
GEM_TRACE("%s: timed out on STOP_RING\n", engine->name);
|
2019-01-16 23:33:04 +08:00
|
|
|
|
|
|
|
I915_WRITE_FW(RING_HEAD(base), I915_READ_FW(RING_TAIL(base)));
|
|
|
|
POSTING_READ_FW(RING_HEAD(base)); /* paranoia */
|
|
|
|
|
|
|
|
I915_WRITE_FW(RING_HEAD(base), 0);
|
|
|
|
I915_WRITE_FW(RING_TAIL(base), 0);
|
|
|
|
POSTING_READ_FW(RING_TAIL(base));
|
|
|
|
|
|
|
|
/* The ring must be empty before it is disabled */
|
|
|
|
I915_WRITE_FW(RING_CTL(base), 0);
|
|
|
|
|
|
|
|
/* Check acts as a post */
|
2019-02-14 07:20:47 +08:00
|
|
|
if (I915_READ_FW(RING_HEAD(base)))
|
|
|
|
GEM_TRACE("%s: ring head [%x] not parked\n",
|
|
|
|
engine->name, I915_READ_FW(RING_HEAD(base)));
|
2019-01-16 23:33:04 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
static void i915_stop_engines(struct drm_i915_private *i915,
|
|
|
|
unsigned int engine_mask)
|
|
|
|
{
|
|
|
|
struct intel_engine_cs *engine;
|
|
|
|
enum intel_engine_id id;
|
|
|
|
|
|
|
|
if (INTEL_GEN(i915) < 3)
|
|
|
|
return;
|
|
|
|
|
|
|
|
for_each_engine_masked(engine, i915, engine_mask, id)
|
|
|
|
gen3_stop_engine(engine);
|
|
|
|
}
|
|
|
|
|
|
|
|
static bool i915_in_reset(struct pci_dev *pdev)
|
|
|
|
{
|
|
|
|
u8 gdrst;
|
|
|
|
|
|
|
|
pci_read_config_byte(pdev, I915_GDRST, &gdrst);
|
|
|
|
return gdrst & GRDOM_RESET_STATUS;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int i915_do_reset(struct drm_i915_private *i915,
|
|
|
|
unsigned int engine_mask,
|
|
|
|
unsigned int retry)
|
|
|
|
{
|
|
|
|
struct pci_dev *pdev = i915->drm.pdev;
|
|
|
|
int err;
|
|
|
|
|
|
|
|
/* Assert reset for at least 20 usec, and wait for acknowledgement. */
|
|
|
|
pci_write_config_byte(pdev, I915_GDRST, GRDOM_RESET_ENABLE);
|
2019-01-25 21:22:26 +08:00
|
|
|
udelay(50);
|
|
|
|
err = wait_for_atomic(i915_in_reset(pdev), 50);
|
2019-01-16 23:33:04 +08:00
|
|
|
|
|
|
|
/* Clear the reset request. */
|
|
|
|
pci_write_config_byte(pdev, I915_GDRST, 0);
|
2019-01-25 21:22:26 +08:00
|
|
|
udelay(50);
|
2019-01-16 23:33:04 +08:00
|
|
|
if (!err)
|
2019-01-25 21:22:26 +08:00
|
|
|
err = wait_for_atomic(!i915_in_reset(pdev), 50);
|
2019-01-16 23:33:04 +08:00
|
|
|
|
|
|
|
return err;
|
|
|
|
}
|
|
|
|
|
|
|
|
static bool g4x_reset_complete(struct pci_dev *pdev)
|
|
|
|
{
|
|
|
|
u8 gdrst;
|
|
|
|
|
|
|
|
pci_read_config_byte(pdev, I915_GDRST, &gdrst);
|
|
|
|
return (gdrst & GRDOM_RESET_ENABLE) == 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int g33_do_reset(struct drm_i915_private *i915,
|
|
|
|
unsigned int engine_mask,
|
|
|
|
unsigned int retry)
|
|
|
|
{
|
|
|
|
struct pci_dev *pdev = i915->drm.pdev;
|
|
|
|
|
|
|
|
pci_write_config_byte(pdev, I915_GDRST, GRDOM_RESET_ENABLE);
|
2019-01-25 21:22:26 +08:00
|
|
|
return wait_for_atomic(g4x_reset_complete(pdev), 50);
|
2019-01-16 23:33:04 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
static int g4x_do_reset(struct drm_i915_private *dev_priv,
|
|
|
|
unsigned int engine_mask,
|
|
|
|
unsigned int retry)
|
|
|
|
{
|
|
|
|
struct pci_dev *pdev = dev_priv->drm.pdev;
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
/* WaVcpClkGateDisableForMediaReset:ctg,elk */
|
2019-01-25 21:22:26 +08:00
|
|
|
I915_WRITE_FW(VDECCLK_GATE_D,
|
|
|
|
I915_READ(VDECCLK_GATE_D) | VCP_UNIT_CLOCK_GATE_DISABLE);
|
|
|
|
POSTING_READ_FW(VDECCLK_GATE_D);
|
2019-01-16 23:33:04 +08:00
|
|
|
|
|
|
|
pci_write_config_byte(pdev, I915_GDRST,
|
|
|
|
GRDOM_MEDIA | GRDOM_RESET_ENABLE);
|
2019-01-25 21:22:26 +08:00
|
|
|
ret = wait_for_atomic(g4x_reset_complete(pdev), 50);
|
2019-01-16 23:33:04 +08:00
|
|
|
if (ret) {
|
|
|
|
DRM_DEBUG_DRIVER("Wait for media reset failed\n");
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
|
|
|
|
pci_write_config_byte(pdev, I915_GDRST,
|
|
|
|
GRDOM_RENDER | GRDOM_RESET_ENABLE);
|
2019-01-25 21:22:26 +08:00
|
|
|
ret = wait_for_atomic(g4x_reset_complete(pdev), 50);
|
2019-01-16 23:33:04 +08:00
|
|
|
if (ret) {
|
|
|
|
DRM_DEBUG_DRIVER("Wait for render reset failed\n");
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
|
|
|
|
out:
|
|
|
|
pci_write_config_byte(pdev, I915_GDRST, 0);
|
|
|
|
|
2019-01-25 21:22:26 +08:00
|
|
|
I915_WRITE_FW(VDECCLK_GATE_D,
|
|
|
|
I915_READ(VDECCLK_GATE_D) & ~VCP_UNIT_CLOCK_GATE_DISABLE);
|
|
|
|
POSTING_READ_FW(VDECCLK_GATE_D);
|
2019-01-16 23:33:04 +08:00
|
|
|
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int ironlake_do_reset(struct drm_i915_private *dev_priv,
|
|
|
|
unsigned int engine_mask,
|
|
|
|
unsigned int retry)
|
|
|
|
{
|
|
|
|
int ret;
|
|
|
|
|
2019-01-25 21:22:26 +08:00
|
|
|
I915_WRITE_FW(ILK_GDSR, ILK_GRDOM_RENDER | ILK_GRDOM_RESET_ENABLE);
|
|
|
|
ret = __intel_wait_for_register_fw(dev_priv, ILK_GDSR,
|
|
|
|
ILK_GRDOM_RESET_ENABLE, 0,
|
|
|
|
5000, 0,
|
|
|
|
NULL);
|
2019-01-16 23:33:04 +08:00
|
|
|
if (ret) {
|
|
|
|
DRM_DEBUG_DRIVER("Wait for render reset failed\n");
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
|
2019-01-25 21:22:26 +08:00
|
|
|
I915_WRITE_FW(ILK_GDSR, ILK_GRDOM_MEDIA | ILK_GRDOM_RESET_ENABLE);
|
|
|
|
ret = __intel_wait_for_register_fw(dev_priv, ILK_GDSR,
|
|
|
|
ILK_GRDOM_RESET_ENABLE, 0,
|
|
|
|
5000, 0,
|
|
|
|
NULL);
|
2019-01-16 23:33:04 +08:00
|
|
|
if (ret) {
|
|
|
|
DRM_DEBUG_DRIVER("Wait for media reset failed\n");
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
|
|
|
|
out:
|
2019-01-25 21:22:26 +08:00
|
|
|
I915_WRITE_FW(ILK_GDSR, 0);
|
|
|
|
POSTING_READ_FW(ILK_GDSR);
|
2019-01-16 23:33:04 +08:00
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Reset the hardware domains (GENX_GRDOM_*) specified by mask */
|
|
|
|
static int gen6_hw_domain_reset(struct drm_i915_private *dev_priv,
|
|
|
|
u32 hw_domain_mask)
|
|
|
|
{
|
|
|
|
int err;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* GEN6_GDRST is not in the gt power well, no need to check
|
|
|
|
* for fifo space for the write or forcewake the chip for
|
|
|
|
* the read
|
|
|
|
*/
|
|
|
|
I915_WRITE_FW(GEN6_GDRST, hw_domain_mask);
|
|
|
|
|
|
|
|
/* Wait for the device to ack the reset requests */
|
|
|
|
err = __intel_wait_for_register_fw(dev_priv,
|
|
|
|
GEN6_GDRST, hw_domain_mask, 0,
|
|
|
|
500, 0,
|
|
|
|
NULL);
|
|
|
|
if (err)
|
|
|
|
DRM_DEBUG_DRIVER("Wait for 0x%08x engines reset failed\n",
|
|
|
|
hw_domain_mask);
|
|
|
|
|
|
|
|
return err;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int gen6_reset_engines(struct drm_i915_private *i915,
|
|
|
|
unsigned int engine_mask,
|
|
|
|
unsigned int retry)
|
|
|
|
{
|
|
|
|
struct intel_engine_cs *engine;
|
2019-03-06 02:03:30 +08:00
|
|
|
const u32 hw_engine_mask[] = {
|
|
|
|
[RCS0] = GEN6_GRDOM_RENDER,
|
|
|
|
[BCS0] = GEN6_GRDOM_BLT,
|
|
|
|
[VCS0] = GEN6_GRDOM_MEDIA,
|
|
|
|
[VCS1] = GEN8_GRDOM_MEDIA2,
|
|
|
|
[VECS0] = GEN6_GRDOM_VECS,
|
2019-01-16 23:33:04 +08:00
|
|
|
};
|
|
|
|
u32 hw_mask;
|
|
|
|
|
|
|
|
if (engine_mask == ALL_ENGINES) {
|
|
|
|
hw_mask = GEN6_GRDOM_FULL;
|
|
|
|
} else {
|
|
|
|
unsigned int tmp;
|
|
|
|
|
|
|
|
hw_mask = 0;
|
2019-03-06 02:03:30 +08:00
|
|
|
for_each_engine_masked(engine, i915, engine_mask, tmp) {
|
|
|
|
GEM_BUG_ON(engine->id >= ARRAY_SIZE(hw_engine_mask));
|
2019-01-16 23:33:04 +08:00
|
|
|
hw_mask |= hw_engine_mask[engine->id];
|
2019-03-06 02:03:30 +08:00
|
|
|
}
|
2019-01-16 23:33:04 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
return gen6_hw_domain_reset(i915, hw_mask);
|
|
|
|
}
|
|
|
|
|
|
|
|
static u32 gen11_lock_sfc(struct drm_i915_private *dev_priv,
|
|
|
|
struct intel_engine_cs *engine)
|
|
|
|
{
|
|
|
|
u8 vdbox_sfc_access = RUNTIME_INFO(dev_priv)->vdbox_sfc_access;
|
|
|
|
i915_reg_t sfc_forced_lock, sfc_forced_lock_ack;
|
|
|
|
u32 sfc_forced_lock_bit, sfc_forced_lock_ack_bit;
|
|
|
|
i915_reg_t sfc_usage;
|
|
|
|
u32 sfc_usage_bit;
|
|
|
|
u32 sfc_reset_bit;
|
|
|
|
|
|
|
|
switch (engine->class) {
|
|
|
|
case VIDEO_DECODE_CLASS:
|
|
|
|
if ((BIT(engine->instance) & vdbox_sfc_access) == 0)
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
sfc_forced_lock = GEN11_VCS_SFC_FORCED_LOCK(engine);
|
|
|
|
sfc_forced_lock_bit = GEN11_VCS_SFC_FORCED_LOCK_BIT;
|
|
|
|
|
|
|
|
sfc_forced_lock_ack = GEN11_VCS_SFC_LOCK_STATUS(engine);
|
|
|
|
sfc_forced_lock_ack_bit = GEN11_VCS_SFC_LOCK_ACK_BIT;
|
|
|
|
|
|
|
|
sfc_usage = GEN11_VCS_SFC_LOCK_STATUS(engine);
|
|
|
|
sfc_usage_bit = GEN11_VCS_SFC_USAGE_BIT;
|
|
|
|
sfc_reset_bit = GEN11_VCS_SFC_RESET_BIT(engine->instance);
|
|
|
|
break;
|
|
|
|
|
|
|
|
case VIDEO_ENHANCEMENT_CLASS:
|
|
|
|
sfc_forced_lock = GEN11_VECS_SFC_FORCED_LOCK(engine);
|
|
|
|
sfc_forced_lock_bit = GEN11_VECS_SFC_FORCED_LOCK_BIT;
|
|
|
|
|
|
|
|
sfc_forced_lock_ack = GEN11_VECS_SFC_LOCK_ACK(engine);
|
|
|
|
sfc_forced_lock_ack_bit = GEN11_VECS_SFC_LOCK_ACK_BIT;
|
|
|
|
|
|
|
|
sfc_usage = GEN11_VECS_SFC_USAGE(engine);
|
|
|
|
sfc_usage_bit = GEN11_VECS_SFC_USAGE_BIT;
|
|
|
|
sfc_reset_bit = GEN11_VECS_SFC_RESET_BIT(engine->instance);
|
|
|
|
break;
|
|
|
|
|
|
|
|
default:
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Tell the engine that a software reset is going to happen. The engine
|
|
|
|
* will then try to force lock the SFC (if currently locked, it will
|
|
|
|
* remain so until we tell the engine it is safe to unlock; if currently
|
|
|
|
* unlocked, it will ignore this and all new lock requests). If SFC
|
|
|
|
* ends up being locked to the engine we want to reset, we have to reset
|
|
|
|
* it as well (we will unlock it once the reset sequence is completed).
|
|
|
|
*/
|
|
|
|
I915_WRITE_FW(sfc_forced_lock,
|
|
|
|
I915_READ_FW(sfc_forced_lock) | sfc_forced_lock_bit);
|
|
|
|
|
|
|
|
if (__intel_wait_for_register_fw(dev_priv,
|
|
|
|
sfc_forced_lock_ack,
|
|
|
|
sfc_forced_lock_ack_bit,
|
|
|
|
sfc_forced_lock_ack_bit,
|
|
|
|
1000, 0, NULL)) {
|
|
|
|
DRM_DEBUG_DRIVER("Wait for SFC forced lock ack failed\n");
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (I915_READ_FW(sfc_usage) & sfc_usage_bit)
|
|
|
|
return sfc_reset_bit;
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void gen11_unlock_sfc(struct drm_i915_private *dev_priv,
|
|
|
|
struct intel_engine_cs *engine)
|
|
|
|
{
|
|
|
|
u8 vdbox_sfc_access = RUNTIME_INFO(dev_priv)->vdbox_sfc_access;
|
|
|
|
i915_reg_t sfc_forced_lock;
|
|
|
|
u32 sfc_forced_lock_bit;
|
|
|
|
|
|
|
|
switch (engine->class) {
|
|
|
|
case VIDEO_DECODE_CLASS:
|
|
|
|
if ((BIT(engine->instance) & vdbox_sfc_access) == 0)
|
|
|
|
return;
|
|
|
|
|
|
|
|
sfc_forced_lock = GEN11_VCS_SFC_FORCED_LOCK(engine);
|
|
|
|
sfc_forced_lock_bit = GEN11_VCS_SFC_FORCED_LOCK_BIT;
|
|
|
|
break;
|
|
|
|
|
|
|
|
case VIDEO_ENHANCEMENT_CLASS:
|
|
|
|
sfc_forced_lock = GEN11_VECS_SFC_FORCED_LOCK(engine);
|
|
|
|
sfc_forced_lock_bit = GEN11_VECS_SFC_FORCED_LOCK_BIT;
|
|
|
|
break;
|
|
|
|
|
|
|
|
default:
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
I915_WRITE_FW(sfc_forced_lock,
|
|
|
|
I915_READ_FW(sfc_forced_lock) & ~sfc_forced_lock_bit);
|
|
|
|
}
|
|
|
|
|
|
|
|
static int gen11_reset_engines(struct drm_i915_private *i915,
|
|
|
|
unsigned int engine_mask,
|
|
|
|
unsigned int retry)
|
|
|
|
{
|
2019-03-06 02:03:30 +08:00
|
|
|
const u32 hw_engine_mask[] = {
|
|
|
|
[RCS0] = GEN11_GRDOM_RENDER,
|
|
|
|
[BCS0] = GEN11_GRDOM_BLT,
|
|
|
|
[VCS0] = GEN11_GRDOM_MEDIA,
|
|
|
|
[VCS1] = GEN11_GRDOM_MEDIA2,
|
|
|
|
[VCS2] = GEN11_GRDOM_MEDIA3,
|
|
|
|
[VCS3] = GEN11_GRDOM_MEDIA4,
|
|
|
|
[VECS0] = GEN11_GRDOM_VECS,
|
|
|
|
[VECS1] = GEN11_GRDOM_VECS2,
|
2019-01-16 23:33:04 +08:00
|
|
|
};
|
|
|
|
struct intel_engine_cs *engine;
|
|
|
|
unsigned int tmp;
|
|
|
|
u32 hw_mask;
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
if (engine_mask == ALL_ENGINES) {
|
|
|
|
hw_mask = GEN11_GRDOM_FULL;
|
|
|
|
} else {
|
|
|
|
hw_mask = 0;
|
|
|
|
for_each_engine_masked(engine, i915, engine_mask, tmp) {
|
2019-03-06 02:03:30 +08:00
|
|
|
GEM_BUG_ON(engine->id >= ARRAY_SIZE(hw_engine_mask));
|
2019-01-16 23:33:04 +08:00
|
|
|
hw_mask |= hw_engine_mask[engine->id];
|
|
|
|
hw_mask |= gen11_lock_sfc(i915, engine);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
ret = gen6_hw_domain_reset(i915, hw_mask);
|
|
|
|
|
|
|
|
if (engine_mask != ALL_ENGINES)
|
|
|
|
for_each_engine_masked(engine, i915, engine_mask, tmp)
|
|
|
|
gen11_unlock_sfc(i915, engine);
|
|
|
|
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int gen8_engine_reset_prepare(struct intel_engine_cs *engine)
|
|
|
|
{
|
|
|
|
struct drm_i915_private *dev_priv = engine->i915;
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
I915_WRITE_FW(RING_RESET_CTL(engine->mmio_base),
|
|
|
|
_MASKED_BIT_ENABLE(RESET_CTL_REQUEST_RESET));
|
|
|
|
|
|
|
|
ret = __intel_wait_for_register_fw(dev_priv,
|
|
|
|
RING_RESET_CTL(engine->mmio_base),
|
|
|
|
RESET_CTL_READY_TO_RESET,
|
|
|
|
RESET_CTL_READY_TO_RESET,
|
|
|
|
700, 0,
|
|
|
|
NULL);
|
|
|
|
if (ret)
|
|
|
|
DRM_ERROR("%s: reset request timeout\n", engine->name);
|
|
|
|
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void gen8_engine_reset_cancel(struct intel_engine_cs *engine)
|
|
|
|
{
|
|
|
|
struct drm_i915_private *dev_priv = engine->i915;
|
|
|
|
|
|
|
|
I915_WRITE_FW(RING_RESET_CTL(engine->mmio_base),
|
|
|
|
_MASKED_BIT_DISABLE(RESET_CTL_REQUEST_RESET));
|
|
|
|
}
|
|
|
|
|
|
|
|
static int gen8_reset_engines(struct drm_i915_private *i915,
|
|
|
|
unsigned int engine_mask,
|
|
|
|
unsigned int retry)
|
|
|
|
{
|
|
|
|
struct intel_engine_cs *engine;
|
|
|
|
const bool reset_non_ready = retry >= 1;
|
|
|
|
unsigned int tmp;
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
for_each_engine_masked(engine, i915, engine_mask, tmp) {
|
|
|
|
ret = gen8_engine_reset_prepare(engine);
|
|
|
|
if (ret && !reset_non_ready)
|
|
|
|
goto skip_reset;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* If this is not the first failed attempt to prepare,
|
|
|
|
* we decide to proceed anyway.
|
|
|
|
*
|
|
|
|
* By doing so we risk context corruption and with
|
|
|
|
* some gens (kbl), possible system hang if reset
|
|
|
|
* happens during active bb execution.
|
|
|
|
*
|
|
|
|
* We rather take context corruption instead of
|
|
|
|
* failed reset with a wedged driver/gpu. And
|
|
|
|
* active bb execution case should be covered by
|
|
|
|
* i915_stop_engines we have before the reset.
|
|
|
|
*/
|
|
|
|
}
|
|
|
|
|
|
|
|
if (INTEL_GEN(i915) >= 11)
|
|
|
|
ret = gen11_reset_engines(i915, engine_mask, retry);
|
|
|
|
else
|
|
|
|
ret = gen6_reset_engines(i915, engine_mask, retry);
|
|
|
|
|
|
|
|
skip_reset:
|
|
|
|
for_each_engine_masked(engine, i915, engine_mask, tmp)
|
|
|
|
gen8_engine_reset_cancel(engine);
|
|
|
|
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
|
|
|
typedef int (*reset_func)(struct drm_i915_private *,
|
|
|
|
unsigned int engine_mask,
|
|
|
|
unsigned int retry);
|
|
|
|
|
|
|
|
static reset_func intel_get_gpu_reset(struct drm_i915_private *i915)
|
|
|
|
{
|
|
|
|
if (INTEL_GEN(i915) >= 8)
|
|
|
|
return gen8_reset_engines;
|
|
|
|
else if (INTEL_GEN(i915) >= 6)
|
|
|
|
return gen6_reset_engines;
|
|
|
|
else if (INTEL_GEN(i915) >= 5)
|
|
|
|
return ironlake_do_reset;
|
|
|
|
else if (IS_G4X(i915))
|
|
|
|
return g4x_do_reset;
|
|
|
|
else if (IS_G33(i915) || IS_PINEVIEW(i915))
|
|
|
|
return g33_do_reset;
|
|
|
|
else if (INTEL_GEN(i915) >= 3)
|
|
|
|
return i915_do_reset;
|
|
|
|
else
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
|
|
|
int intel_gpu_reset(struct drm_i915_private *i915, unsigned int engine_mask)
|
|
|
|
{
|
2019-01-25 21:22:26 +08:00
|
|
|
const int retries = engine_mask == ALL_ENGINES ? RESET_MAX_RETRIES : 1;
|
|
|
|
reset_func reset;
|
|
|
|
int ret = -ETIMEDOUT;
|
2019-01-16 23:33:04 +08:00
|
|
|
int retry;
|
|
|
|
|
2019-01-25 21:22:26 +08:00
|
|
|
reset = intel_get_gpu_reset(i915);
|
|
|
|
if (!reset)
|
|
|
|
return -ENODEV;
|
2019-01-16 23:33:04 +08:00
|
|
|
|
|
|
|
/*
|
|
|
|
* If the power well sleeps during the reset, the reset
|
|
|
|
* request may be dropped and never completes (causing -EIO).
|
|
|
|
*/
|
|
|
|
intel_uncore_forcewake_get(i915, FORCEWAKE_ALL);
|
2019-01-25 21:22:26 +08:00
|
|
|
for (retry = 0; ret == -ETIMEDOUT && retry < retries; retry++) {
|
2019-01-16 23:33:04 +08:00
|
|
|
/*
|
|
|
|
* We stop engines, otherwise we might get failed reset and a
|
|
|
|
* dead gpu (on elk). Also as modern gpu as kbl can suffer
|
|
|
|
* from system hang if batchbuffer is progressing when
|
|
|
|
* the reset is issued, regardless of READY_TO_RESET ack.
|
|
|
|
* Thus assume it is best to stop engines on all gens
|
|
|
|
* where we have a gpu reset.
|
|
|
|
*
|
|
|
|
* WaKBLVECSSemaphoreWaitPoll:kbl (on ALL_ENGINES)
|
|
|
|
*
|
|
|
|
* WaMediaResetMainRingCleanup:ctg,elk (presumably)
|
|
|
|
*
|
|
|
|
* FIXME: Wa for more modern gens needs to be validated
|
|
|
|
*/
|
2019-02-14 07:20:47 +08:00
|
|
|
if (retry)
|
|
|
|
i915_stop_engines(i915, engine_mask);
|
2019-01-16 23:33:04 +08:00
|
|
|
|
2019-01-25 21:22:26 +08:00
|
|
|
GEM_TRACE("engine_mask=%x\n", engine_mask);
|
|
|
|
preempt_disable();
|
|
|
|
ret = reset(i915, engine_mask, retry);
|
|
|
|
preempt_enable();
|
2019-01-16 23:33:04 +08:00
|
|
|
}
|
|
|
|
intel_uncore_forcewake_put(i915, FORCEWAKE_ALL);
|
|
|
|
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
|
|
|
bool intel_has_gpu_reset(struct drm_i915_private *i915)
|
|
|
|
{
|
2019-01-25 21:22:27 +08:00
|
|
|
if (USES_GUC(i915))
|
|
|
|
return false;
|
|
|
|
|
2019-02-08 23:37:04 +08:00
|
|
|
if (!i915_modparams.reset)
|
|
|
|
return NULL;
|
|
|
|
|
2019-01-16 23:33:04 +08:00
|
|
|
return intel_get_gpu_reset(i915);
|
|
|
|
}
|
|
|
|
|
|
|
|
bool intel_has_reset_engine(struct drm_i915_private *i915)
|
|
|
|
{
|
|
|
|
return INTEL_INFO(i915)->has_reset_engine && i915_modparams.reset >= 2;
|
|
|
|
}
|
|
|
|
|
|
|
|
int intel_reset_guc(struct drm_i915_private *i915)
|
|
|
|
{
|
|
|
|
u32 guc_domain =
|
|
|
|
INTEL_GEN(i915) >= 11 ? GEN11_GRDOM_GUC : GEN9_GRDOM_GUC;
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
GEM_BUG_ON(!HAS_GUC(i915));
|
|
|
|
|
|
|
|
intel_uncore_forcewake_get(i915, FORCEWAKE_ALL);
|
|
|
|
ret = gen6_hw_domain_reset(i915, guc_domain);
|
|
|
|
intel_uncore_forcewake_put(i915, FORCEWAKE_ALL);
|
|
|
|
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Ensure irq handler finishes, and not run again.
|
|
|
|
* Also return the active request so that we only search for it once.
|
|
|
|
*/
|
2019-01-25 21:22:28 +08:00
|
|
|
static void reset_prepare_engine(struct intel_engine_cs *engine)
|
2019-01-16 23:33:04 +08:00
|
|
|
{
|
|
|
|
/*
|
|
|
|
* During the reset sequence, we must prevent the engine from
|
|
|
|
* entering RC6. As the context state is undefined until we restart
|
|
|
|
* the engine, if it does enter RC6 during the reset, the state
|
|
|
|
* written to the powercontext is undefined and so we may lose
|
|
|
|
* GPU state upon resume, i.e. fail to restart after a reset.
|
|
|
|
*/
|
|
|
|
intel_uncore_forcewake_get(engine->i915, FORCEWAKE_ALL);
|
2019-01-25 21:22:28 +08:00
|
|
|
engine->reset.prepare(engine);
|
2019-01-16 23:33:04 +08:00
|
|
|
}
|
|
|
|
|
2019-02-08 23:37:03 +08:00
|
|
|
static void revoke_mmaps(struct drm_i915_private *i915)
|
|
|
|
{
|
|
|
|
int i;
|
|
|
|
|
|
|
|
for (i = 0; i < i915->num_fence_regs; i++) {
|
|
|
|
struct drm_vma_offset_node *node;
|
|
|
|
struct i915_vma *vma;
|
|
|
|
u64 vma_offset;
|
|
|
|
|
|
|
|
vma = READ_ONCE(i915->fence_regs[i].vma);
|
|
|
|
if (!vma)
|
|
|
|
continue;
|
|
|
|
|
|
|
|
if (!i915_vma_has_userfault(vma))
|
|
|
|
continue;
|
|
|
|
|
|
|
|
GEM_BUG_ON(vma->fence != &i915->fence_regs[i]);
|
|
|
|
node = &vma->obj->base.vma_node;
|
|
|
|
vma_offset = vma->ggtt_view.partial.offset << PAGE_SHIFT;
|
|
|
|
unmap_mapping_range(i915->drm.anon_inode->i_mapping,
|
|
|
|
drm_vma_node_offset_addr(node) + vma_offset,
|
|
|
|
vma->size,
|
|
|
|
1);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2019-01-25 21:22:28 +08:00
|
|
|
static void reset_prepare(struct drm_i915_private *i915)
|
2019-01-16 23:33:04 +08:00
|
|
|
{
|
|
|
|
struct intel_engine_cs *engine;
|
|
|
|
enum intel_engine_id id;
|
|
|
|
|
2019-01-25 21:22:28 +08:00
|
|
|
for_each_engine(engine, i915, id)
|
|
|
|
reset_prepare_engine(engine);
|
2019-01-16 23:33:04 +08:00
|
|
|
|
2019-02-20 09:39:27 +08:00
|
|
|
intel_uc_reset_prepare(i915);
|
2019-02-08 23:37:03 +08:00
|
|
|
revoke_mmaps(i915);
|
2019-01-16 23:33:04 +08:00
|
|
|
}
|
|
|
|
|
2019-01-25 21:22:28 +08:00
|
|
|
static int gt_reset(struct drm_i915_private *i915, unsigned int stalled_mask)
|
2019-01-16 23:33:04 +08:00
|
|
|
{
|
2019-01-25 21:22:28 +08:00
|
|
|
struct intel_engine_cs *engine;
|
|
|
|
enum intel_engine_id id;
|
|
|
|
int err;
|
|
|
|
|
2019-01-16 23:33:04 +08:00
|
|
|
/*
|
2019-01-25 21:22:28 +08:00
|
|
|
* Everything depends on having the GTT running, so we need to start
|
|
|
|
* there.
|
2019-01-16 23:33:04 +08:00
|
|
|
*/
|
2019-01-25 21:22:28 +08:00
|
|
|
err = i915_ggtt_enable_hw(i915);
|
|
|
|
if (err)
|
|
|
|
return err;
|
2019-01-16 23:33:04 +08:00
|
|
|
|
2019-01-25 21:22:28 +08:00
|
|
|
for_each_engine(engine, i915, id)
|
2019-03-06 02:03:30 +08:00
|
|
|
intel_engine_reset(engine, stalled_mask & engine->mask);
|
2019-01-16 23:33:04 +08:00
|
|
|
|
2019-01-25 21:22:28 +08:00
|
|
|
i915_gem_restore_fences(i915);
|
2019-01-16 23:33:04 +08:00
|
|
|
|
2019-01-25 21:22:28 +08:00
|
|
|
return err;
|
2019-01-16 23:33:04 +08:00
|
|
|
}
|
|
|
|
|
2019-01-25 21:22:28 +08:00
|
|
|
static void reset_finish_engine(struct intel_engine_cs *engine)
|
2019-01-16 23:33:04 +08:00
|
|
|
{
|
2019-01-25 21:22:28 +08:00
|
|
|
engine->reset.finish(engine);
|
|
|
|
intel_uncore_forcewake_put(engine->i915, FORCEWAKE_ALL);
|
2019-01-16 23:33:04 +08:00
|
|
|
}
|
|
|
|
|
2019-01-25 21:22:28 +08:00
|
|
|
struct i915_gpu_restart {
|
|
|
|
struct work_struct work;
|
|
|
|
struct drm_i915_private *i915;
|
|
|
|
};
|
|
|
|
|
|
|
|
static void restart_work(struct work_struct *work)
|
2019-01-16 23:33:04 +08:00
|
|
|
{
|
2019-01-25 21:22:28 +08:00
|
|
|
struct i915_gpu_restart *arg = container_of(work, typeof(*arg), work);
|
|
|
|
struct drm_i915_private *i915 = arg->i915;
|
2019-01-16 23:33:04 +08:00
|
|
|
struct intel_engine_cs *engine;
|
|
|
|
enum intel_engine_id id;
|
2019-01-25 21:22:28 +08:00
|
|
|
intel_wakeref_t wakeref;
|
2019-01-16 23:33:04 +08:00
|
|
|
|
2019-01-25 21:22:28 +08:00
|
|
|
wakeref = intel_runtime_pm_get(i915);
|
|
|
|
mutex_lock(&i915->drm.struct_mutex);
|
|
|
|
WRITE_ONCE(i915->gpu_error.restart, NULL);
|
2019-01-16 23:33:04 +08:00
|
|
|
|
|
|
|
for_each_engine(engine, i915, id) {
|
2019-01-25 21:22:28 +08:00
|
|
|
struct i915_request *rq;
|
2019-01-16 23:33:04 +08:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Ostensibily, we always want a context loaded for powersaving,
|
|
|
|
* so if the engine is idle after the reset, send a request
|
|
|
|
* to load our scratch kernel_context.
|
|
|
|
*/
|
2019-01-25 21:22:28 +08:00
|
|
|
if (!intel_engine_is_idle(engine))
|
|
|
|
continue;
|
2019-01-16 23:33:04 +08:00
|
|
|
|
2019-01-25 21:22:28 +08:00
|
|
|
rq = i915_request_alloc(engine, i915->kernel_context);
|
|
|
|
if (!IS_ERR(rq))
|
|
|
|
i915_request_add(rq);
|
2019-01-16 23:33:04 +08:00
|
|
|
}
|
|
|
|
|
2019-01-25 21:22:28 +08:00
|
|
|
mutex_unlock(&i915->drm.struct_mutex);
|
|
|
|
intel_runtime_pm_put(i915, wakeref);
|
2019-01-16 23:33:04 +08:00
|
|
|
|
2019-01-25 21:22:28 +08:00
|
|
|
kfree(arg);
|
2019-01-16 23:33:04 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
static void reset_finish(struct drm_i915_private *i915)
|
|
|
|
{
|
|
|
|
struct intel_engine_cs *engine;
|
|
|
|
enum intel_engine_id id;
|
|
|
|
|
2019-01-25 21:22:28 +08:00
|
|
|
for_each_engine(engine, i915, id)
|
2019-01-16 23:33:04 +08:00
|
|
|
reset_finish_engine(engine);
|
2019-01-25 21:22:28 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
static void reset_restart(struct drm_i915_private *i915)
|
|
|
|
{
|
|
|
|
struct i915_gpu_restart *arg;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Following the reset, ensure that we always reload context for
|
|
|
|
* powersaving, and to correct engine->last_retired_context. Since
|
|
|
|
* this requires us to submit a request, queue a worker to do that
|
|
|
|
* task for us to evade any locking here.
|
|
|
|
*/
|
|
|
|
if (READ_ONCE(i915->gpu_error.restart))
|
|
|
|
return;
|
|
|
|
|
|
|
|
arg = kmalloc(sizeof(*arg), GFP_KERNEL);
|
|
|
|
if (arg) {
|
|
|
|
arg->i915 = i915;
|
|
|
|
INIT_WORK(&arg->work, restart_work);
|
|
|
|
|
|
|
|
WRITE_ONCE(i915->gpu_error.restart, arg);
|
|
|
|
queue_work(i915->wq, &arg->work);
|
2019-01-16 23:33:04 +08:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
static void nop_submit_request(struct i915_request *request)
|
|
|
|
{
|
drm/i915: Replace global breadcrumbs with per-context interrupt tracking
A few years ago, see commit 688e6c725816 ("drm/i915: Slaughter the
thundering i915_wait_request herd"), the issue of handling multiple
clients waiting in parallel was brought to our attention. The
requirement was that every client should be woken immediately upon its
request being signaled, without incurring any cpu overhead.
To handle certain fragility of our hw meant that we could not do a
simple check inside the irq handler (some generations required almost
unbounded delays before we could be sure of seqno coherency) and so
request completion checking required delegation.
Before commit 688e6c725816, the solution was simple. Every client
waiting on a request would be woken on every interrupt and each would do
a heavyweight check to see if their request was complete. Commit
688e6c725816 introduced an rbtree so that only the earliest waiter on
the global timeline would woken, and would wake the next and so on.
(Along with various complications to handle requests being reordered
along the global timeline, and also a requirement for kthread to provide
a delegate for fence signaling that had no process context.)
The global rbtree depends on knowing the execution timeline (and global
seqno). Without knowing that order, we must instead check all contexts
queued to the HW to see which may have advanced. We trim that list by
only checking queued contexts that are being waited on, but still we
keep a list of all active contexts and their active signalers that we
inspect from inside the irq handler. By moving the waiters onto the fence
signal list, we can combine the client wakeup with the dma_fence
signaling (a dramatic reduction in complexity, but does require the HW
being coherent, the seqno must be visible from the cpu before the
interrupt is raised - we keep a timer backup just in case).
Having previously fixed all the issues with irq-seqno serialisation (by
inserting delays onto the GPU after each request instead of random delays
on the CPU after each interrupt), we can rely on the seqno state to
perfom direct wakeups from the interrupt handler. This allows us to
preserve our single context switch behaviour of the current routine,
with the only downside that we lose the RT priority sorting of wakeups.
In general, direct wakeup latency of multiple clients is about the same
(about 10% better in most cases) with a reduction in total CPU time spent
in the waiter (about 20-50% depending on gen). Average herd behaviour is
improved, but at the cost of not delegating wakeups on task_prio.
v2: Capture fence signaling state for error state and add comments to
warm even the most cold of hearts.
v3: Check if the request is still active before busywaiting
v4: Reduce the amount of pointer misdirection with list_for_each_safe
and using a local i915_request variable inside the loops
v5: Add a missing pluralisation to a purely informative selftest message.
References: 688e6c725816 ("drm/i915: Slaughter the thundering i915_wait_request herd")
Signed-off-by: Chris Wilson <chris@chris-wilson.co.uk>
Cc: Tvrtko Ursulin <tvrtko.ursulin@intel.com>
Reviewed-by: Tvrtko Ursulin <tvrtko.ursulin@intel.com>
Link: https://patchwork.freedesktop.org/patch/msgid/20190129205230.19056-2-chris@chris-wilson.co.uk
2019-01-30 04:52:29 +08:00
|
|
|
struct intel_engine_cs *engine = request->engine;
|
2019-01-16 23:33:04 +08:00
|
|
|
unsigned long flags;
|
|
|
|
|
|
|
|
GEM_TRACE("%s fence %llx:%lld -> -EIO\n",
|
drm/i915: Replace global breadcrumbs with per-context interrupt tracking
A few years ago, see commit 688e6c725816 ("drm/i915: Slaughter the
thundering i915_wait_request herd"), the issue of handling multiple
clients waiting in parallel was brought to our attention. The
requirement was that every client should be woken immediately upon its
request being signaled, without incurring any cpu overhead.
To handle certain fragility of our hw meant that we could not do a
simple check inside the irq handler (some generations required almost
unbounded delays before we could be sure of seqno coherency) and so
request completion checking required delegation.
Before commit 688e6c725816, the solution was simple. Every client
waiting on a request would be woken on every interrupt and each would do
a heavyweight check to see if their request was complete. Commit
688e6c725816 introduced an rbtree so that only the earliest waiter on
the global timeline would woken, and would wake the next and so on.
(Along with various complications to handle requests being reordered
along the global timeline, and also a requirement for kthread to provide
a delegate for fence signaling that had no process context.)
The global rbtree depends on knowing the execution timeline (and global
seqno). Without knowing that order, we must instead check all contexts
queued to the HW to see which may have advanced. We trim that list by
only checking queued contexts that are being waited on, but still we
keep a list of all active contexts and their active signalers that we
inspect from inside the irq handler. By moving the waiters onto the fence
signal list, we can combine the client wakeup with the dma_fence
signaling (a dramatic reduction in complexity, but does require the HW
being coherent, the seqno must be visible from the cpu before the
interrupt is raised - we keep a timer backup just in case).
Having previously fixed all the issues with irq-seqno serialisation (by
inserting delays onto the GPU after each request instead of random delays
on the CPU after each interrupt), we can rely on the seqno state to
perfom direct wakeups from the interrupt handler. This allows us to
preserve our single context switch behaviour of the current routine,
with the only downside that we lose the RT priority sorting of wakeups.
In general, direct wakeup latency of multiple clients is about the same
(about 10% better in most cases) with a reduction in total CPU time spent
in the waiter (about 20-50% depending on gen). Average herd behaviour is
improved, but at the cost of not delegating wakeups on task_prio.
v2: Capture fence signaling state for error state and add comments to
warm even the most cold of hearts.
v3: Check if the request is still active before busywaiting
v4: Reduce the amount of pointer misdirection with list_for_each_safe
and using a local i915_request variable inside the loops
v5: Add a missing pluralisation to a purely informative selftest message.
References: 688e6c725816 ("drm/i915: Slaughter the thundering i915_wait_request herd")
Signed-off-by: Chris Wilson <chris@chris-wilson.co.uk>
Cc: Tvrtko Ursulin <tvrtko.ursulin@intel.com>
Reviewed-by: Tvrtko Ursulin <tvrtko.ursulin@intel.com>
Link: https://patchwork.freedesktop.org/patch/msgid/20190129205230.19056-2-chris@chris-wilson.co.uk
2019-01-30 04:52:29 +08:00
|
|
|
engine->name, request->fence.context, request->fence.seqno);
|
2019-01-16 23:33:04 +08:00
|
|
|
dma_fence_set_error(&request->fence, -EIO);
|
|
|
|
|
drm/i915: Replace global breadcrumbs with per-context interrupt tracking
A few years ago, see commit 688e6c725816 ("drm/i915: Slaughter the
thundering i915_wait_request herd"), the issue of handling multiple
clients waiting in parallel was brought to our attention. The
requirement was that every client should be woken immediately upon its
request being signaled, without incurring any cpu overhead.
To handle certain fragility of our hw meant that we could not do a
simple check inside the irq handler (some generations required almost
unbounded delays before we could be sure of seqno coherency) and so
request completion checking required delegation.
Before commit 688e6c725816, the solution was simple. Every client
waiting on a request would be woken on every interrupt and each would do
a heavyweight check to see if their request was complete. Commit
688e6c725816 introduced an rbtree so that only the earliest waiter on
the global timeline would woken, and would wake the next and so on.
(Along with various complications to handle requests being reordered
along the global timeline, and also a requirement for kthread to provide
a delegate for fence signaling that had no process context.)
The global rbtree depends on knowing the execution timeline (and global
seqno). Without knowing that order, we must instead check all contexts
queued to the HW to see which may have advanced. We trim that list by
only checking queued contexts that are being waited on, but still we
keep a list of all active contexts and their active signalers that we
inspect from inside the irq handler. By moving the waiters onto the fence
signal list, we can combine the client wakeup with the dma_fence
signaling (a dramatic reduction in complexity, but does require the HW
being coherent, the seqno must be visible from the cpu before the
interrupt is raised - we keep a timer backup just in case).
Having previously fixed all the issues with irq-seqno serialisation (by
inserting delays onto the GPU after each request instead of random delays
on the CPU after each interrupt), we can rely on the seqno state to
perfom direct wakeups from the interrupt handler. This allows us to
preserve our single context switch behaviour of the current routine,
with the only downside that we lose the RT priority sorting of wakeups.
In general, direct wakeup latency of multiple clients is about the same
(about 10% better in most cases) with a reduction in total CPU time spent
in the waiter (about 20-50% depending on gen). Average herd behaviour is
improved, but at the cost of not delegating wakeups on task_prio.
v2: Capture fence signaling state for error state and add comments to
warm even the most cold of hearts.
v3: Check if the request is still active before busywaiting
v4: Reduce the amount of pointer misdirection with list_for_each_safe
and using a local i915_request variable inside the loops
v5: Add a missing pluralisation to a purely informative selftest message.
References: 688e6c725816 ("drm/i915: Slaughter the thundering i915_wait_request herd")
Signed-off-by: Chris Wilson <chris@chris-wilson.co.uk>
Cc: Tvrtko Ursulin <tvrtko.ursulin@intel.com>
Reviewed-by: Tvrtko Ursulin <tvrtko.ursulin@intel.com>
Link: https://patchwork.freedesktop.org/patch/msgid/20190129205230.19056-2-chris@chris-wilson.co.uk
2019-01-30 04:52:29 +08:00
|
|
|
spin_lock_irqsave(&engine->timeline.lock, flags);
|
2019-01-16 23:33:04 +08:00
|
|
|
__i915_request_submit(request);
|
2019-01-29 02:18:11 +08:00
|
|
|
i915_request_mark_complete(request);
|
drm/i915: Replace global breadcrumbs with per-context interrupt tracking
A few years ago, see commit 688e6c725816 ("drm/i915: Slaughter the
thundering i915_wait_request herd"), the issue of handling multiple
clients waiting in parallel was brought to our attention. The
requirement was that every client should be woken immediately upon its
request being signaled, without incurring any cpu overhead.
To handle certain fragility of our hw meant that we could not do a
simple check inside the irq handler (some generations required almost
unbounded delays before we could be sure of seqno coherency) and so
request completion checking required delegation.
Before commit 688e6c725816, the solution was simple. Every client
waiting on a request would be woken on every interrupt and each would do
a heavyweight check to see if their request was complete. Commit
688e6c725816 introduced an rbtree so that only the earliest waiter on
the global timeline would woken, and would wake the next and so on.
(Along with various complications to handle requests being reordered
along the global timeline, and also a requirement for kthread to provide
a delegate for fence signaling that had no process context.)
The global rbtree depends on knowing the execution timeline (and global
seqno). Without knowing that order, we must instead check all contexts
queued to the HW to see which may have advanced. We trim that list by
only checking queued contexts that are being waited on, but still we
keep a list of all active contexts and their active signalers that we
inspect from inside the irq handler. By moving the waiters onto the fence
signal list, we can combine the client wakeup with the dma_fence
signaling (a dramatic reduction in complexity, but does require the HW
being coherent, the seqno must be visible from the cpu before the
interrupt is raised - we keep a timer backup just in case).
Having previously fixed all the issues with irq-seqno serialisation (by
inserting delays onto the GPU after each request instead of random delays
on the CPU after each interrupt), we can rely on the seqno state to
perfom direct wakeups from the interrupt handler. This allows us to
preserve our single context switch behaviour of the current routine,
with the only downside that we lose the RT priority sorting of wakeups.
In general, direct wakeup latency of multiple clients is about the same
(about 10% better in most cases) with a reduction in total CPU time spent
in the waiter (about 20-50% depending on gen). Average herd behaviour is
improved, but at the cost of not delegating wakeups on task_prio.
v2: Capture fence signaling state for error state and add comments to
warm even the most cold of hearts.
v3: Check if the request is still active before busywaiting
v4: Reduce the amount of pointer misdirection with list_for_each_safe
and using a local i915_request variable inside the loops
v5: Add a missing pluralisation to a purely informative selftest message.
References: 688e6c725816 ("drm/i915: Slaughter the thundering i915_wait_request herd")
Signed-off-by: Chris Wilson <chris@chris-wilson.co.uk>
Cc: Tvrtko Ursulin <tvrtko.ursulin@intel.com>
Reviewed-by: Tvrtko Ursulin <tvrtko.ursulin@intel.com>
Link: https://patchwork.freedesktop.org/patch/msgid/20190129205230.19056-2-chris@chris-wilson.co.uk
2019-01-30 04:52:29 +08:00
|
|
|
spin_unlock_irqrestore(&engine->timeline.lock, flags);
|
|
|
|
|
|
|
|
intel_engine_queue_breadcrumbs(engine);
|
2019-01-16 23:33:04 +08:00
|
|
|
}
|
|
|
|
|
2019-02-08 23:37:07 +08:00
|
|
|
static void __i915_gem_set_wedged(struct drm_i915_private *i915)
|
2019-01-16 23:33:04 +08:00
|
|
|
{
|
|
|
|
struct i915_gpu_error *error = &i915->gpu_error;
|
|
|
|
struct intel_engine_cs *engine;
|
|
|
|
enum intel_engine_id id;
|
|
|
|
|
2019-02-08 23:37:07 +08:00
|
|
|
if (test_bit(I915_WEDGED, &error->flags))
|
2019-01-16 23:33:04 +08:00
|
|
|
return;
|
|
|
|
|
|
|
|
if (GEM_SHOW_DEBUG() && !intel_engines_are_idle(i915)) {
|
|
|
|
struct drm_printer p = drm_debug_printer(__func__);
|
|
|
|
|
|
|
|
for_each_engine(engine, i915, id)
|
|
|
|
intel_engine_dump(engine, &p, "%s\n", engine->name);
|
|
|
|
}
|
|
|
|
|
|
|
|
GEM_TRACE("start\n");
|
|
|
|
|
|
|
|
/*
|
|
|
|
* First, stop submission to hw, but do not yet complete requests by
|
|
|
|
* rolling the global seqno forward (since this would complete requests
|
|
|
|
* for which we haven't set the fence error to EIO yet).
|
|
|
|
*/
|
|
|
|
for_each_engine(engine, i915, id)
|
|
|
|
reset_prepare_engine(engine);
|
|
|
|
|
|
|
|
/* Even if the GPU reset fails, it should still stop the engines */
|
2019-02-08 23:37:04 +08:00
|
|
|
if (!INTEL_INFO(i915)->gpu_reset_clobbers_display)
|
2019-01-16 23:33:04 +08:00
|
|
|
intel_gpu_reset(i915, ALL_ENGINES);
|
|
|
|
|
|
|
|
for_each_engine(engine, i915, id) {
|
|
|
|
engine->submit_request = nop_submit_request;
|
|
|
|
engine->schedule = NULL;
|
|
|
|
}
|
|
|
|
i915->caps.scheduler = 0;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Make sure no request can slip through without getting completed by
|
|
|
|
* either this call here to intel_engine_write_global_seqno, or the one
|
|
|
|
* in nop_submit_request.
|
|
|
|
*/
|
2019-02-19 20:21:57 +08:00
|
|
|
synchronize_rcu_expedited();
|
2019-01-16 23:33:04 +08:00
|
|
|
|
|
|
|
/* Mark all executing requests as skipped */
|
|
|
|
for_each_engine(engine, i915, id)
|
|
|
|
engine->cancel_requests(engine);
|
|
|
|
|
|
|
|
for_each_engine(engine, i915, id) {
|
|
|
|
reset_finish_engine(engine);
|
drm/i915: Replace global breadcrumbs with per-context interrupt tracking
A few years ago, see commit 688e6c725816 ("drm/i915: Slaughter the
thundering i915_wait_request herd"), the issue of handling multiple
clients waiting in parallel was brought to our attention. The
requirement was that every client should be woken immediately upon its
request being signaled, without incurring any cpu overhead.
To handle certain fragility of our hw meant that we could not do a
simple check inside the irq handler (some generations required almost
unbounded delays before we could be sure of seqno coherency) and so
request completion checking required delegation.
Before commit 688e6c725816, the solution was simple. Every client
waiting on a request would be woken on every interrupt and each would do
a heavyweight check to see if their request was complete. Commit
688e6c725816 introduced an rbtree so that only the earliest waiter on
the global timeline would woken, and would wake the next and so on.
(Along with various complications to handle requests being reordered
along the global timeline, and also a requirement for kthread to provide
a delegate for fence signaling that had no process context.)
The global rbtree depends on knowing the execution timeline (and global
seqno). Without knowing that order, we must instead check all contexts
queued to the HW to see which may have advanced. We trim that list by
only checking queued contexts that are being waited on, but still we
keep a list of all active contexts and their active signalers that we
inspect from inside the irq handler. By moving the waiters onto the fence
signal list, we can combine the client wakeup with the dma_fence
signaling (a dramatic reduction in complexity, but does require the HW
being coherent, the seqno must be visible from the cpu before the
interrupt is raised - we keep a timer backup just in case).
Having previously fixed all the issues with irq-seqno serialisation (by
inserting delays onto the GPU after each request instead of random delays
on the CPU after each interrupt), we can rely on the seqno state to
perfom direct wakeups from the interrupt handler. This allows us to
preserve our single context switch behaviour of the current routine,
with the only downside that we lose the RT priority sorting of wakeups.
In general, direct wakeup latency of multiple clients is about the same
(about 10% better in most cases) with a reduction in total CPU time spent
in the waiter (about 20-50% depending on gen). Average herd behaviour is
improved, but at the cost of not delegating wakeups on task_prio.
v2: Capture fence signaling state for error state and add comments to
warm even the most cold of hearts.
v3: Check if the request is still active before busywaiting
v4: Reduce the amount of pointer misdirection with list_for_each_safe
and using a local i915_request variable inside the loops
v5: Add a missing pluralisation to a purely informative selftest message.
References: 688e6c725816 ("drm/i915: Slaughter the thundering i915_wait_request herd")
Signed-off-by: Chris Wilson <chris@chris-wilson.co.uk>
Cc: Tvrtko Ursulin <tvrtko.ursulin@intel.com>
Reviewed-by: Tvrtko Ursulin <tvrtko.ursulin@intel.com>
Link: https://patchwork.freedesktop.org/patch/msgid/20190129205230.19056-2-chris@chris-wilson.co.uk
2019-01-30 04:52:29 +08:00
|
|
|
intel_engine_signal_breadcrumbs(engine);
|
2019-01-16 23:33:04 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
smp_mb__before_atomic();
|
|
|
|
set_bit(I915_WEDGED, &error->flags);
|
|
|
|
|
|
|
|
GEM_TRACE("end\n");
|
2019-02-08 23:37:07 +08:00
|
|
|
}
|
2019-01-16 23:33:04 +08:00
|
|
|
|
2019-02-08 23:37:07 +08:00
|
|
|
void i915_gem_set_wedged(struct drm_i915_private *i915)
|
|
|
|
{
|
|
|
|
struct i915_gpu_error *error = &i915->gpu_error;
|
|
|
|
|
|
|
|
mutex_lock(&error->wedge_mutex);
|
|
|
|
__i915_gem_set_wedged(i915);
|
|
|
|
mutex_unlock(&error->wedge_mutex);
|
2019-01-16 23:33:04 +08:00
|
|
|
}
|
|
|
|
|
2019-02-08 23:37:07 +08:00
|
|
|
static bool __i915_gem_unset_wedged(struct drm_i915_private *i915)
|
2019-01-16 23:33:04 +08:00
|
|
|
{
|
|
|
|
struct i915_gpu_error *error = &i915->gpu_error;
|
|
|
|
struct i915_timeline *tl;
|
|
|
|
|
|
|
|
if (!test_bit(I915_WEDGED, &error->flags))
|
|
|
|
return true;
|
|
|
|
|
|
|
|
if (!i915->gt.scratch) /* Never full initialised, recovery impossible */
|
|
|
|
return false;
|
|
|
|
|
|
|
|
GEM_TRACE("start\n");
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Before unwedging, make sure that all pending operations
|
|
|
|
* are flushed and errored out - we may have requests waiting upon
|
|
|
|
* third party fences. We marked all inflight requests as EIO, and
|
|
|
|
* every execbuf since returned EIO, for consistency we want all
|
|
|
|
* the currently pending requests to also be marked as EIO, which
|
|
|
|
* is done inside our nop_submit_request - and so we must wait.
|
|
|
|
*
|
|
|
|
* No more can be submitted until we reset the wedged bit.
|
|
|
|
*/
|
2019-01-28 18:23:56 +08:00
|
|
|
mutex_lock(&i915->gt.timelines.mutex);
|
2019-01-29 02:18:12 +08:00
|
|
|
list_for_each_entry(tl, &i915->gt.timelines.active_list, link) {
|
2019-01-16 23:33:04 +08:00
|
|
|
struct i915_request *rq;
|
|
|
|
|
drm/i915: Pull i915_gem_active into the i915_active family
Looking forward, we need to break the struct_mutex dependency on
i915_gem_active. In the meantime, external use of i915_gem_active is
quite beguiling, little do new users suspect that it implies a barrier
as each request it tracks must be ordered wrt the previous one. As one
of many, it can be used to track activity across multiple timelines, a
shared fence, which fits our unordered request submission much better. We
need to steer external users away from the singular, exclusive fence
imposed by i915_gem_active to i915_active instead. As part of that
process, we move i915_gem_active out of i915_request.c into
i915_active.c to start separating the two concepts, and rename it to
i915_active_request (both to tie it to the concept of tracking just one
request, and to give it a longer, less appealing name).
Signed-off-by: Chris Wilson <chris@chris-wilson.co.uk>
Reviewed-by: Tvrtko Ursulin <tvrtko.ursulin@intel.com>
Link: https://patchwork.freedesktop.org/patch/msgid/20190205130005.2807-5-chris@chris-wilson.co.uk
2019-02-05 21:00:05 +08:00
|
|
|
rq = i915_active_request_get_unlocked(&tl->last_request);
|
2019-01-16 23:33:04 +08:00
|
|
|
if (!rq)
|
|
|
|
continue;
|
|
|
|
|
|
|
|
/*
|
2019-02-08 23:37:05 +08:00
|
|
|
* All internal dependencies (i915_requests) will have
|
|
|
|
* been flushed by the set-wedge, but we may be stuck waiting
|
|
|
|
* for external fences. These should all be capped to 10s
|
|
|
|
* (I915_FENCE_TIMEOUT) so this wait should not be unbounded
|
|
|
|
* in the worst case.
|
2019-01-16 23:33:04 +08:00
|
|
|
*/
|
2019-02-08 23:37:05 +08:00
|
|
|
dma_fence_default_wait(&rq->fence, false, MAX_SCHEDULE_TIMEOUT);
|
2019-01-25 21:22:28 +08:00
|
|
|
i915_request_put(rq);
|
2019-01-16 23:33:04 +08:00
|
|
|
}
|
2019-01-28 18:23:56 +08:00
|
|
|
mutex_unlock(&i915->gt.timelines.mutex);
|
2019-01-16 23:33:04 +08:00
|
|
|
|
|
|
|
intel_engines_sanitize(i915, false);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Undo nop_submit_request. We prevent all new i915 requests from
|
|
|
|
* being queued (by disallowing execbuf whilst wedged) so having
|
|
|
|
* waited for all active requests above, we know the system is idle
|
|
|
|
* and do not have to worry about a thread being inside
|
|
|
|
* engine->submit_request() as we swap over. So unlike installing
|
|
|
|
* the nop_submit_request on reset, we can do this from normal
|
|
|
|
* context and do not require stop_machine().
|
|
|
|
*/
|
|
|
|
intel_engines_reset_default_submission(i915);
|
|
|
|
|
|
|
|
GEM_TRACE("end\n");
|
|
|
|
|
|
|
|
smp_mb__before_atomic(); /* complete takeover before enabling execbuf */
|
|
|
|
clear_bit(I915_WEDGED, &i915->gpu_error.flags);
|
2019-02-08 23:37:05 +08:00
|
|
|
|
|
|
|
return true;
|
2019-01-16 23:33:04 +08:00
|
|
|
}
|
|
|
|
|
2019-02-08 23:37:07 +08:00
|
|
|
bool i915_gem_unset_wedged(struct drm_i915_private *i915)
|
|
|
|
{
|
|
|
|
struct i915_gpu_error *error = &i915->gpu_error;
|
|
|
|
bool result;
|
|
|
|
|
|
|
|
mutex_lock(&error->wedge_mutex);
|
|
|
|
result = __i915_gem_unset_wedged(i915);
|
|
|
|
mutex_unlock(&error->wedge_mutex);
|
|
|
|
|
|
|
|
return result;
|
|
|
|
}
|
|
|
|
|
2019-01-25 21:22:28 +08:00
|
|
|
static int do_reset(struct drm_i915_private *i915, unsigned int stalled_mask)
|
|
|
|
{
|
|
|
|
int err, i;
|
|
|
|
|
2019-02-08 23:37:03 +08:00
|
|
|
err = intel_gpu_reset(i915, ALL_ENGINES);
|
2019-01-25 21:22:28 +08:00
|
|
|
for (i = 0; err && i < RESET_MAX_RETRIES; i++) {
|
2019-02-08 23:37:03 +08:00
|
|
|
msleep(10 * (i + 1));
|
|
|
|
err = intel_gpu_reset(i915, ALL_ENGINES);
|
2019-01-25 21:22:28 +08:00
|
|
|
}
|
2019-02-08 23:37:03 +08:00
|
|
|
if (err)
|
|
|
|
return err;
|
2019-01-25 21:22:28 +08:00
|
|
|
|
2019-02-08 23:37:03 +08:00
|
|
|
return gt_reset(i915, stalled_mask);
|
2019-01-25 21:22:28 +08:00
|
|
|
}
|
|
|
|
|
2019-01-16 23:33:04 +08:00
|
|
|
/**
|
|
|
|
* i915_reset - reset chip after a hang
|
|
|
|
* @i915: #drm_i915_private to reset
|
|
|
|
* @stalled_mask: mask of the stalled engines with the guilty requests
|
|
|
|
* @reason: user error message for why we are resetting
|
|
|
|
*
|
|
|
|
* Reset the chip. Useful if a hang is detected. Marks the device as wedged
|
|
|
|
* on failure.
|
|
|
|
*
|
|
|
|
* Procedure is fairly simple:
|
|
|
|
* - reset the chip using the reset reg
|
|
|
|
* - re-init context state
|
|
|
|
* - re-init hardware status page
|
|
|
|
* - re-init ring buffer
|
|
|
|
* - re-init interrupt state
|
|
|
|
* - re-init display
|
|
|
|
*/
|
|
|
|
void i915_reset(struct drm_i915_private *i915,
|
|
|
|
unsigned int stalled_mask,
|
|
|
|
const char *reason)
|
|
|
|
{
|
|
|
|
struct i915_gpu_error *error = &i915->gpu_error;
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
GEM_TRACE("flags=%lx\n", error->flags);
|
|
|
|
|
|
|
|
might_sleep();
|
|
|
|
assert_rpm_wakelock_held(i915);
|
|
|
|
GEM_BUG_ON(!test_bit(I915_RESET_BACKOFF, &error->flags));
|
|
|
|
|
|
|
|
/* Clear any previous failed attempts at recovery. Time to try again. */
|
2019-02-08 23:37:07 +08:00
|
|
|
if (!__i915_gem_unset_wedged(i915))
|
2019-01-25 21:22:28 +08:00
|
|
|
return;
|
2019-01-16 23:33:04 +08:00
|
|
|
|
|
|
|
if (reason)
|
|
|
|
dev_notice(i915->drm.dev, "Resetting chip for %s\n", reason);
|
|
|
|
error->reset_count++;
|
|
|
|
|
2019-01-25 21:22:28 +08:00
|
|
|
reset_prepare(i915);
|
2019-01-16 23:33:04 +08:00
|
|
|
|
|
|
|
if (!intel_has_gpu_reset(i915)) {
|
|
|
|
if (i915_modparams.reset)
|
|
|
|
dev_err(i915->drm.dev, "GPU reset not supported\n");
|
|
|
|
else
|
|
|
|
DRM_DEBUG_DRIVER("GPU reset disabled\n");
|
|
|
|
goto error;
|
|
|
|
}
|
|
|
|
|
2019-02-18 23:31:06 +08:00
|
|
|
if (INTEL_INFO(i915)->gpu_reset_clobbers_display)
|
|
|
|
intel_runtime_pm_disable_interrupts(i915);
|
|
|
|
|
2019-01-25 21:22:28 +08:00
|
|
|
if (do_reset(i915, stalled_mask)) {
|
2019-01-16 23:33:04 +08:00
|
|
|
dev_err(i915->drm.dev, "Failed to reset chip\n");
|
|
|
|
goto taint;
|
|
|
|
}
|
|
|
|
|
2019-02-18 23:31:06 +08:00
|
|
|
if (INTEL_INFO(i915)->gpu_reset_clobbers_display)
|
|
|
|
intel_runtime_pm_enable_interrupts(i915);
|
|
|
|
|
2019-01-16 23:33:04 +08:00
|
|
|
intel_overlay_reset(i915);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Next we need to restore the context, but we don't use those
|
|
|
|
* yet either...
|
|
|
|
*
|
|
|
|
* Ring buffer needs to be re-initialized in the KMS case, or if X
|
|
|
|
* was running at the time of the reset (i.e. we weren't VT
|
|
|
|
* switched away).
|
|
|
|
*/
|
|
|
|
ret = i915_gem_init_hw(i915);
|
|
|
|
if (ret) {
|
|
|
|
DRM_ERROR("Failed to initialise HW following reset (%d)\n",
|
|
|
|
ret);
|
|
|
|
goto error;
|
|
|
|
}
|
|
|
|
|
|
|
|
i915_queue_hangcheck(i915);
|
|
|
|
|
|
|
|
finish:
|
|
|
|
reset_finish(i915);
|
2019-02-20 22:56:37 +08:00
|
|
|
if (!__i915_wedged(error))
|
2019-01-25 21:22:28 +08:00
|
|
|
reset_restart(i915);
|
2019-01-16 23:33:04 +08:00
|
|
|
return;
|
|
|
|
|
|
|
|
taint:
|
|
|
|
/*
|
|
|
|
* History tells us that if we cannot reset the GPU now, we
|
|
|
|
* never will. This then impacts everything that is run
|
|
|
|
* subsequently. On failing the reset, we mark the driver
|
|
|
|
* as wedged, preventing further execution on the GPU.
|
|
|
|
* We also want to go one step further and add a taint to the
|
|
|
|
* kernel so that any subsequent faults can be traced back to
|
|
|
|
* this failure. This is important for CI, where if the
|
|
|
|
* GPU/driver fails we would like to reboot and restart testing
|
|
|
|
* rather than continue on into oblivion. For everyone else,
|
|
|
|
* the system should still plod along, but they have been warned!
|
|
|
|
*/
|
|
|
|
add_taint(TAINT_WARN, LOCKDEP_STILL_OK);
|
|
|
|
error:
|
2019-02-08 23:37:07 +08:00
|
|
|
__i915_gem_set_wedged(i915);
|
2019-01-16 23:33:04 +08:00
|
|
|
goto finish;
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline int intel_gt_reset_engine(struct drm_i915_private *i915,
|
|
|
|
struct intel_engine_cs *engine)
|
|
|
|
{
|
2019-03-06 02:03:30 +08:00
|
|
|
return intel_gpu_reset(i915, engine->mask);
|
2019-01-16 23:33:04 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
|
|
|
* i915_reset_engine - reset GPU engine to recover from a hang
|
|
|
|
* @engine: engine to reset
|
|
|
|
* @msg: reason for GPU reset; or NULL for no dev_notice()
|
|
|
|
*
|
|
|
|
* Reset a specific GPU engine. Useful if a hang is detected.
|
|
|
|
* Returns zero on successful reset or otherwise an error code.
|
|
|
|
*
|
|
|
|
* Procedure is:
|
|
|
|
* - identifies the request that caused the hang and it is dropped
|
|
|
|
* - reset engine (which will force the engine to idle)
|
|
|
|
* - re-init/configure engine
|
|
|
|
*/
|
|
|
|
int i915_reset_engine(struct intel_engine_cs *engine, const char *msg)
|
|
|
|
{
|
|
|
|
struct i915_gpu_error *error = &engine->i915->gpu_error;
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
GEM_TRACE("%s flags=%lx\n", engine->name, error->flags);
|
|
|
|
GEM_BUG_ON(!test_bit(I915_RESET_ENGINE + engine->id, &error->flags));
|
|
|
|
|
2019-01-25 21:22:28 +08:00
|
|
|
reset_prepare_engine(engine);
|
2019-01-16 23:33:04 +08:00
|
|
|
|
|
|
|
if (msg)
|
|
|
|
dev_notice(engine->i915->drm.dev,
|
|
|
|
"Resetting %s for %s\n", engine->name, msg);
|
|
|
|
error->reset_engine_count[engine->id]++;
|
|
|
|
|
|
|
|
if (!engine->i915->guc.execbuf_client)
|
|
|
|
ret = intel_gt_reset_engine(engine->i915, engine);
|
|
|
|
else
|
|
|
|
ret = intel_guc_reset_engine(&engine->i915->guc, engine);
|
|
|
|
if (ret) {
|
|
|
|
/* If we fail here, we expect to fallback to a global reset */
|
|
|
|
DRM_DEBUG_DRIVER("%sFailed to reset %s, ret=%d\n",
|
|
|
|
engine->i915->guc.execbuf_client ? "GuC " : "",
|
|
|
|
engine->name, ret);
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* The request that caused the hang is stuck on elsp, we know the
|
|
|
|
* active request and can drop it, adjust head to skip the offending
|
|
|
|
* request to resume executing remaining requests in the queue.
|
|
|
|
*/
|
2019-01-25 21:22:28 +08:00
|
|
|
intel_engine_reset(engine, true);
|
2019-01-16 23:33:04 +08:00
|
|
|
|
|
|
|
/*
|
|
|
|
* The engine and its registers (and workarounds in case of render)
|
|
|
|
* have been reset to their default values. Follow the init_ring
|
|
|
|
* process to program RING_MODE, HWSP and re-enable submission.
|
|
|
|
*/
|
|
|
|
ret = engine->init_hw(engine);
|
|
|
|
if (ret)
|
|
|
|
goto out;
|
|
|
|
|
|
|
|
out:
|
|
|
|
intel_engine_cancel_stop_cs(engine);
|
|
|
|
reset_finish_engine(engine);
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void i915_reset_device(struct drm_i915_private *i915,
|
|
|
|
u32 engine_mask,
|
|
|
|
const char *reason)
|
|
|
|
{
|
|
|
|
struct i915_gpu_error *error = &i915->gpu_error;
|
|
|
|
struct kobject *kobj = &i915->drm.primary->kdev->kobj;
|
|
|
|
char *error_event[] = { I915_ERROR_UEVENT "=1", NULL };
|
|
|
|
char *reset_event[] = { I915_RESET_UEVENT "=1", NULL };
|
|
|
|
char *reset_done_event[] = { I915_ERROR_UEVENT "=0", NULL };
|
|
|
|
struct i915_wedge_me w;
|
|
|
|
|
|
|
|
kobject_uevent_env(kobj, KOBJ_CHANGE, error_event);
|
|
|
|
|
|
|
|
DRM_DEBUG_DRIVER("resetting chip\n");
|
|
|
|
kobject_uevent_env(kobj, KOBJ_CHANGE, reset_event);
|
|
|
|
|
|
|
|
/* Use a watchdog to ensure that our reset completes */
|
|
|
|
i915_wedge_on_timeout(&w, i915, 5 * HZ) {
|
|
|
|
intel_prepare_reset(i915);
|
|
|
|
|
2019-02-11 21:50:39 +08:00
|
|
|
/* Flush everyone using a resource about to be clobbered */
|
2019-02-11 21:50:40 +08:00
|
|
|
synchronize_srcu_expedited(&error->reset_backoff_srcu);
|
2019-02-11 21:50:39 +08:00
|
|
|
|
2019-02-08 23:37:07 +08:00
|
|
|
mutex_lock(&error->wedge_mutex);
|
2019-01-25 21:22:28 +08:00
|
|
|
i915_reset(i915, engine_mask, reason);
|
2019-02-08 23:37:07 +08:00
|
|
|
mutex_unlock(&error->wedge_mutex);
|
2019-01-16 23:33:04 +08:00
|
|
|
|
|
|
|
intel_finish_reset(i915);
|
|
|
|
}
|
|
|
|
|
|
|
|
if (!test_bit(I915_WEDGED, &error->flags))
|
|
|
|
kobject_uevent_env(kobj, KOBJ_CHANGE, reset_done_event);
|
|
|
|
}
|
|
|
|
|
|
|
|
void i915_clear_error_registers(struct drm_i915_private *dev_priv)
|
|
|
|
{
|
|
|
|
u32 eir;
|
|
|
|
|
|
|
|
if (!IS_GEN(dev_priv, 2))
|
|
|
|
I915_WRITE(PGTBL_ER, I915_READ(PGTBL_ER));
|
|
|
|
|
|
|
|
if (INTEL_GEN(dev_priv) < 4)
|
|
|
|
I915_WRITE(IPEIR, I915_READ(IPEIR));
|
|
|
|
else
|
|
|
|
I915_WRITE(IPEIR_I965, I915_READ(IPEIR_I965));
|
|
|
|
|
|
|
|
I915_WRITE(EIR, I915_READ(EIR));
|
|
|
|
eir = I915_READ(EIR);
|
|
|
|
if (eir) {
|
|
|
|
/*
|
|
|
|
* some errors might have become stuck,
|
|
|
|
* mask them.
|
|
|
|
*/
|
|
|
|
DRM_DEBUG_DRIVER("EIR stuck: 0x%08x, masking\n", eir);
|
|
|
|
I915_WRITE(EMR, I915_READ(EMR) | eir);
|
|
|
|
I915_WRITE(IIR, I915_MASTER_ERROR_INTERRUPT);
|
|
|
|
}
|
|
|
|
|
|
|
|
if (INTEL_GEN(dev_priv) >= 8) {
|
|
|
|
I915_WRITE(GEN8_RING_FAULT_REG,
|
|
|
|
I915_READ(GEN8_RING_FAULT_REG) & ~RING_FAULT_VALID);
|
|
|
|
POSTING_READ(GEN8_RING_FAULT_REG);
|
|
|
|
} else if (INTEL_GEN(dev_priv) >= 6) {
|
|
|
|
struct intel_engine_cs *engine;
|
|
|
|
enum intel_engine_id id;
|
|
|
|
|
|
|
|
for_each_engine(engine, dev_priv, id) {
|
|
|
|
I915_WRITE(RING_FAULT_REG(engine),
|
|
|
|
I915_READ(RING_FAULT_REG(engine)) &
|
|
|
|
~RING_FAULT_VALID);
|
|
|
|
}
|
2019-03-06 02:03:30 +08:00
|
|
|
POSTING_READ(RING_FAULT_REG(dev_priv->engine[RCS0]));
|
2019-01-16 23:33:04 +08:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
|
|
|
* i915_handle_error - handle a gpu error
|
|
|
|
* @i915: i915 device private
|
|
|
|
* @engine_mask: mask representing engines that are hung
|
|
|
|
* @flags: control flags
|
|
|
|
* @fmt: Error message format string
|
|
|
|
*
|
|
|
|
* Do some basic checking of register state at error time and
|
|
|
|
* dump it to the syslog. Also call i915_capture_error_state() to make
|
|
|
|
* sure we get a record and make it available in debugfs. Fire a uevent
|
|
|
|
* so userspace knows something bad happened (should trigger collection
|
|
|
|
* of a ring dump etc.).
|
|
|
|
*/
|
|
|
|
void i915_handle_error(struct drm_i915_private *i915,
|
|
|
|
u32 engine_mask,
|
|
|
|
unsigned long flags,
|
|
|
|
const char *fmt, ...)
|
|
|
|
{
|
2019-02-08 23:37:07 +08:00
|
|
|
struct i915_gpu_error *error = &i915->gpu_error;
|
2019-01-16 23:33:04 +08:00
|
|
|
struct intel_engine_cs *engine;
|
|
|
|
intel_wakeref_t wakeref;
|
|
|
|
unsigned int tmp;
|
|
|
|
char error_msg[80];
|
|
|
|
char *msg = NULL;
|
|
|
|
|
|
|
|
if (fmt) {
|
|
|
|
va_list args;
|
|
|
|
|
|
|
|
va_start(args, fmt);
|
|
|
|
vscnprintf(error_msg, sizeof(error_msg), fmt, args);
|
|
|
|
va_end(args);
|
|
|
|
|
|
|
|
msg = error_msg;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* In most cases it's guaranteed that we get here with an RPM
|
|
|
|
* reference held, for example because there is a pending GPU
|
|
|
|
* request that won't finish until the reset is done. This
|
|
|
|
* isn't the case at least when we get here by doing a
|
|
|
|
* simulated reset via debugfs, so get an RPM reference.
|
|
|
|
*/
|
|
|
|
wakeref = intel_runtime_pm_get(i915);
|
|
|
|
|
2019-03-06 02:03:30 +08:00
|
|
|
engine_mask &= INTEL_INFO(i915)->engine_mask;
|
2019-01-16 23:33:04 +08:00
|
|
|
|
|
|
|
if (flags & I915_ERROR_CAPTURE) {
|
|
|
|
i915_capture_error_state(i915, engine_mask, msg);
|
|
|
|
i915_clear_error_registers(i915);
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Try engine reset when available. We fall back to full reset if
|
|
|
|
* single reset fails.
|
|
|
|
*/
|
2019-02-20 22:56:37 +08:00
|
|
|
if (intel_has_reset_engine(i915) && !__i915_wedged(error)) {
|
2019-01-16 23:33:04 +08:00
|
|
|
for_each_engine_masked(engine, i915, engine_mask, tmp) {
|
|
|
|
BUILD_BUG_ON(I915_RESET_MODESET >= I915_RESET_ENGINE);
|
|
|
|
if (test_and_set_bit(I915_RESET_ENGINE + engine->id,
|
2019-02-08 23:37:07 +08:00
|
|
|
&error->flags))
|
2019-01-16 23:33:04 +08:00
|
|
|
continue;
|
|
|
|
|
|
|
|
if (i915_reset_engine(engine, msg) == 0)
|
2019-03-06 02:03:30 +08:00
|
|
|
engine_mask &= ~engine->mask;
|
2019-01-16 23:33:04 +08:00
|
|
|
|
|
|
|
clear_bit(I915_RESET_ENGINE + engine->id,
|
2019-02-08 23:37:07 +08:00
|
|
|
&error->flags);
|
|
|
|
wake_up_bit(&error->flags,
|
2019-01-16 23:33:04 +08:00
|
|
|
I915_RESET_ENGINE + engine->id);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if (!engine_mask)
|
|
|
|
goto out;
|
|
|
|
|
|
|
|
/* Full reset needs the mutex, stop any other user trying to do so. */
|
2019-02-08 23:37:07 +08:00
|
|
|
if (test_and_set_bit(I915_RESET_BACKOFF, &error->flags)) {
|
|
|
|
wait_event(error->reset_queue,
|
|
|
|
!test_bit(I915_RESET_BACKOFF, &error->flags));
|
2019-02-08 23:37:03 +08:00
|
|
|
goto out; /* piggy-back on the other reset */
|
2019-01-16 23:33:04 +08:00
|
|
|
}
|
|
|
|
|
2019-02-08 23:37:03 +08:00
|
|
|
/* Make sure i915_reset_trylock() sees the I915_RESET_BACKOFF */
|
|
|
|
synchronize_rcu_expedited();
|
|
|
|
|
2019-01-16 23:33:04 +08:00
|
|
|
/* Prevent any other reset-engine attempt. */
|
|
|
|
for_each_engine(engine, i915, tmp) {
|
|
|
|
while (test_and_set_bit(I915_RESET_ENGINE + engine->id,
|
2019-02-08 23:37:07 +08:00
|
|
|
&error->flags))
|
|
|
|
wait_on_bit(&error->flags,
|
2019-01-16 23:33:04 +08:00
|
|
|
I915_RESET_ENGINE + engine->id,
|
|
|
|
TASK_UNINTERRUPTIBLE);
|
|
|
|
}
|
|
|
|
|
|
|
|
i915_reset_device(i915, engine_mask, msg);
|
|
|
|
|
|
|
|
for_each_engine(engine, i915, tmp) {
|
|
|
|
clear_bit(I915_RESET_ENGINE + engine->id,
|
2019-02-08 23:37:07 +08:00
|
|
|
&error->flags);
|
2019-01-16 23:33:04 +08:00
|
|
|
}
|
|
|
|
|
2019-02-08 23:37:07 +08:00
|
|
|
clear_bit(I915_RESET_BACKOFF, &error->flags);
|
|
|
|
wake_up_all(&error->reset_queue);
|
2019-01-16 23:33:04 +08:00
|
|
|
|
|
|
|
out:
|
|
|
|
intel_runtime_pm_put(i915, wakeref);
|
|
|
|
}
|
|
|
|
|
2019-02-08 23:37:03 +08:00
|
|
|
int i915_reset_trylock(struct drm_i915_private *i915)
|
|
|
|
{
|
|
|
|
struct i915_gpu_error *error = &i915->gpu_error;
|
|
|
|
int srcu;
|
|
|
|
|
2019-02-12 21:08:31 +08:00
|
|
|
might_lock(&error->reset_backoff_srcu);
|
|
|
|
might_sleep();
|
|
|
|
|
2019-02-08 23:37:03 +08:00
|
|
|
rcu_read_lock();
|
|
|
|
while (test_bit(I915_RESET_BACKOFF, &error->flags)) {
|
|
|
|
rcu_read_unlock();
|
|
|
|
|
|
|
|
if (wait_event_interruptible(error->reset_queue,
|
|
|
|
!test_bit(I915_RESET_BACKOFF,
|
|
|
|
&error->flags)))
|
|
|
|
return -EINTR;
|
|
|
|
|
|
|
|
rcu_read_lock();
|
|
|
|
}
|
|
|
|
srcu = srcu_read_lock(&error->reset_backoff_srcu);
|
|
|
|
rcu_read_unlock();
|
|
|
|
|
|
|
|
return srcu;
|
|
|
|
}
|
|
|
|
|
|
|
|
void i915_reset_unlock(struct drm_i915_private *i915, int tag)
|
|
|
|
__releases(&i915->gpu_error.reset_backoff_srcu)
|
|
|
|
{
|
|
|
|
struct i915_gpu_error *error = &i915->gpu_error;
|
|
|
|
|
|
|
|
srcu_read_unlock(&error->reset_backoff_srcu, tag);
|
|
|
|
}
|
|
|
|
|
2019-02-20 22:56:37 +08:00
|
|
|
int i915_terminally_wedged(struct drm_i915_private *i915)
|
|
|
|
{
|
|
|
|
struct i915_gpu_error *error = &i915->gpu_error;
|
|
|
|
|
|
|
|
might_sleep();
|
|
|
|
|
|
|
|
if (!__i915_wedged(error))
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
/* Reset still in progress? Maybe we will recover? */
|
|
|
|
if (!test_bit(I915_RESET_BACKOFF, &error->flags))
|
|
|
|
return -EIO;
|
|
|
|
|
|
|
|
/* XXX intel_reset_finish() still takes struct_mutex!!! */
|
|
|
|
if (mutex_is_locked(&i915->drm.struct_mutex))
|
|
|
|
return -EAGAIN;
|
|
|
|
|
|
|
|
if (wait_event_interruptible(error->reset_queue,
|
|
|
|
!test_bit(I915_RESET_BACKOFF,
|
|
|
|
&error->flags)))
|
|
|
|
return -EINTR;
|
|
|
|
|
|
|
|
return __i915_wedged(error) ? -EIO : 0;
|
|
|
|
}
|
|
|
|
|
2019-01-25 21:22:28 +08:00
|
|
|
bool i915_reset_flush(struct drm_i915_private *i915)
|
|
|
|
{
|
|
|
|
int err;
|
|
|
|
|
|
|
|
cancel_delayed_work_sync(&i915->gpu_error.hangcheck_work);
|
|
|
|
|
|
|
|
flush_workqueue(i915->wq);
|
|
|
|
GEM_BUG_ON(READ_ONCE(i915->gpu_error.restart));
|
|
|
|
|
|
|
|
mutex_lock(&i915->drm.struct_mutex);
|
|
|
|
err = i915_gem_wait_for_idle(i915,
|
|
|
|
I915_WAIT_LOCKED |
|
|
|
|
I915_WAIT_FOR_IDLE_BOOST,
|
|
|
|
MAX_SCHEDULE_TIMEOUT);
|
|
|
|
mutex_unlock(&i915->drm.struct_mutex);
|
|
|
|
|
|
|
|
return !err;
|
|
|
|
}
|
|
|
|
|
2019-01-16 23:33:04 +08:00
|
|
|
static void i915_wedge_me(struct work_struct *work)
|
|
|
|
{
|
|
|
|
struct i915_wedge_me *w = container_of(work, typeof(*w), work.work);
|
|
|
|
|
|
|
|
dev_err(w->i915->drm.dev,
|
|
|
|
"%s timed out, cancelling all in-flight rendering.\n",
|
|
|
|
w->name);
|
|
|
|
i915_gem_set_wedged(w->i915);
|
|
|
|
}
|
|
|
|
|
|
|
|
void __i915_init_wedge(struct i915_wedge_me *w,
|
|
|
|
struct drm_i915_private *i915,
|
|
|
|
long timeout,
|
|
|
|
const char *name)
|
|
|
|
{
|
|
|
|
w->i915 = i915;
|
|
|
|
w->name = name;
|
|
|
|
|
|
|
|
INIT_DELAYED_WORK_ONSTACK(&w->work, i915_wedge_me);
|
|
|
|
schedule_delayed_work(&w->work, timeout);
|
|
|
|
}
|
|
|
|
|
|
|
|
void __i915_fini_wedge(struct i915_wedge_me *w)
|
|
|
|
{
|
|
|
|
cancel_delayed_work_sync(&w->work);
|
|
|
|
destroy_delayed_work_on_stack(&w->work);
|
|
|
|
w->i915 = NULL;
|
|
|
|
}
|