mirror of https://gitee.com/openkylin/linux.git
drm/i915/cmdparser: Add support for backward jumps
To keep things manageable, the pre-gen9 cmdparser does not attempt to track any form of nested BB_START's. This did not prevent usermode from using nested starts, or even chained batches because the cmdparser is not strictly enforced pre gen9. Instead, the existence of a nested BB_START would cause the batch to be emitted in insecure mode, and any privileged capabilities would not be available. For Gen9, the cmdparser becomes mandatory (for BCS at least), and so not providing any form of nested BB_START support becomes overly restrictive. Any such batch will simply not run. We make heavy use of backward jumps in igt, and it is much easier to add support for this restricted subset of nested jumps, than to rewrite the whole of our test suite to avoid them. Add the required logic to support limited backward jumps, to instructions that have already been validated by the parser. Note that it's not sufficient to simply approve any BB_START that jumps backwards in the buffer because this would allow an attacker to embed a rogue instruction sequence within the operand words of a harmless instruction (say LRI) and jump to that. We introduce a bit array to track every instr offset successfully validated, and test the target of BB_START against this. If the target offset hits, it is re-written to the same offset in the shadow buffer and the BB_START cmd is allowed. Note: This patch deliberately ignores checkpatch issues in the cmdtables, in order to match the style of the surrounding code. We'll correct the entire file in one go in a later patch. v2: set dispatch secure late (Mika) v3: rebase (Mika) v4: Clear whitelist on each parse Minor review updates (Chris) v5: Correct backward jump batching v6: fix compilation error due to struct eb shuffle (Mika) Cc: Tony Luck <tony.luck@intel.com> Cc: Dave Airlie <airlied@redhat.com> Cc: Takashi Iwai <tiwai@suse.de> Cc: Tyler Hicks <tyhicks@canonical.com> Signed-off-by: Jon Bloomfield <jon.bloomfield@intel.com> Signed-off-by: Mika Kuoppala <mika.kuoppala@linux.intel.com> Reviewed-by: Chris Wilson <chris.p.wilson@intel.com>
This commit is contained in:
parent
0546a29cd8
commit
f8c08d8fae
|
@ -319,6 +319,8 @@ static void i915_gem_context_free(struct i915_gem_context *ctx)
|
||||||
free_engines(rcu_access_pointer(ctx->engines));
|
free_engines(rcu_access_pointer(ctx->engines));
|
||||||
mutex_destroy(&ctx->engines_mutex);
|
mutex_destroy(&ctx->engines_mutex);
|
||||||
|
|
||||||
|
kfree(ctx->jump_whitelist);
|
||||||
|
|
||||||
if (ctx->timeline)
|
if (ctx->timeline)
|
||||||
intel_timeline_put(ctx->timeline);
|
intel_timeline_put(ctx->timeline);
|
||||||
|
|
||||||
|
@ -441,6 +443,9 @@ __create_context(struct drm_i915_private *i915)
|
||||||
for (i = 0; i < ARRAY_SIZE(ctx->hang_timestamp); i++)
|
for (i = 0; i < ARRAY_SIZE(ctx->hang_timestamp); i++)
|
||||||
ctx->hang_timestamp[i] = jiffies - CONTEXT_FAST_HANG_JIFFIES;
|
ctx->hang_timestamp[i] = jiffies - CONTEXT_FAST_HANG_JIFFIES;
|
||||||
|
|
||||||
|
ctx->jump_whitelist = NULL;
|
||||||
|
ctx->jump_whitelist_cmds = 0;
|
||||||
|
|
||||||
return ctx;
|
return ctx;
|
||||||
|
|
||||||
err_free:
|
err_free:
|
||||||
|
|
|
@ -192,6 +192,13 @@ struct i915_gem_context {
|
||||||
* per vm, which may be one per context or shared with the global GTT)
|
* per vm, which may be one per context or shared with the global GTT)
|
||||||
*/
|
*/
|
||||||
struct radix_tree_root handles_vma;
|
struct radix_tree_root handles_vma;
|
||||||
|
|
||||||
|
/** jump_whitelist: Bit array for tracking cmds during cmdparsing
|
||||||
|
* Guarded by struct_mutex
|
||||||
|
*/
|
||||||
|
unsigned long *jump_whitelist;
|
||||||
|
/** jump_whitelist_cmds: No of cmd slots available */
|
||||||
|
u32 jump_whitelist_cmds;
|
||||||
};
|
};
|
||||||
|
|
||||||
#endif /* __I915_GEM_CONTEXT_TYPES_H__ */
|
#endif /* __I915_GEM_CONTEXT_TYPES_H__ */
|
||||||
|
|
|
@ -1972,7 +1972,6 @@ shadow_batch_pin(struct i915_execbuffer *eb, struct drm_i915_gem_object *obj)
|
||||||
if (CMDPARSER_USES_GGTT(dev_priv)) {
|
if (CMDPARSER_USES_GGTT(dev_priv)) {
|
||||||
flags = PIN_GLOBAL;
|
flags = PIN_GLOBAL;
|
||||||
vm = &dev_priv->ggtt.vm;
|
vm = &dev_priv->ggtt.vm;
|
||||||
eb->batch_flags |= I915_DISPATCH_SECURE;
|
|
||||||
} else if (vma->vm->has_read_only) {
|
} else if (vma->vm->has_read_only) {
|
||||||
flags = PIN_USER;
|
flags = PIN_USER;
|
||||||
vm = vma->vm;
|
vm = vma->vm;
|
||||||
|
@ -1989,18 +1988,35 @@ static struct i915_vma *eb_parse(struct i915_execbuffer *eb)
|
||||||
{
|
{
|
||||||
struct intel_engine_pool_node *pool;
|
struct intel_engine_pool_node *pool;
|
||||||
struct i915_vma *vma;
|
struct i915_vma *vma;
|
||||||
|
u64 batch_start;
|
||||||
|
u64 shadow_batch_start;
|
||||||
int err;
|
int err;
|
||||||
|
|
||||||
pool = intel_engine_pool_get(&eb->engine->pool, eb->batch_len);
|
pool = intel_engine_pool_get(&eb->engine->pool, eb->batch_len);
|
||||||
if (IS_ERR(pool))
|
if (IS_ERR(pool))
|
||||||
return ERR_CAST(pool);
|
return ERR_CAST(pool);
|
||||||
|
|
||||||
err = intel_engine_cmd_parser(eb->engine,
|
vma = shadow_batch_pin(eb, pool->obj);
|
||||||
|
if (IS_ERR(vma))
|
||||||
|
goto err;
|
||||||
|
|
||||||
|
batch_start = gen8_canonical_addr(eb->batch->node.start) +
|
||||||
|
eb->batch_start_offset;
|
||||||
|
|
||||||
|
shadow_batch_start = gen8_canonical_addr(vma->node.start);
|
||||||
|
|
||||||
|
err = intel_engine_cmd_parser(eb->gem_context,
|
||||||
|
eb->engine,
|
||||||
eb->batch->obj,
|
eb->batch->obj,
|
||||||
pool->obj,
|
batch_start,
|
||||||
eb->batch_start_offset,
|
eb->batch_start_offset,
|
||||||
eb->batch_len);
|
eb->batch_len,
|
||||||
|
pool->obj,
|
||||||
|
shadow_batch_start);
|
||||||
|
|
||||||
if (err) {
|
if (err) {
|
||||||
|
i915_vma_unpin(vma);
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Unsafe GGTT-backed buffers can still be submitted safely
|
* Unsafe GGTT-backed buffers can still be submitted safely
|
||||||
* as non-secure.
|
* as non-secure.
|
||||||
|
@ -2015,10 +2031,6 @@ static struct i915_vma *eb_parse(struct i915_execbuffer *eb)
|
||||||
goto err;
|
goto err;
|
||||||
}
|
}
|
||||||
|
|
||||||
vma = shadow_batch_pin(eb, pool->obj);
|
|
||||||
if (IS_ERR(vma))
|
|
||||||
goto err;
|
|
||||||
|
|
||||||
eb->vma[eb->buffer_count] = i915_vma_get(vma);
|
eb->vma[eb->buffer_count] = i915_vma_get(vma);
|
||||||
eb->flags[eb->buffer_count] =
|
eb->flags[eb->buffer_count] =
|
||||||
__EXEC_OBJECT_HAS_PIN | __EXEC_OBJECT_HAS_REF;
|
__EXEC_OBJECT_HAS_PIN | __EXEC_OBJECT_HAS_REF;
|
||||||
|
@ -2027,6 +2039,10 @@ static struct i915_vma *eb_parse(struct i915_execbuffer *eb)
|
||||||
|
|
||||||
eb->batch_start_offset = 0;
|
eb->batch_start_offset = 0;
|
||||||
eb->batch = vma;
|
eb->batch = vma;
|
||||||
|
|
||||||
|
if (CMDPARSER_USES_GGTT(eb->i915))
|
||||||
|
eb->batch_flags |= I915_DISPATCH_SECURE;
|
||||||
|
|
||||||
/* eb->batch_len unchanged */
|
/* eb->batch_len unchanged */
|
||||||
|
|
||||||
vma->private = pool;
|
vma->private = pool;
|
||||||
|
|
|
@ -483,6 +483,19 @@ static const struct drm_i915_cmd_descriptor gen9_blt_cmds[] = {
|
||||||
.reg = { .offset = 1, .mask = 0x007FFFFC } ),
|
.reg = { .offset = 1, .mask = 0x007FFFFC } ),
|
||||||
CMD( MI_LOAD_REGISTER_REG, SMI, !F, 0xFF, W,
|
CMD( MI_LOAD_REGISTER_REG, SMI, !F, 0xFF, W,
|
||||||
.reg = { .offset = 1, .mask = 0x007FFFFC, .step = 1 } ),
|
.reg = { .offset = 1, .mask = 0x007FFFFC, .step = 1 } ),
|
||||||
|
|
||||||
|
/*
|
||||||
|
* We allow BB_START but apply further checks. We just sanitize the
|
||||||
|
* basic fields here.
|
||||||
|
*/
|
||||||
|
#define MI_BB_START_OPERAND_MASK GENMASK(SMI-1, 0)
|
||||||
|
#define MI_BB_START_OPERAND_EXPECT (MI_BATCH_PPGTT_HSW | 1)
|
||||||
|
CMD( MI_BATCH_BUFFER_START_GEN8, SMI, !F, 0xFF, B,
|
||||||
|
.bits = {{
|
||||||
|
.offset = 0,
|
||||||
|
.mask = MI_BB_START_OPERAND_MASK,
|
||||||
|
.expected = MI_BB_START_OPERAND_EXPECT,
|
||||||
|
}}, ),
|
||||||
};
|
};
|
||||||
|
|
||||||
static const struct drm_i915_cmd_descriptor noop_desc =
|
static const struct drm_i915_cmd_descriptor noop_desc =
|
||||||
|
@ -1293,15 +1306,113 @@ static bool check_cmd(const struct intel_engine_cs *engine,
|
||||||
return true;
|
return true;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
static int check_bbstart(const struct i915_gem_context *ctx,
|
||||||
|
u32 *cmd, u32 offset, u32 length,
|
||||||
|
u32 batch_len,
|
||||||
|
u64 batch_start,
|
||||||
|
u64 shadow_batch_start)
|
||||||
|
{
|
||||||
|
u64 jump_offset, jump_target;
|
||||||
|
u32 target_cmd_offset, target_cmd_index;
|
||||||
|
|
||||||
|
/* For igt compatibility on older platforms */
|
||||||
|
if (CMDPARSER_USES_GGTT(ctx->i915)) {
|
||||||
|
DRM_DEBUG("CMD: Rejecting BB_START for ggtt based submission\n");
|
||||||
|
return -EACCES;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (length != 3) {
|
||||||
|
DRM_DEBUG("CMD: Recursive BB_START with bad length(%u)\n",
|
||||||
|
length);
|
||||||
|
return -EINVAL;
|
||||||
|
}
|
||||||
|
|
||||||
|
jump_target = *(u64*)(cmd+1);
|
||||||
|
jump_offset = jump_target - batch_start;
|
||||||
|
|
||||||
|
/*
|
||||||
|
* Any underflow of jump_target is guaranteed to be outside the range
|
||||||
|
* of a u32, so >= test catches both too large and too small
|
||||||
|
*/
|
||||||
|
if (jump_offset >= batch_len) {
|
||||||
|
DRM_DEBUG("CMD: BB_START to 0x%llx jumps out of BB\n",
|
||||||
|
jump_target);
|
||||||
|
return -EINVAL;
|
||||||
|
}
|
||||||
|
|
||||||
|
/*
|
||||||
|
* This cannot overflow a u32 because we already checked jump_offset
|
||||||
|
* is within the BB, and the batch_len is a u32
|
||||||
|
*/
|
||||||
|
target_cmd_offset = lower_32_bits(jump_offset);
|
||||||
|
target_cmd_index = target_cmd_offset / sizeof(u32);
|
||||||
|
|
||||||
|
*(u64*)(cmd + 1) = shadow_batch_start + target_cmd_offset;
|
||||||
|
|
||||||
|
if (target_cmd_index == offset)
|
||||||
|
return 0;
|
||||||
|
|
||||||
|
if (ctx->jump_whitelist_cmds <= target_cmd_index) {
|
||||||
|
DRM_DEBUG("CMD: Rejecting BB_START - truncated whitelist array\n");
|
||||||
|
return -EINVAL;
|
||||||
|
} else if (!test_bit(target_cmd_index, ctx->jump_whitelist)) {
|
||||||
|
DRM_DEBUG("CMD: BB_START to 0x%llx not a previously executed cmd\n",
|
||||||
|
jump_target);
|
||||||
|
return -EINVAL;
|
||||||
|
}
|
||||||
|
|
||||||
|
return 0;
|
||||||
|
}
|
||||||
|
|
||||||
|
static void init_whitelist(struct i915_gem_context *ctx, u32 batch_len)
|
||||||
|
{
|
||||||
|
const u32 batch_cmds = DIV_ROUND_UP(batch_len, sizeof(u32));
|
||||||
|
const u32 exact_size = BITS_TO_LONGS(batch_cmds);
|
||||||
|
u32 next_size = BITS_TO_LONGS(roundup_pow_of_two(batch_cmds));
|
||||||
|
unsigned long *next_whitelist;
|
||||||
|
|
||||||
|
if (CMDPARSER_USES_GGTT(ctx->i915))
|
||||||
|
return;
|
||||||
|
|
||||||
|
if (batch_cmds <= ctx->jump_whitelist_cmds) {
|
||||||
|
memset(ctx->jump_whitelist, 0, exact_size * sizeof(u32));
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
|
||||||
|
again:
|
||||||
|
next_whitelist = kcalloc(next_size, sizeof(long), GFP_KERNEL);
|
||||||
|
if (next_whitelist) {
|
||||||
|
kfree(ctx->jump_whitelist);
|
||||||
|
ctx->jump_whitelist = next_whitelist;
|
||||||
|
ctx->jump_whitelist_cmds =
|
||||||
|
next_size * BITS_PER_BYTE * sizeof(long);
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (next_size > exact_size) {
|
||||||
|
next_size = exact_size;
|
||||||
|
goto again;
|
||||||
|
}
|
||||||
|
|
||||||
|
DRM_DEBUG("CMD: Failed to extend whitelist. BB_START may be disallowed\n");
|
||||||
|
memset(ctx->jump_whitelist, 0,
|
||||||
|
BITS_TO_LONGS(ctx->jump_whitelist_cmds) * sizeof(u32));
|
||||||
|
|
||||||
|
return;
|
||||||
|
}
|
||||||
|
|
||||||
#define LENGTH_BIAS 2
|
#define LENGTH_BIAS 2
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* i915_parse_cmds() - parse a submitted batch buffer for privilege violations
|
* i915_parse_cmds() - parse a submitted batch buffer for privilege violations
|
||||||
|
* @ctx: the context in which the batch is to execute
|
||||||
* @engine: the engine on which the batch is to execute
|
* @engine: the engine on which the batch is to execute
|
||||||
* @batch_obj: the batch buffer in question
|
* @batch_obj: the batch buffer in question
|
||||||
* @shadow_batch_obj: copy of the batch buffer in question
|
* @batch_start: Canonical base address of batch
|
||||||
* @batch_start_offset: byte offset in the batch at which execution starts
|
* @batch_start_offset: byte offset in the batch at which execution starts
|
||||||
* @batch_len: length of the commands in batch_obj
|
* @batch_len: length of the commands in batch_obj
|
||||||
|
* @shadow_batch_obj: copy of the batch buffer in question
|
||||||
|
* @shadow_batch_start: Canonical base address of shadow_batch_obj
|
||||||
*
|
*
|
||||||
* Parses the specified batch buffer looking for privilege violations as
|
* Parses the specified batch buffer looking for privilege violations as
|
||||||
* described in the overview.
|
* described in the overview.
|
||||||
|
@ -1309,13 +1420,17 @@ static bool check_cmd(const struct intel_engine_cs *engine,
|
||||||
* Return: non-zero if the parser finds violations or otherwise fails; -EACCES
|
* Return: non-zero if the parser finds violations or otherwise fails; -EACCES
|
||||||
* if the batch appears legal but should use hardware parsing
|
* if the batch appears legal but should use hardware parsing
|
||||||
*/
|
*/
|
||||||
int intel_engine_cmd_parser(struct intel_engine_cs *engine,
|
|
||||||
|
int intel_engine_cmd_parser(struct i915_gem_context *ctx,
|
||||||
|
struct intel_engine_cs *engine,
|
||||||
struct drm_i915_gem_object *batch_obj,
|
struct drm_i915_gem_object *batch_obj,
|
||||||
struct drm_i915_gem_object *shadow_batch_obj,
|
u64 batch_start,
|
||||||
u32 batch_start_offset,
|
u32 batch_start_offset,
|
||||||
u32 batch_len)
|
u32 batch_len,
|
||||||
|
struct drm_i915_gem_object *shadow_batch_obj,
|
||||||
|
u64 shadow_batch_start)
|
||||||
{
|
{
|
||||||
u32 *cmd, *batch_end;
|
u32 *cmd, *batch_end, offset = 0;
|
||||||
struct drm_i915_cmd_descriptor default_desc = noop_desc;
|
struct drm_i915_cmd_descriptor default_desc = noop_desc;
|
||||||
const struct drm_i915_cmd_descriptor *desc = &default_desc;
|
const struct drm_i915_cmd_descriptor *desc = &default_desc;
|
||||||
bool needs_clflush_after = false;
|
bool needs_clflush_after = false;
|
||||||
|
@ -1329,6 +1444,8 @@ int intel_engine_cmd_parser(struct intel_engine_cs *engine,
|
||||||
return PTR_ERR(cmd);
|
return PTR_ERR(cmd);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
init_whitelist(ctx, batch_len);
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* We use the batch length as size because the shadow object is as
|
* We use the batch length as size because the shadow object is as
|
||||||
* large or larger and copy_batch() will write MI_NOPs to the extra
|
* large or larger and copy_batch() will write MI_NOPs to the extra
|
||||||
|
@ -1349,16 +1466,6 @@ int intel_engine_cmd_parser(struct intel_engine_cs *engine,
|
||||||
goto err;
|
goto err;
|
||||||
}
|
}
|
||||||
|
|
||||||
/*
|
|
||||||
* We don't try to handle BATCH_BUFFER_START because it adds
|
|
||||||
* non-trivial complexity. Instead we abort the scan and return
|
|
||||||
* and error to indicate that the batch is unsafe.
|
|
||||||
*/
|
|
||||||
if (desc->cmd.value == MI_BATCH_BUFFER_START) {
|
|
||||||
ret = -EACCES;
|
|
||||||
goto err;
|
|
||||||
}
|
|
||||||
|
|
||||||
if (desc->flags & CMD_DESC_FIXED)
|
if (desc->flags & CMD_DESC_FIXED)
|
||||||
length = desc->length.fixed;
|
length = desc->length.fixed;
|
||||||
else
|
else
|
||||||
|
@ -1378,7 +1485,21 @@ int intel_engine_cmd_parser(struct intel_engine_cs *engine,
|
||||||
goto err;
|
goto err;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
if (desc->cmd.value == MI_BATCH_BUFFER_START) {
|
||||||
|
ret = check_bbstart(ctx, cmd, offset, length,
|
||||||
|
batch_len, batch_start,
|
||||||
|
shadow_batch_start);
|
||||||
|
|
||||||
|
if (ret)
|
||||||
|
goto err;
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
|
||||||
|
if (ctx->jump_whitelist_cmds > offset)
|
||||||
|
set_bit(offset, ctx->jump_whitelist);
|
||||||
|
|
||||||
cmd += length;
|
cmd += length;
|
||||||
|
offset += length;
|
||||||
if (cmd >= batch_end) {
|
if (cmd >= batch_end) {
|
||||||
DRM_DEBUG_DRIVER("CMD: Got to the end of the buffer w/o a BBE cmd!\n");
|
DRM_DEBUG_DRIVER("CMD: Got to the end of the buffer w/o a BBE cmd!\n");
|
||||||
ret = -EINVAL;
|
ret = -EINVAL;
|
||||||
|
|
|
@ -2408,11 +2408,14 @@ const char *i915_cache_level_str(struct drm_i915_private *i915, int type);
|
||||||
int i915_cmd_parser_get_version(struct drm_i915_private *dev_priv);
|
int i915_cmd_parser_get_version(struct drm_i915_private *dev_priv);
|
||||||
void intel_engine_init_cmd_parser(struct intel_engine_cs *engine);
|
void intel_engine_init_cmd_parser(struct intel_engine_cs *engine);
|
||||||
void intel_engine_cleanup_cmd_parser(struct intel_engine_cs *engine);
|
void intel_engine_cleanup_cmd_parser(struct intel_engine_cs *engine);
|
||||||
int intel_engine_cmd_parser(struct intel_engine_cs *engine,
|
int intel_engine_cmd_parser(struct i915_gem_context *cxt,
|
||||||
|
struct intel_engine_cs *engine,
|
||||||
struct drm_i915_gem_object *batch_obj,
|
struct drm_i915_gem_object *batch_obj,
|
||||||
struct drm_i915_gem_object *shadow_batch_obj,
|
u64 user_batch_start,
|
||||||
u32 batch_start_offset,
|
u32 batch_start_offset,
|
||||||
u32 batch_len);
|
u32 batch_len,
|
||||||
|
struct drm_i915_gem_object *shadow_batch_obj,
|
||||||
|
u64 shadow_batch_start);
|
||||||
|
|
||||||
/* intel_device_info.c */
|
/* intel_device_info.c */
|
||||||
static inline struct intel_device_info *
|
static inline struct intel_device_info *
|
||||||
|
|
Loading…
Reference in New Issue