2016-07-13 23:03:40 +08:00
|
|
|
/*
|
|
|
|
* Copyright © 2016 Intel Corporation
|
|
|
|
*
|
|
|
|
* Permission is hereby granted, free of charge, to any person obtaining a
|
|
|
|
* copy of this software and associated documentation files (the "Software"),
|
|
|
|
* to deal in the Software without restriction, including without limitation
|
|
|
|
* the rights to use, copy, modify, merge, publish, distribute, sublicense,
|
|
|
|
* and/or sell copies of the Software, and to permit persons to whom the
|
|
|
|
* Software is furnished to do so, subject to the following conditions:
|
|
|
|
*
|
|
|
|
* The above copyright notice and this permission notice (including the next
|
|
|
|
* paragraph) shall be included in all copies or substantial portions of the
|
|
|
|
* Software.
|
|
|
|
*
|
|
|
|
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
|
|
|
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
|
|
|
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
|
|
|
|
* THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
|
|
|
|
* LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
|
|
|
|
* FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS
|
|
|
|
* IN THE SOFTWARE.
|
|
|
|
*
|
|
|
|
*/
|
|
|
|
|
|
|
|
#include "i915_drv.h"
|
|
|
|
#include "intel_ringbuffer.h"
|
|
|
|
#include "intel_lrc.h"
|
|
|
|
|
|
|
|
static const struct engine_info {
|
|
|
|
const char *name;
|
|
|
|
unsigned exec_id;
|
2016-08-17 00:04:20 +08:00
|
|
|
enum intel_engine_hw_id hw_id;
|
2016-07-13 23:03:40 +08:00
|
|
|
u32 mmio_base;
|
|
|
|
unsigned irq_shift;
|
|
|
|
int (*init_legacy)(struct intel_engine_cs *engine);
|
|
|
|
int (*init_execlists)(struct intel_engine_cs *engine);
|
|
|
|
} intel_engines[] = {
|
|
|
|
[RCS] = {
|
|
|
|
.name = "render ring",
|
|
|
|
.exec_id = I915_EXEC_RENDER,
|
2016-08-17 00:04:20 +08:00
|
|
|
.hw_id = RCS_HW,
|
2016-07-13 23:03:40 +08:00
|
|
|
.mmio_base = RENDER_RING_BASE,
|
|
|
|
.irq_shift = GEN8_RCS_IRQ_SHIFT,
|
|
|
|
.init_execlists = logical_render_ring_init,
|
|
|
|
.init_legacy = intel_init_render_ring_buffer,
|
|
|
|
},
|
|
|
|
[BCS] = {
|
|
|
|
.name = "blitter ring",
|
|
|
|
.exec_id = I915_EXEC_BLT,
|
2016-08-17 00:04:20 +08:00
|
|
|
.hw_id = BCS_HW,
|
2016-07-13 23:03:40 +08:00
|
|
|
.mmio_base = BLT_RING_BASE,
|
|
|
|
.irq_shift = GEN8_BCS_IRQ_SHIFT,
|
|
|
|
.init_execlists = logical_xcs_ring_init,
|
|
|
|
.init_legacy = intel_init_blt_ring_buffer,
|
|
|
|
},
|
|
|
|
[VCS] = {
|
|
|
|
.name = "bsd ring",
|
|
|
|
.exec_id = I915_EXEC_BSD,
|
2016-08-17 00:04:20 +08:00
|
|
|
.hw_id = VCS_HW,
|
2016-07-13 23:03:40 +08:00
|
|
|
.mmio_base = GEN6_BSD_RING_BASE,
|
|
|
|
.irq_shift = GEN8_VCS1_IRQ_SHIFT,
|
|
|
|
.init_execlists = logical_xcs_ring_init,
|
|
|
|
.init_legacy = intel_init_bsd_ring_buffer,
|
|
|
|
},
|
|
|
|
[VCS2] = {
|
|
|
|
.name = "bsd2 ring",
|
|
|
|
.exec_id = I915_EXEC_BSD,
|
2016-08-17 00:04:20 +08:00
|
|
|
.hw_id = VCS2_HW,
|
2016-07-13 23:03:40 +08:00
|
|
|
.mmio_base = GEN8_BSD2_RING_BASE,
|
|
|
|
.irq_shift = GEN8_VCS2_IRQ_SHIFT,
|
|
|
|
.init_execlists = logical_xcs_ring_init,
|
|
|
|
.init_legacy = intel_init_bsd2_ring_buffer,
|
|
|
|
},
|
|
|
|
[VECS] = {
|
|
|
|
.name = "video enhancement ring",
|
|
|
|
.exec_id = I915_EXEC_VEBOX,
|
2016-08-17 00:04:20 +08:00
|
|
|
.hw_id = VECS_HW,
|
2016-07-13 23:03:40 +08:00
|
|
|
.mmio_base = VEBOX_RING_BASE,
|
|
|
|
.irq_shift = GEN8_VECS_IRQ_SHIFT,
|
|
|
|
.init_execlists = logical_xcs_ring_init,
|
|
|
|
.init_legacy = intel_init_vebox_ring_buffer,
|
|
|
|
},
|
|
|
|
};
|
|
|
|
|
|
|
|
static struct intel_engine_cs *
|
|
|
|
intel_engine_setup(struct drm_i915_private *dev_priv,
|
|
|
|
enum intel_engine_id id)
|
|
|
|
{
|
|
|
|
const struct engine_info *info = &intel_engines[id];
|
|
|
|
struct intel_engine_cs *engine = &dev_priv->engine[id];
|
|
|
|
|
|
|
|
engine->id = id;
|
|
|
|
engine->i915 = dev_priv;
|
|
|
|
engine->name = info->name;
|
|
|
|
engine->exec_id = info->exec_id;
|
2016-08-17 00:04:20 +08:00
|
|
|
engine->hw_id = engine->guc_id = info->hw_id;
|
2016-07-13 23:03:40 +08:00
|
|
|
engine->mmio_base = info->mmio_base;
|
|
|
|
engine->irq_shift = info->irq_shift;
|
|
|
|
|
|
|
|
return engine;
|
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
|
|
|
* intel_engines_init() - allocate, populate and init the Engine Command Streamers
|
|
|
|
* @dev: DRM device.
|
|
|
|
*
|
|
|
|
* Return: non-zero if the initialization failed.
|
|
|
|
*/
|
|
|
|
int intel_engines_init(struct drm_device *dev)
|
|
|
|
{
|
|
|
|
struct drm_i915_private *dev_priv = to_i915(dev);
|
2016-08-10 23:22:10 +08:00
|
|
|
struct intel_device_info *device_info = mkwrite_device_info(dev_priv);
|
2016-07-13 23:03:40 +08:00
|
|
|
unsigned int mask = 0;
|
|
|
|
int (*init)(struct intel_engine_cs *engine);
|
|
|
|
unsigned int i;
|
|
|
|
int ret;
|
|
|
|
|
2016-07-29 07:45:35 +08:00
|
|
|
WARN_ON(INTEL_INFO(dev_priv)->ring_mask == 0);
|
2016-07-13 23:03:40 +08:00
|
|
|
WARN_ON(INTEL_INFO(dev_priv)->ring_mask &
|
|
|
|
GENMASK(sizeof(mask) * BITS_PER_BYTE - 1, I915_NUM_ENGINES));
|
|
|
|
|
|
|
|
for (i = 0; i < ARRAY_SIZE(intel_engines); i++) {
|
|
|
|
if (!HAS_ENGINE(dev_priv, i))
|
|
|
|
continue;
|
|
|
|
|
|
|
|
if (i915.enable_execlists)
|
|
|
|
init = intel_engines[i].init_execlists;
|
|
|
|
else
|
|
|
|
init = intel_engines[i].init_legacy;
|
|
|
|
|
|
|
|
if (!init)
|
|
|
|
continue;
|
|
|
|
|
|
|
|
ret = init(intel_engine_setup(dev_priv, i));
|
|
|
|
if (ret)
|
|
|
|
goto cleanup;
|
|
|
|
|
|
|
|
mask |= ENGINE_MASK(i);
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Catch failures to update intel_engines table when the new engines
|
|
|
|
* are added to the driver by a warning and disabling the forgotten
|
|
|
|
* engines.
|
|
|
|
*/
|
2016-08-10 23:22:10 +08:00
|
|
|
if (WARN_ON(mask != INTEL_INFO(dev_priv)->ring_mask))
|
|
|
|
device_info->ring_mask = mask;
|
|
|
|
|
|
|
|
device_info->num_rings = hweight32(mask);
|
2016-07-13 23:03:40 +08:00
|
|
|
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
cleanup:
|
|
|
|
for (i = 0; i < I915_NUM_ENGINES; i++) {
|
|
|
|
if (i915.enable_execlists)
|
|
|
|
intel_logical_ring_cleanup(&dev_priv->engine[i]);
|
|
|
|
else
|
2016-08-03 05:50:21 +08:00
|
|
|
intel_engine_cleanup(&dev_priv->engine[i]);
|
2016-07-13 23:03:40 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
2016-08-15 17:49:00 +08:00
|
|
|
void intel_engine_init_seqno(struct intel_engine_cs *engine, u32 seqno)
|
|
|
|
{
|
|
|
|
struct drm_i915_private *dev_priv = engine->i915;
|
|
|
|
|
|
|
|
/* Our semaphore implementation is strictly monotonic (i.e. we proceed
|
|
|
|
* so long as the semaphore value in the register/page is greater
|
|
|
|
* than the sync value), so whenever we reset the seqno,
|
|
|
|
* so long as we reset the tracking semaphore value to 0, it will
|
|
|
|
* always be before the next request's seqno. If we don't reset
|
|
|
|
* the semaphore value, then when the seqno moves backwards all
|
|
|
|
* future waits will complete instantly (causing rendering corruption).
|
|
|
|
*/
|
|
|
|
if (IS_GEN6(dev_priv) || IS_GEN7(dev_priv)) {
|
|
|
|
I915_WRITE(RING_SYNC_0(engine->mmio_base), 0);
|
|
|
|
I915_WRITE(RING_SYNC_1(engine->mmio_base), 0);
|
|
|
|
if (HAS_VEBOX(dev_priv))
|
|
|
|
I915_WRITE(RING_SYNC_2(engine->mmio_base), 0);
|
|
|
|
}
|
2016-08-15 17:49:02 +08:00
|
|
|
if (dev_priv->semaphore) {
|
|
|
|
struct page *page = i915_vma_first_page(dev_priv->semaphore);
|
|
|
|
void *semaphores;
|
|
|
|
|
|
|
|
/* Semaphores are in noncoherent memory, flush to be safe */
|
|
|
|
semaphores = kmap(page);
|
2016-08-15 17:49:00 +08:00
|
|
|
memset(semaphores + GEN8_SEMAPHORE_OFFSET(engine->id, 0),
|
|
|
|
0, I915_NUM_ENGINES * gen8_semaphore_seqno_size);
|
2016-08-15 17:49:02 +08:00
|
|
|
drm_clflush_virt_range(semaphores + GEN8_SEMAPHORE_OFFSET(engine->id, 0),
|
|
|
|
I915_NUM_ENGINES * gen8_semaphore_seqno_size);
|
2016-08-15 17:49:00 +08:00
|
|
|
kunmap(page);
|
|
|
|
}
|
|
|
|
memset(engine->semaphore.sync_seqno, 0,
|
|
|
|
sizeof(engine->semaphore.sync_seqno));
|
|
|
|
|
|
|
|
intel_write_status_page(engine, I915_GEM_HWS_INDEX, seqno);
|
|
|
|
if (engine->irq_seqno_barrier)
|
|
|
|
engine->irq_seqno_barrier(engine);
|
|
|
|
engine->last_submitted_seqno = seqno;
|
|
|
|
|
|
|
|
engine->hangcheck.seqno = seqno;
|
|
|
|
|
|
|
|
/* After manually advancing the seqno, fake the interrupt in case
|
|
|
|
* there are any waiters for that seqno.
|
|
|
|
*/
|
|
|
|
intel_engine_wakeup(engine);
|
|
|
|
}
|
|
|
|
|
2016-07-13 23:03:41 +08:00
|
|
|
void intel_engine_init_hangcheck(struct intel_engine_cs *engine)
|
|
|
|
{
|
|
|
|
memset(&engine->hangcheck, 0, sizeof(engine->hangcheck));
|
|
|
|
}
|
|
|
|
|
2016-08-05 17:14:11 +08:00
|
|
|
static void intel_engine_init_requests(struct intel_engine_cs *engine)
|
|
|
|
{
|
|
|
|
init_request_active(&engine->last_request, NULL);
|
|
|
|
INIT_LIST_HEAD(&engine->request_list);
|
|
|
|
}
|
|
|
|
|
2016-07-13 23:03:41 +08:00
|
|
|
/**
|
|
|
|
* intel_engines_setup_common - setup engine state not requiring hw access
|
|
|
|
* @engine: Engine to setup.
|
|
|
|
*
|
|
|
|
* Initializes @engine@ structure members shared between legacy and execlists
|
|
|
|
* submission modes which do not require hardware access.
|
|
|
|
*
|
|
|
|
* Typically done early in the submission mode specific engine setup stage.
|
|
|
|
*/
|
|
|
|
void intel_engine_setup_common(struct intel_engine_cs *engine)
|
|
|
|
{
|
|
|
|
INIT_LIST_HEAD(&engine->execlist_queue);
|
|
|
|
spin_lock_init(&engine->execlist_lock);
|
|
|
|
|
2016-07-20 16:21:11 +08:00
|
|
|
engine->fence_context = fence_context_alloc(1);
|
|
|
|
|
2016-08-05 17:14:11 +08:00
|
|
|
intel_engine_init_requests(engine);
|
2016-07-13 23:03:41 +08:00
|
|
|
intel_engine_init_hangcheck(engine);
|
2016-08-04 23:32:19 +08:00
|
|
|
i915_gem_batch_pool_init(engine, &engine->batch_pool);
|
2016-08-19 00:17:10 +08:00
|
|
|
|
|
|
|
intel_engine_init_cmd_parser(engine);
|
2016-07-13 23:03:41 +08:00
|
|
|
}
|
|
|
|
|
2016-08-15 17:48:59 +08:00
|
|
|
int intel_engine_create_scratch(struct intel_engine_cs *engine, int size)
|
|
|
|
{
|
|
|
|
struct drm_i915_gem_object *obj;
|
|
|
|
struct i915_vma *vma;
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
WARN_ON(engine->scratch);
|
|
|
|
|
|
|
|
obj = i915_gem_object_create_stolen(&engine->i915->drm, size);
|
|
|
|
if (!obj)
|
|
|
|
obj = i915_gem_object_create(&engine->i915->drm, size);
|
|
|
|
if (IS_ERR(obj)) {
|
|
|
|
DRM_ERROR("Failed to allocate scratch page\n");
|
|
|
|
return PTR_ERR(obj);
|
|
|
|
}
|
|
|
|
|
|
|
|
vma = i915_vma_create(obj, &engine->i915->ggtt.base, NULL);
|
|
|
|
if (IS_ERR(vma)) {
|
|
|
|
ret = PTR_ERR(vma);
|
|
|
|
goto err_unref;
|
|
|
|
}
|
|
|
|
|
|
|
|
ret = i915_vma_pin(vma, 0, 4096, PIN_GLOBAL | PIN_HIGH);
|
|
|
|
if (ret)
|
|
|
|
goto err_unref;
|
|
|
|
|
|
|
|
engine->scratch = vma;
|
2016-08-15 17:49:07 +08:00
|
|
|
DRM_DEBUG_DRIVER("%s pipe control offset: 0x%08x\n",
|
|
|
|
engine->name, i915_ggtt_offset(vma));
|
2016-08-15 17:48:59 +08:00
|
|
|
return 0;
|
|
|
|
|
|
|
|
err_unref:
|
|
|
|
i915_gem_object_put(obj);
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void intel_engine_cleanup_scratch(struct intel_engine_cs *engine)
|
|
|
|
{
|
2016-08-15 17:49:05 +08:00
|
|
|
i915_vma_unpin_and_release(&engine->scratch);
|
2016-08-15 17:48:59 +08:00
|
|
|
}
|
|
|
|
|
2016-07-13 23:03:41 +08:00
|
|
|
/**
|
|
|
|
* intel_engines_init_common - initialize cengine state which might require hw access
|
|
|
|
* @engine: Engine to initialize.
|
|
|
|
*
|
|
|
|
* Initializes @engine@ structure members shared between legacy and execlists
|
|
|
|
* submission modes which do require hardware access.
|
|
|
|
*
|
|
|
|
* Typcally done at later stages of submission mode specific engine setup.
|
|
|
|
*
|
|
|
|
* Returns zero on success or an error code on failure.
|
|
|
|
*/
|
|
|
|
int intel_engine_init_common(struct intel_engine_cs *engine)
|
|
|
|
{
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
ret = intel_engine_init_breadcrumbs(engine);
|
|
|
|
if (ret)
|
|
|
|
return ret;
|
|
|
|
|
2016-08-19 00:17:10 +08:00
|
|
|
return 0;
|
2016-07-13 23:03:41 +08:00
|
|
|
}
|
2016-08-03 20:19:16 +08:00
|
|
|
|
|
|
|
/**
|
|
|
|
* intel_engines_cleanup_common - cleans up the engine state created by
|
|
|
|
* the common initiailizers.
|
|
|
|
* @engine: Engine to cleanup.
|
|
|
|
*
|
|
|
|
* This cleans up everything created by the common helpers.
|
|
|
|
*/
|
|
|
|
void intel_engine_cleanup_common(struct intel_engine_cs *engine)
|
|
|
|
{
|
2016-08-15 17:48:59 +08:00
|
|
|
intel_engine_cleanup_scratch(engine);
|
|
|
|
|
2016-08-03 20:19:16 +08:00
|
|
|
intel_engine_fini_breadcrumbs(engine);
|
2016-08-19 00:17:10 +08:00
|
|
|
intel_engine_cleanup_cmd_parser(engine);
|
2016-08-03 20:19:16 +08:00
|
|
|
i915_gem_batch_pool_fini(&engine->batch_pool);
|
|
|
|
}
|
2016-10-05 04:11:31 +08:00
|
|
|
|
|
|
|
u64 intel_engine_get_active_head(struct intel_engine_cs *engine)
|
|
|
|
{
|
|
|
|
struct drm_i915_private *dev_priv = engine->i915;
|
|
|
|
u64 acthd;
|
|
|
|
|
|
|
|
if (INTEL_GEN(dev_priv) >= 8)
|
|
|
|
acthd = I915_READ64_2x32(RING_ACTHD(engine->mmio_base),
|
|
|
|
RING_ACTHD_UDW(engine->mmio_base));
|
|
|
|
else if (INTEL_GEN(dev_priv) >= 4)
|
|
|
|
acthd = I915_READ(RING_ACTHD(engine->mmio_base));
|
|
|
|
else
|
|
|
|
acthd = I915_READ(ACTHD);
|
|
|
|
|
|
|
|
return acthd;
|
|
|
|
}
|
|
|
|
|
|
|
|
u64 intel_engine_get_last_batch_head(struct intel_engine_cs *engine)
|
|
|
|
{
|
|
|
|
struct drm_i915_private *dev_priv = engine->i915;
|
|
|
|
u64 bbaddr;
|
|
|
|
|
|
|
|
if (INTEL_GEN(dev_priv) >= 8)
|
|
|
|
bbaddr = I915_READ64_2x32(RING_BBADDR(engine->mmio_base),
|
|
|
|
RING_BBADDR_UDW(engine->mmio_base));
|
|
|
|
else
|
|
|
|
bbaddr = I915_READ(RING_BBADDR(engine->mmio_base));
|
|
|
|
|
|
|
|
return bbaddr;
|
|
|
|
}
|