2013-07-12 21:50:57 +08:00
|
|
|
/*
|
|
|
|
* Copyright (c) 2008 Intel Corporation
|
|
|
|
*
|
|
|
|
* Permission is hereby granted, free of charge, to any person obtaining a
|
|
|
|
* copy of this software and associated documentation files (the "Software"),
|
|
|
|
* to deal in the Software without restriction, including without limitation
|
|
|
|
* the rights to use, copy, modify, merge, publish, distribute, sublicense,
|
|
|
|
* and/or sell copies of the Software, and to permit persons to whom the
|
|
|
|
* Software is furnished to do so, subject to the following conditions:
|
|
|
|
*
|
|
|
|
* The above copyright notice and this permission notice (including the next
|
|
|
|
* paragraph) shall be included in all copies or substantial portions of the
|
|
|
|
* Software.
|
|
|
|
*
|
|
|
|
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR
|
|
|
|
* IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY,
|
|
|
|
* FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL
|
|
|
|
* THE AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER
|
|
|
|
* LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING
|
|
|
|
* FROM, OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS
|
|
|
|
* IN THE SOFTWARE.
|
|
|
|
*
|
|
|
|
* Authors:
|
|
|
|
* Eric Anholt <eric@anholt.net>
|
|
|
|
* Keith Packard <keithp@keithp.com>
|
|
|
|
* Mika Kuoppala <mika.kuoppala@intel.com>
|
|
|
|
*
|
|
|
|
*/
|
|
|
|
|
2018-11-23 21:23:25 +08:00
|
|
|
#include <linux/ascii85.h>
|
|
|
|
#include <linux/nmi.h>
|
|
|
|
#include <linux/scatterlist.h>
|
2016-10-12 17:05:19 +08:00
|
|
|
#include <linux/stop_machine.h>
|
2018-11-23 21:23:25 +08:00
|
|
|
#include <linux/utsname.h>
|
2016-10-12 17:05:22 +08:00
|
|
|
#include <linux/zlib.h>
|
2018-11-23 21:23:25 +08:00
|
|
|
|
2017-10-27 01:36:55 +08:00
|
|
|
#include <drm/drm_print.h>
|
|
|
|
|
2018-03-08 17:50:37 +08:00
|
|
|
#include "i915_gpu_error.h"
|
2013-07-12 21:50:57 +08:00
|
|
|
#include "i915_drv.h"
|
|
|
|
|
2018-01-19 01:52:28 +08:00
|
|
|
static inline const struct intel_engine_cs *
|
|
|
|
engine_lookup(const struct drm_i915_private *i915, unsigned int id)
|
|
|
|
{
|
|
|
|
if (id >= I915_NUM_ENGINES)
|
|
|
|
return NULL;
|
|
|
|
|
|
|
|
return i915->engine[id];
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline const char *
|
|
|
|
__engine_name(const struct intel_engine_cs *engine)
|
|
|
|
{
|
|
|
|
return engine ? engine->name : "";
|
|
|
|
}
|
|
|
|
|
|
|
|
static const char *
|
|
|
|
engine_name(const struct drm_i915_private *i915, unsigned int id)
|
|
|
|
{
|
|
|
|
return __engine_name(engine_lookup(i915, id));
|
2013-07-12 21:50:57 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
static const char *tiling_flag(int tiling)
|
|
|
|
{
|
|
|
|
switch (tiling) {
|
|
|
|
default:
|
|
|
|
case I915_TILING_NONE: return "";
|
|
|
|
case I915_TILING_X: return " X";
|
|
|
|
case I915_TILING_Y: return " Y";
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
static const char *dirty_flag(int dirty)
|
|
|
|
{
|
|
|
|
return dirty ? " dirty" : "";
|
|
|
|
}
|
|
|
|
|
|
|
|
static const char *purgeable_flag(int purgeable)
|
|
|
|
{
|
|
|
|
return purgeable ? " purgeable" : "";
|
|
|
|
}
|
|
|
|
|
2018-11-23 21:23:25 +08:00
|
|
|
static void __sg_set_buf(struct scatterlist *sg,
|
|
|
|
void *addr, unsigned int len, loff_t it)
|
2013-07-12 21:50:57 +08:00
|
|
|
{
|
2018-11-23 21:23:25 +08:00
|
|
|
sg->page_link = (unsigned long)virt_to_page(addr);
|
|
|
|
sg->offset = offset_in_page(addr);
|
|
|
|
sg->length = len;
|
|
|
|
sg->dma_address = it;
|
2013-07-12 21:50:57 +08:00
|
|
|
}
|
|
|
|
|
2018-11-23 21:23:25 +08:00
|
|
|
static bool __i915_error_grow(struct drm_i915_error_state_buf *e, size_t len)
|
2013-07-12 21:50:57 +08:00
|
|
|
{
|
2018-11-23 21:23:25 +08:00
|
|
|
if (!len)
|
2013-07-12 21:50:57 +08:00
|
|
|
return false;
|
|
|
|
|
2018-11-23 21:23:25 +08:00
|
|
|
if (e->bytes + len + 1 <= e->size)
|
|
|
|
return true;
|
|
|
|
|
|
|
|
if (e->bytes) {
|
|
|
|
__sg_set_buf(e->cur++, e->buf, e->bytes, e->iter);
|
|
|
|
e->iter += e->bytes;
|
|
|
|
e->buf = NULL;
|
|
|
|
e->bytes = 0;
|
2013-07-12 21:50:57 +08:00
|
|
|
}
|
|
|
|
|
2018-11-23 21:23:25 +08:00
|
|
|
if (e->cur == e->end) {
|
|
|
|
struct scatterlist *sgl;
|
2013-07-12 21:50:57 +08:00
|
|
|
|
2018-11-23 21:23:25 +08:00
|
|
|
sgl = (typeof(sgl))__get_free_page(GFP_KERNEL);
|
|
|
|
if (!sgl) {
|
|
|
|
e->err = -ENOMEM;
|
|
|
|
return false;
|
|
|
|
}
|
2013-07-12 21:50:57 +08:00
|
|
|
|
2018-11-23 21:23:25 +08:00
|
|
|
if (e->cur) {
|
|
|
|
e->cur->offset = 0;
|
|
|
|
e->cur->length = 0;
|
|
|
|
e->cur->page_link =
|
|
|
|
(unsigned long)sgl | SG_CHAIN;
|
|
|
|
} else {
|
|
|
|
e->sgl = sgl;
|
2013-07-12 21:50:57 +08:00
|
|
|
}
|
|
|
|
|
2018-11-23 21:23:25 +08:00
|
|
|
e->cur = sgl;
|
|
|
|
e->end = sgl + SG_MAX_SINGLE_ALLOC - 1;
|
2013-07-12 21:50:57 +08:00
|
|
|
}
|
|
|
|
|
2018-11-23 21:23:25 +08:00
|
|
|
e->size = ALIGN(len + 1, SZ_64K);
|
|
|
|
e->buf = kmalloc(e->size, GFP_KERNEL | __GFP_NOWARN | __GFP_NORETRY);
|
|
|
|
if (!e->buf) {
|
|
|
|
e->size = PAGE_ALIGN(len + 1);
|
|
|
|
e->buf = kmalloc(e->size, GFP_KERNEL);
|
|
|
|
}
|
|
|
|
if (!e->buf) {
|
|
|
|
e->err = -ENOMEM;
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
|
|
|
return true;
|
2013-07-12 21:50:57 +08:00
|
|
|
}
|
|
|
|
|
2017-01-14 18:51:12 +08:00
|
|
|
__printf(2, 0)
|
2013-07-12 21:50:57 +08:00
|
|
|
static void i915_error_vprintf(struct drm_i915_error_state_buf *e,
|
2018-11-23 21:23:25 +08:00
|
|
|
const char *fmt, va_list args)
|
2013-07-12 21:50:57 +08:00
|
|
|
{
|
2018-11-23 21:23:25 +08:00
|
|
|
va_list ap;
|
|
|
|
int len;
|
2013-07-12 21:50:57 +08:00
|
|
|
|
2018-11-23 21:23:25 +08:00
|
|
|
if (e->err)
|
2013-07-12 21:50:57 +08:00
|
|
|
return;
|
|
|
|
|
2018-11-23 21:23:25 +08:00
|
|
|
va_copy(ap, args);
|
|
|
|
len = vsnprintf(NULL, 0, fmt, ap);
|
|
|
|
va_end(ap);
|
|
|
|
if (len <= 0) {
|
|
|
|
e->err = len;
|
|
|
|
return;
|
2013-07-12 21:50:57 +08:00
|
|
|
}
|
|
|
|
|
2018-11-23 21:23:25 +08:00
|
|
|
if (!__i915_error_grow(e, len))
|
|
|
|
return;
|
2013-07-12 21:50:57 +08:00
|
|
|
|
2018-11-23 21:23:25 +08:00
|
|
|
GEM_BUG_ON(e->bytes >= e->size);
|
|
|
|
len = vscnprintf(e->buf + e->bytes, e->size - e->bytes, fmt, args);
|
|
|
|
if (len < 0) {
|
|
|
|
e->err = len;
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
e->bytes += len;
|
2013-07-12 21:50:57 +08:00
|
|
|
}
|
|
|
|
|
2018-11-23 21:23:25 +08:00
|
|
|
static void i915_error_puts(struct drm_i915_error_state_buf *e, const char *str)
|
2013-07-12 21:50:57 +08:00
|
|
|
{
|
|
|
|
unsigned len;
|
|
|
|
|
2018-11-23 21:23:25 +08:00
|
|
|
if (e->err || !str)
|
2013-07-12 21:50:57 +08:00
|
|
|
return;
|
|
|
|
|
|
|
|
len = strlen(str);
|
2018-11-23 21:23:25 +08:00
|
|
|
if (!__i915_error_grow(e, len))
|
|
|
|
return;
|
2013-07-12 21:50:57 +08:00
|
|
|
|
2018-11-23 21:23:25 +08:00
|
|
|
GEM_BUG_ON(e->bytes + len > e->size);
|
2013-07-12 21:50:57 +08:00
|
|
|
memcpy(e->buf + e->bytes, str, len);
|
2018-11-23 21:23:25 +08:00
|
|
|
e->bytes += len;
|
2013-07-12 21:50:57 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
#define err_printf(e, ...) i915_error_printf(e, __VA_ARGS__)
|
|
|
|
#define err_puts(e, s) i915_error_puts(e, s)
|
|
|
|
|
2017-10-27 01:36:55 +08:00
|
|
|
static void __i915_printfn_error(struct drm_printer *p, struct va_format *vaf)
|
|
|
|
{
|
|
|
|
i915_error_vprintf(p->arg, vaf->fmt, *vaf->va);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline struct drm_printer
|
|
|
|
i915_error_printer(struct drm_i915_error_state_buf *e)
|
|
|
|
{
|
|
|
|
struct drm_printer p = {
|
|
|
|
.printfn = __i915_printfn_error,
|
|
|
|
.arg = e,
|
|
|
|
};
|
|
|
|
return p;
|
|
|
|
}
|
|
|
|
|
2016-10-12 17:05:22 +08:00
|
|
|
#ifdef CONFIG_DRM_I915_COMPRESS_ERROR
|
|
|
|
|
2016-12-06 20:40:51 +08:00
|
|
|
struct compress {
|
|
|
|
struct z_stream_s zstream;
|
|
|
|
void *tmp;
|
|
|
|
};
|
|
|
|
|
|
|
|
static bool compress_init(struct compress *c)
|
2016-10-12 17:05:22 +08:00
|
|
|
{
|
2016-12-06 20:40:51 +08:00
|
|
|
struct z_stream_s *zstream = memset(&c->zstream, 0, sizeof(c->zstream));
|
2016-10-12 17:05:22 +08:00
|
|
|
|
|
|
|
zstream->workspace =
|
|
|
|
kmalloc(zlib_deflate_workspacesize(MAX_WBITS, MAX_MEM_LEVEL),
|
|
|
|
GFP_ATOMIC | __GFP_NOWARN);
|
|
|
|
if (!zstream->workspace)
|
|
|
|
return false;
|
|
|
|
|
|
|
|
if (zlib_deflateInit(zstream, Z_DEFAULT_COMPRESSION) != Z_OK) {
|
|
|
|
kfree(zstream->workspace);
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
2016-12-06 20:40:51 +08:00
|
|
|
c->tmp = NULL;
|
2017-01-06 23:20:09 +08:00
|
|
|
if (i915_has_memcpy_from_wc())
|
2016-12-06 20:40:51 +08:00
|
|
|
c->tmp = (void *)__get_free_page(GFP_ATOMIC | __GFP_NOWARN);
|
|
|
|
|
2016-10-12 17:05:22 +08:00
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2018-10-03 16:24:22 +08:00
|
|
|
static void *compress_next_page(struct drm_i915_error_object *dst)
|
|
|
|
{
|
|
|
|
unsigned long page;
|
|
|
|
|
|
|
|
if (dst->page_count >= dst->num_pages)
|
|
|
|
return ERR_PTR(-ENOSPC);
|
|
|
|
|
|
|
|
page = __get_free_page(GFP_ATOMIC | __GFP_NOWARN);
|
|
|
|
if (!page)
|
|
|
|
return ERR_PTR(-ENOMEM);
|
|
|
|
|
|
|
|
return dst->pages[dst->page_count++] = (void *)page;
|
|
|
|
}
|
|
|
|
|
2016-12-06 20:40:51 +08:00
|
|
|
static int compress_page(struct compress *c,
|
2016-10-12 17:05:22 +08:00
|
|
|
void *src,
|
|
|
|
struct drm_i915_error_object *dst)
|
|
|
|
{
|
2016-12-06 20:40:51 +08:00
|
|
|
struct z_stream_s *zstream = &c->zstream;
|
|
|
|
|
2016-10-12 17:05:22 +08:00
|
|
|
zstream->next_in = src;
|
2016-12-06 20:40:51 +08:00
|
|
|
if (c->tmp && i915_memcpy_from_wc(c->tmp, src, PAGE_SIZE))
|
|
|
|
zstream->next_in = c->tmp;
|
2016-10-12 17:05:22 +08:00
|
|
|
zstream->avail_in = PAGE_SIZE;
|
|
|
|
|
|
|
|
do {
|
|
|
|
if (zstream->avail_out == 0) {
|
2018-10-03 16:24:22 +08:00
|
|
|
zstream->next_out = compress_next_page(dst);
|
|
|
|
if (IS_ERR(zstream->next_out))
|
|
|
|
return PTR_ERR(zstream->next_out);
|
2016-10-12 17:05:22 +08:00
|
|
|
|
|
|
|
zstream->avail_out = PAGE_SIZE;
|
|
|
|
}
|
|
|
|
|
2018-10-03 16:24:22 +08:00
|
|
|
if (zlib_deflate(zstream, Z_NO_FLUSH) != Z_OK)
|
2016-10-12 17:05:22 +08:00
|
|
|
return -EIO;
|
2018-11-23 21:23:25 +08:00
|
|
|
|
|
|
|
touch_nmi_watchdog();
|
2016-10-12 17:05:22 +08:00
|
|
|
} while (zstream->avail_in);
|
|
|
|
|
|
|
|
/* Fallback to uncompressed if we increase size? */
|
|
|
|
if (0 && zstream->total_out > zstream->total_in)
|
|
|
|
return -E2BIG;
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2018-10-03 16:24:22 +08:00
|
|
|
static int compress_flush(struct compress *c,
|
2016-10-12 17:05:22 +08:00
|
|
|
struct drm_i915_error_object *dst)
|
|
|
|
{
|
2016-12-06 20:40:51 +08:00
|
|
|
struct z_stream_s *zstream = &c->zstream;
|
|
|
|
|
2018-10-03 16:24:22 +08:00
|
|
|
do {
|
|
|
|
switch (zlib_deflate(zstream, Z_FINISH)) {
|
|
|
|
case Z_OK: /* more space requested */
|
|
|
|
zstream->next_out = compress_next_page(dst);
|
|
|
|
if (IS_ERR(zstream->next_out))
|
|
|
|
return PTR_ERR(zstream->next_out);
|
|
|
|
|
|
|
|
zstream->avail_out = PAGE_SIZE;
|
|
|
|
break;
|
|
|
|
|
|
|
|
case Z_STREAM_END:
|
|
|
|
goto end;
|
|
|
|
|
|
|
|
default: /* any error */
|
|
|
|
return -EIO;
|
|
|
|
}
|
|
|
|
} while (1);
|
|
|
|
|
|
|
|
end:
|
|
|
|
memset(zstream->next_out, 0, zstream->avail_out);
|
|
|
|
dst->unused = zstream->avail_out;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void compress_fini(struct compress *c,
|
|
|
|
struct drm_i915_error_object *dst)
|
|
|
|
{
|
|
|
|
struct z_stream_s *zstream = &c->zstream;
|
2016-10-12 17:05:22 +08:00
|
|
|
|
|
|
|
zlib_deflateEnd(zstream);
|
|
|
|
kfree(zstream->workspace);
|
2016-12-06 20:40:51 +08:00
|
|
|
if (c->tmp)
|
|
|
|
free_page((unsigned long)c->tmp);
|
2016-10-12 17:05:22 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
static void err_compression_marker(struct drm_i915_error_state_buf *m)
|
|
|
|
{
|
|
|
|
err_puts(m, ":");
|
|
|
|
}
|
|
|
|
|
|
|
|
#else
|
|
|
|
|
2016-12-06 20:40:51 +08:00
|
|
|
struct compress {
|
|
|
|
};
|
|
|
|
|
|
|
|
static bool compress_init(struct compress *c)
|
2016-10-12 17:05:22 +08:00
|
|
|
{
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2016-12-06 20:40:51 +08:00
|
|
|
static int compress_page(struct compress *c,
|
2016-10-12 17:05:22 +08:00
|
|
|
void *src,
|
|
|
|
struct drm_i915_error_object *dst)
|
|
|
|
{
|
|
|
|
unsigned long page;
|
2016-12-06 20:40:51 +08:00
|
|
|
void *ptr;
|
2016-10-12 17:05:22 +08:00
|
|
|
|
|
|
|
page = __get_free_page(GFP_ATOMIC | __GFP_NOWARN);
|
|
|
|
if (!page)
|
|
|
|
return -ENOMEM;
|
|
|
|
|
2016-12-06 20:40:51 +08:00
|
|
|
ptr = (void *)page;
|
|
|
|
if (!i915_memcpy_from_wc(ptr, src, PAGE_SIZE))
|
|
|
|
memcpy(ptr, src, PAGE_SIZE);
|
|
|
|
dst->pages[dst->page_count++] = ptr;
|
2016-10-12 17:05:22 +08:00
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2018-10-03 16:24:22 +08:00
|
|
|
static int compress_flush(struct compress *c,
|
|
|
|
struct drm_i915_error_object *dst)
|
|
|
|
{
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2016-12-06 20:40:51 +08:00
|
|
|
static void compress_fini(struct compress *c,
|
2016-10-12 17:05:22 +08:00
|
|
|
struct drm_i915_error_object *dst)
|
|
|
|
{
|
|
|
|
}
|
|
|
|
|
|
|
|
static void err_compression_marker(struct drm_i915_error_state_buf *m)
|
|
|
|
{
|
|
|
|
err_puts(m, "~");
|
|
|
|
}
|
|
|
|
|
|
|
|
#endif
|
|
|
|
|
2013-07-12 21:50:57 +08:00
|
|
|
static void print_error_buffers(struct drm_i915_error_state_buf *m,
|
|
|
|
const char *name,
|
|
|
|
struct drm_i915_error_buffer *err,
|
|
|
|
int count)
|
|
|
|
{
|
2016-08-15 17:48:41 +08:00
|
|
|
err_printf(m, "%s [%d]:\n", name, count);
|
2013-07-12 21:50:57 +08:00
|
|
|
|
|
|
|
while (count--) {
|
2019-02-26 17:49:21 +08:00
|
|
|
err_printf(m, " %08x_%08x %8u %02x %02x",
|
2015-07-30 00:23:56 +08:00
|
|
|
upper_32_bits(err->gtt_offset),
|
|
|
|
lower_32_bits(err->gtt_offset),
|
2013-07-12 21:50:57 +08:00
|
|
|
err->size,
|
|
|
|
err->read_domains,
|
2019-02-26 17:49:21 +08:00
|
|
|
err->write_domain);
|
2013-07-12 21:50:57 +08:00
|
|
|
err_puts(m, tiling_flag(err->tiling));
|
|
|
|
err_puts(m, dirty_flag(err->dirty));
|
|
|
|
err_puts(m, purgeable_flag(err->purgeable));
|
drm/i915: Introduce mapping of user pages into video memory (userptr) ioctl
By exporting the ability to map user address and inserting PTEs
representing their backing pages into the GTT, we can exploit UMA in order
to utilize normal application data as a texture source or even as a
render target (depending upon the capabilities of the chipset). This has
a number of uses, with zero-copy downloads to the GPU and efficient
readback making the intermixed streaming of CPU and GPU operations
fairly efficient. This ability has many widespread implications from
faster rendering of client-side software rasterisers (chromium),
mitigation of stalls due to read back (firefox) and to faster pipelining
of texture data (such as pixel buffer objects in GL or data blobs in CL).
v2: Compile with CONFIG_MMU_NOTIFIER
v3: We can sleep while performing invalidate-range, which we can utilise
to drop our page references prior to the kernel manipulating the vma
(for either discard or cloning) and so protect normal users.
v4: Only run the invalidate notifier if the range intercepts the bo.
v5: Prevent userspace from attempting to GTT mmap non-page aligned buffers
v6: Recheck after reacquire mutex for lost mmu.
v7: Fix implicit padding of ioctl struct by rounding to next 64bit boundary.
v8: Fix rebasing error after forwarding porting the back port.
v9: Limit the userptr to page aligned entries. We now expect userspace
to handle all the offset-in-page adjustments itself.
v10: Prevent vma from being copied across fork to avoid issues with cow.
v11: Drop vma behaviour changes -- locking is nigh on impossible.
Use a worker to load user pages to avoid lock inversions.
v12: Use get_task_mm()/mmput() for correct refcounting of mm.
v13: Use a worker to release the mmu_notifier to avoid lock inversion
v14: Decouple mmu_notifier from struct_mutex using a custom mmu_notifer
with its own locking and tree of objects for each mm/mmu_notifier.
v15: Prevent overlapping userptr objects, and invalidate all objects
within the mmu_notifier range
v16: Fix a typo for iterating over multiple objects in the range and
rearrange error path to destroy the mmu_notifier locklessly.
Also close a race between invalidate_range and the get_pages_worker.
v17: Close a race between get_pages_worker/invalidate_range and fresh
allocations of the same userptr range - and notice that
struct_mutex was presumed to be held when during creation it wasn't.
v18: Sigh. Fix the refactor of st_set_pages() to allocate enough memory
for the struct sg_table and to clear it before reporting an error.
v19: Always error out on read-only userptr requests as we don't have the
hardware infrastructure to support them at the moment.
v20: Refuse to implement read-only support until we have the required
infrastructure - but reserve the bit in flags for future use.
v21: use_mm() is not required for get_user_pages(). It is only meant to
be used to fix up the kernel thread's current->mm for use with
copy_user().
v22: Use sg_alloc_table_from_pages for that chunky feeling
v23: Export a function for sanity checking dma-buf rather than encode
userptr details elsewhere, and clean up comments based on
suggestions by Bradley.
Signed-off-by: Chris Wilson <chris@chris-wilson.co.uk>
Cc: Tvrtko Ursulin <tvrtko.ursulin@linux.intel.com>
Cc: "Gong, Zhipeng" <zhipeng.gong@intel.com>
Cc: Akash Goel <akash.goel@intel.com>
Cc: "Volkin, Bradley D" <bradley.d.volkin@intel.com>
Reviewed-by: Tvrtko Ursulin <tvrtko.ursulin@linux.intel.com>
Reviewed-by: Brad Volkin <bradley.d.volkin@intel.com>
[danvet: Frob ioctl allocation to pick the next one - will cause a bit
of fuss with create2 apparently, but such are the rules.]
[danvet2: oops, forgot to git add after manual patch application]
[danvet3: Appease sparse.]
Signed-off-by: Daniel Vetter <daniel.vetter@ffwll.ch>
2014-05-16 21:22:37 +08:00
|
|
|
err_puts(m, err->userptr ? " userptr" : "");
|
2014-08-22 21:41:39 +08:00
|
|
|
err_puts(m, i915_cache_level_str(m->i915, err->cache_level));
|
2013-07-12 21:50:57 +08:00
|
|
|
|
|
|
|
if (err->name)
|
|
|
|
err_printf(m, " (name: %d)", err->name);
|
|
|
|
if (err->fence_reg != I915_FENCE_REG_NONE)
|
|
|
|
err_printf(m, " (fence: %d)", err->fence_reg);
|
|
|
|
|
|
|
|
err_puts(m, "\n");
|
|
|
|
err++;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2016-09-20 21:54:32 +08:00
|
|
|
static void error_print_instdone(struct drm_i915_error_state_buf *m,
|
2017-02-15 00:46:11 +08:00
|
|
|
const struct drm_i915_error_engine *ee)
|
2016-09-20 21:54:32 +08:00
|
|
|
{
|
2016-09-20 21:54:33 +08:00
|
|
|
int slice;
|
|
|
|
int subslice;
|
|
|
|
|
2016-09-20 21:54:32 +08:00
|
|
|
err_printf(m, " INSTDONE: 0x%08x\n",
|
|
|
|
ee->instdone.instdone);
|
|
|
|
|
|
|
|
if (ee->engine_id != RCS || INTEL_GEN(m->i915) <= 3)
|
|
|
|
return;
|
|
|
|
|
|
|
|
err_printf(m, " SC_INSTDONE: 0x%08x\n",
|
|
|
|
ee->instdone.slice_common);
|
|
|
|
|
|
|
|
if (INTEL_GEN(m->i915) <= 6)
|
|
|
|
return;
|
|
|
|
|
2016-09-20 21:54:33 +08:00
|
|
|
for_each_instdone_slice_subslice(m->i915, slice, subslice)
|
|
|
|
err_printf(m, " SAMPLER_INSTDONE[%d][%d]: 0x%08x\n",
|
|
|
|
slice, subslice,
|
|
|
|
ee->instdone.sampler[slice][subslice]);
|
|
|
|
|
|
|
|
for_each_instdone_slice_subslice(m->i915, slice, subslice)
|
|
|
|
err_printf(m, " ROW_INSTDONE[%d][%d]: 0x%08x\n",
|
|
|
|
slice, subslice,
|
|
|
|
ee->instdone.row[slice][subslice]);
|
2016-09-20 21:54:32 +08:00
|
|
|
}
|
|
|
|
|
2016-10-13 18:18:14 +08:00
|
|
|
static void error_print_request(struct drm_i915_error_state_buf *m,
|
|
|
|
const char *prefix,
|
2018-04-30 15:52:59 +08:00
|
|
|
const struct drm_i915_error_request *erq,
|
|
|
|
const unsigned long epoch)
|
2016-10-13 18:18:14 +08:00
|
|
|
{
|
|
|
|
if (!erq->seqno)
|
|
|
|
return;
|
|
|
|
|
drm/i915: Use time based guilty context banning
Currently, we accumulate each time a context hangs the GPU, offset
against the number of requests it submits, and if that score exceeds a
certain threshold, we ban that context from submitting any more requests
(cancelling any work in flight). In contrast, we use a simple timer on
the file, that if we see more than a 9 hangs faster than 60s apart in
total across all of its contexts, we will ban the client from creating
any more contexts. This leads to a confusing situation where the file
may be banned before the context, so lets use a simple timer scheme for
each.
If the context submits 3 hanging requests within a 120s period, declare
it forbidden to ever send more requests.
This has the advantage of not being easy to repair by simply sending
empty requests, but has the disadvantage that if the context is idle
then it is forgiven. However, if the context is idle, it is not
disrupting the system, but a hog can evade the request counting and
cause much more severe disruption to the system.
Updating ban_score from request retirement is dubious as the retirement
is purposely not in sync with request submission (i.e. we try and batch
retirement to reduce overhead and avoid latency on submission), which
leads to surprising situations where we can forgive a hang immediately
due to a backlog of requests from before the hang being retired
afterwards.
Signed-off-by: Chris Wilson <chris@chris-wilson.co.uk>
Cc: Mika Kuoppala <mika.kuoppala@intel.com>
Reviewed-by: Mika Kuoppala <mika.kuoppala@intel.com>
Link: https://patchwork.freedesktop.org/patch/msgid/20190219122215.8941-2-chris@chris-wilson.co.uk
2019-02-19 20:21:52 +08:00
|
|
|
err_printf(m, "%s pid %d, seqno %8x:%08x%s%s, prio %d, emitted %dms, start %08x, head %08x, tail %08x\n",
|
|
|
|
prefix, erq->pid, erq->context, erq->seqno,
|
drm/i915: Replace global breadcrumbs with per-context interrupt tracking
A few years ago, see commit 688e6c725816 ("drm/i915: Slaughter the
thundering i915_wait_request herd"), the issue of handling multiple
clients waiting in parallel was brought to our attention. The
requirement was that every client should be woken immediately upon its
request being signaled, without incurring any cpu overhead.
To handle certain fragility of our hw meant that we could not do a
simple check inside the irq handler (some generations required almost
unbounded delays before we could be sure of seqno coherency) and so
request completion checking required delegation.
Before commit 688e6c725816, the solution was simple. Every client
waiting on a request would be woken on every interrupt and each would do
a heavyweight check to see if their request was complete. Commit
688e6c725816 introduced an rbtree so that only the earliest waiter on
the global timeline would woken, and would wake the next and so on.
(Along with various complications to handle requests being reordered
along the global timeline, and also a requirement for kthread to provide
a delegate for fence signaling that had no process context.)
The global rbtree depends on knowing the execution timeline (and global
seqno). Without knowing that order, we must instead check all contexts
queued to the HW to see which may have advanced. We trim that list by
only checking queued contexts that are being waited on, but still we
keep a list of all active contexts and their active signalers that we
inspect from inside the irq handler. By moving the waiters onto the fence
signal list, we can combine the client wakeup with the dma_fence
signaling (a dramatic reduction in complexity, but does require the HW
being coherent, the seqno must be visible from the cpu before the
interrupt is raised - we keep a timer backup just in case).
Having previously fixed all the issues with irq-seqno serialisation (by
inserting delays onto the GPU after each request instead of random delays
on the CPU after each interrupt), we can rely on the seqno state to
perfom direct wakeups from the interrupt handler. This allows us to
preserve our single context switch behaviour of the current routine,
with the only downside that we lose the RT priority sorting of wakeups.
In general, direct wakeup latency of multiple clients is about the same
(about 10% better in most cases) with a reduction in total CPU time spent
in the waiter (about 20-50% depending on gen). Average herd behaviour is
improved, but at the cost of not delegating wakeups on task_prio.
v2: Capture fence signaling state for error state and add comments to
warm even the most cold of hearts.
v3: Check if the request is still active before busywaiting
v4: Reduce the amount of pointer misdirection with list_for_each_safe
and using a local i915_request variable inside the loops
v5: Add a missing pluralisation to a purely informative selftest message.
References: 688e6c725816 ("drm/i915: Slaughter the thundering i915_wait_request herd")
Signed-off-by: Chris Wilson <chris@chris-wilson.co.uk>
Cc: Tvrtko Ursulin <tvrtko.ursulin@intel.com>
Reviewed-by: Tvrtko Ursulin <tvrtko.ursulin@intel.com>
Link: https://patchwork.freedesktop.org/patch/msgid/20190129205230.19056-2-chris@chris-wilson.co.uk
2019-01-30 04:52:29 +08:00
|
|
|
test_bit(DMA_FENCE_FLAG_SIGNALED_BIT,
|
|
|
|
&erq->flags) ? "!" : "",
|
|
|
|
test_bit(DMA_FENCE_FLAG_ENABLE_SIGNAL_BIT,
|
|
|
|
&erq->flags) ? "+" : "",
|
|
|
|
erq->sched_attr.priority,
|
2018-04-30 15:52:59 +08:00
|
|
|
jiffies_to_msecs(erq->jiffies - epoch),
|
2018-05-02 18:41:50 +08:00
|
|
|
erq->start, erq->head, erq->tail);
|
2016-10-13 18:18:14 +08:00
|
|
|
}
|
|
|
|
|
2017-01-29 17:24:33 +08:00
|
|
|
static void error_print_context(struct drm_i915_error_state_buf *m,
|
|
|
|
const char *header,
|
2017-02-15 00:46:11 +08:00
|
|
|
const struct drm_i915_error_context *ctx)
|
2017-01-29 17:24:33 +08:00
|
|
|
{
|
drm/i915: Use time based guilty context banning
Currently, we accumulate each time a context hangs the GPU, offset
against the number of requests it submits, and if that score exceeds a
certain threshold, we ban that context from submitting any more requests
(cancelling any work in flight). In contrast, we use a simple timer on
the file, that if we see more than a 9 hangs faster than 60s apart in
total across all of its contexts, we will ban the client from creating
any more contexts. This leads to a confusing situation where the file
may be banned before the context, so lets use a simple timer scheme for
each.
If the context submits 3 hanging requests within a 120s period, declare
it forbidden to ever send more requests.
This has the advantage of not being easy to repair by simply sending
empty requests, but has the disadvantage that if the context is idle
then it is forgiven. However, if the context is idle, it is not
disrupting the system, but a hog can evade the request counting and
cause much more severe disruption to the system.
Updating ban_score from request retirement is dubious as the retirement
is purposely not in sync with request submission (i.e. we try and batch
retirement to reduce overhead and avoid latency on submission), which
leads to surprising situations where we can forgive a hang immediately
due to a backlog of requests from before the hang being retired
afterwards.
Signed-off-by: Chris Wilson <chris@chris-wilson.co.uk>
Cc: Mika Kuoppala <mika.kuoppala@intel.com>
Reviewed-by: Mika Kuoppala <mika.kuoppala@intel.com>
Link: https://patchwork.freedesktop.org/patch/msgid/20190219122215.8941-2-chris@chris-wilson.co.uk
2019-02-19 20:21:52 +08:00
|
|
|
err_printf(m, "%s%s[%d] user_handle %d hw_id %d, prio %d, guilty %d active %d\n",
|
2017-01-29 17:24:33 +08:00
|
|
|
header, ctx->comm, ctx->pid, ctx->handle, ctx->hw_id,
|
drm/i915: Use time based guilty context banning
Currently, we accumulate each time a context hangs the GPU, offset
against the number of requests it submits, and if that score exceeds a
certain threshold, we ban that context from submitting any more requests
(cancelling any work in flight). In contrast, we use a simple timer on
the file, that if we see more than a 9 hangs faster than 60s apart in
total across all of its contexts, we will ban the client from creating
any more contexts. This leads to a confusing situation where the file
may be banned before the context, so lets use a simple timer scheme for
each.
If the context submits 3 hanging requests within a 120s period, declare
it forbidden to ever send more requests.
This has the advantage of not being easy to repair by simply sending
empty requests, but has the disadvantage that if the context is idle
then it is forgiven. However, if the context is idle, it is not
disrupting the system, but a hog can evade the request counting and
cause much more severe disruption to the system.
Updating ban_score from request retirement is dubious as the retirement
is purposely not in sync with request submission (i.e. we try and batch
retirement to reduce overhead and avoid latency on submission), which
leads to surprising situations where we can forgive a hang immediately
due to a backlog of requests from before the hang being retired
afterwards.
Signed-off-by: Chris Wilson <chris@chris-wilson.co.uk>
Cc: Mika Kuoppala <mika.kuoppala@intel.com>
Reviewed-by: Mika Kuoppala <mika.kuoppala@intel.com>
Link: https://patchwork.freedesktop.org/patch/msgid/20190219122215.8941-2-chris@chris-wilson.co.uk
2019-02-19 20:21:52 +08:00
|
|
|
ctx->sched_attr.priority, ctx->guilty, ctx->active);
|
2017-01-29 17:24:33 +08:00
|
|
|
}
|
|
|
|
|
2016-07-27 16:07:28 +08:00
|
|
|
static void error_print_engine(struct drm_i915_error_state_buf *m,
|
2018-04-30 15:52:59 +08:00
|
|
|
const struct drm_i915_error_engine *ee,
|
|
|
|
const unsigned long epoch)
|
2013-07-12 21:50:57 +08:00
|
|
|
{
|
2017-09-22 20:43:07 +08:00
|
|
|
int n;
|
|
|
|
|
2018-01-19 01:52:28 +08:00
|
|
|
err_printf(m, "%s command stream:\n",
|
|
|
|
engine_name(m->i915, ee->engine_id));
|
2017-12-19 21:14:19 +08:00
|
|
|
err_printf(m, " IDLE?: %s\n", yesno(ee->idle));
|
2016-07-27 16:07:28 +08:00
|
|
|
err_printf(m, " START: 0x%08x\n", ee->start);
|
2016-10-13 18:18:15 +08:00
|
|
|
err_printf(m, " HEAD: 0x%08x [0x%08x]\n", ee->head, ee->rq_head);
|
2016-10-05 04:11:30 +08:00
|
|
|
err_printf(m, " TAIL: 0x%08x [0x%08x, 0x%08x]\n",
|
|
|
|
ee->tail, ee->rq_post, ee->rq_tail);
|
2016-07-27 16:07:28 +08:00
|
|
|
err_printf(m, " CTL: 0x%08x\n", ee->ctl);
|
2016-08-15 17:49:11 +08:00
|
|
|
err_printf(m, " MODE: 0x%08x\n", ee->mode);
|
2016-07-27 16:07:28 +08:00
|
|
|
err_printf(m, " HWS: 0x%08x\n", ee->hws);
|
|
|
|
err_printf(m, " ACTHD: 0x%08x %08x\n",
|
|
|
|
(u32)(ee->acthd>>32), (u32)ee->acthd);
|
|
|
|
err_printf(m, " IPEIR: 0x%08x\n", ee->ipeir);
|
|
|
|
err_printf(m, " IPEHR: 0x%08x\n", ee->ipehr);
|
2016-09-20 21:54:32 +08:00
|
|
|
|
|
|
|
error_print_instdone(m, ee);
|
|
|
|
|
2016-08-15 17:49:09 +08:00
|
|
|
if (ee->batchbuffer) {
|
|
|
|
u64 start = ee->batchbuffer->gtt_offset;
|
|
|
|
u64 end = start + ee->batchbuffer->gtt_size;
|
|
|
|
|
|
|
|
err_printf(m, " batch: [0x%08x_%08x, 0x%08x_%08x]\n",
|
|
|
|
upper_32_bits(start), lower_32_bits(start),
|
|
|
|
upper_32_bits(end), lower_32_bits(end));
|
|
|
|
}
|
2016-07-27 16:07:28 +08:00
|
|
|
if (INTEL_GEN(m->i915) >= 4) {
|
2016-08-15 17:49:09 +08:00
|
|
|
err_printf(m, " BBADDR: 0x%08x_%08x\n",
|
2016-07-27 16:07:28 +08:00
|
|
|
(u32)(ee->bbaddr>>32), (u32)ee->bbaddr);
|
|
|
|
err_printf(m, " BB_STATE: 0x%08x\n", ee->bbstate);
|
|
|
|
err_printf(m, " INSTPS: 0x%08x\n", ee->instps);
|
2013-12-11 03:44:43 +08:00
|
|
|
}
|
2016-07-27 16:07:28 +08:00
|
|
|
err_printf(m, " INSTPM: 0x%08x\n", ee->instpm);
|
|
|
|
err_printf(m, " FADDR: 0x%08x %08x\n", upper_32_bits(ee->faddr),
|
|
|
|
lower_32_bits(ee->faddr));
|
|
|
|
if (INTEL_GEN(m->i915) >= 6) {
|
|
|
|
err_printf(m, " RC PSMI: 0x%08x\n", ee->rc_psmi);
|
|
|
|
err_printf(m, " FAULT_REG: 0x%08x\n", ee->fault_reg);
|
2016-10-28 20:58:53 +08:00
|
|
|
err_printf(m, " SYNC_0: 0x%08x\n",
|
|
|
|
ee->semaphore_mboxes[0]);
|
|
|
|
err_printf(m, " SYNC_1: 0x%08x\n",
|
|
|
|
ee->semaphore_mboxes[1]);
|
|
|
|
if (HAS_VEBOX(m->i915))
|
|
|
|
err_printf(m, " SYNC_2: 0x%08x\n",
|
|
|
|
ee->semaphore_mboxes[2]);
|
2013-07-12 21:50:57 +08:00
|
|
|
}
|
2018-09-27 04:12:22 +08:00
|
|
|
if (HAS_PPGTT(m->i915)) {
|
2016-07-27 16:07:28 +08:00
|
|
|
err_printf(m, " GFX_MODE: 0x%08x\n", ee->vm_info.gfx_mode);
|
2014-01-30 16:19:40 +08:00
|
|
|
|
2016-07-27 16:07:28 +08:00
|
|
|
if (INTEL_GEN(m->i915) >= 8) {
|
2014-01-30 16:19:40 +08:00
|
|
|
int i;
|
|
|
|
for (i = 0; i < 4; i++)
|
|
|
|
err_printf(m, " PDP%d: 0x%016llx\n",
|
2016-07-27 16:07:28 +08:00
|
|
|
i, ee->vm_info.pdp[i]);
|
2014-01-30 16:19:40 +08:00
|
|
|
} else {
|
|
|
|
err_printf(m, " PP_DIR_BASE: 0x%08x\n",
|
2016-07-27 16:07:28 +08:00
|
|
|
ee->vm_info.pp_dir_base);
|
2014-01-30 16:19:40 +08:00
|
|
|
}
|
|
|
|
}
|
2016-07-27 16:07:28 +08:00
|
|
|
err_printf(m, " ring->head: 0x%08x\n", ee->cpu_ring_head);
|
|
|
|
err_printf(m, " ring->tail: 0x%08x\n", ee->cpu_ring_tail);
|
2019-01-25 21:22:28 +08:00
|
|
|
err_printf(m, " hangcheck timestamp: %dms (%lu%s)\n",
|
2018-04-30 15:52:59 +08:00
|
|
|
jiffies_to_msecs(ee->hangcheck_timestamp - epoch),
|
2016-11-18 21:09:04 +08:00
|
|
|
ee->hangcheck_timestamp,
|
2018-04-30 15:52:59 +08:00
|
|
|
ee->hangcheck_timestamp == epoch ? "; epoch" : "");
|
2017-06-20 17:57:48 +08:00
|
|
|
err_printf(m, " engine reset count: %u\n", ee->reset_count);
|
2016-11-18 21:09:04 +08:00
|
|
|
|
2017-09-22 20:43:07 +08:00
|
|
|
for (n = 0; n < ee->num_ports; n++) {
|
|
|
|
err_printf(m, " ELSP[%d]:", n);
|
2018-04-30 15:52:59 +08:00
|
|
|
error_print_request(m, " ", &ee->execlist[n], epoch);
|
2017-09-22 20:43:07 +08:00
|
|
|
}
|
|
|
|
|
2017-01-29 17:24:33 +08:00
|
|
|
error_print_context(m, " Active context: ", &ee->context);
|
2013-07-12 21:50:57 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
void i915_error_printf(struct drm_i915_error_state_buf *e, const char *f, ...)
|
|
|
|
{
|
|
|
|
va_list args;
|
|
|
|
|
|
|
|
va_start(args, f);
|
|
|
|
i915_error_vprintf(e, f, args);
|
|
|
|
va_end(args);
|
|
|
|
}
|
|
|
|
|
2014-02-25 23:11:24 +08:00
|
|
|
static void print_error_obj(struct drm_i915_error_state_buf *m,
|
2016-10-12 17:05:21 +08:00
|
|
|
struct intel_engine_cs *engine,
|
|
|
|
const char *name,
|
2014-02-25 23:11:24 +08:00
|
|
|
struct drm_i915_error_object *obj)
|
|
|
|
{
|
2018-07-25 00:33:19 +08:00
|
|
|
char out[ASCII85_BUFSZ];
|
2016-10-12 17:05:22 +08:00
|
|
|
int page;
|
2014-02-25 23:11:24 +08:00
|
|
|
|
2016-10-12 17:05:21 +08:00
|
|
|
if (!obj)
|
|
|
|
return;
|
|
|
|
|
|
|
|
if (name) {
|
|
|
|
err_printf(m, "%s --- %s = 0x%08x %08x\n",
|
|
|
|
engine ? engine->name : "global", name,
|
|
|
|
upper_32_bits(obj->gtt_offset),
|
|
|
|
lower_32_bits(obj->gtt_offset));
|
|
|
|
}
|
|
|
|
|
2016-10-12 17:05:22 +08:00
|
|
|
err_compression_marker(m);
|
|
|
|
for (page = 0; page < obj->page_count; page++) {
|
|
|
|
int i, len;
|
|
|
|
|
|
|
|
len = PAGE_SIZE;
|
|
|
|
if (page == obj->page_count - 1)
|
|
|
|
len -= obj->unused;
|
|
|
|
len = ascii85_encode_len(len);
|
|
|
|
|
2018-07-25 00:33:19 +08:00
|
|
|
for (i = 0; i < len; i++)
|
|
|
|
err_puts(m, ascii85_encode(obj->pages[page][i], out));
|
2014-02-25 23:11:24 +08:00
|
|
|
}
|
2016-10-12 17:05:22 +08:00
|
|
|
err_puts(m, "\n");
|
2014-02-25 23:11:24 +08:00
|
|
|
}
|
|
|
|
|
2016-08-15 17:48:45 +08:00
|
|
|
static void err_print_capabilities(struct drm_i915_error_state_buf *m,
|
2018-02-08 05:05:43 +08:00
|
|
|
const struct intel_device_info *info,
|
2018-12-31 22:56:41 +08:00
|
|
|
const struct intel_runtime_info *runtime,
|
2018-02-08 05:05:43 +08:00
|
|
|
const struct intel_driver_caps *caps)
|
2016-08-15 17:48:45 +08:00
|
|
|
{
|
2017-12-19 19:43:44 +08:00
|
|
|
struct drm_printer p = i915_error_printer(m);
|
|
|
|
|
|
|
|
intel_device_info_dump_flags(info, &p);
|
2018-02-08 05:05:43 +08:00
|
|
|
intel_driver_caps_print(caps, &p);
|
2018-12-31 22:56:41 +08:00
|
|
|
intel_device_info_dump_topology(&runtime->sseu, &p);
|
2016-08-15 17:48:45 +08:00
|
|
|
}
|
|
|
|
|
2017-02-07 05:36:07 +08:00
|
|
|
static void err_print_params(struct drm_i915_error_state_buf *m,
|
2017-12-19 19:43:46 +08:00
|
|
|
const struct i915_params *params)
|
2017-02-07 05:36:07 +08:00
|
|
|
{
|
2017-12-19 19:43:46 +08:00
|
|
|
struct drm_printer p = i915_error_printer(m);
|
|
|
|
|
|
|
|
i915_params_dump(params, &p);
|
2017-02-07 05:36:07 +08:00
|
|
|
}
|
|
|
|
|
2017-02-15 00:46:11 +08:00
|
|
|
static void err_print_pciid(struct drm_i915_error_state_buf *m,
|
|
|
|
struct drm_i915_private *i915)
|
|
|
|
{
|
|
|
|
struct pci_dev *pdev = i915->drm.pdev;
|
|
|
|
|
|
|
|
err_printf(m, "PCI ID: 0x%04x\n", pdev->device);
|
|
|
|
err_printf(m, "PCI Revision: 0x%02x\n", pdev->revision);
|
|
|
|
err_printf(m, "PCI Subsystem: %04x:%04x\n",
|
|
|
|
pdev->subsystem_vendor,
|
|
|
|
pdev->subsystem_device);
|
|
|
|
}
|
|
|
|
|
2017-10-27 01:36:55 +08:00
|
|
|
static void err_print_uc(struct drm_i915_error_state_buf *m,
|
|
|
|
const struct i915_error_uc *error_uc)
|
|
|
|
{
|
|
|
|
struct drm_printer p = i915_error_printer(m);
|
|
|
|
const struct i915_gpu_state *error =
|
|
|
|
container_of(error_uc, typeof(*error), uc);
|
|
|
|
|
|
|
|
if (!error->device_info.has_guc)
|
|
|
|
return;
|
|
|
|
|
|
|
|
intel_uc_fw_dump(&error_uc->guc_fw, &p);
|
|
|
|
intel_uc_fw_dump(&error_uc->huc_fw, &p);
|
2017-10-27 01:36:56 +08:00
|
|
|
print_error_obj(m, NULL, "GuC log buffer", error_uc->guc_log);
|
2017-10-27 01:36:55 +08:00
|
|
|
}
|
|
|
|
|
2018-11-23 21:23:25 +08:00
|
|
|
static void err_free_sgl(struct scatterlist *sgl)
|
2013-07-12 21:50:57 +08:00
|
|
|
{
|
2018-11-23 21:23:25 +08:00
|
|
|
while (sgl) {
|
|
|
|
struct scatterlist *sg;
|
2013-07-12 21:50:57 +08:00
|
|
|
|
2018-11-23 21:23:25 +08:00
|
|
|
for (sg = sgl; !sg_is_chain(sg); sg++) {
|
|
|
|
kfree(sg_virt(sg));
|
|
|
|
if (sg_is_last(sg))
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
sg = sg_is_last(sg) ? NULL : sg_chain_ptr(sg);
|
|
|
|
free_page((unsigned long)sgl);
|
|
|
|
sgl = sg;
|
2013-07-12 21:50:57 +08:00
|
|
|
}
|
2018-11-23 21:23:25 +08:00
|
|
|
}
|
2013-07-12 21:50:57 +08:00
|
|
|
|
2018-11-23 21:23:25 +08:00
|
|
|
static void __err_print_to_sgl(struct drm_i915_error_state_buf *m,
|
|
|
|
struct i915_gpu_state *error)
|
|
|
|
{
|
|
|
|
struct drm_i915_error_object *obj;
|
|
|
|
struct timespec64 ts;
|
|
|
|
int i, j;
|
2018-11-03 00:12:12 +08:00
|
|
|
|
2017-02-15 00:46:11 +08:00
|
|
|
if (*error->error_msg)
|
|
|
|
err_printf(m, "%s\n", error->error_msg);
|
2019-01-03 18:12:45 +08:00
|
|
|
err_printf(m, "Kernel: %s %s\n",
|
|
|
|
init_utsname()->release,
|
|
|
|
init_utsname()->machine);
|
2018-01-17 23:48:53 +08:00
|
|
|
ts = ktime_to_timespec64(error->time);
|
|
|
|
err_printf(m, "Time: %lld s %ld us\n",
|
|
|
|
(s64)ts.tv_sec, ts.tv_nsec / NSEC_PER_USEC);
|
|
|
|
ts = ktime_to_timespec64(error->boottime);
|
|
|
|
err_printf(m, "Boottime: %lld s %ld us\n",
|
|
|
|
(s64)ts.tv_sec, ts.tv_nsec / NSEC_PER_USEC);
|
|
|
|
ts = ktime_to_timespec64(error->uptime);
|
|
|
|
err_printf(m, "Uptime: %lld s %ld us\n",
|
|
|
|
(s64)ts.tv_sec, ts.tv_nsec / NSEC_PER_USEC);
|
2018-04-30 15:52:59 +08:00
|
|
|
err_printf(m, "Epoch: %lu jiffies (%u HZ)\n", error->epoch, HZ);
|
|
|
|
err_printf(m, "Capture: %lu jiffies; %d ms ago, %d ms after epoch\n",
|
|
|
|
error->capture,
|
|
|
|
jiffies_to_msecs(jiffies - error->capture),
|
|
|
|
jiffies_to_msecs(error->capture - error->epoch));
|
2016-11-18 21:09:04 +08:00
|
|
|
|
2016-07-27 16:07:28 +08:00
|
|
|
for (i = 0; i < ARRAY_SIZE(error->engine); i++) {
|
2019-01-25 21:22:28 +08:00
|
|
|
if (!error->engine[i].context.pid)
|
|
|
|
continue;
|
|
|
|
|
drm/i915: Use time based guilty context banning
Currently, we accumulate each time a context hangs the GPU, offset
against the number of requests it submits, and if that score exceeds a
certain threshold, we ban that context from submitting any more requests
(cancelling any work in flight). In contrast, we use a simple timer on
the file, that if we see more than a 9 hangs faster than 60s apart in
total across all of its contexts, we will ban the client from creating
any more contexts. This leads to a confusing situation where the file
may be banned before the context, so lets use a simple timer scheme for
each.
If the context submits 3 hanging requests within a 120s period, declare
it forbidden to ever send more requests.
This has the advantage of not being easy to repair by simply sending
empty requests, but has the disadvantage that if the context is idle
then it is forgiven. However, if the context is idle, it is not
disrupting the system, but a hog can evade the request counting and
cause much more severe disruption to the system.
Updating ban_score from request retirement is dubious as the retirement
is purposely not in sync with request submission (i.e. we try and batch
retirement to reduce overhead and avoid latency on submission), which
leads to surprising situations where we can forgive a hang immediately
due to a backlog of requests from before the hang being retired
afterwards.
Signed-off-by: Chris Wilson <chris@chris-wilson.co.uk>
Cc: Mika Kuoppala <mika.kuoppala@intel.com>
Reviewed-by: Mika Kuoppala <mika.kuoppala@intel.com>
Link: https://patchwork.freedesktop.org/patch/msgid/20190219122215.8941-2-chris@chris-wilson.co.uk
2019-02-19 20:21:52 +08:00
|
|
|
err_printf(m, "Active process (on ring %s): %s [%d]\n",
|
2019-01-25 21:22:28 +08:00
|
|
|
engine_name(m->i915, i),
|
|
|
|
error->engine[i].context.comm,
|
drm/i915: Use time based guilty context banning
Currently, we accumulate each time a context hangs the GPU, offset
against the number of requests it submits, and if that score exceeds a
certain threshold, we ban that context from submitting any more requests
(cancelling any work in flight). In contrast, we use a simple timer on
the file, that if we see more than a 9 hangs faster than 60s apart in
total across all of its contexts, we will ban the client from creating
any more contexts. This leads to a confusing situation where the file
may be banned before the context, so lets use a simple timer scheme for
each.
If the context submits 3 hanging requests within a 120s period, declare
it forbidden to ever send more requests.
This has the advantage of not being easy to repair by simply sending
empty requests, but has the disadvantage that if the context is idle
then it is forgiven. However, if the context is idle, it is not
disrupting the system, but a hog can evade the request counting and
cause much more severe disruption to the system.
Updating ban_score from request retirement is dubious as the retirement
is purposely not in sync with request submission (i.e. we try and batch
retirement to reduce overhead and avoid latency on submission), which
leads to surprising situations where we can forgive a hang immediately
due to a backlog of requests from before the hang being retired
afterwards.
Signed-off-by: Chris Wilson <chris@chris-wilson.co.uk>
Cc: Mika Kuoppala <mika.kuoppala@intel.com>
Reviewed-by: Mika Kuoppala <mika.kuoppala@intel.com>
Link: https://patchwork.freedesktop.org/patch/msgid/20190219122215.8941-2-chris@chris-wilson.co.uk
2019-02-19 20:21:52 +08:00
|
|
|
error->engine[i].context.pid);
|
2014-02-25 23:11:24 +08:00
|
|
|
}
|
2014-02-25 23:11:27 +08:00
|
|
|
err_printf(m, "Reset count: %u\n", error->reset_count);
|
2014-02-25 23:11:28 +08:00
|
|
|
err_printf(m, "Suspend count: %u\n", error->suspend_count);
|
2016-12-01 20:49:55 +08:00
|
|
|
err_printf(m, "Platform: %s\n", intel_platform_name(error->device_info.platform));
|
2018-11-23 21:23:25 +08:00
|
|
|
err_print_pciid(m, m->i915);
|
2017-02-07 05:36:07 +08:00
|
|
|
|
2015-08-08 03:24:15 +08:00
|
|
|
err_printf(m, "IOMMU enabled?: %d\n", error->iommu);
|
2015-10-29 21:21:19 +08:00
|
|
|
|
2018-11-23 21:23:25 +08:00
|
|
|
if (HAS_CSR(m->i915)) {
|
|
|
|
struct intel_csr *csr = &m->i915->csr;
|
2015-10-29 21:21:19 +08:00
|
|
|
|
|
|
|
err_printf(m, "DMC loaded: %s\n",
|
|
|
|
yesno(csr->dmc_payload != NULL));
|
|
|
|
err_printf(m, "DMC fw version: %d.%d\n",
|
|
|
|
CSR_VERSION_MAJOR(csr->version),
|
|
|
|
CSR_VERSION_MINOR(csr->version));
|
|
|
|
}
|
|
|
|
|
2017-03-02 23:03:56 +08:00
|
|
|
err_printf(m, "GT awake: %s\n", yesno(error->awake));
|
2017-03-02 23:15:44 +08:00
|
|
|
err_printf(m, "RPM wakelock: %s\n", yesno(error->wakelock));
|
|
|
|
err_printf(m, "PM suspended: %s\n", yesno(error->suspended));
|
2013-07-12 21:50:57 +08:00
|
|
|
err_printf(m, "EIR: 0x%08x\n", error->eir);
|
|
|
|
err_printf(m, "IER: 0x%08x\n", error->ier);
|
2017-02-15 00:46:11 +08:00
|
|
|
for (i = 0; i < error->ngtier; i++)
|
|
|
|
err_printf(m, "GTIER[%d]: 0x%08x\n", i, error->gtier[i]);
|
2013-07-12 21:50:57 +08:00
|
|
|
err_printf(m, "PGTBL_ER: 0x%08x\n", error->pgtbl_er);
|
|
|
|
err_printf(m, "FORCEWAKE: 0x%08x\n", error->forcewake);
|
|
|
|
err_printf(m, "DERRMR: 0x%08x\n", error->derrmr);
|
|
|
|
err_printf(m, "CCID: 0x%08x\n", error->ccid);
|
|
|
|
|
2017-02-15 00:46:11 +08:00
|
|
|
for (i = 0; i < error->nfence; i++)
|
2013-07-12 21:50:57 +08:00
|
|
|
err_printf(m, " fence[%d] = %08llx\n", i, error->fence[i]);
|
|
|
|
|
2018-11-23 21:23:25 +08:00
|
|
|
if (INTEL_GEN(m->i915) >= 6) {
|
2013-07-12 21:50:57 +08:00
|
|
|
err_printf(m, "ERROR: 0x%08x\n", error->error);
|
2015-03-24 20:54:19 +08:00
|
|
|
|
2018-11-23 21:23:25 +08:00
|
|
|
if (INTEL_GEN(m->i915) >= 8)
|
2015-03-24 20:54:19 +08:00
|
|
|
err_printf(m, "FAULT_TLB_DATA: 0x%08x 0x%08x\n",
|
|
|
|
error->fault_data1, error->fault_data0);
|
|
|
|
|
2013-07-12 21:50:57 +08:00
|
|
|
err_printf(m, "DONE_REG: 0x%08x\n", error->done_reg);
|
|
|
|
}
|
|
|
|
|
drm/i915: replace IS_GEN<N> with IS_GEN(..., N)
Define IS_GEN() similarly to our IS_GEN_RANGE(). but use gen instead of
gen_mask to do the comparison. Now callers can pass then gen as a parameter,
so we don't require one macro for each gen.
The following spatch was used to convert the users of these macros:
@@
expression e;
@@
(
- IS_GEN2(e)
+ IS_GEN(e, 2)
|
- IS_GEN3(e)
+ IS_GEN(e, 3)
|
- IS_GEN4(e)
+ IS_GEN(e, 4)
|
- IS_GEN5(e)
+ IS_GEN(e, 5)
|
- IS_GEN6(e)
+ IS_GEN(e, 6)
|
- IS_GEN7(e)
+ IS_GEN(e, 7)
|
- IS_GEN8(e)
+ IS_GEN(e, 8)
|
- IS_GEN9(e)
+ IS_GEN(e, 9)
|
- IS_GEN10(e)
+ IS_GEN(e, 10)
|
- IS_GEN11(e)
+ IS_GEN(e, 11)
)
v2: use IS_GEN rather than GT_GEN and compare to info.gen rather than
using the bitmask
Signed-off-by: Lucas De Marchi <lucas.demarchi@intel.com>
Reviewed-by: Jani Nikula <jani.nikula@intel.com>
Signed-off-by: Rodrigo Vivi <rodrigo.vivi@intel.com>
Link: https://patchwork.freedesktop.org/patch/msgid/20181212181044.15886-2-lucas.demarchi@intel.com
2018-12-13 02:10:43 +08:00
|
|
|
if (IS_GEN(m->i915, 7))
|
2013-07-12 21:50:57 +08:00
|
|
|
err_printf(m, "ERR_INT: 0x%08x\n", error->err_int);
|
|
|
|
|
2016-07-27 16:07:28 +08:00
|
|
|
for (i = 0; i < ARRAY_SIZE(error->engine); i++) {
|
|
|
|
if (error->engine[i].engine_id != -1)
|
2018-04-30 15:52:59 +08:00
|
|
|
error_print_engine(m, &error->engine[i], error->epoch);
|
2016-07-27 16:07:28 +08:00
|
|
|
}
|
2013-07-12 21:50:57 +08:00
|
|
|
|
2016-08-15 17:48:41 +08:00
|
|
|
for (i = 0; i < ARRAY_SIZE(error->active_vm); i++) {
|
|
|
|
char buf[128];
|
|
|
|
int len, first = 1;
|
2014-08-13 03:05:47 +08:00
|
|
|
|
2016-08-15 17:48:41 +08:00
|
|
|
if (!error->active_vm[i])
|
|
|
|
break;
|
|
|
|
|
|
|
|
len = scnprintf(buf, sizeof(buf), "Active (");
|
|
|
|
for (j = 0; j < ARRAY_SIZE(error->engine); j++) {
|
|
|
|
if (error->engine[j].vm != error->active_vm[i])
|
|
|
|
continue;
|
|
|
|
|
|
|
|
len += scnprintf(buf + len, sizeof(buf), "%s%s",
|
|
|
|
first ? "" : ", ",
|
2018-11-23 21:23:25 +08:00
|
|
|
m->i915->engine[j]->name);
|
2016-08-15 17:48:41 +08:00
|
|
|
first = 0;
|
|
|
|
}
|
|
|
|
scnprintf(buf + len, sizeof(buf), ")");
|
|
|
|
print_error_buffers(m, buf,
|
2014-08-13 03:05:47 +08:00
|
|
|
error->active_bo[i],
|
|
|
|
error->active_bo_count[i]);
|
|
|
|
}
|
2013-07-12 21:50:57 +08:00
|
|
|
|
2016-08-15 17:48:41 +08:00
|
|
|
print_error_buffers(m, "Pinned (global)",
|
|
|
|
error->pinned_bo,
|
|
|
|
error->pinned_bo_count);
|
|
|
|
|
2016-07-27 16:07:28 +08:00
|
|
|
for (i = 0; i < ARRAY_SIZE(error->engine); i++) {
|
2017-02-15 00:46:11 +08:00
|
|
|
const struct drm_i915_error_engine *ee = &error->engine[i];
|
2016-07-27 16:07:28 +08:00
|
|
|
|
|
|
|
obj = ee->batchbuffer;
|
2014-02-25 23:11:24 +08:00
|
|
|
if (obj) {
|
2018-11-23 21:23:25 +08:00
|
|
|
err_puts(m, m->i915->engine[i]->name);
|
2017-01-29 17:24:33 +08:00
|
|
|
if (ee->context.pid)
|
drm/i915: Use time based guilty context banning
Currently, we accumulate each time a context hangs the GPU, offset
against the number of requests it submits, and if that score exceeds a
certain threshold, we ban that context from submitting any more requests
(cancelling any work in flight). In contrast, we use a simple timer on
the file, that if we see more than a 9 hangs faster than 60s apart in
total across all of its contexts, we will ban the client from creating
any more contexts. This leads to a confusing situation where the file
may be banned before the context, so lets use a simple timer scheme for
each.
If the context submits 3 hanging requests within a 120s period, declare
it forbidden to ever send more requests.
This has the advantage of not being easy to repair by simply sending
empty requests, but has the disadvantage that if the context is idle
then it is forgiven. However, if the context is idle, it is not
disrupting the system, but a hog can evade the request counting and
cause much more severe disruption to the system.
Updating ban_score from request retirement is dubious as the retirement
is purposely not in sync with request submission (i.e. we try and batch
retirement to reduce overhead and avoid latency on submission), which
leads to surprising situations where we can forgive a hang immediately
due to a backlog of requests from before the hang being retired
afterwards.
Signed-off-by: Chris Wilson <chris@chris-wilson.co.uk>
Cc: Mika Kuoppala <mika.kuoppala@intel.com>
Reviewed-by: Mika Kuoppala <mika.kuoppala@intel.com>
Link: https://patchwork.freedesktop.org/patch/msgid/20190219122215.8941-2-chris@chris-wilson.co.uk
2019-02-19 20:21:52 +08:00
|
|
|
err_printf(m, " (submitted by %s [%d], ctx %d [%d])",
|
2017-01-29 17:24:33 +08:00
|
|
|
ee->context.comm,
|
|
|
|
ee->context.pid,
|
|
|
|
ee->context.handle,
|
drm/i915: Use time based guilty context banning
Currently, we accumulate each time a context hangs the GPU, offset
against the number of requests it submits, and if that score exceeds a
certain threshold, we ban that context from submitting any more requests
(cancelling any work in flight). In contrast, we use a simple timer on
the file, that if we see more than a 9 hangs faster than 60s apart in
total across all of its contexts, we will ban the client from creating
any more contexts. This leads to a confusing situation where the file
may be banned before the context, so lets use a simple timer scheme for
each.
If the context submits 3 hanging requests within a 120s period, declare
it forbidden to ever send more requests.
This has the advantage of not being easy to repair by simply sending
empty requests, but has the disadvantage that if the context is idle
then it is forgiven. However, if the context is idle, it is not
disrupting the system, but a hog can evade the request counting and
cause much more severe disruption to the system.
Updating ban_score from request retirement is dubious as the retirement
is purposely not in sync with request submission (i.e. we try and batch
retirement to reduce overhead and avoid latency on submission), which
leads to surprising situations where we can forgive a hang immediately
due to a backlog of requests from before the hang being retired
afterwards.
Signed-off-by: Chris Wilson <chris@chris-wilson.co.uk>
Cc: Mika Kuoppala <mika.kuoppala@intel.com>
Reviewed-by: Mika Kuoppala <mika.kuoppala@intel.com>
Link: https://patchwork.freedesktop.org/patch/msgid/20190219122215.8941-2-chris@chris-wilson.co.uk
2019-02-19 20:21:52 +08:00
|
|
|
ee->context.hw_id);
|
2015-07-30 00:23:56 +08:00
|
|
|
err_printf(m, " --- gtt_offset = 0x%08x %08x\n",
|
|
|
|
upper_32_bits(obj->gtt_offset),
|
|
|
|
lower_32_bits(obj->gtt_offset));
|
2018-11-23 21:23:25 +08:00
|
|
|
print_error_obj(m, m->i915->engine[i], NULL, obj);
|
2013-07-12 21:50:57 +08:00
|
|
|
}
|
|
|
|
|
2017-04-15 17:39:02 +08:00
|
|
|
for (j = 0; j < ee->user_bo_count; j++)
|
2018-11-23 21:23:25 +08:00
|
|
|
print_error_obj(m, m->i915->engine[i],
|
2017-04-15 17:39:02 +08:00
|
|
|
"user", ee->user_bo[j]);
|
|
|
|
|
2016-07-27 16:07:28 +08:00
|
|
|
if (ee->num_requests) {
|
2013-07-12 21:50:57 +08:00
|
|
|
err_printf(m, "%s --- %d requests\n",
|
2018-11-23 21:23:25 +08:00
|
|
|
m->i915->engine[i]->name,
|
2016-07-27 16:07:28 +08:00
|
|
|
ee->num_requests);
|
2016-10-13 18:18:14 +08:00
|
|
|
for (j = 0; j < ee->num_requests; j++)
|
2018-04-30 15:52:59 +08:00
|
|
|
error_print_request(m, " ",
|
|
|
|
&ee->requests[j],
|
|
|
|
error->epoch);
|
2013-07-12 21:50:57 +08:00
|
|
|
}
|
|
|
|
|
2018-11-23 21:23:25 +08:00
|
|
|
print_error_obj(m, m->i915->engine[i],
|
2016-10-12 17:05:21 +08:00
|
|
|
"ringbuffer", ee->ringbuffer);
|
2013-07-12 21:50:57 +08:00
|
|
|
|
2018-11-23 21:23:25 +08:00
|
|
|
print_error_obj(m, m->i915->engine[i],
|
2016-10-12 17:05:21 +08:00
|
|
|
"HW Status", ee->hws_page);
|
2015-09-16 01:03:01 +08:00
|
|
|
|
2018-11-23 21:23:25 +08:00
|
|
|
print_error_obj(m, m->i915->engine[i],
|
2016-10-12 17:05:21 +08:00
|
|
|
"HW context", ee->ctx);
|
2014-01-24 06:40:36 +08:00
|
|
|
|
2018-11-23 21:23:25 +08:00
|
|
|
print_error_obj(m, m->i915->engine[i],
|
2016-10-12 17:05:21 +08:00
|
|
|
"WA context", ee->wa_ctx);
|
2016-03-01 19:24:36 +08:00
|
|
|
|
2018-11-23 21:23:25 +08:00
|
|
|
print_error_obj(m, m->i915->engine[i],
|
2016-10-12 17:05:21 +08:00
|
|
|
"WA batchbuffer", ee->wa_batchbuffer);
|
2017-11-27 06:09:01 +08:00
|
|
|
|
2018-11-23 21:23:25 +08:00
|
|
|
print_error_obj(m, m->i915->engine[i],
|
2017-11-27 06:09:01 +08:00
|
|
|
"NULL context", ee->default_state);
|
2013-07-12 21:50:57 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
if (error->overlay)
|
|
|
|
intel_overlay_print_error_state(m, error->overlay);
|
|
|
|
|
|
|
|
if (error->display)
|
2017-02-15 00:46:11 +08:00
|
|
|
intel_display_print_error_state(m, error->display);
|
2013-07-12 21:50:57 +08:00
|
|
|
|
2018-12-31 22:56:41 +08:00
|
|
|
err_print_capabilities(m, &error->device_info, &error->runtime_info,
|
|
|
|
&error->driver_caps);
|
2017-02-07 05:36:07 +08:00
|
|
|
err_print_params(m, &error->params);
|
2017-10-27 01:36:55 +08:00
|
|
|
err_print_uc(m, &error->uc);
|
2018-11-23 21:23:25 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
static int err_print_to_sgl(struct i915_gpu_state *error)
|
|
|
|
{
|
|
|
|
struct drm_i915_error_state_buf m;
|
|
|
|
|
|
|
|
if (IS_ERR(error))
|
|
|
|
return PTR_ERR(error);
|
|
|
|
|
|
|
|
if (READ_ONCE(error->sgl))
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
memset(&m, 0, sizeof(m));
|
|
|
|
m.i915 = error->i915;
|
|
|
|
|
|
|
|
__err_print_to_sgl(&m, error);
|
|
|
|
|
|
|
|
if (m.buf) {
|
|
|
|
__sg_set_buf(m.cur++, m.buf, m.bytes, m.iter);
|
|
|
|
m.bytes = 0;
|
|
|
|
m.buf = NULL;
|
|
|
|
}
|
|
|
|
if (m.cur) {
|
|
|
|
GEM_BUG_ON(m.end < m.cur);
|
|
|
|
sg_mark_end(m.cur - 1);
|
|
|
|
}
|
|
|
|
GEM_BUG_ON(m.sgl && !m.cur);
|
|
|
|
|
|
|
|
if (m.err) {
|
|
|
|
err_free_sgl(m.sgl);
|
|
|
|
return m.err;
|
|
|
|
}
|
2017-02-07 05:36:07 +08:00
|
|
|
|
2018-11-23 21:23:25 +08:00
|
|
|
if (cmpxchg(&error->sgl, NULL, m.sgl))
|
|
|
|
err_free_sgl(m.sgl);
|
2013-07-12 21:50:57 +08:00
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2018-11-23 21:23:25 +08:00
|
|
|
ssize_t i915_gpu_state_copy_to_buffer(struct i915_gpu_state *error,
|
|
|
|
char *buf, loff_t off, size_t rem)
|
2013-07-12 21:50:57 +08:00
|
|
|
{
|
2018-11-23 21:23:25 +08:00
|
|
|
struct scatterlist *sg;
|
|
|
|
size_t count;
|
|
|
|
loff_t pos;
|
|
|
|
int err;
|
2013-07-12 21:50:57 +08:00
|
|
|
|
2018-11-23 21:23:25 +08:00
|
|
|
if (!error || !rem)
|
|
|
|
return 0;
|
2013-07-12 21:50:57 +08:00
|
|
|
|
2018-11-23 21:23:25 +08:00
|
|
|
err = err_print_to_sgl(error);
|
|
|
|
if (err)
|
|
|
|
return err;
|
2013-07-12 21:50:57 +08:00
|
|
|
|
2018-11-23 21:23:25 +08:00
|
|
|
sg = READ_ONCE(error->fit);
|
|
|
|
if (!sg || off < sg->dma_address)
|
|
|
|
sg = error->sgl;
|
|
|
|
if (!sg)
|
|
|
|
return 0;
|
2013-07-12 21:50:57 +08:00
|
|
|
|
2018-11-23 21:23:25 +08:00
|
|
|
pos = sg->dma_address;
|
|
|
|
count = 0;
|
|
|
|
do {
|
|
|
|
size_t len, start;
|
|
|
|
|
|
|
|
if (sg_is_chain(sg)) {
|
|
|
|
sg = sg_chain_ptr(sg);
|
|
|
|
GEM_BUG_ON(sg_is_chain(sg));
|
|
|
|
}
|
2013-07-12 21:50:57 +08:00
|
|
|
|
2018-11-23 21:23:25 +08:00
|
|
|
len = sg->length;
|
|
|
|
if (pos + len <= off) {
|
|
|
|
pos += len;
|
|
|
|
continue;
|
|
|
|
}
|
2013-07-12 21:50:57 +08:00
|
|
|
|
2018-11-23 21:23:25 +08:00
|
|
|
start = sg->offset;
|
|
|
|
if (pos < off) {
|
|
|
|
GEM_BUG_ON(off - pos > len);
|
|
|
|
len -= off - pos;
|
|
|
|
start += off - pos;
|
|
|
|
pos = off;
|
|
|
|
}
|
|
|
|
|
|
|
|
len = min(len, rem);
|
|
|
|
GEM_BUG_ON(!len || len > sg->length);
|
|
|
|
|
|
|
|
memcpy(buf, page_address(sg_page(sg)) + start, len);
|
|
|
|
|
|
|
|
count += len;
|
|
|
|
pos += len;
|
|
|
|
|
|
|
|
buf += len;
|
|
|
|
rem -= len;
|
|
|
|
if (!rem) {
|
|
|
|
WRITE_ONCE(error->fit, sg);
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
} while (!sg_is_last(sg++));
|
|
|
|
|
|
|
|
return count;
|
2013-07-12 21:50:57 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
static void i915_error_object_free(struct drm_i915_error_object *obj)
|
|
|
|
{
|
|
|
|
int page;
|
|
|
|
|
|
|
|
if (obj == NULL)
|
|
|
|
return;
|
|
|
|
|
|
|
|
for (page = 0; page < obj->page_count; page++)
|
2016-10-12 17:05:20 +08:00
|
|
|
free_page((unsigned long)obj->pages[page]);
|
2013-07-12 21:50:57 +08:00
|
|
|
|
|
|
|
kfree(obj);
|
|
|
|
}
|
|
|
|
|
2017-02-22 00:26:19 +08:00
|
|
|
|
2017-10-27 01:36:57 +08:00
|
|
|
static void cleanup_params(struct i915_gpu_state *error)
|
|
|
|
{
|
2018-12-27 22:33:38 +08:00
|
|
|
i915_params_free(&error->params);
|
2017-10-27 01:36:57 +08:00
|
|
|
}
|
|
|
|
|
2017-10-27 01:36:55 +08:00
|
|
|
static void cleanup_uc_state(struct i915_gpu_state *error)
|
|
|
|
{
|
|
|
|
struct i915_error_uc *error_uc = &error->uc;
|
|
|
|
|
|
|
|
kfree(error_uc->guc_fw.path);
|
|
|
|
kfree(error_uc->huc_fw.path);
|
2017-10-27 01:36:56 +08:00
|
|
|
i915_error_object_free(error_uc->guc_log);
|
2017-10-27 01:36:55 +08:00
|
|
|
}
|
|
|
|
|
2017-02-15 00:46:11 +08:00
|
|
|
void __i915_gpu_state_free(struct kref *error_ref)
|
2013-07-12 21:50:57 +08:00
|
|
|
{
|
2017-02-15 00:46:11 +08:00
|
|
|
struct i915_gpu_state *error =
|
|
|
|
container_of(error_ref, typeof(*error), ref);
|
2017-04-15 17:39:02 +08:00
|
|
|
long i, j;
|
2013-07-12 21:50:57 +08:00
|
|
|
|
2016-07-27 16:07:28 +08:00
|
|
|
for (i = 0; i < ARRAY_SIZE(error->engine); i++) {
|
|
|
|
struct drm_i915_error_engine *ee = &error->engine[i];
|
|
|
|
|
2017-04-15 17:39:02 +08:00
|
|
|
for (j = 0; j < ee->user_bo_count; j++)
|
|
|
|
i915_error_object_free(ee->user_bo[j]);
|
|
|
|
kfree(ee->user_bo);
|
|
|
|
|
2016-07-27 16:07:28 +08:00
|
|
|
i915_error_object_free(ee->batchbuffer);
|
|
|
|
i915_error_object_free(ee->wa_batchbuffer);
|
|
|
|
i915_error_object_free(ee->ringbuffer);
|
|
|
|
i915_error_object_free(ee->hws_page);
|
|
|
|
i915_error_object_free(ee->ctx);
|
|
|
|
i915_error_object_free(ee->wa_ctx);
|
|
|
|
|
|
|
|
kfree(ee->requests);
|
2013-07-12 21:50:57 +08:00
|
|
|
}
|
|
|
|
|
2016-08-15 17:48:41 +08:00
|
|
|
for (i = 0; i < ARRAY_SIZE(error->active_bo); i++)
|
2015-03-20 17:41:03 +08:00
|
|
|
kfree(error->active_bo[i]);
|
|
|
|
kfree(error->pinned_bo);
|
2016-08-15 17:48:41 +08:00
|
|
|
|
2013-07-12 21:50:57 +08:00
|
|
|
kfree(error->overlay);
|
|
|
|
kfree(error->display);
|
2017-02-22 00:26:19 +08:00
|
|
|
|
2017-10-27 01:36:57 +08:00
|
|
|
cleanup_params(error);
|
2017-10-27 01:36:55 +08:00
|
|
|
cleanup_uc_state(error);
|
|
|
|
|
2018-11-23 21:23:25 +08:00
|
|
|
err_free_sgl(error->sgl);
|
2013-07-12 21:50:57 +08:00
|
|
|
kfree(error);
|
|
|
|
}
|
|
|
|
|
|
|
|
static struct drm_i915_error_object *
|
2016-10-12 17:05:20 +08:00
|
|
|
i915_error_object_create(struct drm_i915_private *i915,
|
2016-08-15 17:49:06 +08:00
|
|
|
struct i915_vma *vma)
|
2013-07-12 21:50:57 +08:00
|
|
|
{
|
2016-10-12 17:05:20 +08:00
|
|
|
struct i915_ggtt *ggtt = &i915->ggtt;
|
|
|
|
const u64 slot = ggtt->error_capture.start;
|
2013-07-12 21:50:57 +08:00
|
|
|
struct drm_i915_error_object *dst;
|
2016-12-06 20:40:51 +08:00
|
|
|
struct compress compress;
|
2016-10-12 17:05:20 +08:00
|
|
|
unsigned long num_pages;
|
|
|
|
struct sgt_iter iter;
|
|
|
|
dma_addr_t dma;
|
2018-10-03 16:24:22 +08:00
|
|
|
int ret;
|
2013-07-12 21:50:57 +08:00
|
|
|
|
2019-01-10 19:15:22 +08:00
|
|
|
if (!vma || !vma->pages)
|
2016-08-15 17:49:06 +08:00
|
|
|
return NULL;
|
|
|
|
|
2016-10-12 17:05:20 +08:00
|
|
|
num_pages = min_t(u64, vma->size, vma->obj->base.size) >> PAGE_SHIFT;
|
2016-10-12 17:05:22 +08:00
|
|
|
num_pages = DIV_ROUND_UP(10 * num_pages, 8); /* worstcase zlib growth */
|
2016-10-12 17:05:20 +08:00
|
|
|
dst = kmalloc(sizeof(*dst) + num_pages * sizeof(u32 *),
|
|
|
|
GFP_ATOMIC | __GFP_NOWARN);
|
2016-08-15 17:49:06 +08:00
|
|
|
if (!dst)
|
2013-07-12 21:50:57 +08:00
|
|
|
return NULL;
|
|
|
|
|
2016-08-15 17:49:09 +08:00
|
|
|
dst->gtt_offset = vma->node.start;
|
|
|
|
dst->gtt_size = vma->node.size;
|
2018-10-03 16:24:22 +08:00
|
|
|
dst->num_pages = num_pages;
|
2016-10-12 17:05:20 +08:00
|
|
|
dst->page_count = 0;
|
2016-10-12 17:05:22 +08:00
|
|
|
dst->unused = 0;
|
|
|
|
|
2016-12-06 20:40:51 +08:00
|
|
|
if (!compress_init(&compress)) {
|
2016-10-12 17:05:22 +08:00
|
|
|
kfree(dst);
|
|
|
|
return NULL;
|
|
|
|
}
|
2016-08-15 17:49:09 +08:00
|
|
|
|
2018-10-03 16:24:22 +08:00
|
|
|
ret = -EINVAL;
|
2016-10-12 17:05:20 +08:00
|
|
|
for_each_sgt_dma(dma, iter, vma->pages) {
|
|
|
|
void __iomem *s;
|
2014-08-13 03:05:48 +08:00
|
|
|
|
2018-06-05 23:37:58 +08:00
|
|
|
ggtt->vm.insert_page(&ggtt->vm, dma, slot, I915_CACHE_NONE, 0);
|
2014-08-13 03:05:48 +08:00
|
|
|
|
2017-12-11 23:18:20 +08:00
|
|
|
s = io_mapping_map_atomic_wc(&ggtt->iomap, slot);
|
2016-12-06 20:40:51 +08:00
|
|
|
ret = compress_page(&compress, (void __force *)s, dst);
|
2016-10-12 17:05:20 +08:00
|
|
|
io_mapping_unmap_atomic(s);
|
|
|
|
if (ret)
|
2018-10-03 16:24:22 +08:00
|
|
|
break;
|
2013-07-12 21:50:57 +08:00
|
|
|
}
|
|
|
|
|
2018-10-03 16:24:22 +08:00
|
|
|
if (ret || compress_flush(&compress, dst)) {
|
|
|
|
while (dst->page_count--)
|
|
|
|
free_page((unsigned long)dst->pages[dst->page_count]);
|
|
|
|
kfree(dst);
|
|
|
|
dst = NULL;
|
|
|
|
}
|
2016-10-12 17:05:20 +08:00
|
|
|
|
2016-12-06 20:40:51 +08:00
|
|
|
compress_fini(&compress, dst);
|
2016-10-12 17:05:20 +08:00
|
|
|
return dst;
|
2013-07-12 21:50:57 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
static void capture_bo(struct drm_i915_error_buffer *err,
|
2014-08-13 03:05:47 +08:00
|
|
|
struct i915_vma *vma)
|
2013-07-12 21:50:57 +08:00
|
|
|
{
|
2014-08-13 03:05:47 +08:00
|
|
|
struct drm_i915_gem_object *obj = vma->obj;
|
|
|
|
|
2013-07-12 21:50:57 +08:00
|
|
|
err->size = obj->base.size;
|
|
|
|
err->name = obj->base.name;
|
2016-08-04 14:52:31 +08:00
|
|
|
|
2014-08-13 03:05:47 +08:00
|
|
|
err->gtt_offset = vma->node.start;
|
2018-02-16 20:43:38 +08:00
|
|
|
err->read_domains = obj->read_domains;
|
|
|
|
err->write_domain = obj->write_domain;
|
2016-08-19 00:17:00 +08:00
|
|
|
err->fence_reg = vma->fence ? vma->fence->id : -1;
|
2016-08-05 17:14:23 +08:00
|
|
|
err->tiling = i915_gem_object_get_tiling(obj);
|
2016-10-28 20:58:35 +08:00
|
|
|
err->dirty = obj->mm.dirty;
|
|
|
|
err->purgeable = obj->mm.madv != I915_MADV_WILLNEED;
|
drm/i915: Introduce mapping of user pages into video memory (userptr) ioctl
By exporting the ability to map user address and inserting PTEs
representing their backing pages into the GTT, we can exploit UMA in order
to utilize normal application data as a texture source or even as a
render target (depending upon the capabilities of the chipset). This has
a number of uses, with zero-copy downloads to the GPU and efficient
readback making the intermixed streaming of CPU and GPU operations
fairly efficient. This ability has many widespread implications from
faster rendering of client-side software rasterisers (chromium),
mitigation of stalls due to read back (firefox) and to faster pipelining
of texture data (such as pixel buffer objects in GL or data blobs in CL).
v2: Compile with CONFIG_MMU_NOTIFIER
v3: We can sleep while performing invalidate-range, which we can utilise
to drop our page references prior to the kernel manipulating the vma
(for either discard or cloning) and so protect normal users.
v4: Only run the invalidate notifier if the range intercepts the bo.
v5: Prevent userspace from attempting to GTT mmap non-page aligned buffers
v6: Recheck after reacquire mutex for lost mmu.
v7: Fix implicit padding of ioctl struct by rounding to next 64bit boundary.
v8: Fix rebasing error after forwarding porting the back port.
v9: Limit the userptr to page aligned entries. We now expect userspace
to handle all the offset-in-page adjustments itself.
v10: Prevent vma from being copied across fork to avoid issues with cow.
v11: Drop vma behaviour changes -- locking is nigh on impossible.
Use a worker to load user pages to avoid lock inversions.
v12: Use get_task_mm()/mmput() for correct refcounting of mm.
v13: Use a worker to release the mmu_notifier to avoid lock inversion
v14: Decouple mmu_notifier from struct_mutex using a custom mmu_notifer
with its own locking and tree of objects for each mm/mmu_notifier.
v15: Prevent overlapping userptr objects, and invalidate all objects
within the mmu_notifier range
v16: Fix a typo for iterating over multiple objects in the range and
rearrange error path to destroy the mmu_notifier locklessly.
Also close a race between invalidate_range and the get_pages_worker.
v17: Close a race between get_pages_worker/invalidate_range and fresh
allocations of the same userptr range - and notice that
struct_mutex was presumed to be held when during creation it wasn't.
v18: Sigh. Fix the refactor of st_set_pages() to allocate enough memory
for the struct sg_table and to clear it before reporting an error.
v19: Always error out on read-only userptr requests as we don't have the
hardware infrastructure to support them at the moment.
v20: Refuse to implement read-only support until we have the required
infrastructure - but reserve the bit in flags for future use.
v21: use_mm() is not required for get_user_pages(). It is only meant to
be used to fix up the kernel thread's current->mm for use with
copy_user().
v22: Use sg_alloc_table_from_pages for that chunky feeling
v23: Export a function for sanity checking dma-buf rather than encode
userptr details elsewhere, and clean up comments based on
suggestions by Bradley.
Signed-off-by: Chris Wilson <chris@chris-wilson.co.uk>
Cc: Tvrtko Ursulin <tvrtko.ursulin@linux.intel.com>
Cc: "Gong, Zhipeng" <zhipeng.gong@intel.com>
Cc: Akash Goel <akash.goel@intel.com>
Cc: "Volkin, Bradley D" <bradley.d.volkin@intel.com>
Reviewed-by: Tvrtko Ursulin <tvrtko.ursulin@linux.intel.com>
Reviewed-by: Brad Volkin <bradley.d.volkin@intel.com>
[danvet: Frob ioctl allocation to pick the next one - will cause a bit
of fuss with create2 apparently, but such are the rules.]
[danvet2: oops, forgot to git add after manual patch application]
[danvet3: Appease sparse.]
Signed-off-by: Daniel Vetter <daniel.vetter@ffwll.ch>
2014-05-16 21:22:37 +08:00
|
|
|
err->userptr = obj->userptr.mm != NULL;
|
2013-07-12 21:50:57 +08:00
|
|
|
err->cache_level = obj->cache_level;
|
|
|
|
}
|
|
|
|
|
2016-08-15 17:48:41 +08:00
|
|
|
static u32 capture_error_bo(struct drm_i915_error_buffer *err,
|
|
|
|
int count, struct list_head *head,
|
2019-01-28 18:23:52 +08:00
|
|
|
unsigned int flags)
|
|
|
|
#define ACTIVE_ONLY BIT(0)
|
|
|
|
#define PINNED_ONLY BIT(1)
|
2013-07-12 21:50:57 +08:00
|
|
|
{
|
2013-08-01 08:00:14 +08:00
|
|
|
struct i915_vma *vma;
|
2013-07-12 21:50:57 +08:00
|
|
|
int i = 0;
|
|
|
|
|
2016-02-26 19:03:19 +08:00
|
|
|
list_for_each_entry(vma, head, vm_link) {
|
2018-06-07 23:40:45 +08:00
|
|
|
if (!vma->obj)
|
|
|
|
continue;
|
|
|
|
|
2019-01-28 18:23:52 +08:00
|
|
|
if (flags & ACTIVE_ONLY && !i915_vma_is_active(vma))
|
|
|
|
continue;
|
|
|
|
|
|
|
|
if (flags & PINNED_ONLY && !i915_vma_is_pinned(vma))
|
2016-08-15 17:48:41 +08:00
|
|
|
continue;
|
|
|
|
|
2014-08-13 03:05:47 +08:00
|
|
|
capture_bo(err++, vma);
|
2013-07-12 21:50:57 +08:00
|
|
|
if (++i == count)
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
return i;
|
|
|
|
}
|
|
|
|
|
2019-01-25 21:22:28 +08:00
|
|
|
/*
|
|
|
|
* Generate a semi-unique error code. The code is not meant to have meaning, The
|
2014-02-04 20:18:55 +08:00
|
|
|
* code's only purpose is to try to prevent false duplicated bug reports by
|
|
|
|
* grossly estimating a GPU error state.
|
|
|
|
*
|
|
|
|
* TODO Ideally, hashing the batchbuffer would be a very nice way to determine
|
|
|
|
* the hang if we could strip the GTT offset information from it.
|
|
|
|
*
|
|
|
|
* It's only a small step better than a random number in its current form.
|
|
|
|
*/
|
2019-01-25 21:22:28 +08:00
|
|
|
static u32 i915_error_generate_code(struct i915_gpu_state *error,
|
|
|
|
unsigned long engine_mask)
|
2014-02-04 20:18:55 +08:00
|
|
|
{
|
2019-01-25 21:22:28 +08:00
|
|
|
/*
|
|
|
|
* IPEHR would be an ideal way to detect errors, as it's the gross
|
2014-02-04 20:18:55 +08:00
|
|
|
* measure of "the command that hung." However, has some very common
|
|
|
|
* synchronization commands which almost always appear in the case
|
|
|
|
* strictly a client bug. Use instdone to differentiate those some.
|
|
|
|
*/
|
2019-01-25 21:22:28 +08:00
|
|
|
if (engine_mask) {
|
|
|
|
struct drm_i915_error_engine *ee =
|
|
|
|
&error->engine[ffs(engine_mask)];
|
2014-02-25 23:11:25 +08:00
|
|
|
|
2019-01-25 21:22:28 +08:00
|
|
|
return ee->ipehr ^ ee->instdone.instdone;
|
2014-02-25 23:11:25 +08:00
|
|
|
}
|
2014-02-04 20:18:55 +08:00
|
|
|
|
2019-01-25 21:22:28 +08:00
|
|
|
return 0;
|
2014-02-04 20:18:55 +08:00
|
|
|
}
|
|
|
|
|
2018-03-06 06:21:21 +08:00
|
|
|
static void gem_record_fences(struct i915_gpu_state *error)
|
2013-07-12 21:50:57 +08:00
|
|
|
{
|
2018-03-06 06:21:21 +08:00
|
|
|
struct drm_i915_private *dev_priv = error->i915;
|
2013-07-12 21:50:57 +08:00
|
|
|
int i;
|
|
|
|
|
2017-02-15 00:46:11 +08:00
|
|
|
if (INTEL_GEN(dev_priv) >= 6) {
|
2014-12-04 22:48:10 +08:00
|
|
|
for (i = 0; i < dev_priv->num_fence_regs; i++)
|
2017-02-15 00:46:11 +08:00
|
|
|
error->fence[i] = I915_READ64(FENCE_REG_GEN6_LO(i));
|
|
|
|
} else if (INTEL_GEN(dev_priv) >= 4) {
|
2015-09-21 23:05:14 +08:00
|
|
|
for (i = 0; i < dev_priv->num_fence_regs; i++)
|
|
|
|
error->fence[i] = I915_READ64(FENCE_REG_965_LO(i));
|
2017-02-15 00:46:11 +08:00
|
|
|
} else {
|
2015-09-21 23:05:14 +08:00
|
|
|
for (i = 0; i < dev_priv->num_fence_regs; i++)
|
2017-02-15 00:46:11 +08:00
|
|
|
error->fence[i] = I915_READ(FENCE_REG(i));
|
2015-09-21 23:05:14 +08:00
|
|
|
}
|
2017-02-15 00:46:11 +08:00
|
|
|
error->nfence = i;
|
2013-07-12 21:50:57 +08:00
|
|
|
}
|
|
|
|
|
2016-07-27 16:07:28 +08:00
|
|
|
static void gen6_record_semaphore_state(struct intel_engine_cs *engine,
|
|
|
|
struct drm_i915_error_engine *ee)
|
2014-07-01 00:53:40 +08:00
|
|
|
{
|
2016-07-27 16:07:28 +08:00
|
|
|
struct drm_i915_private *dev_priv = engine->i915;
|
|
|
|
|
|
|
|
ee->semaphore_mboxes[0] = I915_READ(RING_SYNC_0(engine->mmio_base));
|
|
|
|
ee->semaphore_mboxes[1] = I915_READ(RING_SYNC_1(engine->mmio_base));
|
2016-10-28 20:58:53 +08:00
|
|
|
if (HAS_VEBOX(dev_priv))
|
2016-07-27 16:07:28 +08:00
|
|
|
ee->semaphore_mboxes[2] =
|
2016-03-16 19:00:37 +08:00
|
|
|
I915_READ(RING_SYNC_2(engine->mmio_base));
|
2014-07-01 00:53:40 +08:00
|
|
|
}
|
|
|
|
|
2017-02-15 00:46:11 +08:00
|
|
|
static void error_record_engine_registers(struct i915_gpu_state *error,
|
2016-07-27 16:07:28 +08:00
|
|
|
struct intel_engine_cs *engine,
|
|
|
|
struct drm_i915_error_engine *ee)
|
2013-07-12 21:50:57 +08:00
|
|
|
{
|
2016-07-27 16:07:28 +08:00
|
|
|
struct drm_i915_private *dev_priv = engine->i915;
|
|
|
|
|
2016-05-06 22:40:21 +08:00
|
|
|
if (INTEL_GEN(dev_priv) >= 6) {
|
2016-07-27 16:07:28 +08:00
|
|
|
ee->rc_psmi = I915_READ(RING_PSMI_CTL(engine->mmio_base));
|
2017-11-14 01:36:28 +08:00
|
|
|
if (INTEL_GEN(dev_priv) >= 8) {
|
|
|
|
ee->fault_reg = I915_READ(GEN8_RING_FAULT_REG);
|
|
|
|
} else {
|
2016-07-27 16:07:28 +08:00
|
|
|
gen6_record_semaphore_state(engine, ee);
|
2017-11-14 01:36:28 +08:00
|
|
|
ee->fault_reg = I915_READ(RING_FAULT_REG(engine));
|
|
|
|
}
|
2013-08-13 07:53:04 +08:00
|
|
|
}
|
|
|
|
|
2016-05-06 22:40:21 +08:00
|
|
|
if (INTEL_GEN(dev_priv) >= 4) {
|
2016-07-27 16:07:28 +08:00
|
|
|
ee->faddr = I915_READ(RING_DMA_FADD(engine->mmio_base));
|
|
|
|
ee->ipeir = I915_READ(RING_IPEIR(engine->mmio_base));
|
|
|
|
ee->ipehr = I915_READ(RING_IPEHR(engine->mmio_base));
|
|
|
|
ee->instps = I915_READ(RING_INSTPS(engine->mmio_base));
|
|
|
|
ee->bbaddr = I915_READ(RING_BBADDR(engine->mmio_base));
|
2016-05-06 22:40:21 +08:00
|
|
|
if (INTEL_GEN(dev_priv) >= 8) {
|
2016-07-27 16:07:28 +08:00
|
|
|
ee->faddr |= (u64) I915_READ(RING_DMA_FADD_UDW(engine->mmio_base)) << 32;
|
|
|
|
ee->bbaddr |= (u64) I915_READ(RING_BBADDR_UDW(engine->mmio_base)) << 32;
|
2014-04-02 07:31:07 +08:00
|
|
|
}
|
2016-07-27 16:07:28 +08:00
|
|
|
ee->bbstate = I915_READ(RING_BBSTATE(engine->mmio_base));
|
2013-07-12 21:50:57 +08:00
|
|
|
} else {
|
2016-07-27 16:07:28 +08:00
|
|
|
ee->faddr = I915_READ(DMA_FADD_I8XX);
|
|
|
|
ee->ipeir = I915_READ(IPEIR);
|
|
|
|
ee->ipehr = I915_READ(IPEHR);
|
2013-07-12 21:50:57 +08:00
|
|
|
}
|
|
|
|
|
2016-10-12 17:05:17 +08:00
|
|
|
intel_engine_get_instdone(engine, &ee->instdone);
|
2016-09-20 21:54:32 +08:00
|
|
|
|
2016-07-27 16:07:28 +08:00
|
|
|
ee->instpm = I915_READ(RING_INSTPM(engine->mmio_base));
|
2016-08-03 05:50:21 +08:00
|
|
|
ee->acthd = intel_engine_get_active_head(engine);
|
2016-07-27 16:07:28 +08:00
|
|
|
ee->start = I915_READ_START(engine);
|
|
|
|
ee->head = I915_READ_HEAD(engine);
|
|
|
|
ee->tail = I915_READ_TAIL(engine);
|
|
|
|
ee->ctl = I915_READ_CTL(engine);
|
2016-08-15 17:49:11 +08:00
|
|
|
if (INTEL_GEN(dev_priv) > 2)
|
|
|
|
ee->mode = I915_READ_MODE(engine);
|
2013-07-12 21:50:57 +08:00
|
|
|
|
2016-08-18 03:30:56 +08:00
|
|
|
if (!HWS_NEEDS_PHYSICAL(dev_priv)) {
|
drm/i915: Type safe register read/write
Make I915_READ and I915_WRITE more type safe by wrapping the register
offset in a struct. This should eliminate most of the fumbles we've had
with misplaced parens.
This only takes care of normal mmio registers. We could extend the idea
to other register types and define each with its own struct. That way
you wouldn't be able to accidentally pass the wrong thing to a specific
register access function.
The gpio_reg setup is probably the ugliest thing left. But I figure I'd
just leave it for now, and wait for some divine inspiration to strike
before making it nice.
As for the generated code, it's actually a bit better sometimes. Eg.
looking at i915_irq_handler(), we can see the following change:
lea 0x70024(%rdx,%rax,1),%r9d
mov $0x1,%edx
- movslq %r9d,%r9
- mov %r9,%rsi
- mov %r9,-0x58(%rbp)
- callq *0xd8(%rbx)
+ mov %r9d,%esi
+ mov %r9d,-0x48(%rbp)
callq *0xd8(%rbx)
So previously gcc thought the register offset might be signed and
decided to sign extend it, just in case. The rest appears to be
mostly just minor shuffling of instructions.
v2: i915_mmio_reg_{offset,equal,valid}() helpers added
s/_REG/_MMIO/ in the register defines
mo more switch statements left to worry about
ring_emit stuff got sorted in a prep patch
cmd parser, lrc context and w/a batch buildup also in prep patch
vgpu stuff cleaned up and moved to a prep patch
all other unrelated changes split out
v3: Rebased due to BXT DSI/BLC, MOCS, etc.
v4: Rebased due to churn, s/i915_mmio_reg_t/i915_reg_t/
Signed-off-by: Ville Syrjälä <ville.syrjala@linux.intel.com>
Reviewed-by: Chris Wilson <chris@chris-wilson.co.uk>
Link: http://patchwork.freedesktop.org/patch/msgid/1447853606-2751-1-git-send-email-ville.syrjala@linux.intel.com
2015-11-18 21:33:26 +08:00
|
|
|
i915_reg_t mmio;
|
2014-01-24 06:40:36 +08:00
|
|
|
|
drm/i915: replace IS_GEN<N> with IS_GEN(..., N)
Define IS_GEN() similarly to our IS_GEN_RANGE(). but use gen instead of
gen_mask to do the comparison. Now callers can pass then gen as a parameter,
so we don't require one macro for each gen.
The following spatch was used to convert the users of these macros:
@@
expression e;
@@
(
- IS_GEN2(e)
+ IS_GEN(e, 2)
|
- IS_GEN3(e)
+ IS_GEN(e, 3)
|
- IS_GEN4(e)
+ IS_GEN(e, 4)
|
- IS_GEN5(e)
+ IS_GEN(e, 5)
|
- IS_GEN6(e)
+ IS_GEN(e, 6)
|
- IS_GEN7(e)
+ IS_GEN(e, 7)
|
- IS_GEN8(e)
+ IS_GEN(e, 8)
|
- IS_GEN9(e)
+ IS_GEN(e, 9)
|
- IS_GEN10(e)
+ IS_GEN(e, 10)
|
- IS_GEN11(e)
+ IS_GEN(e, 11)
)
v2: use IS_GEN rather than GT_GEN and compare to info.gen rather than
using the bitmask
Signed-off-by: Lucas De Marchi <lucas.demarchi@intel.com>
Reviewed-by: Jani Nikula <jani.nikula@intel.com>
Signed-off-by: Rodrigo Vivi <rodrigo.vivi@intel.com>
Link: https://patchwork.freedesktop.org/patch/msgid/20181212181044.15886-2-lucas.demarchi@intel.com
2018-12-13 02:10:43 +08:00
|
|
|
if (IS_GEN(dev_priv, 7)) {
|
2016-03-16 19:00:37 +08:00
|
|
|
switch (engine->id) {
|
2014-01-24 06:40:36 +08:00
|
|
|
default:
|
|
|
|
case RCS:
|
|
|
|
mmio = RENDER_HWS_PGA_GEN7;
|
|
|
|
break;
|
|
|
|
case BCS:
|
|
|
|
mmio = BLT_HWS_PGA_GEN7;
|
|
|
|
break;
|
|
|
|
case VCS:
|
|
|
|
mmio = BSD_HWS_PGA_GEN7;
|
|
|
|
break;
|
|
|
|
case VECS:
|
|
|
|
mmio = VEBOX_HWS_PGA_GEN7;
|
|
|
|
break;
|
|
|
|
}
|
drm/i915: replace IS_GEN<N> with IS_GEN(..., N)
Define IS_GEN() similarly to our IS_GEN_RANGE(). but use gen instead of
gen_mask to do the comparison. Now callers can pass then gen as a parameter,
so we don't require one macro for each gen.
The following spatch was used to convert the users of these macros:
@@
expression e;
@@
(
- IS_GEN2(e)
+ IS_GEN(e, 2)
|
- IS_GEN3(e)
+ IS_GEN(e, 3)
|
- IS_GEN4(e)
+ IS_GEN(e, 4)
|
- IS_GEN5(e)
+ IS_GEN(e, 5)
|
- IS_GEN6(e)
+ IS_GEN(e, 6)
|
- IS_GEN7(e)
+ IS_GEN(e, 7)
|
- IS_GEN8(e)
+ IS_GEN(e, 8)
|
- IS_GEN9(e)
+ IS_GEN(e, 9)
|
- IS_GEN10(e)
+ IS_GEN(e, 10)
|
- IS_GEN11(e)
+ IS_GEN(e, 11)
)
v2: use IS_GEN rather than GT_GEN and compare to info.gen rather than
using the bitmask
Signed-off-by: Lucas De Marchi <lucas.demarchi@intel.com>
Reviewed-by: Jani Nikula <jani.nikula@intel.com>
Signed-off-by: Rodrigo Vivi <rodrigo.vivi@intel.com>
Link: https://patchwork.freedesktop.org/patch/msgid/20181212181044.15886-2-lucas.demarchi@intel.com
2018-12-13 02:10:43 +08:00
|
|
|
} else if (IS_GEN(engine->i915, 6)) {
|
2016-03-16 19:00:37 +08:00
|
|
|
mmio = RING_HWS_PGA_GEN6(engine->mmio_base);
|
2014-01-24 06:40:36 +08:00
|
|
|
} else {
|
|
|
|
/* XXX: gen8 returns to sanity */
|
2016-03-16 19:00:37 +08:00
|
|
|
mmio = RING_HWS_PGA(engine->mmio_base);
|
2014-01-24 06:40:36 +08:00
|
|
|
}
|
|
|
|
|
2016-07-27 16:07:28 +08:00
|
|
|
ee->hws = I915_READ(mmio);
|
2014-01-24 06:40:36 +08:00
|
|
|
}
|
|
|
|
|
2017-12-19 21:14:19 +08:00
|
|
|
ee->idle = intel_engine_is_idle(engine);
|
2019-01-25 21:22:28 +08:00
|
|
|
if (!ee->idle)
|
|
|
|
ee->hangcheck_timestamp = engine->hangcheck.action_timestamp;
|
2017-06-20 17:57:48 +08:00
|
|
|
ee->reset_count = i915_reset_engine_count(&dev_priv->gpu_error,
|
|
|
|
engine);
|
2014-01-30 16:19:40 +08:00
|
|
|
|
2018-09-27 04:12:22 +08:00
|
|
|
if (HAS_PPGTT(dev_priv)) {
|
2014-01-30 16:19:40 +08:00
|
|
|
int i;
|
|
|
|
|
2016-07-27 16:07:28 +08:00
|
|
|
ee->vm_info.gfx_mode = I915_READ(RING_MODE_GEN7(engine));
|
2014-01-30 16:19:40 +08:00
|
|
|
|
drm/i915: replace IS_GEN<N> with IS_GEN(..., N)
Define IS_GEN() similarly to our IS_GEN_RANGE(). but use gen instead of
gen_mask to do the comparison. Now callers can pass then gen as a parameter,
so we don't require one macro for each gen.
The following spatch was used to convert the users of these macros:
@@
expression e;
@@
(
- IS_GEN2(e)
+ IS_GEN(e, 2)
|
- IS_GEN3(e)
+ IS_GEN(e, 3)
|
- IS_GEN4(e)
+ IS_GEN(e, 4)
|
- IS_GEN5(e)
+ IS_GEN(e, 5)
|
- IS_GEN6(e)
+ IS_GEN(e, 6)
|
- IS_GEN7(e)
+ IS_GEN(e, 7)
|
- IS_GEN8(e)
+ IS_GEN(e, 8)
|
- IS_GEN9(e)
+ IS_GEN(e, 9)
|
- IS_GEN10(e)
+ IS_GEN(e, 10)
|
- IS_GEN11(e)
+ IS_GEN(e, 11)
)
v2: use IS_GEN rather than GT_GEN and compare to info.gen rather than
using the bitmask
Signed-off-by: Lucas De Marchi <lucas.demarchi@intel.com>
Reviewed-by: Jani Nikula <jani.nikula@intel.com>
Signed-off-by: Rodrigo Vivi <rodrigo.vivi@intel.com>
Link: https://patchwork.freedesktop.org/patch/msgid/20181212181044.15886-2-lucas.demarchi@intel.com
2018-12-13 02:10:43 +08:00
|
|
|
if (IS_GEN(dev_priv, 6))
|
2016-07-27 16:07:28 +08:00
|
|
|
ee->vm_info.pp_dir_base =
|
2016-03-16 19:00:37 +08:00
|
|
|
I915_READ(RING_PP_DIR_BASE_READ(engine));
|
drm/i915: replace IS_GEN<N> with IS_GEN(..., N)
Define IS_GEN() similarly to our IS_GEN_RANGE(). but use gen instead of
gen_mask to do the comparison. Now callers can pass then gen as a parameter,
so we don't require one macro for each gen.
The following spatch was used to convert the users of these macros:
@@
expression e;
@@
(
- IS_GEN2(e)
+ IS_GEN(e, 2)
|
- IS_GEN3(e)
+ IS_GEN(e, 3)
|
- IS_GEN4(e)
+ IS_GEN(e, 4)
|
- IS_GEN5(e)
+ IS_GEN(e, 5)
|
- IS_GEN6(e)
+ IS_GEN(e, 6)
|
- IS_GEN7(e)
+ IS_GEN(e, 7)
|
- IS_GEN8(e)
+ IS_GEN(e, 8)
|
- IS_GEN9(e)
+ IS_GEN(e, 9)
|
- IS_GEN10(e)
+ IS_GEN(e, 10)
|
- IS_GEN11(e)
+ IS_GEN(e, 11)
)
v2: use IS_GEN rather than GT_GEN and compare to info.gen rather than
using the bitmask
Signed-off-by: Lucas De Marchi <lucas.demarchi@intel.com>
Reviewed-by: Jani Nikula <jani.nikula@intel.com>
Signed-off-by: Rodrigo Vivi <rodrigo.vivi@intel.com>
Link: https://patchwork.freedesktop.org/patch/msgid/20181212181044.15886-2-lucas.demarchi@intel.com
2018-12-13 02:10:43 +08:00
|
|
|
else if (IS_GEN(dev_priv, 7))
|
2016-07-27 16:07:28 +08:00
|
|
|
ee->vm_info.pp_dir_base =
|
2016-03-16 19:00:37 +08:00
|
|
|
I915_READ(RING_PP_DIR_BASE(engine));
|
2016-05-06 22:40:21 +08:00
|
|
|
else if (INTEL_GEN(dev_priv) >= 8)
|
2014-01-30 16:19:40 +08:00
|
|
|
for (i = 0; i < 4; i++) {
|
2016-07-27 16:07:28 +08:00
|
|
|
ee->vm_info.pdp[i] =
|
2016-03-16 19:00:37 +08:00
|
|
|
I915_READ(GEN8_RING_PDP_UDW(engine, i));
|
2016-07-27 16:07:28 +08:00
|
|
|
ee->vm_info.pdp[i] <<= 32;
|
|
|
|
ee->vm_info.pdp[i] |=
|
2016-03-16 19:00:37 +08:00
|
|
|
I915_READ(GEN8_RING_PDP_LDW(engine, i));
|
2014-01-30 16:19:40 +08:00
|
|
|
}
|
|
|
|
}
|
2013-07-12 21:50:57 +08:00
|
|
|
}
|
|
|
|
|
2018-02-21 17:56:36 +08:00
|
|
|
static void record_request(struct i915_request *request,
|
2016-10-13 18:18:14 +08:00
|
|
|
struct drm_i915_error_request *erq)
|
|
|
|
{
|
2018-05-18 05:26:30 +08:00
|
|
|
struct i915_gem_context *ctx = request->gem_context;
|
|
|
|
|
drm/i915: Replace global breadcrumbs with per-context interrupt tracking
A few years ago, see commit 688e6c725816 ("drm/i915: Slaughter the
thundering i915_wait_request herd"), the issue of handling multiple
clients waiting in parallel was brought to our attention. The
requirement was that every client should be woken immediately upon its
request being signaled, without incurring any cpu overhead.
To handle certain fragility of our hw meant that we could not do a
simple check inside the irq handler (some generations required almost
unbounded delays before we could be sure of seqno coherency) and so
request completion checking required delegation.
Before commit 688e6c725816, the solution was simple. Every client
waiting on a request would be woken on every interrupt and each would do
a heavyweight check to see if their request was complete. Commit
688e6c725816 introduced an rbtree so that only the earliest waiter on
the global timeline would woken, and would wake the next and so on.
(Along with various complications to handle requests being reordered
along the global timeline, and also a requirement for kthread to provide
a delegate for fence signaling that had no process context.)
The global rbtree depends on knowing the execution timeline (and global
seqno). Without knowing that order, we must instead check all contexts
queued to the HW to see which may have advanced. We trim that list by
only checking queued contexts that are being waited on, but still we
keep a list of all active contexts and their active signalers that we
inspect from inside the irq handler. By moving the waiters onto the fence
signal list, we can combine the client wakeup with the dma_fence
signaling (a dramatic reduction in complexity, but does require the HW
being coherent, the seqno must be visible from the cpu before the
interrupt is raised - we keep a timer backup just in case).
Having previously fixed all the issues with irq-seqno serialisation (by
inserting delays onto the GPU after each request instead of random delays
on the CPU after each interrupt), we can rely on the seqno state to
perfom direct wakeups from the interrupt handler. This allows us to
preserve our single context switch behaviour of the current routine,
with the only downside that we lose the RT priority sorting of wakeups.
In general, direct wakeup latency of multiple clients is about the same
(about 10% better in most cases) with a reduction in total CPU time spent
in the waiter (about 20-50% depending on gen). Average herd behaviour is
improved, but at the cost of not delegating wakeups on task_prio.
v2: Capture fence signaling state for error state and add comments to
warm even the most cold of hearts.
v3: Check if the request is still active before busywaiting
v4: Reduce the amount of pointer misdirection with list_for_each_safe
and using a local i915_request variable inside the loops
v5: Add a missing pluralisation to a purely informative selftest message.
References: 688e6c725816 ("drm/i915: Slaughter the thundering i915_wait_request herd")
Signed-off-by: Chris Wilson <chris@chris-wilson.co.uk>
Cc: Tvrtko Ursulin <tvrtko.ursulin@intel.com>
Reviewed-by: Tvrtko Ursulin <tvrtko.ursulin@intel.com>
Link: https://patchwork.freedesktop.org/patch/msgid/20190129205230.19056-2-chris@chris-wilson.co.uk
2019-01-30 04:52:29 +08:00
|
|
|
erq->flags = request->fence.flags;
|
2019-02-26 17:49:21 +08:00
|
|
|
erq->context = request->fence.context;
|
|
|
|
erq->seqno = request->fence.seqno;
|
2018-04-19 02:40:52 +08:00
|
|
|
erq->sched_attr = request->sched.attr;
|
2016-10-13 18:18:14 +08:00
|
|
|
erq->jiffies = request->emitted_jiffies;
|
2018-05-02 18:41:50 +08:00
|
|
|
erq->start = i915_ggtt_offset(request->ring->vma);
|
2016-10-13 18:18:14 +08:00
|
|
|
erq->head = request->head;
|
|
|
|
erq->tail = request->tail;
|
|
|
|
|
|
|
|
rcu_read_lock();
|
2018-05-18 05:26:30 +08:00
|
|
|
erq->pid = ctx->pid ? pid_nr(ctx->pid) : 0;
|
2016-10-13 18:18:14 +08:00
|
|
|
rcu_read_unlock();
|
|
|
|
}
|
|
|
|
|
2016-08-15 17:49:10 +08:00
|
|
|
static void engine_record_requests(struct intel_engine_cs *engine,
|
2018-02-21 17:56:36 +08:00
|
|
|
struct i915_request *first,
|
2016-08-15 17:49:10 +08:00
|
|
|
struct drm_i915_error_engine *ee)
|
|
|
|
{
|
2018-02-21 17:56:36 +08:00
|
|
|
struct i915_request *request;
|
2016-08-15 17:49:10 +08:00
|
|
|
int count;
|
|
|
|
|
|
|
|
count = 0;
|
|
|
|
request = first;
|
2018-05-03 00:38:39 +08:00
|
|
|
list_for_each_entry_from(request, &engine->timeline.requests, link)
|
2016-08-15 17:49:10 +08:00
|
|
|
count++;
|
|
|
|
if (!count)
|
|
|
|
return;
|
|
|
|
|
|
|
|
ee->requests = kcalloc(count, sizeof(*ee->requests), GFP_ATOMIC);
|
|
|
|
if (!ee->requests)
|
|
|
|
return;
|
|
|
|
|
|
|
|
ee->num_requests = count;
|
|
|
|
|
|
|
|
count = 0;
|
|
|
|
request = first;
|
2018-05-03 00:38:39 +08:00
|
|
|
list_for_each_entry_from(request, &engine->timeline.requests, link) {
|
2016-08-15 17:49:10 +08:00
|
|
|
if (count >= ee->num_requests) {
|
|
|
|
/*
|
|
|
|
* If the ring request list was changed in
|
|
|
|
* between the point where the error request
|
|
|
|
* list was created and dimensioned and this
|
|
|
|
* point then just exit early to avoid crashes.
|
|
|
|
*
|
|
|
|
* We don't need to communicate that the
|
|
|
|
* request list changed state during error
|
|
|
|
* state capture and that the error state is
|
|
|
|
* slightly incorrect as a consequence since we
|
|
|
|
* are typically only interested in the request
|
|
|
|
* list state at the point of error state
|
|
|
|
* capture, not in any changes happening during
|
|
|
|
* the capture.
|
|
|
|
*/
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
2016-10-13 18:18:14 +08:00
|
|
|
record_request(request, &ee->requests[count++]);
|
2016-08-15 17:49:10 +08:00
|
|
|
}
|
|
|
|
ee->num_requests = count;
|
|
|
|
}
|
|
|
|
|
2016-10-13 18:18:14 +08:00
|
|
|
static void error_record_engine_execlists(struct intel_engine_cs *engine,
|
|
|
|
struct drm_i915_error_engine *ee)
|
|
|
|
{
|
2017-09-22 20:43:07 +08:00
|
|
|
const struct intel_engine_execlists * const execlists = &engine->execlists;
|
2016-10-13 18:18:14 +08:00
|
|
|
unsigned int n;
|
|
|
|
|
2017-09-22 20:43:07 +08:00
|
|
|
for (n = 0; n < execlists_num_ports(execlists); n++) {
|
2018-02-21 17:56:36 +08:00
|
|
|
struct i915_request *rq = port_request(&execlists->port[n]);
|
2017-05-17 20:10:00 +08:00
|
|
|
|
|
|
|
if (!rq)
|
|
|
|
break;
|
|
|
|
|
|
|
|
record_request(rq, &ee->execlist[n]);
|
|
|
|
}
|
2017-09-22 20:43:07 +08:00
|
|
|
|
|
|
|
ee->num_ports = n;
|
2016-10-13 18:18:14 +08:00
|
|
|
}
|
|
|
|
|
2017-01-29 17:24:33 +08:00
|
|
|
static void record_context(struct drm_i915_error_context *e,
|
|
|
|
struct i915_gem_context *ctx)
|
|
|
|
{
|
|
|
|
if (ctx->pid) {
|
|
|
|
struct task_struct *task;
|
|
|
|
|
|
|
|
rcu_read_lock();
|
|
|
|
task = pid_task(ctx->pid, PIDTYPE_PID);
|
|
|
|
if (task) {
|
|
|
|
strcpy(e->comm, task->comm);
|
|
|
|
e->pid = task->pid;
|
|
|
|
}
|
|
|
|
rcu_read_unlock();
|
|
|
|
}
|
|
|
|
|
|
|
|
e->handle = ctx->user_handle;
|
|
|
|
e->hw_id = ctx->hw_id;
|
2018-04-19 02:40:52 +08:00
|
|
|
e->sched_attr = ctx->sched;
|
2017-07-21 20:32:30 +08:00
|
|
|
e->guilty = atomic_read(&ctx->guilty_count);
|
|
|
|
e->active = atomic_read(&ctx->active_count);
|
2017-01-29 17:24:33 +08:00
|
|
|
}
|
|
|
|
|
2018-02-21 17:56:36 +08:00
|
|
|
static void request_record_user_bo(struct i915_request *request,
|
2017-04-15 17:39:02 +08:00
|
|
|
struct drm_i915_error_engine *ee)
|
|
|
|
{
|
2018-02-21 17:56:36 +08:00
|
|
|
struct i915_capture_list *c;
|
2017-04-15 17:39:02 +08:00
|
|
|
struct drm_i915_error_object **bo;
|
2018-09-11 19:57:47 +08:00
|
|
|
long count, max;
|
2017-04-15 17:39:02 +08:00
|
|
|
|
2018-09-11 19:57:47 +08:00
|
|
|
max = 0;
|
2017-04-15 17:39:02 +08:00
|
|
|
for (c = request->capture_list; c; c = c->next)
|
2018-09-11 19:57:47 +08:00
|
|
|
max++;
|
|
|
|
if (!max)
|
|
|
|
return;
|
2017-04-15 17:39:02 +08:00
|
|
|
|
2018-09-11 19:57:47 +08:00
|
|
|
bo = kmalloc_array(max, sizeof(*bo), GFP_ATOMIC);
|
|
|
|
if (!bo) {
|
|
|
|
/* If we can't capture everything, try to capture something. */
|
|
|
|
max = min_t(long, max, PAGE_SIZE / sizeof(*bo));
|
|
|
|
bo = kmalloc_array(max, sizeof(*bo), GFP_ATOMIC);
|
|
|
|
}
|
2017-04-15 17:39:02 +08:00
|
|
|
if (!bo)
|
|
|
|
return;
|
|
|
|
|
|
|
|
count = 0;
|
|
|
|
for (c = request->capture_list; c; c = c->next) {
|
|
|
|
bo[count] = i915_error_object_create(request->i915, c->vma);
|
|
|
|
if (!bo[count])
|
|
|
|
break;
|
2018-09-11 19:57:47 +08:00
|
|
|
if (++count == max)
|
|
|
|
break;
|
2017-04-15 17:39:02 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
ee->user_bo = bo;
|
|
|
|
ee->user_bo_count = count;
|
|
|
|
}
|
|
|
|
|
2017-11-27 06:09:01 +08:00
|
|
|
static struct drm_i915_error_object *
|
|
|
|
capture_object(struct drm_i915_private *dev_priv,
|
|
|
|
struct drm_i915_gem_object *obj)
|
|
|
|
{
|
|
|
|
if (obj && i915_gem_object_has_pages(obj)) {
|
|
|
|
struct i915_vma fake = {
|
|
|
|
.node = { .start = U64_MAX, .size = obj->base.size },
|
2017-12-01 08:15:36 +08:00
|
|
|
.size = obj->base.size,
|
2017-11-27 06:09:01 +08:00
|
|
|
.pages = obj->mm.pages,
|
|
|
|
.obj = obj,
|
|
|
|
};
|
|
|
|
|
|
|
|
return i915_error_object_create(dev_priv, &fake);
|
|
|
|
} else {
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2018-03-06 06:21:21 +08:00
|
|
|
static void gem_record_rings(struct i915_gpu_state *error)
|
2013-07-12 21:50:57 +08:00
|
|
|
{
|
2018-03-06 06:21:21 +08:00
|
|
|
struct drm_i915_private *i915 = error->i915;
|
|
|
|
struct i915_ggtt *ggtt = &i915->ggtt;
|
2016-08-15 17:49:10 +08:00
|
|
|
int i;
|
2013-07-12 21:50:57 +08:00
|
|
|
|
2016-03-16 19:00:39 +08:00
|
|
|
for (i = 0; i < I915_NUM_ENGINES; i++) {
|
2018-03-06 06:21:21 +08:00
|
|
|
struct intel_engine_cs *engine = i915->engine[i];
|
2016-07-27 16:07:28 +08:00
|
|
|
struct drm_i915_error_engine *ee = &error->engine[i];
|
2018-02-21 17:56:36 +08:00
|
|
|
struct i915_request *request;
|
2014-01-27 21:52:34 +08:00
|
|
|
|
2016-07-27 16:07:28 +08:00
|
|
|
ee->engine_id = -1;
|
2014-06-10 19:09:29 +08:00
|
|
|
|
drm/i915: Allocate intel_engine_cs structure only for the enabled engines
With the possibility of addition of many more number of rings in future,
the drm_i915_private structure could bloat as an array, of type
intel_engine_cs, is embedded inside it.
struct intel_engine_cs engine[I915_NUM_ENGINES];
Though this is still fine as generally there is only a single instance of
drm_i915_private structure used, but not all of the possible rings would be
enabled or active on most of the platforms. Some memory can be saved by
allocating intel_engine_cs structure only for the enabled/active engines.
Currently the engine/ring ID is kept static and dev_priv->engine[] is simply
indexed using the enums defined in intel_engine_id.
To save memory and continue using the static engine/ring IDs, 'engine' is
defined as an array of pointers.
struct intel_engine_cs *engine[I915_NUM_ENGINES];
dev_priv->engine[engine_ID] will be NULL for disabled engine instances.
There is a text size reduction of 928 bytes, from 1028200 to 1027272, for
i915.o file (but for i915.ko file text size remain same as 1193131 bytes).
v2:
- Remove the engine iterator field added in drm_i915_private structure,
instead pass a local iterator variable to the for_each_engine**
macros. (Chris)
- Do away with intel_engine_initialized() and instead directly use the
NULL pointer check on engine pointer. (Chris)
v3:
- Remove for_each_engine_id() macro, as the updated macro for_each_engine()
can be used in place of it. (Chris)
- Protect the access to Render engine Fault register with a NULL check, as
engine specific init is done later in Driver load sequence.
v4:
- Use !!dev_priv->engine[VCS] style for the engine check in getparam. (Chris)
- Kill the superfluous init_engine_lists().
v5:
- Cleanup the intel_engines_init() & intel_engines_setup(), with respect to
allocation of intel_engine_cs structure. (Chris)
v6:
- Rebase.
v7:
- Optimize the for_each_engine_masked() macro. (Chris)
- Change the type of 'iter' local variable to enum intel_engine_id. (Chris)
- Rebase.
v8: Rebase.
v9: Rebase.
v10:
- For index calculation use engine ID instead of pointer based arithmetic in
intel_engine_sync_index() as engine pointers are not contiguous now (Chris)
- For appropriateness, rename local enum variable 'iter' to 'id'. (Joonas)
- Use for_each_engine macro for cleanup in intel_engines_init() and remove
check for NULL engine pointer in cleanup() routines. (Joonas)
v11: Rebase.
Cc: Chris Wilson <chris@chris-wilson.co.uk>
Signed-off-by: Akash Goel <akash.goel@intel.com>
Reviewed-by: Joonas Lahtinen <joonas.lahtinen@linux.intel.com>
Signed-off-by: Tvrtko Ursulin <tvrtko.ursulin@intel.com>
Link: http://patchwork.freedesktop.org/patch/msgid/1476378888-7372-1-git-send-email-akash.goel@intel.com
2016-10-14 01:14:48 +08:00
|
|
|
if (!engine)
|
2014-01-27 21:52:34 +08:00
|
|
|
continue;
|
|
|
|
|
2016-07-27 16:07:28 +08:00
|
|
|
ee->engine_id = i;
|
2014-01-27 21:52:34 +08:00
|
|
|
|
2016-07-27 16:07:28 +08:00
|
|
|
error_record_engine_registers(error, engine, ee);
|
2016-10-13 18:18:14 +08:00
|
|
|
error_record_engine_execlists(engine, ee);
|
2013-07-12 21:50:57 +08:00
|
|
|
|
2016-03-16 19:00:36 +08:00
|
|
|
request = i915_gem_find_active_request(engine);
|
2014-02-25 23:11:24 +08:00
|
|
|
if (request) {
|
2018-05-18 05:26:30 +08:00
|
|
|
struct i915_gem_context *ctx = request->gem_context;
|
2016-08-03 05:50:21 +08:00
|
|
|
struct intel_ring *ring;
|
2014-08-06 21:04:53 +08:00
|
|
|
|
2018-06-05 23:37:58 +08:00
|
|
|
ee->vm = ctx->ppgtt ? &ctx->ppgtt->vm : &ggtt->vm;
|
2014-08-06 21:04:53 +08:00
|
|
|
|
2018-05-18 05:26:30 +08:00
|
|
|
record_context(&ee->context, ctx);
|
2017-01-29 17:24:33 +08:00
|
|
|
|
2014-02-25 23:11:24 +08:00
|
|
|
/* We need to copy these to an anonymous buffer
|
|
|
|
* as the simplest method to avoid being overwritten
|
|
|
|
* by userspace.
|
|
|
|
*/
|
2016-07-27 16:07:28 +08:00
|
|
|
ee->batchbuffer =
|
2018-03-06 06:21:21 +08:00
|
|
|
i915_error_object_create(i915, request->batch);
|
2014-02-25 23:11:24 +08:00
|
|
|
|
2018-03-06 06:21:21 +08:00
|
|
|
if (HAS_BROKEN_CS_TLB(i915))
|
2016-07-27 16:07:28 +08:00
|
|
|
ee->wa_batchbuffer =
|
2018-03-06 06:21:21 +08:00
|
|
|
i915_error_object_create(i915,
|
2018-12-04 22:15:16 +08:00
|
|
|
i915->gt.scratch);
|
2017-04-15 17:39:02 +08:00
|
|
|
request_record_user_bo(request, ee);
|
2014-02-25 23:11:24 +08:00
|
|
|
|
2016-08-15 17:49:06 +08:00
|
|
|
ee->ctx =
|
2018-03-06 06:21:21 +08:00
|
|
|
i915_error_object_create(i915,
|
2018-05-18 05:26:32 +08:00
|
|
|
request->hw_context->state);
|
2016-08-15 17:48:42 +08:00
|
|
|
|
2016-07-04 15:08:39 +08:00
|
|
|
error->simulated |=
|
2018-05-18 05:26:30 +08:00
|
|
|
i915_gem_context_no_error_capture(ctx);
|
2016-07-04 15:08:39 +08:00
|
|
|
|
2016-10-05 04:11:30 +08:00
|
|
|
ee->rq_head = request->head;
|
|
|
|
ee->rq_post = request->postfix;
|
|
|
|
ee->rq_tail = request->tail;
|
|
|
|
|
2016-08-03 05:50:19 +08:00
|
|
|
ring = request->ring;
|
|
|
|
ee->cpu_ring_head = ring->head;
|
|
|
|
ee->cpu_ring_tail = ring->tail;
|
2016-07-27 16:07:28 +08:00
|
|
|
ee->ringbuffer =
|
2018-03-06 06:21:21 +08:00
|
|
|
i915_error_object_create(i915, ring->vma);
|
2016-08-15 17:49:10 +08:00
|
|
|
|
|
|
|
engine_record_requests(engine, request, ee);
|
2016-07-04 15:08:38 +08:00
|
|
|
}
|
2013-07-12 21:50:57 +08:00
|
|
|
|
2016-07-27 16:07:28 +08:00
|
|
|
ee->hws_page =
|
2018-03-06 06:21:21 +08:00
|
|
|
i915_error_object_create(i915,
|
2016-08-15 17:49:06 +08:00
|
|
|
engine->status_page.vma);
|
2013-07-12 21:50:57 +08:00
|
|
|
|
2018-03-06 06:21:21 +08:00
|
|
|
ee->wa_ctx = i915_error_object_create(i915, engine->wa_ctx.vma);
|
2017-11-27 06:09:01 +08:00
|
|
|
|
2018-03-06 06:21:21 +08:00
|
|
|
ee->default_state = capture_object(i915, engine->default_state);
|
2013-07-12 21:50:57 +08:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2018-03-06 06:21:21 +08:00
|
|
|
static void gem_capture_vm(struct i915_gpu_state *error,
|
|
|
|
struct i915_address_space *vm,
|
|
|
|
int idx)
|
2013-07-12 21:50:57 +08:00
|
|
|
{
|
2016-08-15 17:48:41 +08:00
|
|
|
struct drm_i915_error_buffer *active_bo;
|
2013-08-01 08:00:15 +08:00
|
|
|
struct i915_vma *vma;
|
2016-08-15 17:48:41 +08:00
|
|
|
int count;
|
2013-07-12 21:50:57 +08:00
|
|
|
|
2016-08-15 17:48:41 +08:00
|
|
|
count = 0;
|
2019-01-28 18:23:52 +08:00
|
|
|
list_for_each_entry(vma, &vm->bound_list, vm_link)
|
|
|
|
if (i915_vma_is_active(vma))
|
|
|
|
count++;
|
2013-07-12 21:50:57 +08:00
|
|
|
|
2016-08-15 17:48:41 +08:00
|
|
|
active_bo = NULL;
|
|
|
|
if (count)
|
|
|
|
active_bo = kcalloc(count, sizeof(*active_bo), GFP_ATOMIC);
|
2013-08-01 08:00:15 +08:00
|
|
|
if (active_bo)
|
2019-01-28 18:23:52 +08:00
|
|
|
count = capture_error_bo(active_bo,
|
|
|
|
count, &vm->bound_list,
|
|
|
|
ACTIVE_ONLY);
|
2016-08-15 17:48:41 +08:00
|
|
|
else
|
|
|
|
count = 0;
|
|
|
|
|
|
|
|
error->active_vm[idx] = vm;
|
|
|
|
error->active_bo[idx] = active_bo;
|
|
|
|
error->active_bo_count[idx] = count;
|
2013-08-01 08:00:15 +08:00
|
|
|
}
|
|
|
|
|
2018-03-06 06:21:21 +08:00
|
|
|
static void capture_active_buffers(struct i915_gpu_state *error)
|
2013-08-01 08:00:15 +08:00
|
|
|
{
|
2016-08-15 17:48:41 +08:00
|
|
|
int cnt = 0, i, j;
|
|
|
|
|
|
|
|
BUILD_BUG_ON(ARRAY_SIZE(error->engine) > ARRAY_SIZE(error->active_bo));
|
|
|
|
BUILD_BUG_ON(ARRAY_SIZE(error->active_bo) != ARRAY_SIZE(error->active_vm));
|
|
|
|
BUILD_BUG_ON(ARRAY_SIZE(error->active_bo) != ARRAY_SIZE(error->active_bo_count));
|
|
|
|
|
|
|
|
/* Scan each engine looking for unique active contexts/vm */
|
|
|
|
for (i = 0; i < ARRAY_SIZE(error->engine); i++) {
|
|
|
|
struct drm_i915_error_engine *ee = &error->engine[i];
|
|
|
|
bool found;
|
|
|
|
|
|
|
|
if (!ee->vm)
|
|
|
|
continue;
|
2014-08-13 03:05:47 +08:00
|
|
|
|
2016-08-15 17:48:41 +08:00
|
|
|
found = false;
|
|
|
|
for (j = 0; j < i && !found; j++)
|
|
|
|
found = error->engine[j].vm == ee->vm;
|
|
|
|
if (!found)
|
2018-03-06 06:21:21 +08:00
|
|
|
gem_capture_vm(error, ee->vm, cnt++);
|
2014-08-13 03:05:47 +08:00
|
|
|
}
|
2013-07-12 21:50:57 +08:00
|
|
|
}
|
|
|
|
|
2018-03-06 06:21:21 +08:00
|
|
|
static void capture_pinned_buffers(struct i915_gpu_state *error)
|
2016-08-15 17:48:41 +08:00
|
|
|
{
|
2018-06-05 23:37:58 +08:00
|
|
|
struct i915_address_space *vm = &error->i915->ggtt.vm;
|
2016-08-15 17:48:41 +08:00
|
|
|
struct drm_i915_error_buffer *bo;
|
|
|
|
struct i915_vma *vma;
|
2019-01-28 18:23:52 +08:00
|
|
|
int count;
|
2016-08-15 17:48:41 +08:00
|
|
|
|
2019-01-28 18:23:52 +08:00
|
|
|
count = 0;
|
|
|
|
list_for_each_entry(vma, &vm->bound_list, vm_link)
|
|
|
|
count++;
|
2016-08-15 17:48:41 +08:00
|
|
|
|
|
|
|
bo = NULL;
|
2019-01-28 18:23:52 +08:00
|
|
|
if (count)
|
|
|
|
bo = kcalloc(count, sizeof(*bo), GFP_ATOMIC);
|
2016-08-15 17:48:41 +08:00
|
|
|
if (!bo)
|
|
|
|
return;
|
|
|
|
|
2019-01-28 18:23:52 +08:00
|
|
|
error->pinned_bo_count =
|
|
|
|
capture_error_bo(bo, count, &vm->bound_list, PINNED_ONLY);
|
2016-08-15 17:48:41 +08:00
|
|
|
error->pinned_bo = bo;
|
|
|
|
}
|
|
|
|
|
2017-10-27 01:36:55 +08:00
|
|
|
static void capture_uc_state(struct i915_gpu_state *error)
|
|
|
|
{
|
|
|
|
struct drm_i915_private *i915 = error->i915;
|
|
|
|
struct i915_error_uc *error_uc = &error->uc;
|
|
|
|
|
|
|
|
/* Capturing uC state won't be useful if there is no GuC */
|
|
|
|
if (!error->device_info.has_guc)
|
|
|
|
return;
|
|
|
|
|
|
|
|
error_uc->guc_fw = i915->guc.fw;
|
|
|
|
error_uc->huc_fw = i915->huc.fw;
|
|
|
|
|
|
|
|
/* Non-default firmware paths will be specified by the modparam.
|
|
|
|
* As modparams are generally accesible from the userspace make
|
|
|
|
* explicit copies of the firmware paths.
|
|
|
|
*/
|
|
|
|
error_uc->guc_fw.path = kstrdup(i915->guc.fw.path, GFP_ATOMIC);
|
|
|
|
error_uc->huc_fw.path = kstrdup(i915->huc.fw.path, GFP_ATOMIC);
|
2017-10-27 01:36:56 +08:00
|
|
|
error_uc->guc_log = i915_error_object_create(i915, i915->guc.log.vma);
|
2016-10-13 00:24:39 +08:00
|
|
|
}
|
|
|
|
|
2014-01-30 16:19:35 +08:00
|
|
|
/* Capture all registers which don't fit into another category. */
|
2018-03-06 06:21:21 +08:00
|
|
|
static void capture_reg_state(struct i915_gpu_state *error)
|
2013-07-12 21:50:57 +08:00
|
|
|
{
|
2018-03-06 06:21:21 +08:00
|
|
|
struct drm_i915_private *dev_priv = error->i915;
|
2014-08-06 01:07:13 +08:00
|
|
|
int i;
|
2013-07-12 21:50:57 +08:00
|
|
|
|
2014-01-30 16:19:36 +08:00
|
|
|
/* General organization
|
|
|
|
* 1. Registers specific to a single generation
|
|
|
|
* 2. Registers which belong to multiple generations
|
|
|
|
* 3. Feature specific registers.
|
|
|
|
* 4. Everything else
|
|
|
|
* Please try to follow the order.
|
|
|
|
*/
|
2013-07-12 21:50:57 +08:00
|
|
|
|
2014-01-30 16:19:36 +08:00
|
|
|
/* 1: Registers specific to a single generation */
|
2016-10-13 18:03:08 +08:00
|
|
|
if (IS_VALLEYVIEW(dev_priv)) {
|
2014-08-06 01:07:13 +08:00
|
|
|
error->gtier[0] = I915_READ(GTIER);
|
2014-08-02 00:12:27 +08:00
|
|
|
error->ier = I915_READ(VLV_IER);
|
2015-10-22 20:34:57 +08:00
|
|
|
error->forcewake = I915_READ_FW(FORCEWAKE_VLV);
|
2014-01-30 16:19:36 +08:00
|
|
|
}
|
2013-07-12 21:50:57 +08:00
|
|
|
|
drm/i915: replace IS_GEN<N> with IS_GEN(..., N)
Define IS_GEN() similarly to our IS_GEN_RANGE(). but use gen instead of
gen_mask to do the comparison. Now callers can pass then gen as a parameter,
so we don't require one macro for each gen.
The following spatch was used to convert the users of these macros:
@@
expression e;
@@
(
- IS_GEN2(e)
+ IS_GEN(e, 2)
|
- IS_GEN3(e)
+ IS_GEN(e, 3)
|
- IS_GEN4(e)
+ IS_GEN(e, 4)
|
- IS_GEN5(e)
+ IS_GEN(e, 5)
|
- IS_GEN6(e)
+ IS_GEN(e, 6)
|
- IS_GEN7(e)
+ IS_GEN(e, 7)
|
- IS_GEN8(e)
+ IS_GEN(e, 8)
|
- IS_GEN9(e)
+ IS_GEN(e, 9)
|
- IS_GEN10(e)
+ IS_GEN(e, 10)
|
- IS_GEN11(e)
+ IS_GEN(e, 11)
)
v2: use IS_GEN rather than GT_GEN and compare to info.gen rather than
using the bitmask
Signed-off-by: Lucas De Marchi <lucas.demarchi@intel.com>
Reviewed-by: Jani Nikula <jani.nikula@intel.com>
Signed-off-by: Rodrigo Vivi <rodrigo.vivi@intel.com>
Link: https://patchwork.freedesktop.org/patch/msgid/20181212181044.15886-2-lucas.demarchi@intel.com
2018-12-13 02:10:43 +08:00
|
|
|
if (IS_GEN(dev_priv, 7))
|
2014-01-30 16:19:36 +08:00
|
|
|
error->err_int = I915_READ(GEN7_ERR_INT);
|
2013-07-12 21:50:57 +08:00
|
|
|
|
2016-11-16 16:55:37 +08:00
|
|
|
if (INTEL_GEN(dev_priv) >= 8) {
|
2015-03-24 20:54:19 +08:00
|
|
|
error->fault_data0 = I915_READ(GEN8_FAULT_TLB_DATA0);
|
|
|
|
error->fault_data1 = I915_READ(GEN8_FAULT_TLB_DATA1);
|
|
|
|
}
|
|
|
|
|
drm/i915: replace IS_GEN<N> with IS_GEN(..., N)
Define IS_GEN() similarly to our IS_GEN_RANGE(). but use gen instead of
gen_mask to do the comparison. Now callers can pass then gen as a parameter,
so we don't require one macro for each gen.
The following spatch was used to convert the users of these macros:
@@
expression e;
@@
(
- IS_GEN2(e)
+ IS_GEN(e, 2)
|
- IS_GEN3(e)
+ IS_GEN(e, 3)
|
- IS_GEN4(e)
+ IS_GEN(e, 4)
|
- IS_GEN5(e)
+ IS_GEN(e, 5)
|
- IS_GEN6(e)
+ IS_GEN(e, 6)
|
- IS_GEN7(e)
+ IS_GEN(e, 7)
|
- IS_GEN8(e)
+ IS_GEN(e, 8)
|
- IS_GEN9(e)
+ IS_GEN(e, 9)
|
- IS_GEN10(e)
+ IS_GEN(e, 10)
|
- IS_GEN11(e)
+ IS_GEN(e, 11)
)
v2: use IS_GEN rather than GT_GEN and compare to info.gen rather than
using the bitmask
Signed-off-by: Lucas De Marchi <lucas.demarchi@intel.com>
Reviewed-by: Jani Nikula <jani.nikula@intel.com>
Signed-off-by: Rodrigo Vivi <rodrigo.vivi@intel.com>
Link: https://patchwork.freedesktop.org/patch/msgid/20181212181044.15886-2-lucas.demarchi@intel.com
2018-12-13 02:10:43 +08:00
|
|
|
if (IS_GEN(dev_priv, 6)) {
|
2015-10-22 20:34:57 +08:00
|
|
|
error->forcewake = I915_READ_FW(FORCEWAKE);
|
2014-01-30 16:19:39 +08:00
|
|
|
error->gab_ctl = I915_READ(GAB_CTL);
|
|
|
|
error->gfx_mode = I915_READ(GFX_MODE);
|
|
|
|
}
|
2013-07-12 21:50:57 +08:00
|
|
|
|
2014-01-30 16:19:36 +08:00
|
|
|
/* 2: Registers which belong to multiple generations */
|
2016-11-16 16:55:37 +08:00
|
|
|
if (INTEL_GEN(dev_priv) >= 7)
|
2015-10-22 20:34:57 +08:00
|
|
|
error->forcewake = I915_READ_FW(FORCEWAKE_MT);
|
2013-07-12 21:50:57 +08:00
|
|
|
|
2016-11-16 16:55:37 +08:00
|
|
|
if (INTEL_GEN(dev_priv) >= 6) {
|
2014-01-30 16:19:36 +08:00
|
|
|
error->derrmr = I915_READ(DERRMR);
|
2013-07-12 21:50:57 +08:00
|
|
|
error->error = I915_READ(ERROR_GEN6);
|
|
|
|
error->done_reg = I915_READ(DONE_REG);
|
|
|
|
}
|
|
|
|
|
2017-04-28 15:53:39 +08:00
|
|
|
if (INTEL_GEN(dev_priv) >= 5)
|
2017-04-28 15:53:37 +08:00
|
|
|
error->ccid = I915_READ(CCID);
|
|
|
|
|
2014-01-30 16:19:36 +08:00
|
|
|
/* 3: Feature specific registers */
|
drm/i915: merge gen checks to use range
Instead of using IS_GEN() for consecutive gen checks, let's pass the
range to IS_GEN_RANGE(). By code inspection these were the ranges deemed
necessary for spatch:
@@
expression e;
@@
(
- IS_GEN(e, 3) || IS_GEN(e, 2)
+ IS_GEN_RANGE(e, 2, 3)
|
- IS_GEN(e, 3) || IS_GEN(e, 4)
+ IS_GEN_RANGE(e, 3, 4)
|
- IS_GEN(e, 5) || IS_GEN(e, 6)
+ IS_GEN_RANGE(e, 5, 6)
|
- IS_GEN(e, 6) || IS_GEN(e, 7)
+ IS_GEN_RANGE(e, 6, 7)
|
- IS_GEN(e, 7) || IS_GEN(e, 8)
+ IS_GEN_RANGE(e, 7, 8)
|
- IS_GEN(e, 8) || IS_GEN(e, 9)
+ IS_GEN_RANGE(e, 8, 9)
|
- IS_GEN(e, 10) || IS_GEN(e, 9)
+ IS_GEN_RANGE(e, 9, 10)
|
- IS_GEN(e, 9) || IS_GEN(e, 10)
+ IS_GEN_RANGE(e, 9, 10)
)
After conversion, checking we don't have any missing IS_GEN_RANGE() ||
IS_GEN() was also done.
Signed-off-by: Lucas De Marchi <lucas.demarchi@intel.com>
Reviewed-by: Jani Nikula <jani.nikula@intel.com>
Signed-off-by: Rodrigo Vivi <rodrigo.vivi@intel.com>
Link: https://patchwork.freedesktop.org/patch/msgid/20181212181044.15886-3-lucas.demarchi@intel.com
2018-12-13 02:10:44 +08:00
|
|
|
if (IS_GEN_RANGE(dev_priv, 6, 7)) {
|
2014-01-30 16:19:39 +08:00
|
|
|
error->gam_ecochk = I915_READ(GAM_ECOCHK);
|
|
|
|
error->gac_eco = I915_READ(GAC_ECO_BITS);
|
|
|
|
}
|
|
|
|
|
|
|
|
/* 4: Everything else */
|
2018-05-11 05:59:55 +08:00
|
|
|
if (INTEL_GEN(dev_priv) >= 11) {
|
|
|
|
error->ier = I915_READ(GEN8_DE_MISC_IER);
|
|
|
|
error->gtier[0] = I915_READ(GEN11_RENDER_COPY_INTR_ENABLE);
|
|
|
|
error->gtier[1] = I915_READ(GEN11_VCS_VECS_INTR_ENABLE);
|
|
|
|
error->gtier[2] = I915_READ(GEN11_GUC_SG_INTR_ENABLE);
|
|
|
|
error->gtier[3] = I915_READ(GEN11_GPM_WGBOXPERF_INTR_ENABLE);
|
|
|
|
error->gtier[4] = I915_READ(GEN11_CRYPTO_RSVD_INTR_ENABLE);
|
|
|
|
error->gtier[5] = I915_READ(GEN11_GUNIT_CSME_INTR_ENABLE);
|
|
|
|
error->ngtier = 6;
|
|
|
|
} else if (INTEL_GEN(dev_priv) >= 8) {
|
2014-08-06 01:07:13 +08:00
|
|
|
error->ier = I915_READ(GEN8_DE_MISC_IER);
|
|
|
|
for (i = 0; i < 4; i++)
|
|
|
|
error->gtier[i] = I915_READ(GEN8_GT_IER(i));
|
2017-02-15 00:46:11 +08:00
|
|
|
error->ngtier = 4;
|
2016-10-13 18:02:53 +08:00
|
|
|
} else if (HAS_PCH_SPLIT(dev_priv)) {
|
2014-08-02 00:12:27 +08:00
|
|
|
error->ier = I915_READ(DEIER);
|
2014-08-06 01:07:13 +08:00
|
|
|
error->gtier[0] = I915_READ(GTIER);
|
2017-02-15 00:46:11 +08:00
|
|
|
error->ngtier = 1;
|
drm/i915: replace IS_GEN<N> with IS_GEN(..., N)
Define IS_GEN() similarly to our IS_GEN_RANGE(). but use gen instead of
gen_mask to do the comparison. Now callers can pass then gen as a parameter,
so we don't require one macro for each gen.
The following spatch was used to convert the users of these macros:
@@
expression e;
@@
(
- IS_GEN2(e)
+ IS_GEN(e, 2)
|
- IS_GEN3(e)
+ IS_GEN(e, 3)
|
- IS_GEN4(e)
+ IS_GEN(e, 4)
|
- IS_GEN5(e)
+ IS_GEN(e, 5)
|
- IS_GEN6(e)
+ IS_GEN(e, 6)
|
- IS_GEN7(e)
+ IS_GEN(e, 7)
|
- IS_GEN8(e)
+ IS_GEN(e, 8)
|
- IS_GEN9(e)
+ IS_GEN(e, 9)
|
- IS_GEN10(e)
+ IS_GEN(e, 10)
|
- IS_GEN11(e)
+ IS_GEN(e, 11)
)
v2: use IS_GEN rather than GT_GEN and compare to info.gen rather than
using the bitmask
Signed-off-by: Lucas De Marchi <lucas.demarchi@intel.com>
Reviewed-by: Jani Nikula <jani.nikula@intel.com>
Signed-off-by: Rodrigo Vivi <rodrigo.vivi@intel.com>
Link: https://patchwork.freedesktop.org/patch/msgid/20181212181044.15886-2-lucas.demarchi@intel.com
2018-12-13 02:10:43 +08:00
|
|
|
} else if (IS_GEN(dev_priv, 2)) {
|
2014-08-02 00:12:27 +08:00
|
|
|
error->ier = I915_READ16(IER);
|
2016-10-13 18:03:08 +08:00
|
|
|
} else if (!IS_VALLEYVIEW(dev_priv)) {
|
2014-08-02 00:12:27 +08:00
|
|
|
error->ier = I915_READ(IER);
|
2014-01-30 16:19:36 +08:00
|
|
|
}
|
|
|
|
error->eir = I915_READ(EIR);
|
|
|
|
error->pgtbl_er = I915_READ(PGTBL_ER);
|
2014-01-30 16:19:35 +08:00
|
|
|
}
|
|
|
|
|
2019-01-25 21:22:28 +08:00
|
|
|
static const char *
|
|
|
|
error_msg(struct i915_gpu_state *error, unsigned long engines, const char *msg)
|
2014-02-25 23:11:25 +08:00
|
|
|
{
|
2019-01-25 21:22:28 +08:00
|
|
|
int len;
|
|
|
|
int i;
|
2014-02-25 23:11:25 +08:00
|
|
|
|
2019-01-25 21:22:28 +08:00
|
|
|
for (i = 0; i < ARRAY_SIZE(error->engine); i++)
|
|
|
|
if (!error->engine[i].context.pid)
|
|
|
|
engines &= ~BIT(i);
|
2014-02-25 23:11:25 +08:00
|
|
|
|
2014-02-25 23:11:26 +08:00
|
|
|
len = scnprintf(error->error_msg, sizeof(error->error_msg),
|
2019-01-25 21:22:28 +08:00
|
|
|
"GPU HANG: ecode %d:%lx:0x%08x",
|
|
|
|
INTEL_GEN(error->i915), engines,
|
|
|
|
i915_error_generate_code(error, engines));
|
|
|
|
if (engines) {
|
|
|
|
/* Just show the first executing process, more is confusing */
|
|
|
|
i = ffs(engines);
|
2014-02-25 23:11:26 +08:00
|
|
|
len += scnprintf(error->error_msg + len,
|
|
|
|
sizeof(error->error_msg) - len,
|
|
|
|
", in %s [%d]",
|
2019-01-25 21:22:28 +08:00
|
|
|
error->engine[i].context.comm,
|
|
|
|
error->engine[i].context.pid);
|
|
|
|
}
|
|
|
|
if (msg)
|
|
|
|
len += scnprintf(error->error_msg + len,
|
|
|
|
sizeof(error->error_msg) - len,
|
|
|
|
", %s", msg);
|
2014-02-25 23:11:26 +08:00
|
|
|
|
2019-01-25 21:22:28 +08:00
|
|
|
return error->error_msg;
|
2014-02-25 23:11:25 +08:00
|
|
|
}
|
|
|
|
|
2018-03-06 06:21:21 +08:00
|
|
|
static void capture_gen_state(struct i915_gpu_state *error)
|
2014-02-25 23:11:27 +08:00
|
|
|
{
|
2018-03-06 06:21:21 +08:00
|
|
|
struct drm_i915_private *i915 = error->i915;
|
|
|
|
|
|
|
|
error->awake = i915->gt.awake;
|
|
|
|
error->wakelock = atomic_read(&i915->runtime_pm.wakeref_count);
|
|
|
|
error->suspended = i915->runtime_pm.suspended;
|
2017-03-02 23:03:56 +08:00
|
|
|
|
2015-08-08 03:24:15 +08:00
|
|
|
error->iommu = -1;
|
|
|
|
#ifdef CONFIG_INTEL_IOMMU
|
|
|
|
error->iommu = intel_iommu_gfx_mapped;
|
|
|
|
#endif
|
2018-03-06 06:21:21 +08:00
|
|
|
error->reset_count = i915_reset_count(&i915->gpu_error);
|
|
|
|
error->suspend_count = i915->suspend_count;
|
2016-08-15 17:48:45 +08:00
|
|
|
|
|
|
|
memcpy(&error->device_info,
|
2018-03-06 06:21:21 +08:00
|
|
|
INTEL_INFO(i915),
|
2016-08-15 17:48:45 +08:00
|
|
|
sizeof(error->device_info));
|
2018-12-31 22:56:41 +08:00
|
|
|
memcpy(&error->runtime_info,
|
|
|
|
RUNTIME_INFO(i915),
|
|
|
|
sizeof(error->runtime_info));
|
2018-03-06 06:21:21 +08:00
|
|
|
error->driver_caps = i915->caps;
|
2014-02-25 23:11:27 +08:00
|
|
|
}
|
|
|
|
|
2017-10-27 01:36:57 +08:00
|
|
|
static void capture_params(struct i915_gpu_state *error)
|
|
|
|
{
|
2018-12-27 22:33:37 +08:00
|
|
|
i915_params_copy(&error->params, &i915_modparams);
|
2017-10-27 01:36:57 +08:00
|
|
|
}
|
|
|
|
|
2018-04-30 15:52:59 +08:00
|
|
|
static unsigned long capture_find_epoch(const struct i915_gpu_state *error)
|
|
|
|
{
|
|
|
|
unsigned long epoch = error->capture;
|
|
|
|
int i;
|
|
|
|
|
|
|
|
for (i = 0; i < ARRAY_SIZE(error->engine); i++) {
|
|
|
|
const struct drm_i915_error_engine *ee = &error->engine[i];
|
|
|
|
|
2019-01-25 21:22:28 +08:00
|
|
|
if (ee->hangcheck_timestamp &&
|
2018-04-30 15:52:59 +08:00
|
|
|
time_before(ee->hangcheck_timestamp, epoch))
|
|
|
|
epoch = ee->hangcheck_timestamp;
|
|
|
|
}
|
|
|
|
|
|
|
|
return epoch;
|
|
|
|
}
|
|
|
|
|
2018-10-02 03:44:46 +08:00
|
|
|
static void capture_finish(struct i915_gpu_state *error)
|
|
|
|
{
|
|
|
|
struct i915_ggtt *ggtt = &error->i915->ggtt;
|
|
|
|
const u64 slot = ggtt->error_capture.start;
|
|
|
|
|
|
|
|
ggtt->vm.clear_range(&ggtt->vm, slot, PAGE_SIZE);
|
|
|
|
}
|
|
|
|
|
2016-10-12 17:05:19 +08:00
|
|
|
static int capture(void *data)
|
|
|
|
{
|
2017-02-15 00:46:11 +08:00
|
|
|
struct i915_gpu_state *error = data;
|
2016-10-12 17:05:19 +08:00
|
|
|
|
2018-01-17 23:48:53 +08:00
|
|
|
error->time = ktime_get_real();
|
|
|
|
error->boottime = ktime_get_boottime();
|
|
|
|
error->uptime = ktime_sub(ktime_get(),
|
|
|
|
error->i915->gt.last_init_time);
|
2018-04-30 15:52:59 +08:00
|
|
|
error->capture = jiffies;
|
2017-02-07 05:36:07 +08:00
|
|
|
|
2017-10-27 01:36:57 +08:00
|
|
|
capture_params(error);
|
2018-03-06 06:21:21 +08:00
|
|
|
capture_gen_state(error);
|
2018-03-06 06:21:22 +08:00
|
|
|
capture_uc_state(error);
|
2018-03-06 06:21:21 +08:00
|
|
|
capture_reg_state(error);
|
|
|
|
gem_record_fences(error);
|
|
|
|
gem_record_rings(error);
|
|
|
|
capture_active_buffers(error);
|
|
|
|
capture_pinned_buffers(error);
|
2016-10-12 17:05:19 +08:00
|
|
|
|
|
|
|
error->overlay = intel_overlay_capture_error_state(error->i915);
|
|
|
|
error->display = intel_display_capture_error_state(error->i915);
|
|
|
|
|
2018-04-30 15:52:59 +08:00
|
|
|
error->epoch = capture_find_epoch(error);
|
|
|
|
|
2018-10-02 03:44:46 +08:00
|
|
|
capture_finish(error);
|
2016-10-12 17:05:19 +08:00
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2016-10-14 21:44:28 +08:00
|
|
|
#define DAY_AS_SECONDS(x) (24 * 60 * 60 * (x))
|
|
|
|
|
2017-02-15 00:46:11 +08:00
|
|
|
struct i915_gpu_state *
|
|
|
|
i915_capture_gpu_state(struct drm_i915_private *i915)
|
|
|
|
{
|
|
|
|
struct i915_gpu_state *error;
|
|
|
|
|
2018-12-07 19:05:54 +08:00
|
|
|
/* Check if GPU capture has been disabled */
|
|
|
|
error = READ_ONCE(i915->gpu_error.first_error);
|
|
|
|
if (IS_ERR(error))
|
|
|
|
return error;
|
|
|
|
|
2017-02-15 00:46:11 +08:00
|
|
|
error = kzalloc(sizeof(*error), GFP_ATOMIC);
|
2018-12-07 19:05:54 +08:00
|
|
|
if (!error) {
|
|
|
|
i915_disable_error_state(i915, -ENOMEM);
|
|
|
|
return ERR_PTR(-ENOMEM);
|
|
|
|
}
|
2017-02-15 00:46:11 +08:00
|
|
|
|
|
|
|
kref_init(&error->ref);
|
|
|
|
error->i915 = i915;
|
|
|
|
|
|
|
|
stop_machine(capture, error, NULL);
|
|
|
|
|
|
|
|
return error;
|
|
|
|
}
|
|
|
|
|
2014-01-30 16:19:35 +08:00
|
|
|
/**
|
|
|
|
* i915_capture_error_state - capture an error record for later analysis
|
2018-02-08 19:11:05 +08:00
|
|
|
* @i915: i915 device
|
|
|
|
* @engine_mask: the mask of engines triggering the hang
|
2019-01-25 21:22:28 +08:00
|
|
|
* @msg: a message to insert into the error capture header
|
2014-01-30 16:19:35 +08:00
|
|
|
*
|
|
|
|
* Should be called when an error is detected (either a hang or an error
|
|
|
|
* interrupt) to capture error state from the time of the error. Fills
|
|
|
|
* out a structure which becomes available in debugfs for user level tools
|
|
|
|
* to pick up.
|
|
|
|
*/
|
2018-02-08 19:11:05 +08:00
|
|
|
void i915_capture_error_state(struct drm_i915_private *i915,
|
2019-01-25 21:22:28 +08:00
|
|
|
unsigned long engine_mask,
|
|
|
|
const char *msg)
|
2014-01-30 16:19:35 +08:00
|
|
|
{
|
2014-01-30 22:38:15 +08:00
|
|
|
static bool warned;
|
2017-02-15 00:46:11 +08:00
|
|
|
struct i915_gpu_state *error;
|
2014-01-30 16:19:35 +08:00
|
|
|
unsigned long flags;
|
|
|
|
|
2017-09-20 03:38:44 +08:00
|
|
|
if (!i915_modparams.error_capture)
|
2016-10-12 17:05:18 +08:00
|
|
|
return;
|
|
|
|
|
2018-02-08 19:11:05 +08:00
|
|
|
if (READ_ONCE(i915->gpu_error.first_error))
|
2016-07-04 15:48:33 +08:00
|
|
|
return;
|
|
|
|
|
2018-02-08 19:11:05 +08:00
|
|
|
error = i915_capture_gpu_state(i915);
|
2018-12-07 19:05:54 +08:00
|
|
|
if (IS_ERR(error))
|
2014-01-30 16:19:35 +08:00
|
|
|
return;
|
|
|
|
|
2019-01-25 21:22:28 +08:00
|
|
|
dev_info(i915->drm.dev, "%s\n", error_msg(error, engine_mask, msg));
|
2014-02-25 23:11:25 +08:00
|
|
|
|
2016-07-04 15:08:39 +08:00
|
|
|
if (!error->simulated) {
|
2018-02-08 19:11:05 +08:00
|
|
|
spin_lock_irqsave(&i915->gpu_error.lock, flags);
|
|
|
|
if (!i915->gpu_error.first_error) {
|
|
|
|
i915->gpu_error.first_error = error;
|
2016-07-04 15:08:39 +08:00
|
|
|
error = NULL;
|
|
|
|
}
|
2018-02-08 19:11:05 +08:00
|
|
|
spin_unlock_irqrestore(&i915->gpu_error.lock, flags);
|
2013-07-12 21:50:57 +08:00
|
|
|
}
|
|
|
|
|
2014-02-25 23:11:25 +08:00
|
|
|
if (error) {
|
2017-02-15 00:46:11 +08:00
|
|
|
__i915_gpu_state_free(&error->ref);
|
2014-02-25 23:11:25 +08:00
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
2016-10-14 21:44:28 +08:00
|
|
|
if (!warned &&
|
|
|
|
ktime_get_real_seconds() - DRIVER_TIMESTAMP < DAY_AS_SECONDS(180)) {
|
2014-02-25 23:11:25 +08:00
|
|
|
DRM_INFO("GPU hangs can indicate a bug anywhere in the entire gfx stack, including userspace.\n");
|
|
|
|
DRM_INFO("Please file a _new_ bug report on bugs.freedesktop.org against DRI -> DRM/Intel\n");
|
|
|
|
DRM_INFO("drm/i915 developers can then reassign to the right component if it's not a kernel issue.\n");
|
|
|
|
DRM_INFO("The gpu crash dump is required to analyze gpu hangs, so please always attach it.\n");
|
2016-07-05 17:40:23 +08:00
|
|
|
DRM_INFO("GPU crash dump saved to /sys/class/drm/card%d/error\n",
|
2018-02-08 19:11:05 +08:00
|
|
|
i915->drm.primary->index);
|
2014-02-25 23:11:25 +08:00
|
|
|
warned = true;
|
|
|
|
}
|
2013-07-12 21:50:57 +08:00
|
|
|
}
|
|
|
|
|
2017-02-15 00:46:11 +08:00
|
|
|
struct i915_gpu_state *
|
|
|
|
i915_first_error_state(struct drm_i915_private *i915)
|
2013-07-12 21:50:57 +08:00
|
|
|
{
|
2017-02-15 00:46:11 +08:00
|
|
|
struct i915_gpu_state *error;
|
2013-07-12 21:50:57 +08:00
|
|
|
|
2017-02-15 00:46:11 +08:00
|
|
|
spin_lock_irq(&i915->gpu_error.lock);
|
|
|
|
error = i915->gpu_error.first_error;
|
2018-12-07 19:05:54 +08:00
|
|
|
if (!IS_ERR_OR_NULL(error))
|
2017-02-15 00:46:11 +08:00
|
|
|
i915_gpu_state_get(error);
|
|
|
|
spin_unlock_irq(&i915->gpu_error.lock);
|
2013-07-12 21:50:57 +08:00
|
|
|
|
2017-02-15 00:46:11 +08:00
|
|
|
return error;
|
2013-07-12 21:50:57 +08:00
|
|
|
}
|
|
|
|
|
2017-02-15 00:46:11 +08:00
|
|
|
void i915_reset_error_state(struct drm_i915_private *i915)
|
2013-07-12 21:50:57 +08:00
|
|
|
{
|
2017-02-15 00:46:11 +08:00
|
|
|
struct i915_gpu_state *error;
|
2013-07-12 21:50:57 +08:00
|
|
|
|
2017-02-15 00:46:11 +08:00
|
|
|
spin_lock_irq(&i915->gpu_error.lock);
|
|
|
|
error = i915->gpu_error.first_error;
|
2018-12-07 19:05:54 +08:00
|
|
|
if (error != ERR_PTR(-ENODEV)) /* if disabled, always disabled */
|
|
|
|
i915->gpu_error.first_error = NULL;
|
2017-02-15 00:46:11 +08:00
|
|
|
spin_unlock_irq(&i915->gpu_error.lock);
|
2013-07-12 21:50:57 +08:00
|
|
|
|
2018-12-07 19:05:54 +08:00
|
|
|
if (!IS_ERR_OR_NULL(error))
|
2018-11-03 00:12:12 +08:00
|
|
|
i915_gpu_state_put(error);
|
|
|
|
}
|
|
|
|
|
|
|
|
void i915_disable_error_state(struct drm_i915_private *i915, int err)
|
|
|
|
{
|
|
|
|
spin_lock_irq(&i915->gpu_error.lock);
|
|
|
|
if (!i915->gpu_error.first_error)
|
|
|
|
i915->gpu_error.first_error = ERR_PTR(err);
|
|
|
|
spin_unlock_irq(&i915->gpu_error.lock);
|
2013-07-12 21:50:57 +08:00
|
|
|
}
|