Merge tag 'drm-intel-next-fixes-2017-09-07' of git://anongit.freedesktop.org/git/drm-intel
Pull i916 drm fixes from Rodrigo Vivi: "Since Dave is on paternity leave we are sending drm/i915 fixes for v4.14-rc1 directly to you as he had asked us to do. The most critical ones are the GPU reset fix for gen2-4 and GVT fix for a regression that is blocking gvt init to work on your tree. The rest is general fixes for patches coming from drm-next" Acked-by: Dave Airlie <airlied@redhat.com> * tag 'drm-intel-next-fixes-2017-09-07' of git://anongit.freedesktop.org/git/drm-intel: drm/i915: Re-enable GTT following a device reset drm/i915: Annotate user relocs with __user drm/i915: Silence sparse by using gfp_t drm/i915: Add __rcu to radix tree slot pointer drm/i915: Fix the missing PPAT cache attributes on CNL drm/i915/gvt: Remove one duplicated MMIO drm/i915: Fix enum pipe vs. enum transcoder for the PCH transcoder drm/i915: Make i2c lock ops static drm/i915: Make i9xx_load_ycbcr_conversion_matrix() static drm/i915/edp: Increase T12 panel delay to 900 ms to fix DP AUX CH timeouts drm/i915: Ignore duplicate VMA stored within the per-object handle LUT drm/i915: Skip fence alignemnt check for the CCS plane drm/i915: Treat fb->offsets[] as a raw byte offset instead of a linear offset drm/i915: Always wake the device to flush the GTT drm/i915: Recreate vmapping even when the object is pinned drm/i915: Quietly cancel FBC activation if CRTC is turned off before worker
This commit is contained in:
commit
7d95565612
|
@ -2659,7 +2659,6 @@ static int init_skl_mmio_info(struct intel_gvt *gvt)
|
|||
MMIO_D(HSW_PWR_WELL_CTL_BIOS(SKL_DISP_PW_MISC_IO), D_SKL_PLUS);
|
||||
MMIO_DH(HSW_PWR_WELL_CTL_DRIVER(SKL_DISP_PW_MISC_IO), D_SKL_PLUS, NULL,
|
||||
skl_power_well_ctl_write);
|
||||
MMIO_DH(GEN6_PCODE_MAILBOX, D_SKL_PLUS, NULL, mailbox_write);
|
||||
|
||||
MMIO_D(0xa210, D_SKL_PLUS);
|
||||
MMIO_D(GEN9_MEDIA_PG_IDLE_HYSTERESIS, D_SKL_PLUS);
|
||||
|
|
|
@ -1073,7 +1073,7 @@ static u32 *copy_batch(struct drm_i915_gem_object *dst_obj,
|
|||
goto unpin_src;
|
||||
}
|
||||
|
||||
dst = i915_gem_object_pin_map(dst_obj, I915_MAP_WB);
|
||||
dst = i915_gem_object_pin_map(dst_obj, I915_MAP_FORCE_WB);
|
||||
if (IS_ERR(dst))
|
||||
goto unpin_dst;
|
||||
|
||||
|
|
|
@ -1891,9 +1891,15 @@ void i915_reset(struct drm_i915_private *i915, unsigned int flags)
|
|||
|
||||
/*
|
||||
* Everything depends on having the GTT running, so we need to start
|
||||
* there. Fortunately we don't need to do this unless we reset the
|
||||
* chip at a PCI level.
|
||||
*
|
||||
* there.
|
||||
*/
|
||||
ret = i915_ggtt_enable_hw(i915);
|
||||
if (ret) {
|
||||
DRM_ERROR("Failed to re-enable GGTT following reset %d\n", ret);
|
||||
goto error;
|
||||
}
|
||||
|
||||
/*
|
||||
* Next we need to restore the context, but we don't use those
|
||||
* yet either...
|
||||
*
|
||||
|
|
|
@ -3479,6 +3479,9 @@ void __i915_gem_object_invalidate(struct drm_i915_gem_object *obj);
|
|||
enum i915_map_type {
|
||||
I915_MAP_WB = 0,
|
||||
I915_MAP_WC,
|
||||
#define I915_MAP_OVERRIDE BIT(31)
|
||||
I915_MAP_FORCE_WB = I915_MAP_WB | I915_MAP_OVERRIDE,
|
||||
I915_MAP_FORCE_WC = I915_MAP_WC | I915_MAP_OVERRIDE,
|
||||
};
|
||||
|
||||
/**
|
||||
|
|
|
@ -695,12 +695,11 @@ flush_write_domain(struct drm_i915_gem_object *obj, unsigned int flush_domains)
|
|||
switch (obj->base.write_domain) {
|
||||
case I915_GEM_DOMAIN_GTT:
|
||||
if (INTEL_GEN(dev_priv) >= 6 && !HAS_LLC(dev_priv)) {
|
||||
if (intel_runtime_pm_get_if_in_use(dev_priv)) {
|
||||
spin_lock_irq(&dev_priv->uncore.lock);
|
||||
POSTING_READ_FW(RING_ACTHD(dev_priv->engine[RCS]->mmio_base));
|
||||
spin_unlock_irq(&dev_priv->uncore.lock);
|
||||
intel_runtime_pm_put(dev_priv);
|
||||
}
|
||||
intel_runtime_pm_get(dev_priv);
|
||||
spin_lock_irq(&dev_priv->uncore.lock);
|
||||
POSTING_READ_FW(RING_ACTHD(dev_priv->engine[RCS]->mmio_base));
|
||||
spin_unlock_irq(&dev_priv->uncore.lock);
|
||||
intel_runtime_pm_put(dev_priv);
|
||||
}
|
||||
|
||||
intel_fb_obj_flush(obj,
|
||||
|
@ -2213,7 +2212,7 @@ i915_gem_object_put_pages_gtt(struct drm_i915_gem_object *obj,
|
|||
static void __i915_gem_object_reset_page_iter(struct drm_i915_gem_object *obj)
|
||||
{
|
||||
struct radix_tree_iter iter;
|
||||
void **slot;
|
||||
void __rcu **slot;
|
||||
|
||||
radix_tree_for_each_slot(slot, &obj->mm.get_page.radix, &iter, 0)
|
||||
radix_tree_delete(&obj->mm.get_page.radix, iter.index);
|
||||
|
@ -2553,6 +2552,9 @@ static void *i915_gem_object_map(const struct drm_i915_gem_object *obj,
|
|||
GEM_BUG_ON(i != n_pages);
|
||||
|
||||
switch (type) {
|
||||
default:
|
||||
MISSING_CASE(type);
|
||||
/* fallthrough to use PAGE_KERNEL anyway */
|
||||
case I915_MAP_WB:
|
||||
pgprot = PAGE_KERNEL;
|
||||
break;
|
||||
|
@ -2583,7 +2585,9 @@ void *i915_gem_object_pin_map(struct drm_i915_gem_object *obj,
|
|||
if (ret)
|
||||
return ERR_PTR(ret);
|
||||
|
||||
pinned = true;
|
||||
pinned = !(type & I915_MAP_OVERRIDE);
|
||||
type &= ~I915_MAP_OVERRIDE;
|
||||
|
||||
if (!atomic_inc_not_zero(&obj->mm.pages_pin_count)) {
|
||||
if (unlikely(IS_ERR_OR_NULL(obj->mm.pages))) {
|
||||
ret = ____i915_gem_object_get_pages(obj);
|
||||
|
@ -3258,7 +3262,13 @@ void i915_gem_close_object(struct drm_gem_object *gem, struct drm_file *file)
|
|||
|
||||
vma = radix_tree_delete(&ctx->handles_vma, lut->handle);
|
||||
|
||||
if (!i915_vma_is_ggtt(vma))
|
||||
GEM_BUG_ON(vma->obj != obj);
|
||||
|
||||
/* We allow the process to have multiple handles to the same
|
||||
* vma, in the same fd namespace, by virtue of flink/open.
|
||||
*/
|
||||
GEM_BUG_ON(!vma->open_count);
|
||||
if (!--vma->open_count && !i915_vma_is_ggtt(vma))
|
||||
i915_vma_close(vma);
|
||||
|
||||
list_del(&lut->obj_link);
|
||||
|
|
|
@ -285,7 +285,7 @@ static int eb_create(struct i915_execbuffer *eb)
|
|||
* direct lookup.
|
||||
*/
|
||||
do {
|
||||
unsigned int flags;
|
||||
gfp_t flags;
|
||||
|
||||
/* While we can still reduce the allocation size, don't
|
||||
* raise a warning and allow the allocation to fail.
|
||||
|
@ -720,6 +720,7 @@ static int eb_lookup_vmas(struct i915_execbuffer *eb)
|
|||
goto err_obj;
|
||||
}
|
||||
|
||||
vma->open_count++;
|
||||
list_add(&lut->obj_link, &obj->lut_list);
|
||||
list_add(&lut->ctx_link, &eb->ctx->handles_list);
|
||||
lut->ctx = eb->ctx;
|
||||
|
@ -1070,7 +1071,9 @@ static int __reloc_gpu_alloc(struct i915_execbuffer *eb,
|
|||
return PTR_ERR(obj);
|
||||
|
||||
cmd = i915_gem_object_pin_map(obj,
|
||||
cache->has_llc ? I915_MAP_WB : I915_MAP_WC);
|
||||
cache->has_llc ?
|
||||
I915_MAP_FORCE_WB :
|
||||
I915_MAP_FORCE_WC);
|
||||
i915_gem_object_unpin_pages(obj);
|
||||
if (IS_ERR(cmd))
|
||||
return PTR_ERR(cmd);
|
||||
|
@ -1526,7 +1529,7 @@ static int eb_copy_relocations(const struct i915_execbuffer *eb)
|
|||
min_t(u64, BIT_ULL(31), size - copied);
|
||||
|
||||
if (__copy_from_user((char *)relocs + copied,
|
||||
(char *)urelocs + copied,
|
||||
(char __user *)urelocs + copied,
|
||||
len)) {
|
||||
kvfree(relocs);
|
||||
err = -EFAULT;
|
||||
|
|
|
@ -2754,10 +2754,10 @@ static void cnl_setup_private_ppat(struct drm_i915_private *dev_priv)
|
|||
I915_WRITE(GEN10_PAT_INDEX(1), GEN8_PPAT_WC | GEN8_PPAT_LLCELLC);
|
||||
I915_WRITE(GEN10_PAT_INDEX(2), GEN8_PPAT_WT | GEN8_PPAT_LLCELLC);
|
||||
I915_WRITE(GEN10_PAT_INDEX(3), GEN8_PPAT_UC);
|
||||
I915_WRITE(GEN10_PAT_INDEX(4), GEN8_PPAT_LLCELLC | GEN8_PPAT_AGE(0));
|
||||
I915_WRITE(GEN10_PAT_INDEX(5), GEN8_PPAT_LLCELLC | GEN8_PPAT_AGE(1));
|
||||
I915_WRITE(GEN10_PAT_INDEX(6), GEN8_PPAT_LLCELLC | GEN8_PPAT_AGE(2));
|
||||
I915_WRITE(GEN10_PAT_INDEX(7), GEN8_PPAT_LLCELLC | GEN8_PPAT_AGE(3));
|
||||
I915_WRITE(GEN10_PAT_INDEX(4), GEN8_PPAT_WB | GEN8_PPAT_LLCELLC | GEN8_PPAT_AGE(0));
|
||||
I915_WRITE(GEN10_PAT_INDEX(5), GEN8_PPAT_WB | GEN8_PPAT_LLCELLC | GEN8_PPAT_AGE(1));
|
||||
I915_WRITE(GEN10_PAT_INDEX(6), GEN8_PPAT_WB | GEN8_PPAT_LLCELLC | GEN8_PPAT_AGE(2));
|
||||
I915_WRITE(GEN10_PAT_INDEX(7), GEN8_PPAT_WB | GEN8_PPAT_LLCELLC | GEN8_PPAT_AGE(3));
|
||||
}
|
||||
|
||||
/* The GGTT and PPGTT need a private PPAT setup in order to handle cacheability
|
||||
|
|
|
@ -38,7 +38,7 @@ TRACE_EVENT(intel_cpu_fifo_underrun,
|
|||
);
|
||||
|
||||
TRACE_EVENT(intel_pch_fifo_underrun,
|
||||
TP_PROTO(struct drm_i915_private *dev_priv, enum transcoder pch_transcoder),
|
||||
TP_PROTO(struct drm_i915_private *dev_priv, enum pipe pch_transcoder),
|
||||
TP_ARGS(dev_priv, pch_transcoder),
|
||||
|
||||
TP_STRUCT__entry(
|
||||
|
@ -48,7 +48,7 @@ TRACE_EVENT(intel_pch_fifo_underrun,
|
|||
),
|
||||
|
||||
TP_fast_assign(
|
||||
enum pipe pipe = (enum pipe)pch_transcoder;
|
||||
enum pipe pipe = pch_transcoder;
|
||||
__entry->pipe = pipe;
|
||||
__entry->frame = dev_priv->drm.driver->get_vblank_counter(&dev_priv->drm, pipe);
|
||||
__entry->scanline = intel_get_crtc_scanline(intel_get_crtc_for_pipe(dev_priv, pipe));
|
||||
|
|
|
@ -59,6 +59,12 @@ struct i915_vma {
|
|||
u32 fence_size;
|
||||
u32 fence_alignment;
|
||||
|
||||
/**
|
||||
* Count of the number of times this vma has been opened by different
|
||||
* handles (but same file) for execbuf, i.e. the number of aliases
|
||||
* that exist in the ctx->handle_vmas LUT for this vma.
|
||||
*/
|
||||
unsigned int open_count;
|
||||
unsigned int flags;
|
||||
/**
|
||||
* How many users have pinned this object in GTT space. The following
|
||||
|
|
|
@ -107,7 +107,7 @@ static void ctm_mult_by_limited(uint64_t *result, int64_t *input)
|
|||
}
|
||||
}
|
||||
|
||||
void i9xx_load_ycbcr_conversion_matrix(struct intel_crtc *intel_crtc)
|
||||
static void i9xx_load_ycbcr_conversion_matrix(struct intel_crtc *intel_crtc)
|
||||
{
|
||||
int pipe = intel_crtc->pipe;
|
||||
struct drm_i915_private *dev_priv = to_i915(intel_crtc->base.dev);
|
||||
|
|
|
@ -2288,17 +2288,13 @@ void intel_add_fb_offsets(int *x, int *y,
|
|||
}
|
||||
}
|
||||
|
||||
/*
|
||||
* Input tile dimensions and pitch must already be
|
||||
* rotated to match x and y, and in pixel units.
|
||||
*/
|
||||
static u32 _intel_adjust_tile_offset(int *x, int *y,
|
||||
unsigned int tile_width,
|
||||
unsigned int tile_height,
|
||||
unsigned int tile_size,
|
||||
unsigned int pitch_tiles,
|
||||
u32 old_offset,
|
||||
u32 new_offset)
|
||||
static u32 __intel_adjust_tile_offset(int *x, int *y,
|
||||
unsigned int tile_width,
|
||||
unsigned int tile_height,
|
||||
unsigned int tile_size,
|
||||
unsigned int pitch_tiles,
|
||||
u32 old_offset,
|
||||
u32 new_offset)
|
||||
{
|
||||
unsigned int pitch_pixels = pitch_tiles * tile_width;
|
||||
unsigned int tiles;
|
||||
|
@ -2319,18 +2315,13 @@ static u32 _intel_adjust_tile_offset(int *x, int *y,
|
|||
return new_offset;
|
||||
}
|
||||
|
||||
/*
|
||||
* Adjust the tile offset by moving the difference into
|
||||
* the x/y offsets.
|
||||
*/
|
||||
static u32 intel_adjust_tile_offset(int *x, int *y,
|
||||
const struct intel_plane_state *state, int plane,
|
||||
u32 old_offset, u32 new_offset)
|
||||
static u32 _intel_adjust_tile_offset(int *x, int *y,
|
||||
const struct drm_framebuffer *fb, int plane,
|
||||
unsigned int rotation,
|
||||
u32 old_offset, u32 new_offset)
|
||||
{
|
||||
const struct drm_i915_private *dev_priv = to_i915(state->base.plane->dev);
|
||||
const struct drm_framebuffer *fb = state->base.fb;
|
||||
const struct drm_i915_private *dev_priv = to_i915(fb->dev);
|
||||
unsigned int cpp = fb->format->cpp[plane];
|
||||
unsigned int rotation = state->base.rotation;
|
||||
unsigned int pitch = intel_fb_pitch(fb, plane, rotation);
|
||||
|
||||
WARN_ON(new_offset > old_offset);
|
||||
|
@ -2349,9 +2340,9 @@ static u32 intel_adjust_tile_offset(int *x, int *y,
|
|||
pitch_tiles = pitch / (tile_width * cpp);
|
||||
}
|
||||
|
||||
_intel_adjust_tile_offset(x, y, tile_width, tile_height,
|
||||
tile_size, pitch_tiles,
|
||||
old_offset, new_offset);
|
||||
__intel_adjust_tile_offset(x, y, tile_width, tile_height,
|
||||
tile_size, pitch_tiles,
|
||||
old_offset, new_offset);
|
||||
} else {
|
||||
old_offset += *y * pitch + *x * cpp;
|
||||
|
||||
|
@ -2362,6 +2353,19 @@ static u32 intel_adjust_tile_offset(int *x, int *y,
|
|||
return new_offset;
|
||||
}
|
||||
|
||||
/*
|
||||
* Adjust the tile offset by moving the difference into
|
||||
* the x/y offsets.
|
||||
*/
|
||||
static u32 intel_adjust_tile_offset(int *x, int *y,
|
||||
const struct intel_plane_state *state, int plane,
|
||||
u32 old_offset, u32 new_offset)
|
||||
{
|
||||
return _intel_adjust_tile_offset(x, y, state->base.fb, plane,
|
||||
state->base.rotation,
|
||||
old_offset, new_offset);
|
||||
}
|
||||
|
||||
/*
|
||||
* Computes the linear offset to the base tile and adjusts
|
||||
* x, y. bytes per pixel is assumed to be a power-of-two.
|
||||
|
@ -2413,9 +2417,9 @@ static u32 _intel_compute_tile_offset(const struct drm_i915_private *dev_priv,
|
|||
offset = (tile_rows * pitch_tiles + tiles) * tile_size;
|
||||
offset_aligned = offset & ~alignment;
|
||||
|
||||
_intel_adjust_tile_offset(x, y, tile_width, tile_height,
|
||||
tile_size, pitch_tiles,
|
||||
offset, offset_aligned);
|
||||
__intel_adjust_tile_offset(x, y, tile_width, tile_height,
|
||||
tile_size, pitch_tiles,
|
||||
offset, offset_aligned);
|
||||
} else {
|
||||
offset = *y * pitch + *x * cpp;
|
||||
offset_aligned = offset & ~alignment;
|
||||
|
@ -2447,16 +2451,24 @@ u32 intel_compute_tile_offset(int *x, int *y,
|
|||
rotation, alignment);
|
||||
}
|
||||
|
||||
/* Convert the fb->offset[] linear offset into x/y offsets */
|
||||
static void intel_fb_offset_to_xy(int *x, int *y,
|
||||
const struct drm_framebuffer *fb, int plane)
|
||||
/* Convert the fb->offset[] into x/y offsets */
|
||||
static int intel_fb_offset_to_xy(int *x, int *y,
|
||||
const struct drm_framebuffer *fb, int plane)
|
||||
{
|
||||
unsigned int cpp = fb->format->cpp[plane];
|
||||
unsigned int pitch = fb->pitches[plane];
|
||||
u32 linear_offset = fb->offsets[plane];
|
||||
struct drm_i915_private *dev_priv = to_i915(fb->dev);
|
||||
|
||||
*y = linear_offset / pitch;
|
||||
*x = linear_offset % pitch / cpp;
|
||||
if (fb->modifier != DRM_FORMAT_MOD_LINEAR &&
|
||||
fb->offsets[plane] % intel_tile_size(dev_priv))
|
||||
return -EINVAL;
|
||||
|
||||
*x = 0;
|
||||
*y = 0;
|
||||
|
||||
_intel_adjust_tile_offset(x, y,
|
||||
fb, plane, DRM_MODE_ROTATE_0,
|
||||
fb->offsets[plane], 0);
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
static unsigned int intel_fb_modifier_to_tiling(uint64_t fb_modifier)
|
||||
|
@ -2523,12 +2535,18 @@ intel_fill_fb_info(struct drm_i915_private *dev_priv,
|
|||
unsigned int cpp, size;
|
||||
u32 offset;
|
||||
int x, y;
|
||||
int ret;
|
||||
|
||||
cpp = fb->format->cpp[i];
|
||||
width = drm_framebuffer_plane_width(fb->width, fb, i);
|
||||
height = drm_framebuffer_plane_height(fb->height, fb, i);
|
||||
|
||||
intel_fb_offset_to_xy(&x, &y, fb, i);
|
||||
ret = intel_fb_offset_to_xy(&x, &y, fb, i);
|
||||
if (ret) {
|
||||
DRM_DEBUG_KMS("bad fb plane %d offset: 0x%x\n",
|
||||
i, fb->offsets[i]);
|
||||
return ret;
|
||||
}
|
||||
|
||||
if ((fb->modifier == I915_FORMAT_MOD_Y_TILED_CCS ||
|
||||
fb->modifier == I915_FORMAT_MOD_Yf_TILED_CCS) && i == 1) {
|
||||
|
@ -2539,11 +2557,13 @@ intel_fill_fb_info(struct drm_i915_private *dev_priv,
|
|||
int ccs_x, ccs_y;
|
||||
|
||||
intel_tile_dims(fb, i, &tile_width, &tile_height);
|
||||
tile_width *= hsub;
|
||||
tile_height *= vsub;
|
||||
|
||||
ccs_x = (x * hsub) % (tile_width * hsub);
|
||||
ccs_y = (y * vsub) % (tile_height * vsub);
|
||||
main_x = intel_fb->normal[0].x % (tile_width * hsub);
|
||||
main_y = intel_fb->normal[0].y % (tile_height * vsub);
|
||||
ccs_x = (x * hsub) % tile_width;
|
||||
ccs_y = (y * vsub) % tile_height;
|
||||
main_x = intel_fb->normal[0].x % tile_width;
|
||||
main_y = intel_fb->normal[0].y % tile_height;
|
||||
|
||||
/*
|
||||
* CCS doesn't have its own x/y offset register, so the intra CCS tile
|
||||
|
@ -2569,7 +2589,7 @@ intel_fill_fb_info(struct drm_i915_private *dev_priv,
|
|||
* fb layout agrees with the fence layout. We already check that the
|
||||
* fb stride matches the fence stride elsewhere.
|
||||
*/
|
||||
if (i915_gem_object_is_tiled(intel_fb->obj) &&
|
||||
if (i == 0 && i915_gem_object_is_tiled(intel_fb->obj) &&
|
||||
(x + width) * cpp > fb->pitches[i]) {
|
||||
DRM_DEBUG_KMS("bad fb plane %d offset: 0x%x\n",
|
||||
i, fb->offsets[i]);
|
||||
|
@ -2632,10 +2652,10 @@ intel_fill_fb_info(struct drm_i915_private *dev_priv,
|
|||
* We only keep the x/y offsets, so push all of the
|
||||
* gtt offset into the x/y offsets.
|
||||
*/
|
||||
_intel_adjust_tile_offset(&x, &y,
|
||||
tile_width, tile_height,
|
||||
tile_size, pitch_tiles,
|
||||
gtt_offset_rotated * tile_size, 0);
|
||||
__intel_adjust_tile_offset(&x, &y,
|
||||
tile_width, tile_height,
|
||||
tile_size, pitch_tiles,
|
||||
gtt_offset_rotated * tile_size, 0);
|
||||
|
||||
gtt_offset_rotated += rot_info->plane[i].width * rot_info->plane[i].height;
|
||||
|
||||
|
|
|
@ -5273,7 +5273,7 @@ intel_dp_init_panel_power_sequencer(struct drm_device *dev,
|
|||
* seems sufficient to avoid this problem.
|
||||
*/
|
||||
if (dev_priv->quirks & QUIRK_INCREASE_T12_DELAY) {
|
||||
vbt.t11_t12 = max_t(u16, vbt.t11_t12, 800 * 10);
|
||||
vbt.t11_t12 = max_t(u16, vbt.t11_t12, 900 * 10);
|
||||
DRM_DEBUG_KMS("Increasing T12 panel delay as per the quirk to %d\n",
|
||||
vbt.t11_t12);
|
||||
}
|
||||
|
|
|
@ -406,9 +406,7 @@ static void intel_fbc_work_fn(struct work_struct *__work)
|
|||
struct drm_vblank_crtc *vblank = &dev_priv->drm.vblank[crtc->pipe];
|
||||
|
||||
if (drm_crtc_vblank_get(&crtc->base)) {
|
||||
DRM_ERROR("vblank not available for FBC on pipe %c\n",
|
||||
pipe_name(crtc->pipe));
|
||||
|
||||
/* CRTC is now off, leave FBC deactivated */
|
||||
mutex_lock(&fbc->lock);
|
||||
work->scheduled = false;
|
||||
mutex_unlock(&fbc->lock);
|
||||
|
|
|
@ -187,11 +187,11 @@ static void broadwell_set_fifo_underrun_reporting(struct drm_device *dev,
|
|||
}
|
||||
|
||||
static void ibx_set_fifo_underrun_reporting(struct drm_device *dev,
|
||||
enum transcoder pch_transcoder,
|
||||
enum pipe pch_transcoder,
|
||||
bool enable)
|
||||
{
|
||||
struct drm_i915_private *dev_priv = to_i915(dev);
|
||||
uint32_t bit = (pch_transcoder == TRANSCODER_A) ?
|
||||
uint32_t bit = (pch_transcoder == PIPE_A) ?
|
||||
SDE_TRANSA_FIFO_UNDER : SDE_TRANSB_FIFO_UNDER;
|
||||
|
||||
if (enable)
|
||||
|
@ -203,7 +203,7 @@ static void ibx_set_fifo_underrun_reporting(struct drm_device *dev,
|
|||
static void cpt_check_pch_fifo_underruns(struct intel_crtc *crtc)
|
||||
{
|
||||
struct drm_i915_private *dev_priv = to_i915(crtc->base.dev);
|
||||
enum transcoder pch_transcoder = (enum transcoder) crtc->pipe;
|
||||
enum pipe pch_transcoder = crtc->pipe;
|
||||
uint32_t serr_int = I915_READ(SERR_INT);
|
||||
|
||||
lockdep_assert_held(&dev_priv->irq_lock);
|
||||
|
@ -215,12 +215,12 @@ static void cpt_check_pch_fifo_underruns(struct intel_crtc *crtc)
|
|||
POSTING_READ(SERR_INT);
|
||||
|
||||
trace_intel_pch_fifo_underrun(dev_priv, pch_transcoder);
|
||||
DRM_ERROR("pch fifo underrun on pch transcoder %s\n",
|
||||
transcoder_name(pch_transcoder));
|
||||
DRM_ERROR("pch fifo underrun on pch transcoder %c\n",
|
||||
pipe_name(pch_transcoder));
|
||||
}
|
||||
|
||||
static void cpt_set_fifo_underrun_reporting(struct drm_device *dev,
|
||||
enum transcoder pch_transcoder,
|
||||
enum pipe pch_transcoder,
|
||||
bool enable, bool old)
|
||||
{
|
||||
struct drm_i915_private *dev_priv = to_i915(dev);
|
||||
|
@ -238,8 +238,8 @@ static void cpt_set_fifo_underrun_reporting(struct drm_device *dev,
|
|||
|
||||
if (old && I915_READ(SERR_INT) &
|
||||
SERR_INT_TRANS_FIFO_UNDERRUN(pch_transcoder)) {
|
||||
DRM_ERROR("uncleared pch fifo underrun on pch transcoder %s\n",
|
||||
transcoder_name(pch_transcoder));
|
||||
DRM_ERROR("uncleared pch fifo underrun on pch transcoder %c\n",
|
||||
pipe_name(pch_transcoder));
|
||||
}
|
||||
}
|
||||
}
|
||||
|
@ -395,8 +395,8 @@ void intel_pch_fifo_underrun_irq_handler(struct drm_i915_private *dev_priv,
|
|||
if (intel_set_pch_fifo_underrun_reporting(dev_priv, pch_transcoder,
|
||||
false)) {
|
||||
trace_intel_pch_fifo_underrun(dev_priv, pch_transcoder);
|
||||
DRM_ERROR("PCH transcoder %s FIFO underrun\n",
|
||||
transcoder_name(pch_transcoder));
|
||||
DRM_ERROR("PCH transcoder %c FIFO underrun\n",
|
||||
pipe_name(pch_transcoder));
|
||||
}
|
||||
}
|
||||
|
||||
|
|
|
@ -649,7 +649,7 @@ static void gmbus_unlock_bus(struct i2c_adapter *adapter,
|
|||
mutex_unlock(&dev_priv->gmbus_mutex);
|
||||
}
|
||||
|
||||
const struct i2c_lock_operations gmbus_lock_ops = {
|
||||
static const struct i2c_lock_operations gmbus_lock_ops = {
|
||||
.lock_bus = gmbus_lock_bus,
|
||||
.trylock_bus = gmbus_trylock_bus,
|
||||
.unlock_bus = gmbus_unlock_bus,
|
||||
|
|
|
@ -2980,7 +2980,7 @@ static void proxy_unlock_bus(struct i2c_adapter *adapter,
|
|||
sdvo->i2c->lock_ops->unlock_bus(sdvo->i2c, flags);
|
||||
}
|
||||
|
||||
const struct i2c_lock_operations proxy_lock_ops = {
|
||||
static const struct i2c_lock_operations proxy_lock_ops = {
|
||||
.lock_bus = proxy_lock_bus,
|
||||
.trylock_bus = proxy_trylock_bus,
|
||||
.unlock_bus = proxy_unlock_bus,
|
||||
|
|
Loading…
Reference in New Issue