i915: Map status page cached for chips with GTT-based HWS location.
This should improve performance by avoiding uncached reads by the CPU (the point of having a status page), and may improve stability. This patch only affects G33, GM45 and G45 chips as those are the only ones using GTT-based HWS mappings. Signed-off-by: Keith Packard <keithp@keithp.com> Signed-off-by: Eric Anholt <eric@anholt.net> Signed-off-by: Dave Airlie <airlied@redhat.com>
This commit is contained in:
parent
50aa253d82
commit
ba1eb1d825
|
@ -464,7 +464,8 @@ DRM_AGP_MEM *
|
||||||
drm_agp_bind_pages(struct drm_device *dev,
|
drm_agp_bind_pages(struct drm_device *dev,
|
||||||
struct page **pages,
|
struct page **pages,
|
||||||
unsigned long num_pages,
|
unsigned long num_pages,
|
||||||
uint32_t gtt_offset)
|
uint32_t gtt_offset,
|
||||||
|
u32 type)
|
||||||
{
|
{
|
||||||
DRM_AGP_MEM *mem;
|
DRM_AGP_MEM *mem;
|
||||||
int ret, i;
|
int ret, i;
|
||||||
|
@ -472,7 +473,7 @@ drm_agp_bind_pages(struct drm_device *dev,
|
||||||
DRM_DEBUG("\n");
|
DRM_DEBUG("\n");
|
||||||
|
|
||||||
mem = drm_agp_allocate_memory(dev->agp->bridge, num_pages,
|
mem = drm_agp_allocate_memory(dev->agp->bridge, num_pages,
|
||||||
AGP_USER_MEMORY);
|
type);
|
||||||
if (mem == NULL) {
|
if (mem == NULL) {
|
||||||
DRM_ERROR("Failed to allocate memory for %ld pages\n",
|
DRM_ERROR("Failed to allocate memory for %ld pages\n",
|
||||||
num_pages);
|
num_pages);
|
||||||
|
|
|
@ -373,6 +373,9 @@ struct drm_i915_gem_object {
|
||||||
/** Current tiling mode for the object. */
|
/** Current tiling mode for the object. */
|
||||||
uint32_t tiling_mode;
|
uint32_t tiling_mode;
|
||||||
|
|
||||||
|
/** AGP mapping type (AGP_USER_MEMORY or AGP_USER_CACHED_MEMORY */
|
||||||
|
uint32_t agp_type;
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Flagging of which individual pages are valid in GEM_DOMAIN_CPU when
|
* Flagging of which individual pages are valid in GEM_DOMAIN_CPU when
|
||||||
* GEM_DOMAIN_CPU is not in the object's read domain.
|
* GEM_DOMAIN_CPU is not in the object's read domain.
|
||||||
|
|
|
@ -1159,7 +1159,8 @@ i915_gem_object_bind_to_gtt(struct drm_gem_object *obj, unsigned alignment)
|
||||||
obj_priv->agp_mem = drm_agp_bind_pages(dev,
|
obj_priv->agp_mem = drm_agp_bind_pages(dev,
|
||||||
obj_priv->page_list,
|
obj_priv->page_list,
|
||||||
page_count,
|
page_count,
|
||||||
obj_priv->gtt_offset);
|
obj_priv->gtt_offset,
|
||||||
|
obj_priv->agp_type);
|
||||||
if (obj_priv->agp_mem == NULL) {
|
if (obj_priv->agp_mem == NULL) {
|
||||||
i915_gem_object_free_page_list(obj);
|
i915_gem_object_free_page_list(obj);
|
||||||
drm_mm_put_block(obj_priv->gtt_space);
|
drm_mm_put_block(obj_priv->gtt_space);
|
||||||
|
@ -2142,6 +2143,8 @@ int i915_gem_init_object(struct drm_gem_object *obj)
|
||||||
obj->write_domain = I915_GEM_DOMAIN_CPU;
|
obj->write_domain = I915_GEM_DOMAIN_CPU;
|
||||||
obj->read_domains = I915_GEM_DOMAIN_CPU;
|
obj->read_domains = I915_GEM_DOMAIN_CPU;
|
||||||
|
|
||||||
|
obj_priv->agp_type = AGP_USER_MEMORY;
|
||||||
|
|
||||||
obj->driver_private = obj_priv;
|
obj->driver_private = obj_priv;
|
||||||
obj_priv->obj = obj;
|
obj_priv->obj = obj;
|
||||||
INIT_LIST_HEAD(&obj_priv->list);
|
INIT_LIST_HEAD(&obj_priv->list);
|
||||||
|
@ -2311,6 +2314,7 @@ i915_gem_init_hws(struct drm_device *dev)
|
||||||
return -ENOMEM;
|
return -ENOMEM;
|
||||||
}
|
}
|
||||||
obj_priv = obj->driver_private;
|
obj_priv = obj->driver_private;
|
||||||
|
obj_priv->agp_type = AGP_USER_CACHED_MEMORY;
|
||||||
|
|
||||||
ret = i915_gem_object_pin(obj, 4096);
|
ret = i915_gem_object_pin(obj, 4096);
|
||||||
if (ret != 0) {
|
if (ret != 0) {
|
||||||
|
@ -2319,25 +2323,18 @@ i915_gem_init_hws(struct drm_device *dev)
|
||||||
}
|
}
|
||||||
|
|
||||||
dev_priv->status_gfx_addr = obj_priv->gtt_offset;
|
dev_priv->status_gfx_addr = obj_priv->gtt_offset;
|
||||||
dev_priv->hws_map.offset = dev->agp->base + obj_priv->gtt_offset;
|
|
||||||
dev_priv->hws_map.size = 4096;
|
|
||||||
dev_priv->hws_map.type = 0;
|
|
||||||
dev_priv->hws_map.flags = 0;
|
|
||||||
dev_priv->hws_map.mtrr = 0;
|
|
||||||
|
|
||||||
/* Ioremapping here is the wrong thing to do. We want cached access.
|
dev_priv->hw_status_page = kmap(obj_priv->page_list[0]);
|
||||||
*/
|
if (dev_priv->hw_status_page == NULL) {
|
||||||
drm_core_ioremap_wc(&dev_priv->hws_map, dev);
|
|
||||||
if (dev_priv->hws_map.handle == NULL) {
|
|
||||||
DRM_ERROR("Failed to map status page.\n");
|
DRM_ERROR("Failed to map status page.\n");
|
||||||
memset(&dev_priv->hws_map, 0, sizeof(dev_priv->hws_map));
|
memset(&dev_priv->hws_map, 0, sizeof(dev_priv->hws_map));
|
||||||
drm_gem_object_unreference(obj);
|
drm_gem_object_unreference(obj);
|
||||||
return -EINVAL;
|
return -EINVAL;
|
||||||
}
|
}
|
||||||
dev_priv->hws_obj = obj;
|
dev_priv->hws_obj = obj;
|
||||||
dev_priv->hw_status_page = dev_priv->hws_map.handle;
|
|
||||||
memset(dev_priv->hw_status_page, 0, PAGE_SIZE);
|
memset(dev_priv->hw_status_page, 0, PAGE_SIZE);
|
||||||
I915_WRITE(HWS_PGA, dev_priv->status_gfx_addr);
|
I915_WRITE(HWS_PGA, dev_priv->status_gfx_addr);
|
||||||
|
I915_READ(HWS_PGA); /* posting read */
|
||||||
DRM_DEBUG("hws offset: 0x%08x\n", dev_priv->status_gfx_addr);
|
DRM_DEBUG("hws offset: 0x%08x\n", dev_priv->status_gfx_addr);
|
||||||
|
|
||||||
return 0;
|
return 0;
|
||||||
|
@ -2456,10 +2453,15 @@ i915_gem_cleanup_ringbuffer(struct drm_device *dev)
|
||||||
memset(&dev_priv->ring, 0, sizeof(dev_priv->ring));
|
memset(&dev_priv->ring, 0, sizeof(dev_priv->ring));
|
||||||
|
|
||||||
if (dev_priv->hws_obj != NULL) {
|
if (dev_priv->hws_obj != NULL) {
|
||||||
i915_gem_object_unpin(dev_priv->hws_obj);
|
struct drm_gem_object *obj = dev_priv->hws_obj;
|
||||||
drm_gem_object_unreference(dev_priv->hws_obj);
|
struct drm_i915_gem_object *obj_priv = obj->driver_private;
|
||||||
|
|
||||||
|
kunmap(obj_priv->page_list[0]);
|
||||||
|
i915_gem_object_unpin(obj);
|
||||||
|
drm_gem_object_unreference(obj);
|
||||||
dev_priv->hws_obj = NULL;
|
dev_priv->hws_obj = NULL;
|
||||||
memset(&dev_priv->hws_map, 0, sizeof(dev_priv->hws_map));
|
memset(&dev_priv->hws_map, 0, sizeof(dev_priv->hws_map));
|
||||||
|
dev_priv->hw_status_page = NULL;
|
||||||
|
|
||||||
/* Write high address into HWS_PGA when disabling. */
|
/* Write high address into HWS_PGA when disabling. */
|
||||||
I915_WRITE(HWS_PGA, 0x1ffff000);
|
I915_WRITE(HWS_PGA, 0x1ffff000);
|
||||||
|
|
|
@ -1016,7 +1016,8 @@ extern int drm_bind_agp(DRM_AGP_MEM * handle, unsigned int start);
|
||||||
extern DRM_AGP_MEM *drm_agp_bind_pages(struct drm_device *dev,
|
extern DRM_AGP_MEM *drm_agp_bind_pages(struct drm_device *dev,
|
||||||
struct page **pages,
|
struct page **pages,
|
||||||
unsigned long num_pages,
|
unsigned long num_pages,
|
||||||
uint32_t gtt_offset);
|
uint32_t gtt_offset,
|
||||||
|
uint32_t type);
|
||||||
extern int drm_unbind_agp(DRM_AGP_MEM * handle);
|
extern int drm_unbind_agp(DRM_AGP_MEM * handle);
|
||||||
|
|
||||||
/* Misc. IOCTL support (drm_ioctl.h) */
|
/* Misc. IOCTL support (drm_ioctl.h) */
|
||||||
|
|
Loading…
Reference in New Issue