drm: rip out DRIVER_FB_DMA and related code
No driver ever sets that flag, so good riddance! Signed-off-by: Daniel Vetter <daniel.vetter@ffwll.ch> Signed-off-by: Dave Airlie <airlied@redhat.com>
This commit is contained in:
parent
b0e898ac55
commit
687fbb2e4f
|
@ -1130,161 +1130,6 @@ static int drm_addbufs_sg(struct drm_device * dev, struct drm_buf_desc * request
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
static int drm_addbufs_fb(struct drm_device * dev, struct drm_buf_desc * request)
|
|
||||||
{
|
|
||||||
struct drm_device_dma *dma = dev->dma;
|
|
||||||
struct drm_buf_entry *entry;
|
|
||||||
struct drm_buf *buf;
|
|
||||||
unsigned long offset;
|
|
||||||
unsigned long agp_offset;
|
|
||||||
int count;
|
|
||||||
int order;
|
|
||||||
int size;
|
|
||||||
int alignment;
|
|
||||||
int page_order;
|
|
||||||
int total;
|
|
||||||
int byte_count;
|
|
||||||
int i;
|
|
||||||
struct drm_buf **temp_buflist;
|
|
||||||
|
|
||||||
if (!drm_core_check_feature(dev, DRIVER_FB_DMA))
|
|
||||||
return -EINVAL;
|
|
||||||
|
|
||||||
if (!dma)
|
|
||||||
return -EINVAL;
|
|
||||||
|
|
||||||
if (!capable(CAP_SYS_ADMIN))
|
|
||||||
return -EPERM;
|
|
||||||
|
|
||||||
count = request->count;
|
|
||||||
order = order_base_2(request->size);
|
|
||||||
size = 1 << order;
|
|
||||||
|
|
||||||
alignment = (request->flags & _DRM_PAGE_ALIGN)
|
|
||||||
? PAGE_ALIGN(size) : size;
|
|
||||||
page_order = order - PAGE_SHIFT > 0 ? order - PAGE_SHIFT : 0;
|
|
||||||
total = PAGE_SIZE << page_order;
|
|
||||||
|
|
||||||
byte_count = 0;
|
|
||||||
agp_offset = request->agp_start;
|
|
||||||
|
|
||||||
DRM_DEBUG("count: %d\n", count);
|
|
||||||
DRM_DEBUG("order: %d\n", order);
|
|
||||||
DRM_DEBUG("size: %d\n", size);
|
|
||||||
DRM_DEBUG("agp_offset: %lu\n", agp_offset);
|
|
||||||
DRM_DEBUG("alignment: %d\n", alignment);
|
|
||||||
DRM_DEBUG("page_order: %d\n", page_order);
|
|
||||||
DRM_DEBUG("total: %d\n", total);
|
|
||||||
|
|
||||||
if (order < DRM_MIN_ORDER || order > DRM_MAX_ORDER)
|
|
||||||
return -EINVAL;
|
|
||||||
|
|
||||||
spin_lock(&dev->count_lock);
|
|
||||||
if (dev->buf_use) {
|
|
||||||
spin_unlock(&dev->count_lock);
|
|
||||||
return -EBUSY;
|
|
||||||
}
|
|
||||||
atomic_inc(&dev->buf_alloc);
|
|
||||||
spin_unlock(&dev->count_lock);
|
|
||||||
|
|
||||||
mutex_lock(&dev->struct_mutex);
|
|
||||||
entry = &dma->bufs[order];
|
|
||||||
if (entry->buf_count) {
|
|
||||||
mutex_unlock(&dev->struct_mutex);
|
|
||||||
atomic_dec(&dev->buf_alloc);
|
|
||||||
return -ENOMEM; /* May only call once for each order */
|
|
||||||
}
|
|
||||||
|
|
||||||
if (count < 0 || count > 4096) {
|
|
||||||
mutex_unlock(&dev->struct_mutex);
|
|
||||||
atomic_dec(&dev->buf_alloc);
|
|
||||||
return -EINVAL;
|
|
||||||
}
|
|
||||||
|
|
||||||
entry->buflist = kzalloc(count * sizeof(*entry->buflist),
|
|
||||||
GFP_KERNEL);
|
|
||||||
if (!entry->buflist) {
|
|
||||||
mutex_unlock(&dev->struct_mutex);
|
|
||||||
atomic_dec(&dev->buf_alloc);
|
|
||||||
return -ENOMEM;
|
|
||||||
}
|
|
||||||
|
|
||||||
entry->buf_size = size;
|
|
||||||
entry->page_order = page_order;
|
|
||||||
|
|
||||||
offset = 0;
|
|
||||||
|
|
||||||
while (entry->buf_count < count) {
|
|
||||||
buf = &entry->buflist[entry->buf_count];
|
|
||||||
buf->idx = dma->buf_count + entry->buf_count;
|
|
||||||
buf->total = alignment;
|
|
||||||
buf->order = order;
|
|
||||||
buf->used = 0;
|
|
||||||
|
|
||||||
buf->offset = (dma->byte_count + offset);
|
|
||||||
buf->bus_address = agp_offset + offset;
|
|
||||||
buf->address = (void *)(agp_offset + offset);
|
|
||||||
buf->next = NULL;
|
|
||||||
buf->waiting = 0;
|
|
||||||
buf->pending = 0;
|
|
||||||
buf->file_priv = NULL;
|
|
||||||
|
|
||||||
buf->dev_priv_size = dev->driver->dev_priv_size;
|
|
||||||
buf->dev_private = kzalloc(buf->dev_priv_size, GFP_KERNEL);
|
|
||||||
if (!buf->dev_private) {
|
|
||||||
/* Set count correctly so we free the proper amount. */
|
|
||||||
entry->buf_count = count;
|
|
||||||
drm_cleanup_buf_error(dev, entry);
|
|
||||||
mutex_unlock(&dev->struct_mutex);
|
|
||||||
atomic_dec(&dev->buf_alloc);
|
|
||||||
return -ENOMEM;
|
|
||||||
}
|
|
||||||
|
|
||||||
DRM_DEBUG("buffer %d @ %p\n", entry->buf_count, buf->address);
|
|
||||||
|
|
||||||
offset += alignment;
|
|
||||||
entry->buf_count++;
|
|
||||||
byte_count += PAGE_SIZE << page_order;
|
|
||||||
}
|
|
||||||
|
|
||||||
DRM_DEBUG("byte_count: %d\n", byte_count);
|
|
||||||
|
|
||||||
temp_buflist = krealloc(dma->buflist,
|
|
||||||
(dma->buf_count + entry->buf_count) *
|
|
||||||
sizeof(*dma->buflist), GFP_KERNEL);
|
|
||||||
if (!temp_buflist) {
|
|
||||||
/* Free the entry because it isn't valid */
|
|
||||||
drm_cleanup_buf_error(dev, entry);
|
|
||||||
mutex_unlock(&dev->struct_mutex);
|
|
||||||
atomic_dec(&dev->buf_alloc);
|
|
||||||
return -ENOMEM;
|
|
||||||
}
|
|
||||||
dma->buflist = temp_buflist;
|
|
||||||
|
|
||||||
for (i = 0; i < entry->buf_count; i++) {
|
|
||||||
dma->buflist[i + dma->buf_count] = &entry->buflist[i];
|
|
||||||
}
|
|
||||||
|
|
||||||
dma->buf_count += entry->buf_count;
|
|
||||||
dma->seg_count += entry->seg_count;
|
|
||||||
dma->page_count += byte_count >> PAGE_SHIFT;
|
|
||||||
dma->byte_count += byte_count;
|
|
||||||
|
|
||||||
DRM_DEBUG("dma->buf_count : %d\n", dma->buf_count);
|
|
||||||
DRM_DEBUG("entry->buf_count : %d\n", entry->buf_count);
|
|
||||||
|
|
||||||
mutex_unlock(&dev->struct_mutex);
|
|
||||||
|
|
||||||
request->count = entry->buf_count;
|
|
||||||
request->size = size;
|
|
||||||
|
|
||||||
dma->flags = _DRM_DMA_USE_FB;
|
|
||||||
|
|
||||||
atomic_dec(&dev->buf_alloc);
|
|
||||||
return 0;
|
|
||||||
}
|
|
||||||
|
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* Add buffers for DMA transfers (ioctl).
|
* Add buffers for DMA transfers (ioctl).
|
||||||
*
|
*
|
||||||
|
@ -1319,7 +1164,7 @@ int drm_addbufs(struct drm_device *dev, void *data,
|
||||||
if (request->flags & _DRM_SG_BUFFER)
|
if (request->flags & _DRM_SG_BUFFER)
|
||||||
ret = drm_addbufs_sg(dev, request);
|
ret = drm_addbufs_sg(dev, request);
|
||||||
else if (request->flags & _DRM_FB_BUFFER)
|
else if (request->flags & _DRM_FB_BUFFER)
|
||||||
ret = drm_addbufs_fb(dev, request);
|
ret = -EINVAL;
|
||||||
else
|
else
|
||||||
ret = drm_addbufs_pci(dev, request);
|
ret = drm_addbufs_pci(dev, request);
|
||||||
|
|
||||||
|
@ -1556,9 +1401,7 @@ int drm_mapbufs(struct drm_device *dev, void *data,
|
||||||
if (request->count >= dma->buf_count) {
|
if (request->count >= dma->buf_count) {
|
||||||
if ((drm_core_has_AGP(dev) && (dma->flags & _DRM_DMA_USE_AGP))
|
if ((drm_core_has_AGP(dev) && (dma->flags & _DRM_DMA_USE_AGP))
|
||||||
|| (drm_core_check_feature(dev, DRIVER_SG)
|
|| (drm_core_check_feature(dev, DRIVER_SG)
|
||||||
&& (dma->flags & _DRM_DMA_USE_SG))
|
&& (dma->flags & _DRM_DMA_USE_SG))) {
|
||||||
|| (drm_core_check_feature(dev, DRIVER_FB_DMA)
|
|
||||||
&& (dma->flags & _DRM_DMA_USE_FB))) {
|
|
||||||
struct drm_local_map *map = dev->agp_buffer_map;
|
struct drm_local_map *map = dev->agp_buffer_map;
|
||||||
unsigned long token = dev->agp_buffer_token;
|
unsigned long token = dev->agp_buffer_token;
|
||||||
|
|
||||||
|
|
|
@ -147,7 +147,6 @@ int drm_err(const char *func, const char *format, ...);
|
||||||
#define DRIVER_IRQ_SHARED 0x80
|
#define DRIVER_IRQ_SHARED 0x80
|
||||||
#define DRIVER_IRQ_VBL 0x100
|
#define DRIVER_IRQ_VBL 0x100
|
||||||
#define DRIVER_DMA_QUEUE 0x200
|
#define DRIVER_DMA_QUEUE 0x200
|
||||||
#define DRIVER_FB_DMA 0x400
|
|
||||||
#define DRIVER_IRQ_VBL2 0x800
|
#define DRIVER_IRQ_VBL2 0x800
|
||||||
#define DRIVER_GEM 0x1000
|
#define DRIVER_GEM 0x1000
|
||||||
#define DRIVER_MODESET 0x2000
|
#define DRIVER_MODESET 0x2000
|
||||||
|
|
Loading…
Reference in New Issue