staging: bcm2835-camera: Replace open-coded idr with a struct idr.
We just need some integer handles that can map back to our message struct when we're handling a reply, which struct idr is perfect for. v2: Fix error check to look at the right variable. Signed-off-by: Eric Anholt <eric@anholt.net> Signed-off-by: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
This commit is contained in:
parent
6c66a7b097
commit
950fd867c6
|
@ -21,7 +21,6 @@
|
||||||
#include <linux/slab.h>
|
#include <linux/slab.h>
|
||||||
#include <linux/completion.h>
|
#include <linux/completion.h>
|
||||||
#include <linux/vmalloc.h>
|
#include <linux/vmalloc.h>
|
||||||
#include <linux/btree.h>
|
|
||||||
#include <asm/cacheflush.h>
|
#include <asm/cacheflush.h>
|
||||||
#include <media/videobuf2-vmalloc.h>
|
#include <media/videobuf2-vmalloc.h>
|
||||||
|
|
||||||
|
@ -111,7 +110,11 @@ struct vchiq_mmal_instance;
|
||||||
/* normal message context */
|
/* normal message context */
|
||||||
struct mmal_msg_context {
|
struct mmal_msg_context {
|
||||||
struct vchiq_mmal_instance *instance;
|
struct vchiq_mmal_instance *instance;
|
||||||
u32 handle;
|
|
||||||
|
/* Index in the context_map idr so that we can find the
|
||||||
|
* mmal_msg_context again when servicing the VCHI reply.
|
||||||
|
*/
|
||||||
|
int handle;
|
||||||
|
|
||||||
union {
|
union {
|
||||||
struct {
|
struct {
|
||||||
|
@ -149,13 +152,6 @@ struct mmal_msg_context {
|
||||||
|
|
||||||
};
|
};
|
||||||
|
|
||||||
struct vchiq_mmal_context_map {
|
|
||||||
/* ensure serialized access to the btree(contention should be low) */
|
|
||||||
struct mutex lock;
|
|
||||||
struct btree_head32 btree_head;
|
|
||||||
u32 last_handle;
|
|
||||||
};
|
|
||||||
|
|
||||||
struct vchiq_mmal_instance {
|
struct vchiq_mmal_instance {
|
||||||
VCHI_SERVICE_HANDLE_T handle;
|
VCHI_SERVICE_HANDLE_T handle;
|
||||||
|
|
||||||
|
@ -165,92 +161,19 @@ struct vchiq_mmal_instance {
|
||||||
/* vmalloc page to receive scratch bulk xfers into */
|
/* vmalloc page to receive scratch bulk xfers into */
|
||||||
void *bulk_scratch;
|
void *bulk_scratch;
|
||||||
|
|
||||||
/* mapping table between context handles and mmal_msg_contexts */
|
struct idr context_map;
|
||||||
struct vchiq_mmal_context_map context_map;
|
spinlock_t context_map_lock;
|
||||||
|
|
||||||
/* component to use next */
|
/* component to use next */
|
||||||
int component_idx;
|
int component_idx;
|
||||||
struct vchiq_mmal_component component[VCHIQ_MMAL_MAX_COMPONENTS];
|
struct vchiq_mmal_component component[VCHIQ_MMAL_MAX_COMPONENTS];
|
||||||
};
|
};
|
||||||
|
|
||||||
static int __must_check
|
|
||||||
mmal_context_map_init(struct vchiq_mmal_context_map *context_map)
|
|
||||||
{
|
|
||||||
mutex_init(&context_map->lock);
|
|
||||||
context_map->last_handle = 0;
|
|
||||||
return btree_init32(&context_map->btree_head);
|
|
||||||
}
|
|
||||||
|
|
||||||
static void mmal_context_map_destroy(struct vchiq_mmal_context_map *context_map)
|
|
||||||
{
|
|
||||||
mutex_lock(&context_map->lock);
|
|
||||||
btree_destroy32(&context_map->btree_head);
|
|
||||||
mutex_unlock(&context_map->lock);
|
|
||||||
}
|
|
||||||
|
|
||||||
static u32
|
|
||||||
mmal_context_map_create_handle(struct vchiq_mmal_context_map *context_map,
|
|
||||||
struct mmal_msg_context *msg_context,
|
|
||||||
gfp_t gfp)
|
|
||||||
{
|
|
||||||
u32 handle;
|
|
||||||
|
|
||||||
mutex_lock(&context_map->lock);
|
|
||||||
|
|
||||||
while (1) {
|
|
||||||
/* just use a simple count for handles, but do not use 0 */
|
|
||||||
context_map->last_handle++;
|
|
||||||
if (!context_map->last_handle)
|
|
||||||
context_map->last_handle++;
|
|
||||||
|
|
||||||
handle = context_map->last_handle;
|
|
||||||
|
|
||||||
/* check if the handle is already in use */
|
|
||||||
if (!btree_lookup32(&context_map->btree_head, handle))
|
|
||||||
break;
|
|
||||||
}
|
|
||||||
|
|
||||||
if (btree_insert32(&context_map->btree_head, handle,
|
|
||||||
msg_context, gfp)) {
|
|
||||||
/* probably out of memory */
|
|
||||||
mutex_unlock(&context_map->lock);
|
|
||||||
return 0;
|
|
||||||
}
|
|
||||||
|
|
||||||
mutex_unlock(&context_map->lock);
|
|
||||||
return handle;
|
|
||||||
}
|
|
||||||
|
|
||||||
static struct mmal_msg_context *
|
|
||||||
mmal_context_map_lookup_handle(struct vchiq_mmal_context_map *context_map,
|
|
||||||
u32 handle)
|
|
||||||
{
|
|
||||||
struct mmal_msg_context *msg_context;
|
|
||||||
|
|
||||||
if (!handle)
|
|
||||||
return NULL;
|
|
||||||
|
|
||||||
mutex_lock(&context_map->lock);
|
|
||||||
|
|
||||||
msg_context = btree_lookup32(&context_map->btree_head, handle);
|
|
||||||
|
|
||||||
mutex_unlock(&context_map->lock);
|
|
||||||
return msg_context;
|
|
||||||
}
|
|
||||||
|
|
||||||
static void
|
|
||||||
mmal_context_map_destroy_handle(struct vchiq_mmal_context_map *context_map,
|
|
||||||
u32 handle)
|
|
||||||
{
|
|
||||||
mutex_lock(&context_map->lock);
|
|
||||||
btree_remove32(&context_map->btree_head, handle);
|
|
||||||
mutex_unlock(&context_map->lock);
|
|
||||||
}
|
|
||||||
|
|
||||||
static struct mmal_msg_context *
|
static struct mmal_msg_context *
|
||||||
get_msg_context(struct vchiq_mmal_instance *instance)
|
get_msg_context(struct vchiq_mmal_instance *instance)
|
||||||
{
|
{
|
||||||
struct mmal_msg_context *msg_context;
|
struct mmal_msg_context *msg_context;
|
||||||
|
int handle;
|
||||||
|
|
||||||
/* todo: should this be allocated from a pool to avoid kzalloc */
|
/* todo: should this be allocated from a pool to avoid kzalloc */
|
||||||
msg_context = kzalloc(sizeof(*msg_context), GFP_KERNEL);
|
msg_context = kzalloc(sizeof(*msg_context), GFP_KERNEL);
|
||||||
|
@ -258,32 +181,40 @@ get_msg_context(struct vchiq_mmal_instance *instance)
|
||||||
if (!msg_context)
|
if (!msg_context)
|
||||||
return ERR_PTR(-ENOMEM);
|
return ERR_PTR(-ENOMEM);
|
||||||
|
|
||||||
msg_context->instance = instance;
|
/* Create an ID that will be passed along with our message so
|
||||||
msg_context->handle =
|
* that when we service the VCHI reply, we can look up what
|
||||||
mmal_context_map_create_handle(&instance->context_map,
|
* message is being replied to.
|
||||||
msg_context,
|
*/
|
||||||
GFP_KERNEL);
|
spin_lock(&instance->context_map_lock);
|
||||||
|
handle = idr_alloc(&instance->context_map, msg_context,
|
||||||
|
0, 0, GFP_KERNEL);
|
||||||
|
spin_unlock(&instance->context_map_lock);
|
||||||
|
|
||||||
if (!msg_context->handle) {
|
if (handle < 0) {
|
||||||
kfree(msg_context);
|
kfree(msg_context);
|
||||||
return ERR_PTR(-ENOMEM);
|
return ERR_PTR(handle);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
msg_context->instance = instance;
|
||||||
|
msg_context->handle = handle;
|
||||||
|
|
||||||
return msg_context;
|
return msg_context;
|
||||||
}
|
}
|
||||||
|
|
||||||
static struct mmal_msg_context *
|
static struct mmal_msg_context *
|
||||||
lookup_msg_context(struct vchiq_mmal_instance *instance, u32 handle)
|
lookup_msg_context(struct vchiq_mmal_instance *instance, int handle)
|
||||||
{
|
{
|
||||||
return mmal_context_map_lookup_handle(&instance->context_map,
|
return idr_find(&instance->context_map, handle);
|
||||||
handle);
|
|
||||||
}
|
}
|
||||||
|
|
||||||
static void
|
static void
|
||||||
release_msg_context(struct mmal_msg_context *msg_context)
|
release_msg_context(struct mmal_msg_context *msg_context)
|
||||||
{
|
{
|
||||||
mmal_context_map_destroy_handle(&msg_context->instance->context_map,
|
struct vchiq_mmal_instance *instance = msg_context->instance;
|
||||||
msg_context->handle);
|
|
||||||
|
spin_lock(&instance->context_map_lock);
|
||||||
|
idr_remove(&instance->context_map, msg_context->handle);
|
||||||
|
spin_unlock(&instance->context_map_lock);
|
||||||
kfree(msg_context);
|
kfree(msg_context);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
@ -1860,7 +1791,7 @@ int vchiq_mmal_finalise(struct vchiq_mmal_instance *instance)
|
||||||
|
|
||||||
vfree(instance->bulk_scratch);
|
vfree(instance->bulk_scratch);
|
||||||
|
|
||||||
mmal_context_map_destroy(&instance->context_map);
|
idr_destroy(&instance->context_map);
|
||||||
|
|
||||||
kfree(instance);
|
kfree(instance);
|
||||||
|
|
||||||
|
@ -1922,12 +1853,8 @@ int vchiq_mmal_init(struct vchiq_mmal_instance **out_instance)
|
||||||
|
|
||||||
instance->bulk_scratch = vmalloc(PAGE_SIZE);
|
instance->bulk_scratch = vmalloc(PAGE_SIZE);
|
||||||
|
|
||||||
status = mmal_context_map_init(&instance->context_map);
|
spin_lock_init(&instance->context_map_lock);
|
||||||
if (status) {
|
idr_init_base(&instance->context_map, 1);
|
||||||
pr_err("Failed to init context map (status=%d)\n", status);
|
|
||||||
kfree(instance);
|
|
||||||
return status;
|
|
||||||
}
|
|
||||||
|
|
||||||
params.callback_param = instance;
|
params.callback_param = instance;
|
||||||
|
|
||||||
|
|
Loading…
Reference in New Issue