GFS2: Use RCU/hlist_bl based hash for quotas

Prior to this patch, GFS2 kept all the quotas for each
super block in a single linked list. This is rather slow
when there are large numbers of quotas.

This patch introduces a hlist_bl based hash table, similar
to the one used for glocks. The initial look up of the quota
is now lockless in the case where it is already cached,
although we still have to take the per quota spinlock in
order to bump the ref count. Either way though, this is a
big improvement on what was there before.

The qd_lock and the per super block list is preserved, for
the time being. However it is intended that since this is no
longer used for its original role, it should be possible to
shrink the number of items on that list in due course and
remove the requirement to take qd_lock in qd_get.

Signed-off-by: Steven Whitehouse <swhiteho@redhat.com>
Cc: Abhijith Das <adas@redhat.com>
Cc: Paul E. McKenney <paulmck@linux.vnet.ibm.com>
This commit is contained in:
Steven Whitehouse 2013-12-12 10:47:59 +00:00
parent 086352f1aa
commit c754fbbb1b
4 changed files with 127 additions and 49 deletions

View File

@ -428,10 +428,13 @@ enum {
}; };
struct gfs2_quota_data { struct gfs2_quota_data {
struct hlist_bl_node qd_hlist;
struct list_head qd_list; struct list_head qd_list;
struct kqid qd_id; struct kqid qd_id;
struct gfs2_sbd *qd_sbd;
struct lockref qd_lockref; struct lockref qd_lockref;
struct list_head qd_lru; struct list_head qd_lru;
unsigned qd_hash;
unsigned long qd_flags; /* QDF_... */ unsigned long qd_flags; /* QDF_... */
@ -450,6 +453,7 @@ struct gfs2_quota_data {
u64 qd_sync_gen; u64 qd_sync_gen;
unsigned long qd_last_warn; unsigned long qd_last_warn;
struct rcu_head qd_rcu;
}; };
struct gfs2_trans { struct gfs2_trans {

View File

@ -76,6 +76,7 @@ static int __init init_gfs2_fs(void)
gfs2_str2qstr(&gfs2_qdot, "."); gfs2_str2qstr(&gfs2_qdot, ".");
gfs2_str2qstr(&gfs2_qdotdot, ".."); gfs2_str2qstr(&gfs2_qdotdot, "..");
gfs2_quota_hash_init();
error = gfs2_sys_init(); error = gfs2_sys_init();
if (error) if (error)

View File

@ -52,6 +52,10 @@
#include <linux/dqblk_xfs.h> #include <linux/dqblk_xfs.h>
#include <linux/lockref.h> #include <linux/lockref.h>
#include <linux/list_lru.h> #include <linux/list_lru.h>
#include <linux/rcupdate.h>
#include <linux/rculist_bl.h>
#include <linux/bit_spinlock.h>
#include <linux/jhash.h>
#include "gfs2.h" #include "gfs2.h"
#include "incore.h" #include "incore.h"
@ -67,10 +71,43 @@
#include "inode.h" #include "inode.h"
#include "util.h" #include "util.h"
/* Lock order: qd_lock -> qd->lockref.lock -> lru lock */ #define GFS2_QD_HASH_SHIFT 12
#define GFS2_QD_HASH_SIZE (1 << GFS2_QD_HASH_SHIFT)
#define GFS2_QD_HASH_MASK (GFS2_QD_HASH_SIZE - 1)
/* Lock order: qd_lock -> bucket lock -> qd->lockref.lock -> lru lock */
static DEFINE_SPINLOCK(qd_lock); static DEFINE_SPINLOCK(qd_lock);
struct list_lru gfs2_qd_lru; struct list_lru gfs2_qd_lru;
static struct hlist_bl_head qd_hash_table[GFS2_QD_HASH_SIZE];
static unsigned int gfs2_qd_hash(const struct gfs2_sbd *sdp,
const struct kqid qid)
{
unsigned int h;
h = jhash(&sdp, sizeof(struct gfs2_sbd *), 0);
h = jhash(&qid, sizeof(struct kqid), h);
return h & GFS2_QD_HASH_MASK;
}
static inline void spin_lock_bucket(unsigned int hash)
{
hlist_bl_lock(&qd_hash_table[hash]);
}
static inline void spin_unlock_bucket(unsigned int hash)
{
hlist_bl_unlock(&qd_hash_table[hash]);
}
static void gfs2_qd_dealloc(struct rcu_head *rcu)
{
struct gfs2_quota_data *qd = container_of(rcu, struct gfs2_quota_data, qd_rcu);
kmem_cache_free(gfs2_quotad_cachep, qd);
}
static void gfs2_qd_dispose(struct list_head *list) static void gfs2_qd_dispose(struct list_head *list)
{ {
struct gfs2_quota_data *qd; struct gfs2_quota_data *qd;
@ -87,6 +124,10 @@ static void gfs2_qd_dispose(struct list_head *list)
list_del(&qd->qd_list); list_del(&qd->qd_list);
spin_unlock(&qd_lock); spin_unlock(&qd_lock);
spin_lock_bucket(qd->qd_hash);
hlist_bl_del_rcu(&qd->qd_hlist);
spin_unlock_bucket(qd->qd_hash);
gfs2_assert_warn(sdp, !qd->qd_change); gfs2_assert_warn(sdp, !qd->qd_change);
gfs2_assert_warn(sdp, !qd->qd_slot_count); gfs2_assert_warn(sdp, !qd->qd_slot_count);
gfs2_assert_warn(sdp, !qd->qd_bh_count); gfs2_assert_warn(sdp, !qd->qd_bh_count);
@ -95,7 +136,7 @@ static void gfs2_qd_dispose(struct list_head *list)
atomic_dec(&sdp->sd_quota_count); atomic_dec(&sdp->sd_quota_count);
/* Delete it from the common reclaim list */ /* Delete it from the common reclaim list */
kmem_cache_free(gfs2_quotad_cachep, qd); call_rcu(&qd->qd_rcu, gfs2_qd_dealloc);
} }
} }
@ -165,83 +206,95 @@ static u64 qd2offset(struct gfs2_quota_data *qd)
return offset; return offset;
} }
static int qd_alloc(struct gfs2_sbd *sdp, struct kqid qid, static struct gfs2_quota_data *qd_alloc(unsigned hash, struct gfs2_sbd *sdp, struct kqid qid)
struct gfs2_quota_data **qdp)
{ {
struct gfs2_quota_data *qd; struct gfs2_quota_data *qd;
int error; int error;
qd = kmem_cache_zalloc(gfs2_quotad_cachep, GFP_NOFS); qd = kmem_cache_zalloc(gfs2_quotad_cachep, GFP_NOFS);
if (!qd) if (!qd)
return -ENOMEM; return NULL;
qd->qd_sbd = sdp;
qd->qd_lockref.count = 1; qd->qd_lockref.count = 1;
spin_lock_init(&qd->qd_lockref.lock); spin_lock_init(&qd->qd_lockref.lock);
qd->qd_id = qid; qd->qd_id = qid;
qd->qd_slot = -1; qd->qd_slot = -1;
INIT_LIST_HEAD(&qd->qd_lru); INIT_LIST_HEAD(&qd->qd_lru);
qd->qd_hash = hash;
error = gfs2_glock_get(sdp, qd2index(qd), error = gfs2_glock_get(sdp, qd2index(qd),
&gfs2_quota_glops, CREATE, &qd->qd_gl); &gfs2_quota_glops, CREATE, &qd->qd_gl);
if (error) if (error)
goto fail; goto fail;
*qdp = qd; return qd;
return 0;
fail: fail:
kmem_cache_free(gfs2_quotad_cachep, qd); kmem_cache_free(gfs2_quotad_cachep, qd);
return error; return NULL;
} }
static struct gfs2_quota_data *gfs2_qd_search_bucket(unsigned int hash,
const struct gfs2_sbd *sdp,
struct kqid qid)
{
struct gfs2_quota_data *qd;
struct hlist_bl_node *h;
hlist_bl_for_each_entry_rcu(qd, h, &qd_hash_table[hash], qd_hlist) {
if (!qid_eq(qd->qd_id, qid))
continue;
if (qd->qd_sbd != sdp)
continue;
if (lockref_get_not_dead(&qd->qd_lockref)) {
list_lru_del(&gfs2_qd_lru, &qd->qd_lru);
return qd;
}
}
return NULL;
}
static int qd_get(struct gfs2_sbd *sdp, struct kqid qid, static int qd_get(struct gfs2_sbd *sdp, struct kqid qid,
struct gfs2_quota_data **qdp) struct gfs2_quota_data **qdp)
{ {
struct gfs2_quota_data *qd = NULL, *new_qd = NULL; struct gfs2_quota_data *qd, *new_qd;
int error, found; unsigned int hash = gfs2_qd_hash(sdp, qid);
*qdp = NULL; rcu_read_lock();
*qdp = qd = gfs2_qd_search_bucket(hash, sdp, qid);
rcu_read_unlock();
for (;;) { if (qd)
found = 0; return 0;
spin_lock(&qd_lock);
list_for_each_entry(qd, &sdp->sd_quota_list, qd_list) {
if (qid_eq(qd->qd_id, qid) &&
lockref_get_not_dead(&qd->qd_lockref)) {
list_lru_del(&gfs2_qd_lru, &qd->qd_lru);
found = 1;
break;
}
}
if (!found) new_qd = qd_alloc(hash, sdp, qid);
qd = NULL; if (!new_qd)
return -ENOMEM;
if (!qd && new_qd) { spin_lock(&qd_lock);
qd = new_qd; spin_lock_bucket(hash);
list_add(&qd->qd_list, &sdp->sd_quota_list); *qdp = qd = gfs2_qd_search_bucket(hash, sdp, qid);
atomic_inc(&sdp->sd_quota_count); if (qd == NULL) {
new_qd = NULL; *qdp = new_qd;
} list_add(&new_qd->qd_list, &sdp->sd_quota_list);
hlist_bl_add_head_rcu(&new_qd->qd_hlist, &qd_hash_table[hash]);
spin_unlock(&qd_lock); atomic_inc(&sdp->sd_quota_count);
if (qd) {
if (new_qd) {
gfs2_glock_put(new_qd->qd_gl);
kmem_cache_free(gfs2_quotad_cachep, new_qd);
}
*qdp = qd;
return 0;
}
error = qd_alloc(sdp, qid, &new_qd);
if (error)
return error;
} }
spin_unlock_bucket(hash);
spin_unlock(&qd_lock);
if (qd) {
gfs2_glock_put(new_qd->qd_gl);
kmem_cache_free(gfs2_quotad_cachep, new_qd);
}
return 0;
} }
static void qd_hold(struct gfs2_quota_data *qd) static void qd_hold(struct gfs2_quota_data *qd)
{ {
struct gfs2_sbd *sdp = qd->qd_gl->gl_sbd; struct gfs2_sbd *sdp = qd->qd_gl->gl_sbd;
@ -1215,6 +1268,7 @@ int gfs2_quota_init(struct gfs2_sbd *sdp)
unsigned int blocks = size >> sdp->sd_sb.sb_bsize_shift; unsigned int blocks = size >> sdp->sd_sb.sb_bsize_shift;
unsigned int x, slot = 0; unsigned int x, slot = 0;
unsigned int found = 0; unsigned int found = 0;
unsigned int hash;
u64 dblock; u64 dblock;
u32 extlen = 0; u32 extlen = 0;
int error; int error;
@ -1272,8 +1326,9 @@ int gfs2_quota_init(struct gfs2_sbd *sdp)
if (!qc_change) if (!qc_change)
continue; continue;
error = qd_alloc(sdp, qc_id, &qd); hash = gfs2_qd_hash(sdp, qc_id);
if (error) { qd = qd_alloc(hash, sdp, qc_id);
if (qd == NULL) {
brelse(bh); brelse(bh);
goto fail; goto fail;
} }
@ -1289,6 +1344,10 @@ int gfs2_quota_init(struct gfs2_sbd *sdp)
atomic_inc(&sdp->sd_quota_count); atomic_inc(&sdp->sd_quota_count);
spin_unlock(&qd_lock); spin_unlock(&qd_lock);
spin_lock_bucket(hash);
hlist_bl_add_head_rcu(&qd->qd_hlist, &qd_hash_table[hash]);
spin_unlock_bucket(hash);
found++; found++;
} }
@ -1335,11 +1394,16 @@ void gfs2_quota_cleanup(struct gfs2_sbd *sdp)
spin_unlock(&qd->qd_lockref.lock); spin_unlock(&qd->qd_lockref.lock);
list_del(&qd->qd_list); list_del(&qd->qd_list);
/* Also remove if this qd exists in the reclaim list */ /* Also remove if this qd exists in the reclaim list */
list_lru_del(&gfs2_qd_lru, &qd->qd_lru); list_lru_del(&gfs2_qd_lru, &qd->qd_lru);
atomic_dec(&sdp->sd_quota_count); atomic_dec(&sdp->sd_quota_count);
spin_unlock(&qd_lock); spin_unlock(&qd_lock);
spin_lock_bucket(qd->qd_hash);
hlist_bl_del_rcu(&qd->qd_hlist);
spin_unlock_bucket(qd->qd_hash);
if (!qd->qd_lockref.count) { if (!qd->qd_lockref.count) {
gfs2_assert_warn(sdp, !qd->qd_change); gfs2_assert_warn(sdp, !qd->qd_change);
gfs2_assert_warn(sdp, !qd->qd_slot_count); gfs2_assert_warn(sdp, !qd->qd_slot_count);
@ -1348,7 +1412,7 @@ void gfs2_quota_cleanup(struct gfs2_sbd *sdp)
gfs2_assert_warn(sdp, !qd->qd_bh_count); gfs2_assert_warn(sdp, !qd->qd_bh_count);
gfs2_glock_put(qd->qd_gl); gfs2_glock_put(qd->qd_gl);
kmem_cache_free(gfs2_quotad_cachep, qd); call_rcu(&qd->qd_rcu, gfs2_qd_dealloc);
spin_lock(&qd_lock); spin_lock(&qd_lock);
} }
@ -1643,3 +1707,11 @@ const struct quotactl_ops gfs2_quotactl_ops = {
.get_dqblk = gfs2_get_dqblk, .get_dqblk = gfs2_get_dqblk,
.set_dqblk = gfs2_set_dqblk, .set_dqblk = gfs2_set_dqblk,
}; };
void __init gfs2_quota_hash_init(void)
{
unsigned i;
for(i = 0; i < GFS2_QD_HASH_SIZE; i++)
INIT_HLIST_BL_HEAD(&qd_hash_table[i]);
}

View File

@ -57,5 +57,6 @@ static inline int gfs2_quota_lock_check(struct gfs2_inode *ip)
extern const struct quotactl_ops gfs2_quotactl_ops; extern const struct quotactl_ops gfs2_quotactl_ops;
extern struct shrinker gfs2_qd_shrinker; extern struct shrinker gfs2_qd_shrinker;
extern struct list_lru gfs2_qd_lru; extern struct list_lru gfs2_qd_lru;
extern void __init gfs2_quota_hash_init(void);
#endif /* __QUOTA_DOT_H__ */ #endif /* __QUOTA_DOT_H__ */