2005-04-17 06:20:36 +08:00
|
|
|
/*
|
|
|
|
* Copyright (c) 2001 The Regents of the University of Michigan.
|
|
|
|
* All rights reserved.
|
|
|
|
*
|
|
|
|
* Kendrick Smith <kmsmith@umich.edu>
|
|
|
|
* Andy Adamson <kandros@umich.edu>
|
|
|
|
*
|
|
|
|
* Redistribution and use in source and binary forms, with or without
|
|
|
|
* modification, are permitted provided that the following conditions
|
|
|
|
* are met:
|
|
|
|
*
|
|
|
|
* 1. Redistributions of source code must retain the above copyright
|
|
|
|
* notice, this list of conditions and the following disclaimer.
|
|
|
|
* 2. Redistributions in binary form must reproduce the above copyright
|
|
|
|
* notice, this list of conditions and the following disclaimer in the
|
|
|
|
* documentation and/or other materials provided with the distribution.
|
|
|
|
* 3. Neither the name of the University nor the names of its
|
|
|
|
* contributors may be used to endorse or promote products derived
|
|
|
|
* from this software without specific prior written permission.
|
|
|
|
*
|
|
|
|
* THIS SOFTWARE IS PROVIDED ``AS IS'' AND ANY EXPRESS OR IMPLIED
|
|
|
|
* WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF
|
|
|
|
* MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE
|
|
|
|
* DISCLAIMED. IN NO EVENT SHALL THE REGENTS OR CONTRIBUTORS BE LIABLE
|
|
|
|
* FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR
|
|
|
|
* CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF
|
|
|
|
* SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR
|
|
|
|
* BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF
|
|
|
|
* LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING
|
|
|
|
* NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS
|
|
|
|
* SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE.
|
|
|
|
*
|
|
|
|
*/
|
|
|
|
|
2008-02-16 06:37:31 +08:00
|
|
|
#include <linux/file.h>
|
2010-09-18 21:09:31 +08:00
|
|
|
#include <linux/fs.h>
|
include cleanup: Update gfp.h and slab.h includes to prepare for breaking implicit slab.h inclusion from percpu.h
percpu.h is included by sched.h and module.h and thus ends up being
included when building most .c files. percpu.h includes slab.h which
in turn includes gfp.h making everything defined by the two files
universally available and complicating inclusion dependencies.
percpu.h -> slab.h dependency is about to be removed. Prepare for
this change by updating users of gfp and slab facilities include those
headers directly instead of assuming availability. As this conversion
needs to touch large number of source files, the following script is
used as the basis of conversion.
http://userweb.kernel.org/~tj/misc/slabh-sweep.py
The script does the followings.
* Scan files for gfp and slab usages and update includes such that
only the necessary includes are there. ie. if only gfp is used,
gfp.h, if slab is used, slab.h.
* When the script inserts a new include, it looks at the include
blocks and try to put the new include such that its order conforms
to its surrounding. It's put in the include block which contains
core kernel includes, in the same order that the rest are ordered -
alphabetical, Christmas tree, rev-Xmas-tree or at the end if there
doesn't seem to be any matching order.
* If the script can't find a place to put a new include (mostly
because the file doesn't have fitting include block), it prints out
an error message indicating which .h file needs to be added to the
file.
The conversion was done in the following steps.
1. The initial automatic conversion of all .c files updated slightly
over 4000 files, deleting around 700 includes and adding ~480 gfp.h
and ~3000 slab.h inclusions. The script emitted errors for ~400
files.
2. Each error was manually checked. Some didn't need the inclusion,
some needed manual addition while adding it to implementation .h or
embedding .c file was more appropriate for others. This step added
inclusions to around 150 files.
3. The script was run again and the output was compared to the edits
from #2 to make sure no file was left behind.
4. Several build tests were done and a couple of problems were fixed.
e.g. lib/decompress_*.c used malloc/free() wrappers around slab
APIs requiring slab.h to be added manually.
5. The script was run on all .h files but without automatically
editing them as sprinkling gfp.h and slab.h inclusions around .h
files could easily lead to inclusion dependency hell. Most gfp.h
inclusion directives were ignored as stuff from gfp.h was usually
wildly available and often used in preprocessor macros. Each
slab.h inclusion directive was examined and added manually as
necessary.
6. percpu.h was updated not to include slab.h.
7. Build test were done on the following configurations and failures
were fixed. CONFIG_GCOV_KERNEL was turned off for all tests (as my
distributed build env didn't work with gcov compiles) and a few
more options had to be turned off depending on archs to make things
build (like ipr on powerpc/64 which failed due to missing writeq).
* x86 and x86_64 UP and SMP allmodconfig and a custom test config.
* powerpc and powerpc64 SMP allmodconfig
* sparc and sparc64 SMP allmodconfig
* ia64 SMP allmodconfig
* s390 SMP allmodconfig
* alpha SMP allmodconfig
* um on x86_64 SMP allmodconfig
8. percpu.h modifications were reverted so that it could be applied as
a separate patch and serve as bisection point.
Given the fact that I had only a couple of failures from tests on step
6, I'm fairly confident about the coverage of this conversion patch.
If there is a breakage, it's likely to be something in one of the arch
headers which should be easily discoverable easily on most builds of
the specific arch.
Signed-off-by: Tejun Heo <tj@kernel.org>
Guess-its-ok-by: Christoph Lameter <cl@linux-foundation.org>
Cc: Ingo Molnar <mingo@redhat.com>
Cc: Lee Schermerhorn <Lee.Schermerhorn@hp.com>
2010-03-24 16:04:11 +08:00
|
|
|
#include <linux/slab.h>
|
2005-06-24 13:04:32 +08:00
|
|
|
#include <linux/namei.h>
|
2007-07-17 19:04:39 +08:00
|
|
|
#include <linux/swap.h>
|
2011-07-13 22:50:48 +08:00
|
|
|
#include <linux/pagemap.h>
|
2008-12-24 05:17:15 +08:00
|
|
|
#include <linux/sunrpc/svcauth_gss.h>
|
2009-08-15 00:57:56 +08:00
|
|
|
#include <linux/sunrpc/clnt.h>
|
2009-12-04 02:30:56 +08:00
|
|
|
#include "xdr4.h"
|
2009-11-05 07:12:35 +08:00
|
|
|
#include "vfs.h"
|
2005-04-17 06:20:36 +08:00
|
|
|
|
|
|
|
#define NFSDDBG_FACILITY NFSDDBG_PROC
|
|
|
|
|
|
|
|
/* Globals */
|
2010-03-01 12:20:19 +08:00
|
|
|
time_t nfsd4_lease = 90; /* default lease time */
|
2010-03-03 00:04:06 +08:00
|
|
|
time_t nfsd4_grace = 90;
|
2005-06-24 13:04:03 +08:00
|
|
|
static time_t boot_time;
|
2011-12-13 04:00:35 +08:00
|
|
|
|
|
|
|
#define all_ones {{~0,~0},~0}
|
|
|
|
static const stateid_t one_stateid = {
|
|
|
|
.si_generation = ~0,
|
|
|
|
.si_opaque = all_ones,
|
|
|
|
};
|
|
|
|
static const stateid_t zero_stateid = {
|
|
|
|
/* all fields zero */
|
|
|
|
};
|
2012-02-14 05:55:23 +08:00
|
|
|
static const stateid_t currentstateid = {
|
|
|
|
.si_generation = 1,
|
|
|
|
};
|
2011-12-13 04:00:35 +08:00
|
|
|
|
2009-04-03 13:28:28 +08:00
|
|
|
static u64 current_sessionid = 1;
|
2005-06-24 13:04:03 +08:00
|
|
|
|
2011-12-13 04:00:35 +08:00
|
|
|
#define ZERO_STATEID(stateid) (!memcmp((stateid), &zero_stateid, sizeof(stateid_t)))
|
|
|
|
#define ONE_STATEID(stateid) (!memcmp((stateid), &one_stateid, sizeof(stateid_t)))
|
2012-02-14 05:55:23 +08:00
|
|
|
#define CURRENT_STATEID(stateid) (!memcmp((stateid), ¤tstateid, sizeof(stateid_t)))
|
2005-04-17 06:20:36 +08:00
|
|
|
|
|
|
|
/* forward declarations */
|
2011-07-31 11:33:59 +08:00
|
|
|
static int check_for_locks(struct nfs4_file *filp, struct nfs4_lockowner *lowner);
|
2005-04-17 06:20:36 +08:00
|
|
|
|
2009-02-23 06:51:34 +08:00
|
|
|
/* Locking: */
|
|
|
|
|
|
|
|
/* Currently used for almost all code touching nfsv4 state: */
|
2006-03-26 17:37:12 +08:00
|
|
|
static DEFINE_MUTEX(client_mutex);
|
2005-04-17 06:20:36 +08:00
|
|
|
|
2009-02-23 06:51:34 +08:00
|
|
|
/*
|
|
|
|
* Currently used for the del_recall_lru and file hash table. In an
|
|
|
|
* effort to decrease the scope of the client_mutex, this spinlock may
|
|
|
|
* eventually cover more:
|
|
|
|
*/
|
|
|
|
static DEFINE_SPINLOCK(recall_lock);
|
|
|
|
|
2011-07-31 11:33:59 +08:00
|
|
|
static struct kmem_cache *openowner_slab = NULL;
|
|
|
|
static struct kmem_cache *lockowner_slab = NULL;
|
2006-12-07 12:33:20 +08:00
|
|
|
static struct kmem_cache *file_slab = NULL;
|
|
|
|
static struct kmem_cache *stateid_slab = NULL;
|
|
|
|
static struct kmem_cache *deleg_slab = NULL;
|
2005-06-24 13:03:01 +08:00
|
|
|
|
2005-04-17 06:20:36 +08:00
|
|
|
void
|
|
|
|
nfs4_lock_state(void)
|
|
|
|
{
|
2006-03-26 17:37:12 +08:00
|
|
|
mutex_lock(&client_mutex);
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
|
|
|
|
2012-02-24 09:40:52 +08:00
|
|
|
static void free_session(struct kref *);
|
|
|
|
|
|
|
|
/* Must be called under the client_lock */
|
|
|
|
static void nfsd4_put_session_locked(struct nfsd4_session *ses)
|
|
|
|
{
|
|
|
|
kref_put(&ses->se_ref, free_session);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void nfsd4_get_session(struct nfsd4_session *ses)
|
|
|
|
{
|
|
|
|
kref_get(&ses->se_ref);
|
|
|
|
}
|
|
|
|
|
2005-04-17 06:20:36 +08:00
|
|
|
void
|
|
|
|
nfs4_unlock_state(void)
|
|
|
|
{
|
2006-03-26 17:37:12 +08:00
|
|
|
mutex_unlock(&client_mutex);
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
static inline u32
|
|
|
|
opaque_hashval(const void *ptr, int nbytes)
|
|
|
|
{
|
|
|
|
unsigned char *cptr = (unsigned char *) ptr;
|
|
|
|
|
|
|
|
u32 x = 0;
|
|
|
|
while (nbytes--) {
|
|
|
|
x *= 37;
|
|
|
|
x += *cptr++;
|
|
|
|
}
|
|
|
|
return x;
|
|
|
|
}
|
|
|
|
|
|
|
|
static struct list_head del_recall_lru;
|
|
|
|
|
2011-10-14 04:00:16 +08:00
|
|
|
static void nfsd4_free_file(struct nfs4_file *f)
|
|
|
|
{
|
|
|
|
kmem_cache_free(file_slab, f);
|
|
|
|
}
|
|
|
|
|
2005-06-24 13:03:10 +08:00
|
|
|
static inline void
|
|
|
|
put_nfs4_file(struct nfs4_file *fi)
|
|
|
|
{
|
2009-02-23 06:51:34 +08:00
|
|
|
if (atomic_dec_and_lock(&fi->fi_ref, &recall_lock)) {
|
|
|
|
list_del(&fi->fi_hash);
|
|
|
|
spin_unlock(&recall_lock);
|
|
|
|
iput(fi->fi_inode);
|
2011-10-14 04:00:16 +08:00
|
|
|
nfsd4_free_file(fi);
|
2009-02-23 06:51:34 +08:00
|
|
|
}
|
2005-06-24 13:03:10 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
static inline void
|
|
|
|
get_nfs4_file(struct nfs4_file *fi)
|
|
|
|
{
|
2009-02-23 06:51:34 +08:00
|
|
|
atomic_inc(&fi->fi_ref);
|
2005-06-24 13:03:10 +08:00
|
|
|
}
|
|
|
|
|
2006-04-11 13:55:41 +08:00
|
|
|
static int num_delegations;
|
2007-07-17 19:04:39 +08:00
|
|
|
unsigned int max_delegations;
|
2006-04-11 13:55:41 +08:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Open owner state (share locks)
|
|
|
|
*/
|
|
|
|
|
2011-11-08 06:23:30 +08:00
|
|
|
/* hash tables for lock and open owners */
|
|
|
|
#define OWNER_HASH_BITS 8
|
|
|
|
#define OWNER_HASH_SIZE (1 << OWNER_HASH_BITS)
|
|
|
|
#define OWNER_HASH_MASK (OWNER_HASH_SIZE - 1)
|
2006-04-11 13:55:41 +08:00
|
|
|
|
2011-11-08 06:23:30 +08:00
|
|
|
static unsigned int ownerstr_hashval(u32 clientid, struct xdr_netobj *ownername)
|
2011-07-31 11:46:29 +08:00
|
|
|
{
|
|
|
|
unsigned int ret;
|
|
|
|
|
|
|
|
ret = opaque_hashval(ownername->data, ownername->len);
|
|
|
|
ret += clientid;
|
2011-11-08 06:23:30 +08:00
|
|
|
return ret & OWNER_HASH_MASK;
|
2011-07-31 11:46:29 +08:00
|
|
|
}
|
2006-04-11 13:55:41 +08:00
|
|
|
|
2011-11-08 06:23:30 +08:00
|
|
|
static struct list_head ownerstr_hashtbl[OWNER_HASH_SIZE];
|
2006-04-11 13:55:41 +08:00
|
|
|
|
|
|
|
/* hash table for nfs4_file */
|
|
|
|
#define FILE_HASH_BITS 8
|
|
|
|
#define FILE_HASH_SIZE (1 << FILE_HASH_BITS)
|
2011-01-14 17:35:59 +08:00
|
|
|
|
2011-07-31 11:46:29 +08:00
|
|
|
static unsigned int file_hashval(struct inode *ino)
|
|
|
|
{
|
|
|
|
/* XXX: why are we hashing on inode pointer, anyway? */
|
|
|
|
return hash_ptr(ino, FILE_HASH_BITS);
|
|
|
|
}
|
|
|
|
|
2006-04-11 13:55:41 +08:00
|
|
|
static struct list_head file_hashtbl[FILE_HASH_SIZE];
|
|
|
|
|
2010-08-07 21:21:41 +08:00
|
|
|
static void __nfs4_file_get_access(struct nfs4_file *fp, int oflag)
|
2010-07-08 23:02:09 +08:00
|
|
|
{
|
|
|
|
BUG_ON(!(fp->fi_fds[oflag] || fp->fi_fds[O_RDWR]));
|
|
|
|
atomic_inc(&fp->fi_access[oflag]);
|
|
|
|
}
|
|
|
|
|
2010-08-07 21:21:41 +08:00
|
|
|
static void nfs4_file_get_access(struct nfs4_file *fp, int oflag)
|
|
|
|
{
|
|
|
|
if (oflag == O_RDWR) {
|
|
|
|
__nfs4_file_get_access(fp, O_RDONLY);
|
|
|
|
__nfs4_file_get_access(fp, O_WRONLY);
|
|
|
|
} else
|
|
|
|
__nfs4_file_get_access(fp, oflag);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void nfs4_file_put_fd(struct nfs4_file *fp, int oflag)
|
2010-07-08 23:02:09 +08:00
|
|
|
{
|
|
|
|
if (fp->fi_fds[oflag]) {
|
|
|
|
fput(fp->fi_fds[oflag]);
|
|
|
|
fp->fi_fds[oflag] = NULL;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2010-08-07 21:21:41 +08:00
|
|
|
static void __nfs4_file_put_access(struct nfs4_file *fp, int oflag)
|
2010-07-08 23:02:09 +08:00
|
|
|
{
|
|
|
|
if (atomic_dec_and_test(&fp->fi_access[oflag])) {
|
|
|
|
nfs4_file_put_fd(fp, oflag);
|
2011-09-20 03:07:41 +08:00
|
|
|
/*
|
|
|
|
* It's also safe to get rid of the RDWR open *if*
|
|
|
|
* we no longer have need of the other kind of access
|
|
|
|
* or if we already have the other kind of open:
|
|
|
|
*/
|
|
|
|
if (fp->fi_fds[1-oflag]
|
|
|
|
|| atomic_read(&fp->fi_access[1 - oflag]) == 0)
|
|
|
|
nfs4_file_put_fd(fp, O_RDWR);
|
2010-07-08 23:02:09 +08:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2010-08-07 21:21:41 +08:00
|
|
|
static void nfs4_file_put_access(struct nfs4_file *fp, int oflag)
|
|
|
|
{
|
|
|
|
if (oflag == O_RDWR) {
|
|
|
|
__nfs4_file_put_access(fp, O_RDONLY);
|
|
|
|
__nfs4_file_put_access(fp, O_WRONLY);
|
|
|
|
} else
|
|
|
|
__nfs4_file_put_access(fp, oflag);
|
|
|
|
}
|
|
|
|
|
2011-09-24 04:21:15 +08:00
|
|
|
static inline int get_new_stid(struct nfs4_stid *stid)
|
2011-09-09 00:16:03 +08:00
|
|
|
{
|
2011-09-24 04:21:15 +08:00
|
|
|
static int min_stateid = 0;
|
2011-09-24 05:01:19 +08:00
|
|
|
struct idr *stateids = &stid->sc_client->cl_stateids;
|
2011-09-24 04:21:15 +08:00
|
|
|
int new_stid;
|
|
|
|
int error;
|
|
|
|
|
2011-09-24 05:01:19 +08:00
|
|
|
error = idr_get_new_above(stateids, stid, min_stateid, &new_stid);
|
2011-09-24 04:21:15 +08:00
|
|
|
/*
|
2011-10-17 23:14:48 +08:00
|
|
|
* Note: the necessary preallocation was done in
|
|
|
|
* nfs4_alloc_stateid(). The idr code caps the number of
|
|
|
|
* preallocations that can exist at a time, but the state lock
|
|
|
|
* prevents anyone from using ours before we get here:
|
2011-09-24 04:21:15 +08:00
|
|
|
*/
|
|
|
|
BUG_ON(error);
|
|
|
|
/*
|
|
|
|
* It shouldn't be a problem to reuse an opaque stateid value.
|
|
|
|
* I don't think it is for 4.1. But with 4.0 I worry that, for
|
|
|
|
* example, a stray write retransmission could be accepted by
|
|
|
|
* the server when it should have been rejected. Therefore,
|
|
|
|
* adopt a trick from the sctp code to attempt to maximize the
|
|
|
|
* amount of time until an id is reused, by ensuring they always
|
|
|
|
* "increase" (mod INT_MAX):
|
|
|
|
*/
|
2011-09-09 00:16:03 +08:00
|
|
|
|
2011-09-24 04:21:15 +08:00
|
|
|
min_stateid = new_stid+1;
|
|
|
|
if (min_stateid == INT_MAX)
|
|
|
|
min_stateid = 0;
|
|
|
|
return new_stid;
|
2011-09-09 00:16:03 +08:00
|
|
|
}
|
|
|
|
|
2011-10-17 23:14:48 +08:00
|
|
|
static void init_stid(struct nfs4_stid *stid, struct nfs4_client *cl, unsigned char type)
|
2011-09-24 05:20:02 +08:00
|
|
|
{
|
|
|
|
stateid_t *s = &stid->sc_stateid;
|
2011-09-24 04:21:15 +08:00
|
|
|
int new_id;
|
2011-09-24 05:20:02 +08:00
|
|
|
|
|
|
|
stid->sc_type = type;
|
|
|
|
stid->sc_client = cl;
|
|
|
|
s->si_opaque.so_clid = cl->cl_clientid;
|
2011-09-24 04:21:15 +08:00
|
|
|
new_id = get_new_stid(stid);
|
|
|
|
s->si_opaque.so_id = (u32)new_id;
|
2011-09-24 05:20:02 +08:00
|
|
|
/* Will be incremented before return to client: */
|
|
|
|
s->si_generation = 0;
|
2011-10-17 23:14:48 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
static struct nfs4_stid *nfs4_alloc_stid(struct nfs4_client *cl, struct kmem_cache *slab)
|
|
|
|
{
|
|
|
|
struct idr *stateids = &cl->cl_stateids;
|
|
|
|
|
|
|
|
if (!idr_pre_get(stateids, GFP_KERNEL))
|
|
|
|
return NULL;
|
|
|
|
/*
|
|
|
|
* Note: if we fail here (or any time between now and the time
|
|
|
|
* we actually get the new idr), we won't need to undo the idr
|
|
|
|
* preallocation, since the idr code caps the number of
|
|
|
|
* preallocated entries.
|
|
|
|
*/
|
|
|
|
return kmem_cache_alloc(slab, GFP_KERNEL);
|
2011-09-24 05:20:02 +08:00
|
|
|
}
|
|
|
|
|
2011-10-18 03:57:47 +08:00
|
|
|
static struct nfs4_ol_stateid * nfs4_alloc_stateid(struct nfs4_client *clp)
|
|
|
|
{
|
|
|
|
return openlockstateid(nfs4_alloc_stid(clp, stateid_slab));
|
|
|
|
}
|
|
|
|
|
2005-04-17 06:20:36 +08:00
|
|
|
static struct nfs4_delegation *
|
2011-09-08 04:06:42 +08:00
|
|
|
alloc_init_deleg(struct nfs4_client *clp, struct nfs4_ol_stateid *stp, struct svc_fh *current_fh, u32 type)
|
2005-04-17 06:20:36 +08:00
|
|
|
{
|
|
|
|
struct nfs4_delegation *dp;
|
|
|
|
struct nfs4_file *fp = stp->st_file;
|
|
|
|
|
|
|
|
dprintk("NFSD alloc_init_deleg\n");
|
2010-07-28 22:08:57 +08:00
|
|
|
/*
|
|
|
|
* Major work on the lease subsystem (for example, to support
|
|
|
|
* calbacks on stat) will be required before we can support
|
|
|
|
* write delegations properly.
|
|
|
|
*/
|
|
|
|
if (type != NFS4_OPEN_DELEGATE_READ)
|
|
|
|
return NULL;
|
2007-07-17 19:04:40 +08:00
|
|
|
if (fp->fi_had_conflict)
|
|
|
|
return NULL;
|
2007-07-17 19:04:39 +08:00
|
|
|
if (num_delegations > max_delegations)
|
2006-04-11 13:55:41 +08:00
|
|
|
return NULL;
|
2011-10-17 23:14:48 +08:00
|
|
|
dp = delegstateid(nfs4_alloc_stid(clp, deleg_slab));
|
2005-06-24 13:03:04 +08:00
|
|
|
if (dp == NULL)
|
2005-04-17 06:20:36 +08:00
|
|
|
return dp;
|
2011-10-17 23:14:48 +08:00
|
|
|
init_stid(&dp->dl_stid, clp, NFS4_DELEG_STID);
|
2011-09-24 05:20:02 +08:00
|
|
|
/*
|
|
|
|
* delegation seqid's are never incremented. The 4.1 special
|
2011-09-24 04:21:15 +08:00
|
|
|
* meaning of seqid 0 isn't meaningful, really, but let's avoid
|
|
|
|
* 0 anyway just for consistency and use 1:
|
2011-09-24 05:20:02 +08:00
|
|
|
*/
|
|
|
|
dp->dl_stid.sc_stateid.si_generation = 1;
|
2006-04-11 13:55:41 +08:00
|
|
|
num_delegations++;
|
2005-06-24 13:04:17 +08:00
|
|
|
INIT_LIST_HEAD(&dp->dl_perfile);
|
|
|
|
INIT_LIST_HEAD(&dp->dl_perclnt);
|
2005-04-17 06:20:36 +08:00
|
|
|
INIT_LIST_HEAD(&dp->dl_recall_lru);
|
2005-06-24 13:03:10 +08:00
|
|
|
get_nfs4_file(fp);
|
2005-04-17 06:20:36 +08:00
|
|
|
dp->dl_file = fp;
|
|
|
|
dp->dl_type = type;
|
2009-02-03 06:30:51 +08:00
|
|
|
fh_copy_shallow(&dp->dl_fh, ¤t_fh->fh_handle);
|
2005-04-17 06:20:36 +08:00
|
|
|
dp->dl_time = 0;
|
|
|
|
atomic_set(&dp->dl_count, 1);
|
2010-03-04 03:52:55 +08:00
|
|
|
INIT_WORK(&dp->dl_recall.cb_work, nfsd4_do_callback_rpc);
|
2005-04-17 06:20:36 +08:00
|
|
|
return dp;
|
|
|
|
}
|
|
|
|
|
|
|
|
void
|
|
|
|
nfs4_put_delegation(struct nfs4_delegation *dp)
|
|
|
|
{
|
|
|
|
if (atomic_dec_and_test(&dp->dl_count)) {
|
|
|
|
dprintk("NFSD: freeing dp %p\n",dp);
|
2005-06-24 13:03:10 +08:00
|
|
|
put_nfs4_file(dp->dl_file);
|
2005-06-24 13:03:04 +08:00
|
|
|
kmem_cache_free(deleg_slab, dp);
|
2006-04-11 13:55:41 +08:00
|
|
|
num_delegations--;
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2011-02-01 08:20:39 +08:00
|
|
|
static void nfs4_put_deleg_lease(struct nfs4_file *fp)
|
2005-04-17 06:20:36 +08:00
|
|
|
{
|
2011-02-01 08:20:39 +08:00
|
|
|
if (atomic_dec_and_test(&fp->fi_delegees)) {
|
|
|
|
vfs_setlease(fp->fi_deleg_file, F_UNLCK, &fp->fi_lease);
|
|
|
|
fp->fi_lease = NULL;
|
2011-04-16 06:08:26 +08:00
|
|
|
fput(fp->fi_deleg_file);
|
2011-02-01 08:20:39 +08:00
|
|
|
fp->fi_deleg_file = NULL;
|
|
|
|
}
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
|
|
|
|
2011-09-24 04:21:15 +08:00
|
|
|
static void unhash_stid(struct nfs4_stid *s)
|
|
|
|
{
|
2011-09-24 05:01:19 +08:00
|
|
|
struct idr *stateids = &s->sc_client->cl_stateids;
|
|
|
|
|
|
|
|
idr_remove(stateids, s->sc_stateid.si_opaque.so_id);
|
2011-09-24 04:21:15 +08:00
|
|
|
}
|
|
|
|
|
2005-04-17 06:20:36 +08:00
|
|
|
/* Called under the state lock. */
|
|
|
|
static void
|
|
|
|
unhash_delegation(struct nfs4_delegation *dp)
|
|
|
|
{
|
2011-09-24 04:21:15 +08:00
|
|
|
unhash_stid(&dp->dl_stid);
|
2005-06-24 13:04:17 +08:00
|
|
|
list_del_init(&dp->dl_perclnt);
|
2005-04-17 06:20:36 +08:00
|
|
|
spin_lock(&recall_lock);
|
2011-02-08 05:53:46 +08:00
|
|
|
list_del_init(&dp->dl_perfile);
|
2005-04-17 06:20:36 +08:00
|
|
|
list_del_init(&dp->dl_recall_lru);
|
|
|
|
spin_unlock(&recall_lock);
|
2011-02-01 08:20:39 +08:00
|
|
|
nfs4_put_deleg_lease(dp->dl_file);
|
2005-04-17 06:20:36 +08:00
|
|
|
nfs4_put_delegation(dp);
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* SETCLIENTID state
|
|
|
|
*/
|
|
|
|
|
2010-05-12 05:13:04 +08:00
|
|
|
/* client_lock protects the client lru list and session hash table */
|
2010-05-12 05:12:26 +08:00
|
|
|
static DEFINE_SPINLOCK(client_lock);
|
|
|
|
|
2005-04-17 06:20:36 +08:00
|
|
|
/* Hash tables for nfs4_clientid state */
|
|
|
|
#define CLIENT_HASH_BITS 4
|
|
|
|
#define CLIENT_HASH_SIZE (1 << CLIENT_HASH_BITS)
|
|
|
|
#define CLIENT_HASH_MASK (CLIENT_HASH_SIZE - 1)
|
|
|
|
|
2011-07-31 11:46:29 +08:00
|
|
|
static unsigned int clientid_hashval(u32 id)
|
|
|
|
{
|
|
|
|
return id & CLIENT_HASH_MASK;
|
|
|
|
}
|
|
|
|
|
|
|
|
static unsigned int clientstr_hashval(const char *name)
|
|
|
|
{
|
|
|
|
return opaque_hashval(name, 8) & CLIENT_HASH_MASK;
|
|
|
|
}
|
|
|
|
|
2005-04-17 06:20:36 +08:00
|
|
|
/*
|
|
|
|
* reclaim_str_hashtbl[] holds known client info from previous reset/reboot
|
|
|
|
* used in reboot/reset lease grace period processing
|
|
|
|
*
|
|
|
|
* conf_id_hashtbl[], and conf_str_hashtbl[] hold confirmed
|
|
|
|
* setclientid_confirmed info.
|
|
|
|
*
|
|
|
|
* unconf_str_hastbl[] and unconf_id_hashtbl[] hold unconfirmed
|
|
|
|
* setclientid info.
|
|
|
|
*
|
|
|
|
* client_lru holds client queue ordered by nfs4_client.cl_time
|
|
|
|
* for lease renewal.
|
|
|
|
*
|
|
|
|
* close_lru holds (open) stateowner queue ordered by nfs4_stateowner.so_time
|
|
|
|
* for last close replay.
|
|
|
|
*/
|
|
|
|
static struct list_head reclaim_str_hashtbl[CLIENT_HASH_SIZE];
|
|
|
|
static int reclaim_str_hashtbl_size = 0;
|
|
|
|
static struct list_head conf_id_hashtbl[CLIENT_HASH_SIZE];
|
|
|
|
static struct list_head conf_str_hashtbl[CLIENT_HASH_SIZE];
|
|
|
|
static struct list_head unconf_str_hashtbl[CLIENT_HASH_SIZE];
|
|
|
|
static struct list_head unconf_id_hashtbl[CLIENT_HASH_SIZE];
|
|
|
|
static struct list_head client_lru;
|
|
|
|
static struct list_head close_lru;
|
|
|
|
|
2010-07-08 23:02:09 +08:00
|
|
|
/*
|
|
|
|
* We store the NONE, READ, WRITE, and BOTH bits separately in the
|
|
|
|
* st_{access,deny}_bmap field of the stateid, in order to track not
|
|
|
|
* only what share bits are currently in force, but also what
|
|
|
|
* combinations of share bits previous opens have used. This allows us
|
|
|
|
* to enforce the recommendation of rfc 3530 14.2.19 that the server
|
|
|
|
* return an error if the client attempt to downgrade to a combination
|
|
|
|
* of share bits not explicable by closing some of its previous opens.
|
|
|
|
*
|
|
|
|
* XXX: This enforcement is actually incomplete, since we don't keep
|
|
|
|
* track of access/deny bit combinations; so, e.g., we allow:
|
|
|
|
*
|
|
|
|
* OPEN allow read, deny write
|
|
|
|
* OPEN allow both, deny none
|
|
|
|
* DOWNGRADE allow read, deny none
|
|
|
|
*
|
|
|
|
* which we should reject.
|
|
|
|
*/
|
|
|
|
static void
|
|
|
|
set_access(unsigned int *access, unsigned long bmap) {
|
|
|
|
int i;
|
|
|
|
|
|
|
|
*access = 0;
|
|
|
|
for (i = 1; i < 4; i++) {
|
|
|
|
if (test_bit(i, &bmap))
|
|
|
|
*access |= i;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
static void
|
|
|
|
set_deny(unsigned int *deny, unsigned long bmap) {
|
|
|
|
int i;
|
|
|
|
|
|
|
|
*deny = 0;
|
|
|
|
for (i = 0; i < 4; i++) {
|
|
|
|
if (test_bit(i, &bmap))
|
|
|
|
*deny |= i ;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
static int
|
2011-09-08 04:06:42 +08:00
|
|
|
test_share(struct nfs4_ol_stateid *stp, struct nfsd4_open *open) {
|
2010-07-08 23:02:09 +08:00
|
|
|
unsigned int access, deny;
|
|
|
|
|
|
|
|
set_access(&access, stp->st_access_bmap);
|
|
|
|
set_deny(&deny, stp->st_deny_bmap);
|
|
|
|
if ((access & open->op_share_deny) || (deny & open->op_share_access))
|
|
|
|
return 0;
|
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int nfs4_access_to_omode(u32 access)
|
|
|
|
{
|
2010-09-03 03:23:16 +08:00
|
|
|
switch (access & NFS4_SHARE_ACCESS_BOTH) {
|
2010-07-08 23:02:09 +08:00
|
|
|
case NFS4_SHARE_ACCESS_READ:
|
|
|
|
return O_RDONLY;
|
|
|
|
case NFS4_SHARE_ACCESS_WRITE:
|
|
|
|
return O_WRONLY;
|
|
|
|
case NFS4_SHARE_ACCESS_BOTH:
|
|
|
|
return O_RDWR;
|
|
|
|
}
|
|
|
|
BUG();
|
|
|
|
}
|
|
|
|
|
2011-09-08 04:06:42 +08:00
|
|
|
static void unhash_generic_stateid(struct nfs4_ol_stateid *stp)
|
2011-03-03 12:48:33 +08:00
|
|
|
{
|
|
|
|
list_del(&stp->st_perfile);
|
|
|
|
list_del(&stp->st_perstateowner);
|
|
|
|
}
|
|
|
|
|
2011-09-08 04:06:42 +08:00
|
|
|
static void close_generic_stateid(struct nfs4_ol_stateid *stp)
|
2011-03-03 12:48:33 +08:00
|
|
|
{
|
2011-06-28 04:57:12 +08:00
|
|
|
int i;
|
2011-03-03 07:01:35 +08:00
|
|
|
|
2011-03-28 15:15:09 +08:00
|
|
|
if (stp->st_access_bmap) {
|
2011-06-28 04:57:12 +08:00
|
|
|
for (i = 1; i < 4; i++) {
|
|
|
|
if (test_bit(i, &stp->st_access_bmap))
|
|
|
|
nfs4_file_put_access(stp->st_file,
|
|
|
|
nfs4_access_to_omode(i));
|
2011-09-07 02:50:49 +08:00
|
|
|
__clear_bit(i, &stp->st_access_bmap);
|
2011-06-28 04:57:12 +08:00
|
|
|
}
|
2011-03-28 15:15:09 +08:00
|
|
|
}
|
2011-04-18 23:48:55 +08:00
|
|
|
put_nfs4_file(stp->st_file);
|
2011-09-07 02:50:49 +08:00
|
|
|
stp->st_file = NULL;
|
|
|
|
}
|
|
|
|
|
2011-09-08 04:06:42 +08:00
|
|
|
static void free_generic_stateid(struct nfs4_ol_stateid *stp)
|
2011-09-07 02:50:49 +08:00
|
|
|
{
|
2011-03-03 12:48:33 +08:00
|
|
|
kmem_cache_free(stateid_slab, stp);
|
|
|
|
}
|
|
|
|
|
2011-09-08 04:06:42 +08:00
|
|
|
static void release_lock_stateid(struct nfs4_ol_stateid *stp)
|
2011-03-03 12:48:33 +08:00
|
|
|
{
|
|
|
|
struct file *file;
|
|
|
|
|
|
|
|
unhash_generic_stateid(stp);
|
2011-09-24 04:21:15 +08:00
|
|
|
unhash_stid(&stp->st_stid);
|
2011-03-03 12:48:33 +08:00
|
|
|
file = find_any_file(stp->st_file);
|
|
|
|
if (file)
|
2011-07-31 11:33:59 +08:00
|
|
|
locks_remove_posix(file, (fl_owner_t)lockowner(stp->st_stateowner));
|
2011-09-17 05:42:48 +08:00
|
|
|
close_generic_stateid(stp);
|
2011-03-03 12:48:33 +08:00
|
|
|
free_generic_stateid(stp);
|
|
|
|
}
|
|
|
|
|
2011-07-31 11:33:59 +08:00
|
|
|
static void unhash_lockowner(struct nfs4_lockowner *lo)
|
2011-03-03 12:48:33 +08:00
|
|
|
{
|
2011-09-08 04:06:42 +08:00
|
|
|
struct nfs4_ol_stateid *stp;
|
2011-03-03 12:48:33 +08:00
|
|
|
|
2011-07-31 11:33:59 +08:00
|
|
|
list_del(&lo->lo_owner.so_strhash);
|
|
|
|
list_del(&lo->lo_perstateid);
|
2011-11-08 06:40:10 +08:00
|
|
|
list_del(&lo->lo_owner_ino_hash);
|
2011-07-31 11:33:59 +08:00
|
|
|
while (!list_empty(&lo->lo_owner.so_stateids)) {
|
|
|
|
stp = list_first_entry(&lo->lo_owner.so_stateids,
|
2011-09-08 04:06:42 +08:00
|
|
|
struct nfs4_ol_stateid, st_perstateowner);
|
2011-03-03 12:48:33 +08:00
|
|
|
release_lock_stateid(stp);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2011-07-31 11:33:59 +08:00
|
|
|
static void release_lockowner(struct nfs4_lockowner *lo)
|
2011-03-03 12:48:33 +08:00
|
|
|
{
|
2011-07-31 11:33:59 +08:00
|
|
|
unhash_lockowner(lo);
|
|
|
|
nfs4_free_lockowner(lo);
|
2011-03-03 12:48:33 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
static void
|
2011-09-08 04:06:42 +08:00
|
|
|
release_stateid_lockowners(struct nfs4_ol_stateid *open_stp)
|
2011-03-03 12:48:33 +08:00
|
|
|
{
|
2011-07-31 11:33:59 +08:00
|
|
|
struct nfs4_lockowner *lo;
|
2011-03-03 12:48:33 +08:00
|
|
|
|
|
|
|
while (!list_empty(&open_stp->st_lockowners)) {
|
2011-07-31 11:33:59 +08:00
|
|
|
lo = list_entry(open_stp->st_lockowners.next,
|
|
|
|
struct nfs4_lockowner, lo_perstateid);
|
|
|
|
release_lockowner(lo);
|
2011-03-03 12:48:33 +08:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2011-09-17 05:42:48 +08:00
|
|
|
static void unhash_open_stateid(struct nfs4_ol_stateid *stp)
|
2009-01-12 03:27:17 +08:00
|
|
|
{
|
|
|
|
unhash_generic_stateid(stp);
|
|
|
|
release_stateid_lockowners(stp);
|
2011-09-17 05:42:48 +08:00
|
|
|
close_generic_stateid(stp);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void release_open_stateid(struct nfs4_ol_stateid *stp)
|
|
|
|
{
|
|
|
|
unhash_open_stateid(stp);
|
2011-09-24 04:21:15 +08:00
|
|
|
unhash_stid(&stp->st_stid);
|
2009-01-12 03:27:17 +08:00
|
|
|
free_generic_stateid(stp);
|
|
|
|
}
|
|
|
|
|
2011-07-31 11:33:59 +08:00
|
|
|
static void unhash_openowner(struct nfs4_openowner *oo)
|
2009-01-12 03:37:31 +08:00
|
|
|
{
|
2011-09-08 04:06:42 +08:00
|
|
|
struct nfs4_ol_stateid *stp;
|
2009-01-12 03:37:31 +08:00
|
|
|
|
2011-07-31 11:33:59 +08:00
|
|
|
list_del(&oo->oo_owner.so_strhash);
|
|
|
|
list_del(&oo->oo_perclient);
|
|
|
|
while (!list_empty(&oo->oo_owner.so_stateids)) {
|
|
|
|
stp = list_first_entry(&oo->oo_owner.so_stateids,
|
2011-09-08 04:06:42 +08:00
|
|
|
struct nfs4_ol_stateid, st_perstateowner);
|
2009-01-12 04:24:04 +08:00
|
|
|
release_open_stateid(stp);
|
2009-01-12 03:37:31 +08:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2011-09-17 08:12:38 +08:00
|
|
|
static void release_last_closed_stateid(struct nfs4_openowner *oo)
|
|
|
|
{
|
|
|
|
struct nfs4_ol_stateid *s = oo->oo_last_closed_stid;
|
|
|
|
|
|
|
|
if (s) {
|
2011-09-24 04:21:15 +08:00
|
|
|
unhash_stid(&s->st_stid);
|
2011-09-17 08:12:38 +08:00
|
|
|
free_generic_stateid(s);
|
|
|
|
oo->oo_last_closed_stid = NULL;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2011-07-31 11:33:59 +08:00
|
|
|
static void release_openowner(struct nfs4_openowner *oo)
|
2009-01-12 03:37:31 +08:00
|
|
|
{
|
2011-07-31 11:33:59 +08:00
|
|
|
unhash_openowner(oo);
|
|
|
|
list_del(&oo->oo_close_lru);
|
2011-09-17 08:12:38 +08:00
|
|
|
release_last_closed_stateid(oo);
|
2011-07-31 11:33:59 +08:00
|
|
|
nfs4_free_openowner(oo);
|
2009-01-12 03:37:31 +08:00
|
|
|
}
|
|
|
|
|
2009-04-03 13:27:52 +08:00
|
|
|
#define SESSION_HASH_SIZE 512
|
|
|
|
static struct list_head sessionid_hashtbl[SESSION_HASH_SIZE];
|
|
|
|
|
|
|
|
static inline int
|
|
|
|
hash_sessionid(struct nfs4_sessionid *sessionid)
|
|
|
|
{
|
|
|
|
struct nfsd4_sessionid *sid = (struct nfsd4_sessionid *)sessionid;
|
|
|
|
|
|
|
|
return sid->sequence % SESSION_HASH_SIZE;
|
|
|
|
}
|
|
|
|
|
2012-03-21 03:11:17 +08:00
|
|
|
#ifdef NFSD_DEBUG
|
2009-04-03 13:27:52 +08:00
|
|
|
static inline void
|
|
|
|
dump_sessionid(const char *fn, struct nfs4_sessionid *sessionid)
|
|
|
|
{
|
|
|
|
u32 *ptr = (u32 *)(&sessionid->data[0]);
|
|
|
|
dprintk("%s: %u:%u:%u:%u\n", fn, ptr[0], ptr[1], ptr[2], ptr[3]);
|
|
|
|
}
|
2012-03-21 03:11:17 +08:00
|
|
|
#else
|
|
|
|
static inline void
|
|
|
|
dump_sessionid(const char *fn, struct nfs4_sessionid *sessionid)
|
|
|
|
{
|
|
|
|
}
|
|
|
|
#endif
|
|
|
|
|
2009-04-03 13:27:52 +08:00
|
|
|
|
2009-04-03 13:28:28 +08:00
|
|
|
static void
|
|
|
|
gen_sessionid(struct nfsd4_session *ses)
|
|
|
|
{
|
|
|
|
struct nfs4_client *clp = ses->se_client;
|
|
|
|
struct nfsd4_sessionid *sid;
|
|
|
|
|
|
|
|
sid = (struct nfsd4_sessionid *)ses->se_sessionid.data;
|
|
|
|
sid->clientid = clp->cl_clientid;
|
|
|
|
sid->sequence = current_sessionid++;
|
|
|
|
sid->reserved = 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
2009-08-28 20:45:01 +08:00
|
|
|
* The protocol defines ca_maxresponssize_cached to include the size of
|
|
|
|
* the rpc header, but all we need to cache is the data starting after
|
|
|
|
* the end of the initial SEQUENCE operation--the rest we regenerate
|
|
|
|
* each time. Therefore we can advertise a ca_maxresponssize_cached
|
|
|
|
* value that is the number of bytes in our cache plus a few additional
|
|
|
|
* bytes. In order to stay on the safe side, and not promise more than
|
|
|
|
* we can cache, those additional bytes must be the minimum possible: 24
|
|
|
|
* bytes of rpc header (xid through accept state, with AUTH_NULL
|
|
|
|
* verifier), 12 for the compound header (with zero-length tag), and 44
|
|
|
|
* for the SEQUENCE op response:
|
|
|
|
*/
|
|
|
|
#define NFSD_MIN_HDR_SEQ_SZ (24 + 12 + 44)
|
|
|
|
|
2009-08-28 20:45:04 +08:00
|
|
|
static void
|
|
|
|
free_session_slots(struct nfsd4_session *ses)
|
|
|
|
{
|
|
|
|
int i;
|
|
|
|
|
|
|
|
for (i = 0; i < ses->se_fchannel.maxreqs; i++)
|
|
|
|
kfree(ses->se_slots[i]);
|
|
|
|
}
|
|
|
|
|
2009-08-28 20:45:01 +08:00
|
|
|
/*
|
2009-10-25 08:52:16 +08:00
|
|
|
* We don't actually need to cache the rpc and session headers, so we
|
|
|
|
* can allocate a little less for each slot:
|
|
|
|
*/
|
|
|
|
static inline int slot_bytes(struct nfsd4_channel_attrs *ca)
|
|
|
|
{
|
|
|
|
return ca->maxresp_cached - NFSD_MIN_HDR_SEQ_SZ;
|
|
|
|
}
|
|
|
|
|
2010-09-28 05:12:05 +08:00
|
|
|
static int nfsd4_sanitize_slot_size(u32 size)
|
2009-04-03 13:28:28 +08:00
|
|
|
{
|
2010-09-28 05:12:05 +08:00
|
|
|
size -= NFSD_MIN_HDR_SEQ_SZ; /* We don't cache the rpc header */
|
|
|
|
size = min_t(u32, size, NFSD_SLOT_CACHE_SIZE);
|
2009-04-03 13:28:28 +08:00
|
|
|
|
2010-09-28 05:12:05 +08:00
|
|
|
return size;
|
|
|
|
}
|
2009-04-03 13:28:28 +08:00
|
|
|
|
2010-09-28 05:12:05 +08:00
|
|
|
/*
|
|
|
|
* XXX: If we run out of reserved DRC memory we could (up to a point)
|
2009-08-28 20:45:01 +08:00
|
|
|
* re-negotiate active sessions and reduce their slot usage to make
|
2011-11-29 12:31:00 +08:00
|
|
|
* room for new connections. For now we just fail the create session.
|
2009-04-03 13:28:28 +08:00
|
|
|
*/
|
2010-09-28 05:12:05 +08:00
|
|
|
static int nfsd4_get_drc_mem(int slotsize, u32 num)
|
2009-04-03 13:28:28 +08:00
|
|
|
{
|
2010-09-28 05:12:05 +08:00
|
|
|
int avail;
|
2009-04-03 13:28:28 +08:00
|
|
|
|
2010-09-28 05:12:05 +08:00
|
|
|
num = min_t(u32, num, NFSD_MAX_SLOTS_PER_SESSION);
|
2009-06-16 09:19:38 +08:00
|
|
|
|
2010-09-28 05:12:05 +08:00
|
|
|
spin_lock(&nfsd_drc_lock);
|
|
|
|
avail = min_t(int, NFSD_MAX_MEM_PER_SESSION,
|
|
|
|
nfsd_drc_max_mem - nfsd_drc_mem_used);
|
|
|
|
num = min_t(int, num, avail / slotsize);
|
|
|
|
nfsd_drc_mem_used += num * slotsize;
|
|
|
|
spin_unlock(&nfsd_drc_lock);
|
2009-04-03 13:28:28 +08:00
|
|
|
|
2010-09-28 05:12:05 +08:00
|
|
|
return num;
|
|
|
|
}
|
2009-04-03 13:28:28 +08:00
|
|
|
|
2010-09-28 05:12:05 +08:00
|
|
|
static void nfsd4_put_drc_mem(int slotsize, int num)
|
|
|
|
{
|
2009-06-25 03:37:45 +08:00
|
|
|
spin_lock(&nfsd_drc_lock);
|
2010-09-28 05:12:05 +08:00
|
|
|
nfsd_drc_mem_used -= slotsize * num;
|
2009-06-25 03:37:45 +08:00
|
|
|
spin_unlock(&nfsd_drc_lock);
|
2010-09-28 05:12:05 +08:00
|
|
|
}
|
2009-04-03 13:28:28 +08:00
|
|
|
|
2010-09-28 05:12:05 +08:00
|
|
|
static struct nfsd4_session *alloc_session(int slotsize, int numslots)
|
|
|
|
{
|
|
|
|
struct nfsd4_session *new;
|
|
|
|
int mem, i;
|
2009-08-28 20:45:01 +08:00
|
|
|
|
2010-09-28 05:12:05 +08:00
|
|
|
BUILD_BUG_ON(NFSD_MAX_SLOTS_PER_SESSION * sizeof(struct nfsd4_slot *)
|
|
|
|
+ sizeof(struct nfsd4_session) > PAGE_SIZE);
|
|
|
|
mem = numslots * sizeof(struct nfsd4_slot *);
|
2009-04-03 13:28:28 +08:00
|
|
|
|
2010-09-28 05:12:05 +08:00
|
|
|
new = kzalloc(sizeof(*new) + mem, GFP_KERNEL);
|
|
|
|
if (!new)
|
|
|
|
return NULL;
|
2009-08-28 20:45:04 +08:00
|
|
|
/* allocate each struct nfsd4_slot and data cache in one piece */
|
2010-09-28 05:12:05 +08:00
|
|
|
for (i = 0; i < numslots; i++) {
|
|
|
|
mem = sizeof(struct nfsd4_slot) + slotsize;
|
|
|
|
new->se_slots[i] = kzalloc(mem, GFP_KERNEL);
|
|
|
|
if (!new->se_slots[i])
|
2009-08-28 20:45:04 +08:00
|
|
|
goto out_free;
|
|
|
|
}
|
2010-09-28 05:12:05 +08:00
|
|
|
return new;
|
|
|
|
out_free:
|
|
|
|
while (i--)
|
|
|
|
kfree(new->se_slots[i]);
|
|
|
|
kfree(new);
|
|
|
|
return NULL;
|
2009-04-03 13:28:28 +08:00
|
|
|
}
|
|
|
|
|
2010-09-28 05:12:05 +08:00
|
|
|
static void init_forechannel_attrs(struct nfsd4_channel_attrs *new, struct nfsd4_channel_attrs *req, int numslots, int slotsize)
|
2009-04-03 13:28:28 +08:00
|
|
|
{
|
2010-09-28 05:12:05 +08:00
|
|
|
u32 maxrpc = nfsd_serv->sv_max_mesg;
|
2009-04-03 13:28:28 +08:00
|
|
|
|
2010-09-28 05:12:05 +08:00
|
|
|
new->maxreqs = numslots;
|
2011-03-10 17:43:37 +08:00
|
|
|
new->maxresp_cached = min_t(u32, req->maxresp_cached,
|
|
|
|
slotsize + NFSD_MIN_HDR_SEQ_SZ);
|
2010-09-28 05:12:05 +08:00
|
|
|
new->maxreq_sz = min_t(u32, req->maxreq_sz, maxrpc);
|
|
|
|
new->maxresp_sz = min_t(u32, req->maxresp_sz, maxrpc);
|
|
|
|
new->maxops = min_t(u32, req->maxops, NFSD_MAX_OPS_PER_COMPOUND);
|
|
|
|
}
|
2009-04-03 13:28:28 +08:00
|
|
|
|
2010-06-07 06:37:16 +08:00
|
|
|
static void free_conn(struct nfsd4_conn *c)
|
|
|
|
{
|
|
|
|
svc_xprt_put(c->cn_xprt);
|
|
|
|
kfree(c);
|
|
|
|
}
|
2009-04-03 13:28:28 +08:00
|
|
|
|
2010-06-07 06:37:16 +08:00
|
|
|
static void nfsd4_conn_lost(struct svc_xpt_user *u)
|
|
|
|
{
|
|
|
|
struct nfsd4_conn *c = container_of(u, struct nfsd4_conn, cn_xpt_user);
|
|
|
|
struct nfs4_client *clp = c->cn_session->se_client;
|
2009-04-03 13:28:28 +08:00
|
|
|
|
2010-06-07 06:37:16 +08:00
|
|
|
spin_lock(&clp->cl_lock);
|
|
|
|
if (!list_empty(&c->cn_persession)) {
|
|
|
|
list_del(&c->cn_persession);
|
|
|
|
free_conn(c);
|
|
|
|
}
|
|
|
|
spin_unlock(&clp->cl_lock);
|
2010-11-18 21:34:12 +08:00
|
|
|
nfsd4_probe_callback(clp);
|
2010-06-07 06:37:16 +08:00
|
|
|
}
|
2009-04-03 13:28:28 +08:00
|
|
|
|
2010-06-16 05:34:11 +08:00
|
|
|
static struct nfsd4_conn *alloc_conn(struct svc_rqst *rqstp, u32 flags)
|
2010-06-07 06:12:14 +08:00
|
|
|
{
|
|
|
|
struct nfsd4_conn *conn;
|
2009-04-03 13:28:28 +08:00
|
|
|
|
2010-06-07 06:12:14 +08:00
|
|
|
conn = kmalloc(sizeof(struct nfsd4_conn), GFP_KERNEL);
|
|
|
|
if (!conn)
|
2010-09-30 03:29:32 +08:00
|
|
|
return NULL;
|
2010-06-07 06:12:14 +08:00
|
|
|
svc_xprt_get(rqstp->rq_xprt);
|
|
|
|
conn->cn_xprt = rqstp->rq_xprt;
|
2010-06-16 05:34:11 +08:00
|
|
|
conn->cn_flags = flags;
|
2010-09-30 03:29:32 +08:00
|
|
|
INIT_LIST_HEAD(&conn->cn_xpt_user.list);
|
|
|
|
return conn;
|
|
|
|
}
|
2009-08-28 20:45:01 +08:00
|
|
|
|
2010-09-30 04:11:06 +08:00
|
|
|
static void __nfsd4_hash_conn(struct nfsd4_conn *conn, struct nfsd4_session *ses)
|
|
|
|
{
|
|
|
|
conn->cn_session = ses;
|
|
|
|
list_add(&conn->cn_persession, &ses->se_conns);
|
2009-04-03 13:28:28 +08:00
|
|
|
}
|
|
|
|
|
2010-09-30 03:29:32 +08:00
|
|
|
static void nfsd4_hash_conn(struct nfsd4_conn *conn, struct nfsd4_session *ses)
|
2009-08-28 20:45:04 +08:00
|
|
|
{
|
2010-09-30 03:29:32 +08:00
|
|
|
struct nfs4_client *clp = ses->se_client;
|
2009-08-28 20:45:04 +08:00
|
|
|
|
2010-06-07 06:12:14 +08:00
|
|
|
spin_lock(&clp->cl_lock);
|
2010-09-30 04:11:06 +08:00
|
|
|
__nfsd4_hash_conn(conn, ses);
|
2010-06-07 06:12:14 +08:00
|
|
|
spin_unlock(&clp->cl_lock);
|
2009-08-28 20:45:04 +08:00
|
|
|
}
|
|
|
|
|
2010-10-26 22:07:17 +08:00
|
|
|
static int nfsd4_register_conn(struct nfsd4_conn *conn)
|
2009-10-25 08:52:16 +08:00
|
|
|
{
|
2010-06-07 06:37:16 +08:00
|
|
|
conn->cn_xpt_user.callback = nfsd4_conn_lost;
|
2010-10-26 22:07:17 +08:00
|
|
|
return register_xpt_user(conn->cn_xprt, &conn->cn_xpt_user);
|
2009-10-25 08:52:16 +08:00
|
|
|
}
|
|
|
|
|
2010-10-05 11:12:59 +08:00
|
|
|
static __be32 nfsd4_new_conn(struct svc_rqst *rqstp, struct nfsd4_session *ses, u32 dir)
|
2009-04-03 13:28:28 +08:00
|
|
|
{
|
2010-09-30 03:29:32 +08:00
|
|
|
struct nfsd4_conn *conn;
|
2010-10-26 22:07:17 +08:00
|
|
|
int ret;
|
2009-04-03 13:28:28 +08:00
|
|
|
|
2010-10-05 11:12:59 +08:00
|
|
|
conn = alloc_conn(rqstp, dir);
|
2010-09-30 03:29:32 +08:00
|
|
|
if (!conn)
|
|
|
|
return nfserr_jukebox;
|
|
|
|
nfsd4_hash_conn(conn, ses);
|
2010-10-26 22:07:17 +08:00
|
|
|
ret = nfsd4_register_conn(conn);
|
|
|
|
if (ret)
|
|
|
|
/* oops; xprt is already down: */
|
|
|
|
nfsd4_conn_lost(&conn->cn_xpt_user);
|
2010-06-07 06:12:14 +08:00
|
|
|
return nfs_ok;
|
|
|
|
}
|
2009-04-03 13:28:28 +08:00
|
|
|
|
2010-10-05 11:12:59 +08:00
|
|
|
static __be32 nfsd4_new_conn_from_crses(struct svc_rqst *rqstp, struct nfsd4_session *ses)
|
|
|
|
{
|
|
|
|
u32 dir = NFS4_CDFC4_FORE;
|
|
|
|
|
|
|
|
if (ses->se_flags & SESSION4_BACK_CHAN)
|
|
|
|
dir |= NFS4_CDFC4_BACK;
|
|
|
|
|
|
|
|
return nfsd4_new_conn(rqstp, ses, dir);
|
|
|
|
}
|
|
|
|
|
|
|
|
/* must be called under client_lock */
|
2010-06-07 06:37:16 +08:00
|
|
|
static void nfsd4_del_conns(struct nfsd4_session *s)
|
2010-06-07 06:12:14 +08:00
|
|
|
{
|
2010-06-07 06:37:16 +08:00
|
|
|
struct nfs4_client *clp = s->se_client;
|
|
|
|
struct nfsd4_conn *c;
|
2009-04-03 13:28:28 +08:00
|
|
|
|
2010-06-07 06:37:16 +08:00
|
|
|
spin_lock(&clp->cl_lock);
|
|
|
|
while (!list_empty(&s->se_conns)) {
|
|
|
|
c = list_first_entry(&s->se_conns, struct nfsd4_conn, cn_persession);
|
|
|
|
list_del_init(&c->cn_persession);
|
|
|
|
spin_unlock(&clp->cl_lock);
|
2009-08-28 20:45:04 +08:00
|
|
|
|
2010-06-07 06:37:16 +08:00
|
|
|
unregister_xpt_user(c->cn_xprt, &c->cn_xpt_user);
|
|
|
|
free_conn(c);
|
2009-04-03 13:28:28 +08:00
|
|
|
|
2010-06-07 06:37:16 +08:00
|
|
|
spin_lock(&clp->cl_lock);
|
|
|
|
}
|
|
|
|
spin_unlock(&clp->cl_lock);
|
2010-06-07 06:12:14 +08:00
|
|
|
}
|
2009-04-03 13:28:28 +08:00
|
|
|
|
2012-02-24 09:40:52 +08:00
|
|
|
static void free_session(struct kref *kref)
|
2010-06-07 06:12:14 +08:00
|
|
|
{
|
|
|
|
struct nfsd4_session *ses;
|
|
|
|
int mem;
|
|
|
|
|
2012-02-24 09:40:52 +08:00
|
|
|
BUG_ON(!spin_is_locked(&client_lock));
|
2010-06-07 06:12:14 +08:00
|
|
|
ses = container_of(kref, struct nfsd4_session, se_ref);
|
2010-06-07 06:37:16 +08:00
|
|
|
nfsd4_del_conns(ses);
|
2010-06-07 06:12:14 +08:00
|
|
|
spin_lock(&nfsd_drc_lock);
|
|
|
|
mem = ses->se_fchannel.maxreqs * slot_bytes(&ses->se_fchannel);
|
|
|
|
nfsd_drc_mem_used -= mem;
|
|
|
|
spin_unlock(&nfsd_drc_lock);
|
|
|
|
free_session_slots(ses);
|
|
|
|
kfree(ses);
|
|
|
|
}
|
|
|
|
|
2012-02-24 09:40:52 +08:00
|
|
|
void nfsd4_put_session(struct nfsd4_session *ses)
|
|
|
|
{
|
|
|
|
spin_lock(&client_lock);
|
|
|
|
nfsd4_put_session_locked(ses);
|
|
|
|
spin_unlock(&client_lock);
|
|
|
|
}
|
|
|
|
|
2010-06-15 07:01:57 +08:00
|
|
|
static struct nfsd4_session *alloc_init_session(struct svc_rqst *rqstp, struct nfs4_client *clp, struct nfsd4_create_session *cses)
|
2010-09-28 05:12:05 +08:00
|
|
|
{
|
|
|
|
struct nfsd4_session *new;
|
|
|
|
struct nfsd4_channel_attrs *fchan = &cses->fore_channel;
|
|
|
|
int numslots, slotsize;
|
2010-06-07 06:12:14 +08:00
|
|
|
int status;
|
2010-09-28 05:12:05 +08:00
|
|
|
int idx;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Note decreasing slot size below client's request may
|
|
|
|
* make it difficult for client to function correctly, whereas
|
|
|
|
* decreasing the number of slots will (just?) affect
|
|
|
|
* performance. When short on memory we therefore prefer to
|
|
|
|
* decrease number of slots instead of their size.
|
|
|
|
*/
|
|
|
|
slotsize = nfsd4_sanitize_slot_size(fchan->maxresp_cached);
|
|
|
|
numslots = nfsd4_get_drc_mem(slotsize, fchan->maxreqs);
|
2010-11-11 18:03:50 +08:00
|
|
|
if (numslots < 1)
|
|
|
|
return NULL;
|
2010-09-28 05:12:05 +08:00
|
|
|
|
|
|
|
new = alloc_session(slotsize, numslots);
|
|
|
|
if (!new) {
|
|
|
|
nfsd4_put_drc_mem(slotsize, fchan->maxreqs);
|
2010-06-15 07:01:57 +08:00
|
|
|
return NULL;
|
2009-08-28 20:45:04 +08:00
|
|
|
}
|
2010-09-28 05:12:05 +08:00
|
|
|
init_forechannel_attrs(&new->se_fchannel, fchan, numslots, slotsize);
|
2009-08-28 20:45:04 +08:00
|
|
|
|
2009-04-03 13:28:28 +08:00
|
|
|
new->se_client = clp;
|
|
|
|
gen_sessionid(new);
|
|
|
|
|
2010-06-07 06:12:14 +08:00
|
|
|
INIT_LIST_HEAD(&new->se_conns);
|
|
|
|
|
2010-06-15 07:01:57 +08:00
|
|
|
new->se_cb_seq_nr = 1;
|
2009-04-03 13:28:28 +08:00
|
|
|
new->se_flags = cses->flags;
|
2010-10-20 05:31:50 +08:00
|
|
|
new->se_cb_prog = cses->callback_prog;
|
2009-04-03 13:28:28 +08:00
|
|
|
kref_init(&new->se_ref);
|
2010-09-28 05:12:05 +08:00
|
|
|
idx = hash_sessionid(&new->se_sessionid);
|
2010-05-12 05:12:26 +08:00
|
|
|
spin_lock(&client_lock);
|
2009-04-03 13:28:28 +08:00
|
|
|
list_add(&new->se_hash, &sessionid_hashtbl[idx]);
|
2010-06-16 02:22:37 +08:00
|
|
|
spin_lock(&clp->cl_lock);
|
2009-04-03 13:28:28 +08:00
|
|
|
list_add(&new->se_perclnt, &clp->cl_sessions);
|
2010-06-16 02:22:37 +08:00
|
|
|
spin_unlock(&clp->cl_lock);
|
2010-05-12 05:12:26 +08:00
|
|
|
spin_unlock(&client_lock);
|
2009-04-03 13:28:28 +08:00
|
|
|
|
2010-10-05 11:12:59 +08:00
|
|
|
status = nfsd4_new_conn_from_crses(rqstp, new);
|
2010-06-15 07:01:57 +08:00
|
|
|
/* whoops: benny points out, status is ignored! (err, or bogus) */
|
2010-06-07 06:12:14 +08:00
|
|
|
if (status) {
|
2012-02-24 09:40:52 +08:00
|
|
|
spin_lock(&client_lock);
|
2010-06-07 06:12:14 +08:00
|
|
|
free_session(&new->se_ref);
|
2012-02-24 09:40:52 +08:00
|
|
|
spin_unlock(&client_lock);
|
2010-06-15 07:01:57 +08:00
|
|
|
return NULL;
|
2010-06-07 06:12:14 +08:00
|
|
|
}
|
2010-06-16 05:25:45 +08:00
|
|
|
if (cses->flags & SESSION4_BACK_CHAN) {
|
2010-06-15 10:26:31 +08:00
|
|
|
struct sockaddr *sa = svc_addr(rqstp);
|
2010-06-16 05:25:45 +08:00
|
|
|
/*
|
|
|
|
* This is a little silly; with sessions there's no real
|
|
|
|
* use for the callback address. Use the peer address
|
|
|
|
* as a reasonable default for now, but consider fixing
|
|
|
|
* the rpc client not to require an address in the
|
|
|
|
* future:
|
|
|
|
*/
|
2010-06-15 10:26:31 +08:00
|
|
|
rpc_copy_addr((struct sockaddr *)&clp->cl_cb_conn.cb_addr, sa);
|
|
|
|
clp->cl_cb_conn.cb_addrlen = svc_addr_len(sa);
|
|
|
|
}
|
2010-06-16 05:25:45 +08:00
|
|
|
nfsd4_probe_callback(clp);
|
2010-06-15 07:01:57 +08:00
|
|
|
return new;
|
2009-04-03 13:28:28 +08:00
|
|
|
}
|
|
|
|
|
2010-05-12 05:12:26 +08:00
|
|
|
/* caller must hold client_lock */
|
2009-04-03 13:27:52 +08:00
|
|
|
static struct nfsd4_session *
|
|
|
|
find_in_sessionid_hashtbl(struct nfs4_sessionid *sessionid)
|
|
|
|
{
|
|
|
|
struct nfsd4_session *elem;
|
|
|
|
int idx;
|
|
|
|
|
|
|
|
dump_sessionid(__func__, sessionid);
|
|
|
|
idx = hash_sessionid(sessionid);
|
|
|
|
/* Search in the appropriate list */
|
|
|
|
list_for_each_entry(elem, &sessionid_hashtbl[idx], se_hash) {
|
|
|
|
if (!memcmp(elem->se_sessionid.data, sessionid->data,
|
|
|
|
NFS4_MAX_SESSIONID_LEN)) {
|
|
|
|
return elem;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
dprintk("%s: session not found\n", __func__);
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
2010-05-12 05:12:26 +08:00
|
|
|
/* caller must hold client_lock */
|
2009-04-03 13:27:43 +08:00
|
|
|
static void
|
2009-04-03 13:27:52 +08:00
|
|
|
unhash_session(struct nfsd4_session *ses)
|
2009-04-03 13:27:43 +08:00
|
|
|
{
|
|
|
|
list_del(&ses->se_hash);
|
2010-06-16 02:22:37 +08:00
|
|
|
spin_lock(&ses->se_client->cl_lock);
|
2009-04-03 13:27:43 +08:00
|
|
|
list_del(&ses->se_perclnt);
|
2010-06-16 02:22:37 +08:00
|
|
|
spin_unlock(&ses->se_client->cl_lock);
|
2009-04-03 13:27:52 +08:00
|
|
|
}
|
|
|
|
|
2010-05-12 05:13:04 +08:00
|
|
|
/* must be called under the client_lock */
|
2005-04-17 06:20:36 +08:00
|
|
|
static inline void
|
2010-05-12 05:13:04 +08:00
|
|
|
renew_client_locked(struct nfs4_client *clp)
|
2005-04-17 06:20:36 +08:00
|
|
|
{
|
2010-05-12 05:13:41 +08:00
|
|
|
if (is_client_expired(clp)) {
|
|
|
|
dprintk("%s: client (clientid %08x/%08x) already expired\n",
|
|
|
|
__func__,
|
|
|
|
clp->cl_clientid.cl_boot,
|
|
|
|
clp->cl_clientid.cl_id);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
2005-04-17 06:20:36 +08:00
|
|
|
dprintk("renewing client (clientid %08x/%08x)\n",
|
|
|
|
clp->cl_clientid.cl_boot,
|
|
|
|
clp->cl_clientid.cl_id);
|
|
|
|
list_move_tail(&clp->cl_lru, &client_lru);
|
|
|
|
clp->cl_time = get_seconds();
|
|
|
|
}
|
|
|
|
|
2010-05-12 05:13:04 +08:00
|
|
|
static inline void
|
|
|
|
renew_client(struct nfs4_client *clp)
|
|
|
|
{
|
|
|
|
spin_lock(&client_lock);
|
|
|
|
renew_client_locked(clp);
|
|
|
|
spin_unlock(&client_lock);
|
|
|
|
}
|
|
|
|
|
2005-04-17 06:20:36 +08:00
|
|
|
/* SETCLIENTID and SETCLIENTID_CONFIRM Helper functions */
|
|
|
|
static int
|
|
|
|
STALE_CLIENTID(clientid_t *clid)
|
|
|
|
{
|
|
|
|
if (clid->cl_boot == boot_time)
|
|
|
|
return 0;
|
2009-04-03 13:28:50 +08:00
|
|
|
dprintk("NFSD stale clientid (%08x/%08x) boot_time %08lx\n",
|
|
|
|
clid->cl_boot, clid->cl_id, boot_time);
|
2005-04-17 06:20:36 +08:00
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* XXX Should we use a slab cache ?
|
|
|
|
* This type of memory management is somewhat inefficient, but we use it
|
|
|
|
* anyway since SETCLIENTID is not a common operation.
|
|
|
|
*/
|
2007-11-22 11:07:08 +08:00
|
|
|
static struct nfs4_client *alloc_client(struct xdr_netobj name)
|
2005-04-17 06:20:36 +08:00
|
|
|
{
|
|
|
|
struct nfs4_client *clp;
|
|
|
|
|
2007-11-22 11:07:08 +08:00
|
|
|
clp = kzalloc(sizeof(struct nfs4_client), GFP_KERNEL);
|
|
|
|
if (clp == NULL)
|
|
|
|
return NULL;
|
2011-11-18 06:43:40 +08:00
|
|
|
clp->cl_name.data = kmemdup(name.data, name.len, GFP_KERNEL);
|
2007-11-22 11:07:08 +08:00
|
|
|
if (clp->cl_name.data == NULL) {
|
|
|
|
kfree(clp);
|
|
|
|
return NULL;
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
2007-11-22 11:07:08 +08:00
|
|
|
clp->cl_name.len = name.len;
|
2005-04-17 06:20:36 +08:00
|
|
|
return clp;
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline void
|
|
|
|
free_client(struct nfs4_client *clp)
|
|
|
|
{
|
2012-02-24 09:40:52 +08:00
|
|
|
BUG_ON(!spin_is_locked(&client_lock));
|
2010-10-13 07:55:25 +08:00
|
|
|
while (!list_empty(&clp->cl_sessions)) {
|
|
|
|
struct nfsd4_session *ses;
|
|
|
|
ses = list_entry(clp->cl_sessions.next, struct nfsd4_session,
|
|
|
|
se_perclnt);
|
|
|
|
list_del(&ses->se_perclnt);
|
2012-02-24 09:40:52 +08:00
|
|
|
nfsd4_put_session_locked(ses);
|
2010-10-13 07:55:25 +08:00
|
|
|
}
|
2005-04-17 06:20:36 +08:00
|
|
|
if (clp->cl_cred.cr_group_info)
|
|
|
|
put_group_info(clp->cl_cred.cr_group_info);
|
2008-12-24 05:17:15 +08:00
|
|
|
kfree(clp->cl_principal);
|
2005-04-17 06:20:36 +08:00
|
|
|
kfree(clp->cl_name.data);
|
|
|
|
kfree(clp);
|
|
|
|
}
|
|
|
|
|
2010-05-12 05:13:54 +08:00
|
|
|
void
|
|
|
|
release_session_client(struct nfsd4_session *session)
|
|
|
|
{
|
|
|
|
struct nfs4_client *clp = session->se_client;
|
|
|
|
|
|
|
|
if (!atomic_dec_and_lock(&clp->cl_refcount, &client_lock))
|
|
|
|
return;
|
|
|
|
if (is_client_expired(clp)) {
|
|
|
|
free_client(clp);
|
|
|
|
session->se_client = NULL;
|
|
|
|
} else
|
|
|
|
renew_client_locked(clp);
|
|
|
|
spin_unlock(&client_lock);
|
|
|
|
}
|
|
|
|
|
2010-05-12 05:13:16 +08:00
|
|
|
/* must be called under the client_lock */
|
|
|
|
static inline void
|
|
|
|
unhash_client_locked(struct nfs4_client *clp)
|
|
|
|
{
|
2010-10-13 07:55:25 +08:00
|
|
|
struct nfsd4_session *ses;
|
|
|
|
|
2010-05-12 05:13:41 +08:00
|
|
|
mark_client_expired(clp);
|
2010-05-12 05:13:16 +08:00
|
|
|
list_del(&clp->cl_lru);
|
2010-06-16 02:22:37 +08:00
|
|
|
spin_lock(&clp->cl_lock);
|
2010-10-13 07:55:25 +08:00
|
|
|
list_for_each_entry(ses, &clp->cl_sessions, se_perclnt)
|
|
|
|
list_del_init(&ses->se_hash);
|
2010-06-16 02:22:37 +08:00
|
|
|
spin_unlock(&clp->cl_lock);
|
2010-05-12 05:13:16 +08:00
|
|
|
}
|
|
|
|
|
2005-04-17 06:20:36 +08:00
|
|
|
static void
|
|
|
|
expire_client(struct nfs4_client *clp)
|
|
|
|
{
|
2011-07-31 11:33:59 +08:00
|
|
|
struct nfs4_openowner *oo;
|
2005-04-17 06:20:36 +08:00
|
|
|
struct nfs4_delegation *dp;
|
|
|
|
struct list_head reaplist;
|
|
|
|
|
|
|
|
INIT_LIST_HEAD(&reaplist);
|
|
|
|
spin_lock(&recall_lock);
|
2005-06-24 13:04:17 +08:00
|
|
|
while (!list_empty(&clp->cl_delegations)) {
|
|
|
|
dp = list_entry(clp->cl_delegations.next, struct nfs4_delegation, dl_perclnt);
|
|
|
|
list_del_init(&dp->dl_perclnt);
|
2005-04-17 06:20:36 +08:00
|
|
|
list_move(&dp->dl_recall_lru, &reaplist);
|
|
|
|
}
|
|
|
|
spin_unlock(&recall_lock);
|
|
|
|
while (!list_empty(&reaplist)) {
|
|
|
|
dp = list_entry(reaplist.next, struct nfs4_delegation, dl_recall_lru);
|
|
|
|
unhash_delegation(dp);
|
|
|
|
}
|
2005-06-24 13:04:17 +08:00
|
|
|
while (!list_empty(&clp->cl_openowners)) {
|
2011-07-31 11:33:59 +08:00
|
|
|
oo = list_entry(clp->cl_openowners.next, struct nfs4_openowner, oo_perclient);
|
|
|
|
release_openowner(oo);
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
2010-06-05 08:04:45 +08:00
|
|
|
nfsd4_shutdown_callback(clp);
|
2010-05-12 05:13:16 +08:00
|
|
|
if (clp->cl_cb_conn.cb_xprt)
|
|
|
|
svc_xprt_put(clp->cl_cb_conn.cb_xprt);
|
2010-05-12 05:13:04 +08:00
|
|
|
list_del(&clp->cl_idhash);
|
|
|
|
list_del(&clp->cl_strhash);
|
2010-05-12 05:12:39 +08:00
|
|
|
spin_lock(&client_lock);
|
2010-05-12 05:13:16 +08:00
|
|
|
unhash_client_locked(clp);
|
2010-05-12 05:13:29 +08:00
|
|
|
if (atomic_read(&clp->cl_refcount) == 0)
|
|
|
|
free_client(clp);
|
2010-05-12 05:12:39 +08:00
|
|
|
spin_unlock(&client_lock);
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
|
|
|
|
2007-11-22 11:07:08 +08:00
|
|
|
static void copy_verf(struct nfs4_client *target, nfs4_verifier *source)
|
|
|
|
{
|
|
|
|
memcpy(target->cl_verifier.data, source->data,
|
|
|
|
sizeof(target->cl_verifier.data));
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
|
|
|
|
2007-11-22 11:07:08 +08:00
|
|
|
static void copy_clid(struct nfs4_client *target, struct nfs4_client *source)
|
|
|
|
{
|
2005-04-17 06:20:36 +08:00
|
|
|
target->cl_clientid.cl_boot = source->cl_clientid.cl_boot;
|
|
|
|
target->cl_clientid.cl_id = source->cl_clientid.cl_id;
|
|
|
|
}
|
|
|
|
|
2007-11-22 11:07:08 +08:00
|
|
|
static void copy_cred(struct svc_cred *target, struct svc_cred *source)
|
|
|
|
{
|
2005-04-17 06:20:36 +08:00
|
|
|
target->cr_uid = source->cr_uid;
|
|
|
|
target->cr_gid = source->cr_gid;
|
|
|
|
target->cr_group_info = source->cr_group_info;
|
|
|
|
get_group_info(target->cr_group_info);
|
|
|
|
}
|
|
|
|
|
2007-11-22 11:07:08 +08:00
|
|
|
static int same_name(const char *n1, const char *n2)
|
2007-07-27 05:04:54 +08:00
|
|
|
{
|
2005-06-24 13:03:52 +08:00
|
|
|
return 0 == memcmp(n1, n2, HEXDIR_LEN);
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
static int
|
2007-07-27 05:04:54 +08:00
|
|
|
same_verf(nfs4_verifier *v1, nfs4_verifier *v2)
|
|
|
|
{
|
|
|
|
return 0 == memcmp(v1->data, v2->data, sizeof(v1->data));
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
static int
|
2007-07-27 05:04:54 +08:00
|
|
|
same_clid(clientid_t *cl1, clientid_t *cl2)
|
|
|
|
{
|
|
|
|
return (cl1->cl_boot == cl2->cl_boot) && (cl1->cl_id == cl2->cl_id);
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
/* XXX what about NGROUP */
|
|
|
|
static int
|
2007-07-27 05:04:54 +08:00
|
|
|
same_creds(struct svc_cred *cr1, struct svc_cred *cr2)
|
|
|
|
{
|
|
|
|
return cr1->cr_uid == cr2->cr_uid;
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
|
|
|
|
2007-11-22 10:58:56 +08:00
|
|
|
static void gen_clid(struct nfs4_client *clp)
|
|
|
|
{
|
|
|
|
static u32 current_clientid = 1;
|
|
|
|
|
2005-04-17 06:20:36 +08:00
|
|
|
clp->cl_clientid.cl_boot = boot_time;
|
|
|
|
clp->cl_clientid.cl_id = current_clientid++;
|
|
|
|
}
|
|
|
|
|
2007-11-20 09:31:04 +08:00
|
|
|
static void gen_confirm(struct nfs4_client *clp)
|
|
|
|
{
|
2012-03-03 06:13:50 +08:00
|
|
|
__be32 verf[2];
|
2007-11-20 09:31:04 +08:00
|
|
|
static u32 i;
|
2005-04-17 06:20:36 +08:00
|
|
|
|
2012-03-03 06:13:50 +08:00
|
|
|
verf[0] = (__be32)get_seconds();
|
|
|
|
verf[1] = (__be32)i++;
|
|
|
|
memcpy(clp->cl_confirm.data, verf, sizeof(clp->cl_confirm.data));
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
|
|
|
|
2011-09-24 05:01:19 +08:00
|
|
|
static struct nfs4_stid *find_stateid(struct nfs4_client *cl, stateid_t *t)
|
2011-09-07 02:56:09 +08:00
|
|
|
{
|
2011-09-24 05:01:19 +08:00
|
|
|
return idr_find(&cl->cl_stateids, t->si_opaque.so_id);
|
2011-09-07 04:48:57 +08:00
|
|
|
}
|
|
|
|
|
2011-09-24 05:01:19 +08:00
|
|
|
static struct nfs4_stid *find_stateid_by_type(struct nfs4_client *cl, stateid_t *t, char typemask)
|
2011-09-09 21:06:12 +08:00
|
|
|
{
|
|
|
|
struct nfs4_stid *s;
|
2011-09-07 04:48:57 +08:00
|
|
|
|
2011-09-24 05:01:19 +08:00
|
|
|
s = find_stateid(cl, t);
|
2011-09-07 04:48:57 +08:00
|
|
|
if (!s)
|
|
|
|
return NULL;
|
2011-09-09 21:06:12 +08:00
|
|
|
if (typemask & s->sc_type)
|
2011-09-07 02:56:09 +08:00
|
|
|
return s;
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
2009-09-10 17:27:34 +08:00
|
|
|
static struct nfs4_client *create_client(struct xdr_netobj name, char *recdir,
|
|
|
|
struct svc_rqst *rqstp, nfs4_verifier *verf)
|
|
|
|
{
|
|
|
|
struct nfs4_client *clp;
|
|
|
|
struct sockaddr *sa = svc_addr(rqstp);
|
|
|
|
char *princ;
|
|
|
|
|
|
|
|
clp = alloc_client(name);
|
|
|
|
if (clp == NULL)
|
|
|
|
return NULL;
|
|
|
|
|
2010-10-13 07:55:25 +08:00
|
|
|
INIT_LIST_HEAD(&clp->cl_sessions);
|
|
|
|
|
2009-09-10 17:27:34 +08:00
|
|
|
princ = svc_gss_principal(rqstp);
|
|
|
|
if (princ) {
|
|
|
|
clp->cl_principal = kstrdup(princ, GFP_KERNEL);
|
|
|
|
if (clp->cl_principal == NULL) {
|
2012-02-24 09:40:52 +08:00
|
|
|
spin_lock(&client_lock);
|
2009-09-10 17:27:34 +08:00
|
|
|
free_client(clp);
|
2012-02-24 09:40:52 +08:00
|
|
|
spin_unlock(&client_lock);
|
2009-09-10 17:27:34 +08:00
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2011-09-24 05:01:19 +08:00
|
|
|
idr_init(&clp->cl_stateids);
|
2009-09-10 17:27:34 +08:00
|
|
|
memcpy(clp->cl_recdir, recdir, HEXDIR_LEN);
|
2010-05-12 05:13:29 +08:00
|
|
|
atomic_set(&clp->cl_refcount, 0);
|
2010-05-01 06:51:44 +08:00
|
|
|
clp->cl_cb_state = NFSD4_CB_UNKNOWN;
|
2009-09-10 17:27:34 +08:00
|
|
|
INIT_LIST_HEAD(&clp->cl_idhash);
|
|
|
|
INIT_LIST_HEAD(&clp->cl_strhash);
|
|
|
|
INIT_LIST_HEAD(&clp->cl_openowners);
|
|
|
|
INIT_LIST_HEAD(&clp->cl_delegations);
|
|
|
|
INIT_LIST_HEAD(&clp->cl_lru);
|
2011-01-11 05:44:41 +08:00
|
|
|
INIT_LIST_HEAD(&clp->cl_callbacks);
|
2010-06-05 08:04:45 +08:00
|
|
|
spin_lock_init(&clp->cl_lock);
|
2010-05-27 05:52:14 +08:00
|
|
|
INIT_WORK(&clp->cl_cb_null.cb_work, nfsd4_do_callback_rpc);
|
2010-05-12 05:13:41 +08:00
|
|
|
clp->cl_time = get_seconds();
|
2009-09-10 17:27:34 +08:00
|
|
|
clear_bit(0, &clp->cl_cb_slot_busy);
|
|
|
|
rpc_init_wait_queue(&clp->cl_cb_waitq, "Backchannel slot table");
|
|
|
|
copy_verf(clp, verf);
|
|
|
|
rpc_copy_addr((struct sockaddr *) &clp->cl_addr, sa);
|
|
|
|
clp->cl_flavor = rqstp->rq_flavor;
|
|
|
|
copy_cred(&clp->cl_cred, &rqstp->rq_cred);
|
|
|
|
gen_confirm(clp);
|
2010-06-15 10:26:31 +08:00
|
|
|
clp->cl_cb_session = NULL;
|
2009-09-10 17:27:34 +08:00
|
|
|
return clp;
|
|
|
|
}
|
|
|
|
|
2005-06-24 13:04:03 +08:00
|
|
|
static void
|
2005-04-17 06:20:36 +08:00
|
|
|
add_to_unconfirmed(struct nfs4_client *clp, unsigned int strhashval)
|
|
|
|
{
|
|
|
|
unsigned int idhashval;
|
|
|
|
|
|
|
|
list_add(&clp->cl_strhash, &unconf_str_hashtbl[strhashval]);
|
|
|
|
idhashval = clientid_hashval(clp->cl_clientid.cl_id);
|
|
|
|
list_add(&clp->cl_idhash, &unconf_id_hashtbl[idhashval]);
|
2010-05-12 05:13:04 +08:00
|
|
|
renew_client(clp);
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
|
|
|
|
2005-06-24 13:04:03 +08:00
|
|
|
static void
|
2005-04-17 06:20:36 +08:00
|
|
|
move_to_confirmed(struct nfs4_client *clp)
|
|
|
|
{
|
|
|
|
unsigned int idhashval = clientid_hashval(clp->cl_clientid.cl_id);
|
|
|
|
unsigned int strhashval;
|
|
|
|
|
|
|
|
dprintk("NFSD: move_to_confirm nfs4_client %p\n", clp);
|
2006-06-26 15:24:46 +08:00
|
|
|
list_move(&clp->cl_idhash, &conf_id_hashtbl[idhashval]);
|
2005-06-24 13:03:52 +08:00
|
|
|
strhashval = clientstr_hashval(clp->cl_recdir);
|
2010-05-12 05:12:51 +08:00
|
|
|
list_move(&clp->cl_strhash, &conf_str_hashtbl[strhashval]);
|
2005-04-17 06:20:36 +08:00
|
|
|
renew_client(clp);
|
|
|
|
}
|
|
|
|
|
|
|
|
static struct nfs4_client *
|
|
|
|
find_confirmed_client(clientid_t *clid)
|
|
|
|
{
|
|
|
|
struct nfs4_client *clp;
|
|
|
|
unsigned int idhashval = clientid_hashval(clid->cl_id);
|
|
|
|
|
|
|
|
list_for_each_entry(clp, &conf_id_hashtbl[idhashval], cl_idhash) {
|
2011-10-13 04:24:27 +08:00
|
|
|
if (same_clid(&clp->cl_clientid, clid)) {
|
|
|
|
renew_client(clp);
|
2005-04-17 06:20:36 +08:00
|
|
|
return clp;
|
2011-10-13 04:24:27 +08:00
|
|
|
}
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
|
|
|
static struct nfs4_client *
|
|
|
|
find_unconfirmed_client(clientid_t *clid)
|
|
|
|
{
|
|
|
|
struct nfs4_client *clp;
|
|
|
|
unsigned int idhashval = clientid_hashval(clid->cl_id);
|
|
|
|
|
|
|
|
list_for_each_entry(clp, &unconf_id_hashtbl[idhashval], cl_idhash) {
|
2007-07-27 05:04:54 +08:00
|
|
|
if (same_clid(&clp->cl_clientid, clid))
|
2005-04-17 06:20:36 +08:00
|
|
|
return clp;
|
|
|
|
}
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
2010-11-25 06:17:34 +08:00
|
|
|
static bool clp_used_exchangeid(struct nfs4_client *clp)
|
2009-04-03 13:28:05 +08:00
|
|
|
{
|
2010-11-25 06:17:34 +08:00
|
|
|
return clp->cl_exchange_flags != 0;
|
2010-11-25 06:30:54 +08:00
|
|
|
}
|
2009-04-03 13:28:05 +08:00
|
|
|
|
2005-06-24 13:03:56 +08:00
|
|
|
static struct nfs4_client *
|
2010-11-25 06:30:54 +08:00
|
|
|
find_confirmed_client_by_str(const char *dname, unsigned int hashval)
|
2005-06-24 13:03:56 +08:00
|
|
|
{
|
|
|
|
struct nfs4_client *clp;
|
|
|
|
|
|
|
|
list_for_each_entry(clp, &conf_str_hashtbl[hashval], cl_strhash) {
|
2010-11-25 06:30:54 +08:00
|
|
|
if (same_name(clp->cl_recdir, dname))
|
2005-06-24 13:03:56 +08:00
|
|
|
return clp;
|
|
|
|
}
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
|
|
|
static struct nfs4_client *
|
2010-11-25 06:30:54 +08:00
|
|
|
find_unconfirmed_client_by_str(const char *dname, unsigned int hashval)
|
2005-06-24 13:03:56 +08:00
|
|
|
{
|
|
|
|
struct nfs4_client *clp;
|
|
|
|
|
|
|
|
list_for_each_entry(clp, &unconf_str_hashtbl[hashval], cl_strhash) {
|
2010-11-25 06:30:54 +08:00
|
|
|
if (same_name(clp->cl_recdir, dname))
|
2005-06-24 13:03:56 +08:00
|
|
|
return clp;
|
|
|
|
}
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
2005-06-24 13:04:03 +08:00
|
|
|
static void
|
2010-12-15 13:09:01 +08:00
|
|
|
gen_callback(struct nfs4_client *clp, struct nfsd4_setclientid *se, struct svc_rqst *rqstp)
|
2005-04-17 06:20:36 +08:00
|
|
|
{
|
2010-06-01 07:09:40 +08:00
|
|
|
struct nfs4_cb_conn *conn = &clp->cl_cb_conn;
|
2010-12-15 13:09:01 +08:00
|
|
|
struct sockaddr *sa = svc_addr(rqstp);
|
|
|
|
u32 scopeid = rpc_get_scope_id(sa);
|
2009-08-15 00:57:58 +08:00
|
|
|
unsigned short expected_family;
|
|
|
|
|
|
|
|
/* Currently, we only support tcp and tcp6 for the callback channel */
|
|
|
|
if (se->se_callback_netid_len == 3 &&
|
|
|
|
!memcmp(se->se_callback_netid_val, "tcp", 3))
|
|
|
|
expected_family = AF_INET;
|
|
|
|
else if (se->se_callback_netid_len == 4 &&
|
|
|
|
!memcmp(se->se_callback_netid_val, "tcp6", 4))
|
|
|
|
expected_family = AF_INET6;
|
|
|
|
else
|
2005-04-17 06:20:36 +08:00
|
|
|
goto out_err;
|
|
|
|
|
2012-01-13 17:09:27 +08:00
|
|
|
conn->cb_addrlen = rpc_uaddr2sockaddr(&init_net, se->se_callback_addr_val,
|
2009-08-15 00:57:57 +08:00
|
|
|
se->se_callback_addr_len,
|
2010-06-01 07:09:40 +08:00
|
|
|
(struct sockaddr *)&conn->cb_addr,
|
|
|
|
sizeof(conn->cb_addr));
|
2009-08-15 00:57:57 +08:00
|
|
|
|
2010-06-01 07:09:40 +08:00
|
|
|
if (!conn->cb_addrlen || conn->cb_addr.ss_family != expected_family)
|
2005-04-17 06:20:36 +08:00
|
|
|
goto out_err;
|
2009-08-15 00:57:57 +08:00
|
|
|
|
2010-06-01 07:09:40 +08:00
|
|
|
if (conn->cb_addr.ss_family == AF_INET6)
|
|
|
|
((struct sockaddr_in6 *)&conn->cb_addr)->sin6_scope_id = scopeid;
|
2009-08-15 00:57:59 +08:00
|
|
|
|
2010-06-01 07:09:40 +08:00
|
|
|
conn->cb_prog = se->se_callback_prog;
|
|
|
|
conn->cb_ident = se->se_callback_ident;
|
2011-08-30 17:18:41 +08:00
|
|
|
memcpy(&conn->cb_saddr, &rqstp->rq_daddr, rqstp->rq_daddrlen);
|
2005-04-17 06:20:36 +08:00
|
|
|
return;
|
|
|
|
out_err:
|
2010-06-01 07:09:40 +08:00
|
|
|
conn->cb_addr.ss_family = AF_UNSPEC;
|
|
|
|
conn->cb_addrlen = 0;
|
2005-09-13 16:25:36 +08:00
|
|
|
dprintk(KERN_INFO "NFSD: this client (clientid %08x/%08x) "
|
2005-04-17 06:20:36 +08:00
|
|
|
"will not receive delegations\n",
|
|
|
|
clp->cl_clientid.cl_boot, clp->cl_clientid.cl_id);
|
|
|
|
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
2009-04-03 13:28:15 +08:00
|
|
|
/*
|
2009-08-28 20:45:04 +08:00
|
|
|
* Cache a reply. nfsd4_check_drc_limit() has bounded the cache size.
|
2009-04-03 13:28:15 +08:00
|
|
|
*/
|
|
|
|
void
|
|
|
|
nfsd4_store_cache_entry(struct nfsd4_compoundres *resp)
|
|
|
|
{
|
2009-08-28 20:45:04 +08:00
|
|
|
struct nfsd4_slot *slot = resp->cstate.slot;
|
|
|
|
unsigned int base;
|
2009-04-03 13:28:15 +08:00
|
|
|
|
2009-08-28 20:45:04 +08:00
|
|
|
dprintk("--> %s slot %p\n", __func__, slot);
|
2009-04-03 13:28:15 +08:00
|
|
|
|
2009-08-28 20:45:04 +08:00
|
|
|
slot->sl_opcnt = resp->opcnt;
|
|
|
|
slot->sl_status = resp->cstate.status;
|
2009-04-03 13:28:15 +08:00
|
|
|
|
2012-02-14 05:56:19 +08:00
|
|
|
slot->sl_flags |= NFSD4_SLOT_INITIALIZED;
|
2009-04-03 13:28:35 +08:00
|
|
|
if (nfsd4_not_cached(resp)) {
|
2009-08-28 20:45:04 +08:00
|
|
|
slot->sl_datalen = 0;
|
2009-04-03 13:28:35 +08:00
|
|
|
return;
|
2009-04-03 13:28:15 +08:00
|
|
|
}
|
2009-08-28 20:45:04 +08:00
|
|
|
slot->sl_datalen = (char *)resp->p - (char *)resp->cstate.datap;
|
|
|
|
base = (char *)resp->cstate.datap -
|
|
|
|
(char *)resp->xbuf->head[0].iov_base;
|
|
|
|
if (read_bytes_from_xdr_buf(resp->xbuf, base, slot->sl_data,
|
|
|
|
slot->sl_datalen))
|
|
|
|
WARN("%s: sessions DRC could not cache compound\n", __func__);
|
|
|
|
return;
|
2009-04-03 13:28:15 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
2009-07-24 07:02:18 +08:00
|
|
|
* Encode the replay sequence operation from the slot values.
|
|
|
|
* If cachethis is FALSE encode the uncached rep error on the next
|
|
|
|
* operation which sets resp->p and increments resp->opcnt for
|
|
|
|
* nfs4svc_encode_compoundres.
|
2009-04-03 13:28:15 +08:00
|
|
|
*
|
|
|
|
*/
|
2009-07-24 07:02:18 +08:00
|
|
|
static __be32
|
|
|
|
nfsd4_enc_sequence_replay(struct nfsd4_compoundargs *args,
|
|
|
|
struct nfsd4_compoundres *resp)
|
2009-04-03 13:28:15 +08:00
|
|
|
{
|
2009-07-24 07:02:18 +08:00
|
|
|
struct nfsd4_op *op;
|
|
|
|
struct nfsd4_slot *slot = resp->cstate.slot;
|
2009-04-03 13:28:35 +08:00
|
|
|
|
2009-07-24 07:02:18 +08:00
|
|
|
/* Encode the replayed sequence operation */
|
|
|
|
op = &args->ops[resp->opcnt - 1];
|
|
|
|
nfsd4_encode_operation(resp, op);
|
2009-04-03 13:28:35 +08:00
|
|
|
|
2009-07-24 07:02:18 +08:00
|
|
|
/* Return nfserr_retry_uncached_rep in next operation. */
|
2012-02-14 05:39:00 +08:00
|
|
|
if (args->opcnt > 1 && !(slot->sl_flags & NFSD4_SLOT_CACHETHIS)) {
|
2009-07-24 07:02:18 +08:00
|
|
|
op = &args->ops[resp->opcnt++];
|
|
|
|
op->status = nfserr_retry_uncached_rep;
|
|
|
|
nfsd4_encode_operation(resp, op);
|
2009-04-03 13:28:15 +08:00
|
|
|
}
|
2009-07-24 07:02:18 +08:00
|
|
|
return op->status;
|
2009-04-03 13:28:15 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
2009-08-28 20:45:04 +08:00
|
|
|
* The sequence operation is not cached because we can use the slot and
|
|
|
|
* session values.
|
2009-04-03 13:28:15 +08:00
|
|
|
*/
|
|
|
|
__be32
|
2009-04-03 13:28:35 +08:00
|
|
|
nfsd4_replay_cache_entry(struct nfsd4_compoundres *resp,
|
|
|
|
struct nfsd4_sequence *seq)
|
2009-04-03 13:28:15 +08:00
|
|
|
{
|
2009-08-28 20:45:04 +08:00
|
|
|
struct nfsd4_slot *slot = resp->cstate.slot;
|
2009-04-03 13:28:15 +08:00
|
|
|
__be32 status;
|
|
|
|
|
2009-08-28 20:45:04 +08:00
|
|
|
dprintk("--> %s slot %p\n", __func__, slot);
|
2009-04-03 13:28:15 +08:00
|
|
|
|
2009-07-24 07:02:18 +08:00
|
|
|
/* Either returns 0 or nfserr_retry_uncached */
|
|
|
|
status = nfsd4_enc_sequence_replay(resp->rqstp->rq_argp, resp);
|
|
|
|
if (status == nfserr_retry_uncached_rep)
|
|
|
|
return status;
|
2009-04-03 13:28:15 +08:00
|
|
|
|
2009-08-28 20:45:04 +08:00
|
|
|
/* The sequence operation has been encoded, cstate->datap set. */
|
|
|
|
memcpy(resp->cstate.datap, slot->sl_data, slot->sl_datalen);
|
2009-04-03 13:28:15 +08:00
|
|
|
|
2009-08-28 20:45:04 +08:00
|
|
|
resp->opcnt = slot->sl_opcnt;
|
|
|
|
resp->p = resp->cstate.datap + XDR_QUADLEN(slot->sl_datalen);
|
|
|
|
status = slot->sl_status;
|
2009-04-03 13:28:15 +08:00
|
|
|
|
|
|
|
return status;
|
|
|
|
}
|
|
|
|
|
2009-04-03 13:28:01 +08:00
|
|
|
/*
|
|
|
|
* Set the exchange_id flags returned by the server.
|
|
|
|
*/
|
|
|
|
static void
|
|
|
|
nfsd4_set_ex_flags(struct nfs4_client *new, struct nfsd4_exchange_id *clid)
|
|
|
|
{
|
|
|
|
/* pNFS is not supported */
|
|
|
|
new->cl_exchange_flags |= EXCHGID4_FLAG_USE_NON_PNFS;
|
|
|
|
|
|
|
|
/* Referrals are supported, Migration is not. */
|
|
|
|
new->cl_exchange_flags |= EXCHGID4_FLAG_SUPP_MOVED_REFER;
|
|
|
|
|
|
|
|
/* set the wire flags to return to client. */
|
|
|
|
clid->flags = new->cl_exchange_flags;
|
|
|
|
}
|
|
|
|
|
2009-04-03 13:27:58 +08:00
|
|
|
__be32
|
|
|
|
nfsd4_exchange_id(struct svc_rqst *rqstp,
|
|
|
|
struct nfsd4_compound_state *cstate,
|
|
|
|
struct nfsd4_exchange_id *exid)
|
|
|
|
{
|
2009-04-03 13:28:01 +08:00
|
|
|
struct nfs4_client *unconf, *conf, *new;
|
|
|
|
int status;
|
|
|
|
unsigned int strhashval;
|
|
|
|
char dname[HEXDIR_LEN];
|
2009-08-15 00:57:56 +08:00
|
|
|
char addr_str[INET6_ADDRSTRLEN];
|
2009-04-03 13:28:01 +08:00
|
|
|
nfs4_verifier verf = exid->verifier;
|
2009-08-15 00:57:56 +08:00
|
|
|
struct sockaddr *sa = svc_addr(rqstp);
|
2009-04-03 13:28:01 +08:00
|
|
|
|
2009-08-15 00:57:56 +08:00
|
|
|
rpc_ntop(sa, addr_str, sizeof(addr_str));
|
2009-04-03 13:28:01 +08:00
|
|
|
dprintk("%s rqstp=%p exid=%p clname.len=%u clname.data=%p "
|
2009-08-15 00:57:56 +08:00
|
|
|
"ip_addr=%s flags %x, spa_how %d\n",
|
2009-04-03 13:28:01 +08:00
|
|
|
__func__, rqstp, exid, exid->clname.len, exid->clname.data,
|
2009-08-15 00:57:56 +08:00
|
|
|
addr_str, exid->flags, exid->spa_how);
|
2009-04-03 13:28:01 +08:00
|
|
|
|
2011-10-11 03:07:40 +08:00
|
|
|
if (exid->flags & ~EXCHGID4_FLAG_MASK_A)
|
2009-04-03 13:28:01 +08:00
|
|
|
return nfserr_inval;
|
|
|
|
|
|
|
|
/* Currently only support SP4_NONE */
|
|
|
|
switch (exid->spa_how) {
|
|
|
|
case SP4_NONE:
|
|
|
|
break;
|
|
|
|
case SP4_SSV:
|
2010-11-13 03:36:06 +08:00
|
|
|
return nfserr_serverfault;
|
2009-04-03 13:28:01 +08:00
|
|
|
default:
|
|
|
|
BUG(); /* checked by xdr code */
|
|
|
|
case SP4_MACH_CRED:
|
|
|
|
return nfserr_serverfault; /* no excuse :-/ */
|
|
|
|
}
|
|
|
|
|
|
|
|
status = nfs4_make_rec_clidname(dname, &exid->clname);
|
|
|
|
|
|
|
|
if (status)
|
|
|
|
goto error;
|
|
|
|
|
|
|
|
strhashval = clientstr_hashval(dname);
|
|
|
|
|
|
|
|
nfs4_lock_state();
|
|
|
|
status = nfs_ok;
|
|
|
|
|
2010-11-25 06:30:54 +08:00
|
|
|
conf = find_confirmed_client_by_str(dname, strhashval);
|
2009-04-03 13:28:01 +08:00
|
|
|
if (conf) {
|
2010-11-25 06:30:54 +08:00
|
|
|
if (!clp_used_exchangeid(conf)) {
|
|
|
|
status = nfserr_clid_inuse; /* XXX: ? */
|
|
|
|
goto out;
|
|
|
|
}
|
2009-04-03 13:28:01 +08:00
|
|
|
if (!same_verf(&verf, &conf->cl_verifier)) {
|
|
|
|
/* 18.35.4 case 8 */
|
|
|
|
if (exid->flags & EXCHGID4_FLAG_UPD_CONFIRMED_REC_A) {
|
|
|
|
status = nfserr_not_same;
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
/* Client reboot: destroy old state */
|
|
|
|
expire_client(conf);
|
|
|
|
goto out_new;
|
|
|
|
}
|
|
|
|
if (!same_creds(&conf->cl_cred, &rqstp->rq_cred)) {
|
|
|
|
/* 18.35.4 case 9 */
|
|
|
|
if (exid->flags & EXCHGID4_FLAG_UPD_CONFIRMED_REC_A) {
|
|
|
|
status = nfserr_perm;
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
expire_client(conf);
|
|
|
|
goto out_new;
|
|
|
|
}
|
|
|
|
/*
|
|
|
|
* Set bit when the owner id and verifier map to an already
|
|
|
|
* confirmed client id (18.35.3).
|
|
|
|
*/
|
|
|
|
exid->flags |= EXCHGID4_FLAG_CONFIRMED_R;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Falling into 18.35.4 case 2, possible router replay.
|
|
|
|
* Leave confirmed record intact and return same result.
|
|
|
|
*/
|
|
|
|
copy_verf(conf, &verf);
|
|
|
|
new = conf;
|
|
|
|
goto out_copy;
|
2009-06-16 09:20:47 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
/* 18.35.4 case 7 */
|
|
|
|
if (exid->flags & EXCHGID4_FLAG_UPD_CONFIRMED_REC_A) {
|
|
|
|
status = nfserr_noent;
|
|
|
|
goto out;
|
2009-04-03 13:28:01 +08:00
|
|
|
}
|
|
|
|
|
2010-11-25 06:30:54 +08:00
|
|
|
unconf = find_unconfirmed_client_by_str(dname, strhashval);
|
2009-04-03 13:28:01 +08:00
|
|
|
if (unconf) {
|
|
|
|
/*
|
|
|
|
* Possible retry or client restart. Per 18.35.4 case 4,
|
|
|
|
* a new unconfirmed record should be generated regardless
|
|
|
|
* of whether any properties have changed.
|
|
|
|
*/
|
|
|
|
expire_client(unconf);
|
|
|
|
}
|
|
|
|
|
|
|
|
out_new:
|
|
|
|
/* Normal case */
|
2009-09-10 17:27:34 +08:00
|
|
|
new = create_client(exid->clname, dname, rqstp, &verf);
|
2009-04-03 13:28:01 +08:00
|
|
|
if (new == NULL) {
|
2010-06-23 04:17:12 +08:00
|
|
|
status = nfserr_jukebox;
|
2009-04-03 13:28:01 +08:00
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
|
|
|
|
gen_clid(new);
|
|
|
|
add_to_unconfirmed(new, strhashval);
|
|
|
|
out_copy:
|
|
|
|
exid->clientid.cl_boot = new->cl_clientid.cl_boot;
|
|
|
|
exid->clientid.cl_id = new->cl_clientid.cl_id;
|
|
|
|
|
2009-04-03 13:28:32 +08:00
|
|
|
exid->seqid = 1;
|
2009-04-03 13:28:01 +08:00
|
|
|
nfsd4_set_ex_flags(new, exid);
|
|
|
|
|
|
|
|
dprintk("nfsd4_exchange_id seqid %d flags %x\n",
|
2009-07-24 07:02:16 +08:00
|
|
|
new->cl_cs_slot.sl_seqid, new->cl_exchange_flags);
|
2009-04-03 13:28:01 +08:00
|
|
|
status = nfs_ok;
|
|
|
|
|
|
|
|
out:
|
|
|
|
nfs4_unlock_state();
|
|
|
|
error:
|
|
|
|
dprintk("nfsd4_exchange_id returns %d\n", ntohl(status));
|
|
|
|
return status;
|
2009-04-03 13:27:58 +08:00
|
|
|
}
|
|
|
|
|
2009-04-03 13:28:08 +08:00
|
|
|
static int
|
2009-07-24 07:02:15 +08:00
|
|
|
check_slot_seqid(u32 seqid, u32 slot_seqid, int slot_inuse)
|
2009-04-03 13:28:08 +08:00
|
|
|
{
|
2009-07-24 07:02:15 +08:00
|
|
|
dprintk("%s enter. seqid %d slot_seqid %d\n", __func__, seqid,
|
|
|
|
slot_seqid);
|
2009-04-03 13:28:08 +08:00
|
|
|
|
|
|
|
/* The slot is in use, and no response has been sent. */
|
2009-07-24 07:02:15 +08:00
|
|
|
if (slot_inuse) {
|
|
|
|
if (seqid == slot_seqid)
|
2009-04-03 13:28:08 +08:00
|
|
|
return nfserr_jukebox;
|
|
|
|
else
|
|
|
|
return nfserr_seq_misordered;
|
|
|
|
}
|
2012-02-14 05:13:41 +08:00
|
|
|
/* Note unsigned 32-bit arithmetic handles wraparound: */
|
2009-07-24 07:02:15 +08:00
|
|
|
if (likely(seqid == slot_seqid + 1))
|
2009-04-03 13:28:08 +08:00
|
|
|
return nfs_ok;
|
2009-07-24 07:02:15 +08:00
|
|
|
if (seqid == slot_seqid)
|
2009-04-03 13:28:08 +08:00
|
|
|
return nfserr_replay_cache;
|
|
|
|
return nfserr_seq_misordered;
|
|
|
|
}
|
|
|
|
|
2009-07-24 07:02:16 +08:00
|
|
|
/*
|
|
|
|
* Cache the create session result into the create session single DRC
|
|
|
|
* slot cache by saving the xdr structure. sl_seqid has been set.
|
|
|
|
* Do this for solo or embedded create session operations.
|
|
|
|
*/
|
|
|
|
static void
|
|
|
|
nfsd4_cache_create_session(struct nfsd4_create_session *cr_ses,
|
|
|
|
struct nfsd4_clid_slot *slot, int nfserr)
|
|
|
|
{
|
|
|
|
slot->sl_status = nfserr;
|
|
|
|
memcpy(&slot->sl_cr_ses, cr_ses, sizeof(*cr_ses));
|
|
|
|
}
|
|
|
|
|
|
|
|
static __be32
|
|
|
|
nfsd4_replay_create_session(struct nfsd4_create_session *cr_ses,
|
|
|
|
struct nfsd4_clid_slot *slot)
|
|
|
|
{
|
|
|
|
memcpy(cr_ses, &slot->sl_cr_ses, sizeof(*cr_ses));
|
|
|
|
return slot->sl_status;
|
|
|
|
}
|
|
|
|
|
2011-07-14 14:50:17 +08:00
|
|
|
#define NFSD_MIN_REQ_HDR_SEQ_SZ ((\
|
|
|
|
2 * 2 + /* credential,verifier: AUTH_NULL, length 0 */ \
|
|
|
|
1 + /* MIN tag is length with zero, only length */ \
|
|
|
|
3 + /* version, opcount, opcode */ \
|
|
|
|
XDR_QUADLEN(NFS4_MAX_SESSIONID_LEN) + \
|
|
|
|
/* seqid, slotID, slotID, cache */ \
|
|
|
|
4 ) * sizeof(__be32))
|
|
|
|
|
|
|
|
#define NFSD_MIN_RESP_HDR_SEQ_SZ ((\
|
|
|
|
2 + /* verifier: AUTH_NULL, length 0 */\
|
|
|
|
1 + /* status */ \
|
|
|
|
1 + /* MIN tag is length with zero, only length */ \
|
|
|
|
3 + /* opcount, opcode, opstatus*/ \
|
|
|
|
XDR_QUADLEN(NFS4_MAX_SESSIONID_LEN) + \
|
|
|
|
/* seqid, slotID, slotID, slotID, status */ \
|
|
|
|
5 ) * sizeof(__be32))
|
|
|
|
|
|
|
|
static __be32 check_forechannel_attrs(struct nfsd4_channel_attrs fchannel)
|
|
|
|
{
|
|
|
|
return fchannel.maxreq_sz < NFSD_MIN_REQ_HDR_SEQ_SZ
|
|
|
|
|| fchannel.maxresp_sz < NFSD_MIN_RESP_HDR_SEQ_SZ;
|
|
|
|
}
|
|
|
|
|
2009-04-03 13:27:58 +08:00
|
|
|
__be32
|
|
|
|
nfsd4_create_session(struct svc_rqst *rqstp,
|
|
|
|
struct nfsd4_compound_state *cstate,
|
|
|
|
struct nfsd4_create_session *cr_ses)
|
|
|
|
{
|
2009-08-15 00:57:56 +08:00
|
|
|
struct sockaddr *sa = svc_addr(rqstp);
|
2009-04-03 13:28:28 +08:00
|
|
|
struct nfs4_client *conf, *unconf;
|
2010-06-15 07:01:57 +08:00
|
|
|
struct nfsd4_session *new;
|
2009-07-24 07:02:16 +08:00
|
|
|
struct nfsd4_clid_slot *cs_slot = NULL;
|
2010-10-03 05:04:00 +08:00
|
|
|
bool confirm_me = false;
|
2009-04-03 13:28:28 +08:00
|
|
|
int status = 0;
|
|
|
|
|
2011-03-23 17:57:07 +08:00
|
|
|
if (cr_ses->flags & ~SESSION4_FLAG_MASK_A)
|
|
|
|
return nfserr_inval;
|
|
|
|
|
2009-04-03 13:28:28 +08:00
|
|
|
nfs4_lock_state();
|
|
|
|
unconf = find_unconfirmed_client(&cr_ses->clientid);
|
|
|
|
conf = find_confirmed_client(&cr_ses->clientid);
|
|
|
|
|
|
|
|
if (conf) {
|
2009-07-24 07:02:16 +08:00
|
|
|
cs_slot = &conf->cl_cs_slot;
|
|
|
|
status = check_slot_seqid(cr_ses->seqid, cs_slot->sl_seqid, 0);
|
2009-04-03 13:28:32 +08:00
|
|
|
if (status == nfserr_replay_cache) {
|
2009-04-03 13:28:28 +08:00
|
|
|
dprintk("Got a create_session replay! seqid= %d\n",
|
2009-07-24 07:02:16 +08:00
|
|
|
cs_slot->sl_seqid);
|
2009-04-03 13:28:32 +08:00
|
|
|
/* Return the cached reply status */
|
2009-07-24 07:02:16 +08:00
|
|
|
status = nfsd4_replay_create_session(cr_ses, cs_slot);
|
2009-04-03 13:28:32 +08:00
|
|
|
goto out;
|
2009-07-24 07:02:16 +08:00
|
|
|
} else if (cr_ses->seqid != cs_slot->sl_seqid + 1) {
|
2009-04-03 13:28:28 +08:00
|
|
|
status = nfserr_seq_misordered;
|
|
|
|
dprintk("Sequence misordered!\n");
|
|
|
|
dprintk("Expected seqid= %d but got seqid= %d\n",
|
2009-07-24 07:02:16 +08:00
|
|
|
cs_slot->sl_seqid, cr_ses->seqid);
|
2009-04-03 13:28:28 +08:00
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
} else if (unconf) {
|
|
|
|
if (!same_creds(&unconf->cl_cred, &rqstp->rq_cred) ||
|
2009-08-15 00:57:56 +08:00
|
|
|
!rpc_cmp_addr(sa, (struct sockaddr *) &unconf->cl_addr)) {
|
2009-04-03 13:28:28 +08:00
|
|
|
status = nfserr_clid_inuse;
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
|
2009-07-24 07:02:16 +08:00
|
|
|
cs_slot = &unconf->cl_cs_slot;
|
|
|
|
status = check_slot_seqid(cr_ses->seqid, cs_slot->sl_seqid, 0);
|
2009-04-03 13:28:32 +08:00
|
|
|
if (status) {
|
|
|
|
/* an unconfirmed replay returns misordered */
|
2009-04-03 13:28:28 +08:00
|
|
|
status = nfserr_seq_misordered;
|
2010-10-03 05:03:35 +08:00
|
|
|
goto out;
|
2009-04-03 13:28:28 +08:00
|
|
|
}
|
|
|
|
|
2010-10-03 05:04:00 +08:00
|
|
|
confirm_me = true;
|
2009-04-03 13:28:28 +08:00
|
|
|
conf = unconf;
|
|
|
|
} else {
|
|
|
|
status = nfserr_stale_clientid;
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
|
2010-10-20 07:36:51 +08:00
|
|
|
/*
|
|
|
|
* XXX: we should probably set this at creation time, and check
|
|
|
|
* for consistent minorversion use throughout:
|
|
|
|
*/
|
|
|
|
conf->cl_minorversion = 1;
|
2010-04-16 03:11:09 +08:00
|
|
|
/*
|
|
|
|
* We do not support RDMA or persistent sessions
|
|
|
|
*/
|
|
|
|
cr_ses->flags &= ~SESSION4_PERSIST;
|
|
|
|
cr_ses->flags &= ~SESSION4_RDMA;
|
|
|
|
|
2011-07-14 14:50:17 +08:00
|
|
|
status = nfserr_toosmall;
|
|
|
|
if (check_forechannel_attrs(cr_ses->fore_channel))
|
|
|
|
goto out;
|
|
|
|
|
2010-06-15 07:01:57 +08:00
|
|
|
status = nfserr_jukebox;
|
|
|
|
new = alloc_init_session(rqstp, conf, cr_ses);
|
|
|
|
if (!new)
|
2009-04-03 13:28:28 +08:00
|
|
|
goto out;
|
2010-06-15 07:01:57 +08:00
|
|
|
status = nfs_ok;
|
|
|
|
memcpy(cr_ses->sessionid.data, new->se_sessionid.data,
|
2009-04-03 13:28:28 +08:00
|
|
|
NFS4_MAX_SESSIONID_LEN);
|
2010-11-11 18:03:40 +08:00
|
|
|
memcpy(&cr_ses->fore_channel, &new->se_fchannel,
|
|
|
|
sizeof(struct nfsd4_channel_attrs));
|
2010-10-03 05:04:00 +08:00
|
|
|
cs_slot->sl_seqid++;
|
2009-07-24 07:02:16 +08:00
|
|
|
cr_ses->seqid = cs_slot->sl_seqid;
|
2009-04-03 13:28:28 +08:00
|
|
|
|
2009-07-24 07:02:16 +08:00
|
|
|
/* cache solo and embedded create sessions under the state lock */
|
|
|
|
nfsd4_cache_create_session(cr_ses, cs_slot, status);
|
2010-10-03 05:04:00 +08:00
|
|
|
if (confirm_me)
|
|
|
|
move_to_confirmed(conf);
|
2009-04-03 13:28:28 +08:00
|
|
|
out:
|
|
|
|
nfs4_unlock_state();
|
|
|
|
dprintk("%s returns %d\n", __func__, ntohl(status));
|
|
|
|
return status;
|
2009-04-03 13:27:58 +08:00
|
|
|
}
|
|
|
|
|
2010-04-22 00:27:19 +08:00
|
|
|
static bool nfsd4_last_compound_op(struct svc_rqst *rqstp)
|
|
|
|
{
|
|
|
|
struct nfsd4_compoundres *resp = rqstp->rq_resp;
|
|
|
|
struct nfsd4_compoundargs *argp = rqstp->rq_argp;
|
|
|
|
|
|
|
|
return argp->opcnt == resp->opcnt;
|
|
|
|
}
|
|
|
|
|
2010-10-05 11:12:59 +08:00
|
|
|
static __be32 nfsd4_map_bcts_dir(u32 *dir)
|
|
|
|
{
|
|
|
|
switch (*dir) {
|
|
|
|
case NFS4_CDFC4_FORE:
|
|
|
|
case NFS4_CDFC4_BACK:
|
|
|
|
return nfs_ok;
|
|
|
|
case NFS4_CDFC4_FORE_OR_BOTH:
|
|
|
|
case NFS4_CDFC4_BACK_OR_BOTH:
|
|
|
|
*dir = NFS4_CDFC4_BOTH;
|
|
|
|
return nfs_ok;
|
|
|
|
};
|
|
|
|
return nfserr_inval;
|
|
|
|
}
|
|
|
|
|
|
|
|
__be32 nfsd4_bind_conn_to_session(struct svc_rqst *rqstp,
|
|
|
|
struct nfsd4_compound_state *cstate,
|
|
|
|
struct nfsd4_bind_conn_to_session *bcts)
|
|
|
|
{
|
|
|
|
__be32 status;
|
|
|
|
|
|
|
|
if (!nfsd4_last_compound_op(rqstp))
|
|
|
|
return nfserr_not_only_op;
|
|
|
|
spin_lock(&client_lock);
|
|
|
|
cstate->session = find_in_sessionid_hashtbl(&bcts->sessionid);
|
|
|
|
/* Sorta weird: we only need the refcnt'ing because new_conn acquires
|
|
|
|
* client_lock iself: */
|
|
|
|
if (cstate->session) {
|
|
|
|
nfsd4_get_session(cstate->session);
|
|
|
|
atomic_inc(&cstate->session->se_client->cl_refcount);
|
|
|
|
}
|
|
|
|
spin_unlock(&client_lock);
|
|
|
|
if (!cstate->session)
|
|
|
|
return nfserr_badsession;
|
|
|
|
|
|
|
|
status = nfsd4_map_bcts_dir(&bcts->dir);
|
2011-04-28 03:47:15 +08:00
|
|
|
if (!status)
|
|
|
|
nfsd4_new_conn(rqstp, cstate->session, bcts->dir);
|
|
|
|
return status;
|
2010-10-05 11:12:59 +08:00
|
|
|
}
|
|
|
|
|
2010-05-02 00:56:06 +08:00
|
|
|
static bool nfsd4_compound_in_session(struct nfsd4_session *session, struct nfs4_sessionid *sid)
|
|
|
|
{
|
|
|
|
if (!session)
|
|
|
|
return 0;
|
|
|
|
return !memcmp(sid, &session->se_sessionid, sizeof(*sid));
|
|
|
|
}
|
|
|
|
|
2009-04-03 13:27:58 +08:00
|
|
|
__be32
|
|
|
|
nfsd4_destroy_session(struct svc_rqst *r,
|
|
|
|
struct nfsd4_compound_state *cstate,
|
|
|
|
struct nfsd4_destroy_session *sessionid)
|
|
|
|
{
|
2009-04-03 13:28:38 +08:00
|
|
|
struct nfsd4_session *ses;
|
|
|
|
u32 status = nfserr_badsession;
|
|
|
|
|
|
|
|
/* Notes:
|
|
|
|
* - The confirmed nfs4_client->cl_sessionid holds destroyed sessinid
|
|
|
|
* - Should we return nfserr_back_chan_busy if waiting for
|
|
|
|
* callbacks on to-be-destroyed session?
|
|
|
|
* - Do we need to clear any callback info from previous session?
|
|
|
|
*/
|
|
|
|
|
2010-05-02 00:56:06 +08:00
|
|
|
if (nfsd4_compound_in_session(cstate->session, &sessionid->sessionid)) {
|
2010-04-22 00:27:19 +08:00
|
|
|
if (!nfsd4_last_compound_op(r))
|
|
|
|
return nfserr_not_only_op;
|
|
|
|
}
|
2009-04-03 13:28:38 +08:00
|
|
|
dump_sessionid(__func__, &sessionid->sessionid);
|
2010-05-12 05:12:26 +08:00
|
|
|
spin_lock(&client_lock);
|
2009-04-03 13:28:38 +08:00
|
|
|
ses = find_in_sessionid_hashtbl(&sessionid->sessionid);
|
|
|
|
if (!ses) {
|
2010-05-12 05:12:26 +08:00
|
|
|
spin_unlock(&client_lock);
|
2009-04-03 13:28:38 +08:00
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
|
|
|
|
unhash_session(ses);
|
2010-05-12 05:12:26 +08:00
|
|
|
spin_unlock(&client_lock);
|
2009-04-03 13:28:38 +08:00
|
|
|
|
2010-05-12 05:14:06 +08:00
|
|
|
nfs4_lock_state();
|
2010-12-10 04:52:19 +08:00
|
|
|
nfsd4_probe_callback_sync(ses->se_client);
|
2010-05-12 05:14:06 +08:00
|
|
|
nfs4_unlock_state();
|
2010-06-07 06:37:16 +08:00
|
|
|
|
2012-02-24 09:40:52 +08:00
|
|
|
spin_lock(&client_lock);
|
2010-06-07 06:37:16 +08:00
|
|
|
nfsd4_del_conns(ses);
|
2012-02-24 09:40:52 +08:00
|
|
|
nfsd4_put_session_locked(ses);
|
|
|
|
spin_unlock(&client_lock);
|
2009-04-03 13:28:38 +08:00
|
|
|
status = nfs_ok;
|
|
|
|
out:
|
|
|
|
dprintk("%s returns %d\n", __func__, ntohl(status));
|
|
|
|
return status;
|
2009-04-03 13:27:58 +08:00
|
|
|
}
|
|
|
|
|
2010-10-22 05:17:31 +08:00
|
|
|
static struct nfsd4_conn *__nfsd4_find_conn(struct svc_xprt *xpt, struct nfsd4_session *s)
|
2010-09-30 04:11:06 +08:00
|
|
|
{
|
|
|
|
struct nfsd4_conn *c;
|
|
|
|
|
|
|
|
list_for_each_entry(c, &s->se_conns, cn_persession) {
|
2010-10-22 05:17:31 +08:00
|
|
|
if (c->cn_xprt == xpt) {
|
2010-09-30 04:11:06 +08:00
|
|
|
return c;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
2010-10-22 05:17:31 +08:00
|
|
|
static void nfsd4_sequence_check_conn(struct nfsd4_conn *new, struct nfsd4_session *ses)
|
2010-09-30 04:11:06 +08:00
|
|
|
{
|
|
|
|
struct nfs4_client *clp = ses->se_client;
|
2010-10-22 05:17:31 +08:00
|
|
|
struct nfsd4_conn *c;
|
2010-10-26 22:07:17 +08:00
|
|
|
int ret;
|
2010-09-30 04:11:06 +08:00
|
|
|
|
|
|
|
spin_lock(&clp->cl_lock);
|
2010-10-22 05:17:31 +08:00
|
|
|
c = __nfsd4_find_conn(new->cn_xprt, ses);
|
2010-09-30 04:11:06 +08:00
|
|
|
if (c) {
|
|
|
|
spin_unlock(&clp->cl_lock);
|
|
|
|
free_conn(new);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
__nfsd4_hash_conn(new, ses);
|
|
|
|
spin_unlock(&clp->cl_lock);
|
2010-10-26 22:07:17 +08:00
|
|
|
ret = nfsd4_register_conn(new);
|
|
|
|
if (ret)
|
|
|
|
/* oops; xprt is already down: */
|
|
|
|
nfsd4_conn_lost(&new->cn_xpt_user);
|
2010-09-30 04:11:06 +08:00
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
2011-04-27 09:09:58 +08:00
|
|
|
static bool nfsd4_session_too_many_ops(struct svc_rqst *rqstp, struct nfsd4_session *session)
|
|
|
|
{
|
|
|
|
struct nfsd4_compoundargs *args = rqstp->rq_argp;
|
|
|
|
|
|
|
|
return args->opcnt > session->se_fchannel.maxops;
|
|
|
|
}
|
|
|
|
|
2011-07-14 14:56:02 +08:00
|
|
|
static bool nfsd4_request_too_big(struct svc_rqst *rqstp,
|
|
|
|
struct nfsd4_session *session)
|
|
|
|
{
|
|
|
|
struct xdr_buf *xb = &rqstp->rq_arg;
|
|
|
|
|
|
|
|
return xb->len > session->se_fchannel.maxreq_sz;
|
|
|
|
}
|
|
|
|
|
2009-04-03 13:27:58 +08:00
|
|
|
__be32
|
2009-04-03 13:28:08 +08:00
|
|
|
nfsd4_sequence(struct svc_rqst *rqstp,
|
2009-04-03 13:27:58 +08:00
|
|
|
struct nfsd4_compound_state *cstate,
|
|
|
|
struct nfsd4_sequence *seq)
|
|
|
|
{
|
2009-04-03 13:28:12 +08:00
|
|
|
struct nfsd4_compoundres *resp = rqstp->rq_resp;
|
2009-04-03 13:28:08 +08:00
|
|
|
struct nfsd4_session *session;
|
|
|
|
struct nfsd4_slot *slot;
|
2010-10-22 05:17:31 +08:00
|
|
|
struct nfsd4_conn *conn;
|
2009-04-03 13:28:08 +08:00
|
|
|
int status;
|
|
|
|
|
2009-04-03 13:28:12 +08:00
|
|
|
if (resp->opcnt != 1)
|
|
|
|
return nfserr_sequence_pos;
|
|
|
|
|
2010-10-22 05:17:31 +08:00
|
|
|
/*
|
|
|
|
* Will be either used or freed by nfsd4_sequence_check_conn
|
|
|
|
* below.
|
|
|
|
*/
|
|
|
|
conn = alloc_conn(rqstp, NFS4_CDFC4_FORE);
|
|
|
|
if (!conn)
|
|
|
|
return nfserr_jukebox;
|
|
|
|
|
2010-05-12 05:12:26 +08:00
|
|
|
spin_lock(&client_lock);
|
2009-04-03 13:28:08 +08:00
|
|
|
status = nfserr_badsession;
|
|
|
|
session = find_in_sessionid_hashtbl(&seq->sessionid);
|
|
|
|
if (!session)
|
|
|
|
goto out;
|
|
|
|
|
2011-04-27 09:09:58 +08:00
|
|
|
status = nfserr_too_many_ops;
|
|
|
|
if (nfsd4_session_too_many_ops(rqstp, session))
|
|
|
|
goto out;
|
|
|
|
|
2011-07-14 14:56:02 +08:00
|
|
|
status = nfserr_req_too_big;
|
|
|
|
if (nfsd4_request_too_big(rqstp, session))
|
|
|
|
goto out;
|
|
|
|
|
2009-04-03 13:28:08 +08:00
|
|
|
status = nfserr_badslot;
|
2009-06-16 09:19:13 +08:00
|
|
|
if (seq->slotid >= session->se_fchannel.maxreqs)
|
2009-04-03 13:28:08 +08:00
|
|
|
goto out;
|
|
|
|
|
2009-08-28 20:45:04 +08:00
|
|
|
slot = session->se_slots[seq->slotid];
|
2009-04-03 13:28:08 +08:00
|
|
|
dprintk("%s: slotid %d\n", __func__, seq->slotid);
|
|
|
|
|
2009-08-28 20:45:02 +08:00
|
|
|
/* We do not negotiate the number of slots yet, so set the
|
|
|
|
* maxslots to the session maxreqs which is used to encode
|
|
|
|
* sr_highest_slotid and the sr_target_slot id to maxslots */
|
|
|
|
seq->maxslots = session->se_fchannel.maxreqs;
|
|
|
|
|
2012-02-14 05:39:00 +08:00
|
|
|
status = check_slot_seqid(seq->seqid, slot->sl_seqid,
|
|
|
|
slot->sl_flags & NFSD4_SLOT_INUSE);
|
2009-04-03 13:28:08 +08:00
|
|
|
if (status == nfserr_replay_cache) {
|
2012-02-14 05:56:19 +08:00
|
|
|
status = nfserr_seq_misordered;
|
|
|
|
if (!(slot->sl_flags & NFSD4_SLOT_INITIALIZED))
|
|
|
|
goto out;
|
2009-04-03 13:28:08 +08:00
|
|
|
cstate->slot = slot;
|
|
|
|
cstate->session = session;
|
2009-04-03 13:28:22 +08:00
|
|
|
/* Return the cached reply status and set cstate->status
|
2009-08-28 20:45:04 +08:00
|
|
|
* for nfsd4_proc_compound processing */
|
2009-04-03 13:28:35 +08:00
|
|
|
status = nfsd4_replay_cache_entry(resp, seq);
|
2009-04-03 13:28:22 +08:00
|
|
|
cstate->status = nfserr_replay_cache;
|
2009-08-20 08:21:56 +08:00
|
|
|
goto out;
|
2009-04-03 13:28:08 +08:00
|
|
|
}
|
|
|
|
if (status)
|
|
|
|
goto out;
|
|
|
|
|
2010-10-22 05:17:31 +08:00
|
|
|
nfsd4_sequence_check_conn(conn, session);
|
|
|
|
conn = NULL;
|
2010-09-30 04:11:06 +08:00
|
|
|
|
2009-04-03 13:28:08 +08:00
|
|
|
/* Success! bump slot seqid */
|
|
|
|
slot->sl_seqid = seq->seqid;
|
2012-02-14 05:56:19 +08:00
|
|
|
slot->sl_flags |= NFSD4_SLOT_INUSE;
|
2012-02-14 05:39:00 +08:00
|
|
|
if (seq->cachethis)
|
|
|
|
slot->sl_flags |= NFSD4_SLOT_CACHETHIS;
|
2012-02-14 05:56:19 +08:00
|
|
|
else
|
|
|
|
slot->sl_flags &= ~NFSD4_SLOT_CACHETHIS;
|
2009-04-03 13:28:08 +08:00
|
|
|
|
|
|
|
cstate->slot = slot;
|
|
|
|
cstate->session = session;
|
|
|
|
|
|
|
|
out:
|
2010-04-25 03:35:43 +08:00
|
|
|
/* Hold a session reference until done processing the compound. */
|
2009-08-20 08:21:56 +08:00
|
|
|
if (cstate->session) {
|
2010-11-18 21:30:33 +08:00
|
|
|
struct nfs4_client *clp = session->se_client;
|
|
|
|
|
2010-05-12 05:13:04 +08:00
|
|
|
nfsd4_get_session(cstate->session);
|
2010-11-18 21:30:33 +08:00
|
|
|
atomic_inc(&clp->cl_refcount);
|
2011-10-20 10:12:58 +08:00
|
|
|
switch (clp->cl_cb_state) {
|
|
|
|
case NFSD4_CB_DOWN:
|
2011-10-20 10:13:06 +08:00
|
|
|
seq->status_flags = SEQ4_STATUS_CB_PATH_DOWN;
|
2011-10-20 10:12:58 +08:00
|
|
|
break;
|
|
|
|
case NFSD4_CB_FAULT:
|
2011-10-20 10:13:06 +08:00
|
|
|
seq->status_flags = SEQ4_STATUS_BACKCHANNEL_FAULT;
|
2011-10-20 10:12:58 +08:00
|
|
|
break;
|
2011-10-20 10:13:06 +08:00
|
|
|
default:
|
|
|
|
seq->status_flags = 0;
|
2011-10-20 10:12:58 +08:00
|
|
|
}
|
2009-08-20 08:21:56 +08:00
|
|
|
}
|
2010-10-22 05:17:31 +08:00
|
|
|
kfree(conn);
|
2010-05-12 05:13:04 +08:00
|
|
|
spin_unlock(&client_lock);
|
2009-04-03 13:28:08 +08:00
|
|
|
dprintk("%s: return %d\n", __func__, ntohl(status));
|
|
|
|
return status;
|
2009-04-03 13:27:58 +08:00
|
|
|
}
|
|
|
|
|
2011-10-20 17:51:39 +08:00
|
|
|
static inline bool has_resources(struct nfs4_client *clp)
|
|
|
|
{
|
|
|
|
return !list_empty(&clp->cl_openowners)
|
|
|
|
|| !list_empty(&clp->cl_delegations)
|
|
|
|
|| !list_empty(&clp->cl_sessions);
|
|
|
|
}
|
|
|
|
|
|
|
|
__be32
|
|
|
|
nfsd4_destroy_clientid(struct svc_rqst *rqstp, struct nfsd4_compound_state *cstate, struct nfsd4_destroy_clientid *dc)
|
|
|
|
{
|
|
|
|
struct nfs4_client *conf, *unconf, *clp;
|
|
|
|
int status = 0;
|
|
|
|
|
|
|
|
nfs4_lock_state();
|
|
|
|
unconf = find_unconfirmed_client(&dc->clientid);
|
|
|
|
conf = find_confirmed_client(&dc->clientid);
|
|
|
|
|
|
|
|
if (conf) {
|
|
|
|
clp = conf;
|
|
|
|
|
|
|
|
if (!is_client_expired(conf) && has_resources(conf)) {
|
|
|
|
status = nfserr_clientid_busy;
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* rfc5661 18.50.3 */
|
|
|
|
if (cstate->session && conf == cstate->session->se_client) {
|
|
|
|
status = nfserr_clientid_busy;
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
} else if (unconf)
|
|
|
|
clp = unconf;
|
|
|
|
else {
|
|
|
|
status = nfserr_stale_clientid;
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
|
|
|
|
expire_client(clp);
|
|
|
|
out:
|
|
|
|
nfs4_unlock_state();
|
|
|
|
dprintk("%s return %d\n", __func__, ntohl(status));
|
|
|
|
return status;
|
|
|
|
}
|
|
|
|
|
2010-04-20 03:11:28 +08:00
|
|
|
__be32
|
|
|
|
nfsd4_reclaim_complete(struct svc_rqst *rqstp, struct nfsd4_compound_state *cstate, struct nfsd4_reclaim_complete *rc)
|
|
|
|
{
|
2011-04-27 09:14:30 +08:00
|
|
|
int status = 0;
|
|
|
|
|
2010-04-20 03:11:28 +08:00
|
|
|
if (rc->rca_one_fs) {
|
|
|
|
if (!cstate->current_fh.fh_dentry)
|
|
|
|
return nfserr_nofilehandle;
|
|
|
|
/*
|
|
|
|
* We don't take advantage of the rca_one_fs case.
|
|
|
|
* That's OK, it's optional, we can safely ignore it.
|
|
|
|
*/
|
|
|
|
return nfs_ok;
|
|
|
|
}
|
2011-04-27 09:14:30 +08:00
|
|
|
|
2010-04-20 03:11:28 +08:00
|
|
|
nfs4_lock_state();
|
2011-04-27 09:14:30 +08:00
|
|
|
status = nfserr_complete_already;
|
2012-03-21 21:52:02 +08:00
|
|
|
if (test_and_set_bit(NFSD4_CLIENT_RECLAIM_COMPLETE,
|
|
|
|
&cstate->session->se_client->cl_flags))
|
2011-04-27 09:14:30 +08:00
|
|
|
goto out;
|
|
|
|
|
|
|
|
status = nfserr_stale_clientid;
|
|
|
|
if (is_client_expired(cstate->session->se_client))
|
2010-04-20 03:11:28 +08:00
|
|
|
/*
|
|
|
|
* The following error isn't really legal.
|
|
|
|
* But we only get here if the client just explicitly
|
|
|
|
* destroyed the client. Surely it no longer cares what
|
|
|
|
* error it gets back on an operation for the dead
|
|
|
|
* client.
|
|
|
|
*/
|
2011-04-27 09:14:30 +08:00
|
|
|
goto out;
|
|
|
|
|
|
|
|
status = nfs_ok;
|
2012-03-22 04:42:43 +08:00
|
|
|
nfsd4_client_record_create(cstate->session->se_client);
|
2011-04-27 09:14:30 +08:00
|
|
|
out:
|
2010-04-20 03:11:28 +08:00
|
|
|
nfs4_unlock_state();
|
2011-04-27 09:14:30 +08:00
|
|
|
return status;
|
2010-04-20 03:11:28 +08:00
|
|
|
}
|
|
|
|
|
2006-10-20 14:28:59 +08:00
|
|
|
__be32
|
2006-12-13 16:35:38 +08:00
|
|
|
nfsd4_setclientid(struct svc_rqst *rqstp, struct nfsd4_compound_state *cstate,
|
|
|
|
struct nfsd4_setclientid *setclid)
|
2005-04-17 06:20:36 +08:00
|
|
|
{
|
2011-10-11 03:07:40 +08:00
|
|
|
struct xdr_netobj clname = setclid->se_name;
|
2005-04-17 06:20:36 +08:00
|
|
|
nfs4_verifier clverifier = setclid->se_verf;
|
|
|
|
unsigned int strhashval;
|
2005-06-24 13:03:56 +08:00
|
|
|
struct nfs4_client *conf, *unconf, *new;
|
2006-10-20 14:28:59 +08:00
|
|
|
__be32 status;
|
2005-06-24 13:03:52 +08:00
|
|
|
char dname[HEXDIR_LEN];
|
2005-04-17 06:20:36 +08:00
|
|
|
|
2005-06-24 13:03:52 +08:00
|
|
|
status = nfs4_make_rec_clidname(dname, &clname);
|
|
|
|
if (status)
|
2005-09-13 16:25:39 +08:00
|
|
|
return status;
|
2005-06-24 13:03:52 +08:00
|
|
|
|
2005-04-17 06:20:36 +08:00
|
|
|
/*
|
|
|
|
* XXX The Duplicate Request Cache (DRC) has been checked (??)
|
|
|
|
* We get here on a DRC miss.
|
|
|
|
*/
|
|
|
|
|
2005-06-24 13:03:52 +08:00
|
|
|
strhashval = clientstr_hashval(dname);
|
2005-04-17 06:20:36 +08:00
|
|
|
|
|
|
|
nfs4_lock_state();
|
2010-11-25 06:30:54 +08:00
|
|
|
conf = find_confirmed_client_by_str(dname, strhashval);
|
2005-06-24 13:03:56 +08:00
|
|
|
if (conf) {
|
2007-11-21 05:11:27 +08:00
|
|
|
/* RFC 3530 14.2.33 CASE 0: */
|
2005-04-17 06:20:36 +08:00
|
|
|
status = nfserr_clid_inuse;
|
2010-11-25 06:30:54 +08:00
|
|
|
if (clp_used_exchangeid(conf))
|
|
|
|
goto out;
|
2009-03-19 03:06:26 +08:00
|
|
|
if (!same_creds(&conf->cl_cred, &rqstp->rq_cred)) {
|
2009-08-15 00:57:56 +08:00
|
|
|
char addr_str[INET6_ADDRSTRLEN];
|
|
|
|
rpc_ntop((struct sockaddr *) &conf->cl_addr, addr_str,
|
|
|
|
sizeof(addr_str));
|
|
|
|
dprintk("NFSD: setclientid: string in use by client "
|
|
|
|
"at %s\n", addr_str);
|
2005-04-17 06:20:36 +08:00
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
}
|
2007-11-21 05:11:27 +08:00
|
|
|
/*
|
|
|
|
* section 14.2.33 of RFC 3530 (under the heading "IMPLEMENTATION")
|
|
|
|
* has a description of SETCLIENTID request processing consisting
|
|
|
|
* of 5 bullet points, labeled as CASE0 - CASE4 below.
|
|
|
|
*/
|
2010-11-25 06:30:54 +08:00
|
|
|
unconf = find_unconfirmed_client_by_str(dname, strhashval);
|
2011-08-11 07:07:33 +08:00
|
|
|
status = nfserr_jukebox;
|
2005-04-17 06:20:36 +08:00
|
|
|
if (!conf) {
|
2007-11-21 05:11:27 +08:00
|
|
|
/*
|
|
|
|
* RFC 3530 14.2.33 CASE 4:
|
|
|
|
* placed first, because it is the normal case
|
2005-04-17 06:20:36 +08:00
|
|
|
*/
|
|
|
|
if (unconf)
|
|
|
|
expire_client(unconf);
|
2009-09-10 17:27:34 +08:00
|
|
|
new = create_client(clname, dname, rqstp, &clverifier);
|
2005-06-24 13:03:52 +08:00
|
|
|
if (new == NULL)
|
2005-04-17 06:20:36 +08:00
|
|
|
goto out;
|
|
|
|
gen_clid(new);
|
2007-07-27 05:04:54 +08:00
|
|
|
} else if (same_verf(&conf->cl_verifier, &clverifier)) {
|
2005-04-17 06:20:36 +08:00
|
|
|
/*
|
2007-11-21 05:11:27 +08:00
|
|
|
* RFC 3530 14.2.33 CASE 1:
|
|
|
|
* probable callback update
|
2005-04-17 06:20:36 +08:00
|
|
|
*/
|
2005-06-24 13:04:06 +08:00
|
|
|
if (unconf) {
|
|
|
|
/* Note this is removing unconfirmed {*x***},
|
|
|
|
* which is stronger than RFC recommended {vxc**}.
|
|
|
|
* This has the advantage that there is at most
|
|
|
|
* one {*x***} in either list at any time.
|
|
|
|
*/
|
|
|
|
expire_client(unconf);
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
2009-09-10 17:27:34 +08:00
|
|
|
new = create_client(clname, dname, rqstp, &clverifier);
|
2005-06-24 13:03:52 +08:00
|
|
|
if (new == NULL)
|
2005-04-17 06:20:36 +08:00
|
|
|
goto out;
|
|
|
|
copy_clid(new, conf);
|
|
|
|
} else if (!unconf) {
|
|
|
|
/*
|
2007-11-21 05:11:27 +08:00
|
|
|
* RFC 3530 14.2.33 CASE 2:
|
|
|
|
* probable client reboot; state will be removed if
|
|
|
|
* confirmed.
|
2005-04-17 06:20:36 +08:00
|
|
|
*/
|
2009-09-10 17:27:34 +08:00
|
|
|
new = create_client(clname, dname, rqstp, &clverifier);
|
2005-06-24 13:03:52 +08:00
|
|
|
if (new == NULL)
|
2005-04-17 06:20:36 +08:00
|
|
|
goto out;
|
|
|
|
gen_clid(new);
|
2007-11-20 08:09:50 +08:00
|
|
|
} else {
|
2007-11-21 05:11:27 +08:00
|
|
|
/*
|
|
|
|
* RFC 3530 14.2.33 CASE 3:
|
|
|
|
* probable client reboot; state will be removed if
|
|
|
|
* confirmed.
|
2005-04-17 06:20:36 +08:00
|
|
|
*/
|
|
|
|
expire_client(unconf);
|
2009-09-10 17:27:34 +08:00
|
|
|
new = create_client(clname, dname, rqstp, &clverifier);
|
2005-06-24 13:03:52 +08:00
|
|
|
if (new == NULL)
|
2005-04-17 06:20:36 +08:00
|
|
|
goto out;
|
|
|
|
gen_clid(new);
|
|
|
|
}
|
2010-10-20 07:36:51 +08:00
|
|
|
/*
|
|
|
|
* XXX: we should probably set this at creation time, and check
|
|
|
|
* for consistent minorversion use throughout:
|
|
|
|
*/
|
|
|
|
new->cl_minorversion = 0;
|
2010-12-15 13:09:01 +08:00
|
|
|
gen_callback(new, setclid, rqstp);
|
2007-08-10 06:34:32 +08:00
|
|
|
add_to_unconfirmed(new, strhashval);
|
2005-04-17 06:20:36 +08:00
|
|
|
setclid->se_clientid.cl_boot = new->cl_clientid.cl_boot;
|
|
|
|
setclid->se_clientid.cl_id = new->cl_clientid.cl_id;
|
|
|
|
memcpy(setclid->se_confirm.data, new->cl_confirm.data, sizeof(setclid->se_confirm.data));
|
|
|
|
status = nfs_ok;
|
|
|
|
out:
|
|
|
|
nfs4_unlock_state();
|
|
|
|
return status;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
/*
|
2007-11-21 05:11:27 +08:00
|
|
|
* Section 14.2.34 of RFC 3530 (under the heading "IMPLEMENTATION") has
|
|
|
|
* a description of SETCLIENTID_CONFIRM request processing consisting of 4
|
|
|
|
* bullets, labeled as CASE1 - CASE4 below.
|
2005-04-17 06:20:36 +08:00
|
|
|
*/
|
2006-10-20 14:28:59 +08:00
|
|
|
__be32
|
2006-12-13 16:35:38 +08:00
|
|
|
nfsd4_setclientid_confirm(struct svc_rqst *rqstp,
|
|
|
|
struct nfsd4_compound_state *cstate,
|
|
|
|
struct nfsd4_setclientid_confirm *setclientid_confirm)
|
2005-04-17 06:20:36 +08:00
|
|
|
{
|
2009-08-15 00:57:56 +08:00
|
|
|
struct sockaddr *sa = svc_addr(rqstp);
|
2005-06-24 13:04:14 +08:00
|
|
|
struct nfs4_client *conf, *unconf;
|
2005-04-17 06:20:36 +08:00
|
|
|
nfs4_verifier confirm = setclientid_confirm->sc_confirm;
|
|
|
|
clientid_t * clid = &setclientid_confirm->sc_clientid;
|
2006-10-20 14:28:59 +08:00
|
|
|
__be32 status;
|
2005-04-17 06:20:36 +08:00
|
|
|
|
|
|
|
if (STALE_CLIENTID(clid))
|
|
|
|
return nfserr_stale_clientid;
|
|
|
|
/*
|
|
|
|
* XXX The Duplicate Request Cache (DRC) has been checked (??)
|
|
|
|
* We get here on a DRC miss.
|
|
|
|
*/
|
|
|
|
|
|
|
|
nfs4_lock_state();
|
2005-06-24 13:04:14 +08:00
|
|
|
|
|
|
|
conf = find_confirmed_client(clid);
|
|
|
|
unconf = find_unconfirmed_client(clid);
|
|
|
|
|
|
|
|
status = nfserr_clid_inuse;
|
2009-08-15 00:57:56 +08:00
|
|
|
if (conf && !rpc_cmp_addr((struct sockaddr *) &conf->cl_addr, sa))
|
2005-06-24 13:04:14 +08:00
|
|
|
goto out;
|
2009-08-15 00:57:56 +08:00
|
|
|
if (unconf && !rpc_cmp_addr((struct sockaddr *) &unconf->cl_addr, sa))
|
2005-06-24 13:04:14 +08:00
|
|
|
goto out;
|
|
|
|
|
2007-11-21 05:11:27 +08:00
|
|
|
/*
|
|
|
|
* section 14.2.34 of RFC 3530 has a description of
|
|
|
|
* SETCLIENTID_CONFIRM request processing consisting
|
|
|
|
* of 4 bullet points, labeled as CASE1 - CASE4 below.
|
|
|
|
*/
|
2007-11-21 04:54:10 +08:00
|
|
|
if (conf && unconf && same_verf(&confirm, &unconf->cl_confirm)) {
|
2007-11-21 05:11:27 +08:00
|
|
|
/*
|
|
|
|
* RFC 3530 14.2.34 CASE 1:
|
|
|
|
* callback update
|
|
|
|
*/
|
2007-07-27 05:04:54 +08:00
|
|
|
if (!same_creds(&conf->cl_cred, &unconf->cl_cred))
|
2005-04-17 06:20:36 +08:00
|
|
|
status = nfserr_clid_inuse;
|
|
|
|
else {
|
2010-10-20 05:56:52 +08:00
|
|
|
nfsd4_change_callback(conf, &unconf->cl_cb_conn);
|
|
|
|
nfsd4_probe_callback(conf);
|
2005-06-24 13:04:08 +08:00
|
|
|
expire_client(unconf);
|
2005-04-17 06:20:36 +08:00
|
|
|
status = nfs_ok;
|
2005-06-24 13:04:08 +08:00
|
|
|
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
2007-11-21 05:52:07 +08:00
|
|
|
} else if (conf && !unconf) {
|
2007-11-21 05:11:27 +08:00
|
|
|
/*
|
|
|
|
* RFC 3530 14.2.34 CASE 2:
|
|
|
|
* probable retransmitted request; play it safe and
|
|
|
|
* do nothing.
|
2005-06-24 13:04:13 +08:00
|
|
|
*/
|
2007-07-27 05:04:54 +08:00
|
|
|
if (!same_creds(&conf->cl_cred, &rqstp->rq_cred))
|
2005-04-17 06:20:36 +08:00
|
|
|
status = nfserr_clid_inuse;
|
2005-06-24 13:04:14 +08:00
|
|
|
else
|
2005-04-17 06:20:36 +08:00
|
|
|
status = nfs_ok;
|
2005-06-24 13:04:13 +08:00
|
|
|
} else if (!conf && unconf
|
2007-07-27 05:04:54 +08:00
|
|
|
&& same_verf(&unconf->cl_confirm, &confirm)) {
|
2007-11-21 05:11:27 +08:00
|
|
|
/*
|
|
|
|
* RFC 3530 14.2.34 CASE 3:
|
|
|
|
* Normal case; new or rebooted client:
|
2005-06-24 13:04:13 +08:00
|
|
|
*/
|
2007-07-27 05:04:54 +08:00
|
|
|
if (!same_creds(&unconf->cl_cred, &rqstp->rq_cred)) {
|
2005-04-17 06:20:36 +08:00
|
|
|
status = nfserr_clid_inuse;
|
|
|
|
} else {
|
2005-06-24 13:04:08 +08:00
|
|
|
unsigned int hash =
|
|
|
|
clientstr_hashval(unconf->cl_recdir);
|
|
|
|
conf = find_confirmed_client_by_str(unconf->cl_recdir,
|
2010-11-25 06:30:54 +08:00
|
|
|
hash);
|
2005-06-24 13:04:08 +08:00
|
|
|
if (conf) {
|
2012-03-22 04:42:43 +08:00
|
|
|
nfsd4_client_record_remove(conf);
|
2005-06-24 13:04:08 +08:00
|
|
|
expire_client(conf);
|
|
|
|
}
|
2005-04-17 06:20:36 +08:00
|
|
|
move_to_confirmed(unconf);
|
2005-06-24 13:04:14 +08:00
|
|
|
conf = unconf;
|
2010-10-20 05:56:52 +08:00
|
|
|
nfsd4_probe_callback(conf);
|
2005-06-24 13:04:08 +08:00
|
|
|
status = nfs_ok;
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
2007-07-27 05:04:54 +08:00
|
|
|
} else if ((!conf || (conf && !same_verf(&conf->cl_confirm, &confirm)))
|
|
|
|
&& (!unconf || (unconf && !same_verf(&unconf->cl_confirm,
|
2005-06-24 13:04:13 +08:00
|
|
|
&confirm)))) {
|
2007-11-21 05:11:27 +08:00
|
|
|
/*
|
|
|
|
* RFC 3530 14.2.34 CASE 4:
|
|
|
|
* Client probably hasn't noticed that we rebooted yet.
|
2005-06-24 13:04:13 +08:00
|
|
|
*/
|
2005-04-17 06:20:36 +08:00
|
|
|
status = nfserr_stale_clientid;
|
2005-06-24 13:04:13 +08:00
|
|
|
} else {
|
2005-06-24 13:04:11 +08:00
|
|
|
/* check that we have hit one of the cases...*/
|
|
|
|
status = nfserr_clid_inuse;
|
|
|
|
}
|
2005-04-17 06:20:36 +08:00
|
|
|
out:
|
|
|
|
nfs4_unlock_state();
|
|
|
|
return status;
|
|
|
|
}
|
|
|
|
|
2011-10-14 04:00:16 +08:00
|
|
|
static struct nfs4_file *nfsd4_alloc_file(void)
|
|
|
|
{
|
|
|
|
return kmem_cache_alloc(file_slab, GFP_KERNEL);
|
|
|
|
}
|
|
|
|
|
2005-04-17 06:20:36 +08:00
|
|
|
/* OPEN Share state helper functions */
|
2011-10-14 04:00:16 +08:00
|
|
|
static void nfsd4_init_file(struct nfs4_file *fp, struct inode *ino)
|
2005-04-17 06:20:36 +08:00
|
|
|
{
|
|
|
|
unsigned int hashval = file_hashval(ino);
|
|
|
|
|
2011-10-14 04:00:16 +08:00
|
|
|
atomic_set(&fp->fi_ref, 1);
|
|
|
|
INIT_LIST_HEAD(&fp->fi_hash);
|
|
|
|
INIT_LIST_HEAD(&fp->fi_stateids);
|
|
|
|
INIT_LIST_HEAD(&fp->fi_delegations);
|
|
|
|
fp->fi_inode = igrab(ino);
|
|
|
|
fp->fi_had_conflict = false;
|
|
|
|
fp->fi_lease = NULL;
|
|
|
|
memset(fp->fi_fds, 0, sizeof(fp->fi_fds));
|
|
|
|
memset(fp->fi_access, 0, sizeof(fp->fi_access));
|
|
|
|
spin_lock(&recall_lock);
|
|
|
|
list_add(&fp->fi_hash, &file_hashtbl[hashval]);
|
|
|
|
spin_unlock(&recall_lock);
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
|
|
|
|
2005-06-24 13:03:01 +08:00
|
|
|
static void
|
2006-12-07 12:33:20 +08:00
|
|
|
nfsd4_free_slab(struct kmem_cache **slab)
|
2005-04-17 06:20:36 +08:00
|
|
|
{
|
2005-06-24 13:03:01 +08:00
|
|
|
if (*slab == NULL)
|
|
|
|
return;
|
2006-09-27 16:49:40 +08:00
|
|
|
kmem_cache_destroy(*slab);
|
2005-06-24 13:03:01 +08:00
|
|
|
*slab = NULL;
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
|
|
|
|
2007-08-02 03:30:59 +08:00
|
|
|
void
|
2005-04-17 06:20:36 +08:00
|
|
|
nfsd4_free_slabs(void)
|
|
|
|
{
|
2011-07-31 11:33:59 +08:00
|
|
|
nfsd4_free_slab(&openowner_slab);
|
|
|
|
nfsd4_free_slab(&lockowner_slab);
|
2005-06-24 13:03:01 +08:00
|
|
|
nfsd4_free_slab(&file_slab);
|
2005-06-24 13:03:03 +08:00
|
|
|
nfsd4_free_slab(&stateid_slab);
|
2005-06-24 13:03:04 +08:00
|
|
|
nfsd4_free_slab(&deleg_slab);
|
2005-06-24 13:03:01 +08:00
|
|
|
}
|
2005-04-17 06:20:36 +08:00
|
|
|
|
2011-11-02 03:24:59 +08:00
|
|
|
int
|
2005-06-24 13:03:01 +08:00
|
|
|
nfsd4_init_slabs(void)
|
|
|
|
{
|
2011-07-31 11:33:59 +08:00
|
|
|
openowner_slab = kmem_cache_create("nfsd4_openowners",
|
|
|
|
sizeof(struct nfs4_openowner), 0, 0, NULL);
|
|
|
|
if (openowner_slab == NULL)
|
|
|
|
goto out_nomem;
|
|
|
|
lockowner_slab = kmem_cache_create("nfsd4_lockowners",
|
|
|
|
sizeof(struct nfs4_openowner), 0, 0, NULL);
|
|
|
|
if (lockowner_slab == NULL)
|
2005-06-24 13:03:01 +08:00
|
|
|
goto out_nomem;
|
|
|
|
file_slab = kmem_cache_create("nfsd4_files",
|
2007-07-20 09:11:58 +08:00
|
|
|
sizeof(struct nfs4_file), 0, 0, NULL);
|
2005-06-24 13:03:01 +08:00
|
|
|
if (file_slab == NULL)
|
|
|
|
goto out_nomem;
|
2005-06-24 13:03:03 +08:00
|
|
|
stateid_slab = kmem_cache_create("nfsd4_stateids",
|
2011-09-08 04:06:42 +08:00
|
|
|
sizeof(struct nfs4_ol_stateid), 0, 0, NULL);
|
2005-06-24 13:03:03 +08:00
|
|
|
if (stateid_slab == NULL)
|
|
|
|
goto out_nomem;
|
2005-06-24 13:03:04 +08:00
|
|
|
deleg_slab = kmem_cache_create("nfsd4_delegations",
|
2007-07-20 09:11:58 +08:00
|
|
|
sizeof(struct nfs4_delegation), 0, 0, NULL);
|
2005-06-24 13:03:04 +08:00
|
|
|
if (deleg_slab == NULL)
|
|
|
|
goto out_nomem;
|
2005-06-24 13:03:01 +08:00
|
|
|
return 0;
|
|
|
|
out_nomem:
|
|
|
|
nfsd4_free_slabs();
|
|
|
|
dprintk("nfsd4: out of memory while initializing nfsv4\n");
|
|
|
|
return -ENOMEM;
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
|
|
|
|
2011-07-31 11:33:59 +08:00
|
|
|
void nfs4_free_openowner(struct nfs4_openowner *oo)
|
|
|
|
{
|
|
|
|
kfree(oo->oo_owner.so_owner.data);
|
|
|
|
kmem_cache_free(openowner_slab, oo);
|
|
|
|
}
|
|
|
|
|
|
|
|
void nfs4_free_lockowner(struct nfs4_lockowner *lo)
|
2005-04-17 06:20:36 +08:00
|
|
|
{
|
2011-07-31 11:33:59 +08:00
|
|
|
kfree(lo->lo_owner.so_owner.data);
|
|
|
|
kmem_cache_free(lockowner_slab, lo);
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
|
|
|
|
2011-08-12 21:42:57 +08:00
|
|
|
static void init_nfs4_replay(struct nfs4_replay *rp)
|
2005-04-17 06:20:36 +08:00
|
|
|
{
|
2011-08-12 21:42:57 +08:00
|
|
|
rp->rp_status = nfserr_serverfault;
|
|
|
|
rp->rp_buflen = 0;
|
|
|
|
rp->rp_buf = rp->rp_ibuf;
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
|
|
|
|
2011-07-31 11:33:59 +08:00
|
|
|
static inline void *alloc_stateowner(struct kmem_cache *slab, struct xdr_netobj *owner, struct nfs4_client *clp)
|
2011-08-12 21:42:57 +08:00
|
|
|
{
|
2005-04-17 06:20:36 +08:00
|
|
|
struct nfs4_stateowner *sop;
|
|
|
|
|
2011-07-31 11:33:59 +08:00
|
|
|
sop = kmem_cache_alloc(slab, GFP_KERNEL);
|
2011-08-12 21:42:57 +08:00
|
|
|
if (!sop)
|
|
|
|
return NULL;
|
|
|
|
|
|
|
|
sop->so_owner.data = kmemdup(owner->data, owner->len, GFP_KERNEL);
|
|
|
|
if (!sop->so_owner.data) {
|
2011-07-31 11:33:59 +08:00
|
|
|
kmem_cache_free(slab, sop);
|
2005-04-17 06:20:36 +08:00
|
|
|
return NULL;
|
2011-08-12 21:42:57 +08:00
|
|
|
}
|
|
|
|
sop->so_owner.len = owner->len;
|
|
|
|
|
2005-06-24 13:04:17 +08:00
|
|
|
INIT_LIST_HEAD(&sop->so_stateids);
|
2011-08-12 21:42:57 +08:00
|
|
|
sop->so_client = clp;
|
|
|
|
init_nfs4_replay(&sop->so_replay);
|
|
|
|
return sop;
|
|
|
|
}
|
|
|
|
|
2011-07-31 11:33:59 +08:00
|
|
|
static void hash_openowner(struct nfs4_openowner *oo, struct nfs4_client *clp, unsigned int strhashval)
|
2011-08-12 21:42:57 +08:00
|
|
|
{
|
2011-11-08 06:23:30 +08:00
|
|
|
list_add(&oo->oo_owner.so_strhash, &ownerstr_hashtbl[strhashval]);
|
2011-07-31 11:33:59 +08:00
|
|
|
list_add(&oo->oo_perclient, &clp->cl_openowners);
|
2011-08-12 21:42:57 +08:00
|
|
|
}
|
|
|
|
|
2011-07-31 11:33:59 +08:00
|
|
|
static struct nfs4_openowner *
|
2011-08-12 21:42:57 +08:00
|
|
|
alloc_init_open_stateowner(unsigned int strhashval, struct nfs4_client *clp, struct nfsd4_open *open) {
|
2011-07-31 11:33:59 +08:00
|
|
|
struct nfs4_openowner *oo;
|
2011-08-12 21:42:57 +08:00
|
|
|
|
2011-07-31 11:33:59 +08:00
|
|
|
oo = alloc_stateowner(openowner_slab, &open->op_owner, clp);
|
|
|
|
if (!oo)
|
2011-08-12 21:42:57 +08:00
|
|
|
return NULL;
|
2011-07-31 11:33:59 +08:00
|
|
|
oo->oo_owner.so_is_open_owner = 1;
|
|
|
|
oo->oo_owner.so_seqid = open->op_seqid;
|
2011-10-14 03:12:59 +08:00
|
|
|
oo->oo_flags = NFS4_OO_NEW;
|
2011-07-31 11:33:59 +08:00
|
|
|
oo->oo_time = 0;
|
2011-09-17 05:42:48 +08:00
|
|
|
oo->oo_last_closed_stid = NULL;
|
2011-07-31 11:33:59 +08:00
|
|
|
INIT_LIST_HEAD(&oo->oo_close_lru);
|
|
|
|
hash_openowner(oo, clp, strhashval);
|
|
|
|
return oo;
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
|
|
|
|
2011-10-17 23:14:48 +08:00
|
|
|
static void init_open_stateid(struct nfs4_ol_stateid *stp, struct nfs4_file *fp, struct nfsd4_open *open) {
|
2011-07-31 11:33:59 +08:00
|
|
|
struct nfs4_openowner *oo = open->op_openowner;
|
2011-09-16 03:02:41 +08:00
|
|
|
struct nfs4_client *clp = oo->oo_owner.so_client;
|
2005-04-17 06:20:36 +08:00
|
|
|
|
2011-10-17 23:14:48 +08:00
|
|
|
init_stid(&stp->st_stid, clp, NFS4_OPEN_STID);
|
2005-06-24 13:04:17 +08:00
|
|
|
INIT_LIST_HEAD(&stp->st_lockowners);
|
2011-07-31 11:33:59 +08:00
|
|
|
list_add(&stp->st_perstateowner, &oo->oo_owner.so_stateids);
|
2005-06-24 13:03:08 +08:00
|
|
|
list_add(&stp->st_perfile, &fp->fi_stateids);
|
2011-07-31 11:33:59 +08:00
|
|
|
stp->st_stateowner = &oo->oo_owner;
|
2005-06-24 13:03:10 +08:00
|
|
|
get_nfs4_file(fp);
|
2005-04-17 06:20:36 +08:00
|
|
|
stp->st_file = fp;
|
|
|
|
stp->st_access_bmap = 0;
|
|
|
|
stp->st_deny_bmap = 0;
|
2011-10-11 05:44:19 +08:00
|
|
|
__set_bit(open->op_share_access, &stp->st_access_bmap);
|
2005-04-17 06:20:36 +08:00
|
|
|
__set_bit(open->op_share_deny, &stp->st_deny_bmap);
|
2005-07-08 08:59:27 +08:00
|
|
|
stp->st_openstp = NULL;
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
|
|
|
|
2005-06-24 13:04:03 +08:00
|
|
|
static void
|
2011-07-31 11:33:59 +08:00
|
|
|
move_to_close_lru(struct nfs4_openowner *oo)
|
2005-04-17 06:20:36 +08:00
|
|
|
{
|
2011-07-31 11:33:59 +08:00
|
|
|
dprintk("NFSD: move_to_close_lru nfs4_openowner %p\n", oo);
|
2005-04-17 06:20:36 +08:00
|
|
|
|
2011-07-31 11:33:59 +08:00
|
|
|
list_move_tail(&oo->oo_close_lru, &close_lru);
|
|
|
|
oo->oo_time = get_seconds();
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
static int
|
2007-07-27 05:04:54 +08:00
|
|
|
same_owner_str(struct nfs4_stateowner *sop, struct xdr_netobj *owner,
|
|
|
|
clientid_t *clid)
|
|
|
|
{
|
|
|
|
return (sop->so_owner.len == owner->len) &&
|
|
|
|
0 == memcmp(sop->so_owner.data, owner->data, owner->len) &&
|
|
|
|
(sop->so_client->cl_clientid.cl_id == clid->cl_id);
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
|
|
|
|
2011-07-31 11:33:59 +08:00
|
|
|
static struct nfs4_openowner *
|
2005-04-17 06:20:36 +08:00
|
|
|
find_openstateowner_str(unsigned int hashval, struct nfsd4_open *open)
|
|
|
|
{
|
2011-10-13 04:24:27 +08:00
|
|
|
struct nfs4_stateowner *so;
|
|
|
|
struct nfs4_openowner *oo;
|
2005-04-17 06:20:36 +08:00
|
|
|
|
2011-11-08 06:23:30 +08:00
|
|
|
list_for_each_entry(so, &ownerstr_hashtbl[hashval], so_strhash) {
|
|
|
|
if (!so->so_is_open_owner)
|
|
|
|
continue;
|
2011-10-13 04:24:27 +08:00
|
|
|
if (same_owner_str(so, &open->op_owner, &open->op_clientid)) {
|
|
|
|
oo = openowner(so);
|
|
|
|
renew_client(oo->oo_owner.so_client);
|
|
|
|
return oo;
|
|
|
|
}
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* search file_hashtbl[] for file */
|
|
|
|
static struct nfs4_file *
|
|
|
|
find_file(struct inode *ino)
|
|
|
|
{
|
|
|
|
unsigned int hashval = file_hashval(ino);
|
|
|
|
struct nfs4_file *fp;
|
|
|
|
|
2009-02-23 06:51:34 +08:00
|
|
|
spin_lock(&recall_lock);
|
2005-04-17 06:20:36 +08:00
|
|
|
list_for_each_entry(fp, &file_hashtbl[hashval], fi_hash) {
|
2005-06-24 13:03:10 +08:00
|
|
|
if (fp->fi_inode == ino) {
|
|
|
|
get_nfs4_file(fp);
|
2009-02-23 06:51:34 +08:00
|
|
|
spin_unlock(&recall_lock);
|
2005-04-17 06:20:36 +08:00
|
|
|
return fp;
|
2005-06-24 13:03:10 +08:00
|
|
|
}
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
2009-02-23 06:51:34 +08:00
|
|
|
spin_unlock(&recall_lock);
|
2005-04-17 06:20:36 +08:00
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Called to check deny when READ with all zero stateid or
|
|
|
|
* WRITE with all zero or all one stateid
|
|
|
|
*/
|
2006-10-20 14:28:59 +08:00
|
|
|
static __be32
|
2005-04-17 06:20:36 +08:00
|
|
|
nfs4_share_conflict(struct svc_fh *current_fh, unsigned int deny_type)
|
|
|
|
{
|
|
|
|
struct inode *ino = current_fh->fh_dentry->d_inode;
|
|
|
|
struct nfs4_file *fp;
|
2011-09-08 04:06:42 +08:00
|
|
|
struct nfs4_ol_stateid *stp;
|
2006-10-20 14:28:59 +08:00
|
|
|
__be32 ret;
|
2005-04-17 06:20:36 +08:00
|
|
|
|
|
|
|
dprintk("NFSD: nfs4_share_conflict\n");
|
|
|
|
|
|
|
|
fp = find_file(ino);
|
2005-06-24 13:03:10 +08:00
|
|
|
if (!fp)
|
|
|
|
return nfs_ok;
|
2005-07-08 08:59:23 +08:00
|
|
|
ret = nfserr_locked;
|
2005-04-17 06:20:36 +08:00
|
|
|
/* Search for conflicting share reservations */
|
2005-06-24 13:03:10 +08:00
|
|
|
list_for_each_entry(stp, &fp->fi_stateids, st_perfile) {
|
|
|
|
if (test_bit(deny_type, &stp->st_deny_bmap) ||
|
|
|
|
test_bit(NFS4_SHARE_DENY_BOTH, &stp->st_deny_bmap))
|
|
|
|
goto out;
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
2005-06-24 13:03:10 +08:00
|
|
|
ret = nfs_ok;
|
|
|
|
out:
|
|
|
|
put_nfs4_file(fp);
|
|
|
|
return ret;
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
|
|
|
|
2011-02-01 00:54:04 +08:00
|
|
|
static void nfsd_break_one_deleg(struct nfs4_delegation *dp)
|
2005-04-17 06:20:36 +08:00
|
|
|
{
|
|
|
|
/* We're assuming the state code never drops its reference
|
|
|
|
* without first removing the lease. Since we're in this lease
|
|
|
|
* callback (and since the lease code is serialized by the kernel
|
|
|
|
* lock) we know the server hasn't removed the lease yet, we know
|
|
|
|
* it's safe to take a reference: */
|
|
|
|
atomic_inc(&dp->dl_count);
|
|
|
|
|
|
|
|
list_add_tail(&dp->dl_recall_lru, &del_recall_lru);
|
|
|
|
|
2010-11-17 23:26:56 +08:00
|
|
|
/* only place dl_time is set. protected by lock_flocks*/
|
2005-04-17 06:20:36 +08:00
|
|
|
dp->dl_time = get_seconds();
|
|
|
|
|
2011-02-01 00:54:04 +08:00
|
|
|
nfsd4_cb_recall(dp);
|
|
|
|
}
|
|
|
|
|
2011-02-01 08:20:39 +08:00
|
|
|
/* Called from break_lease() with lock_flocks() held. */
|
2011-02-01 00:54:04 +08:00
|
|
|
static void nfsd_break_deleg_cb(struct file_lock *fl)
|
|
|
|
{
|
2011-02-01 08:20:39 +08:00
|
|
|
struct nfs4_file *fp = (struct nfs4_file *)fl->fl_owner;
|
|
|
|
struct nfs4_delegation *dp;
|
2011-02-01 00:54:04 +08:00
|
|
|
|
2011-02-01 08:20:39 +08:00
|
|
|
BUG_ON(!fp);
|
|
|
|
/* We assume break_lease is only called once per lease: */
|
|
|
|
BUG_ON(fp->fi_had_conflict);
|
2007-09-13 06:56:12 +08:00
|
|
|
/*
|
|
|
|
* We don't want the locks code to timeout the lease for us;
|
2011-02-01 08:20:39 +08:00
|
|
|
* we'll remove it ourself if a delegation isn't returned
|
2011-02-01 00:54:04 +08:00
|
|
|
* in time:
|
2007-09-13 06:56:12 +08:00
|
|
|
*/
|
|
|
|
fl->fl_break_time = 0;
|
2005-04-17 06:20:36 +08:00
|
|
|
|
2011-02-08 05:53:46 +08:00
|
|
|
spin_lock(&recall_lock);
|
2011-02-01 08:20:39 +08:00
|
|
|
fp->fi_had_conflict = true;
|
|
|
|
list_for_each_entry(dp, &fp->fi_delegations, dl_perfile)
|
|
|
|
nfsd_break_one_deleg(dp);
|
2011-02-08 05:53:46 +08:00
|
|
|
spin_unlock(&recall_lock);
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
static
|
|
|
|
int nfsd_change_deleg_cb(struct file_lock **onlist, int arg)
|
|
|
|
{
|
|
|
|
if (arg & F_UNLCK)
|
|
|
|
return lease_modify(onlist, arg);
|
|
|
|
else
|
|
|
|
return -EAGAIN;
|
|
|
|
}
|
|
|
|
|
2009-09-22 08:01:12 +08:00
|
|
|
static const struct lock_manager_operations nfsd_lease_mng_ops = {
|
2011-07-21 08:21:59 +08:00
|
|
|
.lm_break = nfsd_break_deleg_cb,
|
|
|
|
.lm_change = nfsd_change_deleg_cb,
|
2005-04-17 06:20:36 +08:00
|
|
|
};
|
|
|
|
|
2011-09-02 21:03:37 +08:00
|
|
|
static __be32 nfsd4_check_seqid(struct nfsd4_compound_state *cstate, struct nfs4_stateowner *so, u32 seqid)
|
|
|
|
{
|
|
|
|
if (nfsd4_has_session(cstate))
|
|
|
|
return nfs_ok;
|
|
|
|
if (seqid == so->so_seqid - 1)
|
|
|
|
return nfserr_replay_me;
|
|
|
|
if (seqid == so->so_seqid)
|
|
|
|
return nfs_ok;
|
|
|
|
return nfserr_bad_seqid;
|
|
|
|
}
|
2005-04-17 06:20:36 +08:00
|
|
|
|
2006-10-20 14:28:59 +08:00
|
|
|
__be32
|
2009-04-03 13:28:45 +08:00
|
|
|
nfsd4_process_open1(struct nfsd4_compound_state *cstate,
|
|
|
|
struct nfsd4_open *open)
|
2005-04-17 06:20:36 +08:00
|
|
|
{
|
|
|
|
clientid_t *clientid = &open->op_clientid;
|
|
|
|
struct nfs4_client *clp = NULL;
|
|
|
|
unsigned int strhashval;
|
2011-07-31 11:33:59 +08:00
|
|
|
struct nfs4_openowner *oo = NULL;
|
2011-10-18 03:57:47 +08:00
|
|
|
__be32 status;
|
2005-04-17 06:20:36 +08:00
|
|
|
|
|
|
|
if (STALE_CLIENTID(&open->op_clientid))
|
|
|
|
return nfserr_stale_clientid;
|
2011-10-14 04:00:16 +08:00
|
|
|
/*
|
|
|
|
* In case we need it later, after we've already created the
|
|
|
|
* file and don't want to risk a further failure:
|
|
|
|
*/
|
|
|
|
open->op_file = nfsd4_alloc_file();
|
|
|
|
if (open->op_file == NULL)
|
|
|
|
return nfserr_jukebox;
|
2005-04-17 06:20:36 +08:00
|
|
|
|
2011-11-08 06:23:30 +08:00
|
|
|
strhashval = ownerstr_hashval(clientid->cl_id, &open->op_owner);
|
2011-07-31 11:33:59 +08:00
|
|
|
oo = find_openstateowner_str(strhashval, open);
|
|
|
|
open->op_openowner = oo;
|
|
|
|
if (!oo) {
|
2005-04-17 06:20:36 +08:00
|
|
|
clp = find_confirmed_client(clientid);
|
|
|
|
if (clp == NULL)
|
2006-01-19 09:43:34 +08:00
|
|
|
return nfserr_expired;
|
2011-10-13 08:44:20 +08:00
|
|
|
goto new_owner;
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
2011-09-13 00:24:13 +08:00
|
|
|
if (!(oo->oo_flags & NFS4_OO_CONFIRMED)) {
|
2006-01-19 09:43:34 +08:00
|
|
|
/* Replace unconfirmed owners without checking for replay. */
|
2011-07-31 11:33:59 +08:00
|
|
|
clp = oo->oo_owner.so_client;
|
|
|
|
release_openowner(oo);
|
|
|
|
open->op_openowner = NULL;
|
2011-10-13 08:44:20 +08:00
|
|
|
goto new_owner;
|
2006-01-19 09:43:34 +08:00
|
|
|
}
|
2011-10-18 03:57:47 +08:00
|
|
|
status = nfsd4_check_seqid(cstate, &oo->oo_owner, open->op_seqid);
|
|
|
|
if (status)
|
|
|
|
return status;
|
|
|
|
clp = oo->oo_owner.so_client;
|
|
|
|
goto alloc_stateid;
|
2011-10-13 08:44:20 +08:00
|
|
|
new_owner:
|
|
|
|
oo = alloc_init_open_stateowner(strhashval, clp, open);
|
|
|
|
if (oo == NULL)
|
|
|
|
return nfserr_jukebox;
|
|
|
|
open->op_openowner = oo;
|
2011-10-18 03:57:47 +08:00
|
|
|
alloc_stateid:
|
|
|
|
open->op_stp = nfs4_alloc_stateid(clp);
|
|
|
|
if (!open->op_stp)
|
|
|
|
return nfserr_jukebox;
|
2006-01-19 09:43:34 +08:00
|
|
|
return nfs_ok;
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
|
|
|
|
2006-10-20 14:28:59 +08:00
|
|
|
static inline __be32
|
2005-06-24 13:02:50 +08:00
|
|
|
nfs4_check_delegmode(struct nfs4_delegation *dp, int flags)
|
|
|
|
{
|
|
|
|
if ((flags & WR_STATE) && (dp->dl_type == NFS4_OPEN_DELEGATE_READ))
|
|
|
|
return nfserr_openmode;
|
|
|
|
else
|
|
|
|
return nfs_ok;
|
|
|
|
}
|
|
|
|
|
2011-09-09 21:06:12 +08:00
|
|
|
static int share_access_to_flags(u32 share_access)
|
2005-06-24 13:02:49 +08:00
|
|
|
{
|
2011-09-09 21:06:12 +08:00
|
|
|
return share_access == NFS4_SHARE_ACCESS_READ ? RD_STATE : WR_STATE;
|
2005-06-24 13:02:49 +08:00
|
|
|
}
|
|
|
|
|
2011-09-24 05:01:19 +08:00
|
|
|
static struct nfs4_delegation *find_deleg_stateid(struct nfs4_client *cl, stateid_t *s)
|
2010-05-19 08:01:35 +08:00
|
|
|
{
|
2011-09-09 21:06:12 +08:00
|
|
|
struct nfs4_stid *ret;
|
2010-05-19 08:01:35 +08:00
|
|
|
|
2011-09-24 05:01:19 +08:00
|
|
|
ret = find_stateid_by_type(cl, s, NFS4_DELEG_STID);
|
2011-09-09 21:06:12 +08:00
|
|
|
if (!ret)
|
|
|
|
return NULL;
|
|
|
|
return delegstateid(ret);
|
2010-05-19 08:01:35 +08:00
|
|
|
}
|
|
|
|
|
2011-10-19 23:52:12 +08:00
|
|
|
static bool nfsd4_is_deleg_cur(struct nfsd4_open *open)
|
|
|
|
{
|
|
|
|
return open->op_claim_type == NFS4_OPEN_CLAIM_DELEGATE_CUR ||
|
|
|
|
open->op_claim_type == NFS4_OPEN_CLAIM_DELEG_CUR_FH;
|
|
|
|
}
|
|
|
|
|
2006-10-20 14:28:59 +08:00
|
|
|
static __be32
|
2011-09-24 05:01:19 +08:00
|
|
|
nfs4_check_deleg(struct nfs4_client *cl, struct nfs4_file *fp, struct nfsd4_open *open,
|
2005-06-24 13:02:53 +08:00
|
|
|
struct nfs4_delegation **dp)
|
|
|
|
{
|
|
|
|
int flags;
|
2006-10-20 14:28:59 +08:00
|
|
|
__be32 status = nfserr_bad_stateid;
|
2005-06-24 13:02:53 +08:00
|
|
|
|
2011-09-24 05:01:19 +08:00
|
|
|
*dp = find_deleg_stateid(cl, &open->op_delegate_stateid);
|
2005-06-24 13:02:53 +08:00
|
|
|
if (*dp == NULL)
|
2005-06-24 13:02:54 +08:00
|
|
|
goto out;
|
2010-05-19 08:01:35 +08:00
|
|
|
flags = share_access_to_flags(open->op_share_access);
|
2005-06-24 13:02:53 +08:00
|
|
|
status = nfs4_check_delegmode(*dp, flags);
|
|
|
|
if (status)
|
|
|
|
*dp = NULL;
|
2005-06-24 13:02:54 +08:00
|
|
|
out:
|
2011-10-19 23:52:12 +08:00
|
|
|
if (!nfsd4_is_deleg_cur(open))
|
2005-06-24 13:02:54 +08:00
|
|
|
return nfs_ok;
|
|
|
|
if (status)
|
|
|
|
return status;
|
2011-09-13 00:24:13 +08:00
|
|
|
open->op_openowner->oo_flags |= NFS4_OO_CONFIRMED;
|
2005-06-24 13:02:54 +08:00
|
|
|
return nfs_ok;
|
2005-06-24 13:02:53 +08:00
|
|
|
}
|
|
|
|
|
2006-10-20 14:28:59 +08:00
|
|
|
static __be32
|
2011-09-08 04:06:42 +08:00
|
|
|
nfs4_check_open(struct nfs4_file *fp, struct nfsd4_open *open, struct nfs4_ol_stateid **stpp)
|
2005-04-17 06:20:36 +08:00
|
|
|
{
|
2011-09-08 04:06:42 +08:00
|
|
|
struct nfs4_ol_stateid *local;
|
2011-07-31 11:33:59 +08:00
|
|
|
struct nfs4_openowner *oo = open->op_openowner;
|
2005-04-17 06:20:36 +08:00
|
|
|
|
2005-06-24 13:03:08 +08:00
|
|
|
list_for_each_entry(local, &fp->fi_stateids, st_perfile) {
|
2005-04-17 06:20:36 +08:00
|
|
|
/* ignore lock owners */
|
|
|
|
if (local->st_stateowner->so_is_open_owner == 0)
|
|
|
|
continue;
|
|
|
|
/* remember if we have seen this open owner */
|
2011-07-31 11:33:59 +08:00
|
|
|
if (local->st_stateowner == &oo->oo_owner)
|
2005-04-17 06:20:36 +08:00
|
|
|
*stpp = local;
|
|
|
|
/* check for conflicting share reservations */
|
|
|
|
if (!test_share(local, open))
|
2011-09-03 00:08:20 +08:00
|
|
|
return nfserr_share_denied;
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
2011-09-03 00:08:20 +08:00
|
|
|
return nfs_ok;
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
|
|
|
|
2011-10-17 23:14:48 +08:00
|
|
|
static void nfs4_free_stateid(struct nfs4_ol_stateid *s)
|
|
|
|
{
|
|
|
|
kmem_cache_free(stateid_slab, s);
|
2005-06-24 13:03:03 +08:00
|
|
|
}
|
|
|
|
|
2010-07-29 00:21:23 +08:00
|
|
|
static inline int nfs4_access_to_access(u32 nfs4_access)
|
|
|
|
{
|
|
|
|
int flags = 0;
|
|
|
|
|
|
|
|
if (nfs4_access & NFS4_SHARE_ACCESS_READ)
|
|
|
|
flags |= NFSD_MAY_READ;
|
|
|
|
if (nfs4_access & NFS4_SHARE_ACCESS_WRITE)
|
|
|
|
flags |= NFSD_MAY_WRITE;
|
|
|
|
return flags;
|
|
|
|
}
|
|
|
|
|
2011-07-24 02:58:10 +08:00
|
|
|
static __be32 nfs4_get_vfs_file(struct svc_rqst *rqstp, struct nfs4_file *fp,
|
|
|
|
struct svc_fh *cur_fh, struct nfsd4_open *open)
|
2010-07-08 23:02:09 +08:00
|
|
|
{
|
|
|
|
__be32 status;
|
2011-07-24 02:58:10 +08:00
|
|
|
int oflag = nfs4_access_to_omode(open->op_share_access);
|
|
|
|
int access = nfs4_access_to_access(open->op_share_access);
|
|
|
|
|
2010-07-08 23:02:09 +08:00
|
|
|
if (!fp->fi_fds[oflag]) {
|
|
|
|
status = nfsd_open(rqstp, cur_fh, S_IFREG, access,
|
|
|
|
&fp->fi_fds[oflag]);
|
|
|
|
if (status)
|
|
|
|
return status;
|
|
|
|
}
|
|
|
|
nfs4_file_get_access(fp, oflag);
|
|
|
|
|
|
|
|
return nfs_ok;
|
|
|
|
}
|
|
|
|
|
2006-10-20 14:28:59 +08:00
|
|
|
static inline __be32
|
2005-04-17 06:20:36 +08:00
|
|
|
nfsd4_truncate(struct svc_rqst *rqstp, struct svc_fh *fh,
|
|
|
|
struct nfsd4_open *open)
|
|
|
|
{
|
|
|
|
struct iattr iattr = {
|
|
|
|
.ia_valid = ATTR_SIZE,
|
|
|
|
.ia_size = 0,
|
|
|
|
};
|
|
|
|
if (!open->op_truncate)
|
|
|
|
return 0;
|
|
|
|
if (!(open->op_share_access & NFS4_SHARE_ACCESS_WRITE))
|
2006-01-19 09:43:46 +08:00
|
|
|
return nfserr_inval;
|
2005-04-17 06:20:36 +08:00
|
|
|
return nfsd_setattr(rqstp, fh, &iattr, 0, (time_t)0);
|
|
|
|
}
|
|
|
|
|
2006-10-20 14:28:59 +08:00
|
|
|
static __be32
|
2011-09-08 04:06:42 +08:00
|
|
|
nfs4_upgrade_open(struct svc_rqst *rqstp, struct nfs4_file *fp, struct svc_fh *cur_fh, struct nfs4_ol_stateid *stp, struct nfsd4_open *open)
|
2005-04-17 06:20:36 +08:00
|
|
|
{
|
2011-10-11 05:44:19 +08:00
|
|
|
u32 op_share_access = open->op_share_access;
|
2010-08-21 06:09:31 +08:00
|
|
|
bool new_access;
|
2006-10-20 14:28:59 +08:00
|
|
|
__be32 status;
|
2005-04-17 06:20:36 +08:00
|
|
|
|
2010-08-21 06:09:31 +08:00
|
|
|
new_access = !test_bit(op_share_access, &stp->st_access_bmap);
|
2010-07-08 23:02:09 +08:00
|
|
|
if (new_access) {
|
2011-07-24 02:58:10 +08:00
|
|
|
status = nfs4_get_vfs_file(rqstp, fp, cur_fh, open);
|
2010-07-08 23:02:09 +08:00
|
|
|
if (status)
|
|
|
|
return status;
|
2006-01-19 09:43:38 +08:00
|
|
|
}
|
2005-04-17 06:20:36 +08:00
|
|
|
status = nfsd4_truncate(rqstp, cur_fh, open);
|
|
|
|
if (status) {
|
2010-07-08 23:02:09 +08:00
|
|
|
if (new_access) {
|
2011-06-29 20:23:50 +08:00
|
|
|
int oflag = nfs4_access_to_omode(op_share_access);
|
2010-07-08 23:02:09 +08:00
|
|
|
nfs4_file_put_access(fp, oflag);
|
|
|
|
}
|
2005-04-17 06:20:36 +08:00
|
|
|
return status;
|
|
|
|
}
|
|
|
|
/* remember the open */
|
2010-05-19 08:01:35 +08:00
|
|
|
__set_bit(op_share_access, &stp->st_access_bmap);
|
2008-04-29 06:22:50 +08:00
|
|
|
__set_bit(open->op_share_deny, &stp->st_deny_bmap);
|
2005-04-17 06:20:36 +08:00
|
|
|
|
|
|
|
return nfs_ok;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
static void
|
2012-03-07 03:35:16 +08:00
|
|
|
nfs4_set_claim_prev(struct nfsd4_open *open, bool has_session)
|
2005-04-17 06:20:36 +08:00
|
|
|
{
|
2011-09-13 00:24:13 +08:00
|
|
|
open->op_openowner->oo_flags |= NFS4_OO_CONFIRMED;
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
|
|
|
|
2010-12-11 08:02:49 +08:00
|
|
|
/* Should we give out recallable state?: */
|
|
|
|
static bool nfsd4_cb_channel_good(struct nfs4_client *clp)
|
|
|
|
{
|
|
|
|
if (clp->cl_cb_state == NFSD4_CB_UP)
|
|
|
|
return true;
|
|
|
|
/*
|
|
|
|
* In the sessions case, since we don't have to establish a
|
|
|
|
* separate connection for callbacks, we assume it's OK
|
|
|
|
* until we hear otherwise:
|
|
|
|
*/
|
|
|
|
return clp->cl_minorversion && clp->cl_cb_state == NFSD4_CB_UNKNOWN;
|
|
|
|
}
|
|
|
|
|
2011-02-01 00:55:12 +08:00
|
|
|
static struct file_lock *nfs4_alloc_init_lease(struct nfs4_delegation *dp, int flag)
|
|
|
|
{
|
|
|
|
struct file_lock *fl;
|
|
|
|
|
|
|
|
fl = locks_alloc_lock();
|
|
|
|
if (!fl)
|
|
|
|
return NULL;
|
|
|
|
locks_init_lock(fl);
|
|
|
|
fl->fl_lmops = &nfsd_lease_mng_ops;
|
|
|
|
fl->fl_flags = FL_LEASE;
|
|
|
|
fl->fl_type = flag == NFS4_OPEN_DELEGATE_READ? F_RDLCK: F_WRLCK;
|
|
|
|
fl->fl_end = OFFSET_MAX;
|
2011-02-01 08:20:39 +08:00
|
|
|
fl->fl_owner = (fl_owner_t)(dp->dl_file);
|
2011-02-01 00:55:12 +08:00
|
|
|
fl->fl_pid = current->tgid;
|
|
|
|
return fl;
|
|
|
|
}
|
|
|
|
|
2011-02-01 06:58:10 +08:00
|
|
|
static int nfs4_setlease(struct nfs4_delegation *dp, int flag)
|
|
|
|
{
|
2011-02-01 08:20:39 +08:00
|
|
|
struct nfs4_file *fp = dp->dl_file;
|
2011-02-01 06:58:10 +08:00
|
|
|
struct file_lock *fl;
|
|
|
|
int status;
|
|
|
|
|
|
|
|
fl = nfs4_alloc_init_lease(dp, flag);
|
|
|
|
if (!fl)
|
|
|
|
return -ENOMEM;
|
2011-02-01 08:20:39 +08:00
|
|
|
fl->fl_file = find_readable_file(fp);
|
2011-09-24 05:20:02 +08:00
|
|
|
list_add(&dp->dl_perclnt, &dp->dl_stid.sc_client->cl_delegations);
|
2011-02-01 08:20:39 +08:00
|
|
|
status = vfs_setlease(fl->fl_file, fl->fl_type, &fl);
|
2011-02-01 06:58:10 +08:00
|
|
|
if (status) {
|
2011-02-01 08:20:39 +08:00
|
|
|
list_del_init(&dp->dl_perclnt);
|
2011-02-01 06:58:10 +08:00
|
|
|
locks_free_lock(fl);
|
|
|
|
return -ENOMEM;
|
|
|
|
}
|
2011-02-01 08:20:39 +08:00
|
|
|
fp->fi_lease = fl;
|
|
|
|
fp->fi_deleg_file = fl->fl_file;
|
|
|
|
get_file(fp->fi_deleg_file);
|
|
|
|
atomic_set(&fp->fi_delegees, 1);
|
|
|
|
list_add(&dp->dl_perfile, &fp->fi_delegations);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int nfs4_set_delegation(struct nfs4_delegation *dp, int flag)
|
|
|
|
{
|
|
|
|
struct nfs4_file *fp = dp->dl_file;
|
|
|
|
|
|
|
|
if (!fp->fi_lease)
|
|
|
|
return nfs4_setlease(dp, flag);
|
|
|
|
spin_lock(&recall_lock);
|
|
|
|
if (fp->fi_had_conflict) {
|
|
|
|
spin_unlock(&recall_lock);
|
|
|
|
return -EAGAIN;
|
|
|
|
}
|
|
|
|
atomic_inc(&fp->fi_delegees);
|
|
|
|
list_add(&dp->dl_perfile, &fp->fi_delegations);
|
|
|
|
spin_unlock(&recall_lock);
|
2011-09-24 05:20:02 +08:00
|
|
|
list_add(&dp->dl_perclnt, &dp->dl_stid.sc_client->cl_delegations);
|
2011-02-01 06:58:10 +08:00
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2012-02-22 06:16:44 +08:00
|
|
|
static void nfsd4_open_deleg_none_ext(struct nfsd4_open *open, int status)
|
|
|
|
{
|
|
|
|
open->op_delegate_type = NFS4_OPEN_DELEGATE_NONE_EXT;
|
|
|
|
if (status == -EAGAIN)
|
|
|
|
open->op_why_no_deleg = WND4_CONTENTION;
|
|
|
|
else {
|
|
|
|
open->op_why_no_deleg = WND4_RESOURCE;
|
|
|
|
switch (open->op_deleg_want) {
|
|
|
|
case NFS4_SHARE_WANT_READ_DELEG:
|
|
|
|
case NFS4_SHARE_WANT_WRITE_DELEG:
|
|
|
|
case NFS4_SHARE_WANT_ANY_DELEG:
|
|
|
|
break;
|
|
|
|
case NFS4_SHARE_WANT_CANCEL:
|
|
|
|
open->op_why_no_deleg = WND4_CANCELLED;
|
|
|
|
break;
|
|
|
|
case NFS4_SHARE_WANT_NO_DELEG:
|
|
|
|
BUG(); /* not supposed to get here */
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2005-04-17 06:20:36 +08:00
|
|
|
/*
|
|
|
|
* Attempt to hand out a delegation.
|
|
|
|
*/
|
|
|
|
static void
|
2011-09-08 04:06:42 +08:00
|
|
|
nfs4_open_delegation(struct svc_fh *fh, struct nfsd4_open *open, struct nfs4_ol_stateid *stp)
|
2005-04-17 06:20:36 +08:00
|
|
|
{
|
|
|
|
struct nfs4_delegation *dp;
|
2011-07-31 11:33:59 +08:00
|
|
|
struct nfs4_openowner *oo = container_of(stp->st_stateowner, struct nfs4_openowner, oo_owner);
|
2010-12-11 08:02:49 +08:00
|
|
|
int cb_up;
|
2012-02-17 02:57:17 +08:00
|
|
|
int status = 0, flag = 0;
|
2005-04-17 06:20:36 +08:00
|
|
|
|
2011-07-31 11:33:59 +08:00
|
|
|
cb_up = nfsd4_cb_channel_good(oo->oo_owner.so_client);
|
2005-04-17 06:20:36 +08:00
|
|
|
flag = NFS4_OPEN_DELEGATE_NONE;
|
2005-06-24 13:03:23 +08:00
|
|
|
open->op_recall = 0;
|
|
|
|
switch (open->op_claim_type) {
|
|
|
|
case NFS4_OPEN_CLAIM_PREVIOUS:
|
2010-03-09 01:37:27 +08:00
|
|
|
if (!cb_up)
|
2005-06-24 13:03:23 +08:00
|
|
|
open->op_recall = 1;
|
|
|
|
flag = open->op_delegate_type;
|
|
|
|
if (flag == NFS4_OPEN_DELEGATE_NONE)
|
|
|
|
goto out;
|
|
|
|
break;
|
|
|
|
case NFS4_OPEN_CLAIM_NULL:
|
|
|
|
/* Let's not give out any delegations till everyone's
|
|
|
|
* had the chance to reclaim theirs.... */
|
2007-09-07 00:34:25 +08:00
|
|
|
if (locks_in_grace())
|
2005-06-24 13:03:23 +08:00
|
|
|
goto out;
|
2011-09-13 00:24:13 +08:00
|
|
|
if (!cb_up || !(oo->oo_flags & NFS4_OO_CONFIRMED))
|
2005-06-24 13:03:23 +08:00
|
|
|
goto out;
|
|
|
|
if (open->op_share_access & NFS4_SHARE_ACCESS_WRITE)
|
|
|
|
flag = NFS4_OPEN_DELEGATE_WRITE;
|
|
|
|
else
|
|
|
|
flag = NFS4_OPEN_DELEGATE_READ;
|
|
|
|
break;
|
|
|
|
default:
|
|
|
|
goto out;
|
|
|
|
}
|
2005-04-17 06:20:36 +08:00
|
|
|
|
2011-07-31 11:33:59 +08:00
|
|
|
dp = alloc_init_deleg(oo->oo_owner.so_client, stp, fh, flag);
|
2011-02-01 06:14:55 +08:00
|
|
|
if (dp == NULL)
|
|
|
|
goto out_no_deleg;
|
2011-02-01 08:20:39 +08:00
|
|
|
status = nfs4_set_delegation(dp, flag);
|
2011-02-01 06:58:10 +08:00
|
|
|
if (status)
|
2011-02-01 06:14:55 +08:00
|
|
|
goto out_free;
|
2005-04-17 06:20:36 +08:00
|
|
|
|
2011-09-09 00:07:44 +08:00
|
|
|
memcpy(&open->op_delegate_stateid, &dp->dl_stid.sc_stateid, sizeof(dp->dl_stid.sc_stateid));
|
2005-04-17 06:20:36 +08:00
|
|
|
|
2009-10-19 18:04:53 +08:00
|
|
|
dprintk("NFSD: delegation stateid=" STATEID_FMT "\n",
|
2011-09-09 00:07:44 +08:00
|
|
|
STATEID_VAL(&dp->dl_stid.sc_stateid));
|
2005-04-17 06:20:36 +08:00
|
|
|
out:
|
|
|
|
open->op_delegate_type = flag;
|
2012-02-17 02:57:17 +08:00
|
|
|
if (flag == NFS4_OPEN_DELEGATE_NONE) {
|
|
|
|
if (open->op_claim_type == NFS4_OPEN_CLAIM_PREVIOUS &&
|
|
|
|
open->op_delegate_type != NFS4_OPEN_DELEGATE_NONE)
|
|
|
|
dprintk("NFSD: WARNING: refusing delegation reclaim\n");
|
|
|
|
|
2012-02-22 06:16:44 +08:00
|
|
|
/* 4.1 client asking for a delegation? */
|
|
|
|
if (open->op_deleg_want)
|
|
|
|
nfsd4_open_deleg_none_ext(open, status);
|
2012-02-17 02:57:17 +08:00
|
|
|
}
|
2011-02-01 06:14:55 +08:00
|
|
|
return;
|
|
|
|
out_free:
|
2011-02-01 08:20:39 +08:00
|
|
|
nfs4_put_delegation(dp);
|
2011-02-01 06:14:55 +08:00
|
|
|
out_no_deleg:
|
|
|
|
flag = NFS4_OPEN_DELEGATE_NONE;
|
|
|
|
goto out;
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
|
|
|
|
2012-02-22 06:16:54 +08:00
|
|
|
static void nfsd4_deleg_xgrade_none_ext(struct nfsd4_open *open,
|
|
|
|
struct nfs4_delegation *dp)
|
|
|
|
{
|
|
|
|
if (open->op_deleg_want == NFS4_SHARE_WANT_READ_DELEG &&
|
|
|
|
dp->dl_type == NFS4_OPEN_DELEGATE_WRITE) {
|
|
|
|
open->op_delegate_type = NFS4_OPEN_DELEGATE_NONE_EXT;
|
|
|
|
open->op_why_no_deleg = WND4_NOT_SUPP_DOWNGRADE;
|
|
|
|
} else if (open->op_deleg_want == NFS4_SHARE_WANT_WRITE_DELEG &&
|
|
|
|
dp->dl_type == NFS4_OPEN_DELEGATE_WRITE) {
|
|
|
|
open->op_delegate_type = NFS4_OPEN_DELEGATE_NONE_EXT;
|
|
|
|
open->op_why_no_deleg = WND4_NOT_SUPP_UPGRADE;
|
|
|
|
}
|
|
|
|
/* Otherwise the client must be confused wanting a delegation
|
|
|
|
* it already has, therefore we don't return
|
|
|
|
* NFS4_OPEN_DELEGATE_NONE_EXT and reason.
|
|
|
|
*/
|
|
|
|
}
|
|
|
|
|
2005-04-17 06:20:36 +08:00
|
|
|
/*
|
|
|
|
* called with nfs4_lock_state() held.
|
|
|
|
*/
|
2006-10-20 14:28:59 +08:00
|
|
|
__be32
|
2005-04-17 06:20:36 +08:00
|
|
|
nfsd4_process_open2(struct svc_rqst *rqstp, struct svc_fh *current_fh, struct nfsd4_open *open)
|
|
|
|
{
|
2009-04-03 13:28:45 +08:00
|
|
|
struct nfsd4_compoundres *resp = rqstp->rq_resp;
|
2011-09-24 05:01:19 +08:00
|
|
|
struct nfs4_client *cl = open->op_openowner->oo_owner.so_client;
|
2005-04-17 06:20:36 +08:00
|
|
|
struct nfs4_file *fp = NULL;
|
|
|
|
struct inode *ino = current_fh->fh_dentry->d_inode;
|
2011-09-08 04:06:42 +08:00
|
|
|
struct nfs4_ol_stateid *stp = NULL;
|
2005-06-24 13:02:53 +08:00
|
|
|
struct nfs4_delegation *dp = NULL;
|
2006-10-20 14:28:59 +08:00
|
|
|
__be32 status;
|
2005-04-17 06:20:36 +08:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Lookup file; if found, lookup stateid and check open request,
|
|
|
|
* and check for delegations in the process of being recalled.
|
|
|
|
* If not found, create the nfs4_file struct
|
|
|
|
*/
|
|
|
|
fp = find_file(ino);
|
|
|
|
if (fp) {
|
|
|
|
if ((status = nfs4_check_open(fp, open, &stp)))
|
|
|
|
goto out;
|
2011-09-24 05:01:19 +08:00
|
|
|
status = nfs4_check_deleg(cl, fp, open, &dp);
|
2005-06-24 13:02:54 +08:00
|
|
|
if (status)
|
|
|
|
goto out;
|
2005-04-17 06:20:36 +08:00
|
|
|
} else {
|
2005-06-24 13:02:54 +08:00
|
|
|
status = nfserr_bad_stateid;
|
2011-10-19 23:52:12 +08:00
|
|
|
if (nfsd4_is_deleg_cur(open))
|
2005-06-24 13:02:54 +08:00
|
|
|
goto out;
|
2011-08-11 07:07:33 +08:00
|
|
|
status = nfserr_jukebox;
|
2011-10-14 04:00:16 +08:00
|
|
|
fp = open->op_file;
|
|
|
|
open->op_file = NULL;
|
|
|
|
nfsd4_init_file(fp, ino);
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* OPEN the file, or upgrade an existing OPEN.
|
|
|
|
* If truncate fails, the OPEN fails.
|
|
|
|
*/
|
|
|
|
if (stp) {
|
|
|
|
/* Stateid was found, this is an OPEN upgrade */
|
2010-07-08 23:02:09 +08:00
|
|
|
status = nfs4_upgrade_open(rqstp, fp, current_fh, stp, open);
|
2005-04-17 06:20:36 +08:00
|
|
|
if (status)
|
|
|
|
goto out;
|
|
|
|
} else {
|
2011-10-18 03:57:47 +08:00
|
|
|
status = nfs4_get_vfs_file(rqstp, fp, current_fh, open);
|
2005-06-24 13:02:53 +08:00
|
|
|
if (status)
|
2005-04-17 06:20:36 +08:00
|
|
|
goto out;
|
2011-10-18 03:57:47 +08:00
|
|
|
stp = open->op_stp;
|
|
|
|
open->op_stp = NULL;
|
2011-10-17 23:14:48 +08:00
|
|
|
init_open_stateid(stp, fp, open);
|
2005-04-17 06:20:36 +08:00
|
|
|
status = nfsd4_truncate(rqstp, current_fh, open);
|
|
|
|
if (status) {
|
2009-01-12 03:27:17 +08:00
|
|
|
release_open_stateid(stp);
|
2005-04-17 06:20:36 +08:00
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
}
|
2011-09-08 04:06:42 +08:00
|
|
|
update_stateid(&stp->st_stid.sc_stateid);
|
|
|
|
memcpy(&open->op_stateid, &stp->st_stid.sc_stateid, sizeof(stateid_t));
|
2005-04-17 06:20:36 +08:00
|
|
|
|
2012-02-17 02:57:17 +08:00
|
|
|
if (nfsd4_has_session(&resp->cstate)) {
|
2011-09-13 00:24:13 +08:00
|
|
|
open->op_openowner->oo_flags |= NFS4_OO_CONFIRMED;
|
2009-04-03 13:28:45 +08:00
|
|
|
|
2012-02-17 02:57:17 +08:00
|
|
|
if (open->op_deleg_want & NFS4_SHARE_WANT_NO_DELEG) {
|
|
|
|
open->op_delegate_type = NFS4_OPEN_DELEGATE_NONE_EXT;
|
|
|
|
open->op_why_no_deleg = WND4_NOT_WANTED;
|
|
|
|
goto nodeleg;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2005-04-17 06:20:36 +08:00
|
|
|
/*
|
|
|
|
* Attempt to hand out a delegation. No error return, because the
|
|
|
|
* OPEN succeeds even if we fail.
|
|
|
|
*/
|
|
|
|
nfs4_open_delegation(current_fh, open, stp);
|
2012-02-17 02:57:17 +08:00
|
|
|
nodeleg:
|
2005-04-17 06:20:36 +08:00
|
|
|
status = nfs_ok;
|
|
|
|
|
2009-10-19 18:04:53 +08:00
|
|
|
dprintk("%s: stateid=" STATEID_FMT "\n", __func__,
|
2011-09-08 04:06:42 +08:00
|
|
|
STATEID_VAL(&stp->st_stid.sc_stateid));
|
2005-04-17 06:20:36 +08:00
|
|
|
out:
|
2012-02-17 02:57:17 +08:00
|
|
|
/* 4.1 client trying to upgrade/downgrade delegation? */
|
|
|
|
if (open->op_delegate_type == NFS4_OPEN_DELEGATE_NONE && dp &&
|
2012-02-22 06:16:54 +08:00
|
|
|
open->op_deleg_want)
|
|
|
|
nfsd4_deleg_xgrade_none_ext(open, dp);
|
2012-02-17 02:57:17 +08:00
|
|
|
|
2005-06-24 13:03:10 +08:00
|
|
|
if (fp)
|
|
|
|
put_nfs4_file(fp);
|
2005-07-08 08:59:16 +08:00
|
|
|
if (status == 0 && open->op_claim_type == NFS4_OPEN_CLAIM_PREVIOUS)
|
2012-03-07 03:35:16 +08:00
|
|
|
nfs4_set_claim_prev(open, nfsd4_has_session(&resp->cstate));
|
2005-04-17 06:20:36 +08:00
|
|
|
/*
|
|
|
|
* To finish the open response, we just need to set the rflags.
|
|
|
|
*/
|
|
|
|
open->op_rflags = NFS4_OPEN_RESULT_LOCKTYPE_POSIX;
|
2011-09-13 00:24:13 +08:00
|
|
|
if (!(open->op_openowner->oo_flags & NFS4_OO_CONFIRMED) &&
|
2009-04-03 13:28:45 +08:00
|
|
|
!nfsd4_has_session(&resp->cstate))
|
2005-04-17 06:20:36 +08:00
|
|
|
open->op_rflags |= NFS4_OPEN_RESULT_CONFIRM;
|
|
|
|
|
|
|
|
return status;
|
|
|
|
}
|
|
|
|
|
2011-10-14 03:12:59 +08:00
|
|
|
void nfsd4_cleanup_open_state(struct nfsd4_open *open, __be32 status)
|
|
|
|
{
|
|
|
|
if (open->op_openowner) {
|
|
|
|
struct nfs4_openowner *oo = open->op_openowner;
|
|
|
|
|
|
|
|
if (!list_empty(&oo->oo_owner.so_stateids))
|
|
|
|
list_del_init(&oo->oo_close_lru);
|
|
|
|
if (oo->oo_flags & NFS4_OO_NEW) {
|
|
|
|
if (status) {
|
|
|
|
release_openowner(oo);
|
|
|
|
open->op_openowner = NULL;
|
|
|
|
} else
|
|
|
|
oo->oo_flags &= ~NFS4_OO_NEW;
|
|
|
|
}
|
|
|
|
}
|
2011-10-14 04:00:16 +08:00
|
|
|
if (open->op_file)
|
|
|
|
nfsd4_free_file(open->op_file);
|
2011-10-18 03:57:47 +08:00
|
|
|
if (open->op_stp)
|
|
|
|
nfs4_free_stateid(open->op_stp);
|
2011-10-14 03:12:59 +08:00
|
|
|
}
|
|
|
|
|
2006-10-20 14:28:59 +08:00
|
|
|
__be32
|
2006-12-13 16:35:38 +08:00
|
|
|
nfsd4_renew(struct svc_rqst *rqstp, struct nfsd4_compound_state *cstate,
|
|
|
|
clientid_t *clid)
|
2005-04-17 06:20:36 +08:00
|
|
|
{
|
|
|
|
struct nfs4_client *clp;
|
2006-10-20 14:28:59 +08:00
|
|
|
__be32 status;
|
2005-04-17 06:20:36 +08:00
|
|
|
|
|
|
|
nfs4_lock_state();
|
|
|
|
dprintk("process_renew(%08x/%08x): starting\n",
|
|
|
|
clid->cl_boot, clid->cl_id);
|
|
|
|
status = nfserr_stale_clientid;
|
|
|
|
if (STALE_CLIENTID(clid))
|
|
|
|
goto out;
|
|
|
|
clp = find_confirmed_client(clid);
|
|
|
|
status = nfserr_expired;
|
|
|
|
if (clp == NULL) {
|
|
|
|
/* We assume the client took too long to RENEW. */
|
|
|
|
dprintk("nfsd4_renew: clientid not found!\n");
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
status = nfserr_cb_path_down;
|
2005-06-24 13:04:17 +08:00
|
|
|
if (!list_empty(&clp->cl_delegations)
|
2010-05-01 06:51:44 +08:00
|
|
|
&& clp->cl_cb_state != NFSD4_CB_UP)
|
2005-04-17 06:20:36 +08:00
|
|
|
goto out;
|
|
|
|
status = nfs_ok;
|
|
|
|
out:
|
|
|
|
nfs4_unlock_state();
|
|
|
|
return status;
|
|
|
|
}
|
|
|
|
|
2011-05-16 22:38:14 +08:00
|
|
|
static struct lock_manager nfsd4_manager = {
|
2007-09-07 00:34:25 +08:00
|
|
|
};
|
|
|
|
|
2012-04-10 23:08:48 +08:00
|
|
|
static bool grace_ended;
|
|
|
|
|
2005-06-24 13:04:01 +08:00
|
|
|
static void
|
2007-09-07 00:34:25 +08:00
|
|
|
nfsd4_end_grace(void)
|
2005-06-24 13:04:01 +08:00
|
|
|
{
|
2012-04-10 23:08:48 +08:00
|
|
|
/* do nothing if grace period already ended */
|
|
|
|
if (grace_ended)
|
|
|
|
return;
|
|
|
|
|
2005-06-24 13:04:01 +08:00
|
|
|
dprintk("NFSD: end of grace period\n");
|
2012-04-10 23:08:48 +08:00
|
|
|
grace_ended = true;
|
2012-03-22 04:42:43 +08:00
|
|
|
nfsd4_record_grace_done(&init_net, boot_time);
|
2007-09-07 00:34:25 +08:00
|
|
|
locks_end_grace(&nfsd4_manager);
|
2010-03-02 08:21:21 +08:00
|
|
|
/*
|
|
|
|
* Now that every NFSv4 client has had the chance to recover and
|
|
|
|
* to see the (possibly new, possibly shorter) lease time, we
|
|
|
|
* can safely set the next grace time to the current lease time:
|
|
|
|
*/
|
|
|
|
nfsd4_grace = nfsd4_lease;
|
2005-06-24 13:04:01 +08:00
|
|
|
}
|
|
|
|
|
2005-06-24 13:04:03 +08:00
|
|
|
static time_t
|
2005-04-17 06:20:36 +08:00
|
|
|
nfs4_laundromat(void)
|
|
|
|
{
|
|
|
|
struct nfs4_client *clp;
|
2011-07-31 11:33:59 +08:00
|
|
|
struct nfs4_openowner *oo;
|
2005-04-17 06:20:36 +08:00
|
|
|
struct nfs4_delegation *dp;
|
|
|
|
struct list_head *pos, *next, reaplist;
|
2010-03-01 12:20:19 +08:00
|
|
|
time_t cutoff = get_seconds() - nfsd4_lease;
|
|
|
|
time_t t, clientid_val = nfsd4_lease;
|
|
|
|
time_t u, test_val = nfsd4_lease;
|
2005-04-17 06:20:36 +08:00
|
|
|
|
|
|
|
nfs4_lock_state();
|
|
|
|
|
|
|
|
dprintk("NFSD: laundromat service - starting\n");
|
2012-04-10 23:08:48 +08:00
|
|
|
nfsd4_end_grace();
|
2010-05-12 05:13:04 +08:00
|
|
|
INIT_LIST_HEAD(&reaplist);
|
|
|
|
spin_lock(&client_lock);
|
2005-04-17 06:20:36 +08:00
|
|
|
list_for_each_safe(pos, next, &client_lru) {
|
|
|
|
clp = list_entry(pos, struct nfs4_client, cl_lru);
|
|
|
|
if (time_after((unsigned long)clp->cl_time, (unsigned long)cutoff)) {
|
|
|
|
t = clp->cl_time - cutoff;
|
|
|
|
if (clientid_val > t)
|
|
|
|
clientid_val = t;
|
|
|
|
break;
|
|
|
|
}
|
2010-05-12 05:13:54 +08:00
|
|
|
if (atomic_read(&clp->cl_refcount)) {
|
|
|
|
dprintk("NFSD: client in use (clientid %08x)\n",
|
|
|
|
clp->cl_clientid.cl_id);
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
unhash_client_locked(clp);
|
|
|
|
list_add(&clp->cl_lru, &reaplist);
|
2010-05-12 05:13:04 +08:00
|
|
|
}
|
|
|
|
spin_unlock(&client_lock);
|
|
|
|
list_for_each_safe(pos, next, &reaplist) {
|
|
|
|
clp = list_entry(pos, struct nfs4_client, cl_lru);
|
2005-04-17 06:20:36 +08:00
|
|
|
dprintk("NFSD: purging unused client (clientid %08x)\n",
|
|
|
|
clp->cl_clientid.cl_id);
|
2012-03-22 04:42:43 +08:00
|
|
|
nfsd4_client_record_remove(clp);
|
2005-04-17 06:20:36 +08:00
|
|
|
expire_client(clp);
|
|
|
|
}
|
|
|
|
spin_lock(&recall_lock);
|
|
|
|
list_for_each_safe(pos, next, &del_recall_lru) {
|
|
|
|
dp = list_entry (pos, struct nfs4_delegation, dl_recall_lru);
|
|
|
|
if (time_after((unsigned long)dp->dl_time, (unsigned long)cutoff)) {
|
|
|
|
u = dp->dl_time - cutoff;
|
|
|
|
if (test_val > u)
|
|
|
|
test_val = u;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
list_move(&dp->dl_recall_lru, &reaplist);
|
|
|
|
}
|
|
|
|
spin_unlock(&recall_lock);
|
|
|
|
list_for_each_safe(pos, next, &reaplist) {
|
|
|
|
dp = list_entry (pos, struct nfs4_delegation, dl_recall_lru);
|
|
|
|
unhash_delegation(dp);
|
|
|
|
}
|
2010-03-01 12:20:19 +08:00
|
|
|
test_val = nfsd4_lease;
|
2005-04-17 06:20:36 +08:00
|
|
|
list_for_each_safe(pos, next, &close_lru) {
|
2011-07-31 11:33:59 +08:00
|
|
|
oo = container_of(pos, struct nfs4_openowner, oo_close_lru);
|
|
|
|
if (time_after((unsigned long)oo->oo_time, (unsigned long)cutoff)) {
|
|
|
|
u = oo->oo_time - cutoff;
|
2005-04-17 06:20:36 +08:00
|
|
|
if (test_val > u)
|
|
|
|
test_val = u;
|
|
|
|
break;
|
|
|
|
}
|
2011-07-31 11:33:59 +08:00
|
|
|
release_openowner(oo);
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
|
|
|
if (clientid_val < NFSD_LAUNDROMAT_MINTIMEOUT)
|
|
|
|
clientid_val = NFSD_LAUNDROMAT_MINTIMEOUT;
|
|
|
|
nfs4_unlock_state();
|
|
|
|
return clientid_val;
|
|
|
|
}
|
|
|
|
|
2008-02-21 04:49:00 +08:00
|
|
|
static struct workqueue_struct *laundry_wq;
|
|
|
|
static void laundromat_main(struct work_struct *);
|
|
|
|
static DECLARE_DELAYED_WORK(laundromat_work, laundromat_main);
|
|
|
|
|
|
|
|
static void
|
2006-11-22 22:57:56 +08:00
|
|
|
laundromat_main(struct work_struct *not_used)
|
2005-04-17 06:20:36 +08:00
|
|
|
{
|
|
|
|
time_t t;
|
|
|
|
|
|
|
|
t = nfs4_laundromat();
|
|
|
|
dprintk("NFSD: laundromat_main - sleeping for %ld seconds\n", t);
|
2005-06-24 13:03:19 +08:00
|
|
|
queue_delayed_work(laundry_wq, &laundromat_work, t*HZ);
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
|
|
|
|
2011-09-17 08:12:38 +08:00
|
|
|
static inline __be32 nfs4_check_fh(struct svc_fh *fhp, struct nfs4_ol_stateid *stp)
|
2005-04-17 06:20:36 +08:00
|
|
|
{
|
2011-09-17 08:12:38 +08:00
|
|
|
if (fhp->fh_dentry->d_inode != stp->st_file->fi_inode)
|
|
|
|
return nfserr_bad_stateid;
|
|
|
|
return nfs_ok;
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
static int
|
|
|
|
STALE_STATEID(stateid_t *stateid)
|
|
|
|
{
|
2011-09-16 03:02:41 +08:00
|
|
|
if (stateid->si_opaque.so_clid.cl_boot == boot_time)
|
2010-04-23 04:21:39 +08:00
|
|
|
return 0;
|
|
|
|
dprintk("NFSD: stale stateid " STATEID_FMT "!\n",
|
2009-10-19 18:04:53 +08:00
|
|
|
STATEID_VAL(stateid));
|
2010-04-23 04:21:39 +08:00
|
|
|
return 1;
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
static inline int
|
|
|
|
access_permit_read(unsigned long access_bmap)
|
|
|
|
{
|
|
|
|
return test_bit(NFS4_SHARE_ACCESS_READ, &access_bmap) ||
|
|
|
|
test_bit(NFS4_SHARE_ACCESS_BOTH, &access_bmap) ||
|
|
|
|
test_bit(NFS4_SHARE_ACCESS_WRITE, &access_bmap);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline int
|
|
|
|
access_permit_write(unsigned long access_bmap)
|
|
|
|
{
|
|
|
|
return test_bit(NFS4_SHARE_ACCESS_WRITE, &access_bmap) ||
|
|
|
|
test_bit(NFS4_SHARE_ACCESS_BOTH, &access_bmap);
|
|
|
|
}
|
|
|
|
|
|
|
|
static
|
2011-09-08 04:06:42 +08:00
|
|
|
__be32 nfs4_check_openmode(struct nfs4_ol_stateid *stp, int flags)
|
2005-04-17 06:20:36 +08:00
|
|
|
{
|
2006-10-20 14:28:59 +08:00
|
|
|
__be32 status = nfserr_openmode;
|
2005-04-17 06:20:36 +08:00
|
|
|
|
2010-07-30 03:16:59 +08:00
|
|
|
/* For lock stateid's, we test the parent open, not the lock: */
|
|
|
|
if (stp->st_openstp)
|
|
|
|
stp = stp->st_openstp;
|
2005-04-17 06:20:36 +08:00
|
|
|
if ((flags & WR_STATE) && (!access_permit_write(stp->st_access_bmap)))
|
|
|
|
goto out;
|
|
|
|
if ((flags & RD_STATE) && (!access_permit_read(stp->st_access_bmap)))
|
|
|
|
goto out;
|
|
|
|
status = nfs_ok;
|
|
|
|
out:
|
|
|
|
return status;
|
|
|
|
}
|
|
|
|
|
2006-10-20 14:28:59 +08:00
|
|
|
static inline __be32
|
2005-04-17 06:20:36 +08:00
|
|
|
check_special_stateids(svc_fh *current_fh, stateid_t *stateid, int flags)
|
|
|
|
{
|
2009-02-22 05:29:14 +08:00
|
|
|
if (ONE_STATEID(stateid) && (flags & RD_STATE))
|
2005-04-17 06:20:36 +08:00
|
|
|
return nfs_ok;
|
2007-09-07 00:34:25 +08:00
|
|
|
else if (locks_in_grace()) {
|
2011-03-31 09:57:33 +08:00
|
|
|
/* Answer in remaining cases depends on existence of
|
2005-04-17 06:20:36 +08:00
|
|
|
* conflicting state; so we must wait out the grace period. */
|
|
|
|
return nfserr_grace;
|
|
|
|
} else if (flags & WR_STATE)
|
|
|
|
return nfs4_share_conflict(current_fh,
|
|
|
|
NFS4_SHARE_DENY_WRITE);
|
|
|
|
else /* (flags & RD_STATE) && ZERO_STATEID(stateid) */
|
|
|
|
return nfs4_share_conflict(current_fh,
|
|
|
|
NFS4_SHARE_DENY_READ);
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Allow READ/WRITE during grace period on recovered state only for files
|
|
|
|
* that are not able to provide mandatory locking.
|
|
|
|
*/
|
|
|
|
static inline int
|
2009-02-22 07:23:01 +08:00
|
|
|
grace_disallows_io(struct inode *inode)
|
2005-04-17 06:20:36 +08:00
|
|
|
{
|
2009-02-22 05:29:14 +08:00
|
|
|
return locks_in_grace() && mandatory_lock(inode);
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
|
|
|
|
2011-08-23 23:03:29 +08:00
|
|
|
/* Returns true iff a is later than b: */
|
|
|
|
static bool stateid_generation_after(stateid_t *a, stateid_t *b)
|
|
|
|
{
|
|
|
|
return (s32)a->si_generation - (s32)b->si_generation > 0;
|
|
|
|
}
|
|
|
|
|
2011-08-22 22:07:12 +08:00
|
|
|
static int check_stateid_generation(stateid_t *in, stateid_t *ref, bool has_session)
|
2008-01-27 08:08:12 +08:00
|
|
|
{
|
2009-04-03 13:28:45 +08:00
|
|
|
/*
|
|
|
|
* When sessions are used the stateid generation number is ignored
|
|
|
|
* when it is zero.
|
|
|
|
*/
|
2011-08-22 22:07:12 +08:00
|
|
|
if (has_session && in->si_generation == 0)
|
2011-08-23 23:03:29 +08:00
|
|
|
return nfs_ok;
|
|
|
|
|
|
|
|
if (in->si_generation == ref->si_generation)
|
|
|
|
return nfs_ok;
|
2009-04-03 13:28:45 +08:00
|
|
|
|
2008-01-27 08:08:12 +08:00
|
|
|
/* If the client sends us a stateid from the future, it's buggy: */
|
2011-08-23 23:03:29 +08:00
|
|
|
if (stateid_generation_after(in, ref))
|
2008-01-27 08:08:12 +08:00
|
|
|
return nfserr_bad_stateid;
|
|
|
|
/*
|
2011-08-23 23:03:29 +08:00
|
|
|
* However, we could see a stateid from the past, even from a
|
|
|
|
* non-buggy client. For example, if the client sends a lock
|
|
|
|
* while some IO is outstanding, the lock may bump si_generation
|
|
|
|
* while the IO is still in flight. The client could avoid that
|
|
|
|
* situation by waiting for responses on all the IO requests,
|
|
|
|
* but better performance may result in retrying IO that
|
|
|
|
* receives an old_stateid error if requests are rarely
|
|
|
|
* reordered in flight:
|
2008-01-27 08:08:12 +08:00
|
|
|
*/
|
2011-08-23 23:03:29 +08:00
|
|
|
return nfserr_old_stateid;
|
2008-01-27 08:08:12 +08:00
|
|
|
}
|
|
|
|
|
2011-09-24 05:01:19 +08:00
|
|
|
__be32 nfs4_validate_stateid(struct nfs4_client *cl, stateid_t *stateid)
|
2011-07-13 22:50:48 +08:00
|
|
|
{
|
2011-09-09 23:26:58 +08:00
|
|
|
struct nfs4_stid *s;
|
|
|
|
struct nfs4_ol_stateid *ols;
|
|
|
|
__be32 status;
|
2011-07-13 22:50:48 +08:00
|
|
|
|
|
|
|
if (STALE_STATEID(stateid))
|
2011-09-09 23:26:58 +08:00
|
|
|
return nfserr_stale_stateid;
|
2011-07-13 22:50:48 +08:00
|
|
|
|
2011-09-24 05:01:19 +08:00
|
|
|
s = find_stateid(cl, stateid);
|
2011-09-09 23:26:58 +08:00
|
|
|
if (!s)
|
|
|
|
return nfserr_stale_stateid;
|
2011-09-27 00:53:00 +08:00
|
|
|
status = check_stateid_generation(stateid, &s->sc_stateid, 1);
|
2011-07-13 22:50:48 +08:00
|
|
|
if (status)
|
2011-09-09 23:26:58 +08:00
|
|
|
return status;
|
|
|
|
if (!(s->sc_type & (NFS4_OPEN_STID | NFS4_LOCK_STID)))
|
|
|
|
return nfs_ok;
|
|
|
|
ols = openlockstateid(s);
|
|
|
|
if (ols->st_stateowner->so_is_open_owner
|
2011-09-13 00:24:13 +08:00
|
|
|
&& !(openowner(ols->st_stateowner)->oo_flags & NFS4_OO_CONFIRMED))
|
2011-09-09 23:26:58 +08:00
|
|
|
return nfserr_bad_stateid;
|
|
|
|
return nfs_ok;
|
2011-07-13 22:50:48 +08:00
|
|
|
}
|
|
|
|
|
2011-09-24 05:01:19 +08:00
|
|
|
static __be32 nfsd4_lookup_stateid(stateid_t *stateid, unsigned char typemask, struct nfs4_stid **s)
|
|
|
|
{
|
|
|
|
struct nfs4_client *cl;
|
|
|
|
|
|
|
|
if (ZERO_STATEID(stateid) || ONE_STATEID(stateid))
|
|
|
|
return nfserr_bad_stateid;
|
|
|
|
if (STALE_STATEID(stateid))
|
|
|
|
return nfserr_stale_stateid;
|
|
|
|
cl = find_confirmed_client(&stateid->si_opaque.so_clid);
|
|
|
|
if (!cl)
|
|
|
|
return nfserr_expired;
|
|
|
|
*s = find_stateid_by_type(cl, stateid, typemask);
|
|
|
|
if (!*s)
|
|
|
|
return nfserr_bad_stateid;
|
|
|
|
return nfs_ok;
|
|
|
|
|
|
|
|
}
|
|
|
|
|
2005-04-17 06:20:36 +08:00
|
|
|
/*
|
|
|
|
* Checks for stateid operations
|
|
|
|
*/
|
2006-10-20 14:28:59 +08:00
|
|
|
__be32
|
2009-04-03 13:28:41 +08:00
|
|
|
nfs4_preprocess_stateid_op(struct nfsd4_compound_state *cstate,
|
|
|
|
stateid_t *stateid, int flags, struct file **filpp)
|
2005-04-17 06:20:36 +08:00
|
|
|
{
|
2011-09-09 23:54:57 +08:00
|
|
|
struct nfs4_stid *s;
|
2011-09-08 04:06:42 +08:00
|
|
|
struct nfs4_ol_stateid *stp = NULL;
|
2005-04-17 06:20:36 +08:00
|
|
|
struct nfs4_delegation *dp = NULL;
|
2009-04-03 13:28:41 +08:00
|
|
|
struct svc_fh *current_fh = &cstate->current_fh;
|
2005-04-17 06:20:36 +08:00
|
|
|
struct inode *ino = current_fh->fh_dentry->d_inode;
|
2006-10-20 14:28:59 +08:00
|
|
|
__be32 status;
|
2005-04-17 06:20:36 +08:00
|
|
|
|
|
|
|
if (filpp)
|
|
|
|
*filpp = NULL;
|
|
|
|
|
2009-02-22 07:23:01 +08:00
|
|
|
if (grace_disallows_io(ino))
|
2005-04-17 06:20:36 +08:00
|
|
|
return nfserr_grace;
|
|
|
|
|
|
|
|
if (ZERO_STATEID(stateid) || ONE_STATEID(stateid))
|
|
|
|
return check_special_stateids(current_fh, stateid, flags);
|
|
|
|
|
2011-09-24 05:01:19 +08:00
|
|
|
status = nfsd4_lookup_stateid(stateid, NFS4_DELEG_STID|NFS4_OPEN_STID|NFS4_LOCK_STID, &s);
|
|
|
|
if (status)
|
|
|
|
return status;
|
2011-09-09 23:54:57 +08:00
|
|
|
status = check_stateid_generation(stateid, &s->sc_stateid, nfsd4_has_session(cstate));
|
|
|
|
if (status)
|
|
|
|
goto out;
|
2011-09-17 08:12:38 +08:00
|
|
|
switch (s->sc_type) {
|
|
|
|
case NFS4_DELEG_STID:
|
2011-09-09 23:54:57 +08:00
|
|
|
dp = delegstateid(s);
|
2009-02-22 03:14:43 +08:00
|
|
|
status = nfs4_check_delegmode(dp, flags);
|
|
|
|
if (status)
|
|
|
|
goto out;
|
2010-10-28 05:19:04 +08:00
|
|
|
if (filpp) {
|
2011-02-01 08:20:39 +08:00
|
|
|
*filpp = dp->dl_file->fi_deleg_file;
|
2010-10-28 05:19:04 +08:00
|
|
|
BUG_ON(!*filpp);
|
|
|
|
}
|
2011-09-17 08:12:38 +08:00
|
|
|
break;
|
|
|
|
case NFS4_OPEN_STID:
|
|
|
|
case NFS4_LOCK_STID:
|
2011-09-09 23:54:57 +08:00
|
|
|
stp = openlockstateid(s);
|
2011-09-17 08:12:38 +08:00
|
|
|
status = nfs4_check_fh(current_fh, stp);
|
|
|
|
if (status)
|
2005-04-17 06:20:36 +08:00
|
|
|
goto out;
|
2011-07-31 11:33:59 +08:00
|
|
|
if (stp->st_stateowner->so_is_open_owner
|
2011-09-13 00:24:13 +08:00
|
|
|
&& !(openowner(stp->st_stateowner)->oo_flags & NFS4_OO_CONFIRMED))
|
2005-04-17 06:20:36 +08:00
|
|
|
goto out;
|
2009-02-22 02:40:22 +08:00
|
|
|
status = nfs4_check_openmode(stp, flags);
|
|
|
|
if (status)
|
2005-04-17 06:20:36 +08:00
|
|
|
goto out;
|
2010-07-08 23:02:09 +08:00
|
|
|
if (filpp) {
|
|
|
|
if (flags & RD_STATE)
|
|
|
|
*filpp = find_readable_file(stp->st_file);
|
|
|
|
else
|
|
|
|
*filpp = find_writeable_file(stp->st_file);
|
|
|
|
}
|
2011-09-17 08:12:38 +08:00
|
|
|
break;
|
|
|
|
default:
|
|
|
|
return nfserr_bad_stateid;
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
|
|
|
status = nfs_ok;
|
|
|
|
out:
|
|
|
|
return status;
|
|
|
|
}
|
|
|
|
|
2011-07-13 23:04:21 +08:00
|
|
|
static __be32
|
2011-09-08 04:06:42 +08:00
|
|
|
nfsd4_free_lock_stateid(struct nfs4_ol_stateid *stp)
|
2011-07-13 23:04:21 +08:00
|
|
|
{
|
2011-07-31 11:33:59 +08:00
|
|
|
if (check_for_locks(stp->st_file, lockowner(stp->st_stateowner)))
|
2011-07-13 23:04:21 +08:00
|
|
|
return nfserr_locks_held;
|
|
|
|
release_lock_stateid(stp);
|
|
|
|
return nfs_ok;
|
|
|
|
}
|
|
|
|
|
2011-07-13 22:50:48 +08:00
|
|
|
/*
|
|
|
|
* Test if the stateid is valid
|
|
|
|
*/
|
|
|
|
__be32
|
|
|
|
nfsd4_test_stateid(struct svc_rqst *rqstp, struct nfsd4_compound_state *cstate,
|
|
|
|
struct nfsd4_test_stateid *test_stateid)
|
|
|
|
{
|
2012-01-27 23:22:49 +08:00
|
|
|
struct nfsd4_test_stateid_id *stateid;
|
|
|
|
struct nfs4_client *cl = cstate->session->se_client;
|
|
|
|
|
|
|
|
nfs4_lock_state();
|
|
|
|
list_for_each_entry(stateid, &test_stateid->ts_stateid_list, ts_id_list)
|
|
|
|
stateid->ts_id_status = nfs4_validate_stateid(cl, &stateid->ts_id_stateid);
|
|
|
|
nfs4_unlock_state();
|
|
|
|
|
2011-07-13 22:50:48 +08:00
|
|
|
return nfs_ok;
|
|
|
|
}
|
|
|
|
|
2011-07-13 23:04:21 +08:00
|
|
|
__be32
|
|
|
|
nfsd4_free_stateid(struct svc_rqst *rqstp, struct nfsd4_compound_state *cstate,
|
|
|
|
struct nfsd4_free_stateid *free_stateid)
|
|
|
|
{
|
|
|
|
stateid_t *stateid = &free_stateid->fr_stateid;
|
2011-09-17 06:56:20 +08:00
|
|
|
struct nfs4_stid *s;
|
2011-09-24 05:01:19 +08:00
|
|
|
struct nfs4_client *cl = cstate->session->se_client;
|
2011-09-17 06:56:20 +08:00
|
|
|
__be32 ret = nfserr_bad_stateid;
|
2011-07-13 23:04:21 +08:00
|
|
|
|
|
|
|
nfs4_lock_state();
|
2011-09-24 05:01:19 +08:00
|
|
|
s = find_stateid(cl, stateid);
|
2011-09-17 06:56:20 +08:00
|
|
|
if (!s)
|
2011-08-23 23:03:29 +08:00
|
|
|
goto out;
|
2011-09-17 06:56:20 +08:00
|
|
|
switch (s->sc_type) {
|
|
|
|
case NFS4_DELEG_STID:
|
2011-07-13 23:04:21 +08:00
|
|
|
ret = nfserr_locks_held;
|
|
|
|
goto out;
|
2011-09-17 06:56:20 +08:00
|
|
|
case NFS4_OPEN_STID:
|
|
|
|
case NFS4_LOCK_STID:
|
|
|
|
ret = check_stateid_generation(stateid, &s->sc_stateid, 1);
|
|
|
|
if (ret)
|
|
|
|
goto out;
|
|
|
|
if (s->sc_type == NFS4_LOCK_STID)
|
|
|
|
ret = nfsd4_free_lock_stateid(openlockstateid(s));
|
|
|
|
else
|
|
|
|
ret = nfserr_locks_held;
|
2011-09-17 08:12:38 +08:00
|
|
|
break;
|
|
|
|
default:
|
|
|
|
ret = nfserr_bad_stateid;
|
2011-07-13 23:04:21 +08:00
|
|
|
}
|
|
|
|
out:
|
|
|
|
nfs4_unlock_state();
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
2005-07-08 08:59:27 +08:00
|
|
|
static inline int
|
|
|
|
setlkflg (int type)
|
|
|
|
{
|
|
|
|
return (type == NFS4_READW_LT || type == NFS4_READ_LT) ?
|
|
|
|
RD_STATE : WR_STATE;
|
|
|
|
}
|
2005-04-17 06:20:36 +08:00
|
|
|
|
2011-09-08 04:06:42 +08:00
|
|
|
static __be32 nfs4_seqid_op_checks(struct nfsd4_compound_state *cstate, stateid_t *stateid, u32 seqid, struct nfs4_ol_stateid *stp)
|
2011-09-07 03:19:46 +08:00
|
|
|
{
|
|
|
|
struct svc_fh *current_fh = &cstate->current_fh;
|
|
|
|
struct nfs4_stateowner *sop = stp->st_stateowner;
|
|
|
|
__be32 status;
|
|
|
|
|
|
|
|
status = nfsd4_check_seqid(cstate, sop, seqid);
|
|
|
|
if (status)
|
|
|
|
return status;
|
2011-09-17 08:12:38 +08:00
|
|
|
if (stp->st_stid.sc_type == NFS4_CLOSED_STID)
|
|
|
|
/*
|
|
|
|
* "Closed" stateid's exist *only* to return
|
|
|
|
* nfserr_replay_me from the previous step.
|
|
|
|
*/
|
|
|
|
return nfserr_bad_stateid;
|
|
|
|
status = check_stateid_generation(stateid, &stp->st_stid.sc_stateid, nfsd4_has_session(cstate));
|
|
|
|
if (status)
|
|
|
|
return status;
|
|
|
|
return nfs4_check_fh(current_fh, stp);
|
2011-09-07 03:19:46 +08:00
|
|
|
}
|
|
|
|
|
2005-04-17 06:20:36 +08:00
|
|
|
/*
|
|
|
|
* Checks for sequence id mutating operations.
|
|
|
|
*/
|
2006-10-20 14:28:59 +08:00
|
|
|
static __be32
|
2009-04-03 13:28:41 +08:00
|
|
|
nfs4_preprocess_seqid_op(struct nfsd4_compound_state *cstate, u32 seqid,
|
2011-09-07 03:50:21 +08:00
|
|
|
stateid_t *stateid, char typemask,
|
2011-09-08 04:06:42 +08:00
|
|
|
struct nfs4_ol_stateid **stpp)
|
2005-04-17 06:20:36 +08:00
|
|
|
{
|
2008-01-27 08:08:12 +08:00
|
|
|
__be32 status;
|
2011-09-24 05:01:19 +08:00
|
|
|
struct nfs4_stid *s;
|
2005-04-17 06:20:36 +08:00
|
|
|
|
2009-10-19 18:04:53 +08:00
|
|
|
dprintk("NFSD: %s: seqid=%d stateid = " STATEID_FMT "\n", __func__,
|
|
|
|
seqid, STATEID_VAL(stateid));
|
2005-07-08 08:59:26 +08:00
|
|
|
|
2005-04-17 06:20:36 +08:00
|
|
|
*stpp = NULL;
|
2011-09-24 05:01:19 +08:00
|
|
|
status = nfsd4_lookup_stateid(stateid, typemask, &s);
|
2011-09-07 03:19:46 +08:00
|
|
|
if (status)
|
|
|
|
return status;
|
2011-09-24 05:01:19 +08:00
|
|
|
*stpp = openlockstateid(s);
|
2011-09-07 03:19:46 +08:00
|
|
|
cstate->replay_owner = (*stpp)->st_stateowner;
|
2005-04-17 06:20:36 +08:00
|
|
|
|
2011-09-07 03:19:46 +08:00
|
|
|
return nfs4_seqid_op_checks(cstate, stateid, seqid, *stpp);
|
|
|
|
}
|
2007-11-27 06:06:39 +08:00
|
|
|
|
2011-09-08 04:06:42 +08:00
|
|
|
static __be32 nfs4_preprocess_confirmed_seqid_op(struct nfsd4_compound_state *cstate, u32 seqid, stateid_t *stateid, struct nfs4_ol_stateid **stpp)
|
2011-09-07 03:19:46 +08:00
|
|
|
{
|
|
|
|
__be32 status;
|
|
|
|
struct nfs4_openowner *oo;
|
2005-04-17 06:20:36 +08:00
|
|
|
|
2011-09-07 03:19:46 +08:00
|
|
|
status = nfs4_preprocess_seqid_op(cstate, seqid, stateid,
|
2011-09-07 03:50:21 +08:00
|
|
|
NFS4_OPEN_STID, stpp);
|
2011-09-02 21:03:37 +08:00
|
|
|
if (status)
|
|
|
|
return status;
|
2011-09-07 03:19:46 +08:00
|
|
|
oo = openowner((*stpp)->st_stateowner);
|
2011-09-13 00:24:13 +08:00
|
|
|
if (!(oo->oo_flags & NFS4_OO_CONFIRMED))
|
2005-07-08 08:59:26 +08:00
|
|
|
return nfserr_bad_stateid;
|
|
|
|
return nfs_ok;
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
|
|
|
|
2006-10-20 14:28:59 +08:00
|
|
|
__be32
|
2006-12-13 16:35:27 +08:00
|
|
|
nfsd4_open_confirm(struct svc_rqst *rqstp, struct nfsd4_compound_state *cstate,
|
2006-12-13 16:35:28 +08:00
|
|
|
struct nfsd4_open_confirm *oc)
|
2005-04-17 06:20:36 +08:00
|
|
|
{
|
2006-10-20 14:28:59 +08:00
|
|
|
__be32 status;
|
2011-07-31 11:33:59 +08:00
|
|
|
struct nfs4_openowner *oo;
|
2011-09-08 04:06:42 +08:00
|
|
|
struct nfs4_ol_stateid *stp;
|
2005-04-17 06:20:36 +08:00
|
|
|
|
|
|
|
dprintk("NFSD: nfsd4_open_confirm on file %.*s\n",
|
2006-12-13 16:35:27 +08:00
|
|
|
(int)cstate->current_fh.fh_dentry->d_name.len,
|
|
|
|
cstate->current_fh.fh_dentry->d_name.name);
|
2005-04-17 06:20:36 +08:00
|
|
|
|
2006-12-13 16:35:27 +08:00
|
|
|
status = fh_verify(rqstp, &cstate->current_fh, S_IFREG, 0);
|
2006-06-30 16:56:13 +08:00
|
|
|
if (status)
|
|
|
|
return status;
|
2005-04-17 06:20:36 +08:00
|
|
|
|
|
|
|
nfs4_lock_state();
|
|
|
|
|
2011-08-25 00:45:03 +08:00
|
|
|
status = nfs4_preprocess_seqid_op(cstate,
|
2006-12-13 16:35:27 +08:00
|
|
|
oc->oc_seqid, &oc->oc_req_stateid,
|
2011-09-07 03:50:21 +08:00
|
|
|
NFS4_OPEN_STID, &stp);
|
2011-08-25 00:45:03 +08:00
|
|
|
if (status)
|
2011-09-03 00:19:43 +08:00
|
|
|
goto out;
|
2011-07-31 11:33:59 +08:00
|
|
|
oo = openowner(stp->st_stateowner);
|
2011-09-03 00:19:43 +08:00
|
|
|
status = nfserr_bad_stateid;
|
2011-09-13 00:24:13 +08:00
|
|
|
if (oo->oo_flags & NFS4_OO_CONFIRMED)
|
2011-09-03 00:19:43 +08:00
|
|
|
goto out;
|
2011-09-13 00:24:13 +08:00
|
|
|
oo->oo_flags |= NFS4_OO_CONFIRMED;
|
2011-09-08 04:06:42 +08:00
|
|
|
update_stateid(&stp->st_stid.sc_stateid);
|
|
|
|
memcpy(&oc->oc_resp_stateid, &stp->st_stid.sc_stateid, sizeof(stateid_t));
|
2009-10-19 18:04:53 +08:00
|
|
|
dprintk("NFSD: %s: success, seqid=%d stateid=" STATEID_FMT "\n",
|
2011-09-08 04:06:42 +08:00
|
|
|
__func__, oc->oc_seqid, STATEID_VAL(&stp->st_stid.sc_stateid));
|
2005-06-24 13:04:30 +08:00
|
|
|
|
2012-03-22 04:42:43 +08:00
|
|
|
nfsd4_client_record_create(oo->oo_owner.so_client);
|
2011-09-03 00:19:43 +08:00
|
|
|
status = nfs_ok;
|
2005-04-17 06:20:36 +08:00
|
|
|
out:
|
2011-08-31 05:02:48 +08:00
|
|
|
if (!cstate->replay_owner)
|
|
|
|
nfs4_unlock_state();
|
2005-04-17 06:20:36 +08:00
|
|
|
return status;
|
|
|
|
}
|
|
|
|
|
2011-09-28 23:37:56 +08:00
|
|
|
static inline void nfs4_stateid_downgrade_bit(struct nfs4_ol_stateid *stp, u32 access)
|
2005-04-17 06:20:36 +08:00
|
|
|
{
|
2011-09-28 23:37:56 +08:00
|
|
|
if (!test_bit(access, &stp->st_access_bmap))
|
|
|
|
return;
|
|
|
|
nfs4_file_put_access(stp->st_file, nfs4_access_to_omode(access));
|
|
|
|
__clear_bit(access, &stp->st_access_bmap);
|
|
|
|
}
|
2011-06-29 20:23:50 +08:00
|
|
|
|
2011-09-28 23:37:56 +08:00
|
|
|
static inline void nfs4_stateid_downgrade(struct nfs4_ol_stateid *stp, u32 to_access)
|
|
|
|
{
|
|
|
|
switch (to_access) {
|
|
|
|
case NFS4_SHARE_ACCESS_READ:
|
|
|
|
nfs4_stateid_downgrade_bit(stp, NFS4_SHARE_ACCESS_WRITE);
|
|
|
|
nfs4_stateid_downgrade_bit(stp, NFS4_SHARE_ACCESS_BOTH);
|
|
|
|
break;
|
|
|
|
case NFS4_SHARE_ACCESS_WRITE:
|
|
|
|
nfs4_stateid_downgrade_bit(stp, NFS4_SHARE_ACCESS_READ);
|
|
|
|
nfs4_stateid_downgrade_bit(stp, NFS4_SHARE_ACCESS_BOTH);
|
|
|
|
break;
|
|
|
|
case NFS4_SHARE_ACCESS_BOTH:
|
|
|
|
break;
|
|
|
|
default:
|
|
|
|
BUG();
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
static void
|
|
|
|
reset_union_bmap_deny(unsigned long deny, unsigned long *bmap)
|
|
|
|
{
|
|
|
|
int i;
|
|
|
|
for (i = 0; i < 4; i++) {
|
|
|
|
if ((i & deny) != i)
|
|
|
|
__clear_bit(i, bmap);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2006-10-20 14:28:59 +08:00
|
|
|
__be32
|
2006-12-13 16:35:27 +08:00
|
|
|
nfsd4_open_downgrade(struct svc_rqst *rqstp,
|
|
|
|
struct nfsd4_compound_state *cstate,
|
2006-12-13 16:35:28 +08:00
|
|
|
struct nfsd4_open_downgrade *od)
|
2005-04-17 06:20:36 +08:00
|
|
|
{
|
2006-10-20 14:28:59 +08:00
|
|
|
__be32 status;
|
2011-09-08 04:06:42 +08:00
|
|
|
struct nfs4_ol_stateid *stp;
|
2005-04-17 06:20:36 +08:00
|
|
|
|
|
|
|
dprintk("NFSD: nfsd4_open_downgrade on file %.*s\n",
|
2006-12-13 16:35:27 +08:00
|
|
|
(int)cstate->current_fh.fh_dentry->d_name.len,
|
|
|
|
cstate->current_fh.fh_dentry->d_name.name);
|
2005-04-17 06:20:36 +08:00
|
|
|
|
2011-10-11 05:34:31 +08:00
|
|
|
/* We don't yet support WANT bits: */
|
2012-02-17 02:57:09 +08:00
|
|
|
if (od->od_deleg_want)
|
|
|
|
dprintk("NFSD: %s: od_deleg_want=0x%x ignored\n", __func__,
|
|
|
|
od->od_deleg_want);
|
2005-04-17 06:20:36 +08:00
|
|
|
|
|
|
|
nfs4_lock_state();
|
2011-09-07 03:19:46 +08:00
|
|
|
status = nfs4_preprocess_confirmed_seqid_op(cstate, od->od_seqid,
|
|
|
|
&od->od_stateid, &stp);
|
2011-08-25 00:45:03 +08:00
|
|
|
if (status)
|
2005-04-17 06:20:36 +08:00
|
|
|
goto out;
|
|
|
|
status = nfserr_inval;
|
|
|
|
if (!test_bit(od->od_share_access, &stp->st_access_bmap)) {
|
|
|
|
dprintk("NFSD:access not a subset current bitmap: 0x%lx, input access=%08x\n",
|
|
|
|
stp->st_access_bmap, od->od_share_access);
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
if (!test_bit(od->od_share_deny, &stp->st_deny_bmap)) {
|
|
|
|
dprintk("NFSD:deny not a subset current bitmap: 0x%lx, input deny=%08x\n",
|
|
|
|
stp->st_deny_bmap, od->od_share_deny);
|
|
|
|
goto out;
|
|
|
|
}
|
2011-09-28 23:37:56 +08:00
|
|
|
nfs4_stateid_downgrade(stp, od->od_share_access);
|
2005-04-17 06:20:36 +08:00
|
|
|
|
|
|
|
reset_union_bmap_deny(od->od_share_deny, &stp->st_deny_bmap);
|
|
|
|
|
2011-09-08 04:06:42 +08:00
|
|
|
update_stateid(&stp->st_stid.sc_stateid);
|
|
|
|
memcpy(&od->od_stateid, &stp->st_stid.sc_stateid, sizeof(stateid_t));
|
2005-04-17 06:20:36 +08:00
|
|
|
status = nfs_ok;
|
|
|
|
out:
|
2011-08-31 05:02:48 +08:00
|
|
|
if (!cstate->replay_owner)
|
|
|
|
nfs4_unlock_state();
|
2005-04-17 06:20:36 +08:00
|
|
|
return status;
|
|
|
|
}
|
|
|
|
|
2011-09-17 05:42:48 +08:00
|
|
|
void nfsd4_purge_closed_stateid(struct nfs4_stateowner *so)
|
|
|
|
{
|
|
|
|
struct nfs4_openowner *oo;
|
|
|
|
struct nfs4_ol_stateid *s;
|
|
|
|
|
|
|
|
if (!so->so_is_open_owner)
|
|
|
|
return;
|
|
|
|
oo = openowner(so);
|
|
|
|
s = oo->oo_last_closed_stid;
|
|
|
|
if (!s)
|
|
|
|
return;
|
|
|
|
if (!(oo->oo_flags & NFS4_OO_PURGE_CLOSE)) {
|
|
|
|
/* Release the last_closed_stid on the next seqid bump: */
|
|
|
|
oo->oo_flags |= NFS4_OO_PURGE_CLOSE;
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
oo->oo_flags &= ~NFS4_OO_PURGE_CLOSE;
|
2011-09-17 08:12:38 +08:00
|
|
|
release_last_closed_stateid(oo);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void nfsd4_close_open_stateid(struct nfs4_ol_stateid *s)
|
|
|
|
{
|
|
|
|
unhash_open_stateid(s);
|
|
|
|
s->st_stid.sc_type = NFS4_CLOSED_STID;
|
2011-09-17 05:42:48 +08:00
|
|
|
}
|
|
|
|
|
2005-04-17 06:20:36 +08:00
|
|
|
/*
|
|
|
|
* nfs4_unlock_state() called after encode
|
|
|
|
*/
|
2006-10-20 14:28:59 +08:00
|
|
|
__be32
|
2006-12-13 16:35:27 +08:00
|
|
|
nfsd4_close(struct svc_rqst *rqstp, struct nfsd4_compound_state *cstate,
|
2006-12-13 16:35:28 +08:00
|
|
|
struct nfsd4_close *close)
|
2005-04-17 06:20:36 +08:00
|
|
|
{
|
2006-10-20 14:28:59 +08:00
|
|
|
__be32 status;
|
2011-07-31 11:33:59 +08:00
|
|
|
struct nfs4_openowner *oo;
|
2011-09-08 04:06:42 +08:00
|
|
|
struct nfs4_ol_stateid *stp;
|
2005-04-17 06:20:36 +08:00
|
|
|
|
|
|
|
dprintk("NFSD: nfsd4_close on file %.*s\n",
|
2006-12-13 16:35:27 +08:00
|
|
|
(int)cstate->current_fh.fh_dentry->d_name.len,
|
|
|
|
cstate->current_fh.fh_dentry->d_name.name);
|
2005-04-17 06:20:36 +08:00
|
|
|
|
|
|
|
nfs4_lock_state();
|
2011-09-17 08:12:38 +08:00
|
|
|
status = nfs4_preprocess_seqid_op(cstate, close->cl_seqid,
|
|
|
|
&close->cl_stateid,
|
|
|
|
NFS4_OPEN_STID|NFS4_CLOSED_STID,
|
|
|
|
&stp);
|
2011-08-25 00:45:03 +08:00
|
|
|
if (status)
|
2005-04-17 06:20:36 +08:00
|
|
|
goto out;
|
2011-07-31 11:33:59 +08:00
|
|
|
oo = openowner(stp->st_stateowner);
|
2005-04-17 06:20:36 +08:00
|
|
|
status = nfs_ok;
|
2011-09-08 04:06:42 +08:00
|
|
|
update_stateid(&stp->st_stid.sc_stateid);
|
|
|
|
memcpy(&close->cl_stateid, &stp->st_stid.sc_stateid, sizeof(stateid_t));
|
2005-04-17 06:20:36 +08:00
|
|
|
|
2011-09-17 08:12:38 +08:00
|
|
|
nfsd4_close_open_stateid(stp);
|
2011-09-17 05:42:48 +08:00
|
|
|
oo->oo_last_closed_stid = stp;
|
2006-01-19 09:43:21 +08:00
|
|
|
|
|
|
|
/* place unused nfs4_stateowners on so_close_lru list to be
|
|
|
|
* released by the laundromat service after the lease period
|
|
|
|
* to enable us to handle CLOSE replay
|
|
|
|
*/
|
2011-07-31 11:33:59 +08:00
|
|
|
if (list_empty(&oo->oo_owner.so_stateids))
|
|
|
|
move_to_close_lru(oo);
|
2005-04-17 06:20:36 +08:00
|
|
|
out:
|
2011-08-31 05:02:48 +08:00
|
|
|
if (!cstate->replay_owner)
|
|
|
|
nfs4_unlock_state();
|
2005-04-17 06:20:36 +08:00
|
|
|
return status;
|
|
|
|
}
|
|
|
|
|
2006-10-20 14:28:59 +08:00
|
|
|
__be32
|
2006-12-13 16:35:27 +08:00
|
|
|
nfsd4_delegreturn(struct svc_rqst *rqstp, struct nfsd4_compound_state *cstate,
|
|
|
|
struct nfsd4_delegreturn *dr)
|
2005-04-17 06:20:36 +08:00
|
|
|
{
|
2009-02-22 05:29:14 +08:00
|
|
|
struct nfs4_delegation *dp;
|
|
|
|
stateid_t *stateid = &dr->dr_stateid;
|
2011-09-24 05:01:19 +08:00
|
|
|
struct nfs4_stid *s;
|
2009-02-22 05:29:14 +08:00
|
|
|
struct inode *inode;
|
2006-10-20 14:28:59 +08:00
|
|
|
__be32 status;
|
2005-04-17 06:20:36 +08:00
|
|
|
|
2006-12-13 16:35:27 +08:00
|
|
|
if ((status = fh_verify(rqstp, &cstate->current_fh, S_IFREG, 0)))
|
2009-02-22 05:29:14 +08:00
|
|
|
return status;
|
|
|
|
inode = cstate->current_fh.fh_dentry->d_inode;
|
2005-04-17 06:20:36 +08:00
|
|
|
|
|
|
|
nfs4_lock_state();
|
2011-09-24 05:01:19 +08:00
|
|
|
status = nfsd4_lookup_stateid(stateid, NFS4_DELEG_STID, &s);
|
|
|
|
if (status)
|
2009-02-22 05:29:14 +08:00
|
|
|
goto out;
|
2011-09-24 05:01:19 +08:00
|
|
|
dp = delegstateid(s);
|
2011-09-09 00:07:44 +08:00
|
|
|
status = check_stateid_generation(stateid, &dp->dl_stid.sc_stateid, nfsd4_has_session(cstate));
|
2009-02-22 05:29:14 +08:00
|
|
|
if (status)
|
|
|
|
goto out;
|
|
|
|
|
|
|
|
unhash_delegation(dp);
|
2005-04-17 06:20:36 +08:00
|
|
|
out:
|
2009-02-22 05:29:14 +08:00
|
|
|
nfs4_unlock_state();
|
|
|
|
|
2005-04-17 06:20:36 +08:00
|
|
|
return status;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
#define LOFF_OVERFLOW(start, len) ((u64)(len) > ~(u64)(start))
|
|
|
|
|
2011-11-08 06:40:10 +08:00
|
|
|
#define LOCKOWNER_INO_HASH_BITS 8
|
|
|
|
#define LOCKOWNER_INO_HASH_SIZE (1 << LOCKOWNER_INO_HASH_BITS)
|
|
|
|
#define LOCKOWNER_INO_HASH_MASK (LOCKOWNER_INO_HASH_SIZE - 1)
|
2005-04-17 06:20:36 +08:00
|
|
|
|
2008-12-16 01:42:03 +08:00
|
|
|
static inline u64
|
|
|
|
end_offset(u64 start, u64 len)
|
|
|
|
{
|
|
|
|
u64 end;
|
|
|
|
|
|
|
|
end = start + len;
|
|
|
|
return end >= start ? end: NFS4_MAX_UINT64;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* last octet in a range */
|
|
|
|
static inline u64
|
|
|
|
last_byte_offset(u64 start, u64 len)
|
|
|
|
{
|
|
|
|
u64 end;
|
|
|
|
|
|
|
|
BUG_ON(!len);
|
|
|
|
end = start + len;
|
|
|
|
return end > start ? end - 1: NFS4_MAX_UINT64;
|
|
|
|
}
|
|
|
|
|
2011-11-08 06:40:10 +08:00
|
|
|
static unsigned int lockowner_ino_hashval(struct inode *inode, u32 cl_id, struct xdr_netobj *ownername)
|
2005-04-17 06:20:36 +08:00
|
|
|
{
|
|
|
|
return (file_hashval(inode) + cl_id
|
|
|
|
+ opaque_hashval(ownername->data, ownername->len))
|
2011-11-08 06:40:10 +08:00
|
|
|
& LOCKOWNER_INO_HASH_MASK;
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
|
|
|
|
2011-11-08 06:40:10 +08:00
|
|
|
static struct list_head lockowner_ino_hashtbl[LOCKOWNER_INO_HASH_SIZE];
|
2005-04-17 06:20:36 +08:00
|
|
|
|
|
|
|
/*
|
|
|
|
* TODO: Linux file offsets are _signed_ 64-bit quantities, which means that
|
|
|
|
* we can't properly handle lock requests that go beyond the (2^63 - 1)-th
|
|
|
|
* byte, because of sign extension problems. Since NFSv4 calls for 64-bit
|
|
|
|
* locking, this prevents us from being completely protocol-compliant. The
|
|
|
|
* real solution to this problem is to start using unsigned file offsets in
|
|
|
|
* the VFS, but this is a very deep change!
|
|
|
|
*/
|
|
|
|
static inline void
|
|
|
|
nfs4_transform_lock_offset(struct file_lock *lock)
|
|
|
|
{
|
|
|
|
if (lock->fl_start < 0)
|
|
|
|
lock->fl_start = OFFSET_MAX;
|
|
|
|
if (lock->fl_end < 0)
|
|
|
|
lock->fl_end = OFFSET_MAX;
|
|
|
|
}
|
|
|
|
|
2006-04-11 13:55:22 +08:00
|
|
|
/* Hack!: For now, we're defining this just so we can use a pointer to it
|
|
|
|
* as a unique cookie to identify our (NFSv4's) posix locks. */
|
2009-09-22 08:01:12 +08:00
|
|
|
static const struct lock_manager_operations nfsd_posix_mng_ops = {
|
2006-04-11 13:55:22 +08:00
|
|
|
};
|
2005-04-17 06:20:36 +08:00
|
|
|
|
|
|
|
static inline void
|
|
|
|
nfs4_set_lock_denied(struct file_lock *fl, struct nfsd4_lock_denied *deny)
|
|
|
|
{
|
2011-07-31 11:33:59 +08:00
|
|
|
struct nfs4_lockowner *lo;
|
2005-04-17 06:20:36 +08:00
|
|
|
|
2006-04-11 13:55:22 +08:00
|
|
|
if (fl->fl_lmops == &nfsd_posix_mng_ops) {
|
2011-07-31 11:33:59 +08:00
|
|
|
lo = (struct nfs4_lockowner *) fl->fl_owner;
|
|
|
|
deny->ld_owner.data = kmemdup(lo->lo_owner.so_owner.data,
|
|
|
|
lo->lo_owner.so_owner.len, GFP_KERNEL);
|
2011-08-31 10:15:47 +08:00
|
|
|
if (!deny->ld_owner.data)
|
|
|
|
/* We just don't care that much */
|
|
|
|
goto nevermind;
|
2011-07-31 11:33:59 +08:00
|
|
|
deny->ld_owner.len = lo->lo_owner.so_owner.len;
|
|
|
|
deny->ld_clientid = lo->lo_owner.so_client->cl_clientid;
|
2006-04-11 13:55:22 +08:00
|
|
|
} else {
|
2011-08-31 10:15:47 +08:00
|
|
|
nevermind:
|
|
|
|
deny->ld_owner.len = 0;
|
|
|
|
deny->ld_owner.data = NULL;
|
2006-04-11 13:55:22 +08:00
|
|
|
deny->ld_clientid.cl_boot = 0;
|
|
|
|
deny->ld_clientid.cl_id = 0;
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
|
|
|
deny->ld_start = fl->fl_start;
|
2008-12-16 01:42:03 +08:00
|
|
|
deny->ld_length = NFS4_MAX_UINT64;
|
|
|
|
if (fl->fl_end != NFS4_MAX_UINT64)
|
2005-04-17 06:20:36 +08:00
|
|
|
deny->ld_length = fl->fl_end - fl->fl_start + 1;
|
|
|
|
deny->ld_type = NFS4_READ_LT;
|
|
|
|
if (fl->fl_type != F_RDLCK)
|
|
|
|
deny->ld_type = NFS4_WRITE_LT;
|
|
|
|
}
|
|
|
|
|
2011-11-08 05:37:57 +08:00
|
|
|
static bool same_lockowner_ino(struct nfs4_lockowner *lo, struct inode *inode, clientid_t *clid, struct xdr_netobj *owner)
|
|
|
|
{
|
|
|
|
struct nfs4_ol_stateid *lst;
|
|
|
|
|
|
|
|
if (!same_owner_str(&lo->lo_owner, owner, clid))
|
|
|
|
return false;
|
|
|
|
lst = list_first_entry(&lo->lo_owner.so_stateids,
|
|
|
|
struct nfs4_ol_stateid, st_perstateowner);
|
|
|
|
return lst->st_file->fi_inode == inode;
|
|
|
|
}
|
|
|
|
|
2011-07-31 11:33:59 +08:00
|
|
|
static struct nfs4_lockowner *
|
|
|
|
find_lockowner_str(struct inode *inode, clientid_t *clid,
|
2005-04-17 06:20:36 +08:00
|
|
|
struct xdr_netobj *owner)
|
|
|
|
{
|
2011-11-08 06:40:10 +08:00
|
|
|
unsigned int hashval = lockowner_ino_hashval(inode, clid->cl_id, owner);
|
2011-11-08 05:37:57 +08:00
|
|
|
struct nfs4_lockowner *lo;
|
2005-04-17 06:20:36 +08:00
|
|
|
|
2011-11-08 06:40:10 +08:00
|
|
|
list_for_each_entry(lo, &lockowner_ino_hashtbl[hashval], lo_owner_ino_hash) {
|
2011-11-08 05:37:57 +08:00
|
|
|
if (same_lockowner_ino(lo, inode, clid, owner))
|
|
|
|
return lo;
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
2011-09-08 04:06:42 +08:00
|
|
|
static void hash_lockowner(struct nfs4_lockowner *lo, unsigned int strhashval, struct nfs4_client *clp, struct nfs4_ol_stateid *open_stp)
|
2011-08-12 21:42:57 +08:00
|
|
|
{
|
2011-11-08 06:40:10 +08:00
|
|
|
struct inode *inode = open_stp->st_file->fi_inode;
|
|
|
|
unsigned int inohash = lockowner_ino_hashval(inode,
|
|
|
|
clp->cl_clientid.cl_id, &lo->lo_owner.so_owner);
|
|
|
|
|
2011-11-08 06:23:30 +08:00
|
|
|
list_add(&lo->lo_owner.so_strhash, &ownerstr_hashtbl[strhashval]);
|
2011-11-08 06:40:10 +08:00
|
|
|
list_add(&lo->lo_owner_ino_hash, &lockowner_ino_hashtbl[inohash]);
|
2011-07-31 11:33:59 +08:00
|
|
|
list_add(&lo->lo_perstateid, &open_stp->st_lockowners);
|
2011-08-12 21:42:57 +08:00
|
|
|
}
|
|
|
|
|
2005-04-17 06:20:36 +08:00
|
|
|
/*
|
|
|
|
* Alloc a lock owner structure.
|
|
|
|
* Called in nfsd4_lock - therefore, OPEN and OPEN_CONFIRM (if needed) has
|
2011-03-31 09:57:33 +08:00
|
|
|
* occurred.
|
2005-04-17 06:20:36 +08:00
|
|
|
*
|
2011-11-08 06:23:30 +08:00
|
|
|
* strhashval = ownerstr_hashval
|
2005-04-17 06:20:36 +08:00
|
|
|
*/
|
|
|
|
|
2011-07-31 11:33:59 +08:00
|
|
|
static struct nfs4_lockowner *
|
2011-09-08 04:06:42 +08:00
|
|
|
alloc_init_lock_stateowner(unsigned int strhashval, struct nfs4_client *clp, struct nfs4_ol_stateid *open_stp, struct nfsd4_lock *lock) {
|
2011-07-31 11:33:59 +08:00
|
|
|
struct nfs4_lockowner *lo;
|
2005-04-17 06:20:36 +08:00
|
|
|
|
2011-07-31 11:33:59 +08:00
|
|
|
lo = alloc_stateowner(lockowner_slab, &lock->lk_new_owner, clp);
|
|
|
|
if (!lo)
|
2005-04-17 06:20:36 +08:00
|
|
|
return NULL;
|
2011-07-31 11:33:59 +08:00
|
|
|
INIT_LIST_HEAD(&lo->lo_owner.so_stateids);
|
|
|
|
lo->lo_owner.so_is_open_owner = 0;
|
2005-09-13 16:25:38 +08:00
|
|
|
/* It is the openowner seqid that will be incremented in encode in the
|
|
|
|
* case of new lockowners; so increment the lock seqid manually: */
|
2011-07-31 11:33:59 +08:00
|
|
|
lo->lo_owner.so_seqid = lock->lk_new_lock_seqid + 1;
|
|
|
|
hash_lockowner(lo, strhashval, clp, open_stp);
|
|
|
|
return lo;
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
|
|
|
|
2011-09-08 04:06:42 +08:00
|
|
|
static struct nfs4_ol_stateid *
|
|
|
|
alloc_init_lock_stateid(struct nfs4_lockowner *lo, struct nfs4_file *fp, struct nfs4_ol_stateid *open_stp)
|
2005-04-17 06:20:36 +08:00
|
|
|
{
|
2011-09-08 04:06:42 +08:00
|
|
|
struct nfs4_ol_stateid *stp;
|
2011-09-16 03:02:41 +08:00
|
|
|
struct nfs4_client *clp = lo->lo_owner.so_client;
|
2005-04-17 06:20:36 +08:00
|
|
|
|
2011-10-17 23:14:48 +08:00
|
|
|
stp = nfs4_alloc_stateid(clp);
|
2005-06-24 13:03:03 +08:00
|
|
|
if (stp == NULL)
|
2011-09-24 04:21:15 +08:00
|
|
|
return NULL;
|
2011-10-17 23:14:48 +08:00
|
|
|
init_stid(&stp->st_stid, clp, NFS4_LOCK_STID);
|
2005-06-24 13:03:08 +08:00
|
|
|
list_add(&stp->st_perfile, &fp->fi_stateids);
|
2011-07-31 11:33:59 +08:00
|
|
|
list_add(&stp->st_perstateowner, &lo->lo_owner.so_stateids);
|
|
|
|
stp->st_stateowner = &lo->lo_owner;
|
2005-06-24 13:03:10 +08:00
|
|
|
get_nfs4_file(fp);
|
2005-04-17 06:20:36 +08:00
|
|
|
stp->st_file = fp;
|
2011-03-03 07:01:35 +08:00
|
|
|
stp->st_access_bmap = 0;
|
2005-04-17 06:20:36 +08:00
|
|
|
stp->st_deny_bmap = open_stp->st_deny_bmap;
|
2005-07-08 08:59:27 +08:00
|
|
|
stp->st_openstp = open_stp;
|
2005-04-17 06:20:36 +08:00
|
|
|
return stp;
|
|
|
|
}
|
|
|
|
|
2005-06-24 13:04:03 +08:00
|
|
|
static int
|
2005-04-17 06:20:36 +08:00
|
|
|
check_lock_length(u64 offset, u64 length)
|
|
|
|
{
|
2008-12-16 01:42:03 +08:00
|
|
|
return ((length == 0) || ((length != NFS4_MAX_UINT64) &&
|
2005-04-17 06:20:36 +08:00
|
|
|
LOFF_OVERFLOW(offset, length)));
|
|
|
|
}
|
|
|
|
|
2011-09-08 04:06:42 +08:00
|
|
|
static void get_lock_access(struct nfs4_ol_stateid *lock_stp, u32 access)
|
2011-03-03 07:01:35 +08:00
|
|
|
{
|
|
|
|
struct nfs4_file *fp = lock_stp->st_file;
|
|
|
|
int oflag = nfs4_access_to_omode(access);
|
|
|
|
|
|
|
|
if (test_bit(access, &lock_stp->st_access_bmap))
|
|
|
|
return;
|
|
|
|
nfs4_file_get_access(fp, oflag);
|
|
|
|
__set_bit(access, &lock_stp->st_access_bmap);
|
|
|
|
}
|
|
|
|
|
2011-10-20 18:57:46 +08:00
|
|
|
__be32 lookup_or_create_lock_state(struct nfsd4_compound_state *cstate, struct nfs4_ol_stateid *ost, struct nfsd4_lock *lock, struct nfs4_ol_stateid **lst, bool *new)
|
|
|
|
{
|
|
|
|
struct nfs4_file *fi = ost->st_file;
|
|
|
|
struct nfs4_openowner *oo = openowner(ost->st_stateowner);
|
|
|
|
struct nfs4_client *cl = oo->oo_owner.so_client;
|
|
|
|
struct nfs4_lockowner *lo;
|
|
|
|
unsigned int strhashval;
|
|
|
|
|
|
|
|
lo = find_lockowner_str(fi->fi_inode, &cl->cl_clientid, &lock->v.new.owner);
|
|
|
|
if (lo) {
|
|
|
|
if (!cstate->minorversion)
|
|
|
|
return nfserr_bad_seqid;
|
|
|
|
/* XXX: a lockowner always has exactly one stateid: */
|
|
|
|
*lst = list_first_entry(&lo->lo_owner.so_stateids,
|
|
|
|
struct nfs4_ol_stateid, st_perstateowner);
|
|
|
|
return nfs_ok;
|
|
|
|
}
|
2011-11-08 06:23:30 +08:00
|
|
|
strhashval = ownerstr_hashval(cl->cl_clientid.cl_id,
|
2011-10-20 18:57:46 +08:00
|
|
|
&lock->v.new.owner);
|
|
|
|
lo = alloc_init_lock_stateowner(strhashval, cl, ost, lock);
|
|
|
|
if (lo == NULL)
|
|
|
|
return nfserr_jukebox;
|
|
|
|
*lst = alloc_init_lock_stateid(lo, fi, ost);
|
|
|
|
if (*lst == NULL) {
|
|
|
|
release_lockowner(lo);
|
|
|
|
return nfserr_jukebox;
|
|
|
|
}
|
|
|
|
*new = true;
|
|
|
|
return nfs_ok;
|
|
|
|
}
|
|
|
|
|
2005-04-17 06:20:36 +08:00
|
|
|
/*
|
|
|
|
* LOCK operation
|
|
|
|
*/
|
2006-10-20 14:28:59 +08:00
|
|
|
__be32
|
2006-12-13 16:35:27 +08:00
|
|
|
nfsd4_lock(struct svc_rqst *rqstp, struct nfsd4_compound_state *cstate,
|
2006-12-13 16:35:28 +08:00
|
|
|
struct nfsd4_lock *lock)
|
2005-04-17 06:20:36 +08:00
|
|
|
{
|
2011-07-31 11:33:59 +08:00
|
|
|
struct nfs4_openowner *open_sop = NULL;
|
|
|
|
struct nfs4_lockowner *lock_sop = NULL;
|
2011-09-08 04:06:42 +08:00
|
|
|
struct nfs4_ol_stateid *lock_stp;
|
2010-08-21 06:09:31 +08:00
|
|
|
struct file *filp = NULL;
|
2005-04-17 06:20:36 +08:00
|
|
|
struct file_lock file_lock;
|
2006-03-21 02:44:26 +08:00
|
|
|
struct file_lock conflock;
|
2006-10-20 14:28:59 +08:00
|
|
|
__be32 status = 0;
|
2011-10-20 18:57:46 +08:00
|
|
|
bool new_state = false;
|
2011-08-23 01:13:31 +08:00
|
|
|
int lkflg;
|
2006-10-20 14:29:01 +08:00
|
|
|
int err;
|
2005-04-17 06:20:36 +08:00
|
|
|
|
|
|
|
dprintk("NFSD: nfsd4_lock: start=%Ld length=%Ld\n",
|
|
|
|
(long long) lock->lk_offset,
|
|
|
|
(long long) lock->lk_length);
|
|
|
|
|
|
|
|
if (check_lock_length(lock->lk_offset, lock->lk_length))
|
|
|
|
return nfserr_inval;
|
|
|
|
|
2006-12-13 16:35:27 +08:00
|
|
|
if ((status = fh_verify(rqstp, &cstate->current_fh,
|
2008-06-16 19:20:29 +08:00
|
|
|
S_IFREG, NFSD_MAY_LOCK))) {
|
2006-01-19 09:43:17 +08:00
|
|
|
dprintk("NFSD: nfsd4_lock: permission denied!\n");
|
|
|
|
return status;
|
|
|
|
}
|
|
|
|
|
2005-04-17 06:20:36 +08:00
|
|
|
nfs4_lock_state();
|
|
|
|
|
|
|
|
if (lock->lk_is_new) {
|
2005-07-08 08:59:17 +08:00
|
|
|
/*
|
|
|
|
* Client indicates that this is a new lockowner.
|
|
|
|
* Use open owner and open stateid to create lock owner and
|
|
|
|
* lock stateid.
|
|
|
|
*/
|
2011-09-08 04:06:42 +08:00
|
|
|
struct nfs4_ol_stateid *open_stp = NULL;
|
2011-11-05 05:08:10 +08:00
|
|
|
|
|
|
|
if (nfsd4_has_session(cstate))
|
|
|
|
/* See rfc 5661 18.10.3: given clientid is ignored: */
|
|
|
|
memcpy(&lock->v.new.clientid,
|
|
|
|
&cstate->session->se_client->cl_clientid,
|
|
|
|
sizeof(clientid_t));
|
|
|
|
|
2005-04-17 06:20:36 +08:00
|
|
|
status = nfserr_stale_clientid;
|
2011-11-05 05:08:10 +08:00
|
|
|
if (STALE_CLIENTID(&lock->lk_new_clientid))
|
2005-04-17 06:20:36 +08:00
|
|
|
goto out;
|
|
|
|
|
|
|
|
/* validate and update open stateid and open seqid */
|
2011-09-07 03:19:46 +08:00
|
|
|
status = nfs4_preprocess_confirmed_seqid_op(cstate,
|
2005-04-17 06:20:36 +08:00
|
|
|
lock->lk_new_open_seqid,
|
|
|
|
&lock->lk_new_open_stateid,
|
2011-09-07 03:19:46 +08:00
|
|
|
&open_stp);
|
2005-07-08 08:59:16 +08:00
|
|
|
if (status)
|
2005-04-17 06:20:36 +08:00
|
|
|
goto out;
|
2011-07-31 11:33:59 +08:00
|
|
|
open_sop = openowner(open_stp->st_stateowner);
|
2011-08-23 01:13:31 +08:00
|
|
|
status = nfserr_bad_stateid;
|
2011-11-05 05:08:10 +08:00
|
|
|
if (!same_clid(&open_sop->oo_owner.so_client->cl_clientid,
|
2011-08-23 01:13:31 +08:00
|
|
|
&lock->v.new.clientid))
|
|
|
|
goto out;
|
2011-10-20 18:57:46 +08:00
|
|
|
status = lookup_or_create_lock_state(cstate, open_stp, lock,
|
|
|
|
&lock_stp, &new_state);
|
|
|
|
if (status)
|
2005-04-17 06:20:36 +08:00
|
|
|
goto out;
|
|
|
|
} else {
|
|
|
|
/* lock (lock owner + lock stateid) already exists */
|
2009-04-03 13:28:41 +08:00
|
|
|
status = nfs4_preprocess_seqid_op(cstate,
|
2011-07-31 11:33:59 +08:00
|
|
|
lock->lk_old_lock_seqid,
|
|
|
|
&lock->lk_old_lock_stateid,
|
2011-09-07 03:50:21 +08:00
|
|
|
NFS4_LOCK_STID, &lock_stp);
|
2005-04-17 06:20:36 +08:00
|
|
|
if (status)
|
|
|
|
goto out;
|
|
|
|
}
|
2011-10-20 18:57:46 +08:00
|
|
|
lock_sop = lockowner(lock_stp->st_stateowner);
|
2005-04-17 06:20:36 +08:00
|
|
|
|
2011-08-23 01:13:31 +08:00
|
|
|
lkflg = setlkflg(lock->lk_type);
|
|
|
|
status = nfs4_check_openmode(lock_stp, lkflg);
|
|
|
|
if (status)
|
|
|
|
goto out;
|
|
|
|
|
2005-07-08 08:59:15 +08:00
|
|
|
status = nfserr_grace;
|
2007-09-07 00:34:25 +08:00
|
|
|
if (locks_in_grace() && !lock->lk_reclaim)
|
2005-07-08 08:59:15 +08:00
|
|
|
goto out;
|
|
|
|
status = nfserr_no_grace;
|
2007-09-07 00:34:25 +08:00
|
|
|
if (!locks_in_grace() && lock->lk_reclaim)
|
2005-07-08 08:59:15 +08:00
|
|
|
goto out;
|
|
|
|
|
2005-04-17 06:20:36 +08:00
|
|
|
locks_init_lock(&file_lock);
|
|
|
|
switch (lock->lk_type) {
|
|
|
|
case NFS4_READ_LT:
|
|
|
|
case NFS4_READW_LT:
|
2011-03-03 07:01:35 +08:00
|
|
|
filp = find_readable_file(lock_stp->st_file);
|
|
|
|
if (filp)
|
|
|
|
get_lock_access(lock_stp, NFS4_SHARE_ACCESS_READ);
|
2005-04-17 06:20:36 +08:00
|
|
|
file_lock.fl_type = F_RDLCK;
|
2011-03-03 12:48:33 +08:00
|
|
|
break;
|
2005-04-17 06:20:36 +08:00
|
|
|
case NFS4_WRITE_LT:
|
|
|
|
case NFS4_WRITEW_LT:
|
2011-03-03 07:01:35 +08:00
|
|
|
filp = find_writeable_file(lock_stp->st_file);
|
|
|
|
if (filp)
|
|
|
|
get_lock_access(lock_stp, NFS4_SHARE_ACCESS_WRITE);
|
2005-04-17 06:20:36 +08:00
|
|
|
file_lock.fl_type = F_WRLCK;
|
2011-03-03 12:48:33 +08:00
|
|
|
break;
|
2005-04-17 06:20:36 +08:00
|
|
|
default:
|
|
|
|
status = nfserr_inval;
|
|
|
|
goto out;
|
|
|
|
}
|
2010-07-08 23:02:09 +08:00
|
|
|
if (!filp) {
|
|
|
|
status = nfserr_openmode;
|
|
|
|
goto out;
|
|
|
|
}
|
2005-09-13 16:25:38 +08:00
|
|
|
file_lock.fl_owner = (fl_owner_t)lock_sop;
|
2005-04-17 06:20:36 +08:00
|
|
|
file_lock.fl_pid = current->tgid;
|
|
|
|
file_lock.fl_file = filp;
|
|
|
|
file_lock.fl_flags = FL_POSIX;
|
2006-04-11 13:55:22 +08:00
|
|
|
file_lock.fl_lmops = &nfsd_posix_mng_ops;
|
2005-04-17 06:20:36 +08:00
|
|
|
|
|
|
|
file_lock.fl_start = lock->lk_offset;
|
2008-12-16 01:42:03 +08:00
|
|
|
file_lock.fl_end = last_byte_offset(lock->lk_offset, lock->lk_length);
|
2005-04-17 06:20:36 +08:00
|
|
|
nfs4_transform_lock_offset(&file_lock);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Try to lock the file in the VFS.
|
|
|
|
* Note: locks.c uses the BKL to protect the inode's lock list.
|
|
|
|
*/
|
|
|
|
|
2011-03-03 12:48:33 +08:00
|
|
|
err = vfs_lock_file(filp, F_SETLK, &file_lock, &conflock);
|
2006-10-20 14:29:01 +08:00
|
|
|
switch (-err) {
|
2005-04-17 06:20:36 +08:00
|
|
|
case 0: /* success! */
|
2011-09-08 04:06:42 +08:00
|
|
|
update_stateid(&lock_stp->st_stid.sc_stateid);
|
|
|
|
memcpy(&lock->lk_resp_stateid, &lock_stp->st_stid.sc_stateid,
|
2005-04-17 06:20:36 +08:00
|
|
|
sizeof(stateid_t));
|
2006-10-20 14:29:01 +08:00
|
|
|
status = 0;
|
2006-03-26 17:37:26 +08:00
|
|
|
break;
|
|
|
|
case (EAGAIN): /* conflock holds conflicting lock */
|
|
|
|
status = nfserr_denied;
|
|
|
|
dprintk("NFSD: nfsd4_lock: conflicting lock found!\n");
|
|
|
|
nfs4_set_lock_denied(&conflock, &lock->lk_denied);
|
|
|
|
break;
|
2005-04-17 06:20:36 +08:00
|
|
|
case (EDEADLK):
|
|
|
|
status = nfserr_deadlock;
|
2006-03-26 17:37:26 +08:00
|
|
|
break;
|
2011-08-11 07:07:33 +08:00
|
|
|
default:
|
2006-11-29 05:26:41 +08:00
|
|
|
dprintk("NFSD: nfsd4_lock: vfs_lock_file() failed! status %d\n",err);
|
2011-08-11 07:07:33 +08:00
|
|
|
status = nfserrno(err);
|
2006-03-26 17:37:26 +08:00
|
|
|
break;
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
|
|
|
out:
|
2011-10-20 18:57:46 +08:00
|
|
|
if (status && new_state)
|
2009-01-12 04:24:04 +08:00
|
|
|
release_lockowner(lock_sop);
|
2011-08-31 05:02:48 +08:00
|
|
|
if (!cstate->replay_owner)
|
|
|
|
nfs4_unlock_state();
|
2005-04-17 06:20:36 +08:00
|
|
|
return status;
|
|
|
|
}
|
|
|
|
|
2008-12-21 03:58:38 +08:00
|
|
|
/*
|
|
|
|
* The NFSv4 spec allows a client to do a LOCKT without holding an OPEN,
|
|
|
|
* so we do a temporary open here just to get an open file to pass to
|
|
|
|
* vfs_test_lock. (Arguably perhaps test_lock should be done with an
|
|
|
|
* inode operation.)
|
|
|
|
*/
|
|
|
|
static int nfsd_test_lock(struct svc_rqst *rqstp, struct svc_fh *fhp, struct file_lock *lock)
|
|
|
|
{
|
|
|
|
struct file *file;
|
|
|
|
int err;
|
|
|
|
|
|
|
|
err = nfsd_open(rqstp, fhp, S_IFREG, NFSD_MAY_READ, &file);
|
|
|
|
if (err)
|
|
|
|
return err;
|
|
|
|
err = vfs_test_lock(file, lock);
|
|
|
|
nfsd_close(file);
|
|
|
|
return err;
|
|
|
|
}
|
|
|
|
|
2005-04-17 06:20:36 +08:00
|
|
|
/*
|
|
|
|
* LOCKT operation
|
|
|
|
*/
|
2006-10-20 14:28:59 +08:00
|
|
|
__be32
|
2006-12-13 16:35:27 +08:00
|
|
|
nfsd4_lockt(struct svc_rqst *rqstp, struct nfsd4_compound_state *cstate,
|
|
|
|
struct nfsd4_lockt *lockt)
|
2005-04-17 06:20:36 +08:00
|
|
|
{
|
|
|
|
struct inode *inode;
|
|
|
|
struct file_lock file_lock;
|
2011-07-31 11:33:59 +08:00
|
|
|
struct nfs4_lockowner *lo;
|
2006-11-29 05:26:41 +08:00
|
|
|
int error;
|
2006-10-20 14:28:59 +08:00
|
|
|
__be32 status;
|
2005-04-17 06:20:36 +08:00
|
|
|
|
2007-09-07 00:34:25 +08:00
|
|
|
if (locks_in_grace())
|
2005-04-17 06:20:36 +08:00
|
|
|
return nfserr_grace;
|
|
|
|
|
|
|
|
if (check_lock_length(lockt->lt_offset, lockt->lt_length))
|
|
|
|
return nfserr_inval;
|
|
|
|
|
|
|
|
nfs4_lock_state();
|
|
|
|
|
|
|
|
status = nfserr_stale_clientid;
|
2009-04-03 13:28:50 +08:00
|
|
|
if (!nfsd4_has_session(cstate) && STALE_CLIENTID(&lockt->lt_clientid))
|
2005-04-17 06:20:36 +08:00
|
|
|
goto out;
|
|
|
|
|
2011-08-16 06:39:32 +08:00
|
|
|
if ((status = fh_verify(rqstp, &cstate->current_fh, S_IFREG, 0)))
|
2005-04-17 06:20:36 +08:00
|
|
|
goto out;
|
|
|
|
|
2006-12-13 16:35:27 +08:00
|
|
|
inode = cstate->current_fh.fh_dentry->d_inode;
|
2005-04-17 06:20:36 +08:00
|
|
|
locks_init_lock(&file_lock);
|
|
|
|
switch (lockt->lt_type) {
|
|
|
|
case NFS4_READ_LT:
|
|
|
|
case NFS4_READW_LT:
|
|
|
|
file_lock.fl_type = F_RDLCK;
|
|
|
|
break;
|
|
|
|
case NFS4_WRITE_LT:
|
|
|
|
case NFS4_WRITEW_LT:
|
|
|
|
file_lock.fl_type = F_WRLCK;
|
|
|
|
break;
|
|
|
|
default:
|
2007-07-28 04:10:37 +08:00
|
|
|
dprintk("NFSD: nfs4_lockt: bad lock type!\n");
|
2005-04-17 06:20:36 +08:00
|
|
|
status = nfserr_inval;
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
|
2011-07-31 11:33:59 +08:00
|
|
|
lo = find_lockowner_str(inode, &lockt->lt_clientid, &lockt->lt_owner);
|
|
|
|
if (lo)
|
|
|
|
file_lock.fl_owner = (fl_owner_t)lo;
|
2005-04-17 06:20:36 +08:00
|
|
|
file_lock.fl_pid = current->tgid;
|
|
|
|
file_lock.fl_flags = FL_POSIX;
|
|
|
|
|
|
|
|
file_lock.fl_start = lockt->lt_offset;
|
2008-12-16 01:42:03 +08:00
|
|
|
file_lock.fl_end = last_byte_offset(lockt->lt_offset, lockt->lt_length);
|
2005-04-17 06:20:36 +08:00
|
|
|
|
|
|
|
nfs4_transform_lock_offset(&file_lock);
|
|
|
|
|
|
|
|
status = nfs_ok;
|
2008-12-21 03:58:38 +08:00
|
|
|
error = nfsd_test_lock(rqstp, &cstate->current_fh, &file_lock);
|
2006-11-29 05:26:41 +08:00
|
|
|
if (error) {
|
|
|
|
status = nfserrno(error);
|
|
|
|
goto out;
|
|
|
|
}
|
2007-02-21 13:55:18 +08:00
|
|
|
if (file_lock.fl_type != F_UNLCK) {
|
2005-04-17 06:20:36 +08:00
|
|
|
status = nfserr_denied;
|
2007-02-21 13:55:18 +08:00
|
|
|
nfs4_set_lock_denied(&file_lock, &lockt->lt_denied);
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
|
|
|
out:
|
|
|
|
nfs4_unlock_state();
|
|
|
|
return status;
|
|
|
|
}
|
|
|
|
|
2006-10-20 14:28:59 +08:00
|
|
|
__be32
|
2006-12-13 16:35:27 +08:00
|
|
|
nfsd4_locku(struct svc_rqst *rqstp, struct nfsd4_compound_state *cstate,
|
2006-12-13 16:35:28 +08:00
|
|
|
struct nfsd4_locku *locku)
|
2005-04-17 06:20:36 +08:00
|
|
|
{
|
2011-09-08 04:06:42 +08:00
|
|
|
struct nfs4_ol_stateid *stp;
|
2005-04-17 06:20:36 +08:00
|
|
|
struct file *filp = NULL;
|
|
|
|
struct file_lock file_lock;
|
2006-10-20 14:28:59 +08:00
|
|
|
__be32 status;
|
2006-10-20 14:29:01 +08:00
|
|
|
int err;
|
2005-04-17 06:20:36 +08:00
|
|
|
|
|
|
|
dprintk("NFSD: nfsd4_locku: start=%Ld length=%Ld\n",
|
|
|
|
(long long) locku->lu_offset,
|
|
|
|
(long long) locku->lu_length);
|
|
|
|
|
|
|
|
if (check_lock_length(locku->lu_offset, locku->lu_length))
|
|
|
|
return nfserr_inval;
|
|
|
|
|
|
|
|
nfs4_lock_state();
|
|
|
|
|
2011-08-25 00:45:03 +08:00
|
|
|
status = nfs4_preprocess_seqid_op(cstate, locku->lu_seqid,
|
2011-09-07 03:50:21 +08:00
|
|
|
&locku->lu_stateid, NFS4_LOCK_STID, &stp);
|
2011-08-25 00:45:03 +08:00
|
|
|
if (status)
|
2005-04-17 06:20:36 +08:00
|
|
|
goto out;
|
2010-07-08 23:02:09 +08:00
|
|
|
filp = find_any_file(stp->st_file);
|
|
|
|
if (!filp) {
|
|
|
|
status = nfserr_lock_range;
|
|
|
|
goto out;
|
|
|
|
}
|
2005-04-17 06:20:36 +08:00
|
|
|
BUG_ON(!filp);
|
|
|
|
locks_init_lock(&file_lock);
|
|
|
|
file_lock.fl_type = F_UNLCK;
|
2011-07-31 11:33:59 +08:00
|
|
|
file_lock.fl_owner = (fl_owner_t)lockowner(stp->st_stateowner);
|
2005-04-17 06:20:36 +08:00
|
|
|
file_lock.fl_pid = current->tgid;
|
|
|
|
file_lock.fl_file = filp;
|
|
|
|
file_lock.fl_flags = FL_POSIX;
|
2006-04-11 13:55:22 +08:00
|
|
|
file_lock.fl_lmops = &nfsd_posix_mng_ops;
|
2005-04-17 06:20:36 +08:00
|
|
|
file_lock.fl_start = locku->lu_offset;
|
|
|
|
|
2008-12-16 01:42:03 +08:00
|
|
|
file_lock.fl_end = last_byte_offset(locku->lu_offset, locku->lu_length);
|
2005-04-17 06:20:36 +08:00
|
|
|
nfs4_transform_lock_offset(&file_lock);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Try to unlock the file in the VFS.
|
|
|
|
*/
|
2006-11-29 05:26:41 +08:00
|
|
|
err = vfs_lock_file(filp, F_SETLK, &file_lock, NULL);
|
2006-10-20 14:29:01 +08:00
|
|
|
if (err) {
|
2006-11-29 05:26:41 +08:00
|
|
|
dprintk("NFSD: nfs4_locku: vfs_lock_file failed!\n");
|
2005-04-17 06:20:36 +08:00
|
|
|
goto out_nfserr;
|
|
|
|
}
|
|
|
|
/*
|
|
|
|
* OK, unlock succeeded; the only thing left to do is update the stateid.
|
|
|
|
*/
|
2011-09-08 04:06:42 +08:00
|
|
|
update_stateid(&stp->st_stid.sc_stateid);
|
|
|
|
memcpy(&locku->lu_stateid, &stp->st_stid.sc_stateid, sizeof(stateid_t));
|
2005-04-17 06:20:36 +08:00
|
|
|
|
|
|
|
out:
|
2011-09-28 09:42:29 +08:00
|
|
|
if (!cstate->replay_owner)
|
|
|
|
nfs4_unlock_state();
|
2005-04-17 06:20:36 +08:00
|
|
|
return status;
|
|
|
|
|
|
|
|
out_nfserr:
|
2006-10-20 14:29:01 +08:00
|
|
|
status = nfserrno(err);
|
2005-04-17 06:20:36 +08:00
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* returns
|
|
|
|
* 1: locks held by lockowner
|
|
|
|
* 0: no locks held by lockowner
|
|
|
|
*/
|
|
|
|
static int
|
2011-07-31 11:33:59 +08:00
|
|
|
check_for_locks(struct nfs4_file *filp, struct nfs4_lockowner *lowner)
|
2005-04-17 06:20:36 +08:00
|
|
|
{
|
|
|
|
struct file_lock **flpp;
|
2010-07-08 23:02:09 +08:00
|
|
|
struct inode *inode = filp->fi_inode;
|
2005-04-17 06:20:36 +08:00
|
|
|
int status = 0;
|
|
|
|
|
2010-09-18 21:09:31 +08:00
|
|
|
lock_flocks();
|
2005-04-17 06:20:36 +08:00
|
|
|
for (flpp = &inode->i_flock; *flpp != NULL; flpp = &(*flpp)->fl_next) {
|
2006-01-19 09:43:22 +08:00
|
|
|
if ((*flpp)->fl_owner == (fl_owner_t)lowner) {
|
2005-04-17 06:20:36 +08:00
|
|
|
status = 1;
|
|
|
|
goto out;
|
2006-01-19 09:43:22 +08:00
|
|
|
}
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
|
|
|
out:
|
2010-09-18 21:09:31 +08:00
|
|
|
unlock_flocks();
|
2005-04-17 06:20:36 +08:00
|
|
|
return status;
|
|
|
|
}
|
|
|
|
|
2006-10-20 14:28:59 +08:00
|
|
|
__be32
|
2006-12-13 16:35:38 +08:00
|
|
|
nfsd4_release_lockowner(struct svc_rqst *rqstp,
|
|
|
|
struct nfsd4_compound_state *cstate,
|
|
|
|
struct nfsd4_release_lockowner *rlockowner)
|
2005-04-17 06:20:36 +08:00
|
|
|
{
|
|
|
|
clientid_t *clid = &rlockowner->rl_clientid;
|
2005-06-24 13:04:20 +08:00
|
|
|
struct nfs4_stateowner *sop;
|
2011-07-31 11:33:59 +08:00
|
|
|
struct nfs4_lockowner *lo;
|
2011-09-08 04:06:42 +08:00
|
|
|
struct nfs4_ol_stateid *stp;
|
2005-04-17 06:20:36 +08:00
|
|
|
struct xdr_netobj *owner = &rlockowner->rl_owner;
|
2005-06-24 13:04:20 +08:00
|
|
|
struct list_head matches;
|
2011-11-08 06:23:30 +08:00
|
|
|
unsigned int hashval = ownerstr_hashval(clid->cl_id, owner);
|
2006-10-20 14:28:59 +08:00
|
|
|
__be32 status;
|
2005-04-17 06:20:36 +08:00
|
|
|
|
|
|
|
dprintk("nfsd4_release_lockowner clientid: (%08x/%08x):\n",
|
|
|
|
clid->cl_boot, clid->cl_id);
|
|
|
|
|
|
|
|
/* XXX check for lease expiration */
|
|
|
|
|
|
|
|
status = nfserr_stale_clientid;
|
2005-09-13 16:25:36 +08:00
|
|
|
if (STALE_CLIENTID(clid))
|
2005-04-17 06:20:36 +08:00
|
|
|
return status;
|
|
|
|
|
|
|
|
nfs4_lock_state();
|
|
|
|
|
2005-06-24 13:04:20 +08:00
|
|
|
status = nfserr_locks_held;
|
|
|
|
INIT_LIST_HEAD(&matches);
|
2011-11-08 05:58:18 +08:00
|
|
|
|
2011-11-08 06:23:30 +08:00
|
|
|
list_for_each_entry(sop, &ownerstr_hashtbl[hashval], so_strhash) {
|
|
|
|
if (sop->so_is_open_owner)
|
|
|
|
continue;
|
2011-11-08 05:58:18 +08:00
|
|
|
if (!same_owner_str(sop, owner, clid))
|
|
|
|
continue;
|
|
|
|
list_for_each_entry(stp, &sop->so_stateids,
|
|
|
|
st_perstateowner) {
|
|
|
|
lo = lockowner(sop);
|
|
|
|
if (check_for_locks(stp->st_file, lo))
|
|
|
|
goto out;
|
|
|
|
list_add(&lo->lo_list, &matches);
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
2005-06-24 13:04:20 +08:00
|
|
|
}
|
|
|
|
/* Clients probably won't expect us to return with some (but not all)
|
|
|
|
* of the lockowner state released; so don't release any until all
|
|
|
|
* have been checked. */
|
|
|
|
status = nfs_ok;
|
2005-07-08 08:59:14 +08:00
|
|
|
while (!list_empty(&matches)) {
|
2011-07-31 11:33:59 +08:00
|
|
|
lo = list_entry(matches.next, struct nfs4_lockowner,
|
|
|
|
lo_list);
|
2005-07-08 08:59:14 +08:00
|
|
|
/* unhash_stateowner deletes so_perclient only
|
|
|
|
* for openowners. */
|
2011-07-31 11:33:59 +08:00
|
|
|
list_del(&lo->lo_list);
|
|
|
|
release_lockowner(lo);
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
|
|
|
out:
|
|
|
|
nfs4_unlock_state();
|
|
|
|
return status;
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline struct nfs4_client_reclaim *
|
2005-06-24 13:03:52 +08:00
|
|
|
alloc_reclaim(void)
|
2005-04-17 06:20:36 +08:00
|
|
|
{
|
2005-06-24 13:03:52 +08:00
|
|
|
return kmalloc(sizeof(struct nfs4_client_reclaim), GFP_KERNEL);
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
|
|
|
|
2005-06-24 13:04:30 +08:00
|
|
|
int
|
2009-04-03 13:28:05 +08:00
|
|
|
nfs4_has_reclaimed_state(const char *name, bool use_exchange_id)
|
2005-06-24 13:04:30 +08:00
|
|
|
{
|
|
|
|
unsigned int strhashval = clientstr_hashval(name);
|
|
|
|
struct nfs4_client *clp;
|
|
|
|
|
2010-11-25 06:30:54 +08:00
|
|
|
clp = find_confirmed_client_by_str(name, strhashval);
|
2012-03-07 03:43:36 +08:00
|
|
|
if (!clp)
|
|
|
|
return 0;
|
2012-03-21 21:52:02 +08:00
|
|
|
return test_bit(NFSD4_CLIENT_STABLE, &clp->cl_flags);
|
2005-06-24 13:04:30 +08:00
|
|
|
}
|
|
|
|
|
2005-04-17 06:20:36 +08:00
|
|
|
/*
|
|
|
|
* failure => all reset bets are off, nfserr_no_grace...
|
|
|
|
*/
|
2005-06-24 13:04:25 +08:00
|
|
|
int
|
|
|
|
nfs4_client_to_reclaim(const char *name)
|
2005-04-17 06:20:36 +08:00
|
|
|
{
|
|
|
|
unsigned int strhashval;
|
|
|
|
struct nfs4_client_reclaim *crp = NULL;
|
|
|
|
|
2005-06-24 13:03:52 +08:00
|
|
|
dprintk("NFSD nfs4_client_to_reclaim NAME: %.*s\n", HEXDIR_LEN, name);
|
|
|
|
crp = alloc_reclaim();
|
2005-04-17 06:20:36 +08:00
|
|
|
if (!crp)
|
|
|
|
return 0;
|
2005-06-24 13:03:52 +08:00
|
|
|
strhashval = clientstr_hashval(name);
|
2005-04-17 06:20:36 +08:00
|
|
|
INIT_LIST_HEAD(&crp->cr_strhash);
|
|
|
|
list_add(&crp->cr_strhash, &reclaim_str_hashtbl[strhashval]);
|
2005-06-24 13:03:52 +08:00
|
|
|
memcpy(crp->cr_recdir, name, HEXDIR_LEN);
|
2005-04-17 06:20:36 +08:00
|
|
|
reclaim_str_hashtbl_size++;
|
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
|
2012-03-22 04:42:43 +08:00
|
|
|
void
|
2005-04-17 06:20:36 +08:00
|
|
|
nfs4_release_reclaim(void)
|
|
|
|
{
|
|
|
|
struct nfs4_client_reclaim *crp = NULL;
|
|
|
|
int i;
|
|
|
|
|
|
|
|
for (i = 0; i < CLIENT_HASH_SIZE; i++) {
|
|
|
|
while (!list_empty(&reclaim_str_hashtbl[i])) {
|
|
|
|
crp = list_entry(reclaim_str_hashtbl[i].next,
|
|
|
|
struct nfs4_client_reclaim, cr_strhash);
|
|
|
|
list_del(&crp->cr_strhash);
|
|
|
|
kfree(crp);
|
|
|
|
reclaim_str_hashtbl_size--;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
BUG_ON(reclaim_str_hashtbl_size);
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* called from OPEN, CLAIM_PREVIOUS with a new clientid. */
|
2012-03-22 04:42:43 +08:00
|
|
|
struct nfs4_client_reclaim *
|
2012-03-21 21:52:02 +08:00
|
|
|
nfsd4_find_reclaim_client(struct nfs4_client *clp)
|
2005-04-17 06:20:36 +08:00
|
|
|
{
|
|
|
|
unsigned int strhashval;
|
|
|
|
struct nfs4_client_reclaim *crp = NULL;
|
|
|
|
|
2005-06-24 13:03:52 +08:00
|
|
|
dprintk("NFSD: nfs4_find_reclaim_client for %.*s with recdir %s\n",
|
|
|
|
clp->cl_name.len, clp->cl_name.data,
|
|
|
|
clp->cl_recdir);
|
2005-04-17 06:20:36 +08:00
|
|
|
|
|
|
|
/* find clp->cl_name in reclaim_str_hashtbl */
|
2005-06-24 13:03:52 +08:00
|
|
|
strhashval = clientstr_hashval(clp->cl_recdir);
|
2005-04-17 06:20:36 +08:00
|
|
|
list_for_each_entry(crp, &reclaim_str_hashtbl[strhashval], cr_strhash) {
|
2005-06-24 13:03:52 +08:00
|
|
|
if (same_name(crp->cr_recdir, clp->cl_recdir)) {
|
2005-04-17 06:20:36 +08:00
|
|
|
return crp;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Called from OPEN. Look for clientid in reclaim list.
|
|
|
|
*/
|
2006-10-20 14:28:59 +08:00
|
|
|
__be32
|
2005-04-17 06:20:36 +08:00
|
|
|
nfs4_check_open_reclaim(clientid_t *clid)
|
|
|
|
{
|
2012-03-21 21:52:02 +08:00
|
|
|
struct nfs4_client *clp;
|
|
|
|
|
|
|
|
/* find clientid in conf_id_hashtbl */
|
|
|
|
clp = find_confirmed_client(clid);
|
|
|
|
if (clp == NULL)
|
|
|
|
return nfserr_reclaim_bad;
|
|
|
|
|
|
|
|
return nfsd4_client_record_check(clp) ? nfserr_reclaim_bad : nfs_ok;
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
|
|
|
|
2011-11-02 01:35:21 +08:00
|
|
|
#ifdef CONFIG_NFSD_FAULT_INJECTION
|
|
|
|
|
|
|
|
void nfsd_forget_clients(u64 num)
|
|
|
|
{
|
|
|
|
struct nfs4_client *clp, *next;
|
|
|
|
int count = 0;
|
|
|
|
|
|
|
|
nfs4_lock_state();
|
|
|
|
list_for_each_entry_safe(clp, next, &client_lru, cl_lru) {
|
2012-03-22 04:42:43 +08:00
|
|
|
nfsd4_client_record_remove(clp);
|
2011-11-02 01:35:21 +08:00
|
|
|
expire_client(clp);
|
|
|
|
if (++count == num)
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
nfs4_unlock_state();
|
|
|
|
|
|
|
|
printk(KERN_INFO "NFSD: Forgot %d clients", count);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void release_lockowner_sop(struct nfs4_stateowner *sop)
|
|
|
|
{
|
|
|
|
release_lockowner(lockowner(sop));
|
|
|
|
}
|
|
|
|
|
|
|
|
static void release_openowner_sop(struct nfs4_stateowner *sop)
|
|
|
|
{
|
|
|
|
release_openowner(openowner(sop));
|
|
|
|
}
|
|
|
|
|
2011-11-16 08:25:03 +08:00
|
|
|
static int nfsd_release_n_owners(u64 num, bool is_open_owner,
|
2011-11-02 01:35:21 +08:00
|
|
|
void (*release_sop)(struct nfs4_stateowner *))
|
|
|
|
{
|
|
|
|
int i, count = 0;
|
|
|
|
struct nfs4_stateowner *sop, *next;
|
|
|
|
|
2011-11-08 06:23:30 +08:00
|
|
|
for (i = 0; i < OWNER_HASH_SIZE; i++) {
|
|
|
|
list_for_each_entry_safe(sop, next, &ownerstr_hashtbl[i], so_strhash) {
|
|
|
|
if (sop->so_is_open_owner != is_open_owner)
|
|
|
|
continue;
|
2011-11-02 01:35:21 +08:00
|
|
|
release_sop(sop);
|
|
|
|
if (++count == num)
|
|
|
|
return count;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return count;
|
|
|
|
}
|
|
|
|
|
|
|
|
void nfsd_forget_locks(u64 num)
|
|
|
|
{
|
|
|
|
int count;
|
|
|
|
|
|
|
|
nfs4_lock_state();
|
2011-11-08 06:23:30 +08:00
|
|
|
count = nfsd_release_n_owners(num, false, release_lockowner_sop);
|
2011-11-02 01:35:21 +08:00
|
|
|
nfs4_unlock_state();
|
|
|
|
|
|
|
|
printk(KERN_INFO "NFSD: Forgot %d locks", count);
|
|
|
|
}
|
|
|
|
|
|
|
|
void nfsd_forget_openowners(u64 num)
|
|
|
|
{
|
|
|
|
int count;
|
|
|
|
|
|
|
|
nfs4_lock_state();
|
2011-11-08 06:23:30 +08:00
|
|
|
count = nfsd_release_n_owners(num, true, release_openowner_sop);
|
2011-11-02 01:35:21 +08:00
|
|
|
nfs4_unlock_state();
|
|
|
|
|
|
|
|
printk(KERN_INFO "NFSD: Forgot %d open owners", count);
|
|
|
|
}
|
|
|
|
|
|
|
|
int nfsd_process_n_delegations(u64 num, void (*deleg_func)(struct nfs4_delegation *))
|
|
|
|
{
|
|
|
|
int i, count = 0;
|
2011-12-15 03:39:56 +08:00
|
|
|
struct nfs4_file *fp, *fnext;
|
|
|
|
struct nfs4_delegation *dp, *dnext;
|
2011-11-02 01:35:21 +08:00
|
|
|
|
|
|
|
for (i = 0; i < FILE_HASH_SIZE; i++) {
|
2011-12-15 03:39:56 +08:00
|
|
|
list_for_each_entry_safe(fp, fnext, &file_hashtbl[i], fi_hash) {
|
|
|
|
list_for_each_entry_safe(dp, dnext, &fp->fi_delegations, dl_perfile) {
|
2011-11-02 01:35:21 +08:00
|
|
|
deleg_func(dp);
|
|
|
|
if (++count == num)
|
|
|
|
return count;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
2011-12-15 03:39:56 +08:00
|
|
|
|
2011-11-02 01:35:21 +08:00
|
|
|
return count;
|
|
|
|
}
|
|
|
|
|
|
|
|
void nfsd_forget_delegations(u64 num)
|
|
|
|
{
|
|
|
|
unsigned int count;
|
|
|
|
|
|
|
|
nfs4_lock_state();
|
|
|
|
count = nfsd_process_n_delegations(num, unhash_delegation);
|
|
|
|
nfs4_unlock_state();
|
|
|
|
|
|
|
|
printk(KERN_INFO "NFSD: Forgot %d delegations", count);
|
|
|
|
}
|
|
|
|
|
|
|
|
void nfsd_recall_delegations(u64 num)
|
|
|
|
{
|
|
|
|
unsigned int count;
|
|
|
|
|
|
|
|
nfs4_lock_state();
|
|
|
|
spin_lock(&recall_lock);
|
|
|
|
count = nfsd_process_n_delegations(num, nfsd_break_one_deleg);
|
|
|
|
spin_unlock(&recall_lock);
|
|
|
|
nfs4_unlock_state();
|
|
|
|
|
|
|
|
printk(KERN_INFO "NFSD: Recalled %d delegations", count);
|
|
|
|
}
|
|
|
|
|
|
|
|
#endif /* CONFIG_NFSD_FAULT_INJECTION */
|
|
|
|
|
2005-06-24 13:03:30 +08:00
|
|
|
/* initialization to perform at module load time: */
|
2005-04-17 06:20:36 +08:00
|
|
|
|
2011-11-02 03:24:59 +08:00
|
|
|
void
|
2005-06-24 13:03:30 +08:00
|
|
|
nfs4_state_init(void)
|
2005-04-17 06:20:36 +08:00
|
|
|
{
|
2011-11-02 03:24:59 +08:00
|
|
|
int i;
|
2005-04-17 06:20:36 +08:00
|
|
|
|
|
|
|
for (i = 0; i < CLIENT_HASH_SIZE; i++) {
|
|
|
|
INIT_LIST_HEAD(&conf_id_hashtbl[i]);
|
|
|
|
INIT_LIST_HEAD(&conf_str_hashtbl[i]);
|
|
|
|
INIT_LIST_HEAD(&unconf_str_hashtbl[i]);
|
|
|
|
INIT_LIST_HEAD(&unconf_id_hashtbl[i]);
|
2009-04-24 15:41:57 +08:00
|
|
|
INIT_LIST_HEAD(&reclaim_str_hashtbl[i]);
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
2009-04-03 13:27:52 +08:00
|
|
|
for (i = 0; i < SESSION_HASH_SIZE; i++)
|
|
|
|
INIT_LIST_HEAD(&sessionid_hashtbl[i]);
|
2005-04-17 06:20:36 +08:00
|
|
|
for (i = 0; i < FILE_HASH_SIZE; i++) {
|
|
|
|
INIT_LIST_HEAD(&file_hashtbl[i]);
|
|
|
|
}
|
2011-11-08 06:23:30 +08:00
|
|
|
for (i = 0; i < OWNER_HASH_SIZE; i++) {
|
|
|
|
INIT_LIST_HEAD(&ownerstr_hashtbl[i]);
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
2011-11-08 06:40:10 +08:00
|
|
|
for (i = 0; i < LOCKOWNER_INO_HASH_SIZE; i++)
|
|
|
|
INIT_LIST_HEAD(&lockowner_ino_hashtbl[i]);
|
2005-04-17 06:20:36 +08:00
|
|
|
INIT_LIST_HEAD(&close_lru);
|
|
|
|
INIT_LIST_HEAD(&client_lru);
|
|
|
|
INIT_LIST_HEAD(&del_recall_lru);
|
2005-06-24 13:03:30 +08:00
|
|
|
reclaim_str_hashtbl_size = 0;
|
|
|
|
}
|
|
|
|
|
2007-07-17 19:04:39 +08:00
|
|
|
/*
|
|
|
|
* Since the lifetime of a delegation isn't limited to that of an open, a
|
|
|
|
* client may quite reasonably hang on to a delegation as long as it has
|
|
|
|
* the inode cached. This becomes an obvious problem the first time a
|
|
|
|
* client's inode cache approaches the size of the server's total memory.
|
|
|
|
*
|
|
|
|
* For now we avoid this problem by imposing a hard limit on the number
|
|
|
|
* of delegations, which varies according to the server's memory size.
|
|
|
|
*/
|
|
|
|
static void
|
|
|
|
set_max_delegations(void)
|
|
|
|
{
|
|
|
|
/*
|
|
|
|
* Allow at most 4 delegations per megabyte of RAM. Quick
|
|
|
|
* estimates suggest that in the worst case (where every delegation
|
|
|
|
* is for a different inode), a delegation could take about 1.5K,
|
|
|
|
* giving a worst case usage of about 6% of memory.
|
|
|
|
*/
|
|
|
|
max_delegations = nr_free_buffer_pages() >> (20 - 2 - PAGE_SHIFT);
|
|
|
|
}
|
|
|
|
|
2005-06-24 13:03:30 +08:00
|
|
|
/* initialization to perform when the nfsd service is started: */
|
|
|
|
|
2012-03-06 00:42:36 +08:00
|
|
|
int
|
|
|
|
nfs4_state_start(void)
|
2005-06-24 13:03:30 +08:00
|
|
|
{
|
2010-03-04 03:52:55 +08:00
|
|
|
int ret;
|
|
|
|
|
2012-03-22 04:42:43 +08:00
|
|
|
/*
|
|
|
|
* FIXME: For now, we hang most of the pernet global stuff off of
|
|
|
|
* init_net until nfsd is fully containerized. Eventually, we'll
|
|
|
|
* need to pass a net pointer into this function, take a reference
|
|
|
|
* to that instead and then do most of the rest of this on a per-net
|
|
|
|
* basis.
|
|
|
|
*/
|
|
|
|
get_net(&init_net);
|
|
|
|
nfsd4_client_tracking_init(&init_net);
|
2005-04-17 06:20:36 +08:00
|
|
|
boot_time = get_seconds();
|
2007-09-07 00:34:25 +08:00
|
|
|
locks_start_grace(&nfsd4_manager);
|
2012-04-10 23:08:48 +08:00
|
|
|
grace_ended = false;
|
2007-07-17 19:04:35 +08:00
|
|
|
printk(KERN_INFO "NFSD: starting %ld-second grace period\n",
|
2010-03-02 08:21:21 +08:00
|
|
|
nfsd4_grace);
|
2010-03-04 03:52:55 +08:00
|
|
|
ret = set_callback_cred();
|
2012-03-06 00:42:36 +08:00
|
|
|
if (ret) {
|
|
|
|
ret = -ENOMEM;
|
|
|
|
goto out_recovery;
|
|
|
|
}
|
2005-06-24 13:03:19 +08:00
|
|
|
laundry_wq = create_singlethread_workqueue("nfsd4");
|
2012-03-06 00:42:36 +08:00
|
|
|
if (laundry_wq == NULL) {
|
|
|
|
ret = -ENOMEM;
|
|
|
|
goto out_recovery;
|
|
|
|
}
|
2010-03-04 03:52:55 +08:00
|
|
|
ret = nfsd4_create_callback_queue();
|
|
|
|
if (ret)
|
|
|
|
goto out_free_laundry;
|
2010-03-02 08:21:21 +08:00
|
|
|
queue_delayed_work(laundry_wq, &laundromat_work, nfsd4_grace * HZ);
|
2007-07-17 19:04:39 +08:00
|
|
|
set_max_delegations();
|
2010-03-04 03:52:55 +08:00
|
|
|
return 0;
|
|
|
|
out_free_laundry:
|
|
|
|
destroy_workqueue(laundry_wq);
|
2012-03-06 00:42:36 +08:00
|
|
|
out_recovery:
|
2012-03-22 04:42:43 +08:00
|
|
|
nfsd4_client_tracking_exit(&init_net);
|
|
|
|
put_net(&init_net);
|
2010-03-04 03:52:55 +08:00
|
|
|
return ret;
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
static void
|
|
|
|
__nfs4_state_shutdown(void)
|
|
|
|
{
|
|
|
|
int i;
|
|
|
|
struct nfs4_client *clp = NULL;
|
|
|
|
struct nfs4_delegation *dp = NULL;
|
|
|
|
struct list_head *pos, *next, reaplist;
|
|
|
|
|
|
|
|
for (i = 0; i < CLIENT_HASH_SIZE; i++) {
|
|
|
|
while (!list_empty(&conf_id_hashtbl[i])) {
|
|
|
|
clp = list_entry(conf_id_hashtbl[i].next, struct nfs4_client, cl_idhash);
|
|
|
|
expire_client(clp);
|
|
|
|
}
|
|
|
|
while (!list_empty(&unconf_str_hashtbl[i])) {
|
|
|
|
clp = list_entry(unconf_str_hashtbl[i].next, struct nfs4_client, cl_strhash);
|
|
|
|
expire_client(clp);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
INIT_LIST_HEAD(&reaplist);
|
|
|
|
spin_lock(&recall_lock);
|
|
|
|
list_for_each_safe(pos, next, &del_recall_lru) {
|
|
|
|
dp = list_entry (pos, struct nfs4_delegation, dl_recall_lru);
|
|
|
|
list_move(&dp->dl_recall_lru, &reaplist);
|
|
|
|
}
|
|
|
|
spin_unlock(&recall_lock);
|
|
|
|
list_for_each_safe(pos, next, &reaplist) {
|
|
|
|
dp = list_entry (pos, struct nfs4_delegation, dl_recall_lru);
|
|
|
|
unhash_delegation(dp);
|
|
|
|
}
|
|
|
|
|
2012-03-22 04:42:43 +08:00
|
|
|
nfsd4_client_tracking_exit(&init_net);
|
|
|
|
put_net(&init_net);
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
void
|
|
|
|
nfs4_state_shutdown(void)
|
|
|
|
{
|
2010-12-14 23:21:17 +08:00
|
|
|
cancel_delayed_work_sync(&laundromat_work);
|
2006-04-11 13:55:37 +08:00
|
|
|
destroy_workqueue(laundry_wq);
|
2008-11-21 04:36:17 +08:00
|
|
|
locks_end_grace(&nfsd4_manager);
|
2005-04-17 06:20:36 +08:00
|
|
|
nfs4_lock_state();
|
|
|
|
__nfs4_state_shutdown();
|
|
|
|
nfs4_unlock_state();
|
2010-06-05 04:42:08 +08:00
|
|
|
nfsd4_destroy_callback_queue();
|
2005-04-17 06:20:36 +08:00
|
|
|
}
|
2012-02-14 05:55:24 +08:00
|
|
|
|
|
|
|
static void
|
|
|
|
get_stateid(struct nfsd4_compound_state *cstate, stateid_t *stateid)
|
|
|
|
{
|
2012-02-14 05:55:32 +08:00
|
|
|
if (HAS_STATE_ID(cstate, CURRENT_STATE_ID_FLAG) && CURRENT_STATEID(stateid))
|
|
|
|
memcpy(stateid, &cstate->current_stateid, sizeof(stateid_t));
|
2012-02-14 05:55:24 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
static void
|
|
|
|
put_stateid(struct nfsd4_compound_state *cstate, stateid_t *stateid)
|
|
|
|
{
|
2012-02-14 05:55:32 +08:00
|
|
|
if (cstate->minorversion) {
|
|
|
|
memcpy(&cstate->current_stateid, stateid, sizeof(stateid_t));
|
|
|
|
SET_STATE_ID(cstate, CURRENT_STATE_ID_FLAG);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
void
|
|
|
|
clear_current_stateid(struct nfsd4_compound_state *cstate)
|
|
|
|
{
|
|
|
|
CLEAR_STATE_ID(cstate, CURRENT_STATE_ID_FLAG);
|
2012-02-14 05:55:24 +08:00
|
|
|
}
|
|
|
|
|
2012-02-14 05:55:25 +08:00
|
|
|
/*
|
|
|
|
* functions to set current state id
|
|
|
|
*/
|
2012-02-14 05:55:31 +08:00
|
|
|
void
|
|
|
|
nfsd4_set_opendowngradestateid(struct nfsd4_compound_state *cstate, struct nfsd4_open_downgrade *odp)
|
|
|
|
{
|
|
|
|
put_stateid(cstate, &odp->od_stateid);
|
|
|
|
}
|
|
|
|
|
2012-02-14 05:55:24 +08:00
|
|
|
void
|
|
|
|
nfsd4_set_openstateid(struct nfsd4_compound_state *cstate, struct nfsd4_open *open)
|
|
|
|
{
|
|
|
|
put_stateid(cstate, &open->op_stateid);
|
|
|
|
}
|
|
|
|
|
2012-02-14 05:55:25 +08:00
|
|
|
void
|
|
|
|
nfsd4_set_closestateid(struct nfsd4_compound_state *cstate, struct nfsd4_close *close)
|
|
|
|
{
|
|
|
|
put_stateid(cstate, &close->cl_stateid);
|
|
|
|
}
|
|
|
|
|
|
|
|
void
|
|
|
|
nfsd4_set_lockstateid(struct nfsd4_compound_state *cstate, struct nfsd4_lock *lock)
|
|
|
|
{
|
|
|
|
put_stateid(cstate, &lock->lk_resp_stateid);
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* functions to consume current state id
|
|
|
|
*/
|
2012-02-14 05:55:30 +08:00
|
|
|
|
2012-02-14 05:55:31 +08:00
|
|
|
void
|
|
|
|
nfsd4_get_opendowngradestateid(struct nfsd4_compound_state *cstate, struct nfsd4_open_downgrade *odp)
|
|
|
|
{
|
|
|
|
get_stateid(cstate, &odp->od_stateid);
|
|
|
|
}
|
|
|
|
|
|
|
|
void
|
|
|
|
nfsd4_get_delegreturnstateid(struct nfsd4_compound_state *cstate, struct nfsd4_delegreturn *drp)
|
|
|
|
{
|
|
|
|
get_stateid(cstate, &drp->dr_stateid);
|
|
|
|
}
|
|
|
|
|
2012-02-14 05:55:30 +08:00
|
|
|
void
|
|
|
|
nfsd4_get_freestateid(struct nfsd4_compound_state *cstate, struct nfsd4_free_stateid *fsp)
|
|
|
|
{
|
|
|
|
get_stateid(cstate, &fsp->fr_stateid);
|
|
|
|
}
|
|
|
|
|
|
|
|
void
|
|
|
|
nfsd4_get_setattrstateid(struct nfsd4_compound_state *cstate, struct nfsd4_setattr *setattr)
|
|
|
|
{
|
|
|
|
get_stateid(cstate, &setattr->sa_stateid);
|
|
|
|
}
|
|
|
|
|
2012-02-14 05:55:24 +08:00
|
|
|
void
|
|
|
|
nfsd4_get_closestateid(struct nfsd4_compound_state *cstate, struct nfsd4_close *close)
|
|
|
|
{
|
|
|
|
get_stateid(cstate, &close->cl_stateid);
|
|
|
|
}
|
|
|
|
|
|
|
|
void
|
2012-02-14 05:55:25 +08:00
|
|
|
nfsd4_get_lockustateid(struct nfsd4_compound_state *cstate, struct nfsd4_locku *locku)
|
2012-02-14 05:55:24 +08:00
|
|
|
{
|
2012-02-14 05:55:25 +08:00
|
|
|
get_stateid(cstate, &locku->lu_stateid);
|
2012-02-14 05:55:24 +08:00
|
|
|
}
|
2012-02-14 05:55:26 +08:00
|
|
|
|
|
|
|
void
|
|
|
|
nfsd4_get_readstateid(struct nfsd4_compound_state *cstate, struct nfsd4_read *read)
|
|
|
|
{
|
|
|
|
get_stateid(cstate, &read->rd_stateid);
|
|
|
|
}
|
|
|
|
|
|
|
|
void
|
|
|
|
nfsd4_get_writestateid(struct nfsd4_compound_state *cstate, struct nfsd4_write *write)
|
|
|
|
{
|
|
|
|
get_stateid(cstate, &write->wr_stateid);
|
|
|
|
}
|