2022-02-17 21:30:38 +08:00
|
|
|
// SPDX-License-Identifier: GPL-2.0-only
|
|
|
|
/* Object lifetime handling and tracing.
|
|
|
|
*
|
|
|
|
* Copyright (C) 2022 Red Hat, Inc. All Rights Reserved.
|
|
|
|
* Written by David Howells (dhowells@redhat.com)
|
|
|
|
*/
|
|
|
|
|
|
|
|
#include <linux/slab.h>
|
|
|
|
#include "internal.h"
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Allocate an I/O request and initialise it.
|
|
|
|
*/
|
2021-08-26 21:24:42 +08:00
|
|
|
struct netfs_io_request *netfs_alloc_request(struct address_space *mapping,
|
|
|
|
struct file *file,
|
|
|
|
loff_t start, size_t len,
|
|
|
|
enum netfs_io_origin origin)
|
2022-02-17 21:30:38 +08:00
|
|
|
{
|
|
|
|
static atomic_t debug_ids;
|
2021-06-30 05:37:05 +08:00
|
|
|
struct inode *inode = file ? file_inode(file) : mapping->host;
|
netfs: Fix gcc-12 warning by embedding vfs inode in netfs_i_context
While randstruct was satisfied with using an open-coded "void *" offset
cast for the netfs_i_context <-> inode casting, __builtin_object_size() as
used by FORTIFY_SOURCE was not as easily fooled. This was causing the
following complaint[1] from gcc v12:
In file included from include/linux/string.h:253,
from include/linux/ceph/ceph_debug.h:7,
from fs/ceph/inode.c:2:
In function 'fortify_memset_chk',
inlined from 'netfs_i_context_init' at include/linux/netfs.h:326:2,
inlined from 'ceph_alloc_inode' at fs/ceph/inode.c:463:2:
include/linux/fortify-string.h:242:25: warning: call to '__write_overflow_field' declared with attribute warning: detected write beyond size of field (1st parameter); maybe use struct_group()? [-Wattribute-warning]
242 | __write_overflow_field(p_size_field, size);
| ^~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
Fix this by embedding a struct inode into struct netfs_i_context (which
should perhaps be renamed to struct netfs_inode). The struct inode
vfs_inode fields are then removed from the 9p, afs, ceph and cifs inode
structs and vfs_inode is then simply changed to "netfs.inode" in those
filesystems.
Further, rename netfs_i_context to netfs_inode, get rid of the
netfs_inode() function that converted a netfs_i_context pointer to an
inode pointer (that can now be done with &ctx->inode) and rename the
netfs_i_context() function to netfs_inode() (which is now a wrapper
around container_of()).
Most of the changes were done with:
perl -p -i -e 's/vfs_inode/netfs.inode/'g \
`git grep -l 'vfs_inode' -- fs/{9p,afs,ceph,cifs}/*.[ch]`
Kees suggested doing it with a pair structure[2] and a special
declarator to insert that into the network filesystem's inode
wrapper[3], but I think it's cleaner to embed it - and then it doesn't
matter if struct randomisation reorders things.
Dave Chinner suggested using a filesystem-specific VFS_I() function in
each filesystem to convert that filesystem's own inode wrapper struct
into the VFS inode struct[4].
Version #2:
- Fix a couple of missed name changes due to a disabled cifs option.
- Rename nfs_i_context to nfs_inode
- Use "netfs" instead of "nic" as the member name in per-fs inode wrapper
structs.
[ This also undoes commit 507160f46c55 ("netfs: gcc-12: temporarily
disable '-Wattribute-warning' for now") that is no longer needed ]
Fixes: bc899ee1c898 ("netfs: Add a netfs inode context")
Reported-by: Jeff Layton <jlayton@kernel.org>
Signed-off-by: David Howells <dhowells@redhat.com>
Reviewed-by: Jeff Layton <jlayton@kernel.org>
Reviewed-by: Kees Cook <keescook@chromium.org>
Reviewed-by: Xiubo Li <xiubli@redhat.com>
cc: Jonathan Corbet <corbet@lwn.net>
cc: Eric Van Hensbergen <ericvh@gmail.com>
cc: Latchesar Ionkov <lucho@ionkov.net>
cc: Dominique Martinet <asmadeus@codewreck.org>
cc: Christian Schoenebeck <linux_oss@crudebyte.com>
cc: Marc Dionne <marc.dionne@auristor.com>
cc: Ilya Dryomov <idryomov@gmail.com>
cc: Steve French <smfrench@gmail.com>
cc: William Kucharski <william.kucharski@oracle.com>
cc: "Matthew Wilcox (Oracle)" <willy@infradead.org>
cc: Dave Chinner <david@fromorbit.com>
cc: linux-doc@vger.kernel.org
cc: v9fs-developer@lists.sourceforge.net
cc: linux-afs@lists.infradead.org
cc: ceph-devel@vger.kernel.org
cc: linux-cifs@vger.kernel.org
cc: samba-technical@lists.samba.org
cc: linux-fsdevel@vger.kernel.org
cc: linux-hardening@vger.kernel.org
Link: https://lore.kernel.org/r/d2ad3a3d7bdd794c6efb562d2f2b655fb67756b9.camel@kernel.org/ [1]
Link: https://lore.kernel.org/r/20220517210230.864239-1-keescook@chromium.org/ [2]
Link: https://lore.kernel.org/r/20220518202212.2322058-1-keescook@chromium.org/ [3]
Link: https://lore.kernel.org/r/20220524101205.GI2306852@dread.disaster.area/ [4]
Link: https://lore.kernel.org/r/165296786831.3591209.12111293034669289733.stgit@warthog.procyon.org.uk/ # v1
Link: https://lore.kernel.org/r/165305805651.4094995.7763502506786714216.stgit@warthog.procyon.org.uk # v2
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2022-06-10 04:46:04 +08:00
|
|
|
struct netfs_inode *ctx = netfs_inode(inode);
|
2022-02-17 21:30:38 +08:00
|
|
|
struct netfs_io_request *rreq;
|
2022-01-21 05:55:46 +08:00
|
|
|
int ret;
|
2022-02-17 21:30:38 +08:00
|
|
|
|
|
|
|
rreq = kzalloc(sizeof(struct netfs_io_request), GFP_KERNEL);
|
2022-01-21 05:55:46 +08:00
|
|
|
if (!rreq)
|
|
|
|
return ERR_PTR(-ENOMEM);
|
|
|
|
|
|
|
|
rreq->start = start;
|
|
|
|
rreq->len = len;
|
|
|
|
rreq->origin = origin;
|
2021-06-30 05:37:05 +08:00
|
|
|
rreq->netfs_ops = ctx->ops;
|
2022-01-21 05:55:46 +08:00
|
|
|
rreq->mapping = mapping;
|
2021-06-30 05:37:05 +08:00
|
|
|
rreq->inode = inode;
|
|
|
|
rreq->i_size = i_size_read(inode);
|
2022-01-21 05:55:46 +08:00
|
|
|
rreq->debug_id = atomic_inc_return(&debug_ids);
|
|
|
|
INIT_LIST_HEAD(&rreq->subrequests);
|
|
|
|
refcount_set(&rreq->ref, 1);
|
|
|
|
__set_bit(NETFS_RREQ_IN_PROGRESS, &rreq->flags);
|
|
|
|
if (rreq->netfs_ops->init_request) {
|
|
|
|
ret = rreq->netfs_ops->init_request(rreq, file);
|
|
|
|
if (ret < 0) {
|
|
|
|
kfree(rreq);
|
|
|
|
return ERR_PTR(ret);
|
|
|
|
}
|
2022-02-17 21:30:38 +08:00
|
|
|
}
|
|
|
|
|
2022-01-21 05:55:46 +08:00
|
|
|
netfs_stat(&netfs_n_rh_rreq);
|
2022-02-17 21:30:38 +08:00
|
|
|
return rreq;
|
|
|
|
}
|
|
|
|
|
2022-02-18 05:13:05 +08:00
|
|
|
void netfs_get_request(struct netfs_io_request *rreq, enum netfs_rreq_ref_trace what)
|
2022-02-17 21:30:38 +08:00
|
|
|
{
|
2022-02-18 05:13:05 +08:00
|
|
|
int r;
|
|
|
|
|
|
|
|
__refcount_inc(&rreq->ref, &r);
|
|
|
|
trace_netfs_rreq_ref(rreq->debug_id, r + 1, what);
|
2022-02-17 21:30:38 +08:00
|
|
|
}
|
|
|
|
|
|
|
|
void netfs_clear_subrequests(struct netfs_io_request *rreq, bool was_async)
|
|
|
|
{
|
|
|
|
struct netfs_io_subrequest *subreq;
|
|
|
|
|
|
|
|
while (!list_empty(&rreq->subrequests)) {
|
|
|
|
subreq = list_first_entry(&rreq->subrequests,
|
|
|
|
struct netfs_io_subrequest, rreq_link);
|
|
|
|
list_del(&subreq->rreq_link);
|
2022-02-17 23:01:24 +08:00
|
|
|
netfs_put_subrequest(subreq, was_async,
|
|
|
|
netfs_sreq_trace_put_clear);
|
2022-02-17 21:30:38 +08:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
static void netfs_free_request(struct work_struct *work)
|
|
|
|
{
|
|
|
|
struct netfs_io_request *rreq =
|
|
|
|
container_of(work, struct netfs_io_request, work);
|
2021-06-30 05:37:05 +08:00
|
|
|
|
2022-02-17 21:30:38 +08:00
|
|
|
trace_netfs_rreq(rreq, netfs_rreq_trace_free);
|
2022-02-25 19:19:14 +08:00
|
|
|
netfs_clear_subrequests(rreq, false);
|
|
|
|
if (rreq->netfs_ops->free_request)
|
|
|
|
rreq->netfs_ops->free_request(rreq);
|
2022-02-17 21:30:38 +08:00
|
|
|
if (rreq->cache_resources.ops)
|
|
|
|
rreq->cache_resources.ops->end_operation(&rreq->cache_resources);
|
|
|
|
kfree(rreq);
|
|
|
|
netfs_stat_d(&netfs_n_rh_rreq);
|
|
|
|
}
|
|
|
|
|
2022-02-18 05:13:05 +08:00
|
|
|
void netfs_put_request(struct netfs_io_request *rreq, bool was_async,
|
|
|
|
enum netfs_rreq_ref_trace what)
|
2022-02-17 21:30:38 +08:00
|
|
|
{
|
2022-02-18 05:13:05 +08:00
|
|
|
unsigned int debug_id = rreq->debug_id;
|
|
|
|
bool dead;
|
|
|
|
int r;
|
|
|
|
|
|
|
|
dead = __refcount_dec_and_test(&rreq->ref, &r);
|
|
|
|
trace_netfs_rreq_ref(debug_id, r - 1, what);
|
|
|
|
if (dead) {
|
2022-02-17 21:30:38 +08:00
|
|
|
if (was_async) {
|
|
|
|
rreq->work.func = netfs_free_request;
|
|
|
|
if (!queue_work(system_unbound_wq, &rreq->work))
|
|
|
|
BUG();
|
|
|
|
} else {
|
|
|
|
netfs_free_request(&rreq->work);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Allocate and partially initialise an I/O request structure.
|
|
|
|
*/
|
|
|
|
struct netfs_io_subrequest *netfs_alloc_subrequest(struct netfs_io_request *rreq)
|
|
|
|
{
|
|
|
|
struct netfs_io_subrequest *subreq;
|
|
|
|
|
|
|
|
subreq = kzalloc(sizeof(struct netfs_io_subrequest), GFP_KERNEL);
|
|
|
|
if (subreq) {
|
|
|
|
INIT_LIST_HEAD(&subreq->rreq_link);
|
2022-02-17 23:01:24 +08:00
|
|
|
refcount_set(&subreq->ref, 2);
|
2022-02-17 21:30:38 +08:00
|
|
|
subreq->rreq = rreq;
|
2022-02-18 05:13:05 +08:00
|
|
|
netfs_get_request(rreq, netfs_rreq_trace_get_subreq);
|
2022-02-17 21:30:38 +08:00
|
|
|
netfs_stat(&netfs_n_rh_sreq);
|
|
|
|
}
|
|
|
|
|
|
|
|
return subreq;
|
|
|
|
}
|
|
|
|
|
2022-02-17 23:01:24 +08:00
|
|
|
void netfs_get_subrequest(struct netfs_io_subrequest *subreq,
|
|
|
|
enum netfs_sreq_ref_trace what)
|
2022-02-17 21:30:38 +08:00
|
|
|
{
|
2022-02-17 23:01:24 +08:00
|
|
|
int r;
|
|
|
|
|
|
|
|
__refcount_inc(&subreq->ref, &r);
|
|
|
|
trace_netfs_sreq_ref(subreq->rreq->debug_id, subreq->debug_index, r + 1,
|
|
|
|
what);
|
2022-02-17 21:30:38 +08:00
|
|
|
}
|
|
|
|
|
2022-02-17 23:01:24 +08:00
|
|
|
static void netfs_free_subrequest(struct netfs_io_subrequest *subreq,
|
|
|
|
bool was_async)
|
2022-02-17 21:30:38 +08:00
|
|
|
{
|
|
|
|
struct netfs_io_request *rreq = subreq->rreq;
|
|
|
|
|
|
|
|
trace_netfs_sreq(subreq, netfs_sreq_trace_free);
|
|
|
|
kfree(subreq);
|
|
|
|
netfs_stat_d(&netfs_n_rh_sreq);
|
2022-02-18 05:13:05 +08:00
|
|
|
netfs_put_request(rreq, was_async, netfs_rreq_trace_put_subreq);
|
2022-02-17 21:30:38 +08:00
|
|
|
}
|
|
|
|
|
2022-02-17 23:01:24 +08:00
|
|
|
void netfs_put_subrequest(struct netfs_io_subrequest *subreq, bool was_async,
|
|
|
|
enum netfs_sreq_ref_trace what)
|
2022-02-17 21:30:38 +08:00
|
|
|
{
|
2022-02-17 23:01:24 +08:00
|
|
|
unsigned int debug_index = subreq->debug_index;
|
|
|
|
unsigned int debug_id = subreq->rreq->debug_id;
|
|
|
|
bool dead;
|
|
|
|
int r;
|
|
|
|
|
|
|
|
dead = __refcount_dec_and_test(&subreq->ref, &r);
|
|
|
|
trace_netfs_sreq_ref(debug_id, debug_index, r - 1, what);
|
|
|
|
if (dead)
|
|
|
|
netfs_free_subrequest(subreq, was_async);
|
2022-02-17 21:30:38 +08:00
|
|
|
}
|