callers of iov_copy_from_user_atomic() don't need pagecache_disable()
... it does that itself (via kmap_atomic()) Signed-off-by: Al Viro <viro@zeniv.linux.org.uk>
This commit is contained in:
parent
c186afb4db
commit
9e8c2af96e
|
@ -425,13 +425,8 @@ static noinline int btrfs_copy_from_user(loff_t pos, int num_pages,
|
|||
struct page *page = prepared_pages[pg];
|
||||
/*
|
||||
* Copy data from userspace to the current page
|
||||
*
|
||||
* Disable pagefault to avoid recursive lock since
|
||||
* the pages are already locked
|
||||
*/
|
||||
pagefault_disable();
|
||||
copied = iov_iter_copy_from_user_atomic(page, i, offset, count);
|
||||
pagefault_enable();
|
||||
|
||||
/* Flush processor's dcache for this page */
|
||||
flush_dcache_page(page);
|
||||
|
|
|
@ -1003,9 +1003,7 @@ static ssize_t fuse_fill_write_pages(struct fuse_req *req,
|
|||
if (mapping_writably_mapped(mapping))
|
||||
flush_dcache_page(page);
|
||||
|
||||
pagefault_disable();
|
||||
tmp = iov_iter_copy_from_user_atomic(page, ii, offset, bytes);
|
||||
pagefault_enable();
|
||||
flush_dcache_page(page);
|
||||
|
||||
mark_page_accessed(page);
|
||||
|
|
|
@ -1974,7 +1974,6 @@ size_t iov_iter_copy_from_user_atomic(struct page *page,
|
|||
char *kaddr;
|
||||
size_t copied;
|
||||
|
||||
BUG_ON(!in_atomic());
|
||||
kaddr = kmap_atomic(page);
|
||||
if (likely(i->nr_segs == 1)) {
|
||||
int left;
|
||||
|
@ -2348,9 +2347,7 @@ again:
|
|||
if (mapping_writably_mapped(mapping))
|
||||
flush_dcache_page(page);
|
||||
|
||||
pagefault_disable();
|
||||
copied = iov_iter_copy_from_user_atomic(page, i, offset, bytes);
|
||||
pagefault_enable();
|
||||
flush_dcache_page(page);
|
||||
|
||||
mark_page_accessed(page);
|
||||
|
|
Loading…
Reference in New Issue