fs/buffer.c: clean up EXPORT* macros
According to Documentation/CodingStyle the EXPORT* macro should follow immediately after the closing function brace line. Also, mark_buffer_async_write_endio() and do_thaw_all() are not used elsewhere so they should be marked as static. In addition, file_fsync() is actually in fs/sync.c so move the EXPORT* to that file. Signed-off-by: H Hartley Sweeten <hsweeten@visionengravers.com> Signed-off-by: Andrew Morton <akpm@linux-foundation.org> Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
This commit is contained in:
parent
88e0fbc452
commit
1fe72eaa0f
57
fs/buffer.c
57
fs/buffer.c
|
@ -52,6 +52,7 @@ init_buffer(struct buffer_head *bh, bh_end_io_t *handler, void *private)
|
|||
bh->b_end_io = handler;
|
||||
bh->b_private = private;
|
||||
}
|
||||
EXPORT_SYMBOL(init_buffer);
|
||||
|
||||
static int sync_buffer(void *word)
|
||||
{
|
||||
|
@ -80,6 +81,7 @@ void unlock_buffer(struct buffer_head *bh)
|
|||
smp_mb__after_clear_bit();
|
||||
wake_up_bit(&bh->b_state, BH_Lock);
|
||||
}
|
||||
EXPORT_SYMBOL(unlock_buffer);
|
||||
|
||||
/*
|
||||
* Block until a buffer comes unlocked. This doesn't stop it
|
||||
|
@ -90,6 +92,7 @@ void __wait_on_buffer(struct buffer_head * bh)
|
|||
{
|
||||
wait_on_bit(&bh->b_state, BH_Lock, sync_buffer, TASK_UNINTERRUPTIBLE);
|
||||
}
|
||||
EXPORT_SYMBOL(__wait_on_buffer);
|
||||
|
||||
static void
|
||||
__clear_page_buffers(struct page *page)
|
||||
|
@ -144,6 +147,7 @@ void end_buffer_read_sync(struct buffer_head *bh, int uptodate)
|
|||
__end_buffer_read_notouch(bh, uptodate);
|
||||
put_bh(bh);
|
||||
}
|
||||
EXPORT_SYMBOL(end_buffer_read_sync);
|
||||
|
||||
void end_buffer_write_sync(struct buffer_head *bh, int uptodate)
|
||||
{
|
||||
|
@ -164,6 +168,7 @@ void end_buffer_write_sync(struct buffer_head *bh, int uptodate)
|
|||
unlock_buffer(bh);
|
||||
put_bh(bh);
|
||||
}
|
||||
EXPORT_SYMBOL(end_buffer_write_sync);
|
||||
|
||||
/*
|
||||
* Various filesystems appear to want __find_get_block to be non-blocking.
|
||||
|
@ -272,6 +277,7 @@ void invalidate_bdev(struct block_device *bdev)
|
|||
invalidate_bh_lrus();
|
||||
invalidate_mapping_pages(mapping, 0, -1);
|
||||
}
|
||||
EXPORT_SYMBOL(invalidate_bdev);
|
||||
|
||||
/*
|
||||
* Kick pdflush then try to free up some ZONE_NORMAL memory.
|
||||
|
@ -410,6 +416,7 @@ still_busy:
|
|||
local_irq_restore(flags);
|
||||
return;
|
||||
}
|
||||
EXPORT_SYMBOL(end_buffer_async_write);
|
||||
|
||||
/*
|
||||
* If a page's buffers are under async readin (end_buffer_async_read
|
||||
|
@ -438,8 +445,8 @@ static void mark_buffer_async_read(struct buffer_head *bh)
|
|||
set_buffer_async_read(bh);
|
||||
}
|
||||
|
||||
void mark_buffer_async_write_endio(struct buffer_head *bh,
|
||||
bh_end_io_t *handler)
|
||||
static void mark_buffer_async_write_endio(struct buffer_head *bh,
|
||||
bh_end_io_t *handler)
|
||||
{
|
||||
bh->b_end_io = handler;
|
||||
set_buffer_async_write(bh);
|
||||
|
@ -553,7 +560,7 @@ repeat:
|
|||
return err;
|
||||
}
|
||||
|
||||
void do_thaw_all(struct work_struct *work)
|
||||
static void do_thaw_all(struct work_struct *work)
|
||||
{
|
||||
struct super_block *sb;
|
||||
char b[BDEVNAME_SIZE];
|
||||
|
@ -1172,6 +1179,7 @@ void mark_buffer_dirty(struct buffer_head *bh)
|
|||
}
|
||||
}
|
||||
}
|
||||
EXPORT_SYMBOL(mark_buffer_dirty);
|
||||
|
||||
/*
|
||||
* Decrement a buffer_head's reference count. If all buffers against a page
|
||||
|
@ -1188,6 +1196,7 @@ void __brelse(struct buffer_head * buf)
|
|||
}
|
||||
WARN(1, KERN_ERR "VFS: brelse: Trying to free free buffer\n");
|
||||
}
|
||||
EXPORT_SYMBOL(__brelse);
|
||||
|
||||
/*
|
||||
* bforget() is like brelse(), except it discards any
|
||||
|
@ -1206,6 +1215,7 @@ void __bforget(struct buffer_head *bh)
|
|||
}
|
||||
__brelse(bh);
|
||||
}
|
||||
EXPORT_SYMBOL(__bforget);
|
||||
|
||||
static struct buffer_head *__bread_slow(struct buffer_head *bh)
|
||||
{
|
||||
|
@ -2218,6 +2228,7 @@ int block_read_full_page(struct page *page, get_block_t *get_block)
|
|||
}
|
||||
return 0;
|
||||
}
|
||||
EXPORT_SYMBOL(block_read_full_page);
|
||||
|
||||
/* utility function for filesystems that need to do work on expanding
|
||||
* truncates. Uses filesystem pagecache writes to allow the filesystem to
|
||||
|
@ -2252,6 +2263,7 @@ int generic_cont_expand_simple(struct inode *inode, loff_t size)
|
|||
out:
|
||||
return err;
|
||||
}
|
||||
EXPORT_SYMBOL(generic_cont_expand_simple);
|
||||
|
||||
static int cont_expand_zero(struct file *file, struct address_space *mapping,
|
||||
loff_t pos, loff_t *bytes)
|
||||
|
@ -2352,6 +2364,7 @@ int cont_write_begin(struct file *file, struct address_space *mapping,
|
|||
out:
|
||||
return err;
|
||||
}
|
||||
EXPORT_SYMBOL(cont_write_begin);
|
||||
|
||||
int block_prepare_write(struct page *page, unsigned from, unsigned to,
|
||||
get_block_t *get_block)
|
||||
|
@ -2362,6 +2375,7 @@ int block_prepare_write(struct page *page, unsigned from, unsigned to,
|
|||
ClearPageUptodate(page);
|
||||
return err;
|
||||
}
|
||||
EXPORT_SYMBOL(block_prepare_write);
|
||||
|
||||
int block_commit_write(struct page *page, unsigned from, unsigned to)
|
||||
{
|
||||
|
@ -2369,6 +2383,7 @@ int block_commit_write(struct page *page, unsigned from, unsigned to)
|
|||
__block_commit_write(inode,page,from,to);
|
||||
return 0;
|
||||
}
|
||||
EXPORT_SYMBOL(block_commit_write);
|
||||
|
||||
/*
|
||||
* block_page_mkwrite() is not allowed to change the file size as it gets
|
||||
|
@ -2426,6 +2441,7 @@ block_page_mkwrite(struct vm_area_struct *vma, struct vm_fault *vmf,
|
|||
out:
|
||||
return ret;
|
||||
}
|
||||
EXPORT_SYMBOL(block_page_mkwrite);
|
||||
|
||||
/*
|
||||
* nobh_write_begin()'s prereads are special: the buffer_heads are freed
|
||||
|
@ -2849,6 +2865,7 @@ unlock:
|
|||
out:
|
||||
return err;
|
||||
}
|
||||
EXPORT_SYMBOL(block_truncate_page);
|
||||
|
||||
/*
|
||||
* The generic ->writepage function for buffer-backed address_spaces
|
||||
|
@ -2890,6 +2907,7 @@ int block_write_full_page_endio(struct page *page, get_block_t *get_block,
|
|||
zero_user_segment(page, offset, PAGE_CACHE_SIZE);
|
||||
return __block_write_full_page(inode, page, get_block, wbc, handler);
|
||||
}
|
||||
EXPORT_SYMBOL(block_write_full_page_endio);
|
||||
|
||||
/*
|
||||
* The generic ->writepage function for buffer-backed address_spaces
|
||||
|
@ -2900,7 +2918,7 @@ int block_write_full_page(struct page *page, get_block_t *get_block,
|
|||
return block_write_full_page_endio(page, get_block, wbc,
|
||||
end_buffer_async_write);
|
||||
}
|
||||
|
||||
EXPORT_SYMBOL(block_write_full_page);
|
||||
|
||||
sector_t generic_block_bmap(struct address_space *mapping, sector_t block,
|
||||
get_block_t *get_block)
|
||||
|
@ -2913,6 +2931,7 @@ sector_t generic_block_bmap(struct address_space *mapping, sector_t block,
|
|||
get_block(inode, block, &tmp, 0);
|
||||
return tmp.b_blocknr;
|
||||
}
|
||||
EXPORT_SYMBOL(generic_block_bmap);
|
||||
|
||||
static void end_bio_bh_io_sync(struct bio *bio, int err)
|
||||
{
|
||||
|
@ -2982,6 +3001,7 @@ int submit_bh(int rw, struct buffer_head * bh)
|
|||
bio_put(bio);
|
||||
return ret;
|
||||
}
|
||||
EXPORT_SYMBOL(submit_bh);
|
||||
|
||||
/**
|
||||
* ll_rw_block: low-level access to block devices (DEPRECATED)
|
||||
|
@ -3043,6 +3063,7 @@ void ll_rw_block(int rw, int nr, struct buffer_head *bhs[])
|
|||
unlock_buffer(bh);
|
||||
}
|
||||
}
|
||||
EXPORT_SYMBOL(ll_rw_block);
|
||||
|
||||
/*
|
||||
* For a data-integrity writeout, we need to wait upon any in-progress I/O
|
||||
|
@ -3071,6 +3092,7 @@ int sync_dirty_buffer(struct buffer_head *bh)
|
|||
}
|
||||
return ret;
|
||||
}
|
||||
EXPORT_SYMBOL(sync_dirty_buffer);
|
||||
|
||||
/*
|
||||
* try_to_free_buffers() checks if all the buffers on this particular page
|
||||
|
@ -3185,6 +3207,7 @@ void block_sync_page(struct page *page)
|
|||
if (mapping)
|
||||
blk_run_backing_dev(mapping->backing_dev_info, page);
|
||||
}
|
||||
EXPORT_SYMBOL(block_sync_page);
|
||||
|
||||
/*
|
||||
* There are no bdflush tunables left. But distributions are
|
||||
|
@ -3361,29 +3384,3 @@ void __init buffer_init(void)
|
|||
max_buffer_heads = nrpages * (PAGE_SIZE / sizeof(struct buffer_head));
|
||||
hotcpu_notifier(buffer_cpu_notify, 0);
|
||||
}
|
||||
|
||||
EXPORT_SYMBOL(__bforget);
|
||||
EXPORT_SYMBOL(__brelse);
|
||||
EXPORT_SYMBOL(__wait_on_buffer);
|
||||
EXPORT_SYMBOL(block_commit_write);
|
||||
EXPORT_SYMBOL(block_prepare_write);
|
||||
EXPORT_SYMBOL(block_page_mkwrite);
|
||||
EXPORT_SYMBOL(block_read_full_page);
|
||||
EXPORT_SYMBOL(block_sync_page);
|
||||
EXPORT_SYMBOL(block_truncate_page);
|
||||
EXPORT_SYMBOL(block_write_full_page);
|
||||
EXPORT_SYMBOL(block_write_full_page_endio);
|
||||
EXPORT_SYMBOL(cont_write_begin);
|
||||
EXPORT_SYMBOL(end_buffer_read_sync);
|
||||
EXPORT_SYMBOL(end_buffer_write_sync);
|
||||
EXPORT_SYMBOL(end_buffer_async_write);
|
||||
EXPORT_SYMBOL(file_fsync);
|
||||
EXPORT_SYMBOL(generic_block_bmap);
|
||||
EXPORT_SYMBOL(generic_cont_expand_simple);
|
||||
EXPORT_SYMBOL(init_buffer);
|
||||
EXPORT_SYMBOL(invalidate_bdev);
|
||||
EXPORT_SYMBOL(ll_rw_block);
|
||||
EXPORT_SYMBOL(mark_buffer_dirty);
|
||||
EXPORT_SYMBOL(submit_bh);
|
||||
EXPORT_SYMBOL(sync_dirty_buffer);
|
||||
EXPORT_SYMBOL(unlock_buffer);
|
||||
|
|
Loading…
Reference in New Issue