Prefix IO_XX flags with XFS_IO_XX to avoid namespace colision.

Add a XFS_ prefix to IO_DIRECT,XFS_IO_DELALLOC, XFS_IO_UNWRITTEN and
XFS_IO_OVERWRITE. This to avoid namespace conflict with other modules.

Signed-off-by: Alain Renaud <arenaud@sgi.com>
Reviewed-by: Rich Johnston <rjohnston@sgi.com>
Signed-off-by: Ben Myers <bpm@sgi.com>
This commit is contained in:
Alain Renaud 2012-05-22 15:56:21 -05:00 committed by Ben Myers
parent 129dbc9a2d
commit 0d882a360b
2 changed files with 31 additions and 31 deletions

View File

@ -179,7 +179,7 @@ xfs_finish_ioend(
if (atomic_dec_and_test(&ioend->io_remaining)) { if (atomic_dec_and_test(&ioend->io_remaining)) {
struct xfs_mount *mp = XFS_I(ioend->io_inode)->i_mount; struct xfs_mount *mp = XFS_I(ioend->io_inode)->i_mount;
if (ioend->io_type == IO_UNWRITTEN) if (ioend->io_type == XFS_IO_UNWRITTEN)
queue_work(mp->m_unwritten_workqueue, &ioend->io_work); queue_work(mp->m_unwritten_workqueue, &ioend->io_work);
else if (ioend->io_append_trans) else if (ioend->io_append_trans)
queue_work(mp->m_data_workqueue, &ioend->io_work); queue_work(mp->m_data_workqueue, &ioend->io_work);
@ -210,7 +210,7 @@ xfs_end_io(
* For unwritten extents we need to issue transactions to convert a * For unwritten extents we need to issue transactions to convert a
* range to normal written extens after the data I/O has finished. * range to normal written extens after the data I/O has finished.
*/ */
if (ioend->io_type == IO_UNWRITTEN) { if (ioend->io_type == XFS_IO_UNWRITTEN) {
/* /*
* For buffered I/O we never preallocate a transaction when * For buffered I/O we never preallocate a transaction when
* doing the unwritten extent conversion, but for direct I/O * doing the unwritten extent conversion, but for direct I/O
@ -312,7 +312,7 @@ xfs_map_blocks(
if (XFS_FORCED_SHUTDOWN(mp)) if (XFS_FORCED_SHUTDOWN(mp))
return -XFS_ERROR(EIO); return -XFS_ERROR(EIO);
if (type == IO_UNWRITTEN) if (type == XFS_IO_UNWRITTEN)
bmapi_flags |= XFS_BMAPI_IGSTATE; bmapi_flags |= XFS_BMAPI_IGSTATE;
if (!xfs_ilock_nowait(ip, XFS_ILOCK_SHARED)) { if (!xfs_ilock_nowait(ip, XFS_ILOCK_SHARED)) {
@ -336,7 +336,7 @@ xfs_map_blocks(
if (error) if (error)
return -XFS_ERROR(error); return -XFS_ERROR(error);
if (type == IO_DELALLOC && if (type == XFS_IO_DELALLOC &&
(!nimaps || isnullstartblock(imap->br_startblock))) { (!nimaps || isnullstartblock(imap->br_startblock))) {
error = xfs_iomap_write_allocate(ip, offset, count, imap); error = xfs_iomap_write_allocate(ip, offset, count, imap);
if (!error) if (!error)
@ -345,7 +345,7 @@ xfs_map_blocks(
} }
#ifdef DEBUG #ifdef DEBUG
if (type == IO_UNWRITTEN) { if (type == XFS_IO_UNWRITTEN) {
ASSERT(nimaps); ASSERT(nimaps);
ASSERT(imap->br_startblock != HOLESTARTBLOCK); ASSERT(imap->br_startblock != HOLESTARTBLOCK);
ASSERT(imap->br_startblock != DELAYSTARTBLOCK); ASSERT(imap->br_startblock != DELAYSTARTBLOCK);
@ -634,11 +634,11 @@ xfs_check_page_type(
bh = head = page_buffers(page); bh = head = page_buffers(page);
do { do {
if (buffer_unwritten(bh)) if (buffer_unwritten(bh))
acceptable += (type == IO_UNWRITTEN); acceptable += (type == XFS_IO_UNWRITTEN);
else if (buffer_delay(bh)) else if (buffer_delay(bh))
acceptable += (type == IO_DELALLOC); acceptable += (type == XFS_IO_DELALLOC);
else if (buffer_dirty(bh) && buffer_mapped(bh)) else if (buffer_dirty(bh) && buffer_mapped(bh))
acceptable += (type == IO_OVERWRITE); acceptable += (type == XFS_IO_OVERWRITE);
else else
break; break;
} while ((bh = bh->b_this_page) != head); } while ((bh = bh->b_this_page) != head);
@ -721,11 +721,11 @@ xfs_convert_page(
if (buffer_unwritten(bh) || buffer_delay(bh) || if (buffer_unwritten(bh) || buffer_delay(bh) ||
buffer_mapped(bh)) { buffer_mapped(bh)) {
if (buffer_unwritten(bh)) if (buffer_unwritten(bh))
type = IO_UNWRITTEN; type = XFS_IO_UNWRITTEN;
else if (buffer_delay(bh)) else if (buffer_delay(bh))
type = IO_DELALLOC; type = XFS_IO_DELALLOC;
else else
type = IO_OVERWRITE; type = XFS_IO_OVERWRITE;
if (!xfs_imap_valid(inode, imap, offset)) { if (!xfs_imap_valid(inode, imap, offset)) {
done = 1; done = 1;
@ -733,7 +733,7 @@ xfs_convert_page(
} }
lock_buffer(bh); lock_buffer(bh);
if (type != IO_OVERWRITE) if (type != XFS_IO_OVERWRITE)
xfs_map_at_offset(inode, bh, imap, offset); xfs_map_at_offset(inode, bh, imap, offset);
xfs_add_to_ioend(inode, bh, offset, type, xfs_add_to_ioend(inode, bh, offset, type,
ioendp, done); ioendp, done);
@ -831,7 +831,7 @@ xfs_aops_discard_page(
struct buffer_head *bh, *head; struct buffer_head *bh, *head;
loff_t offset = page_offset(page); loff_t offset = page_offset(page);
if (!xfs_check_page_type(page, IO_DELALLOC)) if (!xfs_check_page_type(page, XFS_IO_DELALLOC))
goto out_invalidate; goto out_invalidate;
if (XFS_FORCED_SHUTDOWN(ip->i_mount)) if (XFS_FORCED_SHUTDOWN(ip->i_mount))
@ -956,7 +956,7 @@ xfs_vm_writepage(
bh = head = page_buffers(page); bh = head = page_buffers(page);
offset = page_offset(page); offset = page_offset(page);
type = IO_OVERWRITE; type = XFS_IO_OVERWRITE;
if (wbc->sync_mode == WB_SYNC_NONE) if (wbc->sync_mode == WB_SYNC_NONE)
nonblocking = 1; nonblocking = 1;
@ -981,18 +981,18 @@ xfs_vm_writepage(
} }
if (buffer_unwritten(bh)) { if (buffer_unwritten(bh)) {
if (type != IO_UNWRITTEN) { if (type != XFS_IO_UNWRITTEN) {
type = IO_UNWRITTEN; type = XFS_IO_UNWRITTEN;
imap_valid = 0; imap_valid = 0;
} }
} else if (buffer_delay(bh)) { } else if (buffer_delay(bh)) {
if (type != IO_DELALLOC) { if (type != XFS_IO_DELALLOC) {
type = IO_DELALLOC; type = XFS_IO_DELALLOC;
imap_valid = 0; imap_valid = 0;
} }
} else if (buffer_uptodate(bh)) { } else if (buffer_uptodate(bh)) {
if (type != IO_OVERWRITE) { if (type != XFS_IO_OVERWRITE) {
type = IO_OVERWRITE; type = XFS_IO_OVERWRITE;
imap_valid = 0; imap_valid = 0;
} }
} else { } else {
@ -1028,7 +1028,7 @@ xfs_vm_writepage(
} }
if (imap_valid) { if (imap_valid) {
lock_buffer(bh); lock_buffer(bh);
if (type != IO_OVERWRITE) if (type != XFS_IO_OVERWRITE)
xfs_map_at_offset(inode, bh, &imap, offset); xfs_map_at_offset(inode, bh, &imap, offset);
xfs_add_to_ioend(inode, bh, offset, type, &ioend, xfs_add_to_ioend(inode, bh, offset, type, &ioend,
new_ioend); new_ioend);
@ -1069,7 +1069,7 @@ xfs_vm_writepage(
* Reserve log space if we might write beyond the on-disk * Reserve log space if we might write beyond the on-disk
* inode size. * inode size.
*/ */
if (ioend->io_type != IO_UNWRITTEN && if (ioend->io_type != XFS_IO_UNWRITTEN &&
xfs_ioend_is_append(ioend)) { xfs_ioend_is_append(ioend)) {
err = xfs_setfilesize_trans_alloc(ioend); err = xfs_setfilesize_trans_alloc(ioend);
if (err) if (err)
@ -1366,7 +1366,7 @@ xfs_end_io_direct_write(
ioend->io_iocb = iocb; ioend->io_iocb = iocb;
ioend->io_result = ret; ioend->io_result = ret;
if (private && size > 0) if (private && size > 0)
ioend->io_type = IO_UNWRITTEN; ioend->io_type = XFS_IO_UNWRITTEN;
if (is_async) { if (is_async) {
ioend->io_isasync = 1; ioend->io_isasync = 1;
@ -1398,7 +1398,7 @@ xfs_vm_direct_IO(
* and converts at least on unwritten extent we will cancel * and converts at least on unwritten extent we will cancel
* the still clean transaction after the I/O has finished. * the still clean transaction after the I/O has finished.
*/ */
iocb->private = ioend = xfs_alloc_ioend(inode, IO_DIRECT); iocb->private = ioend = xfs_alloc_ioend(inode, XFS_IO_DIRECT);
if (offset + size > XFS_I(inode)->i_d.di_size) { if (offset + size > XFS_I(inode)->i_d.di_size) {
ret = xfs_setfilesize_trans_alloc(ioend); ret = xfs_setfilesize_trans_alloc(ioend);
if (ret) if (ret)

View File

@ -24,17 +24,17 @@ extern mempool_t *xfs_ioend_pool;
* Types of I/O for bmap clustering and I/O completion tracking. * Types of I/O for bmap clustering and I/O completion tracking.
*/ */
enum { enum {
IO_DIRECT = 0, /* special case for direct I/O ioends */ XFS_IO_DIRECT = 0, /* special case for direct I/O ioends */
IO_DELALLOC, /* mapping covers delalloc region */ XFS_IO_DELALLOC, /* covers delalloc region */
IO_UNWRITTEN, /* mapping covers allocated but uninitialized data */ XFS_IO_UNWRITTEN, /* covers allocated but uninitialized data */
IO_OVERWRITE, /* mapping covers already allocated extent */ XFS_IO_OVERWRITE, /* covers already allocated extent */
}; };
#define XFS_IO_TYPES \ #define XFS_IO_TYPES \
{ 0, "" }, \ { 0, "" }, \
{ IO_DELALLOC, "delalloc" }, \ { XFS_IO_DELALLOC, "delalloc" }, \
{ IO_UNWRITTEN, "unwritten" }, \ { XFS_IO_UNWRITTEN, "unwritten" }, \
{ IO_OVERWRITE, "overwrite" } { XFS_IO_OVERWRITE, "overwrite" }
/* /*
* xfs_ioend struct manages large extent writes for XFS. * xfs_ioend struct manages large extent writes for XFS.