ext4: Drop whitespace at end of lines
This patch was generated using: #!/usr/bin/perl -i while (<>) { s/[ ]+$//; print; } Signed-off-by: "Theodore Ts'o" <tytso@mit.edu>
This commit is contained in:
parent
4d92dc0f00
commit
60e6679e28
|
@ -72,9 +72,9 @@ static int add_system_zone(struct ext4_sb_info *sbi,
|
||||||
else if (start_blk >= (entry->start_blk + entry->count))
|
else if (start_blk >= (entry->start_blk + entry->count))
|
||||||
n = &(*n)->rb_right;
|
n = &(*n)->rb_right;
|
||||||
else {
|
else {
|
||||||
if (start_blk + count > (entry->start_blk +
|
if (start_blk + count > (entry->start_blk +
|
||||||
entry->count))
|
entry->count))
|
||||||
entry->count = (start_blk + count -
|
entry->count = (start_blk + count -
|
||||||
entry->start_blk);
|
entry->start_blk);
|
||||||
new_node = *n;
|
new_node = *n;
|
||||||
new_entry = rb_entry(new_node, struct ext4_system_zone,
|
new_entry = rb_entry(new_node, struct ext4_system_zone,
|
||||||
|
|
|
@ -75,7 +75,7 @@ typedef __u32 ext4_lblk_t;
|
||||||
typedef unsigned int ext4_group_t;
|
typedef unsigned int ext4_group_t;
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Flags used in mballoc's allocation_context flags field.
|
* Flags used in mballoc's allocation_context flags field.
|
||||||
*
|
*
|
||||||
* Also used to show what's going on for debugging purposes when the
|
* Also used to show what's going on for debugging purposes when the
|
||||||
* flag field is exported via the traceport interface
|
* flag field is exported via the traceport interface
|
||||||
|
@ -470,7 +470,7 @@ struct ext4_new_group_data {
|
||||||
#define EXT4_GET_BLOCKS_CREATE_UNINIT_EXT (EXT4_GET_BLOCKS_UNINIT_EXT|\
|
#define EXT4_GET_BLOCKS_CREATE_UNINIT_EXT (EXT4_GET_BLOCKS_UNINIT_EXT|\
|
||||||
EXT4_GET_BLOCKS_CREATE)
|
EXT4_GET_BLOCKS_CREATE)
|
||||||
/* Caller is from the delayed allocation writeout path,
|
/* Caller is from the delayed allocation writeout path,
|
||||||
so set the magic i_delalloc_reserve_flag after taking the
|
so set the magic i_delalloc_reserve_flag after taking the
|
||||||
inode allocation semaphore for */
|
inode allocation semaphore for */
|
||||||
#define EXT4_GET_BLOCKS_DELALLOC_RESERVE 0x0004
|
#define EXT4_GET_BLOCKS_DELALLOC_RESERVE 0x0004
|
||||||
/* caller is from the direct IO path, request to creation of an
|
/* caller is from the direct IO path, request to creation of an
|
||||||
|
|
|
@ -182,10 +182,10 @@ static ext4_fsblk_t ext4_ext_find_goal(struct inode *inode,
|
||||||
if (flex_size >= EXT4_FLEX_SIZE_DIR_ALLOC_SCHEME) {
|
if (flex_size >= EXT4_FLEX_SIZE_DIR_ALLOC_SCHEME) {
|
||||||
/*
|
/*
|
||||||
* If there are at least EXT4_FLEX_SIZE_DIR_ALLOC_SCHEME
|
* If there are at least EXT4_FLEX_SIZE_DIR_ALLOC_SCHEME
|
||||||
* block groups per flexgroup, reserve the first block
|
* block groups per flexgroup, reserve the first block
|
||||||
* group for directories and special files. Regular
|
* group for directories and special files. Regular
|
||||||
* files will start at the second block group. This
|
* files will start at the second block group. This
|
||||||
* tends to speed up directory access and improves
|
* tends to speed up directory access and improves
|
||||||
* fsck times.
|
* fsck times.
|
||||||
*/
|
*/
|
||||||
block_group &= ~(flex_size-1);
|
block_group &= ~(flex_size-1);
|
||||||
|
@ -2034,7 +2034,7 @@ ext4_ext_in_cache(struct inode *inode, ext4_lblk_t block,
|
||||||
struct ext4_ext_cache *cex;
|
struct ext4_ext_cache *cex;
|
||||||
int ret = EXT4_EXT_CACHE_NO;
|
int ret = EXT4_EXT_CACHE_NO;
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* We borrow i_block_reservation_lock to protect i_cached_extent
|
* We borrow i_block_reservation_lock to protect i_cached_extent
|
||||||
*/
|
*/
|
||||||
spin_lock(&EXT4_I(inode)->i_block_reservation_lock);
|
spin_lock(&EXT4_I(inode)->i_block_reservation_lock);
|
||||||
|
|
|
@ -66,7 +66,7 @@ int ext4_sync_file(struct file *file, struct dentry *dentry, int datasync)
|
||||||
ret = flush_completed_IO(inode);
|
ret = flush_completed_IO(inode);
|
||||||
if (ret < 0)
|
if (ret < 0)
|
||||||
return ret;
|
return ret;
|
||||||
|
|
||||||
if (!journal)
|
if (!journal)
|
||||||
return simple_fsync(file, dentry, datasync);
|
return simple_fsync(file, dentry, datasync);
|
||||||
|
|
||||||
|
|
|
@ -784,7 +784,7 @@ failed:
|
||||||
/* Allocation failed, free what we already allocated */
|
/* Allocation failed, free what we already allocated */
|
||||||
ext4_free_blocks(handle, inode, 0, new_blocks[0], 1, 0);
|
ext4_free_blocks(handle, inode, 0, new_blocks[0], 1, 0);
|
||||||
for (i = 1; i <= n ; i++) {
|
for (i = 1; i <= n ; i++) {
|
||||||
/*
|
/*
|
||||||
* branch[i].bh is newly allocated, so there is no
|
* branch[i].bh is newly allocated, so there is no
|
||||||
* need to revoke the block, which is why we don't
|
* need to revoke the block, which is why we don't
|
||||||
* need to set EXT4_FREE_BLOCKS_METADATA.
|
* need to set EXT4_FREE_BLOCKS_METADATA.
|
||||||
|
@ -874,7 +874,7 @@ static int ext4_splice_branch(handle_t *handle, struct inode *inode,
|
||||||
|
|
||||||
err_out:
|
err_out:
|
||||||
for (i = 1; i <= num; i++) {
|
for (i = 1; i <= num; i++) {
|
||||||
/*
|
/*
|
||||||
* branch[i].bh is newly allocated, so there is no
|
* branch[i].bh is newly allocated, so there is no
|
||||||
* need to revoke the block, which is why we don't
|
* need to revoke the block, which is why we don't
|
||||||
* need to set EXT4_FREE_BLOCKS_METADATA.
|
* need to set EXT4_FREE_BLOCKS_METADATA.
|
||||||
|
|
|
@ -258,7 +258,7 @@ setversion_out:
|
||||||
if (me.moved_len > 0)
|
if (me.moved_len > 0)
|
||||||
file_remove_suid(donor_filp);
|
file_remove_suid(donor_filp);
|
||||||
|
|
||||||
if (copy_to_user((struct move_extent __user *)arg,
|
if (copy_to_user((struct move_extent __user *)arg,
|
||||||
&me, sizeof(me)))
|
&me, sizeof(me)))
|
||||||
err = -EFAULT;
|
err = -EFAULT;
|
||||||
mext_out:
|
mext_out:
|
||||||
|
|
|
@ -2691,7 +2691,7 @@ int __init init_ext4_mballoc(void)
|
||||||
|
|
||||||
void exit_ext4_mballoc(void)
|
void exit_ext4_mballoc(void)
|
||||||
{
|
{
|
||||||
/*
|
/*
|
||||||
* Wait for completion of call_rcu()'s on ext4_pspace_cachep
|
* Wait for completion of call_rcu()'s on ext4_pspace_cachep
|
||||||
* before destroying the slab cache.
|
* before destroying the slab cache.
|
||||||
*/
|
*/
|
||||||
|
@ -3330,7 +3330,7 @@ static void ext4_mb_put_pa(struct ext4_allocation_context *ac,
|
||||||
spin_unlock(&pa->pa_lock);
|
spin_unlock(&pa->pa_lock);
|
||||||
|
|
||||||
grp_blk = pa->pa_pstart;
|
grp_blk = pa->pa_pstart;
|
||||||
/*
|
/*
|
||||||
* If doing group-based preallocation, pa_pstart may be in the
|
* If doing group-based preallocation, pa_pstart may be in the
|
||||||
* next group when pa is used up
|
* next group when pa is used up
|
||||||
*/
|
*/
|
||||||
|
@ -4534,12 +4534,12 @@ void ext4_free_blocks(handle_t *handle, struct inode *inode,
|
||||||
if (!bh)
|
if (!bh)
|
||||||
tbh = sb_find_get_block(inode->i_sb,
|
tbh = sb_find_get_block(inode->i_sb,
|
||||||
block + i);
|
block + i);
|
||||||
ext4_forget(handle, flags & EXT4_FREE_BLOCKS_METADATA,
|
ext4_forget(handle, flags & EXT4_FREE_BLOCKS_METADATA,
|
||||||
inode, tbh, block + i);
|
inode, tbh, block + i);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* We need to make sure we don't reuse the freed block until
|
* We need to make sure we don't reuse the freed block until
|
||||||
* after the transaction is committed, which we can do by
|
* after the transaction is committed, which we can do by
|
||||||
* treating the block as metadata, below. We make an
|
* treating the block as metadata, below. We make an
|
||||||
|
|
|
@ -187,7 +187,7 @@ unsigned int ext4_rec_len_from_disk(__le16 dlen, unsigned blocksize)
|
||||||
return blocksize;
|
return blocksize;
|
||||||
return (len & 65532) | ((len & 3) << 16);
|
return (len & 65532) | ((len & 3) << 16);
|
||||||
}
|
}
|
||||||
|
|
||||||
__le16 ext4_rec_len_to_disk(unsigned len, unsigned blocksize)
|
__le16 ext4_rec_len_to_disk(unsigned len, unsigned blocksize)
|
||||||
{
|
{
|
||||||
if ((len > blocksize) || (blocksize > (1 << 18)) || (len & 3))
|
if ((len > blocksize) || (blocksize > (1 << 18)) || (len & 3))
|
||||||
|
@ -197,7 +197,7 @@ __le16 ext4_rec_len_to_disk(unsigned len, unsigned blocksize)
|
||||||
if (len == blocksize) {
|
if (len == blocksize) {
|
||||||
if (blocksize == 65536)
|
if (blocksize == 65536)
|
||||||
return cpu_to_le16(EXT4_MAX_REC_LEN);
|
return cpu_to_le16(EXT4_MAX_REC_LEN);
|
||||||
else
|
else
|
||||||
return cpu_to_le16(0);
|
return cpu_to_le16(0);
|
||||||
}
|
}
|
||||||
return cpu_to_le16((len & 65532) | ((len >> 16) & 3));
|
return cpu_to_le16((len & 65532) | ((len >> 16) & 3));
|
||||||
|
@ -349,7 +349,7 @@ struct stats dx_show_entries(struct dx_hash_info *hinfo, struct inode *dir,
|
||||||
brelse(bh);
|
brelse(bh);
|
||||||
}
|
}
|
||||||
if (bcount)
|
if (bcount)
|
||||||
printk(KERN_DEBUG "%snames %u, fullness %u (%u%%)\n",
|
printk(KERN_DEBUG "%snames %u, fullness %u (%u%%)\n",
|
||||||
levels ? "" : " ", names, space/bcount,
|
levels ? "" : " ", names, space/bcount,
|
||||||
(space/bcount)*100/blocksize);
|
(space/bcount)*100/blocksize);
|
||||||
return (struct stats) { names, space, bcount};
|
return (struct stats) { names, space, bcount};
|
||||||
|
@ -653,7 +653,7 @@ int ext4_htree_fill_tree(struct file *dir_file, __u32 start_hash,
|
||||||
int ret, err;
|
int ret, err;
|
||||||
__u32 hashval;
|
__u32 hashval;
|
||||||
|
|
||||||
dxtrace(printk(KERN_DEBUG "In htree_fill_tree, start hash: %x:%x\n",
|
dxtrace(printk(KERN_DEBUG "In htree_fill_tree, start hash: %x:%x\n",
|
||||||
start_hash, start_minor_hash));
|
start_hash, start_minor_hash));
|
||||||
dir = dir_file->f_path.dentry->d_inode;
|
dir = dir_file->f_path.dentry->d_inode;
|
||||||
if (!(ext4_test_inode_flag(dir, EXT4_INODE_INDEX))) {
|
if (!(ext4_test_inode_flag(dir, EXT4_INODE_INDEX))) {
|
||||||
|
@ -1141,7 +1141,7 @@ dx_move_dirents(char *from, char *to, struct dx_map_entry *map, int count,
|
||||||
unsigned rec_len = 0;
|
unsigned rec_len = 0;
|
||||||
|
|
||||||
while (count--) {
|
while (count--) {
|
||||||
struct ext4_dir_entry_2 *de = (struct ext4_dir_entry_2 *)
|
struct ext4_dir_entry_2 *de = (struct ext4_dir_entry_2 *)
|
||||||
(from + (map->offs<<2));
|
(from + (map->offs<<2));
|
||||||
rec_len = EXT4_DIR_REC_LEN(de->name_len);
|
rec_len = EXT4_DIR_REC_LEN(de->name_len);
|
||||||
memcpy (to, de, rec_len);
|
memcpy (to, de, rec_len);
|
||||||
|
|
|
@ -2216,7 +2216,7 @@ static unsigned long ext4_get_stripe_size(struct ext4_sb_info *sbi)
|
||||||
struct ext4_attr {
|
struct ext4_attr {
|
||||||
struct attribute attr;
|
struct attribute attr;
|
||||||
ssize_t (*show)(struct ext4_attr *, struct ext4_sb_info *, char *);
|
ssize_t (*show)(struct ext4_attr *, struct ext4_sb_info *, char *);
|
||||||
ssize_t (*store)(struct ext4_attr *, struct ext4_sb_info *,
|
ssize_t (*store)(struct ext4_attr *, struct ext4_sb_info *,
|
||||||
const char *, size_t);
|
const char *, size_t);
|
||||||
int offset;
|
int offset;
|
||||||
};
|
};
|
||||||
|
@ -3383,7 +3383,7 @@ static int ext4_commit_super(struct super_block *sb, int sync)
|
||||||
if (!(sb->s_flags & MS_RDONLY))
|
if (!(sb->s_flags & MS_RDONLY))
|
||||||
es->s_wtime = cpu_to_le32(get_seconds());
|
es->s_wtime = cpu_to_le32(get_seconds());
|
||||||
es->s_kbytes_written =
|
es->s_kbytes_written =
|
||||||
cpu_to_le64(EXT4_SB(sb)->s_kbytes_written +
|
cpu_to_le64(EXT4_SB(sb)->s_kbytes_written +
|
||||||
((part_stat_read(sb->s_bdev->bd_part, sectors[1]) -
|
((part_stat_read(sb->s_bdev->bd_part, sectors[1]) -
|
||||||
EXT4_SB(sb)->s_sectors_written_start) >> 1));
|
EXT4_SB(sb)->s_sectors_written_start) >> 1));
|
||||||
ext4_free_blocks_count_set(es, percpu_counter_sum_positive(
|
ext4_free_blocks_count_set(es, percpu_counter_sum_positive(
|
||||||
|
|
Loading…
Reference in a new issue