ext4: Drop whitespace at end of lines
This patch was generated using: #!/usr/bin/perl -i while (<>) { s/[ ]+$//; print; } Signed-off-by: "Theodore Ts'o" <tytso@mit.edu>
This commit is contained in:
Родитель
4d92dc0f00
Коммит
60e6679e28
|
@ -72,9 +72,9 @@ static int add_system_zone(struct ext4_sb_info *sbi,
|
|||
else if (start_blk >= (entry->start_blk + entry->count))
|
||||
n = &(*n)->rb_right;
|
||||
else {
|
||||
if (start_blk + count > (entry->start_blk +
|
||||
if (start_blk + count > (entry->start_blk +
|
||||
entry->count))
|
||||
entry->count = (start_blk + count -
|
||||
entry->count = (start_blk + count -
|
||||
entry->start_blk);
|
||||
new_node = *n;
|
||||
new_entry = rb_entry(new_node, struct ext4_system_zone,
|
||||
|
|
|
@ -75,7 +75,7 @@ typedef __u32 ext4_lblk_t;
|
|||
typedef unsigned int ext4_group_t;
|
||||
|
||||
/*
|
||||
* Flags used in mballoc's allocation_context flags field.
|
||||
* Flags used in mballoc's allocation_context flags field.
|
||||
*
|
||||
* Also used to show what's going on for debugging purposes when the
|
||||
* flag field is exported via the traceport interface
|
||||
|
@ -470,7 +470,7 @@ struct ext4_new_group_data {
|
|||
#define EXT4_GET_BLOCKS_CREATE_UNINIT_EXT (EXT4_GET_BLOCKS_UNINIT_EXT|\
|
||||
EXT4_GET_BLOCKS_CREATE)
|
||||
/* Caller is from the delayed allocation writeout path,
|
||||
so set the magic i_delalloc_reserve_flag after taking the
|
||||
so set the magic i_delalloc_reserve_flag after taking the
|
||||
inode allocation semaphore for */
|
||||
#define EXT4_GET_BLOCKS_DELALLOC_RESERVE 0x0004
|
||||
/* caller is from the direct IO path, request to creation of an
|
||||
|
|
|
@ -182,10 +182,10 @@ static ext4_fsblk_t ext4_ext_find_goal(struct inode *inode,
|
|||
if (flex_size >= EXT4_FLEX_SIZE_DIR_ALLOC_SCHEME) {
|
||||
/*
|
||||
* If there are at least EXT4_FLEX_SIZE_DIR_ALLOC_SCHEME
|
||||
* block groups per flexgroup, reserve the first block
|
||||
* group for directories and special files. Regular
|
||||
* block groups per flexgroup, reserve the first block
|
||||
* group for directories and special files. Regular
|
||||
* files will start at the second block group. This
|
||||
* tends to speed up directory access and improves
|
||||
* tends to speed up directory access and improves
|
||||
* fsck times.
|
||||
*/
|
||||
block_group &= ~(flex_size-1);
|
||||
|
@ -2034,7 +2034,7 @@ ext4_ext_in_cache(struct inode *inode, ext4_lblk_t block,
|
|||
struct ext4_ext_cache *cex;
|
||||
int ret = EXT4_EXT_CACHE_NO;
|
||||
|
||||
/*
|
||||
/*
|
||||
* We borrow i_block_reservation_lock to protect i_cached_extent
|
||||
*/
|
||||
spin_lock(&EXT4_I(inode)->i_block_reservation_lock);
|
||||
|
|
|
@ -66,7 +66,7 @@ int ext4_sync_file(struct file *file, struct dentry *dentry, int datasync)
|
|||
ret = flush_completed_IO(inode);
|
||||
if (ret < 0)
|
||||
return ret;
|
||||
|
||||
|
||||
if (!journal)
|
||||
return simple_fsync(file, dentry, datasync);
|
||||
|
||||
|
|
|
@ -784,7 +784,7 @@ failed:
|
|||
/* Allocation failed, free what we already allocated */
|
||||
ext4_free_blocks(handle, inode, 0, new_blocks[0], 1, 0);
|
||||
for (i = 1; i <= n ; i++) {
|
||||
/*
|
||||
/*
|
||||
* branch[i].bh is newly allocated, so there is no
|
||||
* need to revoke the block, which is why we don't
|
||||
* need to set EXT4_FREE_BLOCKS_METADATA.
|
||||
|
@ -874,7 +874,7 @@ static int ext4_splice_branch(handle_t *handle, struct inode *inode,
|
|||
|
||||
err_out:
|
||||
for (i = 1; i <= num; i++) {
|
||||
/*
|
||||
/*
|
||||
* branch[i].bh is newly allocated, so there is no
|
||||
* need to revoke the block, which is why we don't
|
||||
* need to set EXT4_FREE_BLOCKS_METADATA.
|
||||
|
|
|
@ -258,7 +258,7 @@ setversion_out:
|
|||
if (me.moved_len > 0)
|
||||
file_remove_suid(donor_filp);
|
||||
|
||||
if (copy_to_user((struct move_extent __user *)arg,
|
||||
if (copy_to_user((struct move_extent __user *)arg,
|
||||
&me, sizeof(me)))
|
||||
err = -EFAULT;
|
||||
mext_out:
|
||||
|
|
|
@ -2691,7 +2691,7 @@ int __init init_ext4_mballoc(void)
|
|||
|
||||
void exit_ext4_mballoc(void)
|
||||
{
|
||||
/*
|
||||
/*
|
||||
* Wait for completion of call_rcu()'s on ext4_pspace_cachep
|
||||
* before destroying the slab cache.
|
||||
*/
|
||||
|
@ -3330,7 +3330,7 @@ static void ext4_mb_put_pa(struct ext4_allocation_context *ac,
|
|||
spin_unlock(&pa->pa_lock);
|
||||
|
||||
grp_blk = pa->pa_pstart;
|
||||
/*
|
||||
/*
|
||||
* If doing group-based preallocation, pa_pstart may be in the
|
||||
* next group when pa is used up
|
||||
*/
|
||||
|
@ -4534,12 +4534,12 @@ void ext4_free_blocks(handle_t *handle, struct inode *inode,
|
|||
if (!bh)
|
||||
tbh = sb_find_get_block(inode->i_sb,
|
||||
block + i);
|
||||
ext4_forget(handle, flags & EXT4_FREE_BLOCKS_METADATA,
|
||||
ext4_forget(handle, flags & EXT4_FREE_BLOCKS_METADATA,
|
||||
inode, tbh, block + i);
|
||||
}
|
||||
}
|
||||
|
||||
/*
|
||||
/*
|
||||
* We need to make sure we don't reuse the freed block until
|
||||
* after the transaction is committed, which we can do by
|
||||
* treating the block as metadata, below. We make an
|
||||
|
|
|
@ -187,7 +187,7 @@ unsigned int ext4_rec_len_from_disk(__le16 dlen, unsigned blocksize)
|
|||
return blocksize;
|
||||
return (len & 65532) | ((len & 3) << 16);
|
||||
}
|
||||
|
||||
|
||||
__le16 ext4_rec_len_to_disk(unsigned len, unsigned blocksize)
|
||||
{
|
||||
if ((len > blocksize) || (blocksize > (1 << 18)) || (len & 3))
|
||||
|
@ -197,7 +197,7 @@ __le16 ext4_rec_len_to_disk(unsigned len, unsigned blocksize)
|
|||
if (len == blocksize) {
|
||||
if (blocksize == 65536)
|
||||
return cpu_to_le16(EXT4_MAX_REC_LEN);
|
||||
else
|
||||
else
|
||||
return cpu_to_le16(0);
|
||||
}
|
||||
return cpu_to_le16((len & 65532) | ((len >> 16) & 3));
|
||||
|
@ -349,7 +349,7 @@ struct stats dx_show_entries(struct dx_hash_info *hinfo, struct inode *dir,
|
|||
brelse(bh);
|
||||
}
|
||||
if (bcount)
|
||||
printk(KERN_DEBUG "%snames %u, fullness %u (%u%%)\n",
|
||||
printk(KERN_DEBUG "%snames %u, fullness %u (%u%%)\n",
|
||||
levels ? "" : " ", names, space/bcount,
|
||||
(space/bcount)*100/blocksize);
|
||||
return (struct stats) { names, space, bcount};
|
||||
|
@ -653,7 +653,7 @@ int ext4_htree_fill_tree(struct file *dir_file, __u32 start_hash,
|
|||
int ret, err;
|
||||
__u32 hashval;
|
||||
|
||||
dxtrace(printk(KERN_DEBUG "In htree_fill_tree, start hash: %x:%x\n",
|
||||
dxtrace(printk(KERN_DEBUG "In htree_fill_tree, start hash: %x:%x\n",
|
||||
start_hash, start_minor_hash));
|
||||
dir = dir_file->f_path.dentry->d_inode;
|
||||
if (!(ext4_test_inode_flag(dir, EXT4_INODE_INDEX))) {
|
||||
|
@ -1141,7 +1141,7 @@ dx_move_dirents(char *from, char *to, struct dx_map_entry *map, int count,
|
|||
unsigned rec_len = 0;
|
||||
|
||||
while (count--) {
|
||||
struct ext4_dir_entry_2 *de = (struct ext4_dir_entry_2 *)
|
||||
struct ext4_dir_entry_2 *de = (struct ext4_dir_entry_2 *)
|
||||
(from + (map->offs<<2));
|
||||
rec_len = EXT4_DIR_REC_LEN(de->name_len);
|
||||
memcpy (to, de, rec_len);
|
||||
|
|
|
@ -2216,7 +2216,7 @@ static unsigned long ext4_get_stripe_size(struct ext4_sb_info *sbi)
|
|||
struct ext4_attr {
|
||||
struct attribute attr;
|
||||
ssize_t (*show)(struct ext4_attr *, struct ext4_sb_info *, char *);
|
||||
ssize_t (*store)(struct ext4_attr *, struct ext4_sb_info *,
|
||||
ssize_t (*store)(struct ext4_attr *, struct ext4_sb_info *,
|
||||
const char *, size_t);
|
||||
int offset;
|
||||
};
|
||||
|
@ -3383,7 +3383,7 @@ static int ext4_commit_super(struct super_block *sb, int sync)
|
|||
if (!(sb->s_flags & MS_RDONLY))
|
||||
es->s_wtime = cpu_to_le32(get_seconds());
|
||||
es->s_kbytes_written =
|
||||
cpu_to_le64(EXT4_SB(sb)->s_kbytes_written +
|
||||
cpu_to_le64(EXT4_SB(sb)->s_kbytes_written +
|
||||
((part_stat_read(sb->s_bdev->bd_part, sectors[1]) -
|
||||
EXT4_SB(sb)->s_sectors_written_start) >> 1));
|
||||
ext4_free_blocks_count_set(es, percpu_counter_sum_positive(
|
||||
|
|
Загрузка…
Ссылка в новой задаче