fixed all whitespaces in .c files in the fs folder (not in the sub-folders) Signed-off-by: amar-laksh <amarlakshya1@xxxxxxxxx> --- fs/aio.c | 8 ++++---- fs/bad_inode.c | 16 ++++++++-------- fs/binfmt_aout.c | 6 +++--- fs/binfmt_elf.c | 26 +++++++++++++------------- fs/binfmt_em86.c | 4 ++-- fs/binfmt_flat.c | 26 +++++++++++++------------- fs/binfmt_script.c | 8 ++++---- fs/block_dev.c | 8 ++++---- fs/buffer.c | 22 +++++++++++----------- fs/compat.c | 10 +++++----- fs/compat_ioctl.c | 2 +- fs/dcache.c | 32 ++++++++++++++++---------------- fs/dcookies.c | 4 ++-- fs/direct-io.c | 10 +++++----- fs/exec.c | 2 +- fs/fcntl.c | 18 +++++++++--------- fs/file_table.c | 4 ++-- fs/filesystems.c | 10 +++++----- fs/locks.c | 14 +++++++------- fs/mbcache.c | 2 +- fs/mpage.c | 12 ++++++------ fs/namei.c | 4 ++-- fs/namespace.c | 4 ++-- fs/open.c | 4 ++-- fs/pipe.c | 10 +++++----- fs/pnode.c | 2 +- fs/posix_acl.c | 4 ++-- fs/read_write.c | 4 ++-- fs/readdir.c | 2 +- fs/select.c | 4 ++-- fs/signalfd.c | 2 +- fs/stat.c | 4 ++-- fs/super.c | 6 +++--- fs/timerfd.c | 2 +- 34 files changed, 148 insertions(+), 148 deletions(-) diff --git a/fs/aio.c b/fs/aio.c index 155f842..0062687 100644 --- a/fs/aio.c +++ b/fs/aio.c @@ -1304,10 +1304,10 @@ static long read_events(struct kioctx *ctx, long min_nr, long nr, * Create an aio_context capable of receiving at least nr_events. * ctxp must not point to an aio_context that already exists, and * must be initialized to 0 prior to the call. On successful - * creation of the aio_context, *ctxp is filled in with the resulting + * creation of the aio_context, *ctxp is filled in with the resulting * handle. May fail with -EINVAL if *ctxp is not initialized, - * if the specified nr_events exceeds internal limits. May fail - * with -EAGAIN if the specified nr_events exceeds the user's limit + * if the specified nr_events exceeds internal limits. May fail + * with -EAGAIN if the specified nr_events exceeds the user's limit * of available events. May fail with -ENOMEM if insufficient kernel * resources are available. May fail with -EFAULT if an invalid * pointer is passed for ctxp. Will fail with -ENOSYS if not @@ -1344,7 +1344,7 @@ out: } /* sys_io_destroy: - * Destroy the aio_context specified. May cancel any outstanding + * Destroy the aio_context specified. May cancel any outstanding * AIOs and block on completion. Will fail with -ENOSYS if not * implemented. May fail with -EINVAL if the context pointed to * is invalid. diff --git a/fs/bad_inode.c b/fs/bad_inode.c index 103f5d7..a7959f0 100644 --- a/fs/bad_inode.c +++ b/fs/bad_inode.c @@ -152,12 +152,12 @@ static const struct inode_operations bad_inode_ops = /* * When a filesystem is unable to read an inode due to an I/O error in * its read_inode() function, it can call make_bad_inode() to return a - * set of stubs which will return EIO errors as required. + * set of stubs which will return EIO errors as required. * * We only need to do limited initialisation: all other fields are * preinitialised to zero automatically. */ - + /** * make_bad_inode - mark an inode bad due to an I/O error * @inode: Inode to mark bad @@ -166,7 +166,7 @@ static const struct inode_operations bad_inode_ops = * failure this function makes the inode "bad" and causes I/O operations * on it to fail from this point on. */ - + void make_bad_inode(struct inode *inode) { remove_inode_hash(inode); @@ -174,8 +174,8 @@ void make_bad_inode(struct inode *inode) inode->i_mode = S_IFREG; inode->i_atime = inode->i_mtime = inode->i_ctime = current_fs_time(inode->i_sb); - inode->i_op = &bad_inode_ops; - inode->i_fop = &bad_file_ops; + inode->i_op = &bad_inode_ops; + inode->i_fop = &bad_file_ops; } EXPORT_SYMBOL(make_bad_inode); @@ -184,17 +184,17 @@ EXPORT_SYMBOL(make_bad_inode); * &bad_inode_ops to cover the case of invalidated inodes as well as * those created by make_bad_inode() above. */ - + /** * is_bad_inode - is an inode errored * @inode: inode to test * * Returns true if the inode in question has been marked as bad. */ - + bool is_bad_inode(struct inode *inode) { - return (inode->i_op == &bad_inode_ops); + return (inode->i_op == &bad_inode_ops); } EXPORT_SYMBOL(is_bad_inode); diff --git a/fs/binfmt_aout.c b/fs/binfmt_aout.c index 4c55668..1dcba71 100644 --- a/fs/binfmt_aout.c +++ b/fs/binfmt_aout.c @@ -291,7 +291,7 @@ static int load_aout_binary(struct linux_binprm * bprm) if ((fd_offset & ~PAGE_MASK) != 0 && printk_ratelimit()) { - printk(KERN_WARNING + printk(KERN_WARNING "fd_offset is not page aligned. Please convert program: %pD\n", bprm->file); } @@ -374,12 +374,12 @@ static int load_aout_library(struct file *file) if ((N_TXTOFF(ex) & ~PAGE_MASK) != 0) { if (printk_ratelimit()) { - printk(KERN_WARNING + printk(KERN_WARNING "N_TXTOFF is not page aligned. Please convert library: %pD\n", file); } vm_brk(start_addr, ex.a_text + ex.a_data + ex.a_bss); - + read_code(file, start_addr, N_TXTOFF(ex), ex.a_text + ex.a_data); retval = 0; diff --git a/fs/binfmt_elf.c b/fs/binfmt_elf.c index 051ea48..0027233 100644 --- a/fs/binfmt_elf.c +++ b/fs/binfmt_elf.c @@ -224,7 +224,7 @@ create_elf_tables(struct linux_binprm *bprm, struct elfhdr *exec, } while (0) #ifdef ARCH_DLINFO - /* + /* * ARCH_DLINFO must come first so PPC can do its special alignment of * AUXV. * update AT_VECTOR_SIZE_ARCH if the number of NEW_AUX_ENT() in @@ -691,7 +691,7 @@ static int load_elf_binary(struct linux_binprm *bprm) retval = -ENOMEM; goto out_ret; } - + /* Get the exec-header */ loc->elf_ex = *((struct elfhdr *)bprm->buf); @@ -727,7 +727,7 @@ static int load_elf_binary(struct linux_binprm *bprm) * is an a.out format binary */ retval = -ENOEXEC; - if (elf_ppnt->p_filesz > PATH_MAX || + if (elf_ppnt->p_filesz > PATH_MAX || elf_ppnt->p_filesz < 2) goto out_free_ph; @@ -859,7 +859,7 @@ static int load_elf_binary(struct linux_binprm *bprm) executable_stack); if (retval < 0) goto out_free_dentry; - + current->mm->start_stack = bprm->p; /* Now we do a little grungy work by mmapping the ELF image into @@ -875,7 +875,7 @@ static int load_elf_binary(struct linux_binprm *bprm) if (unlikely (elf_brk > elf_bss)) { unsigned long nbyte; - + /* There was a PT_LOAD segment with p_memsz > p_filesz before this one. Map anonymous pages, if needed, and clear the area. */ @@ -1390,7 +1390,7 @@ static void fill_elf_note_phdr(struct elf_phdr *phdr, int sz, loff_t offset) return; } -static void fill_note(struct memelfnote *note, const char *name, int type, +static void fill_note(struct memelfnote *note, const char *name, int type, unsigned int sz, void *data) { note->name = name; @@ -1442,7 +1442,7 @@ static int fill_psinfo(struct elf_prpsinfo *psinfo, struct task_struct *p, { const struct cred *cred; unsigned int i, len; - + /* first copy the parameters from user space */ memset(psinfo, 0, sizeof(struct elf_prpsinfo)); @@ -1476,7 +1476,7 @@ static int fill_psinfo(struct elf_prpsinfo *psinfo, struct task_struct *p, SET_GID(psinfo->pr_gid, from_kgid_munged(cred->user_ns, cred->gid)); rcu_read_unlock(); strncpy(psinfo->pr_fname, p->comm, sizeof(psinfo->pr_fname)); - + return 0; } @@ -1878,8 +1878,8 @@ static int elf_dump_thread_status(long signr, struct elf_thread_status *t) t->num_notes = 0; fill_prstatus(&t->prstatus, p, signr); - elf_core_copy_task_regs(p, &t->prstatus.pr_reg); - + elf_core_copy_task_regs(p, &t->prstatus.pr_reg); + fill_note(&t->notes[0], "CORE", NT_PRSTATUS, sizeof(t->prstatus), &(t->prstatus)); t->num_notes++; @@ -1900,7 +1900,7 @@ static int elf_dump_thread_status(long signr, struct elf_thread_status *t) t->num_notes++; sz += notesize(&t->notes[2]); } -#endif +#endif return sz; } @@ -2143,7 +2143,7 @@ static int elf_core_dump(struct coredump_params *cprm) /* * We no longer stop all VM operations. - * + * * This is because those proceses that could possibly change map_count * or the mmap / vma pages are now blocked in do_exit on current * finishing this core dump. @@ -2152,7 +2152,7 @@ static int elf_core_dump(struct coredump_params *cprm) * the map_count or the pages allocated. So no possibility of crashing * exists while dumping the mm->vm_next areas to the core file. */ - + /* alloc memory for large data structures: too large to be on stack */ elf = kmalloc(sizeof(*elf), GFP_KERNEL); if (!elf) diff --git a/fs/binfmt_em86.c b/fs/binfmt_em86.c index 4905385..f4deca8 100644 --- a/fs/binfmt_em86.c +++ b/fs/binfmt_em86.c @@ -67,11 +67,11 @@ static int load_em86(struct linux_binprm *bprm) */ remove_arg_zero(bprm); retval = copy_strings_kernel(1, &bprm->filename, bprm); - if (retval < 0) return retval; + if (retval < 0) return retval; bprm->argc++; if (i_arg) { retval = copy_strings_kernel(1, &i_arg, bprm); - if (retval < 0) return retval; + if (retval < 0) return retval; bprm->argc++; } retval = copy_strings_kernel(1, &i_name, bprm); diff --git a/fs/binfmt_flat.c b/fs/binfmt_flat.c index f723cd3..eab7b5e 100644 --- a/fs/binfmt_flat.c +++ b/fs/binfmt_flat.c @@ -387,7 +387,7 @@ static void old_reloc(unsigned long rl) #endif flat_v2_reloc_t r; unsigned long *ptr; - + r.value = rl; #if defined(CONFIG_COLDFIRE) ptr = (unsigned long *) (current->mm->start_code + r.reloc.offset); @@ -400,7 +400,7 @@ static void old_reloc(unsigned long rl) "(address %p, currently %x) into segment %s\n", r.reloc.offset, ptr, (int)*ptr, segment[r.reloc.type]); #endif - + switch (r.reloc.type) { case OLD_FLAT_RELOC_TYPE_TEXT: *ptr += current->mm->start_code; @@ -419,7 +419,7 @@ static void old_reloc(unsigned long rl) #ifdef DEBUG printk("Relocation became %x\n", (int)*ptr); #endif -} +} /****************************************************************************/ @@ -478,7 +478,7 @@ static int load_flat_file(struct linux_binprm * bprm, ret = -ENOEXEC; goto err; } - + /* Don't allow old format executables to use shared libraries */ if (rev == OLD_FLAT_VERSION && id != 0) { printk("BINFMT_FLAT: shared libraries are not available before rev 0x%x\n", @@ -580,7 +580,7 @@ static int load_flat_file(struct linux_binprm * bprm, fpos = ntohl(hdr->data_start); #ifdef CONFIG_BINFMT_ZFLAT if (flags & FLAT_FLAG_GZDATA) { - result = decompress_exec(bprm, fpos, (char *) datapos, + result = decompress_exec(bprm, fpos, (char *) datapos, full_data, 0); } else #endif @@ -703,7 +703,7 @@ static int load_flat_file(struct linux_binprm * bprm, libinfo->lib_list[id].loaded = 1; libinfo->lib_list[id].entry = (0x00ffffff & ntohl(hdr->entry)) + textpos; libinfo->lib_list[id].build_date = ntohl(hdr->build_date); - + /* * We just load the allocations into some temporary memory to * help simplify all this mumbo jumbo @@ -783,11 +783,11 @@ static int load_flat_file(struct linux_binprm * bprm, for (i=0; i < relocs; i++) old_reloc(ntohl(reloc[i])); } - + flush_icache_range(start_code, end_code); /* zero the BSS, BRK and stack areas */ - memset((void*)(datapos + data_len), 0, bss_len + + memset((void*)(datapos + data_len), 0, bss_len + (memp + memp_size - stack_len - /* end brk */ libinfo->lib_list[id].start_brk) + /* start brk */ stack_len); @@ -881,11 +881,11 @@ static int load_flat_binary(struct linux_binprm * bprm) stack_len += (bprm->argc + 1) * sizeof(char *); /* the argv array */ stack_len += (bprm->envc + 1) * sizeof(char *); /* the envp array */ stack_len += FLAT_STACK_ALIGN - 1; /* reserve for upcoming alignment */ - + res = load_flat_file(bprm, &libinfo, 0, &stack_len); if (IS_ERR_VALUE(res)) return res; - + /* Update data segment pointers for all libraries */ for (i=0; i<MAX_SHARED_LIBS; i++) if (libinfo.lib_list[i].loaded) @@ -907,7 +907,7 @@ static int load_flat_binary(struct linux_binprm * bprm) ((char *) page_address(bprm->page[i/PAGE_SIZE]))[i % PAGE_SIZE]; sp = (unsigned long *) create_flat_tables(p, bprm); - + /* Fake some return addresses to ensure the call chain will * initialise library in order for us. We are required to call * lib 1 first, then 2, ... and finally the main program (id 0). @@ -923,7 +923,7 @@ static int load_flat_binary(struct linux_binprm * bprm) } } #endif - + /* Stash our initial stack pointer into the mm structure */ current->mm->start_stack = (unsigned long )sp; @@ -932,7 +932,7 @@ static int load_flat_binary(struct linux_binprm * bprm) #endif DBG_FLT("start_thread(regs=0x%x, entry=0x%x, start_stack=0x%x)\n", (int)regs, (int)start_addr, (int)current->mm->start_stack); - + start_thread(regs, start_addr, current->mm->start_stack); return 0; diff --git a/fs/binfmt_script.c b/fs/binfmt_script.c index afdf4e3..4b551a5 100644 --- a/fs/binfmt_script.c +++ b/fs/binfmt_script.c @@ -55,7 +55,7 @@ static int load_script(struct linux_binprm *bprm) break; } for (cp = bprm->buf+2; (*cp == ' ') || (*cp == '\t'); cp++); - if (*cp == '\0') + if (*cp == '\0') return -ENOEXEC; /* No interpreter name found */ i_name = cp; i_arg = NULL; @@ -80,15 +80,15 @@ static int load_script(struct linux_binprm *bprm) if (retval) return retval; retval = copy_strings_kernel(1, &bprm->interp, bprm); - if (retval < 0) return retval; + if (retval < 0) return retval; bprm->argc++; if (i_arg) { retval = copy_strings_kernel(1, &i_arg, bprm); - if (retval < 0) return retval; + if (retval < 0) return retval; bprm->argc++; } retval = copy_strings_kernel(1, &i_name, bprm); - if (retval) return retval; + if (retval) return retval; bprm->argc++; retval = bprm_change_interp(interp, bprm); if (retval < 0) diff --git a/fs/block_dev.c b/fs/block_dev.c index 39b3a17..64f94df 100644 --- a/fs/block_dev.c +++ b/fs/block_dev.c @@ -80,7 +80,7 @@ void kill_bdev(struct block_device *bdev) invalidate_bh_lrus(); truncate_inode_pages(mapping, 0); -} +} EXPORT_SYMBOL(kill_bdev); /* Invalidate clean unused buffers and pagecache. */ @@ -351,13 +351,13 @@ static loff_t block_llseek(struct file *file, loff_t offset, int whence) inode_unlock(bd_inode); return retval; } - + int blkdev_fsync(struct file *filp, loff_t start, loff_t end, int datasync) { struct inode *bd_inode = bdev_file_inode(filp); struct block_device *bdev = I_BDEV(bd_inode); int error; - + error = filemap_write_and_wait_range(filp->f_mapping, start, end); if (error) return error; @@ -692,7 +692,7 @@ void bdput(struct block_device *bdev) } EXPORT_SYMBOL(bdput); - + static struct block_device *bd_acquire(struct inode *inode) { struct block_device *bdev; diff --git a/fs/buffer.c b/fs/buffer.c index e1632ab..a75ca74 100644 --- a/fs/buffer.c +++ b/fs/buffer.c @@ -439,7 +439,7 @@ EXPORT_SYMBOL(mark_buffer_async_write); * try_to_free_buffers() will be operating against the *blockdev* mapping * at the time, not against the S_ISREG file which depends on those buffers. * So the locking for private_list is via the private_lock in the address_space - * which backs the buffers. Which is different from the address_space + * which backs the buffers. Which is different from the address_space * against which the buffers are listed. So for a particular address_space, * mapping->private_lock does *not* protect mapping->private_list! In fact, * mapping->private_list will always be protected by the backing blockdev's @@ -713,7 +713,7 @@ EXPORT_SYMBOL(__set_page_dirty_buffers); * Do this in two main stages: first we copy dirty buffers to a * temporary inode list, queueing the writes as we go. Then we clean * up, waiting for those writes to complete. - * + * * During this second stage, any subsequent updates to the file may end * up refiling the buffer on the original inode's dirty list again, so * there is a chance we will end up with a buffer queued for write but @@ -791,7 +791,7 @@ static int fsync_buffers_list(spinlock_t *lock, struct list_head *list) brelse(bh); spin_lock(lock); } - + spin_unlock(lock); err2 = osync_buffers_list(lock, list); if (err) @@ -901,7 +901,7 @@ no_grow: /* * Return failure for non-async IO requests. Async IO requests * are not allowed to fail, so we have to wait until buffer heads - * become available. But we don't want tasks sleeping with + * become available. But we don't want tasks sleeping with * partially complete buffers, so all were released above. */ if (!retry) @@ -910,7 +910,7 @@ no_grow: /* We're _really_ low on memory. Now we just * wait for old buffer heads to become free due to * finishing IO. Since this is an async request and - * the reserve list is empty, we're sure there are + * the reserve list is empty, we're sure there are * async buffer heads in use. */ free_more_memory(); @@ -946,7 +946,7 @@ static sector_t blkdev_max_block(struct block_device *bdev, unsigned int size) /* * Initialise the state of a blockdev page's buffers. - */ + */ static sector_t init_page_buffers(struct page *page, struct block_device *bdev, sector_t block, int size) @@ -1448,7 +1448,7 @@ static bool has_bh_in_lru(int cpu, void *dummy) { struct bh_lru *b = per_cpu_ptr(&bh_lrus, cpu); int i; - + for (i = 0; i < BH_LRU_SIZE; i++) { if (b->bhs[i]) return 1; @@ -1952,7 +1952,7 @@ int __block_write_begin(struct page *page, loff_t pos, unsigned len, if (PageUptodate(page)) { if (!buffer_uptodate(bh)) set_buffer_uptodate(bh); - continue; + continue; } if (!buffer_uptodate(bh) && !buffer_delay(bh) && !buffer_unwritten(bh) && @@ -2258,7 +2258,7 @@ EXPORT_SYMBOL(block_read_full_page); /* utility function for filesystems that need to do work on expanding * truncates. Uses filesystem pagecache writes to allow the filesystem to - * deal with the hole. + * deal with the hole. */ int generic_cont_expand_simple(struct inode *inode, loff_t size) { @@ -2819,7 +2819,7 @@ int block_truncate_page(struct address_space *mapping, length = blocksize - length; iblock = (sector_t)index << (PAGE_CACHE_SHIFT - inode->i_blkbits); - + page = grab_cache_page(mapping, index); err = -ENOMEM; if (!page) @@ -3069,7 +3069,7 @@ EXPORT_SYMBOL(submit_bh); * * ll_rw_block sets b_end_io to simple completion handler that marks * the buffer up-to-date (if appropriate), unlocks the buffer and wakes - * any waiters. + * any waiters. * * All of the buffers must be for the same device, and must also be a * multiple of the current approved size for the device. diff --git a/fs/compat.c b/fs/compat.c index a71936a..6ce6bdd 100644 --- a/fs/compat.c +++ b/fs/compat.c @@ -7,7 +7,7 @@ * Copyright (C) 2002 Stephen Rothwell, IBM Corporation * Copyright (C) 1997-2000 Jakub Jelinek (jakub@xxxxxxxxxx) * Copyright (C) 1998 Eddie C. Dost (ecd@xxxxxxxxx) - * Copyright (C) 2001,2002 Andi Kleen, SuSE Labs + * Copyright (C) 2001,2002 Andi Kleen, SuSE Labs * Copyright (C) 2003 Pavel Machek (pavel@xxxxxx) * * This program is free software; you can redistribute it and/or modify @@ -211,7 +211,7 @@ COMPAT_SYSCALL_DEFINE2(newfstat, unsigned int, fd, static int put_compat_statfs(struct compat_statfs __user *ubuf, struct kstatfs *kbuf) { - + if (sizeof ubuf->f_blocks == 4) { if ((kbuf->f_blocks | kbuf->f_bfree | kbuf->f_bavail | kbuf->f_bsize | kbuf->f_frsize) & 0xffffffff00000000ULL) @@ -536,7 +536,7 @@ COMPAT_SYSCALL_DEFINE5(io_getevents, compat_aio_context_t, ctx_id, ut = compat_alloc_user_space(sizeof(*ut)); if (copy_to_user(ut, &t, sizeof(t)) ) return -EFAULT; - } + } return sys_io_getevents(ctx_id, min_nr, nr, events, ut); } @@ -636,7 +636,7 @@ copy_iocb(long nr, u32 __user *ptr32, struct iocb __user * __user *ptr64) COMPAT_SYSCALL_DEFINE3(io_submit, compat_aio_context_t, ctx_id, int, nr, u32 __user *, iocb) { - struct iocb __user * __user *iocb64; + struct iocb __user * __user *iocb64; long ret; if (unlikely(nr < 0)) @@ -644,7 +644,7 @@ COMPAT_SYSCALL_DEFINE3(io_submit, compat_aio_context_t, ctx_id, if (nr > MAX_AIO_SUBMITS) nr = MAX_AIO_SUBMITS; - + iocb64 = compat_alloc_user_space(nr * sizeof(*iocb64)); ret = copy_iocb(nr, iocb, iocb64); if (!ret) diff --git a/fs/compat_ioctl.c b/fs/compat_ioctl.c index 6402eaf..c6519f5 100644 --- a/fs/compat_ioctl.c +++ b/fs/compat_ioctl.c @@ -3,7 +3,7 @@ * * Copyright (C) 1997-2000 Jakub Jelinek (jakub@xxxxxxxxxx) * Copyright (C) 1998 Eddie C. Dost (ecd@xxxxxxxxx) - * Copyright (C) 2001,2002 Andi Kleen, SuSE Labs + * Copyright (C) 2001,2002 Andi Kleen, SuSE Labs * Copyright (C) 2003 Pavel Machek (pavel@xxxxxx) * * These routines maintain argument size conversion between 32bit and 64bit diff --git a/fs/dcache.c b/fs/dcache.c index 92d5140..ec44f22 100644 --- a/fs/dcache.c +++ b/fs/dcache.c @@ -254,14 +254,14 @@ static void __d_free(struct rcu_head *head) { struct dentry *dentry = container_of(head, struct dentry, d_u.d_rcu); - kmem_cache_free(dentry_cache, dentry); + kmem_cache_free(dentry_cache, dentry); } static void __d_free_external(struct rcu_head *head) { struct dentry *dentry = container_of(head, struct dentry, d_u.d_rcu); kfree(external_name(dentry)); - kmem_cache_free(dentry_cache, dentry); + kmem_cache_free(dentry_cache, dentry); } static inline int dname_external(const struct dentry *dentry) @@ -731,7 +731,7 @@ static inline bool fast_dput(struct dentry *dentry) } -/* +/* * This is dput * * This is complicated by the fact that we do not want to put @@ -750,7 +750,7 @@ static inline bool fast_dput(struct dentry *dentry) /* * dput - release a dentry - * @dentry: dentry to release + * @dentry: dentry to release * * Release a dentry. This will drop the usage count and if appropriate * call the dentry unlink method as well as removing it from the queues and @@ -1552,7 +1552,7 @@ EXPORT_SYMBOL(d_invalidate); * available. On a success the dentry is returned. The name passed in is * copied and the copy passed in may be reused after this call. */ - + struct dentry *__d_alloc(struct super_block *sb, const struct qstr *name) { struct dentry *dentry; @@ -1574,7 +1574,7 @@ struct dentry *__d_alloc(struct super_block *sb, const struct qstr *name) struct external_name *p = kmalloc(size + name->len, GFP_KERNEL_ACCOUNT); if (!p) { - kmem_cache_free(dentry_cache, dentry); + kmem_cache_free(dentry_cache, dentry); return NULL; } atomic_set(&p->u.count, 1); @@ -1584,7 +1584,7 @@ struct dentry *__d_alloc(struct super_block *sb, const struct qstr *name) round_up(name->len + 1, sizeof(unsigned long))); } else { dname = dentry->d_iname; - } + } dentry->d_name.len = name->len; dentry->d_name.hash = name->hash; @@ -1778,7 +1778,7 @@ static void __d_instantiate(struct dentry *dentry, struct inode *inode) * (or otherwise set) by the caller to indicate that it is now * in use by the dcache. */ - + void d_instantiate(struct dentry *entry, struct inode * inode) { BUG_ON(!hlist_unhashed(&entry->d_u.d_alias)); @@ -2300,7 +2300,7 @@ struct dentry *__d_lookup(const struct dentry *parent, const struct qstr *name) * See Documentation/filesystems/path-lookup.txt for more details. */ rcu_read_lock(); - + hlist_bl_for_each_entry_rcu(dentry, node, b, d_hash) { if (dentry->d_name.hash != hash) @@ -2376,7 +2376,7 @@ EXPORT_SYMBOL(d_hash_and_lookup); * it from the hash queues and waiting for * it to be deleted later when it has no users */ - + /** * d_delete - delete a dentry * @dentry: The dentry to delete @@ -2384,7 +2384,7 @@ EXPORT_SYMBOL(d_hash_and_lookup); * Turn the dentry into a negative dentry if possible, otherwise * remove it from the hash queues so it can be deleted later */ - + void d_delete(struct dentry * dentry) { struct inode *inode; @@ -2437,7 +2437,7 @@ static void _d_rehash(struct dentry * entry) * * Adds a dentry to the hash according to its name. */ - + void d_rehash(struct dentry * entry) { spin_lock(&entry->d_lock); @@ -3136,7 +3136,7 @@ char *simple_dname(struct dentry *dentry, char *buffer, int buflen) /* these dentries are never renamed, so d_lock is not needed */ if (prepend(&end, &buflen, " (deleted)", 11) || prepend(&end, &buflen, dentry->d_name.name, dentry->d_name.len) || - prepend(&end, &buflen, "/", 1)) + prepend(&end, &buflen, "/", 1)) end = ERR_PTR(-ENAMETOOLONG); return end; } @@ -3308,7 +3308,7 @@ out: * Returns false otherwise. * Caller must ensure that "new_dentry" is pinned before calling is_subdir() */ - + bool is_subdir(struct dentry *new_dentry, struct dentry *old_dentry) { bool result; @@ -3410,10 +3410,10 @@ static void __init dcache_init(void) { unsigned int loop; - /* + /* * A constructor could be added for stable state like the lists, * but it is probably not worth it because of the cache nature - * of the dcache. + * of the dcache. */ dentry_cache = KMEM_CACHE(dentry, SLAB_RECLAIM_ACCOUNT|SLAB_PANIC|SLAB_MEM_SPREAD|SLAB_ACCOUNT); diff --git a/fs/dcookies.c b/fs/dcookies.c index ac44a69..ca7f7a6 100644 --- a/fs/dcookies.c +++ b/fs/dcookies.c @@ -187,7 +187,7 @@ SYSCALL_DEFINE3(lookup_dcookie, u64, cookie64, char __user *, buf, size_t, len) } err = -ERANGE; - + pathlen = kbuf + PAGE_SIZE - path; if (pathlen <= len) { err = pathlen; @@ -306,7 +306,7 @@ static void dcookie_exit(void) struct dcookie_user { struct list_head next; }; - + struct dcookie_user * dcookie_register(void) { struct dcookie_user * user; diff --git a/fs/direct-io.c b/fs/direct-io.c index d6a9012..0432f4f 100644 --- a/fs/direct-io.c +++ b/fs/direct-io.c @@ -189,7 +189,7 @@ static inline int dio_refill_pages(struct dio *dio, struct dio_submit *sdio) sdio->to = ((ret - 1) & (PAGE_SIZE - 1)) + 1; return 0; } - return ret; + return ret; } /* @@ -286,7 +286,7 @@ static void dio_aio_complete_work(struct work_struct *work) static int dio_bio_complete(struct dio *dio, struct bio *bio); /* - * Asynchronous IO callback. + * Asynchronous IO callback. */ static void dio_bio_end_aio(struct bio *bio) { @@ -699,7 +699,7 @@ static inline int dio_bio_add_page(struct dio_submit *sdio) } return ret; } - + /* * Put cur_page under IO. The section of cur_page which is described by * cur_page_offset,cur_page_len is put into a BIO. The section of cur_page @@ -761,7 +761,7 @@ out: * An autonomous function to put a chunk of a page under deferred IO. * * The caller doesn't actually know (or care) whether this piece of page is in - * a BIO, or is under IO or whatever. We just take care of all possible + * a BIO, or is under IO or whatever. We just take care of all possible * situations here. The separation between the logic of do_direct_IO() and * that of submit_page_section() is important for clarity. Please don't break. * @@ -879,7 +879,7 @@ static inline void dio_zero_block(struct dio *dio, struct dio_submit *sdio, * We need to zero out part of an fs block. It is either at the * beginning or the end of the fs block. */ - if (end) + if (end) this_chunk_blocks = dio_blocks_per_fs_block - this_chunk_blocks; this_chunk_bytes = this_chunk_blocks << sdio->blkbits; diff --git a/fs/exec.c b/fs/exec.c index dcd4ac7..4cf2252 100644 --- a/fs/exec.c +++ b/fs/exec.c @@ -19,7 +19,7 @@ * current->executable is only used by the procfs. This allows a dispatch * table to check for several different types of binary formats. We keep * trying until we recognize the file or we run out of supported binary - * formats. + * formats. */ #include <linux/slab.h> diff --git a/fs/fcntl.c b/fs/fcntl.c index 350a2c8..ae351e5 100644 --- a/fs/fcntl.c +++ b/fs/fcntl.c @@ -355,7 +355,7 @@ static int check_fcntl_cmd(unsigned cmd) } SYSCALL_DEFINE3(fcntl, unsigned int, fd, unsigned int, cmd, unsigned long, arg) -{ +{ struct fd f = fdget_raw(fd); long err = -EBADF; @@ -380,7 +380,7 @@ out: #if BITS_PER_LONG == 32 SYSCALL_DEFINE3(fcntl64, unsigned int, fd, unsigned int, cmd, unsigned long, arg) -{ +{ struct fd f = fdget_raw(fd); long err = -EBADF; @@ -395,7 +395,7 @@ SYSCALL_DEFINE3(fcntl64, unsigned int, fd, unsigned int, cmd, err = security_file_fcntl(f.file, cmd, arg); if (err) goto out1; - + switch (cmd) { case F_GETLK64: case F_OFD_GETLK: @@ -463,8 +463,8 @@ static void send_sigio_to_task(struct task_struct *p, siginfo_t si; default: /* Queue a rt signal with the appropriate fd as its - value. We use SI_SIGIO as the source, not - SI_KERNEL, since kernel signals always get + value. We use SI_SIGIO as the source, not + SI_KERNEL, since kernel signals always get delivered even if we can't queue. Failure to queue in this case _should_ be reported; we fall back to SIGIO in that case. --sct */ @@ -494,7 +494,7 @@ void send_sigio(struct fown_struct *fown, int fd, int band) enum pid_type type; struct pid *pid; int group = 1; - + read_lock(&fown->lock); type = fown->pid_type; @@ -506,7 +506,7 @@ void send_sigio(struct fown_struct *fown, int fd, int band) pid = fown->pid; if (!pid) goto out_unlock_fown; - + read_lock(&tasklist_lock); do_each_pid_task(pid, type, p) { send_sigio_to_task(p, fown, fd, band, group); @@ -530,7 +530,7 @@ int send_sigurg(struct fown_struct *fown) struct pid *pid; int group = 1; int ret = 0; - + read_lock(&fown->lock); type = fown->pid_type; @@ -544,7 +544,7 @@ int send_sigurg(struct fown_struct *fown) goto out_unlock_fown; ret = 1; - + read_lock(&tasklist_lock); do_each_pid_task(pid, type, p) { send_sigurg_to_task(p, fown, group); diff --git a/fs/file_table.c b/fs/file_table.c index ad17e05..77c98ea 100644 --- a/fs/file_table.c +++ b/fs/file_table.c @@ -310,7 +310,7 @@ void put_filp(struct file *file) } void __init files_init(void) -{ +{ filp_cachep = kmem_cache_create("filp", sizeof(struct file), 0, SLAB_HWCACHE_ALIGN | SLAB_PANIC, NULL); percpu_counter_init(&nr_files, 0, GFP_KERNEL); @@ -329,4 +329,4 @@ void __init files_maxfiles_init(void) n = ((totalram_pages - memreserve) * (PAGE_SIZE / 1024)) / 10; files_stat.max_files = max_t(unsigned long, n, NR_FILE); -} +} diff --git a/fs/filesystems.c b/fs/filesystems.c index c5618db..5e0a168 100644 --- a/fs/filesystems.c +++ b/fs/filesystems.c @@ -61,11 +61,11 @@ static struct file_system_type **find_filesystem(const char *name, unsigned len) * is aware of for mount and other syscalls. Returns 0 on success, * or a negative errno code on an error. * - * The &struct file_system_type that is passed is linked into the kernel + * The &struct file_system_type that is passed is linked into the kernel * structures and must not be freed until the file system has been * unregistered. */ - + int register_filesystem(struct file_system_type * fs) { int res = 0; @@ -93,11 +93,11 @@ EXPORT_SYMBOL(register_filesystem); * Remove a file system that was previously successfully registered * with the kernel. An error is returned if the file system is not found. * Zero is returned on a success. - * + * * Once this function has returned the &struct file_system_type structure * may be freed or reused. */ - + int unregister_filesystem(struct file_system_type * fs) { struct file_system_type ** tmp; @@ -179,7 +179,7 @@ static int fs_maxindex(void) } /* - * Whee.. Weird sysv syscall. + * Whee.. Weird sysv syscall. */ SYSCALL_DEFINE3(sysfs, int, option, unsigned long, arg1, unsigned long, arg2) { diff --git a/fs/locks.c b/fs/locks.c index 7c5f91b..5daec54 100644 --- a/fs/locks.c +++ b/fs/locks.c @@ -11,11 +11,11 @@ * * Miscellaneous edits, and a total rewrite of posix_lock_file() code. * Kai Petzke (wpp@xxxxxxxxxxxxxxxxxxxxxxxxx), 1994 - * + * * Converted file_lock_table to a linked list from an array, which eliminates * the limits on how many active file locks are open. * Chad Page (pageone@xxxxxxxxxx), November 27, 1994 - * + * * Removed dependency on file descriptors. dup()'ed file descriptors now * get the same locks as the original file descriptors, and a close() on * any file descriptor removes ALL the locks on the file for the current @@ -41,7 +41,7 @@ * with a file pointer (filp). As a result they can be shared by a parent * process and its children after a fork(). They are removed when the last * file descriptor referring to the file pointer is closed (unless explicitly - * unlocked). + * unlocked). * * FL_FLOCK locks never deadlock, an existing lock is always removed before * upgrading from shared to exclusive (or vice versa). When this happens @@ -50,7 +50,7 @@ * Andy Walker (andy@xxxxxxxxxxxxxxxxxxx), June 09, 1995 * * Removed some race conditions in flock_lock_file(), marked other possible - * races. Just grep for FIXME to see them. + * races. Just grep for FIXME to see them. * Dmitry Gorodchanin (pgmdsg@xxxxxxx), February 09, 1996. * * Addressed Dmitry's concerns. Deadlock checking no longer recursive. @@ -405,7 +405,7 @@ flock_make_lock(struct file *filp, unsigned int cmd) if (type < 0) return ERR_PTR(type); - + fl = locks_alloc_lock(); if (fl == NULL) return ERR_PTR(-ENOMEM); @@ -416,7 +416,7 @@ flock_make_lock(struct file *filp, unsigned int cmd) fl->fl_flags = FL_FLOCK; fl->fl_type = type; fl->fl_end = OFFSET_MAX; - + return fl; } @@ -2078,7 +2078,7 @@ int fcntl_getlk(struct file *filp, unsigned int cmd, struct flock __user *l) error = vfs_test_lock(filp, &file_lock); if (error) goto out; - + flock.l_type = file_lock.fl_type; if (file_lock.fl_type != F_UNLCK) { error = posix_lock_to_flock(&flock, &file_lock); diff --git a/fs/mbcache.c b/fs/mbcache.c index 187477d..1c519d0 100644 --- a/fs/mbcache.c +++ b/fs/mbcache.c @@ -587,7 +587,7 @@ mb_cache_entry_insert(struct mb_cache_entry *ce, struct block_device *bdev, struct mb_cache_entry *lce; mb_assert(ce); - bucket = hash_long((unsigned long)bdev + (block & 0xffffffff), + bucket = hash_long((unsigned long)bdev + (block & 0xffffffff), cache->c_bucket_bits); block_hash_p = &cache->c_block_hash[bucket]; hlist_bl_lock(block_hash_p); diff --git a/fs/mpage.c b/fs/mpage.c index 1480d3a..0a99fda 100644 --- a/fs/mpage.c +++ b/fs/mpage.c @@ -94,8 +94,8 @@ mpage_alloc(struct block_device *bdev, * them. So when the buffer is up to date and the page size == block size, * this marks the page up to date instead of adding new buffers. */ -static void -map_buffer_to_page(struct page *page, struct buffer_head *bh, int page_block) +static void +map_buffer_to_page(struct page *page, struct buffer_head *bh, int page_block) { struct inode *inode = page->mapping->host; struct buffer_head *page_bh, *head; @@ -106,9 +106,9 @@ map_buffer_to_page(struct page *page, struct buffer_head *bh, int page_block) * don't make any buffers if there is only one buffer on * the page and the page just needs to be set up to date */ - if (inode->i_blkbits == PAGE_CACHE_SHIFT && + if (inode->i_blkbits == PAGE_CACHE_SHIFT && buffer_uptodate(bh)) { - SetPageUptodate(page); + SetPageUptodate(page); return; } create_empty_buffers(page, 1 << inode->i_blkbits, 0); @@ -226,7 +226,7 @@ do_mpage_readpage(struct bio *bio, struct page *page, unsigned nr_pages, map_buffer_to_page(page, map_bh, page_block); goto confused; } - + if (first_hole != blocks_per_page) goto confused; /* hole -> non-hole */ @@ -420,7 +420,7 @@ EXPORT_SYMBOL(mpage_readpage); * * If all blocks are found to be contiguous then the page can go into the * BIO. Otherwise fall back to the mapping's writepage(). - * + * * FIXME: This code wants an estimate of how many pages are still to be * written, so it can intelligently allocate a suitably-sized BIO. For now, * just allocate full-size (16-page) BIOs. diff --git a/fs/namei.c b/fs/namei.c index f624d13..f9cbe3d 100644 --- a/fs/namei.c +++ b/fs/namei.c @@ -49,8 +49,8 @@ * The new code replaces the old recursive symlink resolution with * an iterative one (in case of non-nested symlink chains). It does * this with calls to <fs>_follow_link(). - * As a side effect, dir_namei(), _namei() and follow_link() are now - * replaced with a single function lookup_dentry() that can handle all + * As a side effect, dir_namei(), _namei() and follow_link() are now + * replaced with a single function lookup_dentry() that can handle all * the special cases of the former code. * * With the new dcache, the pathname is stored at each inode, at least as diff --git a/fs/namespace.c b/fs/namespace.c index 4fb1691..806ef24 100644 --- a/fs/namespace.c +++ b/fs/namespace.c @@ -1576,7 +1576,7 @@ out_unlock: namespace_unlock(); } -/* +/* * Is the caller allowed to modify his namespace? */ static inline bool may_mount(void) @@ -2090,7 +2090,7 @@ static int do_loopback(struct path *path, const char *old_name, err = -EINVAL; if (mnt_ns_loop(old_path.dentry)) - goto out; + goto out; mp = lock_mount(path); err = PTR_ERR(mp); diff --git a/fs/open.c b/fs/open.c index 55bdc75..586db20 100644 --- a/fs/open.c +++ b/fs/open.c @@ -875,7 +875,7 @@ struct file *dentry_open(const struct path *path, int flags, fput(f); f = ERR_PTR(error); } - } else { + } else { put_filp(f); f = ERR_PTR(error); } @@ -982,7 +982,7 @@ struct file *filp_open(const char *filename, int flags, umode_t mode) { struct filename *name = getname_kernel(filename); struct file *file = ERR_CAST(name); - + if (!IS_ERR(name)) { file = file_open_name(name, flags, mode); putname(name); diff --git a/fs/pipe.c b/fs/pipe.c index ab8dad3..b72fe79 100644 --- a/fs/pipe.c +++ b/fs/pipe.c @@ -45,10 +45,10 @@ unsigned long pipe_user_pages_hard; unsigned long pipe_user_pages_soft = PIPE_DEF_BUFFERS * INR_OPEN_CUR; /* - * We use a start+len construction, which provides full use of the + * We use a start+len construction, which provides full use of the * allocated memory. * -- Florian Coosmann (FGC) - * + * * Reads with count = 0 should always return 0. * -- Julian Bradfield 1999-06-07. * @@ -842,7 +842,7 @@ SYSCALL_DEFINE1(pipe, int __user *, fildes) static int wait_for_partner(struct pipe_inode_info *pipe, unsigned int *cnt) { - int cur = *cnt; + int cur = *cnt; while (cur == *cnt) { pipe_wait(pipe); @@ -917,7 +917,7 @@ static int fifo_open(struct inode *inode, struct file *filp) } } break; - + case FMODE_WRITE: /* * O_WRONLY @@ -937,7 +937,7 @@ static int fifo_open(struct inode *inode, struct file *filp) goto err_wr; } break; - + case FMODE_READ | FMODE_WRITE: /* * O_RDWR diff --git a/fs/pnode.c b/fs/pnode.c index c524fdd..ced8017 100644 --- a/fs/pnode.c +++ b/fs/pnode.c @@ -240,7 +240,7 @@ static int propagate_one(struct mount *m) if (IS_MNT_SHARED(m)) type |= CL_MAKE_SHARED; } - + /* Notice when we are propagating across user namespaces */ if (m->mnt_ns->user_ns != user_ns) type |= CL_UNPRIVILEGED; diff --git a/fs/posix_acl.c b/fs/posix_acl.c index 711dd51..67e5bc5 100644 --- a/fs/posix_acl.c +++ b/fs/posix_acl.c @@ -675,12 +675,12 @@ posix_acl_from_xattr(struct user_namespace *user_ns, return ERR_PTR(-EINVAL); if (count == 0) return NULL; - + acl = posix_acl_alloc(count, GFP_NOFS); if (!acl) return ERR_PTR(-ENOMEM); acl_e = acl->a_entries; - + for (end = entry + count; entry != end; acl_e++, entry++) { acl_e->e_tag = le16_to_cpu(entry->e_tag); acl_e->e_perm = le16_to_cpu(entry->e_perm); diff --git a/fs/read_write.c b/fs/read_write.c index 0c8782a..904b39d 100644 --- a/fs/read_write.c +++ b/fs/read_write.c @@ -4,7 +4,7 @@ * Copyright (C) 1991, 1992 Linus Torvalds */ -#include <linux/slab.h> +#include <linux/slab.h> #include <linux/stat.h> #include <linux/fcntl.h> #include <linux/file.h> @@ -663,7 +663,7 @@ SYSCALL_DEFINE4(pwrite64, unsigned int, fd, const char __user *, buf, f = fdget(fd); if (f.file) { ret = -ESPIPE; - if (f.file->f_mode & FMODE_PWRITE) + if (f.file->f_mode & FMODE_PWRITE) ret = vfs_write(f.file, buf, count, &pos); fdput(f); } diff --git a/fs/readdir.c b/fs/readdir.c index e69ef3b..15e7e14 100644 --- a/fs/readdir.c +++ b/fs/readdir.c @@ -132,7 +132,7 @@ SYSCALL_DEFINE3(old_readdir, unsigned int, fd, /* * New, all-improved, singing, dancing, iBCS2-compliant getdents() - * interface. + * interface. */ struct linux_dirent { unsigned long d_ino; diff --git a/fs/select.c b/fs/select.c index 79d0d49..c0dd808 100644 --- a/fs/select.c +++ b/fs/select.c @@ -10,7 +10,7 @@ * parameter to reflect time remaining. * * 24 January 2000 - * Changed sys_poll()/do_poll() to use PAGE_SIZE chunk-based allocation + * Changed sys_poll()/do_poll() to use PAGE_SIZE chunk-based allocation * of fds to overcome nfds < 16390 descriptors limit (Tigran Aivazian). */ @@ -570,7 +570,7 @@ int core_sys_select(int n, fd_set __user *inp, fd_set __user *outp, /* * We need 6 bitmaps (in/out/ex for both incoming and outgoing), * since we used fdset we need to allocate memory in units of - * long-words. + * long-words. */ size = FDS_BYTES(n); bits = stack_fds; diff --git a/fs/signalfd.c b/fs/signalfd.c index 270221f..3e01ca6 100644 --- a/fs/signalfd.c +++ b/fs/signalfd.c @@ -117,7 +117,7 @@ static int signalfd_copyinfo(struct signalfd_siginfo __user *uinfo, err |= __put_user(kinfo->si_trapno, &uinfo->ssi_trapno); #endif #ifdef BUS_MCEERR_AO - /* + /* * Other callers might not initialize the si_lsb field, * so check explicitly for the right codes here. */ diff --git a/fs/stat.c b/fs/stat.c index bc045c7..feb37c2 100644 --- a/fs/stat.c +++ b/fs/stat.c @@ -141,7 +141,7 @@ static int cp_old_stat(struct kstat *stat, struct __old_kernel_stat __user * sta { static int warncount = 5; struct __old_kernel_stat tmp; - + if (warncount > 0) { warncount--; printk(KERN_WARNING "VFS: Warning: %s using old stat() call. Recompile your binary.\n", @@ -166,7 +166,7 @@ static int cp_old_stat(struct kstat *stat, struct __old_kernel_stat __user * sta #if BITS_PER_LONG == 32 if (stat->size > MAX_NON_LFS) return -EOVERFLOW; -#endif +#endif tmp.st_size = stat->size; tmp.st_atime = stat->atime.tv_sec; tmp.st_mtime = stat->mtime.tv_sec; diff --git a/fs/super.c b/fs/super.c index 1182af8..d5d28b7 100644 --- a/fs/super.c +++ b/fs/super.c @@ -482,7 +482,7 @@ retry: return ERR_PTR(-ENOMEM); goto retry; } - + err = set(s, data); if (err) { spin_unlock(&sb_lock); @@ -583,7 +583,7 @@ EXPORT_SYMBOL(iterate_supers_type); /** * get_super - get the superblock of a device * @bdev: device to get the superblock for - * + * * Scans the superblock list and finds the superblock of the file system * mounted on the device given. %NULL is returned if no match is found. */ @@ -673,7 +673,7 @@ restart: spin_unlock(&sb_lock); return NULL; } - + struct super_block *user_get_super(dev_t dev) { struct super_block *sb; diff --git a/fs/timerfd.c b/fs/timerfd.c index 053818d..d0ccf0c 100644 --- a/fs/timerfd.c +++ b/fs/timerfd.c @@ -415,7 +415,7 @@ SYSCALL_DEFINE2(timerfd_create, int, clockid, int, flags) return ufd; } -static int do_timerfd_settime(int ufd, int flags, +static int do_timerfd_settime(int ufd, int flags, const struct itimerspec *new, struct itimerspec *old) { -- 1.9.1 -- To unsubscribe from this list: send the line "unsubscribe linux-fsdevel" in the body of a message to majordomo@xxxxxxxxxxxxxxx More majordomo info at http://vger.kernel.org/majordomo-info.html