fs: remove fastcall, it is always empty
[akpm@linux-foundation.org: coding-style fixes] Signed-off-by: Harvey Harrison <harvey.harrison@gmail.com> Signed-off-by: Andrew Morton <akpm@linux-foundation.org> Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
This commit is contained in:
parent
75acb9cd2e
commit
fc9b52cd8f
6 changed files with 26 additions and 27 deletions
17
fs/aio.c
17
fs/aio.c
|
@ -317,7 +317,7 @@ static void wait_for_all_aios(struct kioctx *ctx)
|
||||||
/* wait_on_sync_kiocb:
|
/* wait_on_sync_kiocb:
|
||||||
* Waits on the given sync kiocb to complete.
|
* Waits on the given sync kiocb to complete.
|
||||||
*/
|
*/
|
||||||
ssize_t fastcall wait_on_sync_kiocb(struct kiocb *iocb)
|
ssize_t wait_on_sync_kiocb(struct kiocb *iocb)
|
||||||
{
|
{
|
||||||
while (iocb->ki_users) {
|
while (iocb->ki_users) {
|
||||||
set_current_state(TASK_UNINTERRUPTIBLE);
|
set_current_state(TASK_UNINTERRUPTIBLE);
|
||||||
|
@ -336,7 +336,7 @@ ssize_t fastcall wait_on_sync_kiocb(struct kiocb *iocb)
|
||||||
* go away, they will call put_ioctx and release any pinned memory
|
* go away, they will call put_ioctx and release any pinned memory
|
||||||
* associated with the request (held via struct page * references).
|
* associated with the request (held via struct page * references).
|
||||||
*/
|
*/
|
||||||
void fastcall exit_aio(struct mm_struct *mm)
|
void exit_aio(struct mm_struct *mm)
|
||||||
{
|
{
|
||||||
struct kioctx *ctx = mm->ioctx_list;
|
struct kioctx *ctx = mm->ioctx_list;
|
||||||
mm->ioctx_list = NULL;
|
mm->ioctx_list = NULL;
|
||||||
|
@ -365,7 +365,7 @@ void fastcall exit_aio(struct mm_struct *mm)
|
||||||
* Called when the last user of an aio context has gone away,
|
* Called when the last user of an aio context has gone away,
|
||||||
* and the struct needs to be freed.
|
* and the struct needs to be freed.
|
||||||
*/
|
*/
|
||||||
void fastcall __put_ioctx(struct kioctx *ctx)
|
void __put_ioctx(struct kioctx *ctx)
|
||||||
{
|
{
|
||||||
unsigned nr_events = ctx->max_reqs;
|
unsigned nr_events = ctx->max_reqs;
|
||||||
|
|
||||||
|
@ -397,8 +397,7 @@ void fastcall __put_ioctx(struct kioctx *ctx)
|
||||||
* This prevents races between the aio code path referencing the
|
* This prevents races between the aio code path referencing the
|
||||||
* req (after submitting it) and aio_complete() freeing the req.
|
* req (after submitting it) and aio_complete() freeing the req.
|
||||||
*/
|
*/
|
||||||
static struct kiocb *__aio_get_req(struct kioctx *ctx);
|
static struct kiocb *__aio_get_req(struct kioctx *ctx)
|
||||||
static struct kiocb fastcall *__aio_get_req(struct kioctx *ctx)
|
|
||||||
{
|
{
|
||||||
struct kiocb *req = NULL;
|
struct kiocb *req = NULL;
|
||||||
struct aio_ring *ring;
|
struct aio_ring *ring;
|
||||||
|
@ -533,7 +532,7 @@ static int __aio_put_req(struct kioctx *ctx, struct kiocb *req)
|
||||||
* Returns true if this put was the last user of the kiocb,
|
* Returns true if this put was the last user of the kiocb,
|
||||||
* false if the request is still in use.
|
* false if the request is still in use.
|
||||||
*/
|
*/
|
||||||
int fastcall aio_put_req(struct kiocb *req)
|
int aio_put_req(struct kiocb *req)
|
||||||
{
|
{
|
||||||
struct kioctx *ctx = req->ki_ctx;
|
struct kioctx *ctx = req->ki_ctx;
|
||||||
int ret;
|
int ret;
|
||||||
|
@ -893,7 +892,7 @@ static void try_queue_kicked_iocb(struct kiocb *iocb)
|
||||||
* The retry is usually executed by aio workqueue
|
* The retry is usually executed by aio workqueue
|
||||||
* threads (See aio_kick_handler).
|
* threads (See aio_kick_handler).
|
||||||
*/
|
*/
|
||||||
void fastcall kick_iocb(struct kiocb *iocb)
|
void kick_iocb(struct kiocb *iocb)
|
||||||
{
|
{
|
||||||
/* sync iocbs are easy: they can only ever be executing from a
|
/* sync iocbs are easy: they can only ever be executing from a
|
||||||
* single context. */
|
* single context. */
|
||||||
|
@ -912,7 +911,7 @@ EXPORT_SYMBOL(kick_iocb);
|
||||||
* Returns true if this is the last user of the request. The
|
* Returns true if this is the last user of the request. The
|
||||||
* only other user of the request can be the cancellation code.
|
* only other user of the request can be the cancellation code.
|
||||||
*/
|
*/
|
||||||
int fastcall aio_complete(struct kiocb *iocb, long res, long res2)
|
int aio_complete(struct kiocb *iocb, long res, long res2)
|
||||||
{
|
{
|
||||||
struct kioctx *ctx = iocb->ki_ctx;
|
struct kioctx *ctx = iocb->ki_ctx;
|
||||||
struct aio_ring_info *info;
|
struct aio_ring_info *info;
|
||||||
|
@ -1523,7 +1522,7 @@ static int aio_wake_function(wait_queue_t *wait, unsigned mode,
|
||||||
return 1;
|
return 1;
|
||||||
}
|
}
|
||||||
|
|
||||||
int fastcall io_submit_one(struct kioctx *ctx, struct iocb __user *user_iocb,
|
int io_submit_one(struct kioctx *ctx, struct iocb __user *user_iocb,
|
||||||
struct iocb *iocb)
|
struct iocb *iocb)
|
||||||
{
|
{
|
||||||
struct kiocb *req;
|
struct kiocb *req;
|
||||||
|
|
|
@ -67,14 +67,14 @@ static int sync_buffer(void *word)
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
void fastcall __lock_buffer(struct buffer_head *bh)
|
void __lock_buffer(struct buffer_head *bh)
|
||||||
{
|
{
|
||||||
wait_on_bit_lock(&bh->b_state, BH_Lock, sync_buffer,
|
wait_on_bit_lock(&bh->b_state, BH_Lock, sync_buffer,
|
||||||
TASK_UNINTERRUPTIBLE);
|
TASK_UNINTERRUPTIBLE);
|
||||||
}
|
}
|
||||||
EXPORT_SYMBOL(__lock_buffer);
|
EXPORT_SYMBOL(__lock_buffer);
|
||||||
|
|
||||||
void fastcall unlock_buffer(struct buffer_head *bh)
|
void unlock_buffer(struct buffer_head *bh)
|
||||||
{
|
{
|
||||||
smp_mb__before_clear_bit();
|
smp_mb__before_clear_bit();
|
||||||
clear_buffer_locked(bh);
|
clear_buffer_locked(bh);
|
||||||
|
@ -1164,7 +1164,7 @@ __getblk_slow(struct block_device *bdev, sector_t block, int size)
|
||||||
* mark_buffer_dirty() is atomic. It takes bh->b_page->mapping->private_lock,
|
* mark_buffer_dirty() is atomic. It takes bh->b_page->mapping->private_lock,
|
||||||
* mapping->tree_lock and the global inode_lock.
|
* mapping->tree_lock and the global inode_lock.
|
||||||
*/
|
*/
|
||||||
void fastcall mark_buffer_dirty(struct buffer_head *bh)
|
void mark_buffer_dirty(struct buffer_head *bh)
|
||||||
{
|
{
|
||||||
WARN_ON_ONCE(!buffer_uptodate(bh));
|
WARN_ON_ONCE(!buffer_uptodate(bh));
|
||||||
if (!buffer_dirty(bh) && !test_set_buffer_dirty(bh))
|
if (!buffer_dirty(bh) && !test_set_buffer_dirty(bh))
|
||||||
|
|
|
@ -24,7 +24,7 @@
|
||||||
#include <asm/siginfo.h>
|
#include <asm/siginfo.h>
|
||||||
#include <asm/uaccess.h>
|
#include <asm/uaccess.h>
|
||||||
|
|
||||||
void fastcall set_close_on_exec(unsigned int fd, int flag)
|
void set_close_on_exec(unsigned int fd, int flag)
|
||||||
{
|
{
|
||||||
struct files_struct *files = current->files;
|
struct files_struct *files = current->files;
|
||||||
struct fdtable *fdt;
|
struct fdtable *fdt;
|
||||||
|
|
|
@ -197,7 +197,7 @@ int init_file(struct file *file, struct vfsmount *mnt, struct dentry *dentry,
|
||||||
}
|
}
|
||||||
EXPORT_SYMBOL(init_file);
|
EXPORT_SYMBOL(init_file);
|
||||||
|
|
||||||
void fastcall fput(struct file *file)
|
void fput(struct file *file)
|
||||||
{
|
{
|
||||||
if (atomic_dec_and_test(&file->f_count))
|
if (atomic_dec_and_test(&file->f_count))
|
||||||
__fput(file);
|
__fput(file);
|
||||||
|
@ -208,7 +208,7 @@ EXPORT_SYMBOL(fput);
|
||||||
/* __fput is called from task context when aio completion releases the last
|
/* __fput is called from task context when aio completion releases the last
|
||||||
* last use of a struct file *. Do not use otherwise.
|
* last use of a struct file *. Do not use otherwise.
|
||||||
*/
|
*/
|
||||||
void fastcall __fput(struct file *file)
|
void __fput(struct file *file)
|
||||||
{
|
{
|
||||||
struct dentry *dentry = file->f_path.dentry;
|
struct dentry *dentry = file->f_path.dentry;
|
||||||
struct vfsmount *mnt = file->f_path.mnt;
|
struct vfsmount *mnt = file->f_path.mnt;
|
||||||
|
@ -241,7 +241,7 @@ void fastcall __fput(struct file *file)
|
||||||
mntput(mnt);
|
mntput(mnt);
|
||||||
}
|
}
|
||||||
|
|
||||||
struct file fastcall *fget(unsigned int fd)
|
struct file *fget(unsigned int fd)
|
||||||
{
|
{
|
||||||
struct file *file;
|
struct file *file;
|
||||||
struct files_struct *files = current->files;
|
struct files_struct *files = current->files;
|
||||||
|
@ -269,7 +269,7 @@ EXPORT_SYMBOL(fget);
|
||||||
* and a flag is returned to be passed to the corresponding fput_light().
|
* and a flag is returned to be passed to the corresponding fput_light().
|
||||||
* There must not be a cloning between an fget_light/fput_light pair.
|
* There must not be a cloning between an fget_light/fput_light pair.
|
||||||
*/
|
*/
|
||||||
struct file fastcall *fget_light(unsigned int fd, int *fput_needed)
|
struct file *fget_light(unsigned int fd, int *fput_needed)
|
||||||
{
|
{
|
||||||
struct file *file;
|
struct file *file;
|
||||||
struct files_struct *files = current->files;
|
struct files_struct *files = current->files;
|
||||||
|
|
16
fs/namei.c
16
fs/namei.c
|
@ -106,7 +106,7 @@
|
||||||
* any extra contention...
|
* any extra contention...
|
||||||
*/
|
*/
|
||||||
|
|
||||||
static int fastcall link_path_walk(const char *name, struct nameidata *nd);
|
static int link_path_walk(const char *name, struct nameidata *nd);
|
||||||
|
|
||||||
/* In order to reduce some races, while at the same time doing additional
|
/* In order to reduce some races, while at the same time doing additional
|
||||||
* checking and hopefully speeding things up, we copy filenames to the
|
* checking and hopefully speeding things up, we copy filenames to the
|
||||||
|
@ -823,7 +823,7 @@ static int do_lookup(struct nameidata *nd, struct qstr *name,
|
||||||
* Returns 0 and nd will have valid dentry and mnt on success.
|
* Returns 0 and nd will have valid dentry and mnt on success.
|
||||||
* Returns error and drops reference to input namei data on failure.
|
* Returns error and drops reference to input namei data on failure.
|
||||||
*/
|
*/
|
||||||
static fastcall int __link_path_walk(const char * name, struct nameidata *nd)
|
static int __link_path_walk(const char *name, struct nameidata *nd)
|
||||||
{
|
{
|
||||||
struct path next;
|
struct path next;
|
||||||
struct inode *inode;
|
struct inode *inode;
|
||||||
|
@ -1015,7 +1015,7 @@ static fastcall int __link_path_walk(const char * name, struct nameidata *nd)
|
||||||
* Retry the whole path once, forcing real lookup requests
|
* Retry the whole path once, forcing real lookup requests
|
||||||
* instead of relying on the dcache.
|
* instead of relying on the dcache.
|
||||||
*/
|
*/
|
||||||
static int fastcall link_path_walk(const char *name, struct nameidata *nd)
|
static int link_path_walk(const char *name, struct nameidata *nd)
|
||||||
{
|
{
|
||||||
struct nameidata save = *nd;
|
struct nameidata save = *nd;
|
||||||
int result;
|
int result;
|
||||||
|
@ -1039,7 +1039,7 @@ static int fastcall link_path_walk(const char *name, struct nameidata *nd)
|
||||||
return result;
|
return result;
|
||||||
}
|
}
|
||||||
|
|
||||||
static int fastcall path_walk(const char * name, struct nameidata *nd)
|
static int path_walk(const char *name, struct nameidata *nd)
|
||||||
{
|
{
|
||||||
current->total_link_count = 0;
|
current->total_link_count = 0;
|
||||||
return link_path_walk(name, nd);
|
return link_path_walk(name, nd);
|
||||||
|
@ -1116,7 +1116,7 @@ void set_fs_altroot(void)
|
||||||
}
|
}
|
||||||
|
|
||||||
/* Returns 0 and nd will be valid on success; Retuns error, otherwise. */
|
/* Returns 0 and nd will be valid on success; Retuns error, otherwise. */
|
||||||
static int fastcall do_path_lookup(int dfd, const char *name,
|
static int do_path_lookup(int dfd, const char *name,
|
||||||
unsigned int flags, struct nameidata *nd)
|
unsigned int flags, struct nameidata *nd)
|
||||||
{
|
{
|
||||||
int retval = 0;
|
int retval = 0;
|
||||||
|
@ -1183,7 +1183,7 @@ static int fastcall do_path_lookup(int dfd, const char *name,
|
||||||
goto out_fail;
|
goto out_fail;
|
||||||
}
|
}
|
||||||
|
|
||||||
int fastcall path_lookup(const char *name, unsigned int flags,
|
int path_lookup(const char *name, unsigned int flags,
|
||||||
struct nameidata *nd)
|
struct nameidata *nd)
|
||||||
{
|
{
|
||||||
return do_path_lookup(AT_FDCWD, name, flags, nd);
|
return do_path_lookup(AT_FDCWD, name, flags, nd);
|
||||||
|
@ -1409,7 +1409,7 @@ struct dentry *lookup_one_noperm(const char *name, struct dentry *base)
|
||||||
return __lookup_hash(&this, base, NULL);
|
return __lookup_hash(&this, base, NULL);
|
||||||
}
|
}
|
||||||
|
|
||||||
int fastcall __user_walk_fd(int dfd, const char __user *name, unsigned flags,
|
int __user_walk_fd(int dfd, const char __user *name, unsigned flags,
|
||||||
struct nameidata *nd)
|
struct nameidata *nd)
|
||||||
{
|
{
|
||||||
char *tmp = getname(name);
|
char *tmp = getname(name);
|
||||||
|
@ -1422,7 +1422,7 @@ int fastcall __user_walk_fd(int dfd, const char __user *name, unsigned flags,
|
||||||
return err;
|
return err;
|
||||||
}
|
}
|
||||||
|
|
||||||
int fastcall __user_walk(const char __user *name, unsigned flags, struct nameidata *nd)
|
int __user_walk(const char __user *name, unsigned flags, struct nameidata *nd)
|
||||||
{
|
{
|
||||||
return __user_walk_fd(AT_FDCWD, name, flags, nd);
|
return __user_walk_fd(AT_FDCWD, name, flags, nd);
|
||||||
}
|
}
|
||||||
|
|
|
@ -991,7 +991,7 @@ static void __put_unused_fd(struct files_struct *files, unsigned int fd)
|
||||||
files->next_fd = fd;
|
files->next_fd = fd;
|
||||||
}
|
}
|
||||||
|
|
||||||
void fastcall put_unused_fd(unsigned int fd)
|
void put_unused_fd(unsigned int fd)
|
||||||
{
|
{
|
||||||
struct files_struct *files = current->files;
|
struct files_struct *files = current->files;
|
||||||
spin_lock(&files->file_lock);
|
spin_lock(&files->file_lock);
|
||||||
|
@ -1014,7 +1014,7 @@ EXPORT_SYMBOL(put_unused_fd);
|
||||||
* will follow.
|
* will follow.
|
||||||
*/
|
*/
|
||||||
|
|
||||||
void fastcall fd_install(unsigned int fd, struct file * file)
|
void fd_install(unsigned int fd, struct file *file)
|
||||||
{
|
{
|
||||||
struct files_struct *files = current->files;
|
struct files_struct *files = current->files;
|
||||||
struct fdtable *fdt;
|
struct fdtable *fdt;
|
||||||
|
|
Loading…
Reference in a new issue