1
0
Fork 0

prctl: avoid using mmap_sem for exe_file serialization

Oleg cleverly suggested using xchg() to set the new mm->exe_file instead
of calling set_mm_exe_file() which requires some form of serialization --
mmap_sem in this case.  For archs that do not have atomic rmw instructions
we still fallback to a spinlock alternative, so this should always be
safe.  As such, we only need the mmap_sem for looking up the backing
vm_file, which can be done sharing the lock.  Naturally, this means we
need to manually deal with both the new and old file reference counting,
and we need not worry about the MMF_EXE_FILE_CHANGED bits, which can
probably be deleted in the future anyway.

Signed-off-by: Davidlohr Bueso <dbueso@suse.de>
Suggested-by: Oleg Nesterov <oleg@redhat.com>
Acked-by: Oleg Nesterov <oleg@redhat.com>
Reviewed-by: Konstantin Khlebnikov <khlebnikov@yandex-team.ru>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
hifive-unleashed-5.1
Davidlohr Bueso 2015-04-16 12:47:59 -07:00 committed by Linus Torvalds
parent 90f31d0ea8
commit 6e399cd144
3 changed files with 47 additions and 25 deletions

View File

@ -1078,7 +1078,13 @@ int flush_old_exec(struct linux_binprm * bprm)
if (retval) if (retval)
goto out; goto out;
/*
* Must be called _before_ exec_mmap() as bprm->mm is
* not visibile until then. This also enables the update
* to be lockless.
*/
set_mm_exe_file(bprm->mm, bprm->file); set_mm_exe_file(bprm->mm, bprm->file);
/* /*
* Release all of the old mmap stuff * Release all of the old mmap stuff
*/ */

View File

@ -711,15 +711,22 @@ EXPORT_SYMBOL_GPL(mmput);
* *
* This changes mm's executable file (shown as symlink /proc/[pid]/exe). * This changes mm's executable file (shown as symlink /proc/[pid]/exe).
* *
* Main users are mmput(), sys_execve() and sys_prctl(PR_SET_MM_MAP/EXE_FILE). * Main users are mmput() and sys_execve(). Callers prevent concurrent
* Callers prevent concurrent invocations: in mmput() nobody alive left, * invocations: in mmput() nobody alive left, in execve task is single
* in execve task is single-threaded, prctl holds mmap_sem exclusively. * threaded. sys_prctl(PR_SET_MM_MAP/EXE_FILE) also needs to set the
* mm->exe_file, but does so without using set_mm_exe_file() in order
* to do avoid the need for any locks.
*/ */
void set_mm_exe_file(struct mm_struct *mm, struct file *new_exe_file) void set_mm_exe_file(struct mm_struct *mm, struct file *new_exe_file)
{ {
struct file *old_exe_file = rcu_dereference_protected(mm->exe_file, struct file *old_exe_file;
!atomic_read(&mm->mm_users) || current->in_execve ||
lockdep_is_held(&mm->mmap_sem)); /*
* It is safe to dereference the exe_file without RCU as
* this function is only called if nobody else can access
* this mm -- see comment above for justification.
*/
old_exe_file = rcu_dereference_raw(mm->exe_file);
if (new_exe_file) if (new_exe_file)
get_file(new_exe_file); get_file(new_exe_file);

View File

@ -1649,14 +1649,13 @@ SYSCALL_DEFINE1(umask, int, mask)
return mask; return mask;
} }
static int prctl_set_mm_exe_file_locked(struct mm_struct *mm, unsigned int fd) static int prctl_set_mm_exe_file(struct mm_struct *mm, unsigned int fd)
{ {
struct fd exe; struct fd exe;
struct file *old_exe, *exe_file;
struct inode *inode; struct inode *inode;
int err; int err;
VM_BUG_ON_MM(!rwsem_is_locked(&mm->mmap_sem), mm);
exe = fdget(fd); exe = fdget(fd);
if (!exe.file) if (!exe.file)
return -EBADF; return -EBADF;
@ -1680,15 +1679,22 @@ static int prctl_set_mm_exe_file_locked(struct mm_struct *mm, unsigned int fd)
/* /*
* Forbid mm->exe_file change if old file still mapped. * Forbid mm->exe_file change if old file still mapped.
*/ */
exe_file = get_mm_exe_file(mm);
err = -EBUSY; err = -EBUSY;
if (mm->exe_file) { if (exe_file) {
struct vm_area_struct *vma; struct vm_area_struct *vma;
for (vma = mm->mmap; vma; vma = vma->vm_next) down_read(&mm->mmap_sem);
if (vma->vm_file && for (vma = mm->mmap; vma; vma = vma->vm_next) {
path_equal(&vma->vm_file->f_path, if (!vma->vm_file)
&mm->exe_file->f_path)) continue;
goto exit; if (path_equal(&vma->vm_file->f_path,
&exe_file->f_path))
goto exit_err;
}
up_read(&mm->mmap_sem);
fput(exe_file);
} }
/* /*
@ -1702,10 +1708,18 @@ static int prctl_set_mm_exe_file_locked(struct mm_struct *mm, unsigned int fd)
goto exit; goto exit;
err = 0; err = 0;
set_mm_exe_file(mm, exe.file); /* this grabs a reference to exe.file */ /* set the new file, lockless */
get_file(exe.file);
old_exe = xchg(&mm->exe_file, exe.file);
if (old_exe)
fput(old_exe);
exit: exit:
fdput(exe); fdput(exe);
return err; return err;
exit_err:
up_read(&mm->mmap_sem);
fput(exe_file);
goto exit;
} }
#ifdef CONFIG_CHECKPOINT_RESTORE #ifdef CONFIG_CHECKPOINT_RESTORE
@ -1840,10 +1854,9 @@ static int prctl_set_mm_map(int opt, const void __user *addr, unsigned long data
user_auxv[AT_VECTOR_SIZE - 1] = AT_NULL; user_auxv[AT_VECTOR_SIZE - 1] = AT_NULL;
} }
down_write(&mm->mmap_sem);
if (prctl_map.exe_fd != (u32)-1) if (prctl_map.exe_fd != (u32)-1)
error = prctl_set_mm_exe_file_locked(mm, prctl_map.exe_fd); error = prctl_set_mm_exe_file(mm, prctl_map.exe_fd);
downgrade_write(&mm->mmap_sem); down_read(&mm->mmap_sem);
if (error) if (error)
goto out; goto out;
@ -1909,12 +1922,8 @@ static int prctl_set_mm(int opt, unsigned long addr,
if (!capable(CAP_SYS_RESOURCE)) if (!capable(CAP_SYS_RESOURCE))
return -EPERM; return -EPERM;
if (opt == PR_SET_MM_EXE_FILE) { if (opt == PR_SET_MM_EXE_FILE)
down_write(&mm->mmap_sem); return prctl_set_mm_exe_file(mm, (unsigned int)addr);
error = prctl_set_mm_exe_file_locked(mm, (unsigned int)addr);
up_write(&mm->mmap_sem);
return error;
}
if (addr >= TASK_SIZE || addr < mmap_min_addr) if (addr >= TASK_SIZE || addr < mmap_min_addr)
return -EINVAL; return -EINVAL;