2005-04-17 00:20:36 +02:00
|
|
|
/*
|
|
|
|
* linux/fs/namespace.c
|
|
|
|
*
|
|
|
|
* (C) Copyright Al Viro 2000, 2001
|
|
|
|
* Released under GPL v2.
|
|
|
|
*
|
|
|
|
* Based on code from fs/super.c, copyright Linus Torvalds and others.
|
|
|
|
* Heavily rewritten.
|
|
|
|
*/
|
|
|
|
|
|
|
|
#include <linux/syscalls.h>
|
2011-12-07 19:06:11 +01:00
|
|
|
#include <linux/export.h>
|
2006-01-11 21:17:46 +01:00
|
|
|
#include <linux/capability.h>
|
2006-12-08 11:37:56 +01:00
|
|
|
#include <linux/mnt_namespace.h>
|
2012-07-27 06:08:32 +02:00
|
|
|
#include <linux/user_namespace.h>
|
2005-04-17 00:20:36 +02:00
|
|
|
#include <linux/namei.h>
|
|
|
|
#include <linux/security.h>
|
2017-02-02 17:54:15 +01:00
|
|
|
#include <linux/cred.h>
|
2008-03-26 22:11:34 +01:00
|
|
|
#include <linux/idr.h>
|
2013-09-11 23:26:10 +02:00
|
|
|
#include <linux/init.h> /* init_rootfs */
|
2011-12-07 19:06:11 +01:00
|
|
|
#include <linux/fs_struct.h> /* get_fs_root et.al. */
|
|
|
|
#include <linux/fsnotify.h> /* fsnotify_vfsmount_delete */
|
|
|
|
#include <linux/uaccess.h>
|
2013-04-12 02:50:06 +02:00
|
|
|
#include <linux/proc_ns.h>
|
2013-05-02 02:51:54 +02:00
|
|
|
#include <linux/magic.h>
|
2014-02-28 19:46:44 +01:00
|
|
|
#include <linux/bootmem.h>
|
2014-08-08 19:08:20 +02:00
|
|
|
#include <linux/task_work.h>
|
2017-02-04 01:20:53 +01:00
|
|
|
#include <linux/sched/task.h>
|
|
|
|
|
2005-11-07 23:19:07 +01:00
|
|
|
#include "pnode.h"
|
2007-07-16 08:41:25 +02:00
|
|
|
#include "internal.h"
|
2005-04-17 00:20:36 +02:00
|
|
|
|
2016-09-28 07:27:17 +02:00
|
|
|
/* Maximum number of mounts in a mount namespace */
|
|
|
|
unsigned int sysctl_mount_max __read_mostly = 100000;
|
|
|
|
|
2014-02-28 19:46:44 +01:00
|
|
|
static unsigned int m_hash_mask __read_mostly;
|
|
|
|
static unsigned int m_hash_shift __read_mostly;
|
|
|
|
static unsigned int mp_hash_mask __read_mostly;
|
|
|
|
static unsigned int mp_hash_shift __read_mostly;
|
|
|
|
|
|
|
|
static __initdata unsigned long mhash_entries;
|
|
|
|
static int __init set_mhash_entries(char *str)
|
|
|
|
{
|
|
|
|
if (!str)
|
|
|
|
return 0;
|
|
|
|
mhash_entries = simple_strtoul(str, &str, 0);
|
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
__setup("mhash_entries=", set_mhash_entries);
|
|
|
|
|
|
|
|
static __initdata unsigned long mphash_entries;
|
|
|
|
static int __init set_mphash_entries(char *str)
|
|
|
|
{
|
|
|
|
if (!str)
|
|
|
|
return 0;
|
|
|
|
mphash_entries = simple_strtoul(str, &str, 0);
|
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
__setup("mphash_entries=", set_mphash_entries);
|
2008-02-06 10:37:57 +01:00
|
|
|
|
2014-02-27 20:40:10 +01:00
|
|
|
static u64 event;
|
2008-03-26 22:11:34 +01:00
|
|
|
static DEFINE_IDA(mnt_id_ida);
|
2008-03-27 13:06:23 +01:00
|
|
|
static DEFINE_IDA(mnt_group_ida);
|
fs: brlock vfsmount_lock
fs: brlock vfsmount_lock
Use a brlock for the vfsmount lock. It must be taken for write whenever
modifying the mount hash or associated fields, and may be taken for read when
performing mount hash lookups.
A new lock is added for the mnt-id allocator, so it doesn't need to take
the heavy vfsmount write-lock.
The number of atomics should remain the same for fastpath rlock cases, though
code would be slightly slower due to per-cpu access. Scalability is not not be
much improved in common cases yet, due to other locks (ie. dcache_lock) getting
in the way. However path lookups crossing mountpoints should be one case where
scalability is improved (currently requiring the global lock).
The slowpath is slower due to use of brlock. On a 64 core, 64 socket, 32 node
Altix system (high latency to remote nodes), a simple umount microbenchmark
(mount --bind mnt mnt2 ; umount mnt2 loop 1000 times), before this patch it
took 6.8s, afterwards took 7.1s, about 5% slower.
Cc: Al Viro <viro@ZenIV.linux.org.uk>
Signed-off-by: Nick Piggin <npiggin@kernel.dk>
Signed-off-by: Al Viro <viro@zeniv.linux.org.uk>
2010-08-17 20:37:39 +02:00
|
|
|
static DEFINE_SPINLOCK(mnt_id_lock);
|
2009-06-24 09:12:00 +02:00
|
|
|
static int mnt_id_start = 0;
|
|
|
|
static int mnt_group_start = 1;
|
2005-04-17 00:20:36 +02:00
|
|
|
|
2014-03-21 02:10:51 +01:00
|
|
|
static struct hlist_head *mount_hashtable __read_mostly;
|
2014-02-28 19:46:44 +01:00
|
|
|
static struct hlist_head *mountpoint_hashtable __read_mostly;
|
2006-12-07 05:33:20 +01:00
|
|
|
static struct kmem_cache *mnt_cache __read_mostly;
|
2013-09-17 03:34:53 +02:00
|
|
|
static DECLARE_RWSEM(namespace_sem);
|
2005-04-17 00:20:36 +02:00
|
|
|
|
2006-01-17 07:14:23 +01:00
|
|
|
/* /sys/fs */
|
2007-10-29 21:17:23 +01:00
|
|
|
struct kobject *fs_kobj;
|
|
|
|
EXPORT_SYMBOL_GPL(fs_kobj);
|
2006-01-17 07:14:23 +01:00
|
|
|
|
fs: brlock vfsmount_lock
fs: brlock vfsmount_lock
Use a brlock for the vfsmount lock. It must be taken for write whenever
modifying the mount hash or associated fields, and may be taken for read when
performing mount hash lookups.
A new lock is added for the mnt-id allocator, so it doesn't need to take
the heavy vfsmount write-lock.
The number of atomics should remain the same for fastpath rlock cases, though
code would be slightly slower due to per-cpu access. Scalability is not not be
much improved in common cases yet, due to other locks (ie. dcache_lock) getting
in the way. However path lookups crossing mountpoints should be one case where
scalability is improved (currently requiring the global lock).
The slowpath is slower due to use of brlock. On a 64 core, 64 socket, 32 node
Altix system (high latency to remote nodes), a simple umount microbenchmark
(mount --bind mnt mnt2 ; umount mnt2 loop 1000 times), before this patch it
took 6.8s, afterwards took 7.1s, about 5% slower.
Cc: Al Viro <viro@ZenIV.linux.org.uk>
Signed-off-by: Nick Piggin <npiggin@kernel.dk>
Signed-off-by: Al Viro <viro@zeniv.linux.org.uk>
2010-08-17 20:37:39 +02:00
|
|
|
/*
|
|
|
|
* vfsmount lock may be taken for read to prevent changes to the
|
|
|
|
* vfsmount hash, ie. during mountpoint lookups or walking back
|
|
|
|
* up the tree.
|
|
|
|
*
|
|
|
|
* It should be taken for write in all cases where the vfsmount
|
|
|
|
* tree or hash is modified or when a vfsmount structure is modified.
|
|
|
|
*/
|
2013-09-30 04:06:07 +02:00
|
|
|
__cacheline_aligned_in_smp DEFINE_SEQLOCK(mount_lock);
|
fs: brlock vfsmount_lock
fs: brlock vfsmount_lock
Use a brlock for the vfsmount lock. It must be taken for write whenever
modifying the mount hash or associated fields, and may be taken for read when
performing mount hash lookups.
A new lock is added for the mnt-id allocator, so it doesn't need to take
the heavy vfsmount write-lock.
The number of atomics should remain the same for fastpath rlock cases, though
code would be slightly slower due to per-cpu access. Scalability is not not be
much improved in common cases yet, due to other locks (ie. dcache_lock) getting
in the way. However path lookups crossing mountpoints should be one case where
scalability is improved (currently requiring the global lock).
The slowpath is slower due to use of brlock. On a 64 core, 64 socket, 32 node
Altix system (high latency to remote nodes), a simple umount microbenchmark
(mount --bind mnt mnt2 ; umount mnt2 loop 1000 times), before this patch it
took 6.8s, afterwards took 7.1s, about 5% slower.
Cc: Al Viro <viro@ZenIV.linux.org.uk>
Signed-off-by: Nick Piggin <npiggin@kernel.dk>
Signed-off-by: Al Viro <viro@zeniv.linux.org.uk>
2010-08-17 20:37:39 +02:00
|
|
|
|
2014-03-21 02:10:51 +01:00
|
|
|
static inline struct hlist_head *m_hash(struct vfsmount *mnt, struct dentry *dentry)
|
2005-04-17 00:20:36 +02:00
|
|
|
{
|
2005-11-07 23:16:09 +01:00
|
|
|
unsigned long tmp = ((unsigned long)mnt / L1_CACHE_BYTES);
|
|
|
|
tmp += ((unsigned long)dentry / L1_CACHE_BYTES);
|
2014-02-28 19:46:44 +01:00
|
|
|
tmp = tmp + (tmp >> m_hash_shift);
|
|
|
|
return &mount_hashtable[tmp & m_hash_mask];
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline struct hlist_head *mp_hash(struct dentry *dentry)
|
|
|
|
{
|
|
|
|
unsigned long tmp = ((unsigned long)dentry / L1_CACHE_BYTES);
|
|
|
|
tmp = tmp + (tmp >> mp_hash_shift);
|
|
|
|
return &mountpoint_hashtable[tmp & mp_hash_mask];
|
2005-04-17 00:20:36 +02:00
|
|
|
}
|
|
|
|
|
2011-11-25 02:38:33 +01:00
|
|
|
static int mnt_alloc_id(struct mount *mnt)
|
2008-03-26 22:11:34 +01:00
|
|
|
{
|
|
|
|
int res;
|
|
|
|
|
|
|
|
retry:
|
|
|
|
ida_pre_get(&mnt_id_ida, GFP_KERNEL);
|
fs: brlock vfsmount_lock
fs: brlock vfsmount_lock
Use a brlock for the vfsmount lock. It must be taken for write whenever
modifying the mount hash or associated fields, and may be taken for read when
performing mount hash lookups.
A new lock is added for the mnt-id allocator, so it doesn't need to take
the heavy vfsmount write-lock.
The number of atomics should remain the same for fastpath rlock cases, though
code would be slightly slower due to per-cpu access. Scalability is not not be
much improved in common cases yet, due to other locks (ie. dcache_lock) getting
in the way. However path lookups crossing mountpoints should be one case where
scalability is improved (currently requiring the global lock).
The slowpath is slower due to use of brlock. On a 64 core, 64 socket, 32 node
Altix system (high latency to remote nodes), a simple umount microbenchmark
(mount --bind mnt mnt2 ; umount mnt2 loop 1000 times), before this patch it
took 6.8s, afterwards took 7.1s, about 5% slower.
Cc: Al Viro <viro@ZenIV.linux.org.uk>
Signed-off-by: Nick Piggin <npiggin@kernel.dk>
Signed-off-by: Al Viro <viro@zeniv.linux.org.uk>
2010-08-17 20:37:39 +02:00
|
|
|
spin_lock(&mnt_id_lock);
|
2011-11-25 06:50:41 +01:00
|
|
|
res = ida_get_new_above(&mnt_id_ida, mnt_id_start, &mnt->mnt_id);
|
2009-06-24 09:12:00 +02:00
|
|
|
if (!res)
|
2011-11-25 06:50:41 +01:00
|
|
|
mnt_id_start = mnt->mnt_id + 1;
|
fs: brlock vfsmount_lock
fs: brlock vfsmount_lock
Use a brlock for the vfsmount lock. It must be taken for write whenever
modifying the mount hash or associated fields, and may be taken for read when
performing mount hash lookups.
A new lock is added for the mnt-id allocator, so it doesn't need to take
the heavy vfsmount write-lock.
The number of atomics should remain the same for fastpath rlock cases, though
code would be slightly slower due to per-cpu access. Scalability is not not be
much improved in common cases yet, due to other locks (ie. dcache_lock) getting
in the way. However path lookups crossing mountpoints should be one case where
scalability is improved (currently requiring the global lock).
The slowpath is slower due to use of brlock. On a 64 core, 64 socket, 32 node
Altix system (high latency to remote nodes), a simple umount microbenchmark
(mount --bind mnt mnt2 ; umount mnt2 loop 1000 times), before this patch it
took 6.8s, afterwards took 7.1s, about 5% slower.
Cc: Al Viro <viro@ZenIV.linux.org.uk>
Signed-off-by: Nick Piggin <npiggin@kernel.dk>
Signed-off-by: Al Viro <viro@zeniv.linux.org.uk>
2010-08-17 20:37:39 +02:00
|
|
|
spin_unlock(&mnt_id_lock);
|
2008-03-26 22:11:34 +01:00
|
|
|
if (res == -EAGAIN)
|
|
|
|
goto retry;
|
|
|
|
|
|
|
|
return res;
|
|
|
|
}
|
|
|
|
|
2011-11-25 02:38:33 +01:00
|
|
|
static void mnt_free_id(struct mount *mnt)
|
2008-03-26 22:11:34 +01:00
|
|
|
{
|
2011-11-25 06:50:41 +01:00
|
|
|
int id = mnt->mnt_id;
|
fs: brlock vfsmount_lock
fs: brlock vfsmount_lock
Use a brlock for the vfsmount lock. It must be taken for write whenever
modifying the mount hash or associated fields, and may be taken for read when
performing mount hash lookups.
A new lock is added for the mnt-id allocator, so it doesn't need to take
the heavy vfsmount write-lock.
The number of atomics should remain the same for fastpath rlock cases, though
code would be slightly slower due to per-cpu access. Scalability is not not be
much improved in common cases yet, due to other locks (ie. dcache_lock) getting
in the way. However path lookups crossing mountpoints should be one case where
scalability is improved (currently requiring the global lock).
The slowpath is slower due to use of brlock. On a 64 core, 64 socket, 32 node
Altix system (high latency to remote nodes), a simple umount microbenchmark
(mount --bind mnt mnt2 ; umount mnt2 loop 1000 times), before this patch it
took 6.8s, afterwards took 7.1s, about 5% slower.
Cc: Al Viro <viro@ZenIV.linux.org.uk>
Signed-off-by: Nick Piggin <npiggin@kernel.dk>
Signed-off-by: Al Viro <viro@zeniv.linux.org.uk>
2010-08-17 20:37:39 +02:00
|
|
|
spin_lock(&mnt_id_lock);
|
2009-06-24 09:12:00 +02:00
|
|
|
ida_remove(&mnt_id_ida, id);
|
|
|
|
if (mnt_id_start > id)
|
|
|
|
mnt_id_start = id;
|
fs: brlock vfsmount_lock
fs: brlock vfsmount_lock
Use a brlock for the vfsmount lock. It must be taken for write whenever
modifying the mount hash or associated fields, and may be taken for read when
performing mount hash lookups.
A new lock is added for the mnt-id allocator, so it doesn't need to take
the heavy vfsmount write-lock.
The number of atomics should remain the same for fastpath rlock cases, though
code would be slightly slower due to per-cpu access. Scalability is not not be
much improved in common cases yet, due to other locks (ie. dcache_lock) getting
in the way. However path lookups crossing mountpoints should be one case where
scalability is improved (currently requiring the global lock).
The slowpath is slower due to use of brlock. On a 64 core, 64 socket, 32 node
Altix system (high latency to remote nodes), a simple umount microbenchmark
(mount --bind mnt mnt2 ; umount mnt2 loop 1000 times), before this patch it
took 6.8s, afterwards took 7.1s, about 5% slower.
Cc: Al Viro <viro@ZenIV.linux.org.uk>
Signed-off-by: Nick Piggin <npiggin@kernel.dk>
Signed-off-by: Al Viro <viro@zeniv.linux.org.uk>
2010-08-17 20:37:39 +02:00
|
|
|
spin_unlock(&mnt_id_lock);
|
2008-03-26 22:11:34 +01:00
|
|
|
}
|
|
|
|
|
2008-03-27 13:06:23 +01:00
|
|
|
/*
|
|
|
|
* Allocate a new peer group ID
|
|
|
|
*
|
|
|
|
* mnt_group_ida is protected by namespace_sem
|
|
|
|
*/
|
2011-11-25 01:54:23 +01:00
|
|
|
static int mnt_alloc_group_id(struct mount *mnt)
|
2008-03-27 13:06:23 +01:00
|
|
|
{
|
2009-06-24 09:12:00 +02:00
|
|
|
int res;
|
|
|
|
|
2008-03-27 13:06:23 +01:00
|
|
|
if (!ida_pre_get(&mnt_group_ida, GFP_KERNEL))
|
|
|
|
return -ENOMEM;
|
|
|
|
|
2009-06-24 09:12:00 +02:00
|
|
|
res = ida_get_new_above(&mnt_group_ida,
|
|
|
|
mnt_group_start,
|
2011-11-25 06:50:41 +01:00
|
|
|
&mnt->mnt_group_id);
|
2009-06-24 09:12:00 +02:00
|
|
|
if (!res)
|
2011-11-25 06:50:41 +01:00
|
|
|
mnt_group_start = mnt->mnt_group_id + 1;
|
2009-06-24 09:12:00 +02:00
|
|
|
|
|
|
|
return res;
|
2008-03-27 13:06:23 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Release a peer group ID
|
|
|
|
*/
|
2011-11-25 01:54:23 +01:00
|
|
|
void mnt_release_group_id(struct mount *mnt)
|
2008-03-27 13:06:23 +01:00
|
|
|
{
|
2011-11-25 06:50:41 +01:00
|
|
|
int id = mnt->mnt_group_id;
|
2009-06-24 09:12:00 +02:00
|
|
|
ida_remove(&mnt_group_ida, id);
|
|
|
|
if (mnt_group_start > id)
|
|
|
|
mnt_group_start = id;
|
2011-11-25 06:50:41 +01:00
|
|
|
mnt->mnt_group_id = 0;
|
2008-03-27 13:06:23 +01:00
|
|
|
}
|
|
|
|
|
fs: scale mntget/mntput
The problem that this patch aims to fix is vfsmount refcounting scalability.
We need to take a reference on the vfsmount for every successful path lookup,
which often go to the same mount point.
The fundamental difficulty is that a "simple" reference count can never be made
scalable, because any time a reference is dropped, we must check whether that
was the last reference. To do that requires communication with all other CPUs
that may have taken a reference count.
We can make refcounts more scalable in a couple of ways, involving keeping
distributed counters, and checking for the global-zero condition less
frequently.
- check the global sum once every interval (this will delay zero detection
for some interval, so it's probably a showstopper for vfsmounts).
- keep a local count and only taking the global sum when local reaches 0 (this
is difficult for vfsmounts, because we can't hold preempt off for the life of
a reference, so a counter would need to be per-thread or tied strongly to a
particular CPU which requires more locking).
- keep a local difference of increments and decrements, which allows us to sum
the total difference and hence find the refcount when summing all CPUs. Then,
keep a single integer "long" refcount for slow and long lasting references,
and only take the global sum of local counters when the long refcount is 0.
This last scheme is what I implemented here. Attached mounts and process root
and working directory references are "long" references, and everything else is
a short reference.
This allows scalable vfsmount references during path walking over mounted
subtrees and unattached (lazy umounted) mounts with processes still running
in them.
This results in one fewer atomic op in the fastpath: mntget is now just a
per-CPU inc, rather than an atomic inc; and mntput just requires a spinlock
and non-atomic decrement in the common case. However code is otherwise bigger
and heavier, so single threaded performance is basically a wash.
Signed-off-by: Nick Piggin <npiggin@kernel.dk>
2011-01-07 07:50:11 +01:00
|
|
|
/*
|
|
|
|
* vfsmount lock must be held for read
|
|
|
|
*/
|
2011-11-25 04:37:54 +01:00
|
|
|
static inline void mnt_add_count(struct mount *mnt, int n)
|
fs: scale mntget/mntput
The problem that this patch aims to fix is vfsmount refcounting scalability.
We need to take a reference on the vfsmount for every successful path lookup,
which often go to the same mount point.
The fundamental difficulty is that a "simple" reference count can never be made
scalable, because any time a reference is dropped, we must check whether that
was the last reference. To do that requires communication with all other CPUs
that may have taken a reference count.
We can make refcounts more scalable in a couple of ways, involving keeping
distributed counters, and checking for the global-zero condition less
frequently.
- check the global sum once every interval (this will delay zero detection
for some interval, so it's probably a showstopper for vfsmounts).
- keep a local count and only taking the global sum when local reaches 0 (this
is difficult for vfsmounts, because we can't hold preempt off for the life of
a reference, so a counter would need to be per-thread or tied strongly to a
particular CPU which requires more locking).
- keep a local difference of increments and decrements, which allows us to sum
the total difference and hence find the refcount when summing all CPUs. Then,
keep a single integer "long" refcount for slow and long lasting references,
and only take the global sum of local counters when the long refcount is 0.
This last scheme is what I implemented here. Attached mounts and process root
and working directory references are "long" references, and everything else is
a short reference.
This allows scalable vfsmount references during path walking over mounted
subtrees and unattached (lazy umounted) mounts with processes still running
in them.
This results in one fewer atomic op in the fastpath: mntget is now just a
per-CPU inc, rather than an atomic inc; and mntput just requires a spinlock
and non-atomic decrement in the common case. However code is otherwise bigger
and heavier, so single threaded performance is basically a wash.
Signed-off-by: Nick Piggin <npiggin@kernel.dk>
2011-01-07 07:50:11 +01:00
|
|
|
{
|
|
|
|
#ifdef CONFIG_SMP
|
2011-11-25 04:53:09 +01:00
|
|
|
this_cpu_add(mnt->mnt_pcp->mnt_count, n);
|
fs: scale mntget/mntput
The problem that this patch aims to fix is vfsmount refcounting scalability.
We need to take a reference on the vfsmount for every successful path lookup,
which often go to the same mount point.
The fundamental difficulty is that a "simple" reference count can never be made
scalable, because any time a reference is dropped, we must check whether that
was the last reference. To do that requires communication with all other CPUs
that may have taken a reference count.
We can make refcounts more scalable in a couple of ways, involving keeping
distributed counters, and checking for the global-zero condition less
frequently.
- check the global sum once every interval (this will delay zero detection
for some interval, so it's probably a showstopper for vfsmounts).
- keep a local count and only taking the global sum when local reaches 0 (this
is difficult for vfsmounts, because we can't hold preempt off for the life of
a reference, so a counter would need to be per-thread or tied strongly to a
particular CPU which requires more locking).
- keep a local difference of increments and decrements, which allows us to sum
the total difference and hence find the refcount when summing all CPUs. Then,
keep a single integer "long" refcount for slow and long lasting references,
and only take the global sum of local counters when the long refcount is 0.
This last scheme is what I implemented here. Attached mounts and process root
and working directory references are "long" references, and everything else is
a short reference.
This allows scalable vfsmount references during path walking over mounted
subtrees and unattached (lazy umounted) mounts with processes still running
in them.
This results in one fewer atomic op in the fastpath: mntget is now just a
per-CPU inc, rather than an atomic inc; and mntput just requires a spinlock
and non-atomic decrement in the common case. However code is otherwise bigger
and heavier, so single threaded performance is basically a wash.
Signed-off-by: Nick Piggin <npiggin@kernel.dk>
2011-01-07 07:50:11 +01:00
|
|
|
#else
|
|
|
|
preempt_disable();
|
2011-11-25 04:53:09 +01:00
|
|
|
mnt->mnt_count += n;
|
fs: scale mntget/mntput
The problem that this patch aims to fix is vfsmount refcounting scalability.
We need to take a reference on the vfsmount for every successful path lookup,
which often go to the same mount point.
The fundamental difficulty is that a "simple" reference count can never be made
scalable, because any time a reference is dropped, we must check whether that
was the last reference. To do that requires communication with all other CPUs
that may have taken a reference count.
We can make refcounts more scalable in a couple of ways, involving keeping
distributed counters, and checking for the global-zero condition less
frequently.
- check the global sum once every interval (this will delay zero detection
for some interval, so it's probably a showstopper for vfsmounts).
- keep a local count and only taking the global sum when local reaches 0 (this
is difficult for vfsmounts, because we can't hold preempt off for the life of
a reference, so a counter would need to be per-thread or tied strongly to a
particular CPU which requires more locking).
- keep a local difference of increments and decrements, which allows us to sum
the total difference and hence find the refcount when summing all CPUs. Then,
keep a single integer "long" refcount for slow and long lasting references,
and only take the global sum of local counters when the long refcount is 0.
This last scheme is what I implemented here. Attached mounts and process root
and working directory references are "long" references, and everything else is
a short reference.
This allows scalable vfsmount references during path walking over mounted
subtrees and unattached (lazy umounted) mounts with processes still running
in them.
This results in one fewer atomic op in the fastpath: mntget is now just a
per-CPU inc, rather than an atomic inc; and mntput just requires a spinlock
and non-atomic decrement in the common case. However code is otherwise bigger
and heavier, so single threaded performance is basically a wash.
Signed-off-by: Nick Piggin <npiggin@kernel.dk>
2011-01-07 07:50:11 +01:00
|
|
|
preempt_enable();
|
|
|
|
#endif
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* vfsmount lock must be held for write
|
|
|
|
*/
|
2011-11-25 04:37:54 +01:00
|
|
|
unsigned int mnt_get_count(struct mount *mnt)
|
fs: scale mntget/mntput
The problem that this patch aims to fix is vfsmount refcounting scalability.
We need to take a reference on the vfsmount for every successful path lookup,
which often go to the same mount point.
The fundamental difficulty is that a "simple" reference count can never be made
scalable, because any time a reference is dropped, we must check whether that
was the last reference. To do that requires communication with all other CPUs
that may have taken a reference count.
We can make refcounts more scalable in a couple of ways, involving keeping
distributed counters, and checking for the global-zero condition less
frequently.
- check the global sum once every interval (this will delay zero detection
for some interval, so it's probably a showstopper for vfsmounts).
- keep a local count and only taking the global sum when local reaches 0 (this
is difficult for vfsmounts, because we can't hold preempt off for the life of
a reference, so a counter would need to be per-thread or tied strongly to a
particular CPU which requires more locking).
- keep a local difference of increments and decrements, which allows us to sum
the total difference and hence find the refcount when summing all CPUs. Then,
keep a single integer "long" refcount for slow and long lasting references,
and only take the global sum of local counters when the long refcount is 0.
This last scheme is what I implemented here. Attached mounts and process root
and working directory references are "long" references, and everything else is
a short reference.
This allows scalable vfsmount references during path walking over mounted
subtrees and unattached (lazy umounted) mounts with processes still running
in them.
This results in one fewer atomic op in the fastpath: mntget is now just a
per-CPU inc, rather than an atomic inc; and mntput just requires a spinlock
and non-atomic decrement in the common case. However code is otherwise bigger
and heavier, so single threaded performance is basically a wash.
Signed-off-by: Nick Piggin <npiggin@kernel.dk>
2011-01-07 07:50:11 +01:00
|
|
|
{
|
|
|
|
#ifdef CONFIG_SMP
|
2011-01-15 04:30:21 +01:00
|
|
|
unsigned int count = 0;
|
fs: scale mntget/mntput
The problem that this patch aims to fix is vfsmount refcounting scalability.
We need to take a reference on the vfsmount for every successful path lookup,
which often go to the same mount point.
The fundamental difficulty is that a "simple" reference count can never be made
scalable, because any time a reference is dropped, we must check whether that
was the last reference. To do that requires communication with all other CPUs
that may have taken a reference count.
We can make refcounts more scalable in a couple of ways, involving keeping
distributed counters, and checking for the global-zero condition less
frequently.
- check the global sum once every interval (this will delay zero detection
for some interval, so it's probably a showstopper for vfsmounts).
- keep a local count and only taking the global sum when local reaches 0 (this
is difficult for vfsmounts, because we can't hold preempt off for the life of
a reference, so a counter would need to be per-thread or tied strongly to a
particular CPU which requires more locking).
- keep a local difference of increments and decrements, which allows us to sum
the total difference and hence find the refcount when summing all CPUs. Then,
keep a single integer "long" refcount for slow and long lasting references,
and only take the global sum of local counters when the long refcount is 0.
This last scheme is what I implemented here. Attached mounts and process root
and working directory references are "long" references, and everything else is
a short reference.
This allows scalable vfsmount references during path walking over mounted
subtrees and unattached (lazy umounted) mounts with processes still running
in them.
This results in one fewer atomic op in the fastpath: mntget is now just a
per-CPU inc, rather than an atomic inc; and mntput just requires a spinlock
and non-atomic decrement in the common case. However code is otherwise bigger
and heavier, so single threaded performance is basically a wash.
Signed-off-by: Nick Piggin <npiggin@kernel.dk>
2011-01-07 07:50:11 +01:00
|
|
|
int cpu;
|
|
|
|
|
|
|
|
for_each_possible_cpu(cpu) {
|
2011-11-25 04:53:09 +01:00
|
|
|
count += per_cpu_ptr(mnt->mnt_pcp, cpu)->mnt_count;
|
fs: scale mntget/mntput
The problem that this patch aims to fix is vfsmount refcounting scalability.
We need to take a reference on the vfsmount for every successful path lookup,
which often go to the same mount point.
The fundamental difficulty is that a "simple" reference count can never be made
scalable, because any time a reference is dropped, we must check whether that
was the last reference. To do that requires communication with all other CPUs
that may have taken a reference count.
We can make refcounts more scalable in a couple of ways, involving keeping
distributed counters, and checking for the global-zero condition less
frequently.
- check the global sum once every interval (this will delay zero detection
for some interval, so it's probably a showstopper for vfsmounts).
- keep a local count and only taking the global sum when local reaches 0 (this
is difficult for vfsmounts, because we can't hold preempt off for the life of
a reference, so a counter would need to be per-thread or tied strongly to a
particular CPU which requires more locking).
- keep a local difference of increments and decrements, which allows us to sum
the total difference and hence find the refcount when summing all CPUs. Then,
keep a single integer "long" refcount for slow and long lasting references,
and only take the global sum of local counters when the long refcount is 0.
This last scheme is what I implemented here. Attached mounts and process root
and working directory references are "long" references, and everything else is
a short reference.
This allows scalable vfsmount references during path walking over mounted
subtrees and unattached (lazy umounted) mounts with processes still running
in them.
This results in one fewer atomic op in the fastpath: mntget is now just a
per-CPU inc, rather than an atomic inc; and mntput just requires a spinlock
and non-atomic decrement in the common case. However code is otherwise bigger
and heavier, so single threaded performance is basically a wash.
Signed-off-by: Nick Piggin <npiggin@kernel.dk>
2011-01-07 07:50:11 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
return count;
|
|
|
|
#else
|
2011-11-25 04:53:09 +01:00
|
|
|
return mnt->mnt_count;
|
fs: scale mntget/mntput
The problem that this patch aims to fix is vfsmount refcounting scalability.
We need to take a reference on the vfsmount for every successful path lookup,
which often go to the same mount point.
The fundamental difficulty is that a "simple" reference count can never be made
scalable, because any time a reference is dropped, we must check whether that
was the last reference. To do that requires communication with all other CPUs
that may have taken a reference count.
We can make refcounts more scalable in a couple of ways, involving keeping
distributed counters, and checking for the global-zero condition less
frequently.
- check the global sum once every interval (this will delay zero detection
for some interval, so it's probably a showstopper for vfsmounts).
- keep a local count and only taking the global sum when local reaches 0 (this
is difficult for vfsmounts, because we can't hold preempt off for the life of
a reference, so a counter would need to be per-thread or tied strongly to a
particular CPU which requires more locking).
- keep a local difference of increments and decrements, which allows us to sum
the total difference and hence find the refcount when summing all CPUs. Then,
keep a single integer "long" refcount for slow and long lasting references,
and only take the global sum of local counters when the long refcount is 0.
This last scheme is what I implemented here. Attached mounts and process root
and working directory references are "long" references, and everything else is
a short reference.
This allows scalable vfsmount references during path walking over mounted
subtrees and unattached (lazy umounted) mounts with processes still running
in them.
This results in one fewer atomic op in the fastpath: mntget is now just a
per-CPU inc, rather than an atomic inc; and mntput just requires a spinlock
and non-atomic decrement in the common case. However code is otherwise bigger
and heavier, so single threaded performance is basically a wash.
Signed-off-by: Nick Piggin <npiggin@kernel.dk>
2011-01-07 07:50:11 +01:00
|
|
|
#endif
|
|
|
|
}
|
|
|
|
|
2015-01-11 01:01:08 +01:00
|
|
|
static void drop_mountpoint(struct fs_pin *p)
|
|
|
|
{
|
|
|
|
struct mount *m = container_of(p, struct mount, mnt_umount);
|
|
|
|
dput(m->mnt_ex_mountpoint);
|
|
|
|
pin_remove(p);
|
|
|
|
mntput(&m->mnt);
|
|
|
|
}
|
|
|
|
|
2011-11-25 02:38:33 +01:00
|
|
|
static struct mount *alloc_vfsmnt(const char *name)
|
2005-04-17 00:20:36 +02:00
|
|
|
{
|
2011-11-25 08:35:16 +01:00
|
|
|
struct mount *mnt = kmem_cache_zalloc(mnt_cache, GFP_KERNEL);
|
|
|
|
if (mnt) {
|
2008-03-26 22:11:34 +01:00
|
|
|
int err;
|
|
|
|
|
2011-11-25 08:35:16 +01:00
|
|
|
err = mnt_alloc_id(mnt);
|
2008-07-21 12:06:36 +02:00
|
|
|
if (err)
|
|
|
|
goto out_free_cache;
|
|
|
|
|
|
|
|
if (name) {
|
2015-02-13 23:36:41 +01:00
|
|
|
mnt->mnt_devname = kstrdup_const(name, GFP_KERNEL);
|
2011-11-25 08:35:16 +01:00
|
|
|
if (!mnt->mnt_devname)
|
2008-07-21 12:06:36 +02:00
|
|
|
goto out_free_id;
|
2008-03-26 22:11:34 +01:00
|
|
|
}
|
|
|
|
|
fs: scale mntget/mntput
The problem that this patch aims to fix is vfsmount refcounting scalability.
We need to take a reference on the vfsmount for every successful path lookup,
which often go to the same mount point.
The fundamental difficulty is that a "simple" reference count can never be made
scalable, because any time a reference is dropped, we must check whether that
was the last reference. To do that requires communication with all other CPUs
that may have taken a reference count.
We can make refcounts more scalable in a couple of ways, involving keeping
distributed counters, and checking for the global-zero condition less
frequently.
- check the global sum once every interval (this will delay zero detection
for some interval, so it's probably a showstopper for vfsmounts).
- keep a local count and only taking the global sum when local reaches 0 (this
is difficult for vfsmounts, because we can't hold preempt off for the life of
a reference, so a counter would need to be per-thread or tied strongly to a
particular CPU which requires more locking).
- keep a local difference of increments and decrements, which allows us to sum
the total difference and hence find the refcount when summing all CPUs. Then,
keep a single integer "long" refcount for slow and long lasting references,
and only take the global sum of local counters when the long refcount is 0.
This last scheme is what I implemented here. Attached mounts and process root
and working directory references are "long" references, and everything else is
a short reference.
This allows scalable vfsmount references during path walking over mounted
subtrees and unattached (lazy umounted) mounts with processes still running
in them.
This results in one fewer atomic op in the fastpath: mntget is now just a
per-CPU inc, rather than an atomic inc; and mntput just requires a spinlock
and non-atomic decrement in the common case. However code is otherwise bigger
and heavier, so single threaded performance is basically a wash.
Signed-off-by: Nick Piggin <npiggin@kernel.dk>
2011-01-07 07:50:11 +01:00
|
|
|
#ifdef CONFIG_SMP
|
2011-11-25 08:35:16 +01:00
|
|
|
mnt->mnt_pcp = alloc_percpu(struct mnt_pcp);
|
|
|
|
if (!mnt->mnt_pcp)
|
fs: scale mntget/mntput
The problem that this patch aims to fix is vfsmount refcounting scalability.
We need to take a reference on the vfsmount for every successful path lookup,
which often go to the same mount point.
The fundamental difficulty is that a "simple" reference count can never be made
scalable, because any time a reference is dropped, we must check whether that
was the last reference. To do that requires communication with all other CPUs
that may have taken a reference count.
We can make refcounts more scalable in a couple of ways, involving keeping
distributed counters, and checking for the global-zero condition less
frequently.
- check the global sum once every interval (this will delay zero detection
for some interval, so it's probably a showstopper for vfsmounts).
- keep a local count and only taking the global sum when local reaches 0 (this
is difficult for vfsmounts, because we can't hold preempt off for the life of
a reference, so a counter would need to be per-thread or tied strongly to a
particular CPU which requires more locking).
- keep a local difference of increments and decrements, which allows us to sum
the total difference and hence find the refcount when summing all CPUs. Then,
keep a single integer "long" refcount for slow and long lasting references,
and only take the global sum of local counters when the long refcount is 0.
This last scheme is what I implemented here. Attached mounts and process root
and working directory references are "long" references, and everything else is
a short reference.
This allows scalable vfsmount references during path walking over mounted
subtrees and unattached (lazy umounted) mounts with processes still running
in them.
This results in one fewer atomic op in the fastpath: mntget is now just a
per-CPU inc, rather than an atomic inc; and mntput just requires a spinlock
and non-atomic decrement in the common case. However code is otherwise bigger
and heavier, so single threaded performance is basically a wash.
Signed-off-by: Nick Piggin <npiggin@kernel.dk>
2011-01-07 07:50:11 +01:00
|
|
|
goto out_free_devname;
|
|
|
|
|
2011-11-25 08:35:16 +01:00
|
|
|
this_cpu_add(mnt->mnt_pcp->mnt_count, 1);
|
fs: scale mntget/mntput
The problem that this patch aims to fix is vfsmount refcounting scalability.
We need to take a reference on the vfsmount for every successful path lookup,
which often go to the same mount point.
The fundamental difficulty is that a "simple" reference count can never be made
scalable, because any time a reference is dropped, we must check whether that
was the last reference. To do that requires communication with all other CPUs
that may have taken a reference count.
We can make refcounts more scalable in a couple of ways, involving keeping
distributed counters, and checking for the global-zero condition less
frequently.
- check the global sum once every interval (this will delay zero detection
for some interval, so it's probably a showstopper for vfsmounts).
- keep a local count and only taking the global sum when local reaches 0 (this
is difficult for vfsmounts, because we can't hold preempt off for the life of
a reference, so a counter would need to be per-thread or tied strongly to a
particular CPU which requires more locking).
- keep a local difference of increments and decrements, which allows us to sum
the total difference and hence find the refcount when summing all CPUs. Then,
keep a single integer "long" refcount for slow and long lasting references,
and only take the global sum of local counters when the long refcount is 0.
This last scheme is what I implemented here. Attached mounts and process root
and working directory references are "long" references, and everything else is
a short reference.
This allows scalable vfsmount references during path walking over mounted
subtrees and unattached (lazy umounted) mounts with processes still running
in them.
This results in one fewer atomic op in the fastpath: mntget is now just a
per-CPU inc, rather than an atomic inc; and mntput just requires a spinlock
and non-atomic decrement in the common case. However code is otherwise bigger
and heavier, so single threaded performance is basically a wash.
Signed-off-by: Nick Piggin <npiggin@kernel.dk>
2011-01-07 07:50:11 +01:00
|
|
|
#else
|
2011-11-25 08:35:16 +01:00
|
|
|
mnt->mnt_count = 1;
|
|
|
|
mnt->mnt_writers = 0;
|
fs: scale mntget/mntput
The problem that this patch aims to fix is vfsmount refcounting scalability.
We need to take a reference on the vfsmount for every successful path lookup,
which often go to the same mount point.
The fundamental difficulty is that a "simple" reference count can never be made
scalable, because any time a reference is dropped, we must check whether that
was the last reference. To do that requires communication with all other CPUs
that may have taken a reference count.
We can make refcounts more scalable in a couple of ways, involving keeping
distributed counters, and checking for the global-zero condition less
frequently.
- check the global sum once every interval (this will delay zero detection
for some interval, so it's probably a showstopper for vfsmounts).
- keep a local count and only taking the global sum when local reaches 0 (this
is difficult for vfsmounts, because we can't hold preempt off for the life of
a reference, so a counter would need to be per-thread or tied strongly to a
particular CPU which requires more locking).
- keep a local difference of increments and decrements, which allows us to sum
the total difference and hence find the refcount when summing all CPUs. Then,
keep a single integer "long" refcount for slow and long lasting references,
and only take the global sum of local counters when the long refcount is 0.
This last scheme is what I implemented here. Attached mounts and process root
and working directory references are "long" references, and everything else is
a short reference.
This allows scalable vfsmount references during path walking over mounted
subtrees and unattached (lazy umounted) mounts with processes still running
in them.
This results in one fewer atomic op in the fastpath: mntget is now just a
per-CPU inc, rather than an atomic inc; and mntput just requires a spinlock
and non-atomic decrement in the common case. However code is otherwise bigger
and heavier, so single threaded performance is basically a wash.
Signed-off-by: Nick Piggin <npiggin@kernel.dk>
2011-01-07 07:50:11 +01:00
|
|
|
#endif
|
|
|
|
|
2014-03-21 02:10:51 +01:00
|
|
|
INIT_HLIST_NODE(&mnt->mnt_hash);
|
2011-11-25 08:35:16 +01:00
|
|
|
INIT_LIST_HEAD(&mnt->mnt_child);
|
|
|
|
INIT_LIST_HEAD(&mnt->mnt_mounts);
|
|
|
|
INIT_LIST_HEAD(&mnt->mnt_list);
|
|
|
|
INIT_LIST_HEAD(&mnt->mnt_expire);
|
|
|
|
INIT_LIST_HEAD(&mnt->mnt_share);
|
|
|
|
INIT_LIST_HEAD(&mnt->mnt_slave_list);
|
|
|
|
INIT_LIST_HEAD(&mnt->mnt_slave);
|
2013-09-23 04:37:01 +02:00
|
|
|
INIT_HLIST_NODE(&mnt->mnt_mp_list);
|
2016-10-24 23:16:13 +02:00
|
|
|
INIT_LIST_HEAD(&mnt->mnt_umounting);
|
2015-01-11 01:01:08 +01:00
|
|
|
init_fs_pin(&mnt->mnt_umount, drop_mountpoint);
|
2005-04-17 00:20:36 +02:00
|
|
|
}
|
2011-11-25 08:35:16 +01:00
|
|
|
return mnt;
|
2008-07-21 12:06:36 +02:00
|
|
|
|
2009-04-26 12:25:54 +02:00
|
|
|
#ifdef CONFIG_SMP
|
|
|
|
out_free_devname:
|
2015-02-13 23:36:41 +01:00
|
|
|
kfree_const(mnt->mnt_devname);
|
2009-04-26 12:25:54 +02:00
|
|
|
#endif
|
2008-07-21 12:06:36 +02:00
|
|
|
out_free_id:
|
2011-11-25 08:35:16 +01:00
|
|
|
mnt_free_id(mnt);
|
2008-07-21 12:06:36 +02:00
|
|
|
out_free_cache:
|
2011-11-25 08:35:16 +01:00
|
|
|
kmem_cache_free(mnt_cache, mnt);
|
2008-07-21 12:06:36 +02:00
|
|
|
return NULL;
|
2005-04-17 00:20:36 +02:00
|
|
|
}
|
|
|
|
|
[PATCH] r/o bind mounts: track numbers of writers to mounts
This is the real meat of the entire series. It actually
implements the tracking of the number of writers to a mount.
However, it causes scalability problems because there can be
hundreds of cpus doing open()/close() on files on the same mnt at
the same time. Even an atomic_t in the mnt has massive scalaing
problems because the cacheline gets so terribly contended.
This uses a statically-allocated percpu variable. All want/drop
operations are local to a cpu as long that cpu operates on the same
mount, and there are no writer count imbalances. Writer count
imbalances happen when a write is taken on one cpu, and released
on another, like when an open/close pair is performed on two
Upon a remount,ro request, all of the data from the percpu
variables is collected (expensive, but very rare) and we determine
if there are any outstanding writers to the mount.
I've written a little benchmark to sit in a loop for a couple of
seconds in several cpus in parallel doing open/write/close loops.
http://sr71.net/~dave/linux/openbench.c
The code in here is a a worst-possible case for this patch. It
does opens on a _pair_ of files in two different mounts in parallel.
This should cause my code to lose its "operate on the same mount"
optimization completely. This worst-case scenario causes a 3%
degredation in the benchmark.
I could probably get rid of even this 3%, but it would be more
complex than what I have here, and I think this is getting into
acceptable territory. In practice, I expect writing more than 3
bytes to a file, as well as disk I/O to mask any effects that this
has.
(To get rid of that 3%, we could have an #defined number of mounts
in the percpu variable. So, instead of a CPU getting operate only
on percpu data when it accesses only one mount, it could stay on
percpu data when it only accesses N or fewer mounts.)
[AV] merged fix for __clear_mnt_mount() stepping on freed vfsmount
Acked-by: Al Viro <viro@ZenIV.linux.org.uk>
Signed-off-by: Christoph Hellwig <hch@infradead.org>
Signed-off-by: Dave Hansen <haveblue@us.ibm.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Al Viro <viro@zeniv.linux.org.uk>
2008-02-15 23:37:59 +01:00
|
|
|
/*
|
|
|
|
* Most r/o checks on a fs are for operations that take
|
|
|
|
* discrete amounts of time, like a write() or unlink().
|
|
|
|
* We must keep track of when those operations start
|
|
|
|
* (for permission checks) and when they end, so that
|
|
|
|
* we can determine when writes are able to occur to
|
|
|
|
* a filesystem.
|
|
|
|
*/
|
|
|
|
/*
|
|
|
|
* __mnt_is_readonly: check whether a mount is read-only
|
|
|
|
* @mnt: the mount to check for its write status
|
|
|
|
*
|
|
|
|
* This shouldn't be used directly ouside of the VFS.
|
|
|
|
* It does not guarantee that the filesystem will stay
|
|
|
|
* r/w, just that it is right *now*. This can not and
|
|
|
|
* should not be used in place of IS_RDONLY(inode).
|
|
|
|
* mnt_want/drop_write() will _keep_ the filesystem
|
|
|
|
* r/w.
|
|
|
|
*/
|
|
|
|
int __mnt_is_readonly(struct vfsmount *mnt)
|
|
|
|
{
|
2008-02-15 23:38:00 +01:00
|
|
|
if (mnt->mnt_flags & MNT_READONLY)
|
|
|
|
return 1;
|
2017-07-17 09:45:34 +02:00
|
|
|
if (sb_rdonly(mnt->mnt_sb))
|
2008-02-15 23:38:00 +01:00
|
|
|
return 1;
|
|
|
|
return 0;
|
[PATCH] r/o bind mounts: track numbers of writers to mounts
This is the real meat of the entire series. It actually
implements the tracking of the number of writers to a mount.
However, it causes scalability problems because there can be
hundreds of cpus doing open()/close() on files on the same mnt at
the same time. Even an atomic_t in the mnt has massive scalaing
problems because the cacheline gets so terribly contended.
This uses a statically-allocated percpu variable. All want/drop
operations are local to a cpu as long that cpu operates on the same
mount, and there are no writer count imbalances. Writer count
imbalances happen when a write is taken on one cpu, and released
on another, like when an open/close pair is performed on two
Upon a remount,ro request, all of the data from the percpu
variables is collected (expensive, but very rare) and we determine
if there are any outstanding writers to the mount.
I've written a little benchmark to sit in a loop for a couple of
seconds in several cpus in parallel doing open/write/close loops.
http://sr71.net/~dave/linux/openbench.c
The code in here is a a worst-possible case for this patch. It
does opens on a _pair_ of files in two different mounts in parallel.
This should cause my code to lose its "operate on the same mount"
optimization completely. This worst-case scenario causes a 3%
degredation in the benchmark.
I could probably get rid of even this 3%, but it would be more
complex than what I have here, and I think this is getting into
acceptable territory. In practice, I expect writing more than 3
bytes to a file, as well as disk I/O to mask any effects that this
has.
(To get rid of that 3%, we could have an #defined number of mounts
in the percpu variable. So, instead of a CPU getting operate only
on percpu data when it accesses only one mount, it could stay on
percpu data when it only accesses N or fewer mounts.)
[AV] merged fix for __clear_mnt_mount() stepping on freed vfsmount
Acked-by: Al Viro <viro@ZenIV.linux.org.uk>
Signed-off-by: Christoph Hellwig <hch@infradead.org>
Signed-off-by: Dave Hansen <haveblue@us.ibm.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Al Viro <viro@zeniv.linux.org.uk>
2008-02-15 23:37:59 +01:00
|
|
|
}
|
|
|
|
EXPORT_SYMBOL_GPL(__mnt_is_readonly);
|
|
|
|
|
2011-11-25 04:37:54 +01:00
|
|
|
static inline void mnt_inc_writers(struct mount *mnt)
|
2009-04-26 12:25:54 +02:00
|
|
|
{
|
|
|
|
#ifdef CONFIG_SMP
|
2011-11-25 04:53:09 +01:00
|
|
|
this_cpu_inc(mnt->mnt_pcp->mnt_writers);
|
2009-04-26 12:25:54 +02:00
|
|
|
#else
|
2011-11-25 04:53:09 +01:00
|
|
|
mnt->mnt_writers++;
|
2009-04-26 12:25:54 +02:00
|
|
|
#endif
|
|
|
|
}
|
[PATCH] r/o bind mounts: track numbers of writers to mounts
This is the real meat of the entire series. It actually
implements the tracking of the number of writers to a mount.
However, it causes scalability problems because there can be
hundreds of cpus doing open()/close() on files on the same mnt at
the same time. Even an atomic_t in the mnt has massive scalaing
problems because the cacheline gets so terribly contended.
This uses a statically-allocated percpu variable. All want/drop
operations are local to a cpu as long that cpu operates on the same
mount, and there are no writer count imbalances. Writer count
imbalances happen when a write is taken on one cpu, and released
on another, like when an open/close pair is performed on two
Upon a remount,ro request, all of the data from the percpu
variables is collected (expensive, but very rare) and we determine
if there are any outstanding writers to the mount.
I've written a little benchmark to sit in a loop for a couple of
seconds in several cpus in parallel doing open/write/close loops.
http://sr71.net/~dave/linux/openbench.c
The code in here is a a worst-possible case for this patch. It
does opens on a _pair_ of files in two different mounts in parallel.
This should cause my code to lose its "operate on the same mount"
optimization completely. This worst-case scenario causes a 3%
degredation in the benchmark.
I could probably get rid of even this 3%, but it would be more
complex than what I have here, and I think this is getting into
acceptable territory. In practice, I expect writing more than 3
bytes to a file, as well as disk I/O to mask any effects that this
has.
(To get rid of that 3%, we could have an #defined number of mounts
in the percpu variable. So, instead of a CPU getting operate only
on percpu data when it accesses only one mount, it could stay on
percpu data when it only accesses N or fewer mounts.)
[AV] merged fix for __clear_mnt_mount() stepping on freed vfsmount
Acked-by: Al Viro <viro@ZenIV.linux.org.uk>
Signed-off-by: Christoph Hellwig <hch@infradead.org>
Signed-off-by: Dave Hansen <haveblue@us.ibm.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Al Viro <viro@zeniv.linux.org.uk>
2008-02-15 23:37:59 +01:00
|
|
|
|
2011-11-25 04:37:54 +01:00
|
|
|
static inline void mnt_dec_writers(struct mount *mnt)
|
[PATCH] r/o bind mounts: track numbers of writers to mounts
This is the real meat of the entire series. It actually
implements the tracking of the number of writers to a mount.
However, it causes scalability problems because there can be
hundreds of cpus doing open()/close() on files on the same mnt at
the same time. Even an atomic_t in the mnt has massive scalaing
problems because the cacheline gets so terribly contended.
This uses a statically-allocated percpu variable. All want/drop
operations are local to a cpu as long that cpu operates on the same
mount, and there are no writer count imbalances. Writer count
imbalances happen when a write is taken on one cpu, and released
on another, like when an open/close pair is performed on two
Upon a remount,ro request, all of the data from the percpu
variables is collected (expensive, but very rare) and we determine
if there are any outstanding writers to the mount.
I've written a little benchmark to sit in a loop for a couple of
seconds in several cpus in parallel doing open/write/close loops.
http://sr71.net/~dave/linux/openbench.c
The code in here is a a worst-possible case for this patch. It
does opens on a _pair_ of files in two different mounts in parallel.
This should cause my code to lose its "operate on the same mount"
optimization completely. This worst-case scenario causes a 3%
degredation in the benchmark.
I could probably get rid of even this 3%, but it would be more
complex than what I have here, and I think this is getting into
acceptable territory. In practice, I expect writing more than 3
bytes to a file, as well as disk I/O to mask any effects that this
has.
(To get rid of that 3%, we could have an #defined number of mounts
in the percpu variable. So, instead of a CPU getting operate only
on percpu data when it accesses only one mount, it could stay on
percpu data when it only accesses N or fewer mounts.)
[AV] merged fix for __clear_mnt_mount() stepping on freed vfsmount
Acked-by: Al Viro <viro@ZenIV.linux.org.uk>
Signed-off-by: Christoph Hellwig <hch@infradead.org>
Signed-off-by: Dave Hansen <haveblue@us.ibm.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Al Viro <viro@zeniv.linux.org.uk>
2008-02-15 23:37:59 +01:00
|
|
|
{
|
2009-04-26 12:25:54 +02:00
|
|
|
#ifdef CONFIG_SMP
|
2011-11-25 04:53:09 +01:00
|
|
|
this_cpu_dec(mnt->mnt_pcp->mnt_writers);
|
2009-04-26 12:25:54 +02:00
|
|
|
#else
|
2011-11-25 04:53:09 +01:00
|
|
|
mnt->mnt_writers--;
|
2009-04-26 12:25:54 +02:00
|
|
|
#endif
|
[PATCH] r/o bind mounts: track numbers of writers to mounts
This is the real meat of the entire series. It actually
implements the tracking of the number of writers to a mount.
However, it causes scalability problems because there can be
hundreds of cpus doing open()/close() on files on the same mnt at
the same time. Even an atomic_t in the mnt has massive scalaing
problems because the cacheline gets so terribly contended.
This uses a statically-allocated percpu variable. All want/drop
operations are local to a cpu as long that cpu operates on the same
mount, and there are no writer count imbalances. Writer count
imbalances happen when a write is taken on one cpu, and released
on another, like when an open/close pair is performed on two
Upon a remount,ro request, all of the data from the percpu
variables is collected (expensive, but very rare) and we determine
if there are any outstanding writers to the mount.
I've written a little benchmark to sit in a loop for a couple of
seconds in several cpus in parallel doing open/write/close loops.
http://sr71.net/~dave/linux/openbench.c
The code in here is a a worst-possible case for this patch. It
does opens on a _pair_ of files in two different mounts in parallel.
This should cause my code to lose its "operate on the same mount"
optimization completely. This worst-case scenario causes a 3%
degredation in the benchmark.
I could probably get rid of even this 3%, but it would be more
complex than what I have here, and I think this is getting into
acceptable territory. In practice, I expect writing more than 3
bytes to a file, as well as disk I/O to mask any effects that this
has.
(To get rid of that 3%, we could have an #defined number of mounts
in the percpu variable. So, instead of a CPU getting operate only
on percpu data when it accesses only one mount, it could stay on
percpu data when it only accesses N or fewer mounts.)
[AV] merged fix for __clear_mnt_mount() stepping on freed vfsmount
Acked-by: Al Viro <viro@ZenIV.linux.org.uk>
Signed-off-by: Christoph Hellwig <hch@infradead.org>
Signed-off-by: Dave Hansen <haveblue@us.ibm.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Al Viro <viro@zeniv.linux.org.uk>
2008-02-15 23:37:59 +01:00
|
|
|
}
|
|
|
|
|
2011-11-25 04:37:54 +01:00
|
|
|
static unsigned int mnt_get_writers(struct mount *mnt)
|
[PATCH] r/o bind mounts: track numbers of writers to mounts
This is the real meat of the entire series. It actually
implements the tracking of the number of writers to a mount.
However, it causes scalability problems because there can be
hundreds of cpus doing open()/close() on files on the same mnt at
the same time. Even an atomic_t in the mnt has massive scalaing
problems because the cacheline gets so terribly contended.
This uses a statically-allocated percpu variable. All want/drop
operations are local to a cpu as long that cpu operates on the same
mount, and there are no writer count imbalances. Writer count
imbalances happen when a write is taken on one cpu, and released
on another, like when an open/close pair is performed on two
Upon a remount,ro request, all of the data from the percpu
variables is collected (expensive, but very rare) and we determine
if there are any outstanding writers to the mount.
I've written a little benchmark to sit in a loop for a couple of
seconds in several cpus in parallel doing open/write/close loops.
http://sr71.net/~dave/linux/openbench.c
The code in here is a a worst-possible case for this patch. It
does opens on a _pair_ of files in two different mounts in parallel.
This should cause my code to lose its "operate on the same mount"
optimization completely. This worst-case scenario causes a 3%
degredation in the benchmark.
I could probably get rid of even this 3%, but it would be more
complex than what I have here, and I think this is getting into
acceptable territory. In practice, I expect writing more than 3
bytes to a file, as well as disk I/O to mask any effects that this
has.
(To get rid of that 3%, we could have an #defined number of mounts
in the percpu variable. So, instead of a CPU getting operate only
on percpu data when it accesses only one mount, it could stay on
percpu data when it only accesses N or fewer mounts.)
[AV] merged fix for __clear_mnt_mount() stepping on freed vfsmount
Acked-by: Al Viro <viro@ZenIV.linux.org.uk>
Signed-off-by: Christoph Hellwig <hch@infradead.org>
Signed-off-by: Dave Hansen <haveblue@us.ibm.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Al Viro <viro@zeniv.linux.org.uk>
2008-02-15 23:37:59 +01:00
|
|
|
{
|
2009-04-26 12:25:54 +02:00
|
|
|
#ifdef CONFIG_SMP
|
|
|
|
unsigned int count = 0;
|
[PATCH] r/o bind mounts: track numbers of writers to mounts
This is the real meat of the entire series. It actually
implements the tracking of the number of writers to a mount.
However, it causes scalability problems because there can be
hundreds of cpus doing open()/close() on files on the same mnt at
the same time. Even an atomic_t in the mnt has massive scalaing
problems because the cacheline gets so terribly contended.
This uses a statically-allocated percpu variable. All want/drop
operations are local to a cpu as long that cpu operates on the same
mount, and there are no writer count imbalances. Writer count
imbalances happen when a write is taken on one cpu, and released
on another, like when an open/close pair is performed on two
Upon a remount,ro request, all of the data from the percpu
variables is collected (expensive, but very rare) and we determine
if there are any outstanding writers to the mount.
I've written a little benchmark to sit in a loop for a couple of
seconds in several cpus in parallel doing open/write/close loops.
http://sr71.net/~dave/linux/openbench.c
The code in here is a a worst-possible case for this patch. It
does opens on a _pair_ of files in two different mounts in parallel.
This should cause my code to lose its "operate on the same mount"
optimization completely. This worst-case scenario causes a 3%
degredation in the benchmark.
I could probably get rid of even this 3%, but it would be more
complex than what I have here, and I think this is getting into
acceptable territory. In practice, I expect writing more than 3
bytes to a file, as well as disk I/O to mask any effects that this
has.
(To get rid of that 3%, we could have an #defined number of mounts
in the percpu variable. So, instead of a CPU getting operate only
on percpu data when it accesses only one mount, it could stay on
percpu data when it only accesses N or fewer mounts.)
[AV] merged fix for __clear_mnt_mount() stepping on freed vfsmount
Acked-by: Al Viro <viro@ZenIV.linux.org.uk>
Signed-off-by: Christoph Hellwig <hch@infradead.org>
Signed-off-by: Dave Hansen <haveblue@us.ibm.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Al Viro <viro@zeniv.linux.org.uk>
2008-02-15 23:37:59 +01:00
|
|
|
int cpu;
|
|
|
|
|
|
|
|
for_each_possible_cpu(cpu) {
|
2011-11-25 04:53:09 +01:00
|
|
|
count += per_cpu_ptr(mnt->mnt_pcp, cpu)->mnt_writers;
|
[PATCH] r/o bind mounts: track numbers of writers to mounts
This is the real meat of the entire series. It actually
implements the tracking of the number of writers to a mount.
However, it causes scalability problems because there can be
hundreds of cpus doing open()/close() on files on the same mnt at
the same time. Even an atomic_t in the mnt has massive scalaing
problems because the cacheline gets so terribly contended.
This uses a statically-allocated percpu variable. All want/drop
operations are local to a cpu as long that cpu operates on the same
mount, and there are no writer count imbalances. Writer count
imbalances happen when a write is taken on one cpu, and released
on another, like when an open/close pair is performed on two
Upon a remount,ro request, all of the data from the percpu
variables is collected (expensive, but very rare) and we determine
if there are any outstanding writers to the mount.
I've written a little benchmark to sit in a loop for a couple of
seconds in several cpus in parallel doing open/write/close loops.
http://sr71.net/~dave/linux/openbench.c
The code in here is a a worst-possible case for this patch. It
does opens on a _pair_ of files in two different mounts in parallel.
This should cause my code to lose its "operate on the same mount"
optimization completely. This worst-case scenario causes a 3%
degredation in the benchmark.
I could probably get rid of even this 3%, but it would be more
complex than what I have here, and I think this is getting into
acceptable territory. In practice, I expect writing more than 3
bytes to a file, as well as disk I/O to mask any effects that this
has.
(To get rid of that 3%, we could have an #defined number of mounts
in the percpu variable. So, instead of a CPU getting operate only
on percpu data when it accesses only one mount, it could stay on
percpu data when it only accesses N or fewer mounts.)
[AV] merged fix for __clear_mnt_mount() stepping on freed vfsmount
Acked-by: Al Viro <viro@ZenIV.linux.org.uk>
Signed-off-by: Christoph Hellwig <hch@infradead.org>
Signed-off-by: Dave Hansen <haveblue@us.ibm.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Al Viro <viro@zeniv.linux.org.uk>
2008-02-15 23:37:59 +01:00
|
|
|
}
|
|
|
|
|
2009-04-26 12:25:54 +02:00
|
|
|
return count;
|
|
|
|
#else
|
|
|
|
return mnt->mnt_writers;
|
|
|
|
#endif
|
[PATCH] r/o bind mounts: track numbers of writers to mounts
This is the real meat of the entire series. It actually
implements the tracking of the number of writers to a mount.
However, it causes scalability problems because there can be
hundreds of cpus doing open()/close() on files on the same mnt at
the same time. Even an atomic_t in the mnt has massive scalaing
problems because the cacheline gets so terribly contended.
This uses a statically-allocated percpu variable. All want/drop
operations are local to a cpu as long that cpu operates on the same
mount, and there are no writer count imbalances. Writer count
imbalances happen when a write is taken on one cpu, and released
on another, like when an open/close pair is performed on two
Upon a remount,ro request, all of the data from the percpu
variables is collected (expensive, but very rare) and we determine
if there are any outstanding writers to the mount.
I've written a little benchmark to sit in a loop for a couple of
seconds in several cpus in parallel doing open/write/close loops.
http://sr71.net/~dave/linux/openbench.c
The code in here is a a worst-possible case for this patch. It
does opens on a _pair_ of files in two different mounts in parallel.
This should cause my code to lose its "operate on the same mount"
optimization completely. This worst-case scenario causes a 3%
degredation in the benchmark.
I could probably get rid of even this 3%, but it would be more
complex than what I have here, and I think this is getting into
acceptable territory. In practice, I expect writing more than 3
bytes to a file, as well as disk I/O to mask any effects that this
has.
(To get rid of that 3%, we could have an #defined number of mounts
in the percpu variable. So, instead of a CPU getting operate only
on percpu data when it accesses only one mount, it could stay on
percpu data when it only accesses N or fewer mounts.)
[AV] merged fix for __clear_mnt_mount() stepping on freed vfsmount
Acked-by: Al Viro <viro@ZenIV.linux.org.uk>
Signed-off-by: Christoph Hellwig <hch@infradead.org>
Signed-off-by: Dave Hansen <haveblue@us.ibm.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Al Viro <viro@zeniv.linux.org.uk>
2008-02-15 23:37:59 +01:00
|
|
|
}
|
|
|
|
|
2011-11-21 12:11:31 +01:00
|
|
|
static int mnt_is_readonly(struct vfsmount *mnt)
|
|
|
|
{
|
|
|
|
if (mnt->mnt_sb->s_readonly_remount)
|
|
|
|
return 1;
|
|
|
|
/* Order wrt setting s_flags/s_readonly_remount in do_remount() */
|
|
|
|
smp_rmb();
|
|
|
|
return __mnt_is_readonly(mnt);
|
|
|
|
}
|
|
|
|
|
2008-02-15 23:37:30 +01:00
|
|
|
/*
|
2012-06-12 16:20:35 +02:00
|
|
|
* Most r/o & frozen checks on a fs are for operations that take discrete
|
|
|
|
* amounts of time, like a write() or unlink(). We must keep track of when
|
|
|
|
* those operations start (for permission checks) and when they end, so that we
|
|
|
|
* can determine when writes are able to occur to a filesystem.
|
2008-02-15 23:37:30 +01:00
|
|
|
*/
|
|
|
|
/**
|
2012-06-12 16:20:35 +02:00
|
|
|
* __mnt_want_write - get write access to a mount without freeze protection
|
2011-11-25 04:37:54 +01:00
|
|
|
* @m: the mount on which to take a write
|
2008-02-15 23:37:30 +01:00
|
|
|
*
|
2012-06-12 16:20:35 +02:00
|
|
|
* This tells the low-level filesystem that a write is about to be performed to
|
|
|
|
* it, and makes sure that writes are allowed (mnt it read-write) before
|
|
|
|
* returning success. This operation does not protect against filesystem being
|
|
|
|
* frozen. When the write operation is finished, __mnt_drop_write() must be
|
|
|
|
* called. This is effectively a refcount.
|
2008-02-15 23:37:30 +01:00
|
|
|
*/
|
2012-06-12 16:20:35 +02:00
|
|
|
int __mnt_want_write(struct vfsmount *m)
|
2008-02-15 23:37:30 +01:00
|
|
|
{
|
2011-11-25 04:37:54 +01:00
|
|
|
struct mount *mnt = real_mount(m);
|
[PATCH] r/o bind mounts: track numbers of writers to mounts
This is the real meat of the entire series. It actually
implements the tracking of the number of writers to a mount.
However, it causes scalability problems because there can be
hundreds of cpus doing open()/close() on files on the same mnt at
the same time. Even an atomic_t in the mnt has massive scalaing
problems because the cacheline gets so terribly contended.
This uses a statically-allocated percpu variable. All want/drop
operations are local to a cpu as long that cpu operates on the same
mount, and there are no writer count imbalances. Writer count
imbalances happen when a write is taken on one cpu, and released
on another, like when an open/close pair is performed on two
Upon a remount,ro request, all of the data from the percpu
variables is collected (expensive, but very rare) and we determine
if there are any outstanding writers to the mount.
I've written a little benchmark to sit in a loop for a couple of
seconds in several cpus in parallel doing open/write/close loops.
http://sr71.net/~dave/linux/openbench.c
The code in here is a a worst-possible case for this patch. It
does opens on a _pair_ of files in two different mounts in parallel.
This should cause my code to lose its "operate on the same mount"
optimization completely. This worst-case scenario causes a 3%
degredation in the benchmark.
I could probably get rid of even this 3%, but it would be more
complex than what I have here, and I think this is getting into
acceptable territory. In practice, I expect writing more than 3
bytes to a file, as well as disk I/O to mask any effects that this
has.
(To get rid of that 3%, we could have an #defined number of mounts
in the percpu variable. So, instead of a CPU getting operate only
on percpu data when it accesses only one mount, it could stay on
percpu data when it only accesses N or fewer mounts.)
[AV] merged fix for __clear_mnt_mount() stepping on freed vfsmount
Acked-by: Al Viro <viro@ZenIV.linux.org.uk>
Signed-off-by: Christoph Hellwig <hch@infradead.org>
Signed-off-by: Dave Hansen <haveblue@us.ibm.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Al Viro <viro@zeniv.linux.org.uk>
2008-02-15 23:37:59 +01:00
|
|
|
int ret = 0;
|
|
|
|
|
2009-04-26 12:25:54 +02:00
|
|
|
preempt_disable();
|
2011-01-07 07:50:10 +01:00
|
|
|
mnt_inc_writers(mnt);
|
2009-04-26 12:25:54 +02:00
|
|
|
/*
|
2011-01-07 07:50:10 +01:00
|
|
|
* The store to mnt_inc_writers must be visible before we pass
|
2009-04-26 12:25:54 +02:00
|
|
|
* MNT_WRITE_HOLD loop below, so that the slowpath can see our
|
|
|
|
* incremented count after it has set MNT_WRITE_HOLD.
|
|
|
|
*/
|
|
|
|
smp_mb();
|
2012-11-16 10:23:50 +01:00
|
|
|
while (ACCESS_ONCE(mnt->mnt.mnt_flags) & MNT_WRITE_HOLD)
|
2009-04-26 12:25:54 +02:00
|
|
|
cpu_relax();
|
|
|
|
/*
|
|
|
|
* After the slowpath clears MNT_WRITE_HOLD, mnt_is_readonly will
|
|
|
|
* be set to match its requirements. So we must not load that until
|
|
|
|
* MNT_WRITE_HOLD is cleared.
|
|
|
|
*/
|
|
|
|
smp_rmb();
|
2011-11-21 12:11:31 +01:00
|
|
|
if (mnt_is_readonly(m)) {
|
2011-01-07 07:50:10 +01:00
|
|
|
mnt_dec_writers(mnt);
|
[PATCH] r/o bind mounts: track numbers of writers to mounts
This is the real meat of the entire series. It actually
implements the tracking of the number of writers to a mount.
However, it causes scalability problems because there can be
hundreds of cpus doing open()/close() on files on the same mnt at
the same time. Even an atomic_t in the mnt has massive scalaing
problems because the cacheline gets so terribly contended.
This uses a statically-allocated percpu variable. All want/drop
operations are local to a cpu as long that cpu operates on the same
mount, and there are no writer count imbalances. Writer count
imbalances happen when a write is taken on one cpu, and released
on another, like when an open/close pair is performed on two
Upon a remount,ro request, all of the data from the percpu
variables is collected (expensive, but very rare) and we determine
if there are any outstanding writers to the mount.
I've written a little benchmark to sit in a loop for a couple of
seconds in several cpus in parallel doing open/write/close loops.
http://sr71.net/~dave/linux/openbench.c
The code in here is a a worst-possible case for this patch. It
does opens on a _pair_ of files in two different mounts in parallel.
This should cause my code to lose its "operate on the same mount"
optimization completely. This worst-case scenario causes a 3%
degredation in the benchmark.
I could probably get rid of even this 3%, but it would be more
complex than what I have here, and I think this is getting into
acceptable territory. In practice, I expect writing more than 3
bytes to a file, as well as disk I/O to mask any effects that this
has.
(To get rid of that 3%, we could have an #defined number of mounts
in the percpu variable. So, instead of a CPU getting operate only
on percpu data when it accesses only one mount, it could stay on
percpu data when it only accesses N or fewer mounts.)
[AV] merged fix for __clear_mnt_mount() stepping on freed vfsmount
Acked-by: Al Viro <viro@ZenIV.linux.org.uk>
Signed-off-by: Christoph Hellwig <hch@infradead.org>
Signed-off-by: Dave Hansen <haveblue@us.ibm.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Al Viro <viro@zeniv.linux.org.uk>
2008-02-15 23:37:59 +01:00
|
|
|
ret = -EROFS;
|
|
|
|
}
|
2009-04-26 12:25:54 +02:00
|
|
|
preempt_enable();
|
2012-06-12 16:20:35 +02:00
|
|
|
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
|
|
|
* mnt_want_write - get write access to a mount
|
|
|
|
* @m: the mount on which to take a write
|
|
|
|
*
|
|
|
|
* This tells the low-level filesystem that a write is about to be performed to
|
|
|
|
* it, and makes sure that writes are allowed (mount is read-write, filesystem
|
|
|
|
* is not frozen) before returning success. When the write operation is
|
|
|
|
* finished, mnt_drop_write() must be called. This is effectively a refcount.
|
|
|
|
*/
|
|
|
|
int mnt_want_write(struct vfsmount *m)
|
|
|
|
{
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
sb_start_write(m->mnt_sb);
|
|
|
|
ret = __mnt_want_write(m);
|
|
|
|
if (ret)
|
|
|
|
sb_end_write(m->mnt_sb);
|
[PATCH] r/o bind mounts: track numbers of writers to mounts
This is the real meat of the entire series. It actually
implements the tracking of the number of writers to a mount.
However, it causes scalability problems because there can be
hundreds of cpus doing open()/close() on files on the same mnt at
the same time. Even an atomic_t in the mnt has massive scalaing
problems because the cacheline gets so terribly contended.
This uses a statically-allocated percpu variable. All want/drop
operations are local to a cpu as long that cpu operates on the same
mount, and there are no writer count imbalances. Writer count
imbalances happen when a write is taken on one cpu, and released
on another, like when an open/close pair is performed on two
Upon a remount,ro request, all of the data from the percpu
variables is collected (expensive, but very rare) and we determine
if there are any outstanding writers to the mount.
I've written a little benchmark to sit in a loop for a couple of
seconds in several cpus in parallel doing open/write/close loops.
http://sr71.net/~dave/linux/openbench.c
The code in here is a a worst-possible case for this patch. It
does opens on a _pair_ of files in two different mounts in parallel.
This should cause my code to lose its "operate on the same mount"
optimization completely. This worst-case scenario causes a 3%
degredation in the benchmark.
I could probably get rid of even this 3%, but it would be more
complex than what I have here, and I think this is getting into
acceptable territory. In practice, I expect writing more than 3
bytes to a file, as well as disk I/O to mask any effects that this
has.
(To get rid of that 3%, we could have an #defined number of mounts
in the percpu variable. So, instead of a CPU getting operate only
on percpu data when it accesses only one mount, it could stay on
percpu data when it only accesses N or fewer mounts.)
[AV] merged fix for __clear_mnt_mount() stepping on freed vfsmount
Acked-by: Al Viro <viro@ZenIV.linux.org.uk>
Signed-off-by: Christoph Hellwig <hch@infradead.org>
Signed-off-by: Dave Hansen <haveblue@us.ibm.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Al Viro <viro@zeniv.linux.org.uk>
2008-02-15 23:37:59 +01:00
|
|
|
return ret;
|
2008-02-15 23:37:30 +01:00
|
|
|
}
|
|
|
|
EXPORT_SYMBOL_GPL(mnt_want_write);
|
|
|
|
|
2009-04-26 12:25:55 +02:00
|
|
|
/**
|
|
|
|
* mnt_clone_write - get write access to a mount
|
|
|
|
* @mnt: the mount on which to take a write
|
|
|
|
*
|
|
|
|
* This is effectively like mnt_want_write, except
|
|
|
|
* it must only be used to take an extra write reference
|
|
|
|
* on a mountpoint that we already know has a write reference
|
|
|
|
* on it. This allows some optimisation.
|
|
|
|
*
|
|
|
|
* After finished, mnt_drop_write must be called as usual to
|
|
|
|
* drop the reference.
|
|
|
|
*/
|
|
|
|
int mnt_clone_write(struct vfsmount *mnt)
|
|
|
|
{
|
|
|
|
/* superblock may be r/o */
|
|
|
|
if (__mnt_is_readonly(mnt))
|
|
|
|
return -EROFS;
|
|
|
|
preempt_disable();
|
2011-11-25 04:37:54 +01:00
|
|
|
mnt_inc_writers(real_mount(mnt));
|
2009-04-26 12:25:55 +02:00
|
|
|
preempt_enable();
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL_GPL(mnt_clone_write);
|
|
|
|
|
|
|
|
/**
|
2012-06-12 16:20:35 +02:00
|
|
|
* __mnt_want_write_file - get write access to a file's mount
|
2009-04-26 12:25:55 +02:00
|
|
|
* @file: the file who's mount on which to take a write
|
|
|
|
*
|
2012-06-12 16:20:35 +02:00
|
|
|
* This is like __mnt_want_write, but it takes a file and can
|
2009-04-26 12:25:55 +02:00
|
|
|
* do some optimisations if the file is open for write already
|
|
|
|
*/
|
2012-06-12 16:20:35 +02:00
|
|
|
int __mnt_want_write_file(struct file *file)
|
2009-04-26 12:25:55 +02:00
|
|
|
{
|
2014-03-14 17:02:47 +01:00
|
|
|
if (!(file->f_mode & FMODE_WRITER))
|
2012-06-12 16:20:35 +02:00
|
|
|
return __mnt_want_write(file->f_path.mnt);
|
2009-04-26 12:25:55 +02:00
|
|
|
else
|
|
|
|
return mnt_clone_write(file->f_path.mnt);
|
|
|
|
}
|
2012-06-12 16:20:35 +02:00
|
|
|
|
|
|
|
/**
|
2017-09-05 12:53:12 +02:00
|
|
|
* mnt_want_write_file_path - get write access to a file's mount
|
2012-06-12 16:20:35 +02:00
|
|
|
* @file: the file who's mount on which to take a write
|
|
|
|
*
|
|
|
|
* This is like mnt_want_write, but it takes a file and can
|
|
|
|
* do some optimisations if the file is open for write already
|
2017-09-05 12:53:12 +02:00
|
|
|
*
|
|
|
|
* Called by the vfs for cases when we have an open file at hand, but will do an
|
|
|
|
* inode operation on it (important distinction for files opened on overlayfs,
|
|
|
|
* since the file operations will come from the real underlying file, while
|
|
|
|
* inode operations come from the overlay).
|
2012-06-12 16:20:35 +02:00
|
|
|
*/
|
2017-09-05 12:53:12 +02:00
|
|
|
int mnt_want_write_file_path(struct file *file)
|
2012-06-12 16:20:35 +02:00
|
|
|
{
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
sb_start_write(file->f_path.mnt->mnt_sb);
|
|
|
|
ret = __mnt_want_write_file(file);
|
|
|
|
if (ret)
|
|
|
|
sb_end_write(file->f_path.mnt->mnt_sb);
|
|
|
|
return ret;
|
|
|
|
}
|
2017-09-05 12:53:12 +02:00
|
|
|
|
|
|
|
static inline int may_write_real(struct file *file)
|
|
|
|
{
|
|
|
|
struct dentry *dentry = file->f_path.dentry;
|
|
|
|
struct dentry *upperdentry;
|
|
|
|
|
|
|
|
/* Writable file? */
|
|
|
|
if (file->f_mode & FMODE_WRITER)
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
/* Not overlayfs? */
|
|
|
|
if (likely(!(dentry->d_flags & DCACHE_OP_REAL)))
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
/* File refers to upper, writable layer? */
|
|
|
|
upperdentry = d_real(dentry, NULL, 0, D_REAL_UPPER);
|
|
|
|
if (upperdentry && file_inode(file) == d_inode(upperdentry))
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
/* Lower layer: can't write to real file, sorry... */
|
|
|
|
return -EPERM;
|
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
|
|
|
* mnt_want_write_file - get write access to a file's mount
|
|
|
|
* @file: the file who's mount on which to take a write
|
|
|
|
*
|
|
|
|
* This is like mnt_want_write, but it takes a file and can
|
|
|
|
* do some optimisations if the file is open for write already
|
|
|
|
*
|
|
|
|
* Mostly called by filesystems from their ioctl operation before performing
|
|
|
|
* modification. On overlayfs this needs to check if the file is on a read-only
|
|
|
|
* lower layer and deny access in that case.
|
|
|
|
*/
|
|
|
|
int mnt_want_write_file(struct file *file)
|
|
|
|
{
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
ret = may_write_real(file);
|
|
|
|
if (!ret) {
|
|
|
|
sb_start_write(file_inode(file)->i_sb);
|
|
|
|
ret = __mnt_want_write_file(file);
|
|
|
|
if (ret)
|
|
|
|
sb_end_write(file_inode(file)->i_sb);
|
|
|
|
}
|
|
|
|
return ret;
|
|
|
|
}
|
2009-04-26 12:25:55 +02:00
|
|
|
EXPORT_SYMBOL_GPL(mnt_want_write_file);
|
|
|
|
|
2008-02-15 23:37:30 +01:00
|
|
|
/**
|
2012-06-12 16:20:35 +02:00
|
|
|
* __mnt_drop_write - give up write access to a mount
|
2008-02-15 23:37:30 +01:00
|
|
|
* @mnt: the mount on which to give up write access
|
|
|
|
*
|
|
|
|
* Tells the low-level filesystem that we are done
|
|
|
|
* performing writes to it. Must be matched with
|
2012-06-12 16:20:35 +02:00
|
|
|
* __mnt_want_write() call above.
|
2008-02-15 23:37:30 +01:00
|
|
|
*/
|
2012-06-12 16:20:35 +02:00
|
|
|
void __mnt_drop_write(struct vfsmount *mnt)
|
2008-02-15 23:37:30 +01:00
|
|
|
{
|
2009-04-26 12:25:54 +02:00
|
|
|
preempt_disable();
|
2011-11-25 04:37:54 +01:00
|
|
|
mnt_dec_writers(real_mount(mnt));
|
2009-04-26 12:25:54 +02:00
|
|
|
preempt_enable();
|
2008-02-15 23:37:30 +01:00
|
|
|
}
|
2012-06-12 16:20:35 +02:00
|
|
|
|
|
|
|
/**
|
|
|
|
* mnt_drop_write - give up write access to a mount
|
|
|
|
* @mnt: the mount on which to give up write access
|
|
|
|
*
|
|
|
|
* Tells the low-level filesystem that we are done performing writes to it and
|
|
|
|
* also allows filesystem to be frozen again. Must be matched with
|
|
|
|
* mnt_want_write() call above.
|
|
|
|
*/
|
|
|
|
void mnt_drop_write(struct vfsmount *mnt)
|
|
|
|
{
|
|
|
|
__mnt_drop_write(mnt);
|
|
|
|
sb_end_write(mnt->mnt_sb);
|
|
|
|
}
|
2008-02-15 23:37:30 +01:00
|
|
|
EXPORT_SYMBOL_GPL(mnt_drop_write);
|
|
|
|
|
2012-06-12 16:20:35 +02:00
|
|
|
void __mnt_drop_write_file(struct file *file)
|
|
|
|
{
|
|
|
|
__mnt_drop_write(file->f_path.mnt);
|
|
|
|
}
|
|
|
|
|
2017-09-05 12:53:12 +02:00
|
|
|
void mnt_drop_write_file_path(struct file *file)
|
2011-12-09 14:06:57 +01:00
|
|
|
{
|
|
|
|
mnt_drop_write(file->f_path.mnt);
|
|
|
|
}
|
2017-09-05 12:53:12 +02:00
|
|
|
|
|
|
|
void mnt_drop_write_file(struct file *file)
|
|
|
|
{
|
|
|
|
__mnt_drop_write(file->f_path.mnt);
|
|
|
|
sb_end_write(file_inode(file)->i_sb);
|
|
|
|
}
|
2011-12-09 14:06:57 +01:00
|
|
|
EXPORT_SYMBOL(mnt_drop_write_file);
|
|
|
|
|
2011-11-25 04:37:54 +01:00
|
|
|
static int mnt_make_readonly(struct mount *mnt)
|
2008-02-15 23:37:30 +01:00
|
|
|
{
|
[PATCH] r/o bind mounts: track numbers of writers to mounts
This is the real meat of the entire series. It actually
implements the tracking of the number of writers to a mount.
However, it causes scalability problems because there can be
hundreds of cpus doing open()/close() on files on the same mnt at
the same time. Even an atomic_t in the mnt has massive scalaing
problems because the cacheline gets so terribly contended.
This uses a statically-allocated percpu variable. All want/drop
operations are local to a cpu as long that cpu operates on the same
mount, and there are no writer count imbalances. Writer count
imbalances happen when a write is taken on one cpu, and released
on another, like when an open/close pair is performed on two
Upon a remount,ro request, all of the data from the percpu
variables is collected (expensive, but very rare) and we determine
if there are any outstanding writers to the mount.
I've written a little benchmark to sit in a loop for a couple of
seconds in several cpus in parallel doing open/write/close loops.
http://sr71.net/~dave/linux/openbench.c
The code in here is a a worst-possible case for this patch. It
does opens on a _pair_ of files in two different mounts in parallel.
This should cause my code to lose its "operate on the same mount"
optimization completely. This worst-case scenario causes a 3%
degredation in the benchmark.
I could probably get rid of even this 3%, but it would be more
complex than what I have here, and I think this is getting into
acceptable territory. In practice, I expect writing more than 3
bytes to a file, as well as disk I/O to mask any effects that this
has.
(To get rid of that 3%, we could have an #defined number of mounts
in the percpu variable. So, instead of a CPU getting operate only
on percpu data when it accesses only one mount, it could stay on
percpu data when it only accesses N or fewer mounts.)
[AV] merged fix for __clear_mnt_mount() stepping on freed vfsmount
Acked-by: Al Viro <viro@ZenIV.linux.org.uk>
Signed-off-by: Christoph Hellwig <hch@infradead.org>
Signed-off-by: Dave Hansen <haveblue@us.ibm.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Al Viro <viro@zeniv.linux.org.uk>
2008-02-15 23:37:59 +01:00
|
|
|
int ret = 0;
|
|
|
|
|
2013-09-29 17:24:49 +02:00
|
|
|
lock_mount_hash();
|
2011-11-25 04:37:54 +01:00
|
|
|
mnt->mnt.mnt_flags |= MNT_WRITE_HOLD;
|
[PATCH] r/o bind mounts: track numbers of writers to mounts
This is the real meat of the entire series. It actually
implements the tracking of the number of writers to a mount.
However, it causes scalability problems because there can be
hundreds of cpus doing open()/close() on files on the same mnt at
the same time. Even an atomic_t in the mnt has massive scalaing
problems because the cacheline gets so terribly contended.
This uses a statically-allocated percpu variable. All want/drop
operations are local to a cpu as long that cpu operates on the same
mount, and there are no writer count imbalances. Writer count
imbalances happen when a write is taken on one cpu, and released
on another, like when an open/close pair is performed on two
Upon a remount,ro request, all of the data from the percpu
variables is collected (expensive, but very rare) and we determine
if there are any outstanding writers to the mount.
I've written a little benchmark to sit in a loop for a couple of
seconds in several cpus in parallel doing open/write/close loops.
http://sr71.net/~dave/linux/openbench.c
The code in here is a a worst-possible case for this patch. It
does opens on a _pair_ of files in two different mounts in parallel.
This should cause my code to lose its "operate on the same mount"
optimization completely. This worst-case scenario causes a 3%
degredation in the benchmark.
I could probably get rid of even this 3%, but it would be more
complex than what I have here, and I think this is getting into
acceptable territory. In practice, I expect writing more than 3
bytes to a file, as well as disk I/O to mask any effects that this
has.
(To get rid of that 3%, we could have an #defined number of mounts
in the percpu variable. So, instead of a CPU getting operate only
on percpu data when it accesses only one mount, it could stay on
percpu data when it only accesses N or fewer mounts.)
[AV] merged fix for __clear_mnt_mount() stepping on freed vfsmount
Acked-by: Al Viro <viro@ZenIV.linux.org.uk>
Signed-off-by: Christoph Hellwig <hch@infradead.org>
Signed-off-by: Dave Hansen <haveblue@us.ibm.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Al Viro <viro@zeniv.linux.org.uk>
2008-02-15 23:37:59 +01:00
|
|
|
/*
|
2009-04-26 12:25:54 +02:00
|
|
|
* After storing MNT_WRITE_HOLD, we'll read the counters. This store
|
|
|
|
* should be visible before we do.
|
[PATCH] r/o bind mounts: track numbers of writers to mounts
This is the real meat of the entire series. It actually
implements the tracking of the number of writers to a mount.
However, it causes scalability problems because there can be
hundreds of cpus doing open()/close() on files on the same mnt at
the same time. Even an atomic_t in the mnt has massive scalaing
problems because the cacheline gets so terribly contended.
This uses a statically-allocated percpu variable. All want/drop
operations are local to a cpu as long that cpu operates on the same
mount, and there are no writer count imbalances. Writer count
imbalances happen when a write is taken on one cpu, and released
on another, like when an open/close pair is performed on two
Upon a remount,ro request, all of the data from the percpu
variables is collected (expensive, but very rare) and we determine
if there are any outstanding writers to the mount.
I've written a little benchmark to sit in a loop for a couple of
seconds in several cpus in parallel doing open/write/close loops.
http://sr71.net/~dave/linux/openbench.c
The code in here is a a worst-possible case for this patch. It
does opens on a _pair_ of files in two different mounts in parallel.
This should cause my code to lose its "operate on the same mount"
optimization completely. This worst-case scenario causes a 3%
degredation in the benchmark.
I could probably get rid of even this 3%, but it would be more
complex than what I have here, and I think this is getting into
acceptable territory. In practice, I expect writing more than 3
bytes to a file, as well as disk I/O to mask any effects that this
has.
(To get rid of that 3%, we could have an #defined number of mounts
in the percpu variable. So, instead of a CPU getting operate only
on percpu data when it accesses only one mount, it could stay on
percpu data when it only accesses N or fewer mounts.)
[AV] merged fix for __clear_mnt_mount() stepping on freed vfsmount
Acked-by: Al Viro <viro@ZenIV.linux.org.uk>
Signed-off-by: Christoph Hellwig <hch@infradead.org>
Signed-off-by: Dave Hansen <haveblue@us.ibm.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Al Viro <viro@zeniv.linux.org.uk>
2008-02-15 23:37:59 +01:00
|
|
|
*/
|
2009-04-26 12:25:54 +02:00
|
|
|
smp_mb();
|
|
|
|
|
[PATCH] r/o bind mounts: track numbers of writers to mounts
This is the real meat of the entire series. It actually
implements the tracking of the number of writers to a mount.
However, it causes scalability problems because there can be
hundreds of cpus doing open()/close() on files on the same mnt at
the same time. Even an atomic_t in the mnt has massive scalaing
problems because the cacheline gets so terribly contended.
This uses a statically-allocated percpu variable. All want/drop
operations are local to a cpu as long that cpu operates on the same
mount, and there are no writer count imbalances. Writer count
imbalances happen when a write is taken on one cpu, and released
on another, like when an open/close pair is performed on two
Upon a remount,ro request, all of the data from the percpu
variables is collected (expensive, but very rare) and we determine
if there are any outstanding writers to the mount.
I've written a little benchmark to sit in a loop for a couple of
seconds in several cpus in parallel doing open/write/close loops.
http://sr71.net/~dave/linux/openbench.c
The code in here is a a worst-possible case for this patch. It
does opens on a _pair_ of files in two different mounts in parallel.
This should cause my code to lose its "operate on the same mount"
optimization completely. This worst-case scenario causes a 3%
degredation in the benchmark.
I could probably get rid of even this 3%, but it would be more
complex than what I have here, and I think this is getting into
acceptable territory. In practice, I expect writing more than 3
bytes to a file, as well as disk I/O to mask any effects that this
has.
(To get rid of that 3%, we could have an #defined number of mounts
in the percpu variable. So, instead of a CPU getting operate only
on percpu data when it accesses only one mount, it could stay on
percpu data when it only accesses N or fewer mounts.)
[AV] merged fix for __clear_mnt_mount() stepping on freed vfsmount
Acked-by: Al Viro <viro@ZenIV.linux.org.uk>
Signed-off-by: Christoph Hellwig <hch@infradead.org>
Signed-off-by: Dave Hansen <haveblue@us.ibm.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Al Viro <viro@zeniv.linux.org.uk>
2008-02-15 23:37:59 +01:00
|
|
|
/*
|
2009-04-26 12:25:54 +02:00
|
|
|
* With writers on hold, if this value is zero, then there are
|
|
|
|
* definitely no active writers (although held writers may subsequently
|
|
|
|
* increment the count, they'll have to wait, and decrement it after
|
|
|
|
* seeing MNT_READONLY).
|
|
|
|
*
|
|
|
|
* It is OK to have counter incremented on one CPU and decremented on
|
|
|
|
* another: the sum will add up correctly. The danger would be when we
|
|
|
|
* sum up each counter, if we read a counter before it is incremented,
|
|
|
|
* but then read another CPU's count which it has been subsequently
|
|
|
|
* decremented from -- we would see more decrements than we should.
|
|
|
|
* MNT_WRITE_HOLD protects against this scenario, because
|
|
|
|
* mnt_want_write first increments count, then smp_mb, then spins on
|
|
|
|
* MNT_WRITE_HOLD, so it can't be decremented by another CPU while
|
|
|
|
* we're counting up here.
|
[PATCH] r/o bind mounts: track numbers of writers to mounts
This is the real meat of the entire series. It actually
implements the tracking of the number of writers to a mount.
However, it causes scalability problems because there can be
hundreds of cpus doing open()/close() on files on the same mnt at
the same time. Even an atomic_t in the mnt has massive scalaing
problems because the cacheline gets so terribly contended.
This uses a statically-allocated percpu variable. All want/drop
operations are local to a cpu as long that cpu operates on the same
mount, and there are no writer count imbalances. Writer count
imbalances happen when a write is taken on one cpu, and released
on another, like when an open/close pair is performed on two
Upon a remount,ro request, all of the data from the percpu
variables is collected (expensive, but very rare) and we determine
if there are any outstanding writers to the mount.
I've written a little benchmark to sit in a loop for a couple of
seconds in several cpus in parallel doing open/write/close loops.
http://sr71.net/~dave/linux/openbench.c
The code in here is a a worst-possible case for this patch. It
does opens on a _pair_ of files in two different mounts in parallel.
This should cause my code to lose its "operate on the same mount"
optimization completely. This worst-case scenario causes a 3%
degredation in the benchmark.
I could probably get rid of even this 3%, but it would be more
complex than what I have here, and I think this is getting into
acceptable territory. In practice, I expect writing more than 3
bytes to a file, as well as disk I/O to mask any effects that this
has.
(To get rid of that 3%, we could have an #defined number of mounts
in the percpu variable. So, instead of a CPU getting operate only
on percpu data when it accesses only one mount, it could stay on
percpu data when it only accesses N or fewer mounts.)
[AV] merged fix for __clear_mnt_mount() stepping on freed vfsmount
Acked-by: Al Viro <viro@ZenIV.linux.org.uk>
Signed-off-by: Christoph Hellwig <hch@infradead.org>
Signed-off-by: Dave Hansen <haveblue@us.ibm.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Al Viro <viro@zeniv.linux.org.uk>
2008-02-15 23:37:59 +01:00
|
|
|
*/
|
2011-01-07 07:50:10 +01:00
|
|
|
if (mnt_get_writers(mnt) > 0)
|
2009-04-26 12:25:54 +02:00
|
|
|
ret = -EBUSY;
|
|
|
|
else
|
2011-11-25 04:37:54 +01:00
|
|
|
mnt->mnt.mnt_flags |= MNT_READONLY;
|
2009-04-26 12:25:54 +02:00
|
|
|
/*
|
|
|
|
* MNT_READONLY must become visible before ~MNT_WRITE_HOLD, so writers
|
|
|
|
* that become unheld will see MNT_READONLY.
|
|
|
|
*/
|
|
|
|
smp_wmb();
|
2011-11-25 04:37:54 +01:00
|
|
|
mnt->mnt.mnt_flags &= ~MNT_WRITE_HOLD;
|
2013-09-29 17:24:49 +02:00
|
|
|
unlock_mount_hash();
|
[PATCH] r/o bind mounts: track numbers of writers to mounts
This is the real meat of the entire series. It actually
implements the tracking of the number of writers to a mount.
However, it causes scalability problems because there can be
hundreds of cpus doing open()/close() on files on the same mnt at
the same time. Even an atomic_t in the mnt has massive scalaing
problems because the cacheline gets so terribly contended.
This uses a statically-allocated percpu variable. All want/drop
operations are local to a cpu as long that cpu operates on the same
mount, and there are no writer count imbalances. Writer count
imbalances happen when a write is taken on one cpu, and released
on another, like when an open/close pair is performed on two
Upon a remount,ro request, all of the data from the percpu
variables is collected (expensive, but very rare) and we determine
if there are any outstanding writers to the mount.
I've written a little benchmark to sit in a loop for a couple of
seconds in several cpus in parallel doing open/write/close loops.
http://sr71.net/~dave/linux/openbench.c
The code in here is a a worst-possible case for this patch. It
does opens on a _pair_ of files in two different mounts in parallel.
This should cause my code to lose its "operate on the same mount"
optimization completely. This worst-case scenario causes a 3%
degredation in the benchmark.
I could probably get rid of even this 3%, but it would be more
complex than what I have here, and I think this is getting into
acceptable territory. In practice, I expect writing more than 3
bytes to a file, as well as disk I/O to mask any effects that this
has.
(To get rid of that 3%, we could have an #defined number of mounts
in the percpu variable. So, instead of a CPU getting operate only
on percpu data when it accesses only one mount, it could stay on
percpu data when it only accesses N or fewer mounts.)
[AV] merged fix for __clear_mnt_mount() stepping on freed vfsmount
Acked-by: Al Viro <viro@ZenIV.linux.org.uk>
Signed-off-by: Christoph Hellwig <hch@infradead.org>
Signed-off-by: Dave Hansen <haveblue@us.ibm.com>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Al Viro <viro@zeniv.linux.org.uk>
2008-02-15 23:37:59 +01:00
|
|
|
return ret;
|
2008-02-15 23:37:30 +01:00
|
|
|
}
|
|
|
|
|
2011-11-25 04:37:54 +01:00
|
|
|
static void __mnt_unmake_readonly(struct mount *mnt)
|
2008-02-15 23:38:00 +01:00
|
|
|
{
|
2013-09-29 17:24:49 +02:00
|
|
|
lock_mount_hash();
|
2011-11-25 04:37:54 +01:00
|
|
|
mnt->mnt.mnt_flags &= ~MNT_READONLY;
|
2013-09-29 17:24:49 +02:00
|
|
|
unlock_mount_hash();
|
2008-02-15 23:38:00 +01:00
|
|
|
}
|
|
|
|
|
2011-11-21 12:11:31 +01:00
|
|
|
int sb_prepare_remount_readonly(struct super_block *sb)
|
|
|
|
{
|
|
|
|
struct mount *mnt;
|
|
|
|
int err = 0;
|
|
|
|
|
2011-11-21 12:11:33 +01:00
|
|
|
/* Racy optimization. Recheck the counter under MNT_WRITE_HOLD */
|
|
|
|
if (atomic_long_read(&sb->s_remove_count))
|
|
|
|
return -EBUSY;
|
|
|
|
|
2013-09-29 17:24:49 +02:00
|
|
|
lock_mount_hash();
|
2011-11-21 12:11:31 +01:00
|
|
|
list_for_each_entry(mnt, &sb->s_mounts, mnt_instance) {
|
|
|
|
if (!(mnt->mnt.mnt_flags & MNT_READONLY)) {
|
|
|
|
mnt->mnt.mnt_flags |= MNT_WRITE_HOLD;
|
|
|
|
smp_mb();
|
|
|
|
if (mnt_get_writers(mnt) > 0) {
|
|
|
|
err = -EBUSY;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
2011-11-21 12:11:33 +01:00
|
|
|
if (!err && atomic_long_read(&sb->s_remove_count))
|
|
|
|
err = -EBUSY;
|
|
|
|
|
2011-11-21 12:11:31 +01:00
|
|
|
if (!err) {
|
|
|
|
sb->s_readonly_remount = 1;
|
|
|
|
smp_wmb();
|
|
|
|
}
|
|
|
|
list_for_each_entry(mnt, &sb->s_mounts, mnt_instance) {
|
|
|
|
if (mnt->mnt.mnt_flags & MNT_WRITE_HOLD)
|
|
|
|
mnt->mnt.mnt_flags &= ~MNT_WRITE_HOLD;
|
|
|
|
}
|
2013-09-29 17:24:49 +02:00
|
|
|
unlock_mount_hash();
|
2011-11-21 12:11:31 +01:00
|
|
|
|
|
|
|
return err;
|
|
|
|
}
|
|
|
|
|
2011-11-25 02:38:33 +01:00
|
|
|
static void free_vfsmnt(struct mount *mnt)
|
2005-04-17 00:20:36 +02:00
|
|
|
{
|
2015-02-13 23:36:41 +01:00
|
|
|
kfree_const(mnt->mnt_devname);
|
2009-04-26 12:25:54 +02:00
|
|
|
#ifdef CONFIG_SMP
|
2011-11-25 04:53:09 +01:00
|
|
|
free_percpu(mnt->mnt_pcp);
|
2009-04-26 12:25:54 +02:00
|
|
|
#endif
|
2011-11-25 02:38:33 +01:00
|
|
|
kmem_cache_free(mnt_cache, mnt);
|
2005-04-17 00:20:36 +02:00
|
|
|
}
|
|
|
|
|
2014-01-24 13:17:54 +01:00
|
|
|
static void delayed_free_vfsmnt(struct rcu_head *head)
|
|
|
|
{
|
|
|
|
free_vfsmnt(container_of(head, struct mount, mnt_rcu));
|
|
|
|
}
|
|
|
|
|
2013-09-30 04:06:07 +02:00
|
|
|
/* call under rcu_read_lock */
|
2015-05-08 17:43:53 +02:00
|
|
|
int __legitimize_mnt(struct vfsmount *bastard, unsigned seq)
|
2013-09-30 04:06:07 +02:00
|
|
|
{
|
|
|
|
struct mount *mnt;
|
|
|
|
if (read_seqretry(&mount_lock, seq))
|
2015-05-08 17:43:53 +02:00
|
|
|
return 1;
|
2013-09-30 04:06:07 +02:00
|
|
|
if (bastard == NULL)
|
2015-05-08 17:43:53 +02:00
|
|
|
return 0;
|
2013-09-30 04:06:07 +02:00
|
|
|
mnt = real_mount(bastard);
|
|
|
|
mnt_add_count(mnt, 1);
|
|
|
|
if (likely(!read_seqretry(&mount_lock, seq)))
|
2015-05-08 17:43:53 +02:00
|
|
|
return 0;
|
2013-09-30 04:06:07 +02:00
|
|
|
if (bastard->mnt_flags & MNT_SYNC_UMOUNT) {
|
|
|
|
mnt_add_count(mnt, -1);
|
2015-05-08 17:43:53 +02:00
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
return -1;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* call under rcu_read_lock */
|
|
|
|
bool legitimize_mnt(struct vfsmount *bastard, unsigned seq)
|
|
|
|
{
|
|
|
|
int res = __legitimize_mnt(bastard, seq);
|
|
|
|
if (likely(!res))
|
|
|
|
return true;
|
|
|
|
if (unlikely(res < 0)) {
|
|
|
|
rcu_read_unlock();
|
|
|
|
mntput(bastard);
|
|
|
|
rcu_read_lock();
|
2013-09-30 04:06:07 +02:00
|
|
|
}
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
2005-04-17 00:20:36 +02:00
|
|
|
/*
|
2013-10-01 22:11:26 +02:00
|
|
|
* find the first mount at @dentry on vfsmount @mnt.
|
2013-09-30 04:06:07 +02:00
|
|
|
* call under rcu_read_lock()
|
2005-04-17 00:20:36 +02:00
|
|
|
*/
|
2013-10-01 22:11:26 +02:00
|
|
|
struct mount *__lookup_mnt(struct vfsmount *mnt, struct dentry *dentry)
|
2005-04-17 00:20:36 +02:00
|
|
|
{
|
2014-03-21 02:10:51 +01:00
|
|
|
struct hlist_head *head = m_hash(mnt, dentry);
|
2013-10-01 22:11:26 +02:00
|
|
|
struct mount *p;
|
|
|
|
|
2014-03-21 02:10:51 +01:00
|
|
|
hlist_for_each_entry_rcu(p, head, mnt_hash)
|
2013-10-01 22:11:26 +02:00
|
|
|
if (&p->mnt_parent->mnt == mnt && p->mnt_mountpoint == dentry)
|
|
|
|
return p;
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
2005-11-07 23:20:17 +01:00
|
|
|
/*
|
2012-06-25 13:55:28 +02:00
|
|
|
* lookup_mnt - Return the first child mount mounted at path
|
|
|
|
*
|
|
|
|
* "First" means first mounted chronologically. If you create the
|
|
|
|
* following mounts:
|
|
|
|
*
|
|
|
|
* mount /dev/sda1 /mnt
|
|
|
|
* mount /dev/sda2 /mnt
|
|
|
|
* mount /dev/sda3 /mnt
|
|
|
|
*
|
|
|
|
* Then lookup_mnt() on the base /mnt dentry in the root mount will
|
|
|
|
* return successively the root dentry and vfsmount of /dev/sda1, then
|
|
|
|
* /dev/sda2, then /dev/sda3, then NULL.
|
|
|
|
*
|
|
|
|
* lookup_mnt takes a reference to the found vfsmount.
|
2005-11-07 23:20:17 +01:00
|
|
|
*/
|
2016-11-21 01:45:28 +01:00
|
|
|
struct vfsmount *lookup_mnt(const struct path *path)
|
2005-11-07 23:20:17 +01:00
|
|
|
{
|
2011-11-25 00:22:03 +01:00
|
|
|
struct mount *child_mnt;
|
2013-09-30 04:06:07 +02:00
|
|
|
struct vfsmount *m;
|
|
|
|
unsigned seq;
|
fs: brlock vfsmount_lock
fs: brlock vfsmount_lock
Use a brlock for the vfsmount lock. It must be taken for write whenever
modifying the mount hash or associated fields, and may be taken for read when
performing mount hash lookups.
A new lock is added for the mnt-id allocator, so it doesn't need to take
the heavy vfsmount write-lock.
The number of atomics should remain the same for fastpath rlock cases, though
code would be slightly slower due to per-cpu access. Scalability is not not be
much improved in common cases yet, due to other locks (ie. dcache_lock) getting
in the way. However path lookups crossing mountpoints should be one case where
scalability is improved (currently requiring the global lock).
The slowpath is slower due to use of brlock. On a 64 core, 64 socket, 32 node
Altix system (high latency to remote nodes), a simple umount microbenchmark
(mount --bind mnt mnt2 ; umount mnt2 loop 1000 times), before this patch it
took 6.8s, afterwards took 7.1s, about 5% slower.
Cc: Al Viro <viro@ZenIV.linux.org.uk>
Signed-off-by: Nick Piggin <npiggin@kernel.dk>
Signed-off-by: Al Viro <viro@zeniv.linux.org.uk>
2010-08-17 20:37:39 +02:00
|
|
|
|
2013-09-30 04:06:07 +02:00
|
|
|
rcu_read_lock();
|
|
|
|
do {
|
|
|
|
seq = read_seqbegin(&mount_lock);
|
|
|
|
child_mnt = __lookup_mnt(path->mnt, path->dentry);
|
|
|
|
m = child_mnt ? &child_mnt->mnt : NULL;
|
|
|
|
} while (!legitimize_mnt(m, seq));
|
|
|
|
rcu_read_unlock();
|
|
|
|
return m;
|
2005-11-07 23:20:17 +01:00
|
|
|
}
|
|
|
|
|
2013-10-05 04:15:13 +02:00
|
|
|
/*
|
|
|
|
* __is_local_mountpoint - Test to see if dentry is a mountpoint in the
|
|
|
|
* current mount namespace.
|
|
|
|
*
|
|
|
|
* The common case is dentries are not mountpoints at all and that
|
|
|
|
* test is handled inline. For the slow case when we are actually
|
|
|
|
* dealing with a mountpoint of some kind, walk through all of the
|
|
|
|
* mounts in the current mount namespace and test to see if the dentry
|
|
|
|
* is a mountpoint.
|
|
|
|
*
|
|
|
|
* The mount_hashtable is not usable in the context because we
|
|
|
|
* need to identify all mounts that may be in the current mount
|
|
|
|
* namespace not just a mount that happens to have some specified
|
|
|
|
* parent mount.
|
|
|
|
*/
|
|
|
|
bool __is_local_mountpoint(struct dentry *dentry)
|
|
|
|
{
|
|
|
|
struct mnt_namespace *ns = current->nsproxy->mnt_ns;
|
|
|
|
struct mount *mnt;
|
|
|
|
bool is_covered = false;
|
|
|
|
|
|
|
|
if (!d_mountpoint(dentry))
|
|
|
|
goto out;
|
|
|
|
|
|
|
|
down_read(&namespace_sem);
|
|
|
|
list_for_each_entry(mnt, &ns->list, mnt_list) {
|
|
|
|
is_covered = (mnt->mnt_mountpoint == dentry);
|
|
|
|
if (is_covered)
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
up_read(&namespace_sem);
|
|
|
|
out:
|
|
|
|
return is_covered;
|
|
|
|
}
|
|
|
|
|
2014-02-25 02:32:34 +01:00
|
|
|
static struct mountpoint *lookup_mountpoint(struct dentry *dentry)
|
2013-03-15 15:53:28 +01:00
|
|
|
{
|
2014-02-28 19:46:44 +01:00
|
|
|
struct hlist_head *chain = mp_hash(dentry);
|
2013-03-15 15:53:28 +01:00
|
|
|
struct mountpoint *mp;
|
|
|
|
|
2014-02-28 19:46:44 +01:00
|
|
|
hlist_for_each_entry(mp, chain, m_hash) {
|
2013-03-15 15:53:28 +01:00
|
|
|
if (mp->m_dentry == dentry) {
|
|
|
|
/* might be worth a WARN_ON() */
|
|
|
|
if (d_unlinked(dentry))
|
|
|
|
return ERR_PTR(-ENOENT);
|
|
|
|
mp->m_count++;
|
|
|
|
return mp;
|
|
|
|
}
|
|
|
|
}
|
2014-02-25 02:32:34 +01:00
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
2017-01-03 02:18:43 +01:00
|
|
|
static struct mountpoint *get_mountpoint(struct dentry *dentry)
|
2014-02-25 02:32:34 +01:00
|
|
|
{
|
2017-01-03 02:18:43 +01:00
|
|
|
struct mountpoint *mp, *new = NULL;
|
2014-02-25 02:32:34 +01:00
|
|
|
int ret;
|
2013-03-15 15:53:28 +01:00
|
|
|
|
2017-01-03 02:18:43 +01:00
|
|
|
if (d_mountpoint(dentry)) {
|
|
|
|
mountpoint:
|
|
|
|
read_seqlock_excl(&mount_lock);
|
|
|
|
mp = lookup_mountpoint(dentry);
|
|
|
|
read_sequnlock_excl(&mount_lock);
|
|
|
|
if (mp)
|
|
|
|
goto done;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (!new)
|
|
|
|
new = kmalloc(sizeof(struct mountpoint), GFP_KERNEL);
|
|
|
|
if (!new)
|
2013-03-15 15:53:28 +01:00
|
|
|
return ERR_PTR(-ENOMEM);
|
|
|
|
|
2017-01-03 02:18:43 +01:00
|
|
|
|
|
|
|
/* Exactly one processes may set d_mounted */
|
2013-09-05 14:39:11 +02:00
|
|
|
ret = d_set_mounted(dentry);
|
|
|
|
|
2017-01-03 02:18:43 +01:00
|
|
|
/* Someone else set d_mounted? */
|
|
|
|
if (ret == -EBUSY)
|
|
|
|
goto mountpoint;
|
|
|
|
|
|
|
|
/* The dentry is not available as a mountpoint? */
|
|
|
|
mp = ERR_PTR(ret);
|
|
|
|
if (ret)
|
|
|
|
goto done;
|
|
|
|
|
|
|
|
/* Add the new mountpoint to the hash table */
|
|
|
|
read_seqlock_excl(&mount_lock);
|
|
|
|
new->m_dentry = dentry;
|
|
|
|
new->m_count = 1;
|
|
|
|
hlist_add_head(&new->m_hash, mp_hash(dentry));
|
|
|
|
INIT_HLIST_HEAD(&new->m_list);
|
|
|
|
read_sequnlock_excl(&mount_lock);
|
|
|
|
|
|
|
|
mp = new;
|
|
|
|
new = NULL;
|
|
|
|
done:
|
|
|
|
kfree(new);
|
2013-03-15 15:53:28 +01:00
|
|
|
return mp;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void put_mountpoint(struct mountpoint *mp)
|
|
|
|
{
|
|
|
|
if (!--mp->m_count) {
|
|
|
|
struct dentry *dentry = mp->m_dentry;
|
2013-09-23 04:37:01 +02:00
|
|
|
BUG_ON(!hlist_empty(&mp->m_list));
|
2013-03-15 15:53:28 +01:00
|
|
|
spin_lock(&dentry->d_lock);
|
|
|
|
dentry->d_flags &= ~DCACHE_MOUNTED;
|
|
|
|
spin_unlock(&dentry->d_lock);
|
2014-02-28 19:46:44 +01:00
|
|
|
hlist_del(&mp->m_hash);
|
2013-03-15 15:53:28 +01:00
|
|
|
kfree(mp);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2011-11-25 06:46:35 +01:00
|
|
|
static inline int check_mnt(struct mount *mnt)
|
2005-04-17 00:20:36 +02:00
|
|
|
{
|
2006-12-08 11:37:56 +01:00
|
|
|
return mnt->mnt_ns == current->nsproxy->mnt_ns;
|
2005-04-17 00:20:36 +02:00
|
|
|
}
|
|
|
|
|
fs: brlock vfsmount_lock
fs: brlock vfsmount_lock
Use a brlock for the vfsmount lock. It must be taken for write whenever
modifying the mount hash or associated fields, and may be taken for read when
performing mount hash lookups.
A new lock is added for the mnt-id allocator, so it doesn't need to take
the heavy vfsmount write-lock.
The number of atomics should remain the same for fastpath rlock cases, though
code would be slightly slower due to per-cpu access. Scalability is not not be
much improved in common cases yet, due to other locks (ie. dcache_lock) getting
in the way. However path lookups crossing mountpoints should be one case where
scalability is improved (currently requiring the global lock).
The slowpath is slower due to use of brlock. On a 64 core, 64 socket, 32 node
Altix system (high latency to remote nodes), a simple umount microbenchmark
(mount --bind mnt mnt2 ; umount mnt2 loop 1000 times), before this patch it
took 6.8s, afterwards took 7.1s, about 5% slower.
Cc: Al Viro <viro@ZenIV.linux.org.uk>
Signed-off-by: Nick Piggin <npiggin@kernel.dk>
Signed-off-by: Al Viro <viro@zeniv.linux.org.uk>
2010-08-17 20:37:39 +02:00
|
|
|
/*
|
|
|
|
* vfsmount lock must be held for write
|
|
|
|
*/
|
2006-12-08 11:37:56 +01:00
|
|
|
static void touch_mnt_namespace(struct mnt_namespace *ns)
|
2005-11-07 23:15:49 +01:00
|
|
|
{
|
|
|
|
if (ns) {
|
|
|
|
ns->event = ++event;
|
|
|
|
wake_up_interruptible(&ns->poll);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
fs: brlock vfsmount_lock
fs: brlock vfsmount_lock
Use a brlock for the vfsmount lock. It must be taken for write whenever
modifying the mount hash or associated fields, and may be taken for read when
performing mount hash lookups.
A new lock is added for the mnt-id allocator, so it doesn't need to take
the heavy vfsmount write-lock.
The number of atomics should remain the same for fastpath rlock cases, though
code would be slightly slower due to per-cpu access. Scalability is not not be
much improved in common cases yet, due to other locks (ie. dcache_lock) getting
in the way. However path lookups crossing mountpoints should be one case where
scalability is improved (currently requiring the global lock).
The slowpath is slower due to use of brlock. On a 64 core, 64 socket, 32 node
Altix system (high latency to remote nodes), a simple umount microbenchmark
(mount --bind mnt mnt2 ; umount mnt2 loop 1000 times), before this patch it
took 6.8s, afterwards took 7.1s, about 5% slower.
Cc: Al Viro <viro@ZenIV.linux.org.uk>
Signed-off-by: Nick Piggin <npiggin@kernel.dk>
Signed-off-by: Al Viro <viro@zeniv.linux.org.uk>
2010-08-17 20:37:39 +02:00
|
|
|
/*
|
|
|
|
* vfsmount lock must be held for write
|
|
|
|
*/
|
2006-12-08 11:37:56 +01:00
|
|
|
static void __touch_mnt_namespace(struct mnt_namespace *ns)
|
2005-11-07 23:15:49 +01:00
|
|
|
{
|
|
|
|
if (ns && ns->event != event) {
|
|
|
|
ns->event = event;
|
|
|
|
wake_up_interruptible(&ns->poll);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
fs: brlock vfsmount_lock
fs: brlock vfsmount_lock
Use a brlock for the vfsmount lock. It must be taken for write whenever
modifying the mount hash or associated fields, and may be taken for read when
performing mount hash lookups.
A new lock is added for the mnt-id allocator, so it doesn't need to take
the heavy vfsmount write-lock.
The number of atomics should remain the same for fastpath rlock cases, though
code would be slightly slower due to per-cpu access. Scalability is not not be
much improved in common cases yet, due to other locks (ie. dcache_lock) getting
in the way. However path lookups crossing mountpoints should be one case where
scalability is improved (currently requiring the global lock).
The slowpath is slower due to use of brlock. On a 64 core, 64 socket, 32 node
Altix system (high latency to remote nodes), a simple umount microbenchmark
(mount --bind mnt mnt2 ; umount mnt2 loop 1000 times), before this patch it
took 6.8s, afterwards took 7.1s, about 5% slower.
Cc: Al Viro <viro@ZenIV.linux.org.uk>
Signed-off-by: Nick Piggin <npiggin@kernel.dk>
Signed-off-by: Al Viro <viro@zeniv.linux.org.uk>
2010-08-17 20:37:39 +02:00
|
|
|
/*
|
|
|
|
* vfsmount lock must be held for write
|
|
|
|
*/
|
2014-12-29 20:03:41 +01:00
|
|
|
static void unhash_mnt(struct mount *mnt)
|
2011-11-25 01:41:16 +01:00
|
|
|
{
|
2011-11-25 04:19:58 +01:00
|
|
|
mnt->mnt_parent = mnt;
|
2011-11-25 04:25:07 +01:00
|
|
|
mnt->mnt_mountpoint = mnt->mnt.mnt_root;
|
2011-11-25 05:24:33 +01:00
|
|
|
list_del_init(&mnt->mnt_child);
|
2014-03-21 02:10:51 +01:00
|
|
|
hlist_del_init_rcu(&mnt->mnt_hash);
|
2013-09-23 04:37:01 +02:00
|
|
|
hlist_del_init(&mnt->mnt_mp_list);
|
2013-03-15 15:53:28 +01:00
|
|
|
put_mountpoint(mnt->mnt_mp);
|
|
|
|
mnt->mnt_mp = NULL;
|
2005-04-17 00:20:36 +02:00
|
|
|
}
|
|
|
|
|
2014-12-29 20:03:41 +01:00
|
|
|
/*
|
|
|
|
* vfsmount lock must be held for write
|
|
|
|
*/
|
|
|
|
static void detach_mnt(struct mount *mnt, struct path *old_path)
|
|
|
|
{
|
|
|
|
old_path->dentry = mnt->mnt_mountpoint;
|
|
|
|
old_path->mnt = &mnt->mnt_parent->mnt;
|
|
|
|
unhash_mnt(mnt);
|
|
|
|
}
|
|
|
|
|
2015-01-16 05:58:33 +01:00
|
|
|
/*
|
|
|
|
* vfsmount lock must be held for write
|
|
|
|
*/
|
|
|
|
static void umount_mnt(struct mount *mnt)
|
|
|
|
{
|
|
|
|
/* old mountpoint will be dropped when we can do that */
|
|
|
|
mnt->mnt_ex_mountpoint = mnt->mnt_mountpoint;
|
|
|
|
unhash_mnt(mnt);
|
|
|
|
}
|
|
|
|
|
fs: brlock vfsmount_lock
fs: brlock vfsmount_lock
Use a brlock for the vfsmount lock. It must be taken for write whenever
modifying the mount hash or associated fields, and may be taken for read when
performing mount hash lookups.
A new lock is added for the mnt-id allocator, so it doesn't need to take
the heavy vfsmount write-lock.
The number of atomics should remain the same for fastpath rlock cases, though
code would be slightly slower due to per-cpu access. Scalability is not not be
much improved in common cases yet, due to other locks (ie. dcache_lock) getting
in the way. However path lookups crossing mountpoints should be one case where
scalability is improved (currently requiring the global lock).
The slowpath is slower due to use of brlock. On a 64 core, 64 socket, 32 node
Altix system (high latency to remote nodes), a simple umount microbenchmark
(mount --bind mnt mnt2 ; umount mnt2 loop 1000 times), before this patch it
took 6.8s, afterwards took 7.1s, about 5% slower.
Cc: Al Viro <viro@ZenIV.linux.org.uk>
Signed-off-by: Nick Piggin <npiggin@kernel.dk>
Signed-off-by: Al Viro <viro@zeniv.linux.org.uk>
2010-08-17 20:37:39 +02:00
|
|
|
/*
|
|
|
|
* vfsmount lock must be held for write
|
|
|
|
*/
|
2013-03-15 15:53:28 +01:00
|
|
|
void mnt_set_mountpoint(struct mount *mnt,
|
|
|
|
struct mountpoint *mp,
|
2011-11-25 03:28:22 +01:00
|
|
|
struct mount *child_mnt)
|
2005-11-07 23:19:50 +01:00
|
|
|
{
|
2013-03-15 15:53:28 +01:00
|
|
|
mp->m_count++;
|
2011-11-25 09:19:09 +01:00
|
|
|
mnt_add_count(mnt, 1); /* essentially, that's mntget */
|
2013-03-15 15:53:28 +01:00
|
|
|
child_mnt->mnt_mountpoint = dget(mp->m_dentry);
|
2011-11-25 09:19:09 +01:00
|
|
|
child_mnt->mnt_parent = mnt;
|
2013-03-15 15:53:28 +01:00
|
|
|
child_mnt->mnt_mp = mp;
|
2013-09-23 04:37:01 +02:00
|
|
|
hlist_add_head(&child_mnt->mnt_mp_list, &mp->m_list);
|
2005-11-07 23:19:50 +01:00
|
|
|
}
|
|
|
|
|
2017-01-20 06:28:35 +01:00
|
|
|
static void __attach_mnt(struct mount *mnt, struct mount *parent)
|
|
|
|
{
|
|
|
|
hlist_add_head_rcu(&mnt->mnt_hash,
|
|
|
|
m_hash(&parent->mnt, mnt->mnt_mountpoint));
|
|
|
|
list_add_tail(&mnt->mnt_child, &parent->mnt_mounts);
|
|
|
|
}
|
|
|
|
|
fs: brlock vfsmount_lock
fs: brlock vfsmount_lock
Use a brlock for the vfsmount lock. It must be taken for write whenever
modifying the mount hash or associated fields, and may be taken for read when
performing mount hash lookups.
A new lock is added for the mnt-id allocator, so it doesn't need to take
the heavy vfsmount write-lock.
The number of atomics should remain the same for fastpath rlock cases, though
code would be slightly slower due to per-cpu access. Scalability is not not be
much improved in common cases yet, due to other locks (ie. dcache_lock) getting
in the way. However path lookups crossing mountpoints should be one case where
scalability is improved (currently requiring the global lock).
The slowpath is slower due to use of brlock. On a 64 core, 64 socket, 32 node
Altix system (high latency to remote nodes), a simple umount microbenchmark
(mount --bind mnt mnt2 ; umount mnt2 loop 1000 times), before this patch it
took 6.8s, afterwards took 7.1s, about 5% slower.
Cc: Al Viro <viro@ZenIV.linux.org.uk>
Signed-off-by: Nick Piggin <npiggin@kernel.dk>
Signed-off-by: Al Viro <viro@zeniv.linux.org.uk>
2010-08-17 20:37:39 +02:00
|
|
|
/*
|
|
|
|
* vfsmount lock must be held for write
|
|
|
|
*/
|
2013-03-15 15:53:28 +01:00
|
|
|
static void attach_mnt(struct mount *mnt,
|
|
|
|
struct mount *parent,
|
|
|
|
struct mountpoint *mp)
|
2005-04-17 00:20:36 +02:00
|
|
|
{
|
2013-03-15 15:53:28 +01:00
|
|
|
mnt_set_mountpoint(parent, mp, mnt);
|
2017-01-20 06:28:35 +01:00
|
|
|
__attach_mnt(mnt, parent);
|
2005-11-07 23:19:50 +01:00
|
|
|
}
|
|
|
|
|
2017-01-20 06:28:35 +01:00
|
|
|
void mnt_change_mountpoint(struct mount *parent, struct mountpoint *mp, struct mount *mnt)
|
2014-08-10 09:44:55 +02:00
|
|
|
{
|
2017-01-20 06:28:35 +01:00
|
|
|
struct mountpoint *old_mp = mnt->mnt_mp;
|
|
|
|
struct dentry *old_mountpoint = mnt->mnt_mountpoint;
|
|
|
|
struct mount *old_parent = mnt->mnt_parent;
|
|
|
|
|
|
|
|
list_del_init(&mnt->mnt_child);
|
|
|
|
hlist_del_init(&mnt->mnt_mp_list);
|
|
|
|
hlist_del_init_rcu(&mnt->mnt_hash);
|
|
|
|
|
|
|
|
attach_mnt(mnt, parent, mp);
|
|
|
|
|
|
|
|
put_mountpoint(old_mp);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Safely avoid even the suggestion this code might sleep or
|
|
|
|
* lock the mount hash by taking advantage of the knowledge that
|
|
|
|
* mnt_change_mountpoint will not release the final reference
|
|
|
|
* to a mountpoint.
|
|
|
|
*
|
|
|
|
* During mounting, the mount passed in as the parent mount will
|
|
|
|
* continue to use the old mountpoint and during unmounting, the
|
|
|
|
* old mountpoint will continue to exist until namespace_unlock,
|
|
|
|
* which happens well after mnt_change_mountpoint.
|
|
|
|
*/
|
|
|
|
spin_lock(&old_mountpoint->d_lock);
|
|
|
|
old_mountpoint->d_lockref.count--;
|
|
|
|
spin_unlock(&old_mountpoint->d_lock);
|
|
|
|
|
|
|
|
mnt_add_count(old_parent, -1);
|
2014-08-10 09:44:55 +02:00
|
|
|
}
|
|
|
|
|
2005-11-07 23:19:50 +01:00
|
|
|
/*
|
fs: brlock vfsmount_lock
fs: brlock vfsmount_lock
Use a brlock for the vfsmount lock. It must be taken for write whenever
modifying the mount hash or associated fields, and may be taken for read when
performing mount hash lookups.
A new lock is added for the mnt-id allocator, so it doesn't need to take
the heavy vfsmount write-lock.
The number of atomics should remain the same for fastpath rlock cases, though
code would be slightly slower due to per-cpu access. Scalability is not not be
much improved in common cases yet, due to other locks (ie. dcache_lock) getting
in the way. However path lookups crossing mountpoints should be one case where
scalability is improved (currently requiring the global lock).
The slowpath is slower due to use of brlock. On a 64 core, 64 socket, 32 node
Altix system (high latency to remote nodes), a simple umount microbenchmark
(mount --bind mnt mnt2 ; umount mnt2 loop 1000 times), before this patch it
took 6.8s, afterwards took 7.1s, about 5% slower.
Cc: Al Viro <viro@ZenIV.linux.org.uk>
Signed-off-by: Nick Piggin <npiggin@kernel.dk>
Signed-off-by: Al Viro <viro@zeniv.linux.org.uk>
2010-08-17 20:37:39 +02:00
|
|
|
* vfsmount lock must be held for write
|
2005-11-07 23:19:50 +01:00
|
|
|
*/
|
2017-01-20 06:28:35 +01:00
|
|
|
static void commit_tree(struct mount *mnt)
|
2005-11-07 23:19:50 +01:00
|
|
|
{
|
2011-11-25 04:19:58 +01:00
|
|
|
struct mount *parent = mnt->mnt_parent;
|
2011-11-25 04:37:54 +01:00
|
|
|
struct mount *m;
|
2005-11-07 23:19:50 +01:00
|
|
|
LIST_HEAD(head);
|
2011-11-25 06:46:35 +01:00
|
|
|
struct mnt_namespace *n = parent->mnt_ns;
|
2005-11-07 23:19:50 +01:00
|
|
|
|
2011-11-25 04:19:58 +01:00
|
|
|
BUG_ON(parent == mnt);
|
2005-11-07 23:19:50 +01:00
|
|
|
|
2011-11-25 08:19:55 +01:00
|
|
|
list_add_tail(&head, &mnt->mnt_list);
|
2012-06-09 06:59:08 +02:00
|
|
|
list_for_each_entry(m, &head, mnt_list)
|
2011-11-25 06:46:35 +01:00
|
|
|
m->mnt_ns = n;
|
2011-01-15 04:30:21 +01:00
|
|
|
|
2005-11-07 23:19:50 +01:00
|
|
|
list_splice(&head, n->list.prev);
|
|
|
|
|
2016-09-28 07:27:17 +02:00
|
|
|
n->mounts += n->pending_mounts;
|
|
|
|
n->pending_mounts = 0;
|
|
|
|
|
2017-01-20 06:28:35 +01:00
|
|
|
__attach_mnt(mnt, parent);
|
2006-12-08 11:37:56 +01:00
|
|
|
touch_mnt_namespace(n);
|
2005-04-17 00:20:36 +02:00
|
|
|
}
|
|
|
|
|
2011-11-25 09:06:56 +01:00
|
|
|
static struct mount *next_mnt(struct mount *p, struct mount *root)
|
2005-04-17 00:20:36 +02:00
|
|
|
{
|
2011-11-25 05:24:33 +01:00
|
|
|
struct list_head *next = p->mnt_mounts.next;
|
|
|
|
if (next == &p->mnt_mounts) {
|
2005-04-17 00:20:36 +02:00
|
|
|
while (1) {
|
2011-11-25 09:06:56 +01:00
|
|
|
if (p == root)
|
2005-04-17 00:20:36 +02:00
|
|
|
return NULL;
|
2011-11-25 05:24:33 +01:00
|
|
|
next = p->mnt_child.next;
|
|
|
|
if (next != &p->mnt_parent->mnt_mounts)
|
2005-04-17 00:20:36 +02:00
|
|
|
break;
|
2011-11-25 04:19:58 +01:00
|
|
|
p = p->mnt_parent;
|
2005-04-17 00:20:36 +02:00
|
|
|
}
|
|
|
|
}
|
2011-11-25 05:24:33 +01:00
|
|
|
return list_entry(next, struct mount, mnt_child);
|
2005-04-17 00:20:36 +02:00
|
|
|
}
|
|
|
|
|
2011-11-25 00:57:30 +01:00
|
|
|
static struct mount *skip_mnt_tree(struct mount *p)
|
2005-11-07 23:21:20 +01:00
|
|
|
{
|
2011-11-25 05:24:33 +01:00
|
|
|
struct list_head *prev = p->mnt_mounts.prev;
|
|
|
|
while (prev != &p->mnt_mounts) {
|
|
|
|
p = list_entry(prev, struct mount, mnt_child);
|
|
|
|
prev = p->mnt_mounts.prev;
|
2005-11-07 23:21:20 +01:00
|
|
|
}
|
|
|
|
return p;
|
|
|
|
}
|
|
|
|
|
2011-03-18 03:08:28 +01:00
|
|
|
struct vfsmount *
|
|
|
|
vfs_kern_mount(struct file_system_type *type, int flags, const char *name, void *data)
|
|
|
|
{
|
2011-11-25 02:38:33 +01:00
|
|
|
struct mount *mnt;
|
2011-03-18 03:08:28 +01:00
|
|
|
struct dentry *root;
|
|
|
|
|
|
|
|
if (!type)
|
|
|
|
return ERR_PTR(-ENODEV);
|
|
|
|
|
|
|
|
mnt = alloc_vfsmnt(name);
|
|
|
|
if (!mnt)
|
|
|
|
return ERR_PTR(-ENOMEM);
|
|
|
|
|
2017-07-17 09:45:35 +02:00
|
|
|
if (flags & SB_KERNMOUNT)
|
2011-11-25 02:38:33 +01:00
|
|
|
mnt->mnt.mnt_flags = MNT_INTERNAL;
|
2011-03-18 03:08:28 +01:00
|
|
|
|
|
|
|
root = mount_fs(type, flags, name, data);
|
|
|
|
if (IS_ERR(root)) {
|
2014-01-24 13:17:54 +01:00
|
|
|
mnt_free_id(mnt);
|
2011-03-18 03:08:28 +01:00
|
|
|
free_vfsmnt(mnt);
|
|
|
|
return ERR_CAST(root);
|
|
|
|
}
|
|
|
|
|
2011-11-25 02:38:33 +01:00
|
|
|
mnt->mnt.mnt_root = root;
|
|
|
|
mnt->mnt.mnt_sb = root->d_sb;
|
2011-11-25 04:25:07 +01:00
|
|
|
mnt->mnt_mountpoint = mnt->mnt.mnt_root;
|
2011-11-25 04:19:58 +01:00
|
|
|
mnt->mnt_parent = mnt;
|
2013-09-29 17:24:49 +02:00
|
|
|
lock_mount_hash();
|
2011-11-21 12:11:30 +01:00
|
|
|
list_add_tail(&mnt->mnt_instance, &root->d_sb->s_mounts);
|
2013-09-29 17:24:49 +02:00
|
|
|
unlock_mount_hash();
|
2011-11-25 02:38:33 +01:00
|
|
|
return &mnt->mnt;
|
2011-03-18 03:08:28 +01:00
|
|
|
}
|
|
|
|
EXPORT_SYMBOL_GPL(vfs_kern_mount);
|
|
|
|
|
2017-01-31 18:06:16 +01:00
|
|
|
struct vfsmount *
|
|
|
|
vfs_submount(const struct dentry *mountpoint, struct file_system_type *type,
|
|
|
|
const char *name, void *data)
|
|
|
|
{
|
|
|
|
/* Until it is worked out how to pass the user namespace
|
|
|
|
* through from the parent mount to the submount don't support
|
|
|
|
* unprivileged mounts with submounts.
|
|
|
|
*/
|
|
|
|
if (mountpoint->d_sb->s_user_ns != &init_user_ns)
|
|
|
|
return ERR_PTR(-EPERM);
|
|
|
|
|
2017-07-17 09:45:35 +02:00
|
|
|
return vfs_kern_mount(type, SB_SUBMOUNT, name, data);
|
2017-01-31 18:06:16 +01:00
|
|
|
}
|
|
|
|
EXPORT_SYMBOL_GPL(vfs_submount);
|
|
|
|
|
2011-11-25 03:24:27 +01:00
|
|
|
static struct mount *clone_mnt(struct mount *old, struct dentry *root,
|
2005-11-07 23:17:22 +01:00
|
|
|
int flag)
|
2005-04-17 00:20:36 +02:00
|
|
|
{
|
2011-11-25 03:24:27 +01:00
|
|
|
struct super_block *sb = old->mnt.mnt_sb;
|
2012-06-25 13:55:18 +02:00
|
|
|
struct mount *mnt;
|
|
|
|
int err;
|
2005-04-17 00:20:36 +02:00
|
|
|
|
2012-06-25 13:55:18 +02:00
|
|
|
mnt = alloc_vfsmnt(old->mnt_devname);
|
|
|
|
if (!mnt)
|
|
|
|
return ERR_PTR(-ENOMEM);
|
2008-03-27 13:06:23 +01:00
|
|
|
|
2012-07-31 22:13:04 +02:00
|
|
|
if (flag & (CL_SLAVE | CL_PRIVATE | CL_SHARED_TO_SLAVE))
|
2012-06-25 13:55:18 +02:00
|
|
|
mnt->mnt_group_id = 0; /* not a peer of original */
|
|
|
|
else
|
|
|
|
mnt->mnt_group_id = old->mnt_group_id;
|
2005-11-07 23:19:50 +01:00
|
|
|
|
2012-06-25 13:55:18 +02:00
|
|
|
if ((flag & CL_MAKE_SHARED) && !mnt->mnt_group_id) {
|
|
|
|
err = mnt_alloc_group_id(mnt);
|
|
|
|
if (err)
|
|
|
|
goto out_free;
|
2005-04-17 00:20:36 +02:00
|
|
|
}
|
2012-06-25 13:55:18 +02:00
|
|
|
|
smarter propagate_mnt()
The current mainline has copies propagated to *all* nodes, then
tears down the copies we made for nodes that do not contain
counterparts of the desired mountpoint. That sets the right
propagation graph for the copies (at teardown time we move
the slaves of removed node to a surviving peer or directly
to master), but we end up paying a fairly steep price in
useless allocations. It's fairly easy to create a situation
where N calls of mount(2) create exactly N bindings, with
O(N^2) vfsmounts allocated and freed in process.
Fortunately, it is possible to avoid those allocations/freeings.
The trick is to create copies in the right order and find which
one would've eventually become a master with the current algorithm.
It turns out to be possible in O(nodes getting propagation) time
and with no extra allocations at all.
One part is that we need to make sure that eventual master will be
created before its slaves, so we need to walk the propagation
tree in a different order - by peer groups. And iterate through
the peers before dealing with the next group.
Another thing is finding the (earlier) copy that will be a master
of one we are about to create; to do that we are (temporary) marking
the masters of mountpoints we are attaching the copies to.
Either we are in a peer of the last mountpoint we'd dealt with,
or we have the following situation: we are attaching to mountpoint M,
the last copy S_0 had been attached to M_0 and there are sequences
S_0...S_n, M_0...M_n such that S_{i+1} is a master of S_{i},
S_{i} mounted on M{i} and we need to create a slave of the first S_{k}
such that M is getting propagation from M_{k}. It means that the master
of M_{k} will be among the sequence of masters of M. On the
other hand, the nearest marked node in that sequence will either
be the master of M_{k} or the master of M_{k-1} (the latter -
in the case if M_{k-1} is a slave of something M gets propagation
from, but in a wrong peer group).
So we go through the sequence of masters of M until we find
a marked one (P). Let N be the one before it. Then we go through
the sequence of masters of S_0 until we find one (say, S) mounted
on a node D that has P as master and check if D is a peer of N.
If it is, S will be the master of new copy, if not - the master of S
will be.
That's it for the hard part; the rest is fairly simple. Iterator
is in next_group(), handling of one prospective mountpoint is
propagate_one().
It seems to survive all tests and gives a noticably better performance
than the current mainline for setups that are seriously using shared
subtrees.
Cc: stable@vger.kernel.org
Signed-off-by: Al Viro <viro@zeniv.linux.org.uk>
2014-02-27 15:35:45 +01:00
|
|
|
mnt->mnt.mnt_flags = old->mnt.mnt_flags & ~(MNT_WRITE_HOLD|MNT_MARKED);
|
2013-03-22 12:08:05 +01:00
|
|
|
/* Don't allow unprivileged users to change mount flags */
|
mnt: Correct permission checks in do_remount
While invesgiating the issue where in "mount --bind -oremount,ro ..."
would result in later "mount --bind -oremount,rw" succeeding even if
the mount started off locked I realized that there are several
additional mount flags that should be locked and are not.
In particular MNT_NOSUID, MNT_NODEV, MNT_NOEXEC, and the atime
flags in addition to MNT_READONLY should all be locked. These
flags are all per superblock, can all be changed with MS_BIND,
and should not be changable if set by a more privileged user.
The following additions to the current logic are added in this patch.
- nosuid may not be clearable by a less privileged user.
- nodev may not be clearable by a less privielged user.
- noexec may not be clearable by a less privileged user.
- atime flags may not be changeable by a less privileged user.
The logic with atime is that always setting atime on access is a
global policy and backup software and auditing software could break if
atime bits are not updated (when they are configured to be updated),
and serious performance degradation could result (DOS attack) if atime
updates happen when they have been explicitly disabled. Therefore an
unprivileged user should not be able to mess with the atime bits set
by a more privileged user.
The additional restrictions are implemented with the addition of
MNT_LOCK_NOSUID, MNT_LOCK_NODEV, MNT_LOCK_NOEXEC, and MNT_LOCK_ATIME
mnt flags.
Taken together these changes and the fixes for MNT_LOCK_READONLY
should make it safe for an unprivileged user to create a user
namespace and to call "mount --bind -o remount,... ..." without
the danger of mount flags being changed maliciously.
Cc: stable@vger.kernel.org
Acked-by: Serge E. Hallyn <serge.hallyn@ubuntu.com>
Signed-off-by: "Eric W. Biederman" <ebiederm@xmission.com>
2014-07-29 02:26:07 +02:00
|
|
|
if (flag & CL_UNPRIVILEGED) {
|
|
|
|
mnt->mnt.mnt_flags |= MNT_LOCK_ATIME;
|
|
|
|
|
|
|
|
if (mnt->mnt.mnt_flags & MNT_READONLY)
|
|
|
|
mnt->mnt.mnt_flags |= MNT_LOCK_READONLY;
|
|
|
|
|
|
|
|
if (mnt->mnt.mnt_flags & MNT_NODEV)
|
|
|
|
mnt->mnt.mnt_flags |= MNT_LOCK_NODEV;
|
|
|
|
|
|
|
|
if (mnt->mnt.mnt_flags & MNT_NOSUID)
|
|
|
|
mnt->mnt.mnt_flags |= MNT_LOCK_NOSUID;
|
|
|
|
|
|
|
|
if (mnt->mnt.mnt_flags & MNT_NOEXEC)
|
|
|
|
mnt->mnt.mnt_flags |= MNT_LOCK_NOEXEC;
|
|
|
|
}
|
2013-03-22 12:08:05 +01:00
|
|
|
|
2013-03-30 05:04:39 +01:00
|
|
|
/* Don't allow unprivileged users to reveal what is under a mount */
|
2014-10-08 02:11:46 +02:00
|
|
|
if ((flag & CL_UNPRIVILEGED) &&
|
|
|
|
(!(flag & CL_EXPIRE) || list_empty(&old->mnt_expire)))
|
2013-03-30 05:04:39 +01:00
|
|
|
mnt->mnt.mnt_flags |= MNT_LOCKED;
|
|
|
|
|
2012-06-25 13:55:18 +02:00
|
|
|
atomic_inc(&sb->s_active);
|
|
|
|
mnt->mnt.mnt_sb = sb;
|
|
|
|
mnt->mnt.mnt_root = dget(root);
|
|
|
|
mnt->mnt_mountpoint = mnt->mnt.mnt_root;
|
|
|
|
mnt->mnt_parent = mnt;
|
2013-09-29 17:24:49 +02:00
|
|
|
lock_mount_hash();
|
2012-06-25 13:55:18 +02:00
|
|
|
list_add_tail(&mnt->mnt_instance, &sb->s_mounts);
|
2013-09-29 17:24:49 +02:00
|
|
|
unlock_mount_hash();
|
2012-06-25 13:55:18 +02:00
|
|
|
|
2012-07-31 22:13:04 +02:00
|
|
|
if ((flag & CL_SLAVE) ||
|
|
|
|
((flag & CL_SHARED_TO_SLAVE) && IS_MNT_SHARED(old))) {
|
2012-06-25 13:55:18 +02:00
|
|
|
list_add(&mnt->mnt_slave, &old->mnt_slave_list);
|
|
|
|
mnt->mnt_master = old;
|
|
|
|
CLEAR_MNT_SHARED(mnt);
|
|
|
|
} else if (!(flag & CL_PRIVATE)) {
|
|
|
|
if ((flag & CL_MAKE_SHARED) || IS_MNT_SHARED(old))
|
|
|
|
list_add(&mnt->mnt_share, &old->mnt_share);
|
|
|
|
if (IS_MNT_SLAVE(old))
|
|
|
|
list_add(&mnt->mnt_slave, &old->mnt_slave);
|
|
|
|
mnt->mnt_master = old->mnt_master;
|
2016-11-21 01:33:09 +01:00
|
|
|
} else {
|
|
|
|
CLEAR_MNT_SHARED(mnt);
|
2012-06-25 13:55:18 +02:00
|
|
|
}
|
|
|
|
if (flag & CL_MAKE_SHARED)
|
|
|
|
set_mnt_shared(mnt);
|
|
|
|
|
|
|
|
/* stick the duplicate mount on the same expiry list
|
|
|
|
* as the original if that was on one */
|
|
|
|
if (flag & CL_EXPIRE) {
|
|
|
|
if (!list_empty(&old->mnt_expire))
|
|
|
|
list_add(&mnt->mnt_expire, &old->mnt_expire);
|
|
|
|
}
|
|
|
|
|
2011-11-25 02:55:08 +01:00
|
|
|
return mnt;
|
2008-03-27 13:06:23 +01:00
|
|
|
|
|
|
|
out_free:
|
2014-01-24 13:17:54 +01:00
|
|
|
mnt_free_id(mnt);
|
2008-03-27 13:06:23 +01:00
|
|
|
free_vfsmnt(mnt);
|
2012-06-25 13:55:18 +02:00
|
|
|
return ERR_PTR(err);
|
2005-04-17 00:20:36 +02:00
|
|
|
}
|
|
|
|
|
2014-08-08 19:08:20 +02:00
|
|
|
static void cleanup_mnt(struct mount *mnt)
|
|
|
|
{
|
|
|
|
/*
|
|
|
|
* This probably indicates that somebody messed
|
|
|
|
* up a mnt_want/drop_write() pair. If this
|
|
|
|
* happens, the filesystem was probably unable
|
|
|
|
* to make r/w->r/o transitions.
|
|
|
|
*/
|
|
|
|
/*
|
|
|
|
* The locking used to deal with mnt_count decrement provides barriers,
|
|
|
|
* so mnt_get_writers() below is safe.
|
|
|
|
*/
|
|
|
|
WARN_ON(mnt_get_writers(mnt));
|
|
|
|
if (unlikely(mnt->mnt_pins.first))
|
|
|
|
mnt_pin_kill(mnt);
|
|
|
|
fsnotify_vfsmount_delete(&mnt->mnt);
|
|
|
|
dput(mnt->mnt.mnt_root);
|
|
|
|
deactivate_super(mnt->mnt.mnt_sb);
|
|
|
|
mnt_free_id(mnt);
|
|
|
|
call_rcu(&mnt->mnt_rcu, delayed_free_vfsmnt);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void __cleanup_mnt(struct rcu_head *head)
|
|
|
|
{
|
|
|
|
cleanup_mnt(container_of(head, struct mount, mnt_rcu));
|
|
|
|
}
|
|
|
|
|
|
|
|
static LLIST_HEAD(delayed_mntput_list);
|
|
|
|
static void delayed_mntput(struct work_struct *unused)
|
|
|
|
{
|
|
|
|
struct llist_node *node = llist_del_all(&delayed_mntput_list);
|
2017-08-07 10:44:45 +02:00
|
|
|
struct mount *m, *t;
|
2014-08-08 19:08:20 +02:00
|
|
|
|
2017-08-07 10:44:45 +02:00
|
|
|
llist_for_each_entry_safe(m, t, node, mnt_llist)
|
|
|
|
cleanup_mnt(m);
|
2014-08-08 19:08:20 +02:00
|
|
|
}
|
|
|
|
static DECLARE_DELAYED_WORK(delayed_mntput_work, delayed_mntput);
|
|
|
|
|
2011-11-25 06:33:11 +01:00
|
|
|
static void mntput_no_expire(struct mount *mnt)
|
fs: scale mntget/mntput
The problem that this patch aims to fix is vfsmount refcounting scalability.
We need to take a reference on the vfsmount for every successful path lookup,
which often go to the same mount point.
The fundamental difficulty is that a "simple" reference count can never be made
scalable, because any time a reference is dropped, we must check whether that
was the last reference. To do that requires communication with all other CPUs
that may have taken a reference count.
We can make refcounts more scalable in a couple of ways, involving keeping
distributed counters, and checking for the global-zero condition less
frequently.
- check the global sum once every interval (this will delay zero detection
for some interval, so it's probably a showstopper for vfsmounts).
- keep a local count and only taking the global sum when local reaches 0 (this
is difficult for vfsmounts, because we can't hold preempt off for the life of
a reference, so a counter would need to be per-thread or tied strongly to a
particular CPU which requires more locking).
- keep a local difference of increments and decrements, which allows us to sum
the total difference and hence find the refcount when summing all CPUs. Then,
keep a single integer "long" refcount for slow and long lasting references,
and only take the global sum of local counters when the long refcount is 0.
This last scheme is what I implemented here. Attached mounts and process root
and working directory references are "long" references, and everything else is
a short reference.
This allows scalable vfsmount references during path walking over mounted
subtrees and unattached (lazy umounted) mounts with processes still running
in them.
This results in one fewer atomic op in the fastpath: mntget is now just a
per-CPU inc, rather than an atomic inc; and mntput just requires a spinlock
and non-atomic decrement in the common case. However code is otherwise bigger
and heavier, so single threaded performance is basically a wash.
Signed-off-by: Nick Piggin <npiggin@kernel.dk>
2011-01-07 07:50:11 +01:00
|
|
|
{
|
2013-09-30 04:06:07 +02:00
|
|
|
rcu_read_lock();
|
|
|
|
mnt_add_count(mnt, -1);
|
|
|
|
if (likely(mnt->mnt_ns)) { /* shouldn't be the last one */
|
|
|
|
rcu_read_unlock();
|
2011-01-15 04:30:21 +01:00
|
|
|
return;
|
fs: scale mntget/mntput
The problem that this patch aims to fix is vfsmount refcounting scalability.
We need to take a reference on the vfsmount for every successful path lookup,
which often go to the same mount point.
The fundamental difficulty is that a "simple" reference count can never be made
scalable, because any time a reference is dropped, we must check whether that
was the last reference. To do that requires communication with all other CPUs
that may have taken a reference count.
We can make refcounts more scalable in a couple of ways, involving keeping
distributed counters, and checking for the global-zero condition less
frequently.
- check the global sum once every interval (this will delay zero detection
for some interval, so it's probably a showstopper for vfsmounts).
- keep a local count and only taking the global sum when local reaches 0 (this
is difficult for vfsmounts, because we can't hold preempt off for the life of
a reference, so a counter would need to be per-thread or tied strongly to a
particular CPU which requires more locking).
- keep a local difference of increments and decrements, which allows us to sum
the total difference and hence find the refcount when summing all CPUs. Then,
keep a single integer "long" refcount for slow and long lasting references,
and only take the global sum of local counters when the long refcount is 0.
This last scheme is what I implemented here. Attached mounts and process root
and working directory references are "long" references, and everything else is
a short reference.
This allows scalable vfsmount references during path walking over mounted
subtrees and unattached (lazy umounted) mounts with processes still running
in them.
This results in one fewer atomic op in the fastpath: mntget is now just a
per-CPU inc, rather than an atomic inc; and mntput just requires a spinlock
and non-atomic decrement in the common case. However code is otherwise bigger
and heavier, so single threaded performance is basically a wash.
Signed-off-by: Nick Piggin <npiggin@kernel.dk>
2011-01-07 07:50:11 +01:00
|
|
|
}
|
2013-09-29 17:24:49 +02:00
|
|
|
lock_mount_hash();
|
fs: scale mntget/mntput
The problem that this patch aims to fix is vfsmount refcounting scalability.
We need to take a reference on the vfsmount for every successful path lookup,
which often go to the same mount point.
The fundamental difficulty is that a "simple" reference count can never be made
scalable, because any time a reference is dropped, we must check whether that
was the last reference. To do that requires communication with all other CPUs
that may have taken a reference count.
We can make refcounts more scalable in a couple of ways, involving keeping
distributed counters, and checking for the global-zero condition less
frequently.
- check the global sum once every interval (this will delay zero detection
for some interval, so it's probably a showstopper for vfsmounts).
- keep a local count and only taking the global sum when local reaches 0 (this
is difficult for vfsmounts, because we can't hold preempt off for the life of
a reference, so a counter would need to be per-thread or tied strongly to a
particular CPU which requires more locking).
- keep a local difference of increments and decrements, which allows us to sum
the total difference and hence find the refcount when summing all CPUs. Then,
keep a single integer "long" refcount for slow and long lasting references,
and only take the global sum of local counters when the long refcount is 0.
This last scheme is what I implemented here. Attached mounts and process root
and working directory references are "long" references, and everything else is
a short reference.
This allows scalable vfsmount references during path walking over mounted
subtrees and unattached (lazy umounted) mounts with processes still running
in them.
This results in one fewer atomic op in the fastpath: mntget is now just a
per-CPU inc, rather than an atomic inc; and mntput just requires a spinlock
and non-atomic decrement in the common case. However code is otherwise bigger
and heavier, so single threaded performance is basically a wash.
Signed-off-by: Nick Piggin <npiggin@kernel.dk>
2011-01-07 07:50:11 +01:00
|
|
|
if (mnt_get_count(mnt)) {
|
2013-09-30 04:06:07 +02:00
|
|
|
rcu_read_unlock();
|
2013-09-29 17:24:49 +02:00
|
|
|
unlock_mount_hash();
|
fs: brlock vfsmount_lock
fs: brlock vfsmount_lock
Use a brlock for the vfsmount lock. It must be taken for write whenever
modifying the mount hash or associated fields, and may be taken for read when
performing mount hash lookups.
A new lock is added for the mnt-id allocator, so it doesn't need to take
the heavy vfsmount write-lock.
The number of atomics should remain the same for fastpath rlock cases, though
code would be slightly slower due to per-cpu access. Scalability is not not be
much improved in common cases yet, due to other locks (ie. dcache_lock) getting
in the way. However path lookups crossing mountpoints should be one case where
scalability is improved (currently requiring the global lock).
The slowpath is slower due to use of brlock. On a 64 core, 64 socket, 32 node
Altix system (high latency to remote nodes), a simple umount microbenchmark
(mount --bind mnt mnt2 ; umount mnt2 loop 1000 times), before this patch it
took 6.8s, afterwards took 7.1s, about 5% slower.
Cc: Al Viro <viro@ZenIV.linux.org.uk>
Signed-off-by: Nick Piggin <npiggin@kernel.dk>
Signed-off-by: Al Viro <viro@zeniv.linux.org.uk>
2010-08-17 20:37:39 +02:00
|
|
|
return;
|
|
|
|
}
|
2013-09-30 04:06:07 +02:00
|
|
|
if (unlikely(mnt->mnt.mnt_flags & MNT_DOOMED)) {
|
|
|
|
rcu_read_unlock();
|
|
|
|
unlock_mount_hash();
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
mnt->mnt.mnt_flags |= MNT_DOOMED;
|
|
|
|
rcu_read_unlock();
|
2012-05-08 06:02:02 +02:00
|
|
|
|
2011-11-21 12:11:30 +01:00
|
|
|
list_del(&mnt->mnt_instance);
|
2014-12-24 04:37:03 +01:00
|
|
|
|
|
|
|
if (unlikely(!list_empty(&mnt->mnt_mounts))) {
|
|
|
|
struct mount *p, *tmp;
|
|
|
|
list_for_each_entry_safe(p, tmp, &mnt->mnt_mounts, mnt_child) {
|
|
|
|
umount_mnt(p);
|
|
|
|
}
|
|
|
|
}
|
2013-09-29 17:24:49 +02:00
|
|
|
unlock_mount_hash();
|
2013-09-28 18:41:25 +02:00
|
|
|
|
2014-08-08 19:08:20 +02:00
|
|
|
if (likely(!(mnt->mnt.mnt_flags & MNT_INTERNAL))) {
|
|
|
|
struct task_struct *task = current;
|
|
|
|
if (likely(!(task->flags & PF_KTHREAD))) {
|
|
|
|
init_task_work(&mnt->mnt_rcu, __cleanup_mnt);
|
|
|
|
if (!task_work_add(task, &mnt->mnt_rcu, true))
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
if (llist_add(&mnt->mnt_llist, &delayed_mntput_list))
|
|
|
|
schedule_delayed_work(&delayed_mntput_work, 1);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
cleanup_mnt(mnt);
|
fs: scale mntget/mntput
The problem that this patch aims to fix is vfsmount refcounting scalability.
We need to take a reference on the vfsmount for every successful path lookup,
which often go to the same mount point.
The fundamental difficulty is that a "simple" reference count can never be made
scalable, because any time a reference is dropped, we must check whether that
was the last reference. To do that requires communication with all other CPUs
that may have taken a reference count.
We can make refcounts more scalable in a couple of ways, involving keeping
distributed counters, and checking for the global-zero condition less
frequently.
- check the global sum once every interval (this will delay zero detection
for some interval, so it's probably a showstopper for vfsmounts).
- keep a local count and only taking the global sum when local reaches 0 (this
is difficult for vfsmounts, because we can't hold preempt off for the life of
a reference, so a counter would need to be per-thread or tied strongly to a
particular CPU which requires more locking).
- keep a local difference of increments and decrements, which allows us to sum
the total difference and hence find the refcount when summing all CPUs. Then,
keep a single integer "long" refcount for slow and long lasting references,
and only take the global sum of local counters when the long refcount is 0.
This last scheme is what I implemented here. Attached mounts and process root
and working directory references are "long" references, and everything else is
a short reference.
This allows scalable vfsmount references during path walking over mounted
subtrees and unattached (lazy umounted) mounts with processes still running
in them.
This results in one fewer atomic op in the fastpath: mntget is now just a
per-CPU inc, rather than an atomic inc; and mntput just requires a spinlock
and non-atomic decrement in the common case. However code is otherwise bigger
and heavier, so single threaded performance is basically a wash.
Signed-off-by: Nick Piggin <npiggin@kernel.dk>
2011-01-07 07:50:11 +01:00
|
|
|
}
|
|
|
|
|
|
|
|
void mntput(struct vfsmount *mnt)
|
|
|
|
{
|
|
|
|
if (mnt) {
|
2011-11-25 06:57:42 +01:00
|
|
|
struct mount *m = real_mount(mnt);
|
fs: scale mntget/mntput
The problem that this patch aims to fix is vfsmount refcounting scalability.
We need to take a reference on the vfsmount for every successful path lookup,
which often go to the same mount point.
The fundamental difficulty is that a "simple" reference count can never be made
scalable, because any time a reference is dropped, we must check whether that
was the last reference. To do that requires communication with all other CPUs
that may have taken a reference count.
We can make refcounts more scalable in a couple of ways, involving keeping
distributed counters, and checking for the global-zero condition less
frequently.
- check the global sum once every interval (this will delay zero detection
for some interval, so it's probably a showstopper for vfsmounts).
- keep a local count and only taking the global sum when local reaches 0 (this
is difficult for vfsmounts, because we can't hold preempt off for the life of
a reference, so a counter would need to be per-thread or tied strongly to a
particular CPU which requires more locking).
- keep a local difference of increments and decrements, which allows us to sum
the total difference and hence find the refcount when summing all CPUs. Then,
keep a single integer "long" refcount for slow and long lasting references,
and only take the global sum of local counters when the long refcount is 0.
This last scheme is what I implemented here. Attached mounts and process root
and working directory references are "long" references, and everything else is
a short reference.
This allows scalable vfsmount references during path walking over mounted
subtrees and unattached (lazy umounted) mounts with processes still running
in them.
This results in one fewer atomic op in the fastpath: mntget is now just a
per-CPU inc, rather than an atomic inc; and mntput just requires a spinlock
and non-atomic decrement in the common case. However code is otherwise bigger
and heavier, so single threaded performance is basically a wash.
Signed-off-by: Nick Piggin <npiggin@kernel.dk>
2011-01-07 07:50:11 +01:00
|
|
|
/* avoid cacheline pingpong, hope gcc doesn't get "smart" */
|
2011-11-25 06:57:42 +01:00
|
|
|
if (unlikely(m->mnt_expiry_mark))
|
|
|
|
m->mnt_expiry_mark = 0;
|
|
|
|
mntput_no_expire(m);
|
fs: scale mntget/mntput
The problem that this patch aims to fix is vfsmount refcounting scalability.
We need to take a reference on the vfsmount for every successful path lookup,
which often go to the same mount point.
The fundamental difficulty is that a "simple" reference count can never be made
scalable, because any time a reference is dropped, we must check whether that
was the last reference. To do that requires communication with all other CPUs
that may have taken a reference count.
We can make refcounts more scalable in a couple of ways, involving keeping
distributed counters, and checking for the global-zero condition less
frequently.
- check the global sum once every interval (this will delay zero detection
for some interval, so it's probably a showstopper for vfsmounts).
- keep a local count and only taking the global sum when local reaches 0 (this
is difficult for vfsmounts, because we can't hold preempt off for the life of
a reference, so a counter would need to be per-thread or tied strongly to a
particular CPU which requires more locking).
- keep a local difference of increments and decrements, which allows us to sum
the total difference and hence find the refcount when summing all CPUs. Then,
keep a single integer "long" refcount for slow and long lasting references,
and only take the global sum of local counters when the long refcount is 0.
This last scheme is what I implemented here. Attached mounts and process root
and working directory references are "long" references, and everything else is
a short reference.
This allows scalable vfsmount references during path walking over mounted
subtrees and unattached (lazy umounted) mounts with processes still running
in them.
This results in one fewer atomic op in the fastpath: mntget is now just a
per-CPU inc, rather than an atomic inc; and mntput just requires a spinlock
and non-atomic decrement in the common case. However code is otherwise bigger
and heavier, so single threaded performance is basically a wash.
Signed-off-by: Nick Piggin <npiggin@kernel.dk>
2011-01-07 07:50:11 +01:00
|
|
|
}
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL(mntput);
|
|
|
|
|
|
|
|
struct vfsmount *mntget(struct vfsmount *mnt)
|
|
|
|
{
|
|
|
|
if (mnt)
|
2011-11-25 04:37:54 +01:00
|
|
|
mnt_add_count(real_mount(mnt), 1);
|
fs: scale mntget/mntput
The problem that this patch aims to fix is vfsmount refcounting scalability.
We need to take a reference on the vfsmount for every successful path lookup,
which often go to the same mount point.
The fundamental difficulty is that a "simple" reference count can never be made
scalable, because any time a reference is dropped, we must check whether that
was the last reference. To do that requires communication with all other CPUs
that may have taken a reference count.
We can make refcounts more scalable in a couple of ways, involving keeping
distributed counters, and checking for the global-zero condition less
frequently.
- check the global sum once every interval (this will delay zero detection
for some interval, so it's probably a showstopper for vfsmounts).
- keep a local count and only taking the global sum when local reaches 0 (this
is difficult for vfsmounts, because we can't hold preempt off for the life of
a reference, so a counter would need to be per-thread or tied strongly to a
particular CPU which requires more locking).
- keep a local difference of increments and decrements, which allows us to sum
the total difference and hence find the refcount when summing all CPUs. Then,
keep a single integer "long" refcount for slow and long lasting references,
and only take the global sum of local counters when the long refcount is 0.
This last scheme is what I implemented here. Attached mounts and process root
and working directory references are "long" references, and everything else is
a short reference.
This allows scalable vfsmount references during path walking over mounted
subtrees and unattached (lazy umounted) mounts with processes still running
in them.
This results in one fewer atomic op in the fastpath: mntget is now just a
per-CPU inc, rather than an atomic inc; and mntput just requires a spinlock
and non-atomic decrement in the common case. However code is otherwise bigger
and heavier, so single threaded performance is basically a wash.
Signed-off-by: Nick Piggin <npiggin@kernel.dk>
2011-01-07 07:50:11 +01:00
|
|
|
return mnt;
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL(mntget);
|
|
|
|
|
2016-11-23 22:03:41 +01:00
|
|
|
/* path_is_mountpoint() - Check if path is a mount in the current
|
|
|
|
* namespace.
|
|
|
|
*
|
|
|
|
* d_mountpoint() can only be used reliably to establish if a dentry is
|
|
|
|
* not mounted in any namespace and that common case is handled inline.
|
|
|
|
* d_mountpoint() isn't aware of the possibility there may be multiple
|
|
|
|
* mounts using a given dentry in a different namespace. This function
|
|
|
|
* checks if the passed in path is a mountpoint rather than the dentry
|
|
|
|
* alone.
|
|
|
|
*/
|
|
|
|
bool path_is_mountpoint(const struct path *path)
|
|
|
|
{
|
|
|
|
unsigned seq;
|
|
|
|
bool res;
|
|
|
|
|
|
|
|
if (!d_mountpoint(path->dentry))
|
|
|
|
return false;
|
|
|
|
|
|
|
|
rcu_read_lock();
|
|
|
|
do {
|
|
|
|
seq = read_seqbegin(&mount_lock);
|
|
|
|
res = __path_is_mountpoint(path);
|
|
|
|
} while (read_seqretry(&mount_lock, seq));
|
|
|
|
rcu_read_unlock();
|
|
|
|
|
|
|
|
return res;
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL(path_is_mountpoint);
|
|
|
|
|
2016-11-21 01:45:28 +01:00
|
|
|
struct vfsmount *mnt_clone_internal(const struct path *path)
|
2005-11-07 23:13:39 +01:00
|
|
|
{
|
2014-08-07 15:12:31 +02:00
|
|
|
struct mount *p;
|
|
|
|
p = clone_mnt(real_mount(path->mnt), path->dentry, CL_PRIVATE);
|
|
|
|
if (IS_ERR(p))
|
|
|
|
return ERR_CAST(p);
|
|
|
|
p->mnt.mnt_flags |= MNT_INTERNAL;
|
|
|
|
return &p->mnt;
|
2005-11-07 23:13:39 +01:00
|
|
|
}
|
2005-04-17 00:20:36 +02:00
|
|
|
|
2008-03-27 13:06:24 +01:00
|
|
|
#ifdef CONFIG_PROC_FS
|
2011-12-06 18:21:54 +01:00
|
|
|
/* iterator; we want it to have access to namespace_sem, thus here... */
|
2005-04-17 00:20:36 +02:00
|
|
|
static void *m_start(struct seq_file *m, loff_t *pos)
|
|
|
|
{
|
fs: use seq_open_private() for proc_mounts
A patchset to remove support for passing pre-allocated struct seq_file to
seq_open(). Such feature is undocumented and prone to error.
In particular, if seq_release() is used in release handler, it will
kfree() a pointer which was not allocated by seq_open().
So this patchset drops support for pre-allocated struct seq_file: it's
only of use in proc_namespace.c and can be easily replaced by using
seq_open_private()/seq_release_private().
Additionally, it documents the use of file->private_data to hold pointer
to struct seq_file by seq_open().
This patch (of 3):
Since patch described below, from v2.6.15-rc1, seq_open() could use a
struct seq_file already allocated by the caller if the pointer to the
structure is stored in file->private_data before calling the function.
Commit 1abe77b0fc4b485927f1f798ae81a752677e1d05
Author: Al Viro <viro@zeniv.linux.org.uk>
Date: Mon Nov 7 17:15:34 2005 -0500
[PATCH] allow callers of seq_open do allocation themselves
Allow caller of seq_open() to kmalloc() seq_file + whatever else they
want and set ->private_data to it. seq_open() will then abstain from
doing allocation itself.
Such behavior is only used by mounts_open_common().
In order to drop support for such uncommon feature, proc_mounts is
converted to use seq_open_private(), which take care of allocating the
proc_mounts structure, making it available through ->private in struct
seq_file.
Conversely, proc_mounts is converted to use seq_release_private(), in
order to release the private structure allocated by seq_open_private().
Then, ->private is used directly instead of proc_mounts() macro to access
to the proc_mounts structure.
Link: http://lkml.kernel.org/r/cover.1433193673.git.ydroneaud@opteya.com
Signed-off-by: Yann Droneaud <ydroneaud@opteya.com>
Cc: Al Viro <viro@zeniv.linux.org.uk>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2015-06-30 23:57:30 +02:00
|
|
|
struct proc_mounts *p = m->private;
|
2005-04-17 00:20:36 +02:00
|
|
|
|
2005-11-07 23:17:51 +01:00
|
|
|
down_read(&namespace_sem);
|
2014-02-27 20:40:10 +01:00
|
|
|
if (p->cached_event == p->ns->event) {
|
|
|
|
void *v = p->cached_mount;
|
|
|
|
if (*pos == p->cached_index)
|
|
|
|
return v;
|
|
|
|
if (*pos == p->cached_index + 1) {
|
|
|
|
v = seq_list_next(v, &p->ns->list, &p->cached_index);
|
|
|
|
return p->cached_mount = v;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
p->cached_event = p->ns->event;
|
|
|
|
p->cached_mount = seq_list_start(&p->ns->list, *pos);
|
|
|
|
p->cached_index = *pos;
|
|
|
|
return p->cached_mount;
|
2005-04-17 00:20:36 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
static void *m_next(struct seq_file *m, void *v, loff_t *pos)
|
|
|
|
{
|
fs: use seq_open_private() for proc_mounts
A patchset to remove support for passing pre-allocated struct seq_file to
seq_open(). Such feature is undocumented and prone to error.
In particular, if seq_release() is used in release handler, it will
kfree() a pointer which was not allocated by seq_open().
So this patchset drops support for pre-allocated struct seq_file: it's
only of use in proc_namespace.c and can be easily replaced by using
seq_open_private()/seq_release_private().
Additionally, it documents the use of file->private_data to hold pointer
to struct seq_file by seq_open().
This patch (of 3):
Since patch described below, from v2.6.15-rc1, seq_open() could use a
struct seq_file already allocated by the caller if the pointer to the
structure is stored in file->private_data before calling the function.
Commit 1abe77b0fc4b485927f1f798ae81a752677e1d05
Author: Al Viro <viro@zeniv.linux.org.uk>
Date: Mon Nov 7 17:15:34 2005 -0500
[PATCH] allow callers of seq_open do allocation themselves
Allow caller of seq_open() to kmalloc() seq_file + whatever else they
want and set ->private_data to it. seq_open() will then abstain from
doing allocation itself.
Such behavior is only used by mounts_open_common().
In order to drop support for such uncommon feature, proc_mounts is
converted to use seq_open_private(), which take care of allocating the
proc_mounts structure, making it available through ->private in struct
seq_file.
Conversely, proc_mounts is converted to use seq_release_private(), in
order to release the private structure allocated by seq_open_private().
Then, ->private is used directly instead of proc_mounts() macro to access
to the proc_mounts structure.
Link: http://lkml.kernel.org/r/cover.1433193673.git.ydroneaud@opteya.com
Signed-off-by: Yann Droneaud <ydroneaud@opteya.com>
Cc: Al Viro <viro@zeniv.linux.org.uk>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2015-06-30 23:57:30 +02:00
|
|
|
struct proc_mounts *p = m->private;
|
2007-07-16 08:39:55 +02:00
|
|
|
|
2014-02-27 20:40:10 +01:00
|
|
|
p->cached_mount = seq_list_next(v, &p->ns->list, pos);
|
|
|
|
p->cached_index = *pos;
|
|
|
|
return p->cached_mount;
|
2005-04-17 00:20:36 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
static void m_stop(struct seq_file *m, void *v)
|
|
|
|
{
|
2005-11-07 23:17:51 +01:00
|
|
|
up_read(&namespace_sem);
|
2005-04-17 00:20:36 +02:00
|
|
|
}
|
|
|
|
|
2011-12-06 18:21:54 +01:00
|
|
|
static int m_show(struct seq_file *m, void *v)
|
2008-03-27 13:06:25 +01:00
|
|
|
{
|
fs: use seq_open_private() for proc_mounts
A patchset to remove support for passing pre-allocated struct seq_file to
seq_open(). Such feature is undocumented and prone to error.
In particular, if seq_release() is used in release handler, it will
kfree() a pointer which was not allocated by seq_open().
So this patchset drops support for pre-allocated struct seq_file: it's
only of use in proc_namespace.c and can be easily replaced by using
seq_open_private()/seq_release_private().
Additionally, it documents the use of file->private_data to hold pointer
to struct seq_file by seq_open().
This patch (of 3):
Since patch described below, from v2.6.15-rc1, seq_open() could use a
struct seq_file already allocated by the caller if the pointer to the
structure is stored in file->private_data before calling the function.
Commit 1abe77b0fc4b485927f1f798ae81a752677e1d05
Author: Al Viro <viro@zeniv.linux.org.uk>
Date: Mon Nov 7 17:15:34 2005 -0500
[PATCH] allow callers of seq_open do allocation themselves
Allow caller of seq_open() to kmalloc() seq_file + whatever else they
want and set ->private_data to it. seq_open() will then abstain from
doing allocation itself.
Such behavior is only used by mounts_open_common().
In order to drop support for such uncommon feature, proc_mounts is
converted to use seq_open_private(), which take care of allocating the
proc_mounts structure, making it available through ->private in struct
seq_file.
Conversely, proc_mounts is converted to use seq_release_private(), in
order to release the private structure allocated by seq_open_private().
Then, ->private is used directly instead of proc_mounts() macro to access
to the proc_mounts structure.
Link: http://lkml.kernel.org/r/cover.1433193673.git.ydroneaud@opteya.com
Signed-off-by: Yann Droneaud <ydroneaud@opteya.com>
Cc: Al Viro <viro@zeniv.linux.org.uk>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2015-06-30 23:57:30 +02:00
|
|
|
struct proc_mounts *p = m->private;
|
2011-11-25 08:19:55 +01:00
|
|
|
struct mount *r = list_entry(v, struct mount, mnt_list);
|
2011-12-06 18:21:54 +01:00
|
|
|
return p->show(m, &r->mnt);
|
2005-04-17 00:20:36 +02:00
|
|
|
}
|
|
|
|
|
2008-03-27 13:06:24 +01:00
|
|
|
const struct seq_operations mounts_op = {
|
2005-04-17 00:20:36 +02:00
|
|
|
.start = m_start,
|
|
|
|
.next = m_next,
|
|
|
|
.stop = m_stop,
|
2011-12-06 18:21:54 +01:00
|
|
|
.show = m_show,
|
2006-03-20 19:44:12 +01:00
|
|
|
};
|
2008-03-27 13:06:24 +01:00
|
|
|
#endif /* CONFIG_PROC_FS */
|
2006-03-20 19:44:12 +01:00
|
|
|
|
2005-04-17 00:20:36 +02:00
|
|
|
/**
|
|
|
|
* may_umount_tree - check if a mount tree is busy
|
|
|
|
* @mnt: root of mount tree
|
|
|
|
*
|
|
|
|
* This is called to check if a tree of mounts has any
|
|
|
|
* open files, pwds, chroots or sub mounts that are
|
|
|
|
* busy.
|
|
|
|
*/
|
2011-11-25 09:06:56 +01:00
|
|
|
int may_umount_tree(struct vfsmount *m)
|
2005-04-17 00:20:36 +02:00
|
|
|
{
|
2011-11-25 09:06:56 +01:00
|
|
|
struct mount *mnt = real_mount(m);
|
2005-11-07 23:17:22 +01:00
|
|
|
int actual_refs = 0;
|
|
|
|
int minimum_refs = 0;
|
2011-11-25 00:57:30 +01:00
|
|
|
struct mount *p;
|
2011-11-25 09:06:56 +01:00
|
|
|
BUG_ON(!m);
|
2005-04-17 00:20:36 +02:00
|
|
|
|
fs: scale mntget/mntput
The problem that this patch aims to fix is vfsmount refcounting scalability.
We need to take a reference on the vfsmount for every successful path lookup,
which often go to the same mount point.
The fundamental difficulty is that a "simple" reference count can never be made
scalable, because any time a reference is dropped, we must check whether that
was the last reference. To do that requires communication with all other CPUs
that may have taken a reference count.
We can make refcounts more scalable in a couple of ways, involving keeping
distributed counters, and checking for the global-zero condition less
frequently.
- check the global sum once every interval (this will delay zero detection
for some interval, so it's probably a showstopper for vfsmounts).
- keep a local count and only taking the global sum when local reaches 0 (this
is difficult for vfsmounts, because we can't hold preempt off for the life of
a reference, so a counter would need to be per-thread or tied strongly to a
particular CPU which requires more locking).
- keep a local difference of increments and decrements, which allows us to sum
the total difference and hence find the refcount when summing all CPUs. Then,
keep a single integer "long" refcount for slow and long lasting references,
and only take the global sum of local counters when the long refcount is 0.
This last scheme is what I implemented here. Attached mounts and process root
and working directory references are "long" references, and everything else is
a short reference.
This allows scalable vfsmount references during path walking over mounted
subtrees and unattached (lazy umounted) mounts with processes still running
in them.
This results in one fewer atomic op in the fastpath: mntget is now just a
per-CPU inc, rather than an atomic inc; and mntput just requires a spinlock
and non-atomic decrement in the common case. However code is otherwise bigger
and heavier, so single threaded performance is basically a wash.
Signed-off-by: Nick Piggin <npiggin@kernel.dk>
2011-01-07 07:50:11 +01:00
|
|
|
/* write lock needed for mnt_get_count */
|
2013-09-29 17:24:49 +02:00
|
|
|
lock_mount_hash();
|
2011-11-25 09:06:56 +01:00
|
|
|
for (p = mnt; p; p = next_mnt(p, mnt)) {
|
2011-11-25 04:37:54 +01:00
|
|
|
actual_refs += mnt_get_count(p);
|
2005-04-17 00:20:36 +02:00
|
|
|
minimum_refs += 2;
|
|
|
|
}
|
2013-09-29 17:24:49 +02:00
|
|
|
unlock_mount_hash();
|
2005-04-17 00:20:36 +02:00
|
|
|
|
|
|
|
if (actual_refs > minimum_refs)
|
2006-03-27 11:14:51 +02:00
|
|
|
return 0;
|
2005-04-17 00:20:36 +02:00
|
|
|
|
2006-03-27 11:14:51 +02:00
|
|
|
return 1;
|
2005-04-17 00:20:36 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
EXPORT_SYMBOL(may_umount_tree);
|
|
|
|
|
|
|
|
/**
|
|
|
|
* may_umount - check if a mount point is busy
|
|
|
|
* @mnt: root of mount
|
|
|
|
*
|
|
|
|
* This is called to check if a mount point has any
|
|
|
|
* open files, pwds, chroots or sub mounts. If the
|
|
|
|
* mount has sub mounts this will return busy
|
|
|
|
* regardless of whether the sub mounts are busy.
|
|
|
|
*
|
|
|
|
* Doesn't take quota and stuff into account. IOW, in some cases it will
|
|
|
|
* give false negatives. The main reason why it's here is that we need
|
|
|
|
* a non-destructive way to look for easily umountable filesystems.
|
|
|
|
*/
|
|
|
|
int may_umount(struct vfsmount *mnt)
|
|
|
|
{
|
2006-03-27 11:14:51 +02:00
|
|
|
int ret = 1;
|
2010-01-16 18:56:08 +01:00
|
|
|
down_read(&namespace_sem);
|
2013-09-29 17:24:49 +02:00
|
|
|
lock_mount_hash();
|
2011-11-25 03:35:16 +01:00
|
|
|
if (propagate_mount_busy(real_mount(mnt), 2))
|
2006-03-27 11:14:51 +02:00
|
|
|
ret = 0;
|
2013-09-29 17:24:49 +02:00
|
|
|
unlock_mount_hash();
|
2010-01-16 18:56:08 +01:00
|
|
|
up_read(&namespace_sem);
|
2005-11-07 23:20:17 +01:00
|
|
|
return ret;
|
2005-04-17 00:20:36 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
EXPORT_SYMBOL(may_umount);
|
|
|
|
|
2014-03-21 02:10:51 +01:00
|
|
|
static HLIST_HEAD(unmounted); /* protected by namespace_sem */
|
2013-03-16 19:35:16 +01:00
|
|
|
|
2013-03-16 20:12:40 +01:00
|
|
|
static void namespace_unlock(void)
|
2005-11-07 23:17:04 +01:00
|
|
|
{
|
2015-04-03 03:33:53 +02:00
|
|
|
struct hlist_head head;
|
2013-03-16 20:12:40 +01:00
|
|
|
|
2015-04-03 03:33:53 +02:00
|
|
|
hlist_move_list(&unmounted, &head);
|
2013-03-16 20:12:40 +01:00
|
|
|
|
|
|
|
up_write(&namespace_sem);
|
|
|
|
|
2015-04-03 03:33:53 +02:00
|
|
|
if (likely(hlist_empty(&head)))
|
|
|
|
return;
|
|
|
|
|
2013-09-30 04:06:07 +02:00
|
|
|
synchronize_rcu();
|
|
|
|
|
2015-01-11 01:01:08 +01:00
|
|
|
group_pin_kill(&head);
|
2005-11-07 23:17:04 +01:00
|
|
|
}
|
|
|
|
|
2013-03-16 20:12:40 +01:00
|
|
|
static inline void namespace_lock(void)
|
2013-03-16 19:35:16 +01:00
|
|
|
{
|
2013-03-16 20:12:40 +01:00
|
|
|
down_write(&namespace_sem);
|
2013-03-16 19:35:16 +01:00
|
|
|
}
|
|
|
|
|
2014-12-24 14:20:01 +01:00
|
|
|
enum umount_tree_flags {
|
|
|
|
UMOUNT_SYNC = 1,
|
|
|
|
UMOUNT_PROPAGATE = 2,
|
2015-04-02 01:30:06 +02:00
|
|
|
UMOUNT_CONNECTED = 4,
|
2014-12-24 14:20:01 +01:00
|
|
|
};
|
2015-07-17 21:15:30 +02:00
|
|
|
|
|
|
|
static bool disconnect_mount(struct mount *mnt, enum umount_tree_flags how)
|
|
|
|
{
|
|
|
|
/* Leaving mounts connected is only valid for lazy umounts */
|
|
|
|
if (how & UMOUNT_SYNC)
|
|
|
|
return true;
|
|
|
|
|
|
|
|
/* A mount without a parent has nothing to be connected to */
|
|
|
|
if (!mnt_has_parent(mnt))
|
|
|
|
return true;
|
|
|
|
|
|
|
|
/* Because the reference counting rules change when mounts are
|
|
|
|
* unmounted and connected, umounted mounts may not be
|
|
|
|
* connected to mounted mounts.
|
|
|
|
*/
|
|
|
|
if (!(mnt->mnt_parent->mnt.mnt_flags & MNT_UMOUNT))
|
|
|
|
return true;
|
|
|
|
|
|
|
|
/* Has it been requested that the mount remain connected? */
|
|
|
|
if (how & UMOUNT_CONNECTED)
|
|
|
|
return false;
|
|
|
|
|
|
|
|
/* Is the mount locked such that it needs to remain connected? */
|
|
|
|
if (IS_MNT_LOCKED(mnt))
|
|
|
|
return false;
|
|
|
|
|
|
|
|
/* By default disconnect the mount */
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
fs: brlock vfsmount_lock
fs: brlock vfsmount_lock
Use a brlock for the vfsmount lock. It must be taken for write whenever
modifying the mount hash or associated fields, and may be taken for read when
performing mount hash lookups.
A new lock is added for the mnt-id allocator, so it doesn't need to take
the heavy vfsmount write-lock.
The number of atomics should remain the same for fastpath rlock cases, though
code would be slightly slower due to per-cpu access. Scalability is not not be
much improved in common cases yet, due to other locks (ie. dcache_lock) getting
in the way. However path lookups crossing mountpoints should be one case where
scalability is improved (currently requiring the global lock).
The slowpath is slower due to use of brlock. On a 64 core, 64 socket, 32 node
Altix system (high latency to remote nodes), a simple umount microbenchmark
(mount --bind mnt mnt2 ; umount mnt2 loop 1000 times), before this patch it
took 6.8s, afterwards took 7.1s, about 5% slower.
Cc: Al Viro <viro@ZenIV.linux.org.uk>
Signed-off-by: Nick Piggin <npiggin@kernel.dk>
Signed-off-by: Al Viro <viro@zeniv.linux.org.uk>
2010-08-17 20:37:39 +02:00
|
|
|
/*
|
2013-09-30 04:06:07 +02:00
|
|
|
* mount_lock must be held
|
fs: brlock vfsmount_lock
fs: brlock vfsmount_lock
Use a brlock for the vfsmount lock. It must be taken for write whenever
modifying the mount hash or associated fields, and may be taken for read when
performing mount hash lookups.
A new lock is added for the mnt-id allocator, so it doesn't need to take
the heavy vfsmount write-lock.
The number of atomics should remain the same for fastpath rlock cases, though
code would be slightly slower due to per-cpu access. Scalability is not not be
much improved in common cases yet, due to other locks (ie. dcache_lock) getting
in the way. However path lookups crossing mountpoints should be one case where
scalability is improved (currently requiring the global lock).
The slowpath is slower due to use of brlock. On a 64 core, 64 socket, 32 node
Altix system (high latency to remote nodes), a simple umount microbenchmark
(mount --bind mnt mnt2 ; umount mnt2 loop 1000 times), before this patch it
took 6.8s, afterwards took 7.1s, about 5% slower.
Cc: Al Viro <viro@ZenIV.linux.org.uk>
Signed-off-by: Nick Piggin <npiggin@kernel.dk>
Signed-off-by: Al Viro <viro@zeniv.linux.org.uk>
2010-08-17 20:37:39 +02:00
|
|
|
* namespace_sem must be held for write
|
|
|
|
*/
|
2014-12-24 14:20:01 +01:00
|
|
|
static void umount_tree(struct mount *mnt, enum umount_tree_flags how)
|
2005-04-17 00:20:36 +02:00
|
|
|
{
|
2014-12-18 20:10:48 +01:00
|
|
|
LIST_HEAD(tmp_list);
|
2011-11-25 00:57:30 +01:00
|
|
|
struct mount *p;
|
2005-04-17 00:20:36 +02:00
|
|
|
|
2015-01-03 12:39:35 +01:00
|
|
|
if (how & UMOUNT_PROPAGATE)
|
|
|
|
propagate_mount_unlock(mnt);
|
|
|
|
|
2014-12-18 20:10:48 +01:00
|
|
|
/* Gather the mounts to umount */
|
2014-12-23 01:30:08 +01:00
|
|
|
for (p = mnt; p; p = next_mnt(p, mnt)) {
|
|
|
|
p->mnt.mnt_flags |= MNT_UMOUNT;
|
2014-12-18 20:10:48 +01:00
|
|
|
list_move(&p->mnt_list, &tmp_list);
|
2014-12-23 01:30:08 +01:00
|
|
|
}
|
2005-04-17 00:20:36 +02:00
|
|
|
|
2014-12-23 02:12:07 +01:00
|
|
|
/* Hide the mounts from mnt_mounts */
|
2014-12-18 20:10:48 +01:00
|
|
|
list_for_each_entry(p, &tmp_list, mnt_list) {
|
2014-08-18 21:09:26 +02:00
|
|
|
list_del_init(&p->mnt_child);
|
2014-12-18 20:10:48 +01:00
|
|
|
}
|
2014-08-18 21:09:26 +02:00
|
|
|
|
2014-12-18 20:10:48 +01:00
|
|
|
/* Add propogated mounts to the tmp_list */
|
2014-12-24 14:20:01 +01:00
|
|
|
if (how & UMOUNT_PROPAGATE)
|
2011-01-16 02:08:44 +01:00
|
|
|
propagate_umount(&tmp_list);
|
2005-11-07 23:20:17 +01:00
|
|
|
|
2014-12-18 20:10:48 +01:00
|
|
|
while (!list_empty(&tmp_list)) {
|
2016-09-28 07:27:17 +02:00
|
|
|
struct mnt_namespace *ns;
|
2014-12-24 04:37:03 +01:00
|
|
|
bool disconnect;
|
2014-12-18 20:10:48 +01:00
|
|
|
p = list_first_entry(&tmp_list, struct mount, mnt_list);
|
2011-11-25 06:22:05 +01:00
|
|
|
list_del_init(&p->mnt_expire);
|
2011-11-25 08:19:55 +01:00
|
|
|
list_del_init(&p->mnt_list);
|
2016-09-28 07:27:17 +02:00
|
|
|
ns = p->mnt_ns;
|
|
|
|
if (ns) {
|
|
|
|
ns->mounts--;
|
|
|
|
__touch_mnt_namespace(ns);
|
|
|
|
}
|
2011-11-25 06:46:35 +01:00
|
|
|
p->mnt_ns = NULL;
|
2014-12-24 14:20:01 +01:00
|
|
|
if (how & UMOUNT_SYNC)
|
2013-09-30 04:06:07 +02:00
|
|
|
p->mnt.mnt_flags |= MNT_SYNC_UMOUNT;
|
2015-01-11 01:01:08 +01:00
|
|
|
|
2015-07-17 21:15:30 +02:00
|
|
|
disconnect = disconnect_mount(p, how);
|
2014-12-24 04:37:03 +01:00
|
|
|
|
|
|
|
pin_insert_group(&p->mnt_umount, &p->mnt_parent->mnt,
|
|
|
|
disconnect ? &unmounted : NULL);
|
2011-11-25 03:47:05 +01:00
|
|
|
if (mnt_has_parent(p)) {
|
2014-08-31 00:32:05 +02:00
|
|
|
mnt_add_count(p->mnt_parent, -1);
|
2014-12-24 04:37:03 +01:00
|
|
|
if (!disconnect) {
|
|
|
|
/* Don't forget about p */
|
|
|
|
list_add_tail(&p->mnt_child, &p->mnt_parent->mnt_mounts);
|
|
|
|
} else {
|
|
|
|
umount_mnt(p);
|
|
|
|
}
|
2008-03-22 04:59:49 +01:00
|
|
|
}
|
2011-11-25 02:43:10 +01:00
|
|
|
change_mnt_propagation(p, MS_PRIVATE);
|
2005-04-17 00:20:36 +02:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2013-03-16 19:39:34 +01:00
|
|
|
static void shrink_submounts(struct mount *mnt);
|
2008-03-22 05:46:23 +01:00
|
|
|
|
2011-11-25 03:35:16 +01:00
|
|
|
static int do_umount(struct mount *mnt, int flags)
|
2005-04-17 00:20:36 +02:00
|
|
|
{
|
2011-11-25 03:35:16 +01:00
|
|
|
struct super_block *sb = mnt->mnt.mnt_sb;
|
2005-04-17 00:20:36 +02:00
|
|
|
int retval;
|
|
|
|
|
2011-11-25 03:35:16 +01:00
|
|
|
retval = security_sb_umount(&mnt->mnt, flags);
|
2005-04-17 00:20:36 +02:00
|
|
|
if (retval)
|
|
|
|
return retval;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Allow userspace to request a mountpoint be expired rather than
|
|
|
|
* unmounting unconditionally. Unmount only happens if:
|
|
|
|
* (1) the mark is already set (the mark is cleared by mntput())
|
|
|
|
* (2) the usage count == 1 [parent vfsmount] + 1 [sys_umount]
|
|
|
|
*/
|
|
|
|
if (flags & MNT_EXPIRE) {
|
2011-11-25 03:35:16 +01:00
|
|
|
if (&mnt->mnt == current->fs->root.mnt ||
|
2005-04-17 00:20:36 +02:00
|
|
|
flags & (MNT_FORCE | MNT_DETACH))
|
|
|
|
return -EINVAL;
|
|
|
|
|
fs: scale mntget/mntput
The problem that this patch aims to fix is vfsmount refcounting scalability.
We need to take a reference on the vfsmount for every successful path lookup,
which often go to the same mount point.
The fundamental difficulty is that a "simple" reference count can never be made
scalable, because any time a reference is dropped, we must check whether that
was the last reference. To do that requires communication with all other CPUs
that may have taken a reference count.
We can make refcounts more scalable in a couple of ways, involving keeping
distributed counters, and checking for the global-zero condition less
frequently.
- check the global sum once every interval (this will delay zero detection
for some interval, so it's probably a showstopper for vfsmounts).
- keep a local count and only taking the global sum when local reaches 0 (this
is difficult for vfsmounts, because we can't hold preempt off for the life of
a reference, so a counter would need to be per-thread or tied strongly to a
particular CPU which requires more locking).
- keep a local difference of increments and decrements, which allows us to sum
the total difference and hence find the refcount when summing all CPUs. Then,
keep a single integer "long" refcount for slow and long lasting references,
and only take the global sum of local counters when the long refcount is 0.
This last scheme is what I implemented here. Attached mounts and process root
and working directory references are "long" references, and everything else is
a short reference.
This allows scalable vfsmount references during path walking over mounted
subtrees and unattached (lazy umounted) mounts with processes still running
in them.
This results in one fewer atomic op in the fastpath: mntget is now just a
per-CPU inc, rather than an atomic inc; and mntput just requires a spinlock
and non-atomic decrement in the common case. However code is otherwise bigger
and heavier, so single threaded performance is basically a wash.
Signed-off-by: Nick Piggin <npiggin@kernel.dk>
2011-01-07 07:50:11 +01:00
|
|
|
/*
|
|
|
|
* probably don't strictly need the lock here if we examined
|
|
|
|
* all race cases, but it's a slowpath.
|
|
|
|
*/
|
2013-09-29 17:24:49 +02:00
|
|
|
lock_mount_hash();
|
2011-11-25 04:37:54 +01:00
|
|
|
if (mnt_get_count(mnt) != 2) {
|
2013-09-29 17:24:49 +02:00
|
|
|
unlock_mount_hash();
|
2005-04-17 00:20:36 +02:00
|
|
|
return -EBUSY;
|
fs: scale mntget/mntput
The problem that this patch aims to fix is vfsmount refcounting scalability.
We need to take a reference on the vfsmount for every successful path lookup,
which often go to the same mount point.
The fundamental difficulty is that a "simple" reference count can never be made
scalable, because any time a reference is dropped, we must check whether that
was the last reference. To do that requires communication with all other CPUs
that may have taken a reference count.
We can make refcounts more scalable in a couple of ways, involving keeping
distributed counters, and checking for the global-zero condition less
frequently.
- check the global sum once every interval (this will delay zero detection
for some interval, so it's probably a showstopper for vfsmounts).
- keep a local count and only taking the global sum when local reaches 0 (this
is difficult for vfsmounts, because we can't hold preempt off for the life of
a reference, so a counter would need to be per-thread or tied strongly to a
particular CPU which requires more locking).
- keep a local difference of increments and decrements, which allows us to sum
the total difference and hence find the refcount when summing all CPUs. Then,
keep a single integer "long" refcount for slow and long lasting references,
and only take the global sum of local counters when the long refcount is 0.
This last scheme is what I implemented here. Attached mounts and process root
and working directory references are "long" references, and everything else is
a short reference.
This allows scalable vfsmount references during path walking over mounted
subtrees and unattached (lazy umounted) mounts with processes still running
in them.
This results in one fewer atomic op in the fastpath: mntget is now just a
per-CPU inc, rather than an atomic inc; and mntput just requires a spinlock
and non-atomic decrement in the common case. However code is otherwise bigger
and heavier, so single threaded performance is basically a wash.
Signed-off-by: Nick Piggin <npiggin@kernel.dk>
2011-01-07 07:50:11 +01:00
|
|
|
}
|
2013-09-29 17:24:49 +02:00
|
|
|
unlock_mount_hash();
|
2005-04-17 00:20:36 +02:00
|
|
|
|
2011-11-25 06:57:42 +01:00
|
|
|
if (!xchg(&mnt->mnt_expiry_mark, 1))
|
2005-04-17 00:20:36 +02:00
|
|
|
return -EAGAIN;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* If we may have to abort operations to get out of this
|
|
|
|
* mount, and they will themselves hold resources we must
|
|
|
|
* allow the fs to do things. In the Unix tradition of
|
|
|
|
* 'Gee thats tricky lets do it in userspace' the umount_begin
|
|
|
|
* might fail to complete on the first run through as other tasks
|
|
|
|
* must return, and the like. Thats for the mount program to worry
|
|
|
|
* about for the moment.
|
|
|
|
*/
|
|
|
|
|
2008-04-24 13:21:56 +02:00
|
|
|
if (flags & MNT_FORCE && sb->s_op->umount_begin) {
|
|
|
|
sb->s_op->umount_begin(sb);
|
|
|
|
}
|
2005-04-17 00:20:36 +02:00
|
|
|
|
|
|
|
/*
|
|
|
|
* No sense to grab the lock for this test, but test itself looks
|
|
|
|
* somewhat bogus. Suggestions for better replacement?
|
|
|
|
* Ho-hum... In principle, we might treat that as umount + switch
|
|
|
|
* to rootfs. GC would eventually take care of the old vfsmount.
|
|
|
|
* Actually it makes sense, especially if rootfs would contain a
|
|
|
|
* /reboot - static binary that would close all descriptors and
|
|
|
|
* call reboot(9). Then init(8) could umount root and exec /reboot.
|
|
|
|
*/
|
2011-11-25 03:35:16 +01:00
|
|
|
if (&mnt->mnt == current->fs->root.mnt && !(flags & MNT_DETACH)) {
|
2005-04-17 00:20:36 +02:00
|
|
|
/*
|
|
|
|
* Special case for "unmounting" root ...
|
|
|
|
* we just try to remount it readonly.
|
|
|
|
*/
|
2014-10-08 21:32:47 +02:00
|
|
|
if (!capable(CAP_SYS_ADMIN))
|
|
|
|
return -EPERM;
|
2005-04-17 00:20:36 +02:00
|
|
|
down_write(&sb->s_umount);
|
2017-07-17 09:45:34 +02:00
|
|
|
if (!sb_rdonly(sb))
|
2017-07-17 09:45:35 +02:00
|
|
|
retval = do_remount_sb(sb, SB_RDONLY, NULL, 0);
|
2005-04-17 00:20:36 +02:00
|
|
|
up_write(&sb->s_umount);
|
|
|
|
return retval;
|
|
|
|
}
|
|
|
|
|
2013-03-16 20:12:40 +01:00
|
|
|
namespace_lock();
|
2013-09-29 17:24:49 +02:00
|
|
|
lock_mount_hash();
|
2005-11-07 23:15:49 +01:00
|
|
|
event++;
|
2005-04-17 00:20:36 +02:00
|
|
|
|
2013-09-30 04:06:07 +02:00
|
|
|
if (flags & MNT_DETACH) {
|
2011-11-25 08:19:55 +01:00
|
|
|
if (!list_empty(&mnt->mnt_list))
|
2014-12-24 14:20:01 +01:00
|
|
|
umount_tree(mnt, UMOUNT_PROPAGATE);
|
2005-04-17 00:20:36 +02:00
|
|
|
retval = 0;
|
2013-09-30 04:06:07 +02:00
|
|
|
} else {
|
|
|
|
shrink_submounts(mnt);
|
|
|
|
retval = -EBUSY;
|
|
|
|
if (!propagate_mount_busy(mnt, 2)) {
|
|
|
|
if (!list_empty(&mnt->mnt_list))
|
2014-12-24 14:20:01 +01:00
|
|
|
umount_tree(mnt, UMOUNT_PROPAGATE|UMOUNT_SYNC);
|
2013-09-30 04:06:07 +02:00
|
|
|
retval = 0;
|
|
|
|
}
|
2005-04-17 00:20:36 +02:00
|
|
|
}
|
2013-09-29 17:24:49 +02:00
|
|
|
unlock_mount_hash();
|
2013-03-16 19:35:16 +01:00
|
|
|
namespace_unlock();
|
2005-04-17 00:20:36 +02:00
|
|
|
return retval;
|
|
|
|
}
|
|
|
|
|
2013-10-03 10:31:18 +02:00
|
|
|
/*
|
|
|
|
* __detach_mounts - lazily unmount all mounts on the specified dentry
|
|
|
|
*
|
|
|
|
* During unlink, rmdir, and d_drop it is possible to loose the path
|
|
|
|
* to an existing mountpoint, and wind up leaking the mount.
|
|
|
|
* detach_mounts allows lazily unmounting those mounts instead of
|
|
|
|
* leaking them.
|
|
|
|
*
|
|
|
|
* The caller may hold dentry->d_inode->i_mutex.
|
|
|
|
*/
|
|
|
|
void __detach_mounts(struct dentry *dentry)
|
|
|
|
{
|
|
|
|
struct mountpoint *mp;
|
|
|
|
struct mount *mnt;
|
|
|
|
|
|
|
|
namespace_lock();
|
2017-01-03 02:18:43 +01:00
|
|
|
lock_mount_hash();
|
2013-10-03 10:31:18 +02:00
|
|
|
mp = lookup_mountpoint(dentry);
|
2015-01-19 18:48:45 +01:00
|
|
|
if (IS_ERR_OR_NULL(mp))
|
2013-10-03 10:31:18 +02:00
|
|
|
goto out_unlock;
|
|
|
|
|
2016-04-15 23:24:41 +02:00
|
|
|
event++;
|
2013-10-03 10:31:18 +02:00
|
|
|
while (!hlist_empty(&mp->m_list)) {
|
|
|
|
mnt = hlist_entry(mp->m_list.first, struct mount, mnt_mp_list);
|
2014-12-24 04:37:03 +01:00
|
|
|
if (mnt->mnt.mnt_flags & MNT_UMOUNT) {
|
2015-07-17 21:54:27 +02:00
|
|
|
hlist_add_head(&mnt->mnt_umount.s_list, &unmounted);
|
|
|
|
umount_mnt(mnt);
|
2014-12-24 04:37:03 +01:00
|
|
|
}
|
2015-04-02 01:30:06 +02:00
|
|
|
else umount_tree(mnt, UMOUNT_CONNECTED);
|
2013-10-03 10:31:18 +02:00
|
|
|
}
|
|
|
|
put_mountpoint(mp);
|
|
|
|
out_unlock:
|
2017-01-03 02:18:43 +01:00
|
|
|
unlock_mount_hash();
|
2013-10-03 10:31:18 +02:00
|
|
|
namespace_unlock();
|
|
|
|
}
|
|
|
|
|
2017-07-04 18:25:09 +02:00
|
|
|
/*
|
2013-02-23 04:45:42 +01:00
|
|
|
* Is the caller allowed to modify his namespace?
|
|
|
|
*/
|
|
|
|
static inline bool may_mount(void)
|
|
|
|
{
|
|
|
|
return ns_capable(current->nsproxy->mnt_ns->user_ns, CAP_SYS_ADMIN);
|
|
|
|
}
|
|
|
|
|
2015-11-16 15:49:34 +01:00
|
|
|
static inline bool may_mandlock(void)
|
|
|
|
{
|
|
|
|
#ifndef CONFIG_MANDATORY_FILE_LOCKING
|
|
|
|
return false;
|
|
|
|
#endif
|
2015-11-12 00:22:33 +01:00
|
|
|
return capable(CAP_SYS_ADMIN);
|
2015-11-16 15:49:34 +01:00
|
|
|
}
|
|
|
|
|
2005-04-17 00:20:36 +02:00
|
|
|
/*
|
|
|
|
* Now umount can handle mount points as well as block devices.
|
|
|
|
* This is important for filesystems which use unnamed block devices.
|
|
|
|
*
|
|
|
|
* We now support a flag for forced unmount like the other 'big iron'
|
|
|
|
* unixes. Our API is identical to OSF/1 to avoid making a mess of AMD
|
|
|
|
*/
|
|
|
|
|
2009-01-14 14:14:12 +01:00
|
|
|
SYSCALL_DEFINE2(umount, char __user *, name, int, flags)
|
2005-04-17 00:20:36 +02:00
|
|
|
{
|
2008-07-22 15:59:21 +02:00
|
|
|
struct path path;
|
2011-11-25 06:33:11 +01:00
|
|
|
struct mount *mnt;
|
2005-04-17 00:20:36 +02:00
|
|
|
int retval;
|
2010-02-10 12:15:53 +01:00
|
|
|
int lookup_flags = 0;
|
2005-04-17 00:20:36 +02:00
|
|
|
|
2010-02-10 12:15:53 +01:00
|
|
|
if (flags & ~(MNT_FORCE | MNT_DETACH | MNT_EXPIRE | UMOUNT_NOFOLLOW))
|
|
|
|
return -EINVAL;
|
|
|
|
|
2013-02-23 04:45:42 +01:00
|
|
|
if (!may_mount())
|
|
|
|
return -EPERM;
|
|
|
|
|
2010-02-10 12:15:53 +01:00
|
|
|
if (!(flags & UMOUNT_NOFOLLOW))
|
|
|
|
lookup_flags |= LOOKUP_FOLLOW;
|
|
|
|
|
2013-09-08 20:03:27 +02:00
|
|
|
retval = user_path_mountpoint_at(AT_FDCWD, name, lookup_flags, &path);
|
2005-04-17 00:20:36 +02:00
|
|
|
if (retval)
|
|
|
|
goto out;
|
2011-11-25 06:33:11 +01:00
|
|
|
mnt = real_mount(path.mnt);
|
2005-04-17 00:20:36 +02:00
|
|
|
retval = -EINVAL;
|
2008-07-22 15:59:21 +02:00
|
|
|
if (path.dentry != path.mnt->mnt_root)
|
2005-04-17 00:20:36 +02:00
|
|
|
goto dput_and_out;
|
2011-11-25 06:46:35 +01:00
|
|
|
if (!check_mnt(mnt))
|
2005-04-17 00:20:36 +02:00
|
|
|
goto dput_and_out;
|
2013-03-30 05:04:39 +01:00
|
|
|
if (mnt->mnt.mnt_flags & MNT_LOCKED)
|
|
|
|
goto dput_and_out;
|
2014-10-04 23:44:03 +02:00
|
|
|
retval = -EPERM;
|
|
|
|
if (flags & MNT_FORCE && !capable(CAP_SYS_ADMIN))
|
|
|
|
goto dput_and_out;
|
2005-04-17 00:20:36 +02:00
|
|
|
|
2011-11-25 06:33:11 +01:00
|
|
|
retval = do_umount(mnt, flags);
|
2005-04-17 00:20:36 +02:00
|
|
|
dput_and_out:
|
2008-02-15 04:34:31 +01:00
|
|
|
/* we mustn't call path_put() as that would clear mnt_expiry_mark */
|
2008-07-22 15:59:21 +02:00
|
|
|
dput(path.dentry);
|
2011-11-25 06:33:11 +01:00
|
|
|
mntput_no_expire(mnt);
|
2005-04-17 00:20:36 +02:00
|
|
|
out:
|
|
|
|
return retval;
|
|
|
|
}
|
|
|
|
|
|
|
|
#ifdef __ARCH_WANT_SYS_OLDUMOUNT
|
|
|
|
|
|
|
|
/*
|
2005-11-07 23:16:09 +01:00
|
|
|
* The 2.0 compatible umount. No flags.
|
2005-04-17 00:20:36 +02:00
|
|
|
*/
|
2009-01-14 14:14:12 +01:00
|
|
|
SYSCALL_DEFINE1(oldumount, char __user *, name)
|
2005-04-17 00:20:36 +02:00
|
|
|
{
|
2005-11-07 23:16:09 +01:00
|
|
|
return sys_umount(name, 0);
|
2005-04-17 00:20:36 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
#endif
|
|
|
|
|
2013-03-30 09:35:18 +01:00
|
|
|
static bool is_mnt_ns_file(struct dentry *dentry)
|
2010-03-08 03:49:36 +01:00
|
|
|
{
|
2013-03-30 09:35:18 +01:00
|
|
|
/* Is this a proxy for a mount namespace? */
|
take the targets of /proc/*/ns/* symlinks to separate fs
New pseudo-filesystem: nsfs. Targets of /proc/*/ns/* live there now.
It's not mountable (not even registered, so it's not in /proc/filesystems,
etc.). Files on it *are* bindable - we explicitly permit that in do_loopback().
This stuff lives in fs/nsfs.c now; proc_ns_fget() moved there as well.
get_proc_ns() is a macro now (it's simply returning ->i_private; would
have been an inline, if not for header ordering headache).
proc_ns_inode() is an ex-parrot. The interface used in procfs is
ns_get_path(path, task, ops) and ns_get_name(buf, size, task, ops).
Dentries and inodes are never hashed; a non-counting reference to dentry
is stashed in ns_common (removed by ->d_prune()) and reused by ns_get_path()
if present. See ns_get_path()/ns_prune_dentry/nsfs_evict() for details
of that mechanism.
As the result, proc_ns_follow_link() has stopped poking in nd->path.mnt;
it does nd_jump_link() on a consistent <vfsmount,dentry> pair it gets
from ns_get_path().
Signed-off-by: Al Viro <viro@zeniv.linux.org.uk>
2014-11-01 15:57:28 +01:00
|
|
|
return dentry->d_op == &ns_dentry_operations &&
|
|
|
|
dentry->d_fsdata == &mntns_operations;
|
2013-03-30 09:35:18 +01:00
|
|
|
}
|
|
|
|
|
2014-11-01 05:00:23 +01:00
|
|
|
struct mnt_namespace *to_mnt_ns(struct ns_common *ns)
|
|
|
|
{
|
|
|
|
return container_of(ns, struct mnt_namespace, ns);
|
|
|
|
}
|
|
|
|
|
2013-03-30 09:35:18 +01:00
|
|
|
static bool mnt_ns_loop(struct dentry *dentry)
|
|
|
|
{
|
|
|
|
/* Could bind mounting the mount namespace inode cause a
|
|
|
|
* mount namespace loop?
|
|
|
|
*/
|
|
|
|
struct mnt_namespace *mnt_ns;
|
|
|
|
if (!is_mnt_ns_file(dentry))
|
|
|
|
return false;
|
|
|
|
|
2014-11-01 08:13:17 +01:00
|
|
|
mnt_ns = to_mnt_ns(get_proc_ns(dentry->d_inode));
|
2010-03-08 03:49:36 +01:00
|
|
|
return current->nsproxy->mnt_ns->seq >= mnt_ns->seq;
|
|
|
|
}
|
|
|
|
|
2011-11-25 03:24:27 +01:00
|
|
|
struct mount *copy_tree(struct mount *mnt, struct dentry *dentry,
|
2005-11-07 23:17:22 +01:00
|
|
|
int flag)
|
2005-04-17 00:20:36 +02:00
|
|
|
{
|
2013-03-15 15:53:28 +01:00
|
|
|
struct mount *res, *p, *q, *r, *parent;
|
2005-04-17 00:20:36 +02:00
|
|
|
|
2013-03-30 09:35:18 +01:00
|
|
|
if (!(flag & CL_COPY_UNBINDABLE) && IS_MNT_UNBINDABLE(mnt))
|
|
|
|
return ERR_PTR(-EINVAL);
|
|
|
|
|
|
|
|
if (!(flag & CL_COPY_MNT_NS_FILE) && is_mnt_ns_file(dentry))
|
2012-06-25 13:55:18 +02:00
|
|
|
return ERR_PTR(-EINVAL);
|
2005-11-07 23:21:20 +01:00
|
|
|
|
2005-11-07 23:17:22 +01:00
|
|
|
res = q = clone_mnt(mnt, dentry, flag);
|
2012-06-25 13:55:18 +02:00
|
|
|
if (IS_ERR(q))
|
|
|
|
return q;
|
|
|
|
|
2011-11-25 04:25:07 +01:00
|
|
|
q->mnt_mountpoint = mnt->mnt_mountpoint;
|
2005-04-17 00:20:36 +02:00
|
|
|
|
|
|
|
p = mnt;
|
2011-11-25 05:24:33 +01:00
|
|
|
list_for_each_entry(r, &mnt->mnt_mounts, mnt_child) {
|
2011-11-25 00:57:30 +01:00
|
|
|
struct mount *s;
|
2008-04-29 09:59:40 +02:00
|
|
|
if (!is_subdir(r->mnt_mountpoint, dentry))
|
2005-04-17 00:20:36 +02:00
|
|
|
continue;
|
|
|
|
|
2011-11-25 09:06:56 +01:00
|
|
|
for (s = r; s; s = next_mnt(s, r)) {
|
2013-03-30 09:35:18 +01:00
|
|
|
if (!(flag & CL_COPY_UNBINDABLE) &&
|
|
|
|
IS_MNT_UNBINDABLE(s)) {
|
|
|
|
s = skip_mnt_tree(s);
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
if (!(flag & CL_COPY_MNT_NS_FILE) &&
|
|
|
|
is_mnt_ns_file(s->mnt.mnt_root)) {
|
2005-11-07 23:21:20 +01:00
|
|
|
s = skip_mnt_tree(s);
|
|
|
|
continue;
|
|
|
|
}
|
2011-11-25 04:19:58 +01:00
|
|
|
while (p != s->mnt_parent) {
|
|
|
|
p = p->mnt_parent;
|
|
|
|
q = q->mnt_parent;
|
2005-04-17 00:20:36 +02:00
|
|
|
}
|
2011-11-25 03:24:27 +01:00
|
|
|
p = s;
|
2013-03-15 15:53:28 +01:00
|
|
|
parent = q;
|
2011-11-25 03:24:27 +01:00
|
|
|
q = clone_mnt(p, p->mnt.mnt_root, flag);
|
2012-06-25 13:55:18 +02:00
|
|
|
if (IS_ERR(q))
|
|
|
|
goto out;
|
2013-09-29 17:24:49 +02:00
|
|
|
lock_mount_hash();
|
2011-11-25 08:19:55 +01:00
|
|
|
list_add_tail(&q->mnt_list, &res->mnt_list);
|
2017-01-20 06:28:35 +01:00
|
|
|
attach_mnt(q, parent, p->mnt_mp);
|
2013-09-29 17:24:49 +02:00
|
|
|
unlock_mount_hash();
|
2005-04-17 00:20:36 +02:00
|
|
|
}
|
|
|
|
}
|
|
|
|
return res;
|
2012-06-25 13:55:18 +02:00
|
|
|
out:
|
2005-04-17 00:20:36 +02:00
|
|
|
if (res) {
|
2013-09-29 17:24:49 +02:00
|
|
|
lock_mount_hash();
|
2014-12-24 14:20:01 +01:00
|
|
|
umount_tree(res, UMOUNT_SYNC);
|
2013-09-29 17:24:49 +02:00
|
|
|
unlock_mount_hash();
|
2005-04-17 00:20:36 +02:00
|
|
|
}
|
2012-06-25 13:55:18 +02:00
|
|
|
return q;
|
2005-04-17 00:20:36 +02:00
|
|
|
}
|
|
|
|
|
2012-06-25 13:55:18 +02:00
|
|
|
/* Caller should check returned pointer for errors */
|
|
|
|
|
2016-11-21 01:45:28 +01:00
|
|
|
struct vfsmount *collect_mounts(const struct path *path)
|
2007-06-07 18:20:32 +02:00
|
|
|
{
|
2011-11-25 02:55:08 +01:00
|
|
|
struct mount *tree;
|
2013-03-16 20:12:40 +01:00
|
|
|
namespace_lock();
|
2015-01-07 21:28:26 +01:00
|
|
|
if (!check_mnt(real_mount(path->mnt)))
|
|
|
|
tree = ERR_PTR(-EINVAL);
|
|
|
|
else
|
|
|
|
tree = copy_tree(real_mount(path->mnt), path->dentry,
|
|
|
|
CL_COPY_ALL | CL_PRIVATE);
|
2013-03-16 19:49:45 +01:00
|
|
|
namespace_unlock();
|
2012-06-25 13:55:18 +02:00
|
|
|
if (IS_ERR(tree))
|
2013-08-14 11:44:39 +02:00
|
|
|
return ERR_CAST(tree);
|
2012-06-25 13:55:18 +02:00
|
|
|
return &tree->mnt;
|
2007-06-07 18:20:32 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
void drop_collected_mounts(struct vfsmount *mnt)
|
|
|
|
{
|
2013-03-16 20:12:40 +01:00
|
|
|
namespace_lock();
|
2013-09-29 17:24:49 +02:00
|
|
|
lock_mount_hash();
|
2014-12-24 14:20:01 +01:00
|
|
|
umount_tree(real_mount(mnt), UMOUNT_SYNC);
|
2013-09-29 17:24:49 +02:00
|
|
|
unlock_mount_hash();
|
2013-03-16 19:42:19 +01:00
|
|
|
namespace_unlock();
|
2007-06-07 18:20:32 +02:00
|
|
|
}
|
|
|
|
|
2014-10-24 00:14:36 +02:00
|
|
|
/**
|
|
|
|
* clone_private_mount - create a private clone of a path
|
|
|
|
*
|
|
|
|
* This creates a new vfsmount, which will be the clone of @path. The new will
|
|
|
|
* not be attached anywhere in the namespace and will be private (i.e. changes
|
|
|
|
* to the originating mount won't be propagated into this).
|
|
|
|
*
|
|
|
|
* Release with mntput().
|
|
|
|
*/
|
2016-11-21 01:45:28 +01:00
|
|
|
struct vfsmount *clone_private_mount(const struct path *path)
|
2014-10-24 00:14:36 +02:00
|
|
|
{
|
|
|
|
struct mount *old_mnt = real_mount(path->mnt);
|
|
|
|
struct mount *new_mnt;
|
|
|
|
|
|
|
|
if (IS_MNT_UNBINDABLE(old_mnt))
|
|
|
|
return ERR_PTR(-EINVAL);
|
|
|
|
|
|
|
|
new_mnt = clone_mnt(old_mnt, path->dentry, CL_PRIVATE);
|
|
|
|
if (IS_ERR(new_mnt))
|
|
|
|
return ERR_CAST(new_mnt);
|
|
|
|
|
|
|
|
return &new_mnt->mnt;
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL_GPL(clone_private_mount);
|
|
|
|
|
2010-01-31 04:51:25 +01:00
|
|
|
int iterate_mounts(int (*f)(struct vfsmount *, void *), void *arg,
|
|
|
|
struct vfsmount *root)
|
|
|
|
{
|
2011-11-25 08:19:55 +01:00
|
|
|
struct mount *mnt;
|
2010-01-31 04:51:25 +01:00
|
|
|
int res = f(root, arg);
|
|
|
|
if (res)
|
|
|
|
return res;
|
2011-11-25 08:19:55 +01:00
|
|
|
list_for_each_entry(mnt, &real_mount(root)->mnt_list, mnt_list) {
|
|
|
|
res = f(&mnt->mnt, arg);
|
2010-01-31 04:51:25 +01:00
|
|
|
if (res)
|
|
|
|
return res;
|
|
|
|
}
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2011-11-25 01:54:23 +01:00
|
|
|
static void cleanup_group_ids(struct mount *mnt, struct mount *end)
|
2008-03-27 13:06:23 +01:00
|
|
|
{
|
2011-11-25 00:57:30 +01:00
|
|
|
struct mount *p;
|
2008-03-27 13:06:23 +01:00
|
|
|
|
2011-11-25 09:06:56 +01:00
|
|
|
for (p = mnt; p != end; p = next_mnt(p, mnt)) {
|
2011-11-25 07:05:37 +01:00
|
|
|
if (p->mnt_group_id && !IS_MNT_SHARED(p))
|
2011-11-25 01:54:23 +01:00
|
|
|
mnt_release_group_id(p);
|
2008-03-27 13:06:23 +01:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2011-11-25 01:54:23 +01:00
|
|
|
static int invent_group_ids(struct mount *mnt, bool recurse)
|
2008-03-27 13:06:23 +01:00
|
|
|
{
|
2011-11-25 00:57:30 +01:00
|
|
|
struct mount *p;
|
2008-03-27 13:06:23 +01:00
|
|
|
|
2011-11-25 09:06:56 +01:00
|
|
|
for (p = mnt; p; p = recurse ? next_mnt(p, mnt) : NULL) {
|
2011-11-25 07:05:37 +01:00
|
|
|
if (!p->mnt_group_id && !IS_MNT_SHARED(p)) {
|
2011-11-25 01:54:23 +01:00
|
|
|
int err = mnt_alloc_group_id(p);
|
2008-03-27 13:06:23 +01:00
|
|
|
if (err) {
|
2011-11-25 01:54:23 +01:00
|
|
|
cleanup_group_ids(mnt, p);
|
2008-03-27 13:06:23 +01:00
|
|
|
return err;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2016-09-28 07:27:17 +02:00
|
|
|
int count_mounts(struct mnt_namespace *ns, struct mount *mnt)
|
|
|
|
{
|
|
|
|
unsigned int max = READ_ONCE(sysctl_mount_max);
|
|
|
|
unsigned int mounts = 0, old, pending, sum;
|
|
|
|
struct mount *p;
|
|
|
|
|
|
|
|
for (p = mnt; p; p = next_mnt(p, mnt))
|
|
|
|
mounts++;
|
|
|
|
|
|
|
|
old = ns->mounts;
|
|
|
|
pending = ns->pending_mounts;
|
|
|
|
sum = old + pending;
|
|
|
|
if ((old > sum) ||
|
|
|
|
(pending > sum) ||
|
|
|
|
(max < sum) ||
|
|
|
|
(mounts > (max - sum)))
|
|
|
|
return -ENOSPC;
|
|
|
|
|
|
|
|
ns->pending_mounts = pending + mounts;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2005-11-07 23:19:50 +01:00
|
|
|
/*
|
|
|
|
* @source_mnt : mount tree to be attached
|
2005-11-07 23:20:03 +01:00
|
|
|
* @nd : place the mount tree @source_mnt is attached
|
|
|
|
* @parent_nd : if non-null, detach the source_mnt from its parent and
|
|
|
|
* store the parent mount and mountpoint dentry.
|
|
|
|
* (done when source_mnt is moved)
|
2005-11-07 23:19:50 +01:00
|
|
|
*
|
|
|
|
* NOTE: in the table below explains the semantics when a source mount
|
|
|
|
* of a given type is attached to a destination mount of a given type.
|
2005-11-07 23:21:20 +01:00
|
|
|
* ---------------------------------------------------------------------------
|
|
|
|
* | BIND MOUNT OPERATION |
|
|
|
|
* |**************************************************************************
|
|
|
|
* | source-->| shared | private | slave | unbindable |
|
|
|
|
* | dest | | | | |
|
|
|
|
* | | | | | | |
|
|
|
|
* | v | | | | |
|
|
|
|
* |**************************************************************************
|
|
|
|
* | shared | shared (++) | shared (+) | shared(+++)| invalid |
|
|
|
|
* | | | | | |
|
|
|
|
* |non-shared| shared (+) | private | slave (*) | invalid |
|
|
|
|
* ***************************************************************************
|
2005-11-07 23:19:50 +01:00
|
|
|
* A bind operation clones the source mount and mounts the clone on the
|
|
|
|
* destination mount.
|
|
|
|
*
|
|
|
|
* (++) the cloned mount is propagated to all the mounts in the propagation
|
|
|
|
* tree of the destination mount and the cloned mount is added to
|
|
|
|
* the peer group of the source mount.
|
|
|
|
* (+) the cloned mount is created under the destination mount and is marked
|
|
|
|
* as shared. The cloned mount is added to the peer group of the source
|
|
|
|
* mount.
|
2005-11-07 23:21:01 +01:00
|
|
|
* (+++) the mount is propagated to all the mounts in the propagation tree
|
|
|
|
* of the destination mount and the cloned mount is made slave
|
|
|
|
* of the same master as that of the source mount. The cloned mount
|
|
|
|
* is marked as 'shared and slave'.
|
|
|
|
* (*) the cloned mount is made a slave of the same master as that of the
|
|
|
|
* source mount.
|
|
|
|
*
|
2005-11-07 23:21:20 +01:00
|
|
|
* ---------------------------------------------------------------------------
|
|
|
|
* | MOVE MOUNT OPERATION |
|
|
|
|
* |**************************************************************************
|
|
|
|
* | source-->| shared | private | slave | unbindable |
|
|
|
|
* | dest | | | | |
|
|
|
|
* | | | | | | |
|
|
|
|
* | v | | | | |
|
|
|
|
* |**************************************************************************
|
|
|
|
* | shared | shared (+) | shared (+) | shared(+++) | invalid |
|
|
|
|
* | | | | | |
|
|
|
|
* |non-shared| shared (+*) | private | slave (*) | unbindable |
|
|
|
|
* ***************************************************************************
|
2005-11-07 23:21:01 +01:00
|
|
|
*
|
|
|
|
* (+) the mount is moved to the destination. And is then propagated to
|
|
|
|
* all the mounts in the propagation tree of the destination mount.
|
2005-11-07 23:20:03 +01:00
|
|
|
* (+*) the mount is moved to the destination.
|
2005-11-07 23:21:01 +01:00
|
|
|
* (+++) the mount is moved to the destination and is then propagated to
|
|
|
|
* all the mounts belonging to the destination mount's propagation tree.
|
|
|
|
* the mount is marked as 'shared and slave'.
|
|
|
|
* (*) the mount continues to be a slave at the new location.
|
2005-11-07 23:19:50 +01:00
|
|
|
*
|
|
|
|
* if the source mount is a tree, the operations explained above is
|
|
|
|
* applied to each mount in the tree.
|
|
|
|
* Must be called without spinlocks held, since this function can sleep
|
|
|
|
* in allocations.
|
|
|
|
*/
|
2011-11-25 01:59:16 +01:00
|
|
|
static int attach_recursive_mnt(struct mount *source_mnt,
|
2013-03-15 15:53:28 +01:00
|
|
|
struct mount *dest_mnt,
|
|
|
|
struct mountpoint *dest_mp,
|
|
|
|
struct path *parent_path)
|
2005-11-07 23:19:50 +01:00
|
|
|
{
|
2014-03-21 02:10:51 +01:00
|
|
|
HLIST_HEAD(tree_list);
|
2016-09-28 07:27:17 +02:00
|
|
|
struct mnt_namespace *ns = dest_mnt->mnt_ns;
|
2017-01-20 06:28:35 +01:00
|
|
|
struct mountpoint *smp;
|
2011-11-25 00:57:30 +01:00
|
|
|
struct mount *child, *p;
|
2014-03-21 02:10:51 +01:00
|
|
|
struct hlist_node *n;
|
2008-03-27 13:06:23 +01:00
|
|
|
int err;
|
2005-11-07 23:19:50 +01:00
|
|
|
|
2017-01-20 06:28:35 +01:00
|
|
|
/* Preallocate a mountpoint in case the new mounts need
|
|
|
|
* to be tucked under other mounts.
|
|
|
|
*/
|
|
|
|
smp = get_mountpoint(source_mnt->mnt.mnt_root);
|
|
|
|
if (IS_ERR(smp))
|
|
|
|
return PTR_ERR(smp);
|
|
|
|
|
2016-09-28 07:27:17 +02:00
|
|
|
/* Is there space to add these mounts to the mount namespace? */
|
|
|
|
if (!parent_path) {
|
|
|
|
err = count_mounts(ns, source_mnt);
|
|
|
|
if (err)
|
|
|
|
goto out;
|
|
|
|
}
|
|
|
|
|
2011-11-25 07:05:37 +01:00
|
|
|
if (IS_MNT_SHARED(dest_mnt)) {
|
2011-11-25 01:59:16 +01:00
|
|
|
err = invent_group_ids(source_mnt, true);
|
2008-03-27 13:06:23 +01:00
|
|
|
if (err)
|
|
|
|
goto out;
|
2014-03-21 15:14:08 +01:00
|
|
|
err = propagate_mnt(dest_mnt, dest_mp, source_mnt, &tree_list);
|
smarter propagate_mnt()
The current mainline has copies propagated to *all* nodes, then
tears down the copies we made for nodes that do not contain
counterparts of the desired mountpoint. That sets the right
propagation graph for the copies (at teardown time we move
the slaves of removed node to a surviving peer or directly
to master), but we end up paying a fairly steep price in
useless allocations. It's fairly easy to create a situation
where N calls of mount(2) create exactly N bindings, with
O(N^2) vfsmounts allocated and freed in process.
Fortunately, it is possible to avoid those allocations/freeings.
The trick is to create copies in the right order and find which
one would've eventually become a master with the current algorithm.
It turns out to be possible in O(nodes getting propagation) time
and with no extra allocations at all.
One part is that we need to make sure that eventual master will be
created before its slaves, so we need to walk the propagation
tree in a different order - by peer groups. And iterate through
the peers before dealing with the next group.
Another thing is finding the (earlier) copy that will be a master
of one we are about to create; to do that we are (temporary) marking
the masters of mountpoints we are attaching the copies to.
Either we are in a peer of the last mountpoint we'd dealt with,
or we have the following situation: we are attaching to mountpoint M,
the last copy S_0 had been attached to M_0 and there are sequences
S_0...S_n, M_0...M_n such that S_{i+1} is a master of S_{i},
S_{i} mounted on M{i} and we need to create a slave of the first S_{k}
such that M is getting propagation from M_{k}. It means that the master
of M_{k} will be among the sequence of masters of M. On the
other hand, the nearest marked node in that sequence will either
be the master of M_{k} or the master of M_{k-1} (the latter -
in the case if M_{k-1} is a slave of something M gets propagation
from, but in a wrong peer group).
So we go through the sequence of masters of M until we find
a marked one (P). Let N be the one before it. Then we go through
the sequence of masters of S_0 until we find one (say, S) mounted
on a node D that has P as master and check if D is a peer of N.
If it is, S will be the master of new copy, if not - the master of S
will be.
That's it for the hard part; the rest is fairly simple. Iterator
is in next_group(), handling of one prospective mountpoint is
propagate_one().
It seems to survive all tests and gives a noticably better performance
than the current mainline for setups that are seriously using shared
subtrees.
Cc: stable@vger.kernel.org
Signed-off-by: Al Viro <viro@zeniv.linux.org.uk>
2014-02-27 15:35:45 +01:00
|
|
|
lock_mount_hash();
|
2014-03-21 15:14:08 +01:00
|
|
|
if (err)
|
|
|
|
goto out_cleanup_ids;
|
2011-11-25 09:06:56 +01:00
|
|
|
for (p = source_mnt; p; p = next_mnt(p, source_mnt))
|
2011-11-25 02:43:10 +01:00
|
|
|
set_mnt_shared(p);
|
2014-03-21 15:14:08 +01:00
|
|
|
} else {
|
|
|
|
lock_mount_hash();
|
2005-11-07 23:19:50 +01:00
|
|
|
}
|
2008-03-22 01:48:19 +01:00
|
|
|
if (parent_path) {
|
2011-11-25 01:59:16 +01:00
|
|
|
detach_mnt(source_mnt, parent_path);
|
2013-03-15 15:53:28 +01:00
|
|
|
attach_mnt(source_mnt, dest_mnt, dest_mp);
|
2011-11-25 06:46:35 +01:00
|
|
|
touch_mnt_namespace(source_mnt->mnt_ns);
|
2005-11-07 23:20:03 +01:00
|
|
|
} else {
|
2013-03-15 15:53:28 +01:00
|
|
|
mnt_set_mountpoint(dest_mnt, dest_mp, source_mnt);
|
2017-01-20 06:28:35 +01:00
|
|
|
commit_tree(source_mnt);
|
2005-11-07 23:20:03 +01:00
|
|
|
}
|
2005-11-07 23:19:50 +01:00
|
|
|
|
2014-03-21 02:10:51 +01:00
|
|
|
hlist_for_each_entry_safe(child, n, &tree_list, mnt_hash) {
|
2014-03-21 01:34:43 +01:00
|
|
|
struct mount *q;
|
2014-03-21 02:10:51 +01:00
|
|
|
hlist_del_init(&child->mnt_hash);
|
2017-01-20 06:28:35 +01:00
|
|
|
q = __lookup_mnt(&child->mnt_parent->mnt,
|
|
|
|
child->mnt_mountpoint);
|
|
|
|
if (q)
|
|
|
|
mnt_change_mountpoint(child, smp, q);
|
|
|
|
commit_tree(child);
|
2005-11-07 23:19:50 +01:00
|
|
|
}
|
2017-01-20 06:28:35 +01:00
|
|
|
put_mountpoint(smp);
|
2013-09-29 17:24:49 +02:00
|
|
|
unlock_mount_hash();
|
fs: brlock vfsmount_lock
fs: brlock vfsmount_lock
Use a brlock for the vfsmount lock. It must be taken for write whenever
modifying the mount hash or associated fields, and may be taken for read when
performing mount hash lookups.
A new lock is added for the mnt-id allocator, so it doesn't need to take
the heavy vfsmount write-lock.
The number of atomics should remain the same for fastpath rlock cases, though
code would be slightly slower due to per-cpu access. Scalability is not not be
much improved in common cases yet, due to other locks (ie. dcache_lock) getting
in the way. However path lookups crossing mountpoints should be one case where
scalability is improved (currently requiring the global lock).
The slowpath is slower due to use of brlock. On a 64 core, 64 socket, 32 node
Altix system (high latency to remote nodes), a simple umount microbenchmark
(mount --bind mnt mnt2 ; umount mnt2 loop 1000 times), before this patch it
took 6.8s, afterwards took 7.1s, about 5% slower.
Cc: Al Viro <viro@ZenIV.linux.org.uk>
Signed-off-by: Nick Piggin <npiggin@kernel.dk>
Signed-off-by: Al Viro <viro@zeniv.linux.org.uk>
2010-08-17 20:37:39 +02:00
|
|
|
|
2005-11-07 23:19:50 +01:00
|
|
|
return 0;
|
2008-03-27 13:06:23 +01:00
|
|
|
|
|
|
|
out_cleanup_ids:
|
smarter propagate_mnt()
The current mainline has copies propagated to *all* nodes, then
tears down the copies we made for nodes that do not contain
counterparts of the desired mountpoint. That sets the right
propagation graph for the copies (at teardown time we move
the slaves of removed node to a surviving peer or directly
to master), but we end up paying a fairly steep price in
useless allocations. It's fairly easy to create a situation
where N calls of mount(2) create exactly N bindings, with
O(N^2) vfsmounts allocated and freed in process.
Fortunately, it is possible to avoid those allocations/freeings.
The trick is to create copies in the right order and find which
one would've eventually become a master with the current algorithm.
It turns out to be possible in O(nodes getting propagation) time
and with no extra allocations at all.
One part is that we need to make sure that eventual master will be
created before its slaves, so we need to walk the propagation
tree in a different order - by peer groups. And iterate through
the peers before dealing with the next group.
Another thing is finding the (earlier) copy that will be a master
of one we are about to create; to do that we are (temporary) marking
the masters of mountpoints we are attaching the copies to.
Either we are in a peer of the last mountpoint we'd dealt with,
or we have the following situation: we are attaching to mountpoint M,
the last copy S_0 had been attached to M_0 and there are sequences
S_0...S_n, M_0...M_n such that S_{i+1} is a master of S_{i},
S_{i} mounted on M{i} and we need to create a slave of the first S_{k}
such that M is getting propagation from M_{k}. It means that the master
of M_{k} will be among the sequence of masters of M. On the
other hand, the nearest marked node in that sequence will either
be the master of M_{k} or the master of M_{k-1} (the latter -
in the case if M_{k-1} is a slave of something M gets propagation
from, but in a wrong peer group).
So we go through the sequence of masters of M until we find
a marked one (P). Let N be the one before it. Then we go through
the sequence of masters of S_0 until we find one (say, S) mounted
on a node D that has P as master and check if D is a peer of N.
If it is, S will be the master of new copy, if not - the master of S
will be.
That's it for the hard part; the rest is fairly simple. Iterator
is in next_group(), handling of one prospective mountpoint is
propagate_one().
It seems to survive all tests and gives a noticably better performance
than the current mainline for setups that are seriously using shared
subtrees.
Cc: stable@vger.kernel.org
Signed-off-by: Al Viro <viro@zeniv.linux.org.uk>
2014-02-27 15:35:45 +01:00
|
|
|
while (!hlist_empty(&tree_list)) {
|
|
|
|
child = hlist_entry(tree_list.first, struct mount, mnt_hash);
|
2016-09-28 07:27:17 +02:00
|
|
|
child->mnt_parent->mnt_ns->pending_mounts = 0;
|
2014-12-24 14:20:01 +01:00
|
|
|
umount_tree(child, UMOUNT_SYNC);
|
smarter propagate_mnt()
The current mainline has copies propagated to *all* nodes, then
tears down the copies we made for nodes that do not contain
counterparts of the desired mountpoint. That sets the right
propagation graph for the copies (at teardown time we move
the slaves of removed node to a surviving peer or directly
to master), but we end up paying a fairly steep price in
useless allocations. It's fairly easy to create a situation
where N calls of mount(2) create exactly N bindings, with
O(N^2) vfsmounts allocated and freed in process.
Fortunately, it is possible to avoid those allocations/freeings.
The trick is to create copies in the right order and find which
one would've eventually become a master with the current algorithm.
It turns out to be possible in O(nodes getting propagation) time
and with no extra allocations at all.
One part is that we need to make sure that eventual master will be
created before its slaves, so we need to walk the propagation
tree in a different order - by peer groups. And iterate through
the peers before dealing with the next group.
Another thing is finding the (earlier) copy that will be a master
of one we are about to create; to do that we are (temporary) marking
the masters of mountpoints we are attaching the copies to.
Either we are in a peer of the last mountpoint we'd dealt with,
or we have the following situation: we are attaching to mountpoint M,
the last copy S_0 had been attached to M_0 and there are sequences
S_0...S_n, M_0...M_n such that S_{i+1} is a master of S_{i},
S_{i} mounted on M{i} and we need to create a slave of the first S_{k}
such that M is getting propagation from M_{k}. It means that the master
of M_{k} will be among the sequence of masters of M. On the
other hand, the nearest marked node in that sequence will either
be the master of M_{k} or the master of M_{k-1} (the latter -
in the case if M_{k-1} is a slave of something M gets propagation
from, but in a wrong peer group).
So we go through the sequence of masters of M until we find
a marked one (P). Let N be the one before it. Then we go through
the sequence of masters of S_0 until we find one (say, S) mounted
on a node D that has P as master and check if D is a peer of N.
If it is, S will be the master of new copy, if not - the master of S
will be.
That's it for the hard part; the rest is fairly simple. Iterator
is in next_group(), handling of one prospective mountpoint is
propagate_one().
It seems to survive all tests and gives a noticably better performance
than the current mainline for setups that are seriously using shared
subtrees.
Cc: stable@vger.kernel.org
Signed-off-by: Al Viro <viro@zeniv.linux.org.uk>
2014-02-27 15:35:45 +01:00
|
|
|
}
|
|
|
|
unlock_mount_hash();
|
2014-03-21 15:14:08 +01:00
|
|
|
cleanup_group_ids(source_mnt, NULL);
|
2008-03-27 13:06:23 +01:00
|
|
|
out:
|
2016-09-28 07:27:17 +02:00
|
|
|
ns->pending_mounts = 0;
|
2017-01-20 06:28:35 +01:00
|
|
|
|
|
|
|
read_seqlock_excl(&mount_lock);
|
|
|
|
put_mountpoint(smp);
|
|
|
|
read_sequnlock_excl(&mount_lock);
|
|
|
|
|
2008-03-27 13:06:23 +01:00
|
|
|
return err;
|
2005-11-07 23:19:50 +01:00
|
|
|
}
|
|
|
|
|
2013-03-15 15:53:28 +01:00
|
|
|
static struct mountpoint *lock_mount(struct path *path)
|
2011-03-18 13:55:38 +01:00
|
|
|
{
|
|
|
|
struct vfsmount *mnt;
|
2013-03-15 15:53:28 +01:00
|
|
|
struct dentry *dentry = path->dentry;
|
2011-03-18 13:55:38 +01:00
|
|
|
retry:
|
2016-01-22 21:40:57 +01:00
|
|
|
inode_lock(dentry->d_inode);
|
2013-03-15 15:53:28 +01:00
|
|
|
if (unlikely(cant_mount(dentry))) {
|
2016-01-22 21:40:57 +01:00
|
|
|
inode_unlock(dentry->d_inode);
|
2013-03-15 15:53:28 +01:00
|
|
|
return ERR_PTR(-ENOENT);
|
2011-03-18 13:55:38 +01:00
|
|
|
}
|
2013-03-16 20:12:40 +01:00
|
|
|
namespace_lock();
|
2011-03-18 13:55:38 +01:00
|
|
|
mnt = lookup_mnt(path);
|
2013-03-15 15:53:28 +01:00
|
|
|
if (likely(!mnt)) {
|
2017-01-03 02:18:43 +01:00
|
|
|
struct mountpoint *mp = get_mountpoint(dentry);
|
2013-03-15 15:53:28 +01:00
|
|
|
if (IS_ERR(mp)) {
|
2013-03-16 20:12:40 +01:00
|
|
|
namespace_unlock();
|
2016-01-22 21:40:57 +01:00
|
|
|
inode_unlock(dentry->d_inode);
|
2013-03-15 15:53:28 +01:00
|
|
|
return mp;
|
|
|
|
}
|
|
|
|
return mp;
|
|
|
|
}
|
2013-03-16 20:12:40 +01:00
|
|
|
namespace_unlock();
|
2016-01-22 21:40:57 +01:00
|
|
|
inode_unlock(path->dentry->d_inode);
|
2011-03-18 13:55:38 +01:00
|
|
|
path_put(path);
|
|
|
|
path->mnt = mnt;
|
2013-03-15 15:53:28 +01:00
|
|
|
dentry = path->dentry = dget(mnt->mnt_root);
|
2011-03-18 13:55:38 +01:00
|
|
|
goto retry;
|
|
|
|
}
|
|
|
|
|
2013-03-15 15:53:28 +01:00
|
|
|
static void unlock_mount(struct mountpoint *where)
|
2011-03-18 13:55:38 +01:00
|
|
|
{
|
2013-03-15 15:53:28 +01:00
|
|
|
struct dentry *dentry = where->m_dentry;
|
2017-01-03 02:18:43 +01:00
|
|
|
|
|
|
|
read_seqlock_excl(&mount_lock);
|
2013-03-15 15:53:28 +01:00
|
|
|
put_mountpoint(where);
|
2017-01-03 02:18:43 +01:00
|
|
|
read_sequnlock_excl(&mount_lock);
|
|
|
|
|
2013-03-16 19:49:45 +01:00
|
|
|
namespace_unlock();
|
2016-01-22 21:40:57 +01:00
|
|
|
inode_unlock(dentry->d_inode);
|
2011-03-18 13:55:38 +01:00
|
|
|
}
|
|
|
|
|
2013-03-15 15:53:28 +01:00
|
|
|
static int graft_tree(struct mount *mnt, struct mount *p, struct mountpoint *mp)
|
2005-04-17 00:20:36 +02:00
|
|
|
{
|
2017-07-17 09:45:35 +02:00
|
|
|
if (mnt->mnt.mnt_sb->s_flags & SB_NOUSER)
|
2005-04-17 00:20:36 +02:00
|
|
|
return -EINVAL;
|
|
|
|
|
VFS: (Scripted) Convert S_ISLNK/DIR/REG(dentry->d_inode) to d_is_*(dentry)
Convert the following where appropriate:
(1) S_ISLNK(dentry->d_inode) to d_is_symlink(dentry).
(2) S_ISREG(dentry->d_inode) to d_is_reg(dentry).
(3) S_ISDIR(dentry->d_inode) to d_is_dir(dentry). This is actually more
complicated than it appears as some calls should be converted to
d_can_lookup() instead. The difference is whether the directory in
question is a real dir with a ->lookup op or whether it's a fake dir with
a ->d_automount op.
In some circumstances, we can subsume checks for dentry->d_inode not being
NULL into this, provided we the code isn't in a filesystem that expects
d_inode to be NULL if the dirent really *is* negative (ie. if we're going to
use d_inode() rather than d_backing_inode() to get the inode pointer).
Note that the dentry type field may be set to something other than
DCACHE_MISS_TYPE when d_inode is NULL in the case of unionmount, where the VFS
manages the fall-through from a negative dentry to a lower layer. In such a
case, the dentry type of the negative union dentry is set to the same as the
type of the lower dentry.
However, if you know d_inode is not NULL at the call site, then you can use
the d_is_xxx() functions even in a filesystem.
There is one further complication: a 0,0 chardev dentry may be labelled
DCACHE_WHITEOUT_TYPE rather than DCACHE_SPECIAL_TYPE. Strictly, this was
intended for special directory entry types that don't have attached inodes.
The following perl+coccinelle script was used:
use strict;
my @callers;
open($fd, 'git grep -l \'S_IS[A-Z].*->d_inode\' |') ||
die "Can't grep for S_ISDIR and co. callers";
@callers = <$fd>;
close($fd);
unless (@callers) {
print "No matches\n";
exit(0);
}
my @cocci = (
'@@',
'expression E;',
'@@',
'',
'- S_ISLNK(E->d_inode->i_mode)',
'+ d_is_symlink(E)',
'',
'@@',
'expression E;',
'@@',
'',
'- S_ISDIR(E->d_inode->i_mode)',
'+ d_is_dir(E)',
'',
'@@',
'expression E;',
'@@',
'',
'- S_ISREG(E->d_inode->i_mode)',
'+ d_is_reg(E)' );
my $coccifile = "tmp.sp.cocci";
open($fd, ">$coccifile") || die $coccifile;
print($fd "$_\n") || die $coccifile foreach (@cocci);
close($fd);
foreach my $file (@callers) {
chomp $file;
print "Processing ", $file, "\n";
system("spatch", "--sp-file", $coccifile, $file, "--in-place", "--no-show-diff") == 0 ||
die "spatch failed";
}
[AV: overlayfs parts skipped]
Signed-off-by: David Howells <dhowells@redhat.com>
Signed-off-by: Al Viro <viro@zeniv.linux.org.uk>
2015-01-29 13:02:35 +01:00
|
|
|
if (d_is_dir(mp->m_dentry) !=
|
|
|
|
d_is_dir(mnt->mnt.mnt_root))
|
2005-04-17 00:20:36 +02:00
|
|
|
return -ENOTDIR;
|
|
|
|
|
2013-03-15 15:53:28 +01:00
|
|
|
return attach_recursive_mnt(mnt, p, mp, NULL);
|
2005-04-17 00:20:36 +02:00
|
|
|
}
|
|
|
|
|
2010-08-26 20:07:22 +02:00
|
|
|
/*
|
|
|
|
* Sanity check the flags to change_mnt_propagation.
|
|
|
|
*/
|
|
|
|
|
2017-07-17 09:45:35 +02:00
|
|
|
static int flags_to_propagation_type(int ms_flags)
|
2010-08-26 20:07:22 +02:00
|
|
|
{
|
2017-07-17 09:45:35 +02:00
|
|
|
int type = ms_flags & ~(MS_REC | MS_SILENT);
|
2010-08-26 20:07:22 +02:00
|
|
|
|
|
|
|
/* Fail if any non-propagation flags are set */
|
|
|
|
if (type & ~(MS_SHARED | MS_PRIVATE | MS_SLAVE | MS_UNBINDABLE))
|
|
|
|
return 0;
|
|
|
|
/* Only one propagation flag should be set */
|
|
|
|
if (!is_power_of_2(type))
|
|
|
|
return 0;
|
|
|
|
return type;
|
|
|
|
}
|
|
|
|
|
2005-11-07 23:19:07 +01:00
|
|
|
/*
|
|
|
|
* recursively change the type of the mountpoint.
|
|
|
|
*/
|
2017-07-17 09:45:35 +02:00
|
|
|
static int do_change_type(struct path *path, int ms_flags)
|
2005-11-07 23:19:07 +01:00
|
|
|
{
|
2011-11-25 00:57:30 +01:00
|
|
|
struct mount *m;
|
2011-11-25 01:54:23 +01:00
|
|
|
struct mount *mnt = real_mount(path->mnt);
|
2017-07-17 09:45:35 +02:00
|
|
|
int recurse = ms_flags & MS_REC;
|
2010-08-26 20:07:22 +02:00
|
|
|
int type;
|
2008-03-27 13:06:23 +01:00
|
|
|
int err = 0;
|
2005-11-07 23:19:07 +01:00
|
|
|
|
2008-08-02 06:51:11 +02:00
|
|
|
if (path->dentry != path->mnt->mnt_root)
|
2005-11-07 23:19:07 +01:00
|
|
|
return -EINVAL;
|
|
|
|
|
2017-07-17 09:45:35 +02:00
|
|
|
type = flags_to_propagation_type(ms_flags);
|
2010-08-26 20:07:22 +02:00
|
|
|
if (!type)
|
|
|
|
return -EINVAL;
|
|
|
|
|
2013-03-16 20:12:40 +01:00
|
|
|
namespace_lock();
|
2008-03-27 13:06:23 +01:00
|
|
|
if (type == MS_SHARED) {
|
|
|
|
err = invent_group_ids(mnt, recurse);
|
|
|
|
if (err)
|
|
|
|
goto out_unlock;
|
|
|
|
}
|
|
|
|
|
2013-09-29 17:24:49 +02:00
|
|
|
lock_mount_hash();
|
2011-11-25 09:06:56 +01:00
|
|
|
for (m = mnt; m; m = (recurse ? next_mnt(m, mnt) : NULL))
|
2011-11-25 02:43:10 +01:00
|
|
|
change_mnt_propagation(m, type);
|
2013-09-29 17:24:49 +02:00
|
|
|
unlock_mount_hash();
|
2008-03-27 13:06:23 +01:00
|
|
|
|
|
|
|
out_unlock:
|
2013-03-16 20:12:40 +01:00
|
|
|
namespace_unlock();
|
2008-03-27 13:06:23 +01:00
|
|
|
return err;
|
2005-11-07 23:19:07 +01:00
|
|
|
}
|
|
|
|
|
2013-03-30 05:04:39 +01:00
|
|
|
static bool has_locked_children(struct mount *mnt, struct dentry *dentry)
|
|
|
|
{
|
|
|
|
struct mount *child;
|
|
|
|
list_for_each_entry(child, &mnt->mnt_mounts, mnt_child) {
|
|
|
|
if (!is_subdir(child->mnt_mountpoint, dentry))
|
|
|
|
continue;
|
|
|
|
|
|
|
|
if (child->mnt.mnt_flags & MNT_LOCKED)
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
return false;
|
|
|
|
}
|
|
|
|
|
2005-04-17 00:20:36 +02:00
|
|
|
/*
|
|
|
|
* do loopback mount.
|
|
|
|
*/
|
2012-10-11 17:42:01 +02:00
|
|
|
static int do_loopback(struct path *path, const char *old_name,
|
2008-02-08 13:22:12 +01:00
|
|
|
int recurse)
|
2005-04-17 00:20:36 +02:00
|
|
|
{
|
2008-08-02 06:51:11 +02:00
|
|
|
struct path old_path;
|
2013-03-15 15:53:28 +01:00
|
|
|
struct mount *mnt = NULL, *old, *parent;
|
|
|
|
struct mountpoint *mp;
|
2013-02-23 04:49:10 +01:00
|
|
|
int err;
|
2005-04-17 00:20:36 +02:00
|
|
|
if (!old_name || !*old_name)
|
|
|
|
return -EINVAL;
|
2011-09-27 02:36:09 +02:00
|
|
|
err = kern_path(old_name, LOOKUP_FOLLOW|LOOKUP_AUTOMOUNT, &old_path);
|
2005-04-17 00:20:36 +02:00
|
|
|
if (err)
|
|
|
|
return err;
|
|
|
|
|
2010-03-08 03:49:36 +01:00
|
|
|
err = -EINVAL;
|
2013-03-30 09:35:18 +01:00
|
|
|
if (mnt_ns_loop(old_path.dentry))
|
2017-07-04 18:25:09 +02:00
|
|
|
goto out;
|
2010-03-08 03:49:36 +01:00
|
|
|
|
2013-03-15 15:53:28 +01:00
|
|
|
mp = lock_mount(path);
|
|
|
|
err = PTR_ERR(mp);
|
|
|
|
if (IS_ERR(mp))
|
2011-03-18 13:55:38 +01:00
|
|
|
goto out;
|
|
|
|
|
2011-11-25 03:24:27 +01:00
|
|
|
old = real_mount(old_path.mnt);
|
2013-03-15 15:53:28 +01:00
|
|
|
parent = real_mount(path->mnt);
|
2011-11-25 03:24:27 +01:00
|
|
|
|
2005-04-17 00:20:36 +02:00
|
|
|
err = -EINVAL;
|
2011-11-25 07:05:37 +01:00
|
|
|
if (IS_MNT_UNBINDABLE(old))
|
2011-03-18 13:55:38 +01:00
|
|
|
goto out2;
|
2005-11-07 23:21:20 +01:00
|
|
|
|
take the targets of /proc/*/ns/* symlinks to separate fs
New pseudo-filesystem: nsfs. Targets of /proc/*/ns/* live there now.
It's not mountable (not even registered, so it's not in /proc/filesystems,
etc.). Files on it *are* bindable - we explicitly permit that in do_loopback().
This stuff lives in fs/nsfs.c now; proc_ns_fget() moved there as well.
get_proc_ns() is a macro now (it's simply returning ->i_private; would
have been an inline, if not for header ordering headache).
proc_ns_inode() is an ex-parrot. The interface used in procfs is
ns_get_path(path, task, ops) and ns_get_name(buf, size, task, ops).
Dentries and inodes are never hashed; a non-counting reference to dentry
is stashed in ns_common (removed by ->d_prune()) and reused by ns_get_path()
if present. See ns_get_path()/ns_prune_dentry/nsfs_evict() for details
of that mechanism.
As the result, proc_ns_follow_link() has stopped poking in nd->path.mnt;
it does nd_jump_link() on a consistent <vfsmount,dentry> pair it gets
from ns_get_path().
Signed-off-by: Al Viro <viro@zeniv.linux.org.uk>
2014-11-01 15:57:28 +01:00
|
|
|
if (!check_mnt(parent))
|
|
|
|
goto out2;
|
|
|
|
|
|
|
|
if (!check_mnt(old) && old_path.dentry->d_op != &ns_dentry_operations)
|
2011-03-18 13:55:38 +01:00
|
|
|
goto out2;
|
2005-04-17 00:20:36 +02:00
|
|
|
|
2013-03-30 05:04:39 +01:00
|
|
|
if (!recurse && has_locked_children(old, old_path.dentry))
|
|
|
|
goto out2;
|
|
|
|
|
2005-11-07 23:15:04 +01:00
|
|
|
if (recurse)
|
2013-03-30 09:35:18 +01:00
|
|
|
mnt = copy_tree(old, old_path.dentry, CL_COPY_MNT_NS_FILE);
|
2005-11-07 23:15:04 +01:00
|
|
|
else
|
2011-11-25 03:24:27 +01:00
|
|
|
mnt = clone_mnt(old, old_path.dentry, 0);
|
2005-11-07 23:15:04 +01:00
|
|
|
|
2012-06-25 13:55:18 +02:00
|
|
|
if (IS_ERR(mnt)) {
|
|
|
|
err = PTR_ERR(mnt);
|
2013-04-09 15:33:29 +02:00
|
|
|
goto out2;
|
2012-06-25 13:55:18 +02:00
|
|
|
}
|
2005-11-07 23:15:04 +01:00
|
|
|
|
2013-03-30 05:04:39 +01:00
|
|
|
mnt->mnt.mnt_flags &= ~MNT_LOCKED;
|
|
|
|
|
2013-03-15 15:53:28 +01:00
|
|
|
err = graft_tree(mnt, parent, mp);
|
2005-11-07 23:15:04 +01:00
|
|
|
if (err) {
|
2013-09-29 17:24:49 +02:00
|
|
|
lock_mount_hash();
|
2014-12-24 14:20:01 +01:00
|
|
|
umount_tree(mnt, UMOUNT_SYNC);
|
2013-09-29 17:24:49 +02:00
|
|
|
unlock_mount_hash();
|
2005-11-07 23:16:29 +01:00
|
|
|
}
|
2011-03-18 13:55:38 +01:00
|
|
|
out2:
|
2013-03-15 15:53:28 +01:00
|
|
|
unlock_mount(mp);
|
2005-11-07 23:15:04 +01:00
|
|
|
out:
|
2008-08-02 06:51:11 +02:00
|
|
|
path_put(&old_path);
|
2005-04-17 00:20:36 +02:00
|
|
|
return err;
|
|
|
|
}
|
|
|
|
|
2008-02-15 23:38:00 +01:00
|
|
|
static int change_mount_flags(struct vfsmount *mnt, int ms_flags)
|
|
|
|
{
|
|
|
|
int error = 0;
|
|
|
|
int readonly_request = 0;
|
|
|
|
|
|
|
|
if (ms_flags & MS_RDONLY)
|
|
|
|
readonly_request = 1;
|
|
|
|
if (readonly_request == __mnt_is_readonly(mnt))
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
if (readonly_request)
|
2011-11-25 04:37:54 +01:00
|
|
|
error = mnt_make_readonly(real_mount(mnt));
|
2008-02-15 23:38:00 +01:00
|
|
|
else
|
2011-11-25 04:37:54 +01:00
|
|
|
__mnt_unmake_readonly(real_mount(mnt));
|
2008-02-15 23:38:00 +01:00
|
|
|
return error;
|
|
|
|
}
|
|
|
|
|
2005-04-17 00:20:36 +02:00
|
|
|
/*
|
|
|
|
* change filesystem flags. dir should be a physical root of filesystem.
|
|
|
|
* If you've mounted a non-root directory somewhere and want to do remount
|
|
|
|
* on it - tough luck.
|
|
|
|
*/
|
2017-07-17 09:45:35 +02:00
|
|
|
static int do_remount(struct path *path, int ms_flags, int sb_flags,
|
|
|
|
int mnt_flags, void *data)
|
2005-04-17 00:20:36 +02:00
|
|
|
{
|
|
|
|
int err;
|
2008-08-02 06:51:11 +02:00
|
|
|
struct super_block *sb = path->mnt->mnt_sb;
|
2011-11-25 06:46:35 +01:00
|
|
|
struct mount *mnt = real_mount(path->mnt);
|
2005-04-17 00:20:36 +02:00
|
|
|
|
2011-11-25 06:46:35 +01:00
|
|
|
if (!check_mnt(mnt))
|
2005-04-17 00:20:36 +02:00
|
|
|
return -EINVAL;
|
|
|
|
|
2008-08-02 06:51:11 +02:00
|
|
|
if (path->dentry != path->mnt->mnt_root)
|
2005-04-17 00:20:36 +02:00
|
|
|
return -EINVAL;
|
|
|
|
|
2014-07-29 02:10:56 +02:00
|
|
|
/* Don't allow changing of locked mnt flags.
|
|
|
|
*
|
|
|
|
* No locks need to be held here while testing the various
|
|
|
|
* MNT_LOCK flags because those flags can never be cleared
|
|
|
|
* once they are set.
|
|
|
|
*/
|
|
|
|
if ((mnt->mnt.mnt_flags & MNT_LOCK_READONLY) &&
|
|
|
|
!(mnt_flags & MNT_READONLY)) {
|
|
|
|
return -EPERM;
|
|
|
|
}
|
mnt: Correct permission checks in do_remount
While invesgiating the issue where in "mount --bind -oremount,ro ..."
would result in later "mount --bind -oremount,rw" succeeding even if
the mount started off locked I realized that there are several
additional mount flags that should be locked and are not.
In particular MNT_NOSUID, MNT_NODEV, MNT_NOEXEC, and the atime
flags in addition to MNT_READONLY should all be locked. These
flags are all per superblock, can all be changed with MS_BIND,
and should not be changable if set by a more privileged user.
The following additions to the current logic are added in this patch.
- nosuid may not be clearable by a less privileged user.
- nodev may not be clearable by a less privielged user.
- noexec may not be clearable by a less privileged user.
- atime flags may not be changeable by a less privileged user.
The logic with atime is that always setting atime on access is a
global policy and backup software and auditing software could break if
atime bits are not updated (when they are configured to be updated),
and serious performance degradation could result (DOS attack) if atime
updates happen when they have been explicitly disabled. Therefore an
unprivileged user should not be able to mess with the atime bits set
by a more privileged user.
The additional restrictions are implemented with the addition of
MNT_LOCK_NOSUID, MNT_LOCK_NODEV, MNT_LOCK_NOEXEC, and MNT_LOCK_ATIME
mnt flags.
Taken together these changes and the fixes for MNT_LOCK_READONLY
should make it safe for an unprivileged user to create a user
namespace and to call "mount --bind -o remount,... ..." without
the danger of mount flags being changed maliciously.
Cc: stable@vger.kernel.org
Acked-by: Serge E. Hallyn <serge.hallyn@ubuntu.com>
Signed-off-by: "Eric W. Biederman" <ebiederm@xmission.com>
2014-07-29 02:26:07 +02:00
|
|
|
if ((mnt->mnt.mnt_flags & MNT_LOCK_NODEV) &&
|
|
|
|
!(mnt_flags & MNT_NODEV)) {
|
2016-05-18 20:50:06 +02:00
|
|
|
return -EPERM;
|
mnt: Correct permission checks in do_remount
While invesgiating the issue where in "mount --bind -oremount,ro ..."
would result in later "mount --bind -oremount,rw" succeeding even if
the mount started off locked I realized that there are several
additional mount flags that should be locked and are not.
In particular MNT_NOSUID, MNT_NODEV, MNT_NOEXEC, and the atime
flags in addition to MNT_READONLY should all be locked. These
flags are all per superblock, can all be changed with MS_BIND,
and should not be changable if set by a more privileged user.
The following additions to the current logic are added in this patch.
- nosuid may not be clearable by a less privileged user.
- nodev may not be clearable by a less privielged user.
- noexec may not be clearable by a less privileged user.
- atime flags may not be changeable by a less privileged user.
The logic with atime is that always setting atime on access is a
global policy and backup software and auditing software could break if
atime bits are not updated (when they are configured to be updated),
and serious performance degradation could result (DOS attack) if atime
updates happen when they have been explicitly disabled. Therefore an
unprivileged user should not be able to mess with the atime bits set
by a more privileged user.
The additional restrictions are implemented with the addition of
MNT_LOCK_NOSUID, MNT_LOCK_NODEV, MNT_LOCK_NOEXEC, and MNT_LOCK_ATIME
mnt flags.
Taken together these changes and the fixes for MNT_LOCK_READONLY
should make it safe for an unprivileged user to create a user
namespace and to call "mount --bind -o remount,... ..." without
the danger of mount flags being changed maliciously.
Cc: stable@vger.kernel.org
Acked-by: Serge E. Hallyn <serge.hallyn@ubuntu.com>
Signed-off-by: "Eric W. Biederman" <ebiederm@xmission.com>
2014-07-29 02:26:07 +02:00
|
|
|
}
|
|
|
|
if ((mnt->mnt.mnt_flags & MNT_LOCK_NOSUID) &&
|
|
|
|
!(mnt_flags & MNT_NOSUID)) {
|
|
|
|
return -EPERM;
|
|
|
|
}
|
|
|
|
if ((mnt->mnt.mnt_flags & MNT_LOCK_NOEXEC) &&
|
|
|
|
!(mnt_flags & MNT_NOEXEC)) {
|
|
|
|
return -EPERM;
|
|
|
|
}
|
|
|
|
if ((mnt->mnt.mnt_flags & MNT_LOCK_ATIME) &&
|
|
|
|
((mnt->mnt.mnt_flags & MNT_ATIME_MASK) != (mnt_flags & MNT_ATIME_MASK))) {
|
|
|
|
return -EPERM;
|
|
|
|
}
|
|
|
|
|
2011-03-03 22:09:14 +01:00
|
|
|
err = security_sb_remount(sb, data);
|
|
|
|
if (err)
|
|
|
|
return err;
|
|
|
|
|
2005-04-17 00:20:36 +02:00
|
|
|
down_write(&sb->s_umount);
|
2017-07-17 09:45:35 +02:00
|
|
|
if (ms_flags & MS_BIND)
|
|
|
|
err = change_mount_flags(path->mnt, ms_flags);
|
2013-02-23 04:49:10 +01:00
|
|
|
else if (!capable(CAP_SYS_ADMIN))
|
|
|
|
err = -EPERM;
|
2009-05-08 19:36:58 +02:00
|
|
|
else
|
2017-07-17 09:45:35 +02:00
|
|
|
err = do_remount_sb(sb, sb_flags, data, 0);
|
2010-01-16 19:01:26 +01:00
|
|
|
if (!err) {
|
2013-09-29 17:24:49 +02:00
|
|
|
lock_mount_hash();
|
2014-07-29 01:26:53 +02:00
|
|
|
mnt_flags |= mnt->mnt.mnt_flags & ~MNT_USER_SETTABLE_MASK;
|
2011-11-25 06:46:35 +01:00
|
|
|
mnt->mnt.mnt_flags = mnt_flags;
|
|
|
|
touch_mnt_namespace(mnt->mnt_ns);
|
2013-09-29 17:24:49 +02:00
|
|
|
unlock_mount_hash();
|
2008-09-27 04:01:20 +02:00
|
|
|
}
|
2013-09-17 04:41:01 +02:00
|
|
|
up_write(&sb->s_umount);
|
2005-04-17 00:20:36 +02:00
|
|
|
return err;
|
|
|
|
}
|
|
|
|
|
2011-11-25 02:01:19 +01:00
|
|
|
static inline int tree_contains_unbindable(struct mount *mnt)
|
2005-11-07 23:21:20 +01:00
|
|
|
{
|
2011-11-25 00:57:30 +01:00
|
|
|
struct mount *p;
|
2011-11-25 09:06:56 +01:00
|
|
|
for (p = mnt; p; p = next_mnt(p, mnt)) {
|
2011-11-25 07:05:37 +01:00
|
|
|
if (IS_MNT_UNBINDABLE(p))
|
2005-11-07 23:21:20 +01:00
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2012-10-11 17:42:01 +02:00
|
|
|
static int do_move_mount(struct path *path, const char *old_name)
|
2005-04-17 00:20:36 +02:00
|
|
|
{
|
2008-08-02 06:51:11 +02:00
|
|
|
struct path old_path, parent_path;
|
2011-11-25 03:47:05 +01:00
|
|
|
struct mount *p;
|
2011-11-25 01:59:16 +01:00
|
|
|
struct mount *old;
|
2013-03-15 15:53:28 +01:00
|
|
|
struct mountpoint *mp;
|
2013-02-23 04:49:10 +01:00
|
|
|
int err;
|
2005-04-17 00:20:36 +02:00
|
|
|
if (!old_name || !*old_name)
|
|
|
|
return -EINVAL;
|
2008-08-02 06:51:11 +02:00
|
|
|
err = kern_path(old_name, LOOKUP_FOLLOW, &old_path);
|
2005-04-17 00:20:36 +02:00
|
|
|
if (err)
|
|
|
|
return err;
|
|
|
|
|
2013-03-15 15:53:28 +01:00
|
|
|
mp = lock_mount(path);
|
|
|
|
err = PTR_ERR(mp);
|
|
|
|
if (IS_ERR(mp))
|
Add a dentry op to allow processes to be held during pathwalk transit
Add a dentry op (d_manage) to permit a filesystem to hold a process and make it
sleep when it tries to transit away from one of that filesystem's directories
during a pathwalk. The operation is keyed off a new dentry flag
(DCACHE_MANAGE_TRANSIT).
The filesystem is allowed to be selective about which processes it holds and
which it permits to continue on or prohibits from transiting from each flagged
directory. This will allow autofs to hold up client processes whilst letting
its userspace daemon through to maintain the directory or the stuff behind it
or mounted upon it.
The ->d_manage() dentry operation:
int (*d_manage)(struct path *path, bool mounting_here);
takes a pointer to the directory about to be transited away from and a flag
indicating whether the transit is undertaken by do_add_mount() or
do_move_mount() skipping through a pile of filesystems mounted on a mountpoint.
It should return 0 if successful and to let the process continue on its way;
-EISDIR to prohibit the caller from skipping to overmounted filesystems or
automounting, and to use this directory; or some other error code to return to
the user.
->d_manage() is called with namespace_sem writelocked if mounting_here is true
and no other locks held, so it may sleep. However, if mounting_here is true,
it may not initiate or wait for a mount or unmount upon the parameter
directory, even if the act is actually performed by userspace.
Within fs/namei.c, follow_managed() is extended to check with d_manage() first
on each managed directory, before transiting away from it or attempting to
automount upon it.
follow_down() is renamed follow_down_one() and should only be used where the
filesystem deliberately intends to avoid management steps (e.g. autofs).
A new follow_down() is added that incorporates the loop done by all other
callers of follow_down() (do_add/move_mount(), autofs and NFSD; whilst AFS, NFS
and CIFS do use it, their use is removed by converting them to use
d_automount()). The new follow_down() calls d_manage() as appropriate. It
also takes an extra parameter to indicate if it is being called from mount code
(with namespace_sem writelocked) which it passes to d_manage(). follow_down()
ignores automount points so that it can be used to mount on them.
__follow_mount_rcu() is made to abort rcu-walk mode if it hits a directory with
DCACHE_MANAGE_TRANSIT set on the basis that we're probably going to have to
sleep. It would be possible to enter d_manage() in rcu-walk mode too, and have
that determine whether to abort or not itself. That would allow the autofs
daemon to continue on in rcu-walk mode.
Note that DCACHE_MANAGE_TRANSIT on a directory should be cleared when it isn't
required as every tranist from that directory will cause d_manage() to be
invoked. It can always be set again when necessary.
==========================
WHAT THIS MEANS FOR AUTOFS
==========================
Autofs currently uses the lookup() inode op and the d_revalidate() dentry op to
trigger the automounting of indirect mounts, and both of these can be called
with i_mutex held.
autofs knows that the i_mutex will be held by the caller in lookup(), and so
can drop it before invoking the daemon - but this isn't so for d_revalidate(),
since the lock is only held on _some_ of the code paths that call it. This
means that autofs can't risk dropping i_mutex from its d_revalidate() function
before it calls the daemon.
The bug could manifest itself as, for example, a process that's trying to
validate an automount dentry that gets made to wait because that dentry is
expired and needs cleaning up:
mkdir S ffffffff8014e05a 0 32580 24956
Call Trace:
[<ffffffff885371fd>] :autofs4:autofs4_wait+0x674/0x897
[<ffffffff80127f7d>] avc_has_perm+0x46/0x58
[<ffffffff8009fdcf>] autoremove_wake_function+0x0/0x2e
[<ffffffff88537be6>] :autofs4:autofs4_expire_wait+0x41/0x6b
[<ffffffff88535cfc>] :autofs4:autofs4_revalidate+0x91/0x149
[<ffffffff80036d96>] __lookup_hash+0xa0/0x12f
[<ffffffff80057a2f>] lookup_create+0x46/0x80
[<ffffffff800e6e31>] sys_mkdirat+0x56/0xe4
versus the automount daemon which wants to remove that dentry, but can't
because the normal process is holding the i_mutex lock:
automount D ffffffff8014e05a 0 32581 1 32561
Call Trace:
[<ffffffff80063c3f>] __mutex_lock_slowpath+0x60/0x9b
[<ffffffff8000ccf1>] do_path_lookup+0x2ca/0x2f1
[<ffffffff80063c89>] .text.lock.mutex+0xf/0x14
[<ffffffff800e6d55>] do_rmdir+0x77/0xde
[<ffffffff8005d229>] tracesys+0x71/0xe0
[<ffffffff8005d28d>] tracesys+0xd5/0xe0
which means that the system is deadlocked.
This patch allows autofs to hold up normal processes whilst the daemon goes
ahead and does things to the dentry tree behind the automouter point without
risking a deadlock as almost no locks are held in d_manage() and none in
d_automount().
Signed-off-by: David Howells <dhowells@redhat.com>
Was-Acked-by: Ian Kent <raven@themaw.net>
Signed-off-by: Al Viro <viro@zeniv.linux.org.uk>
2011-01-14 19:45:26 +01:00
|
|
|
goto out;
|
|
|
|
|
2011-11-25 06:46:35 +01:00
|
|
|
old = real_mount(old_path.mnt);
|
2011-11-25 07:05:37 +01:00
|
|
|
p = real_mount(path->mnt);
|
2011-11-25 06:46:35 +01:00
|
|
|
|
2005-04-17 00:20:36 +02:00
|
|
|
err = -EINVAL;
|
2011-11-25 07:05:37 +01:00
|
|
|
if (!check_mnt(p) || !check_mnt(old))
|
2005-04-17 00:20:36 +02:00
|
|
|
goto out1;
|
|
|
|
|
2013-03-30 05:04:39 +01:00
|
|
|
if (old->mnt.mnt_flags & MNT_LOCKED)
|
|
|
|
goto out1;
|
|
|
|
|
2005-04-17 00:20:36 +02:00
|
|
|
err = -EINVAL;
|
2008-08-02 06:51:11 +02:00
|
|
|
if (old_path.dentry != old_path.mnt->mnt_root)
|
2005-11-07 23:20:03 +01:00
|
|
|
goto out1;
|
2005-04-17 00:20:36 +02:00
|
|
|
|
2011-11-25 03:47:05 +01:00
|
|
|
if (!mnt_has_parent(old))
|
2005-11-07 23:20:03 +01:00
|
|
|
goto out1;
|
2005-04-17 00:20:36 +02:00
|
|
|
|
VFS: (Scripted) Convert S_ISLNK/DIR/REG(dentry->d_inode) to d_is_*(dentry)
Convert the following where appropriate:
(1) S_ISLNK(dentry->d_inode) to d_is_symlink(dentry).
(2) S_ISREG(dentry->d_inode) to d_is_reg(dentry).
(3) S_ISDIR(dentry->d_inode) to d_is_dir(dentry). This is actually more
complicated than it appears as some calls should be converted to
d_can_lookup() instead. The difference is whether the directory in
question is a real dir with a ->lookup op or whether it's a fake dir with
a ->d_automount op.
In some circumstances, we can subsume checks for dentry->d_inode not being
NULL into this, provided we the code isn't in a filesystem that expects
d_inode to be NULL if the dirent really *is* negative (ie. if we're going to
use d_inode() rather than d_backing_inode() to get the inode pointer).
Note that the dentry type field may be set to something other than
DCACHE_MISS_TYPE when d_inode is NULL in the case of unionmount, where the VFS
manages the fall-through from a negative dentry to a lower layer. In such a
case, the dentry type of the negative union dentry is set to the same as the
type of the lower dentry.
However, if you know d_inode is not NULL at the call site, then you can use
the d_is_xxx() functions even in a filesystem.
There is one further complication: a 0,0 chardev dentry may be labelled
DCACHE_WHITEOUT_TYPE rather than DCACHE_SPECIAL_TYPE. Strictly, this was
intended for special directory entry types that don't have attached inodes.
The following perl+coccinelle script was used:
use strict;
my @callers;
open($fd, 'git grep -l \'S_IS[A-Z].*->d_inode\' |') ||
die "Can't grep for S_ISDIR and co. callers";
@callers = <$fd>;
close($fd);
unless (@callers) {
print "No matches\n";
exit(0);
}
my @cocci = (
'@@',
'expression E;',
'@@',
'',
'- S_ISLNK(E->d_inode->i_mode)',
'+ d_is_symlink(E)',
'',
'@@',
'expression E;',
'@@',
'',
'- S_ISDIR(E->d_inode->i_mode)',
'+ d_is_dir(E)',
'',
'@@',
'expression E;',
'@@',
'',
'- S_ISREG(E->d_inode->i_mode)',
'+ d_is_reg(E)' );
my $coccifile = "tmp.sp.cocci";
open($fd, ">$coccifile") || die $coccifile;
print($fd "$_\n") || die $coccifile foreach (@cocci);
close($fd);
foreach my $file (@callers) {
chomp $file;
print "Processing ", $file, "\n";
system("spatch", "--sp-file", $coccifile, $file, "--in-place", "--no-show-diff") == 0 ||
die "spatch failed";
}
[AV: overlayfs parts skipped]
Signed-off-by: David Howells <dhowells@redhat.com>
Signed-off-by: Al Viro <viro@zeniv.linux.org.uk>
2015-01-29 13:02:35 +01:00
|
|
|
if (d_is_dir(path->dentry) !=
|
|
|
|
d_is_dir(old_path.dentry))
|
2005-11-07 23:20:03 +01:00
|
|
|
goto out1;
|
|
|
|
/*
|
|
|
|
* Don't move a mount residing in a shared parent.
|
|
|
|
*/
|
2011-11-25 07:05:37 +01:00
|
|
|
if (IS_MNT_SHARED(old->mnt_parent))
|
2005-11-07 23:20:03 +01:00
|
|
|
goto out1;
|
2005-11-07 23:21:20 +01:00
|
|
|
/*
|
|
|
|
* Don't move a mount tree containing unbindable mounts to a destination
|
|
|
|
* mount which is shared.
|
|
|
|
*/
|
2011-11-25 07:05:37 +01:00
|
|
|
if (IS_MNT_SHARED(p) && tree_contains_unbindable(old))
|
2005-11-07 23:21:20 +01:00
|
|
|
goto out1;
|
2005-04-17 00:20:36 +02:00
|
|
|
err = -ELOOP;
|
2011-11-25 07:05:37 +01:00
|
|
|
for (; mnt_has_parent(p); p = p->mnt_parent)
|
2011-11-25 03:47:05 +01:00
|
|
|
if (p == old)
|
2005-11-07 23:20:03 +01:00
|
|
|
goto out1;
|
2005-04-17 00:20:36 +02:00
|
|
|
|
2013-03-15 15:53:28 +01:00
|
|
|
err = attach_recursive_mnt(old, real_mount(path->mnt), mp, &parent_path);
|
2008-02-15 04:34:32 +01:00
|
|
|
if (err)
|
2005-11-07 23:20:03 +01:00
|
|
|
goto out1;
|
2005-04-17 00:20:36 +02:00
|
|
|
|
|
|
|
/* if the mount is moved, it should no longer be expire
|
|
|
|
* automatically */
|
2011-11-25 06:22:05 +01:00
|
|
|
list_del_init(&old->mnt_expire);
|
2005-04-17 00:20:36 +02:00
|
|
|
out1:
|
2013-03-15 15:53:28 +01:00
|
|
|
unlock_mount(mp);
|
2005-04-17 00:20:36 +02:00
|
|
|
out:
|
|
|
|
if (!err)
|
2008-03-22 01:48:19 +01:00
|
|
|
path_put(&parent_path);
|
2008-08-02 06:51:11 +02:00
|
|
|
path_put(&old_path);
|
2005-04-17 00:20:36 +02:00
|
|
|
return err;
|
|
|
|
}
|
|
|
|
|
2011-03-18 03:08:28 +01:00
|
|
|
static struct vfsmount *fs_set_subtype(struct vfsmount *mnt, const char *fstype)
|
|
|
|
{
|
|
|
|
int err;
|
|
|
|
const char *subtype = strchr(fstype, '.');
|
|
|
|
if (subtype) {
|
|
|
|
subtype++;
|
|
|
|
err = -EINVAL;
|
|
|
|
if (!subtype[0])
|
|
|
|
goto err;
|
|
|
|
} else
|
|
|
|
subtype = "";
|
|
|
|
|
|
|
|
mnt->mnt_sb->s_subtype = kstrdup(subtype, GFP_KERNEL);
|
|
|
|
err = -ENOMEM;
|
|
|
|
if (!mnt->mnt_sb->s_subtype)
|
|
|
|
goto err;
|
|
|
|
return mnt;
|
|
|
|
|
|
|
|
err:
|
|
|
|
mntput(mnt);
|
|
|
|
return ERR_PTR(err);
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* add a mount into a namespace's mount tree
|
|
|
|
*/
|
2011-11-25 06:30:56 +01:00
|
|
|
static int do_add_mount(struct mount *newmnt, struct path *path, int mnt_flags)
|
2011-03-18 03:08:28 +01:00
|
|
|
{
|
2013-03-15 15:53:28 +01:00
|
|
|
struct mountpoint *mp;
|
|
|
|
struct mount *parent;
|
2011-03-18 03:08:28 +01:00
|
|
|
int err;
|
|
|
|
|
smarter propagate_mnt()
The current mainline has copies propagated to *all* nodes, then
tears down the copies we made for nodes that do not contain
counterparts of the desired mountpoint. That sets the right
propagation graph for the copies (at teardown time we move
the slaves of removed node to a surviving peer or directly
to master), but we end up paying a fairly steep price in
useless allocations. It's fairly easy to create a situation
where N calls of mount(2) create exactly N bindings, with
O(N^2) vfsmounts allocated and freed in process.
Fortunately, it is possible to avoid those allocations/freeings.
The trick is to create copies in the right order and find which
one would've eventually become a master with the current algorithm.
It turns out to be possible in O(nodes getting propagation) time
and with no extra allocations at all.
One part is that we need to make sure that eventual master will be
created before its slaves, so we need to walk the propagation
tree in a different order - by peer groups. And iterate through
the peers before dealing with the next group.
Another thing is finding the (earlier) copy that will be a master
of one we are about to create; to do that we are (temporary) marking
the masters of mountpoints we are attaching the copies to.
Either we are in a peer of the last mountpoint we'd dealt with,
or we have the following situation: we are attaching to mountpoint M,
the last copy S_0 had been attached to M_0 and there are sequences
S_0...S_n, M_0...M_n such that S_{i+1} is a master of S_{i},
S_{i} mounted on M{i} and we need to create a slave of the first S_{k}
such that M is getting propagation from M_{k}. It means that the master
of M_{k} will be among the sequence of masters of M. On the
other hand, the nearest marked node in that sequence will either
be the master of M_{k} or the master of M_{k-1} (the latter -
in the case if M_{k-1} is a slave of something M gets propagation
from, but in a wrong peer group).
So we go through the sequence of masters of M until we find
a marked one (P). Let N be the one before it. Then we go through
the sequence of masters of S_0 until we find one (say, S) mounted
on a node D that has P as master and check if D is a peer of N.
If it is, S will be the master of new copy, if not - the master of S
will be.
That's it for the hard part; the rest is fairly simple. Iterator
is in next_group(), handling of one prospective mountpoint is
propagate_one().
It seems to survive all tests and gives a noticably better performance
than the current mainline for setups that are seriously using shared
subtrees.
Cc: stable@vger.kernel.org
Signed-off-by: Al Viro <viro@zeniv.linux.org.uk>
2014-02-27 15:35:45 +01:00
|
|
|
mnt_flags &= ~MNT_INTERNAL_FLAGS;
|
2011-03-18 03:08:28 +01:00
|
|
|
|
2013-03-15 15:53:28 +01:00
|
|
|
mp = lock_mount(path);
|
|
|
|
if (IS_ERR(mp))
|
|
|
|
return PTR_ERR(mp);
|
2011-03-18 03:08:28 +01:00
|
|
|
|
2013-03-15 15:53:28 +01:00
|
|
|
parent = real_mount(path->mnt);
|
2011-03-18 03:08:28 +01:00
|
|
|
err = -EINVAL;
|
2013-03-15 15:53:28 +01:00
|
|
|
if (unlikely(!check_mnt(parent))) {
|
2012-09-21 14:19:02 +02:00
|
|
|
/* that's acceptable only for automounts done in private ns */
|
|
|
|
if (!(mnt_flags & MNT_SHRINKABLE))
|
|
|
|
goto unlock;
|
|
|
|
/* ... and for those we'd better have mountpoint still alive */
|
2013-03-15 15:53:28 +01:00
|
|
|
if (!parent->mnt_ns)
|
2012-09-21 14:19:02 +02:00
|
|
|
goto unlock;
|
|
|
|
}
|
2011-03-18 03:08:28 +01:00
|
|
|
|
|
|
|
/* Refuse the same filesystem on the same mount point */
|
|
|
|
err = -EBUSY;
|
2011-11-25 06:30:56 +01:00
|
|
|
if (path->mnt->mnt_sb == newmnt->mnt.mnt_sb &&
|
2011-03-18 03:08:28 +01:00
|
|
|
path->mnt->mnt_root == path->dentry)
|
|
|
|
goto unlock;
|
|
|
|
|
|
|
|
err = -EINVAL;
|
VFS: (Scripted) Convert S_ISLNK/DIR/REG(dentry->d_inode) to d_is_*(dentry)
Convert the following where appropriate:
(1) S_ISLNK(dentry->d_inode) to d_is_symlink(dentry).
(2) S_ISREG(dentry->d_inode) to d_is_reg(dentry).
(3) S_ISDIR(dentry->d_inode) to d_is_dir(dentry). This is actually more
complicated than it appears as some calls should be converted to
d_can_lookup() instead. The difference is whether the directory in
question is a real dir with a ->lookup op or whether it's a fake dir with
a ->d_automount op.
In some circumstances, we can subsume checks for dentry->d_inode not being
NULL into this, provided we the code isn't in a filesystem that expects
d_inode to be NULL if the dirent really *is* negative (ie. if we're going to
use d_inode() rather than d_backing_inode() to get the inode pointer).
Note that the dentry type field may be set to something other than
DCACHE_MISS_TYPE when d_inode is NULL in the case of unionmount, where the VFS
manages the fall-through from a negative dentry to a lower layer. In such a
case, the dentry type of the negative union dentry is set to the same as the
type of the lower dentry.
However, if you know d_inode is not NULL at the call site, then you can use
the d_is_xxx() functions even in a filesystem.
There is one further complication: a 0,0 chardev dentry may be labelled
DCACHE_WHITEOUT_TYPE rather than DCACHE_SPECIAL_TYPE. Strictly, this was
intended for special directory entry types that don't have attached inodes.
The following perl+coccinelle script was used:
use strict;
my @callers;
open($fd, 'git grep -l \'S_IS[A-Z].*->d_inode\' |') ||
die "Can't grep for S_ISDIR and co. callers";
@callers = <$fd>;
close($fd);
unless (@callers) {
print "No matches\n";
exit(0);
}
my @cocci = (
'@@',
'expression E;',
'@@',
'',
'- S_ISLNK(E->d_inode->i_mode)',
'+ d_is_symlink(E)',
'',
'@@',
'expression E;',
'@@',
'',
'- S_ISDIR(E->d_inode->i_mode)',
'+ d_is_dir(E)',
'',
'@@',
'expression E;',
'@@',
'',
'- S_ISREG(E->d_inode->i_mode)',
'+ d_is_reg(E)' );
my $coccifile = "tmp.sp.cocci";
open($fd, ">$coccifile") || die $coccifile;
print($fd "$_\n") || die $coccifile foreach (@cocci);
close($fd);
foreach my $file (@callers) {
chomp $file;
print "Processing ", $file, "\n";
system("spatch", "--sp-file", $coccifile, $file, "--in-place", "--no-show-diff") == 0 ||
die "spatch failed";
}
[AV: overlayfs parts skipped]
Signed-off-by: David Howells <dhowells@redhat.com>
Signed-off-by: Al Viro <viro@zeniv.linux.org.uk>
2015-01-29 13:02:35 +01:00
|
|
|
if (d_is_symlink(newmnt->mnt.mnt_root))
|
2011-03-18 03:08:28 +01:00
|
|
|
goto unlock;
|
|
|
|
|
2011-11-25 06:30:56 +01:00
|
|
|
newmnt->mnt.mnt_flags = mnt_flags;
|
2013-03-15 15:53:28 +01:00
|
|
|
err = graft_tree(newmnt, parent, mp);
|
2011-03-18 03:08:28 +01:00
|
|
|
|
|
|
|
unlock:
|
2013-03-15 15:53:28 +01:00
|
|
|
unlock_mount(mp);
|
2011-03-18 03:08:28 +01:00
|
|
|
return err;
|
|
|
|
}
|
2011-01-17 07:47:59 +01:00
|
|
|
|
2016-06-09 23:06:06 +02:00
|
|
|
static bool mount_too_revealing(struct vfsmount *mnt, int *new_mnt_flags);
|
2015-05-09 06:22:29 +02:00
|
|
|
|
2005-04-17 00:20:36 +02:00
|
|
|
/*
|
|
|
|
* create a new mount for userspace and request it to be added into the
|
|
|
|
* namespace's tree
|
|
|
|
*/
|
2017-07-17 09:45:35 +02:00
|
|
|
static int do_new_mount(struct path *path, const char *fstype, int sb_flags,
|
2012-10-11 17:42:01 +02:00
|
|
|
int mnt_flags, const char *name, void *data)
|
2005-04-17 00:20:36 +02:00
|
|
|
{
|
2012-07-27 06:42:03 +02:00
|
|
|
struct file_system_type *type;
|
2005-04-17 00:20:36 +02:00
|
|
|
struct vfsmount *mnt;
|
2011-01-17 07:41:58 +01:00
|
|
|
int err;
|
2005-04-17 00:20:36 +02:00
|
|
|
|
2012-07-27 06:42:03 +02:00
|
|
|
if (!fstype)
|
2005-04-17 00:20:36 +02:00
|
|
|
return -EINVAL;
|
|
|
|
|
2012-07-27 06:42:03 +02:00
|
|
|
type = get_fs_type(fstype);
|
|
|
|
if (!type)
|
|
|
|
return -ENODEV;
|
|
|
|
|
2017-07-17 09:45:35 +02:00
|
|
|
mnt = vfs_kern_mount(type, sb_flags, name, data);
|
2012-07-27 06:42:03 +02:00
|
|
|
if (!IS_ERR(mnt) && (type->fs_flags & FS_HAS_SUBTYPE) &&
|
|
|
|
!mnt->mnt_sb->s_subtype)
|
|
|
|
mnt = fs_set_subtype(mnt, fstype);
|
|
|
|
|
|
|
|
put_filesystem(type);
|
2005-04-17 00:20:36 +02:00
|
|
|
if (IS_ERR(mnt))
|
|
|
|
return PTR_ERR(mnt);
|
|
|
|
|
2016-06-09 23:06:06 +02:00
|
|
|
if (mount_too_revealing(mnt, &mnt_flags)) {
|
|
|
|
mntput(mnt);
|
|
|
|
return -EPERM;
|
|
|
|
}
|
|
|
|
|
2011-11-25 06:30:56 +01:00
|
|
|
err = do_add_mount(real_mount(mnt), path, mnt_flags);
|
2011-01-17 07:41:58 +01:00
|
|
|
if (err)
|
|
|
|
mntput(mnt);
|
|
|
|
return err;
|
2005-04-17 00:20:36 +02:00
|
|
|
}
|
|
|
|
|
2011-01-17 07:35:23 +01:00
|
|
|
int finish_automount(struct vfsmount *m, struct path *path)
|
|
|
|
{
|
2011-11-25 06:22:05 +01:00
|
|
|
struct mount *mnt = real_mount(m);
|
2011-01-17 07:35:23 +01:00
|
|
|
int err;
|
|
|
|
/* The new mount record should have at least 2 refs to prevent it being
|
|
|
|
* expired before we get a chance to add it
|
|
|
|
*/
|
2011-11-25 06:22:05 +01:00
|
|
|
BUG_ON(mnt_get_count(mnt) < 2);
|
2011-01-17 07:35:23 +01:00
|
|
|
|
|
|
|
if (m->mnt_sb == path->mnt->mnt_sb &&
|
|
|
|
m->mnt_root == path->dentry) {
|
2011-01-17 07:47:59 +01:00
|
|
|
err = -ELOOP;
|
|
|
|
goto fail;
|
2011-01-17 07:35:23 +01:00
|
|
|
}
|
|
|
|
|
2011-11-25 06:30:56 +01:00
|
|
|
err = do_add_mount(mnt, path, path->mnt->mnt_flags | MNT_SHRINKABLE);
|
2011-01-17 07:47:59 +01:00
|
|
|
if (!err)
|
|
|
|
return 0;
|
|
|
|
fail:
|
|
|
|
/* remove m from any expiration list it may be on */
|
2011-11-25 06:22:05 +01:00
|
|
|
if (!list_empty(&mnt->mnt_expire)) {
|
2013-03-16 20:12:40 +01:00
|
|
|
namespace_lock();
|
2011-11-25 06:22:05 +01:00
|
|
|
list_del_init(&mnt->mnt_expire);
|
2013-03-16 20:12:40 +01:00
|
|
|
namespace_unlock();
|
2011-01-17 07:35:23 +01:00
|
|
|
}
|
2011-01-17 07:47:59 +01:00
|
|
|
mntput(m);
|
|
|
|
mntput(m);
|
2011-01-17 07:35:23 +01:00
|
|
|
return err;
|
|
|
|
}
|
|
|
|
|
2011-01-14 20:10:03 +01:00
|
|
|
/**
|
|
|
|
* mnt_set_expiry - Put a mount on an expiration list
|
|
|
|
* @mnt: The mount to list.
|
|
|
|
* @expiry_list: The list to add the mount to.
|
|
|
|
*/
|
|
|
|
void mnt_set_expiry(struct vfsmount *mnt, struct list_head *expiry_list)
|
|
|
|
{
|
2013-03-16 20:12:40 +01:00
|
|
|
namespace_lock();
|
2011-01-14 20:10:03 +01:00
|
|
|
|
2011-11-25 06:22:05 +01:00
|
|
|
list_add_tail(&real_mount(mnt)->mnt_expire, expiry_list);
|
2011-01-14 20:10:03 +01:00
|
|
|
|
2013-03-16 20:12:40 +01:00
|
|
|
namespace_unlock();
|
2011-01-14 20:10:03 +01:00
|
|
|
}
|
|
|
|
EXPORT_SYMBOL(mnt_set_expiry);
|
|
|
|
|
2005-04-17 00:20:36 +02:00
|
|
|
/*
|
|
|
|
* process a list of expirable mountpoints with the intent of discarding any
|
|
|
|
* mountpoints that aren't in use and haven't been touched since last we came
|
|
|
|
* here
|
|
|
|
*/
|
|
|
|
void mark_mounts_for_expiry(struct list_head *mounts)
|
|
|
|
{
|
2011-11-25 03:07:43 +01:00
|
|
|
struct mount *mnt, *next;
|
2005-04-17 00:20:36 +02:00
|
|
|
LIST_HEAD(graveyard);
|
|
|
|
|
|
|
|
if (list_empty(mounts))
|
|
|
|
return;
|
|
|
|
|
2013-03-16 20:12:40 +01:00
|
|
|
namespace_lock();
|
2013-09-29 17:24:49 +02:00
|
|
|
lock_mount_hash();
|
2005-04-17 00:20:36 +02:00
|
|
|
|
|
|
|
/* extract from the expiration list every vfsmount that matches the
|
|
|
|
* following criteria:
|
|
|
|
* - only referenced by its parent vfsmount
|
|
|
|
* - still marked for expiry (marked on the last call here; marks are
|
|
|
|
* cleared by mntput())
|
|
|
|
*/
|
2011-11-25 06:22:05 +01:00
|
|
|
list_for_each_entry_safe(mnt, next, mounts, mnt_expire) {
|
2011-11-25 06:57:42 +01:00
|
|
|
if (!xchg(&mnt->mnt_expiry_mark, 1) ||
|
2011-11-25 03:35:16 +01:00
|
|
|
propagate_mount_busy(mnt, 1))
|
2005-04-17 00:20:36 +02:00
|
|
|
continue;
|
2011-11-25 06:22:05 +01:00
|
|
|
list_move(&mnt->mnt_expire, &graveyard);
|
2005-04-17 00:20:36 +02:00
|
|
|
}
|
2008-03-22 05:21:53 +01:00
|
|
|
while (!list_empty(&graveyard)) {
|
2011-11-25 06:22:05 +01:00
|
|
|
mnt = list_first_entry(&graveyard, struct mount, mnt_expire);
|
2011-11-25 06:46:35 +01:00
|
|
|
touch_mnt_namespace(mnt->mnt_ns);
|
2014-12-24 14:20:01 +01:00
|
|
|
umount_tree(mnt, UMOUNT_PROPAGATE|UMOUNT_SYNC);
|
2008-03-22 05:21:53 +01:00
|
|
|
}
|
2013-09-29 17:24:49 +02:00
|
|
|
unlock_mount_hash();
|
2013-03-16 19:42:19 +01:00
|
|
|
namespace_unlock();
|
2006-06-09 15:34:17 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
EXPORT_SYMBOL_GPL(mark_mounts_for_expiry);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Ripoff of 'select_parent()'
|
|
|
|
*
|
|
|
|
* search the list of submounts for a given mountpoint, and move any
|
|
|
|
* shrinkable submounts to the 'graveyard' list.
|
|
|
|
*/
|
2011-11-25 03:15:14 +01:00
|
|
|
static int select_submounts(struct mount *parent, struct list_head *graveyard)
|
2006-06-09 15:34:17 +02:00
|
|
|
{
|
2011-11-25 03:15:14 +01:00
|
|
|
struct mount *this_parent = parent;
|
2006-06-09 15:34:17 +02:00
|
|
|
struct list_head *next;
|
|
|
|
int found = 0;
|
|
|
|
|
|
|
|
repeat:
|
2011-11-25 05:24:33 +01:00
|
|
|
next = this_parent->mnt_mounts.next;
|
2006-06-09 15:34:17 +02:00
|
|
|
resume:
|
2011-11-25 05:24:33 +01:00
|
|
|
while (next != &this_parent->mnt_mounts) {
|
2006-06-09 15:34:17 +02:00
|
|
|
struct list_head *tmp = next;
|
2011-11-25 05:24:33 +01:00
|
|
|
struct mount *mnt = list_entry(tmp, struct mount, mnt_child);
|
2006-06-09 15:34:17 +02:00
|
|
|
|
|
|
|
next = tmp->next;
|
2011-11-25 03:15:14 +01:00
|
|
|
if (!(mnt->mnt.mnt_flags & MNT_SHRINKABLE))
|
2005-04-17 00:20:36 +02:00
|
|
|
continue;
|
2006-06-09 15:34:17 +02:00
|
|
|
/*
|
|
|
|
* Descend a level if the d_mounts list is non-empty.
|
|
|
|
*/
|
2011-11-25 05:24:33 +01:00
|
|
|
if (!list_empty(&mnt->mnt_mounts)) {
|
2006-06-09 15:34:17 +02:00
|
|
|
this_parent = mnt;
|
|
|
|
goto repeat;
|
|
|
|
}
|
2005-04-17 00:20:36 +02:00
|
|
|
|
2011-11-25 03:35:16 +01:00
|
|
|
if (!propagate_mount_busy(mnt, 1)) {
|
2011-11-25 06:22:05 +01:00
|
|
|
list_move_tail(&mnt->mnt_expire, graveyard);
|
2006-06-09 15:34:17 +02:00
|
|
|
found++;
|
|
|
|
}
|
2005-04-17 00:20:36 +02:00
|
|
|
}
|
2006-06-09 15:34:17 +02:00
|
|
|
/*
|
|
|
|
* All done at this level ... ascend and resume the search
|
|
|
|
*/
|
|
|
|
if (this_parent != parent) {
|
2011-11-25 05:24:33 +01:00
|
|
|
next = this_parent->mnt_child.next;
|
2011-11-25 04:19:58 +01:00
|
|
|
this_parent = this_parent->mnt_parent;
|
2006-06-09 15:34:17 +02:00
|
|
|
goto resume;
|
|
|
|
}
|
|
|
|
return found;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* process a list of expirable mountpoints with the intent of discarding any
|
|
|
|
* submounts of a specific parent mountpoint
|
fs: brlock vfsmount_lock
fs: brlock vfsmount_lock
Use a brlock for the vfsmount lock. It must be taken for write whenever
modifying the mount hash or associated fields, and may be taken for read when
performing mount hash lookups.
A new lock is added for the mnt-id allocator, so it doesn't need to take
the heavy vfsmount write-lock.
The number of atomics should remain the same for fastpath rlock cases, though
code would be slightly slower due to per-cpu access. Scalability is not not be
much improved in common cases yet, due to other locks (ie. dcache_lock) getting
in the way. However path lookups crossing mountpoints should be one case where
scalability is improved (currently requiring the global lock).
The slowpath is slower due to use of brlock. On a 64 core, 64 socket, 32 node
Altix system (high latency to remote nodes), a simple umount microbenchmark
(mount --bind mnt mnt2 ; umount mnt2 loop 1000 times), before this patch it
took 6.8s, afterwards took 7.1s, about 5% slower.
Cc: Al Viro <viro@ZenIV.linux.org.uk>
Signed-off-by: Nick Piggin <npiggin@kernel.dk>
Signed-off-by: Al Viro <viro@zeniv.linux.org.uk>
2010-08-17 20:37:39 +02:00
|
|
|
*
|
2013-09-30 04:06:07 +02:00
|
|
|
* mount_lock must be held for write
|
2006-06-09 15:34:17 +02:00
|
|
|
*/
|
2013-03-16 19:39:34 +01:00
|
|
|
static void shrink_submounts(struct mount *mnt)
|
2006-06-09 15:34:17 +02:00
|
|
|
{
|
|
|
|
LIST_HEAD(graveyard);
|
2011-11-25 03:07:43 +01:00
|
|
|
struct mount *m;
|
2006-06-09 15:34:17 +02:00
|
|
|
|
|
|
|
/* extract submounts of 'mountpoint' from the expiration list */
|
2008-03-22 05:46:23 +01:00
|
|
|
while (select_submounts(mnt, &graveyard)) {
|
2008-03-22 05:21:53 +01:00
|
|
|
while (!list_empty(&graveyard)) {
|
2011-11-25 03:07:43 +01:00
|
|
|
m = list_first_entry(&graveyard, struct mount,
|
2011-11-25 06:22:05 +01:00
|
|
|
mnt_expire);
|
2011-11-25 06:46:35 +01:00
|
|
|
touch_mnt_namespace(m->mnt_ns);
|
2014-12-24 14:20:01 +01:00
|
|
|
umount_tree(m, UMOUNT_PROPAGATE|UMOUNT_SYNC);
|
2008-03-22 05:21:53 +01:00
|
|
|
}
|
|
|
|
}
|
2005-04-17 00:20:36 +02:00
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Some copy_from_user() implementations do not return the exact number of
|
|
|
|
* bytes remaining to copy on a fault. But copy_mount_options() requires that.
|
|
|
|
* Note that this function differs from copy_from_user() in that it will oops
|
|
|
|
* on bad values of `to', rather than returning a short copy.
|
|
|
|
*/
|
2005-11-07 23:16:09 +01:00
|
|
|
static long exact_copy_from_user(void *to, const void __user * from,
|
|
|
|
unsigned long n)
|
2005-04-17 00:20:36 +02:00
|
|
|
{
|
|
|
|
char *t = to;
|
|
|
|
const char __user *f = from;
|
|
|
|
char c;
|
|
|
|
|
|
|
|
if (!access_ok(VERIFY_READ, from, n))
|
|
|
|
return n;
|
|
|
|
|
|
|
|
while (n) {
|
|
|
|
if (__get_user(c, f)) {
|
|
|
|
memset(t, 0, n);
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
*t++ = c;
|
|
|
|
f++;
|
|
|
|
n--;
|
|
|
|
}
|
|
|
|
return n;
|
|
|
|
}
|
|
|
|
|
2015-12-15 00:44:44 +01:00
|
|
|
void *copy_mount_options(const void __user * data)
|
2005-04-17 00:20:36 +02:00
|
|
|
{
|
|
|
|
int i;
|
|
|
|
unsigned long size;
|
2015-12-15 00:44:44 +01:00
|
|
|
char *copy;
|
2005-11-07 23:16:09 +01:00
|
|
|
|
2005-04-17 00:20:36 +02:00
|
|
|
if (!data)
|
2015-12-15 00:44:44 +01:00
|
|
|
return NULL;
|
2005-04-17 00:20:36 +02:00
|
|
|
|
2015-12-15 00:44:44 +01:00
|
|
|
copy = kmalloc(PAGE_SIZE, GFP_KERNEL);
|
|
|
|
if (!copy)
|
|
|
|
return ERR_PTR(-ENOMEM);
|
2005-04-17 00:20:36 +02:00
|
|
|
|
|
|
|
/* We only care that *some* data at the address the user
|
|
|
|
* gave us is valid. Just in case, we'll zero
|
|
|
|
* the remainder of the page.
|
|
|
|
*/
|
|
|
|
/* copy_from_user cannot cross TASK_SIZE ! */
|
|
|
|
size = TASK_SIZE - (unsigned long)data;
|
|
|
|
if (size > PAGE_SIZE)
|
|
|
|
size = PAGE_SIZE;
|
|
|
|
|
2015-12-15 00:44:44 +01:00
|
|
|
i = size - exact_copy_from_user(copy, data, size);
|
2005-04-17 00:20:36 +02:00
|
|
|
if (!i) {
|
2015-12-15 00:44:44 +01:00
|
|
|
kfree(copy);
|
|
|
|
return ERR_PTR(-EFAULT);
|
2005-04-17 00:20:36 +02:00
|
|
|
}
|
|
|
|
if (i != PAGE_SIZE)
|
2015-12-15 00:44:44 +01:00
|
|
|
memset(copy + i, 0, PAGE_SIZE - i);
|
|
|
|
return copy;
|
2005-04-17 00:20:36 +02:00
|
|
|
}
|
|
|
|
|
2014-08-28 19:26:03 +02:00
|
|
|
char *copy_mount_string(const void __user *data)
|
fs: fix overflow in sys_mount() for in-kernel calls
sys_mount() reads/copies a whole page for its "type" parameter. When
do_mount_root() passes a kernel address that points to an object which is
smaller than a whole page, copy_mount_options() will happily go past this
memory object, possibly dereferencing "wild" pointers that could be in any
state (hence the kmemcheck warning, which shows that parts of the next
page are not even allocated).
(The likelihood of something going wrong here is pretty low -- first of
all this only applies to kernel calls to sys_mount(), which are mostly
found in the boot code. Secondly, I guess if the page was not mapped,
exact_copy_from_user() _would_ in fact handle it correctly because of its
access_ok(), etc. checks.)
But it is much nicer to avoid the dubious reads altogether, by stopping as
soon as we find a NUL byte. Is there a good reason why we can't do
something like this, using the already existing strndup_from_user()?
[akpm@linux-foundation.org: make copy_mount_string() static]
[AV: fix compat mount breakage, which involves undoing akpm's change above]
Reported-by: Ingo Molnar <mingo@elte.hu>
Signed-off-by: Vegard Nossum <vegard.nossum@gmail.com>
Cc: Al Viro <viro@zeniv.linux.org.uk>
Cc: Pekka Enberg <penberg@cs.helsinki.fi>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: al <al@dizzy.pdmi.ras.ru>
2009-09-18 22:05:45 +02:00
|
|
|
{
|
2014-08-28 19:26:03 +02:00
|
|
|
return data ? strndup_user(data, PAGE_SIZE) : NULL;
|
fs: fix overflow in sys_mount() for in-kernel calls
sys_mount() reads/copies a whole page for its "type" parameter. When
do_mount_root() passes a kernel address that points to an object which is
smaller than a whole page, copy_mount_options() will happily go past this
memory object, possibly dereferencing "wild" pointers that could be in any
state (hence the kmemcheck warning, which shows that parts of the next
page are not even allocated).
(The likelihood of something going wrong here is pretty low -- first of
all this only applies to kernel calls to sys_mount(), which are mostly
found in the boot code. Secondly, I guess if the page was not mapped,
exact_copy_from_user() _would_ in fact handle it correctly because of its
access_ok(), etc. checks.)
But it is much nicer to avoid the dubious reads altogether, by stopping as
soon as we find a NUL byte. Is there a good reason why we can't do
something like this, using the already existing strndup_from_user()?
[akpm@linux-foundation.org: make copy_mount_string() static]
[AV: fix compat mount breakage, which involves undoing akpm's change above]
Reported-by: Ingo Molnar <mingo@elte.hu>
Signed-off-by: Vegard Nossum <vegard.nossum@gmail.com>
Cc: Al Viro <viro@zeniv.linux.org.uk>
Cc: Pekka Enberg <penberg@cs.helsinki.fi>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: al <al@dizzy.pdmi.ras.ru>
2009-09-18 22:05:45 +02:00
|
|
|
}
|
|
|
|
|
2005-04-17 00:20:36 +02:00
|
|
|
/*
|
|
|
|
* Flags is a 32-bit value that allows up to 31 non-fs dependent flags to
|
|
|
|
* be given to the mount() call (ie: read-only, no-dev, no-suid etc).
|
|
|
|
*
|
|
|
|
* data is a (void *) that can point to any structure up to
|
|
|
|
* PAGE_SIZE-1 bytes, which can contain arbitrary fs-dependent
|
|
|
|
* information (or be NULL).
|
|
|
|
*
|
|
|
|
* Pre-0.97 versions of mount() didn't have a flags word.
|
|
|
|
* When the flags word was introduced its top half was required
|
|
|
|
* to have the magic value 0xC0ED, and this remained so until 2.4.0-test9.
|
|
|
|
* Therefore, if this magic number is present, it carries no information
|
|
|
|
* and must be discarded.
|
|
|
|
*/
|
2014-09-14 15:15:10 +02:00
|
|
|
long do_mount(const char *dev_name, const char __user *dir_name,
|
2012-10-11 17:42:01 +02:00
|
|
|
const char *type_page, unsigned long flags, void *data_page)
|
2005-04-17 00:20:36 +02:00
|
|
|
{
|
2008-08-02 06:51:11 +02:00
|
|
|
struct path path;
|
2017-07-17 09:45:35 +02:00
|
|
|
unsigned int mnt_flags = 0, sb_flags;
|
2005-04-17 00:20:36 +02:00
|
|
|
int retval = 0;
|
|
|
|
|
|
|
|
/* Discard magic */
|
|
|
|
if ((flags & MS_MGC_MSK) == MS_MGC_VAL)
|
|
|
|
flags &= ~MS_MGC_MSK;
|
|
|
|
|
|
|
|
/* Basic sanity checks */
|
|
|
|
if (data_page)
|
|
|
|
((char *)data_page)[PAGE_SIZE - 1] = 0;
|
|
|
|
|
2017-07-17 09:45:35 +02:00
|
|
|
if (flags & MS_NOUSER)
|
|
|
|
return -EINVAL;
|
|
|
|
|
2009-10-04 14:49:49 +02:00
|
|
|
/* ... and get the mountpoint */
|
2014-09-14 15:15:10 +02:00
|
|
|
retval = user_path(dir_name, &path);
|
2009-10-04 14:49:49 +02:00
|
|
|
if (retval)
|
|
|
|
return retval;
|
|
|
|
|
|
|
|
retval = security_sb_mount(dev_name, &path,
|
|
|
|
type_page, flags, data_page);
|
2013-05-04 20:40:51 +02:00
|
|
|
if (!retval && !may_mount())
|
|
|
|
retval = -EPERM;
|
2017-07-17 09:45:35 +02:00
|
|
|
if (!retval && (flags & SB_MANDLOCK) && !may_mandlock())
|
2015-11-16 15:49:34 +01:00
|
|
|
retval = -EPERM;
|
2009-10-04 14:49:49 +02:00
|
|
|
if (retval)
|
|
|
|
goto dput_out;
|
|
|
|
|
2009-04-19 18:40:43 +02:00
|
|
|
/* Default to relatime unless overriden */
|
|
|
|
if (!(flags & MS_NOATIME))
|
|
|
|
mnt_flags |= MNT_RELATIME;
|
2009-03-26 18:53:14 +01:00
|
|
|
|
2005-04-17 00:20:36 +02:00
|
|
|
/* Separate the per-mountpoint flags */
|
|
|
|
if (flags & MS_NOSUID)
|
|
|
|
mnt_flags |= MNT_NOSUID;
|
|
|
|
if (flags & MS_NODEV)
|
|
|
|
mnt_flags |= MNT_NODEV;
|
|
|
|
if (flags & MS_NOEXEC)
|
|
|
|
mnt_flags |= MNT_NOEXEC;
|
2006-01-10 05:52:17 +01:00
|
|
|
if (flags & MS_NOATIME)
|
|
|
|
mnt_flags |= MNT_NOATIME;
|
|
|
|
if (flags & MS_NODIRATIME)
|
|
|
|
mnt_flags |= MNT_NODIRATIME;
|
2009-03-26 18:49:56 +01:00
|
|
|
if (flags & MS_STRICTATIME)
|
|
|
|
mnt_flags &= ~(MNT_RELATIME | MNT_NOATIME);
|
2017-07-17 09:45:35 +02:00
|
|
|
if (flags & SB_RDONLY)
|
2008-02-15 23:38:00 +01:00
|
|
|
mnt_flags |= MNT_READONLY;
|
2006-01-10 05:52:17 +01:00
|
|
|
|
2014-07-29 02:36:04 +02:00
|
|
|
/* The default atime for remount is preservation */
|
|
|
|
if ((flags & MS_REMOUNT) &&
|
|
|
|
((flags & (MS_NOATIME | MS_NODIRATIME | MS_RELATIME |
|
|
|
|
MS_STRICTATIME)) == 0)) {
|
|
|
|
mnt_flags &= ~MNT_ATIME_MASK;
|
|
|
|
mnt_flags |= path.mnt->mnt_flags & MNT_ATIME_MASK;
|
|
|
|
}
|
|
|
|
|
2017-07-17 09:45:35 +02:00
|
|
|
sb_flags = flags & (SB_RDONLY |
|
|
|
|
SB_SYNCHRONOUS |
|
|
|
|
SB_MANDLOCK |
|
|
|
|
SB_DIRSYNC |
|
|
|
|
SB_SILENT |
|
|
|
|
SB_POSIXACL);
|
2005-04-17 00:20:36 +02:00
|
|
|
|
|
|
|
if (flags & MS_REMOUNT)
|
2017-07-17 09:45:35 +02:00
|
|
|
retval = do_remount(&path, flags, sb_flags, mnt_flags,
|
2005-04-17 00:20:36 +02:00
|
|
|
data_page);
|
|
|
|
else if (flags & MS_BIND)
|
2008-08-02 06:51:11 +02:00
|
|
|
retval = do_loopback(&path, dev_name, flags & MS_REC);
|
2005-11-07 23:21:20 +01:00
|
|
|
else if (flags & (MS_SHARED | MS_PRIVATE | MS_SLAVE | MS_UNBINDABLE))
|
2008-08-02 06:51:11 +02:00
|
|
|
retval = do_change_type(&path, flags);
|
2005-04-17 00:20:36 +02:00
|
|
|
else if (flags & MS_MOVE)
|
2008-08-02 06:51:11 +02:00
|
|
|
retval = do_move_mount(&path, dev_name);
|
2005-04-17 00:20:36 +02:00
|
|
|
else
|
2017-07-17 09:45:35 +02:00
|
|
|
retval = do_new_mount(&path, type_page, sb_flags, mnt_flags,
|
2005-04-17 00:20:36 +02:00
|
|
|
dev_name, data_page);
|
|
|
|
dput_out:
|
2008-08-02 06:51:11 +02:00
|
|
|
path_put(&path);
|
2005-04-17 00:20:36 +02:00
|
|
|
return retval;
|
|
|
|
}
|
|
|
|
|
2016-08-08 21:37:37 +02:00
|
|
|
static struct ucounts *inc_mnt_namespaces(struct user_namespace *ns)
|
|
|
|
{
|
|
|
|
return inc_ucount(ns, current_euid(), UCOUNT_MNT_NAMESPACES);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void dec_mnt_namespaces(struct ucounts *ucounts)
|
|
|
|
{
|
|
|
|
dec_ucount(ucounts, UCOUNT_MNT_NAMESPACES);
|
|
|
|
}
|
|
|
|
|
2012-07-27 06:08:32 +02:00
|
|
|
static void free_mnt_ns(struct mnt_namespace *ns)
|
|
|
|
{
|
2014-11-01 05:45:45 +01:00
|
|
|
ns_free_inum(&ns->ns);
|
2016-08-08 21:37:37 +02:00
|
|
|
dec_mnt_namespaces(ns->ucounts);
|
2012-07-27 06:08:32 +02:00
|
|
|
put_user_ns(ns->user_ns);
|
|
|
|
kfree(ns);
|
|
|
|
}
|
|
|
|
|
2010-03-08 03:49:36 +01:00
|
|
|
/*
|
|
|
|
* Assign a sequence number so we can detect when we attempt to bind
|
|
|
|
* mount a reference to an older mount namespace into the current
|
|
|
|
* mount namespace, preventing reference counting loops. A 64bit
|
|
|
|
* number incrementing at 10Ghz will take 12,427 years to wrap which
|
|
|
|
* is effectively never, so we can ignore the possibility.
|
|
|
|
*/
|
|
|
|
static atomic64_t mnt_ns_seq = ATOMIC64_INIT(1);
|
|
|
|
|
2012-07-27 06:08:32 +02:00
|
|
|
static struct mnt_namespace *alloc_mnt_ns(struct user_namespace *user_ns)
|
2009-06-22 21:09:13 +02:00
|
|
|
{
|
|
|
|
struct mnt_namespace *new_ns;
|
2016-08-08 21:37:37 +02:00
|
|
|
struct ucounts *ucounts;
|
2011-06-15 19:21:48 +02:00
|
|
|
int ret;
|
2009-06-22 21:09:13 +02:00
|
|
|
|
2016-08-08 21:37:37 +02:00
|
|
|
ucounts = inc_mnt_namespaces(user_ns);
|
|
|
|
if (!ucounts)
|
2016-09-22 20:08:36 +02:00
|
|
|
return ERR_PTR(-ENOSPC);
|
2016-08-08 21:37:37 +02:00
|
|
|
|
2009-06-22 21:09:13 +02:00
|
|
|
new_ns = kmalloc(sizeof(struct mnt_namespace), GFP_KERNEL);
|
2016-08-08 21:37:37 +02:00
|
|
|
if (!new_ns) {
|
|
|
|
dec_mnt_namespaces(ucounts);
|
2009-06-22 21:09:13 +02:00
|
|
|
return ERR_PTR(-ENOMEM);
|
2016-08-08 21:37:37 +02:00
|
|
|
}
|
2014-11-01 05:45:45 +01:00
|
|
|
ret = ns_alloc_inum(&new_ns->ns);
|
2011-06-15 19:21:48 +02:00
|
|
|
if (ret) {
|
|
|
|
kfree(new_ns);
|
2016-08-08 21:37:37 +02:00
|
|
|
dec_mnt_namespaces(ucounts);
|
2011-06-15 19:21:48 +02:00
|
|
|
return ERR_PTR(ret);
|
|
|
|
}
|
2014-11-01 07:32:53 +01:00
|
|
|
new_ns->ns.ops = &mntns_operations;
|
2010-03-08 03:49:36 +01:00
|
|
|
new_ns->seq = atomic64_add_return(1, &mnt_ns_seq);
|
2009-06-22 21:09:13 +02:00
|
|
|
atomic_set(&new_ns->count, 1);
|
|
|
|
new_ns->root = NULL;
|
|
|
|
INIT_LIST_HEAD(&new_ns->list);
|
|
|
|
init_waitqueue_head(&new_ns->poll);
|
|
|
|
new_ns->event = 0;
|
2012-07-27 06:08:32 +02:00
|
|
|
new_ns->user_ns = get_user_ns(user_ns);
|
2016-08-08 21:37:37 +02:00
|
|
|
new_ns->ucounts = ucounts;
|
2016-09-28 07:27:17 +02:00
|
|
|
new_ns->mounts = 0;
|
|
|
|
new_ns->pending_mounts = 0;
|
2009-06-22 21:09:13 +02:00
|
|
|
return new_ns;
|
|
|
|
}
|
|
|
|
|
2016-06-20 20:42:34 +02:00
|
|
|
__latent_entropy
|
2013-09-29 02:47:57 +02:00
|
|
|
struct mnt_namespace *copy_mnt_ns(unsigned long flags, struct mnt_namespace *ns,
|
|
|
|
struct user_namespace *user_ns, struct fs_struct *new_fs)
|
2005-04-17 00:20:36 +02:00
|
|
|
{
|
2006-12-08 11:37:56 +01:00
|
|
|
struct mnt_namespace *new_ns;
|
2008-05-11 02:44:54 +02:00
|
|
|
struct vfsmount *rootmnt = NULL, *pwdmnt = NULL;
|
2011-11-25 00:57:30 +01:00
|
|
|
struct mount *p, *q;
|
2013-09-29 02:47:57 +02:00
|
|
|
struct mount *old;
|
2011-11-25 02:55:08 +01:00
|
|
|
struct mount *new;
|
2012-07-31 22:13:04 +02:00
|
|
|
int copy_flags;
|
2005-04-17 00:20:36 +02:00
|
|
|
|
2013-09-29 02:47:57 +02:00
|
|
|
BUG_ON(!ns);
|
|
|
|
|
|
|
|
if (likely(!(flags & CLONE_NEWNS))) {
|
|
|
|
get_mnt_ns(ns);
|
|
|
|
return ns;
|
|
|
|
}
|
|
|
|
|
|
|
|
old = ns->root;
|
|
|
|
|
2012-07-27 06:08:32 +02:00
|
|
|
new_ns = alloc_mnt_ns(user_ns);
|
2009-06-22 21:09:13 +02:00
|
|
|
if (IS_ERR(new_ns))
|
|
|
|
return new_ns;
|
2005-04-17 00:20:36 +02:00
|
|
|
|
2013-03-16 20:12:40 +01:00
|
|
|
namespace_lock();
|
2005-04-17 00:20:36 +02:00
|
|
|
/* First pass: copy the tree topology */
|
2013-03-30 09:35:18 +01:00
|
|
|
copy_flags = CL_COPY_UNBINDABLE | CL_EXPIRE;
|
2013-09-29 02:47:57 +02:00
|
|
|
if (user_ns != ns->user_ns)
|
2013-03-22 12:08:05 +01:00
|
|
|
copy_flags |= CL_SHARED_TO_SLAVE | CL_UNPRIVILEGED;
|
2012-07-31 22:13:04 +02:00
|
|
|
new = copy_tree(old, old->mnt.mnt_root, copy_flags);
|
2012-06-25 13:55:18 +02:00
|
|
|
if (IS_ERR(new)) {
|
2013-03-16 19:49:45 +01:00
|
|
|
namespace_unlock();
|
2012-07-27 06:08:32 +02:00
|
|
|
free_mnt_ns(new_ns);
|
2012-06-25 13:55:18 +02:00
|
|
|
return ERR_CAST(new);
|
2005-04-17 00:20:36 +02:00
|
|
|
}
|
2011-12-06 19:32:36 +01:00
|
|
|
new_ns->root = new;
|
2011-11-25 08:19:55 +01:00
|
|
|
list_add_tail(&new_ns->list, &new->mnt_list);
|
2005-04-17 00:20:36 +02:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Second pass: switch the tsk->fs->* elements and mark new vfsmounts
|
|
|
|
* as belonging to new namespace. We have already acquired a private
|
|
|
|
* fs_struct, so tsk->fs->lock is not needed.
|
|
|
|
*/
|
2011-11-25 09:06:56 +01:00
|
|
|
p = old;
|
2011-11-25 02:55:08 +01:00
|
|
|
q = new;
|
2005-04-17 00:20:36 +02:00
|
|
|
while (p) {
|
2011-11-25 06:46:35 +01:00
|
|
|
q->mnt_ns = new_ns;
|
2016-09-28 07:27:17 +02:00
|
|
|
new_ns->mounts++;
|
2013-09-29 02:47:57 +02:00
|
|
|
if (new_fs) {
|
|
|
|
if (&p->mnt == new_fs->root.mnt) {
|
|
|
|
new_fs->root.mnt = mntget(&q->mnt);
|
2011-11-25 00:57:30 +01:00
|
|
|
rootmnt = &p->mnt;
|
2005-04-17 00:20:36 +02:00
|
|
|
}
|
2013-09-29 02:47:57 +02:00
|
|
|
if (&p->mnt == new_fs->pwd.mnt) {
|
|
|
|
new_fs->pwd.mnt = mntget(&q->mnt);
|
2011-11-25 00:57:30 +01:00
|
|
|
pwdmnt = &p->mnt;
|
2005-04-17 00:20:36 +02:00
|
|
|
}
|
|
|
|
}
|
2011-11-25 09:06:56 +01:00
|
|
|
p = next_mnt(p, old);
|
|
|
|
q = next_mnt(q, new);
|
2013-03-30 09:35:18 +01:00
|
|
|
if (!q)
|
|
|
|
break;
|
|
|
|
while (p->mnt.mnt_root != q->mnt.mnt_root)
|
|
|
|
p = next_mnt(p, old);
|
2005-04-17 00:20:36 +02:00
|
|
|
}
|
2013-03-16 19:49:45 +01:00
|
|
|
namespace_unlock();
|
2005-04-17 00:20:36 +02:00
|
|
|
|
|
|
|
if (rootmnt)
|
2011-01-15 04:30:21 +01:00
|
|
|
mntput(rootmnt);
|
2005-04-17 00:20:36 +02:00
|
|
|
if (pwdmnt)
|
2011-01-15 04:30:21 +01:00
|
|
|
mntput(pwdmnt);
|
2005-04-17 00:20:36 +02:00
|
|
|
|
2006-02-07 21:59:00 +01:00
|
|
|
return new_ns;
|
2005-04-17 00:20:36 +02:00
|
|
|
}
|
|
|
|
|
2009-06-22 21:09:13 +02:00
|
|
|
/**
|
|
|
|
* create_mnt_ns - creates a private namespace and adds a root filesystem
|
|
|
|
* @mnt: pointer to the new root filesystem mountpoint
|
|
|
|
*/
|
2011-11-25 08:19:55 +01:00
|
|
|
static struct mnt_namespace *create_mnt_ns(struct vfsmount *m)
|
2009-06-22 21:09:13 +02:00
|
|
|
{
|
2012-07-27 06:08:32 +02:00
|
|
|
struct mnt_namespace *new_ns = alloc_mnt_ns(&init_user_ns);
|
2009-06-22 21:09:13 +02:00
|
|
|
if (!IS_ERR(new_ns)) {
|
2011-11-25 08:19:55 +01:00
|
|
|
struct mount *mnt = real_mount(m);
|
|
|
|
mnt->mnt_ns = new_ns;
|
2011-12-06 19:32:36 +01:00
|
|
|
new_ns->root = mnt;
|
2016-09-28 07:27:17 +02:00
|
|
|
new_ns->mounts++;
|
2013-05-04 21:18:53 +02:00
|
|
|
list_add(&mnt->mnt_list, &new_ns->list);
|
2011-11-16 22:12:14 +01:00
|
|
|
} else {
|
2011-11-25 08:19:55 +01:00
|
|
|
mntput(m);
|
2009-06-22 21:09:13 +02:00
|
|
|
}
|
|
|
|
return new_ns;
|
|
|
|
}
|
|
|
|
|
2011-11-17 03:43:59 +01:00
|
|
|
struct dentry *mount_subtree(struct vfsmount *mnt, const char *name)
|
|
|
|
{
|
|
|
|
struct mnt_namespace *ns;
|
2011-11-22 18:31:21 +01:00
|
|
|
struct super_block *s;
|
2011-11-17 03:43:59 +01:00
|
|
|
struct path path;
|
|
|
|
int err;
|
|
|
|
|
|
|
|
ns = create_mnt_ns(mnt);
|
|
|
|
if (IS_ERR(ns))
|
|
|
|
return ERR_CAST(ns);
|
|
|
|
|
|
|
|
err = vfs_path_lookup(mnt->mnt_root, mnt,
|
|
|
|
name, LOOKUP_FOLLOW|LOOKUP_AUTOMOUNT, &path);
|
|
|
|
|
|
|
|
put_mnt_ns(ns);
|
|
|
|
|
|
|
|
if (err)
|
|
|
|
return ERR_PTR(err);
|
|
|
|
|
|
|
|
/* trade a vfsmount reference for active sb one */
|
2011-11-22 18:31:21 +01:00
|
|
|
s = path.mnt->mnt_sb;
|
|
|
|
atomic_inc(&s->s_active);
|
2011-11-17 03:43:59 +01:00
|
|
|
mntput(path.mnt);
|
|
|
|
/* lock the sucker */
|
2011-11-22 18:31:21 +01:00
|
|
|
down_write(&s->s_umount);
|
2011-11-17 03:43:59 +01:00
|
|
|
/* ... and return the root of (sub)tree on it */
|
|
|
|
return path.dentry;
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL(mount_subtree);
|
|
|
|
|
2009-01-14 14:14:12 +01:00
|
|
|
SYSCALL_DEFINE5(mount, char __user *, dev_name, char __user *, dir_name,
|
|
|
|
char __user *, type, unsigned long, flags, void __user *, data)
|
2005-04-17 00:20:36 +02:00
|
|
|
{
|
fs: fix overflow in sys_mount() for in-kernel calls
sys_mount() reads/copies a whole page for its "type" parameter. When
do_mount_root() passes a kernel address that points to an object which is
smaller than a whole page, copy_mount_options() will happily go past this
memory object, possibly dereferencing "wild" pointers that could be in any
state (hence the kmemcheck warning, which shows that parts of the next
page are not even allocated).
(The likelihood of something going wrong here is pretty low -- first of
all this only applies to kernel calls to sys_mount(), which are mostly
found in the boot code. Secondly, I guess if the page was not mapped,
exact_copy_from_user() _would_ in fact handle it correctly because of its
access_ok(), etc. checks.)
But it is much nicer to avoid the dubious reads altogether, by stopping as
soon as we find a NUL byte. Is there a good reason why we can't do
something like this, using the already existing strndup_from_user()?
[akpm@linux-foundation.org: make copy_mount_string() static]
[AV: fix compat mount breakage, which involves undoing akpm's change above]
Reported-by: Ingo Molnar <mingo@elte.hu>
Signed-off-by: Vegard Nossum <vegard.nossum@gmail.com>
Cc: Al Viro <viro@zeniv.linux.org.uk>
Cc: Pekka Enberg <penberg@cs.helsinki.fi>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: al <al@dizzy.pdmi.ras.ru>
2009-09-18 22:05:45 +02:00
|
|
|
int ret;
|
|
|
|
char *kernel_type;
|
|
|
|
char *kernel_dev;
|
2015-12-15 00:44:44 +01:00
|
|
|
void *options;
|
2005-04-17 00:20:36 +02:00
|
|
|
|
2014-08-28 19:26:03 +02:00
|
|
|
kernel_type = copy_mount_string(type);
|
|
|
|
ret = PTR_ERR(kernel_type);
|
|
|
|
if (IS_ERR(kernel_type))
|
fs: fix overflow in sys_mount() for in-kernel calls
sys_mount() reads/copies a whole page for its "type" parameter. When
do_mount_root() passes a kernel address that points to an object which is
smaller than a whole page, copy_mount_options() will happily go past this
memory object, possibly dereferencing "wild" pointers that could be in any
state (hence the kmemcheck warning, which shows that parts of the next
page are not even allocated).
(The likelihood of something going wrong here is pretty low -- first of
all this only applies to kernel calls to sys_mount(), which are mostly
found in the boot code. Secondly, I guess if the page was not mapped,
exact_copy_from_user() _would_ in fact handle it correctly because of its
access_ok(), etc. checks.)
But it is much nicer to avoid the dubious reads altogether, by stopping as
soon as we find a NUL byte. Is there a good reason why we can't do
something like this, using the already existing strndup_from_user()?
[akpm@linux-foundation.org: make copy_mount_string() static]
[AV: fix compat mount breakage, which involves undoing akpm's change above]
Reported-by: Ingo Molnar <mingo@elte.hu>
Signed-off-by: Vegard Nossum <vegard.nossum@gmail.com>
Cc: Al Viro <viro@zeniv.linux.org.uk>
Cc: Pekka Enberg <penberg@cs.helsinki.fi>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: al <al@dizzy.pdmi.ras.ru>
2009-09-18 22:05:45 +02:00
|
|
|
goto out_type;
|
2005-04-17 00:20:36 +02:00
|
|
|
|
2014-08-28 19:26:03 +02:00
|
|
|
kernel_dev = copy_mount_string(dev_name);
|
|
|
|
ret = PTR_ERR(kernel_dev);
|
|
|
|
if (IS_ERR(kernel_dev))
|
fs: fix overflow in sys_mount() for in-kernel calls
sys_mount() reads/copies a whole page for its "type" parameter. When
do_mount_root() passes a kernel address that points to an object which is
smaller than a whole page, copy_mount_options() will happily go past this
memory object, possibly dereferencing "wild" pointers that could be in any
state (hence the kmemcheck warning, which shows that parts of the next
page are not even allocated).
(The likelihood of something going wrong here is pretty low -- first of
all this only applies to kernel calls to sys_mount(), which are mostly
found in the boot code. Secondly, I guess if the page was not mapped,
exact_copy_from_user() _would_ in fact handle it correctly because of its
access_ok(), etc. checks.)
But it is much nicer to avoid the dubious reads altogether, by stopping as
soon as we find a NUL byte. Is there a good reason why we can't do
something like this, using the already existing strndup_from_user()?
[akpm@linux-foundation.org: make copy_mount_string() static]
[AV: fix compat mount breakage, which involves undoing akpm's change above]
Reported-by: Ingo Molnar <mingo@elte.hu>
Signed-off-by: Vegard Nossum <vegard.nossum@gmail.com>
Cc: Al Viro <viro@zeniv.linux.org.uk>
Cc: Pekka Enberg <penberg@cs.helsinki.fi>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: al <al@dizzy.pdmi.ras.ru>
2009-09-18 22:05:45 +02:00
|
|
|
goto out_dev;
|
2005-04-17 00:20:36 +02:00
|
|
|
|
2015-12-15 00:44:44 +01:00
|
|
|
options = copy_mount_options(data);
|
|
|
|
ret = PTR_ERR(options);
|
|
|
|
if (IS_ERR(options))
|
fs: fix overflow in sys_mount() for in-kernel calls
sys_mount() reads/copies a whole page for its "type" parameter. When
do_mount_root() passes a kernel address that points to an object which is
smaller than a whole page, copy_mount_options() will happily go past this
memory object, possibly dereferencing "wild" pointers that could be in any
state (hence the kmemcheck warning, which shows that parts of the next
page are not even allocated).
(The likelihood of something going wrong here is pretty low -- first of
all this only applies to kernel calls to sys_mount(), which are mostly
found in the boot code. Secondly, I guess if the page was not mapped,
exact_copy_from_user() _would_ in fact handle it correctly because of its
access_ok(), etc. checks.)
But it is much nicer to avoid the dubious reads altogether, by stopping as
soon as we find a NUL byte. Is there a good reason why we can't do
something like this, using the already existing strndup_from_user()?
[akpm@linux-foundation.org: make copy_mount_string() static]
[AV: fix compat mount breakage, which involves undoing akpm's change above]
Reported-by: Ingo Molnar <mingo@elte.hu>
Signed-off-by: Vegard Nossum <vegard.nossum@gmail.com>
Cc: Al Viro <viro@zeniv.linux.org.uk>
Cc: Pekka Enberg <penberg@cs.helsinki.fi>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: al <al@dizzy.pdmi.ras.ru>
2009-09-18 22:05:45 +02:00
|
|
|
goto out_data;
|
2005-04-17 00:20:36 +02:00
|
|
|
|
2015-12-15 00:44:44 +01:00
|
|
|
ret = do_mount(kernel_dev, dir_name, kernel_type, flags, options);
|
2005-04-17 00:20:36 +02:00
|
|
|
|
2015-12-15 00:44:44 +01:00
|
|
|
kfree(options);
|
fs: fix overflow in sys_mount() for in-kernel calls
sys_mount() reads/copies a whole page for its "type" parameter. When
do_mount_root() passes a kernel address that points to an object which is
smaller than a whole page, copy_mount_options() will happily go past this
memory object, possibly dereferencing "wild" pointers that could be in any
state (hence the kmemcheck warning, which shows that parts of the next
page are not even allocated).
(The likelihood of something going wrong here is pretty low -- first of
all this only applies to kernel calls to sys_mount(), which are mostly
found in the boot code. Secondly, I guess if the page was not mapped,
exact_copy_from_user() _would_ in fact handle it correctly because of its
access_ok(), etc. checks.)
But it is much nicer to avoid the dubious reads altogether, by stopping as
soon as we find a NUL byte. Is there a good reason why we can't do
something like this, using the already existing strndup_from_user()?
[akpm@linux-foundation.org: make copy_mount_string() static]
[AV: fix compat mount breakage, which involves undoing akpm's change above]
Reported-by: Ingo Molnar <mingo@elte.hu>
Signed-off-by: Vegard Nossum <vegard.nossum@gmail.com>
Cc: Al Viro <viro@zeniv.linux.org.uk>
Cc: Pekka Enberg <penberg@cs.helsinki.fi>
Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
Signed-off-by: al <al@dizzy.pdmi.ras.ru>
2009-09-18 22:05:45 +02:00
|
|
|
out_data:
|
|
|
|
kfree(kernel_dev);
|
|
|
|
out_dev:
|
|
|
|
kfree(kernel_type);
|
|
|
|
out_type:
|
|
|
|
return ret;
|
2005-04-17 00:20:36 +02:00
|
|
|
}
|
|
|
|
|
2011-11-24 01:34:49 +01:00
|
|
|
/*
|
|
|
|
* Return true if path is reachable from root
|
|
|
|
*
|
2013-09-30 04:06:07 +02:00
|
|
|
* namespace_sem or mount_lock is held
|
2011-11-24 01:34:49 +01:00
|
|
|
*/
|
2011-11-25 04:00:28 +01:00
|
|
|
bool is_path_reachable(struct mount *mnt, struct dentry *dentry,
|
2011-11-24 01:34:49 +01:00
|
|
|
const struct path *root)
|
|
|
|
{
|
2011-11-25 04:00:28 +01:00
|
|
|
while (&mnt->mnt != root->mnt && mnt_has_parent(mnt)) {
|
2011-11-25 04:25:07 +01:00
|
|
|
dentry = mnt->mnt_mountpoint;
|
2011-11-25 04:19:58 +01:00
|
|
|
mnt = mnt->mnt_parent;
|
2011-11-24 01:34:49 +01:00
|
|
|
}
|
2011-11-25 04:00:28 +01:00
|
|
|
return &mnt->mnt == root->mnt && is_subdir(dentry, root->dentry);
|
2011-11-24 01:34:49 +01:00
|
|
|
}
|
|
|
|
|
2016-11-14 22:14:35 +01:00
|
|
|
bool path_is_under(const struct path *path1, const struct path *path2)
|
2011-11-24 01:34:49 +01:00
|
|
|
{
|
2015-11-17 07:40:10 +01:00
|
|
|
bool res;
|
2013-09-30 04:06:07 +02:00
|
|
|
read_seqlock_excl(&mount_lock);
|
2011-11-25 04:00:28 +01:00
|
|
|
res = is_path_reachable(real_mount(path1->mnt), path1->dentry, path2);
|
2013-09-30 04:06:07 +02:00
|
|
|
read_sequnlock_excl(&mount_lock);
|
2011-11-24 01:34:49 +01:00
|
|
|
return res;
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL(path_is_under);
|
|
|
|
|
2005-04-17 00:20:36 +02:00
|
|
|
/*
|
|
|
|
* pivot_root Semantics:
|
|
|
|
* Moves the root file system of the current process to the directory put_old,
|
|
|
|
* makes new_root as the new root file system of the current process, and sets
|
|
|
|
* root/cwd of all processes which had them on the current root to new_root.
|
|
|
|
*
|
|
|
|
* Restrictions:
|
|
|
|
* The new_root and put_old must be directories, and must not be on the
|
|
|
|
* same file system as the current process root. The put_old must be
|
|
|
|
* underneath new_root, i.e. adding a non-zero number of /.. to the string
|
|
|
|
* pointed to by put_old must yield the same directory as new_root. No other
|
|
|
|
* file system may be mounted on put_old. After all, new_root is a mountpoint.
|
|
|
|
*
|
2006-01-08 10:03:18 +01:00
|
|
|
* Also, the current root cannot be on the 'rootfs' (initial ramfs) filesystem.
|
|
|
|
* See Documentation/filesystems/ramfs-rootfs-initramfs.txt for alternatives
|
|
|
|
* in this situation.
|
|
|
|
*
|
2005-04-17 00:20:36 +02:00
|
|
|
* Notes:
|
|
|
|
* - we don't move root/cwd if they are not at the root (reason: if something
|
|
|
|
* cared enough to change them, it's probably wrong to force them elsewhere)
|
|
|
|
* - it's okay to pick a root that isn't the root of a file system, e.g.
|
|
|
|
* /nfs/my_root where /nfs is the mount point. It must be a mountpoint,
|
|
|
|
* though, so you may need to say mount --bind /nfs/my_root /nfs/my_root
|
|
|
|
* first.
|
|
|
|
*/
|
2009-01-14 14:14:16 +01:00
|
|
|
SYSCALL_DEFINE2(pivot_root, const char __user *, new_root,
|
|
|
|
const char __user *, put_old)
|
2005-04-17 00:20:36 +02:00
|
|
|
{
|
2008-07-22 15:59:21 +02:00
|
|
|
struct path new, old, parent_path, root_parent, root;
|
2013-03-15 15:53:28 +01:00
|
|
|
struct mount *new_mnt, *root_mnt, *old_mnt;
|
|
|
|
struct mountpoint *old_mp, *root_mp;
|
2005-04-17 00:20:36 +02:00
|
|
|
int error;
|
|
|
|
|
2013-02-23 04:45:42 +01:00
|
|
|
if (!may_mount())
|
2005-04-17 00:20:36 +02:00
|
|
|
return -EPERM;
|
|
|
|
|
2008-07-22 15:59:21 +02:00
|
|
|
error = user_path_dir(new_root, &new);
|
2005-04-17 00:20:36 +02:00
|
|
|
if (error)
|
|
|
|
goto out0;
|
|
|
|
|
2008-07-22 15:59:21 +02:00
|
|
|
error = user_path_dir(put_old, &old);
|
2005-04-17 00:20:36 +02:00
|
|
|
if (error)
|
|
|
|
goto out1;
|
|
|
|
|
2008-07-22 15:59:21 +02:00
|
|
|
error = security_sb_pivotroot(&old, &new);
|
2011-03-18 13:55:38 +01:00
|
|
|
if (error)
|
|
|
|
goto out2;
|
2005-04-17 00:20:36 +02:00
|
|
|
|
2010-08-10 11:41:36 +02:00
|
|
|
get_fs_root(current->fs, &root);
|
2013-03-15 15:53:28 +01:00
|
|
|
old_mp = lock_mount(&old);
|
|
|
|
error = PTR_ERR(old_mp);
|
|
|
|
if (IS_ERR(old_mp))
|
2011-03-18 13:55:38 +01:00
|
|
|
goto out3;
|
|
|
|
|
2005-04-17 00:20:36 +02:00
|
|
|
error = -EINVAL;
|
2011-11-25 01:41:16 +01:00
|
|
|
new_mnt = real_mount(new.mnt);
|
|
|
|
root_mnt = real_mount(root.mnt);
|
2013-03-15 15:53:28 +01:00
|
|
|
old_mnt = real_mount(old.mnt);
|
|
|
|
if (IS_MNT_SHARED(old_mnt) ||
|
2011-11-25 07:05:37 +01:00
|
|
|
IS_MNT_SHARED(new_mnt->mnt_parent) ||
|
|
|
|
IS_MNT_SHARED(root_mnt->mnt_parent))
|
2011-03-18 13:55:38 +01:00
|
|
|
goto out4;
|
2011-11-25 06:46:35 +01:00
|
|
|
if (!check_mnt(root_mnt) || !check_mnt(new_mnt))
|
2011-03-18 13:55:38 +01:00
|
|
|
goto out4;
|
2013-03-30 05:04:39 +01:00
|
|
|
if (new_mnt->mnt.mnt_flags & MNT_LOCKED)
|
|
|
|
goto out4;
|
2005-04-17 00:20:36 +02:00
|
|
|
error = -ENOENT;
|
2009-05-04 01:32:03 +02:00
|
|
|
if (d_unlinked(new.dentry))
|
2011-03-18 13:55:38 +01:00
|
|
|
goto out4;
|
2005-04-17 00:20:36 +02:00
|
|
|
error = -EBUSY;
|
2013-03-15 15:53:28 +01:00
|
|
|
if (new_mnt == root_mnt || old_mnt == root_mnt)
|
2011-03-18 13:55:38 +01:00
|
|
|
goto out4; /* loop, on the same file system */
|
2005-04-17 00:20:36 +02:00
|
|
|
error = -EINVAL;
|
2008-03-22 23:00:39 +01:00
|
|
|
if (root.mnt->mnt_root != root.dentry)
|
2011-03-18 13:55:38 +01:00
|
|
|
goto out4; /* not a mountpoint */
|
2011-11-25 03:47:05 +01:00
|
|
|
if (!mnt_has_parent(root_mnt))
|
2011-03-18 13:55:38 +01:00
|
|
|
goto out4; /* not attached */
|
2013-03-15 15:53:28 +01:00
|
|
|
root_mp = root_mnt->mnt_mp;
|
2008-07-22 15:59:21 +02:00
|
|
|
if (new.mnt->mnt_root != new.dentry)
|
2011-03-18 13:55:38 +01:00
|
|
|
goto out4; /* not a mountpoint */
|
2011-11-25 03:47:05 +01:00
|
|
|
if (!mnt_has_parent(new_mnt))
|
2011-03-18 13:55:38 +01:00
|
|
|
goto out4; /* not attached */
|
2008-02-15 04:34:32 +01:00
|
|
|
/* make sure we can reach put_old from new_root */
|
2013-03-15 15:53:28 +01:00
|
|
|
if (!is_path_reachable(old_mnt, old.dentry, &new))
|
2011-03-18 13:55:38 +01:00
|
|
|
goto out4;
|
2014-10-08 19:42:27 +02:00
|
|
|
/* make certain new is below the root */
|
|
|
|
if (!is_path_reachable(new_mnt, new.dentry, &root))
|
|
|
|
goto out4;
|
2013-03-15 15:53:28 +01:00
|
|
|
root_mp->m_count++; /* pin it so it won't go away */
|
2013-09-29 17:24:49 +02:00
|
|
|
lock_mount_hash();
|
2011-11-25 01:41:16 +01:00
|
|
|
detach_mnt(new_mnt, &parent_path);
|
|
|
|
detach_mnt(root_mnt, &root_parent);
|
2013-03-30 05:04:39 +01:00
|
|
|
if (root_mnt->mnt.mnt_flags & MNT_LOCKED) {
|
|
|
|
new_mnt->mnt.mnt_flags |= MNT_LOCKED;
|
|
|
|
root_mnt->mnt.mnt_flags &= ~MNT_LOCKED;
|
|
|
|
}
|
2008-02-15 04:34:32 +01:00
|
|
|
/* mount old root on put_old */
|
2013-03-15 15:53:28 +01:00
|
|
|
attach_mnt(root_mnt, old_mnt, old_mp);
|
2008-02-15 04:34:32 +01:00
|
|
|
/* mount new_root on / */
|
2013-03-15 15:53:28 +01:00
|
|
|
attach_mnt(new_mnt, real_mount(root_parent.mnt), root_mp);
|
2006-12-08 11:37:56 +01:00
|
|
|
touch_mnt_namespace(current->nsproxy->mnt_ns);
|
2014-10-08 19:42:57 +02:00
|
|
|
/* A moved mount should not expire automatically */
|
|
|
|
list_del_init(&new_mnt->mnt_expire);
|
2017-01-03 02:18:43 +01:00
|
|
|
put_mountpoint(root_mp);
|
2013-09-29 17:24:49 +02:00
|
|
|
unlock_mount_hash();
|
2008-07-22 15:59:21 +02:00
|
|
|
chroot_fs_refs(&root, &new);
|
2005-04-17 00:20:36 +02:00
|
|
|
error = 0;
|
2011-03-18 13:55:38 +01:00
|
|
|
out4:
|
2013-03-15 15:53:28 +01:00
|
|
|
unlock_mount(old_mp);
|
2011-03-18 13:55:38 +01:00
|
|
|
if (!error) {
|
|
|
|
path_put(&root_parent);
|
|
|
|
path_put(&parent_path);
|
|
|
|
}
|
|
|
|
out3:
|
2008-03-22 23:00:39 +01:00
|
|
|
path_put(&root);
|
2011-03-18 13:55:38 +01:00
|
|
|
out2:
|
2008-07-22 15:59:21 +02:00
|
|
|
path_put(&old);
|
2005-04-17 00:20:36 +02:00
|
|
|
out1:
|
2008-07-22 15:59:21 +02:00
|
|
|
path_put(&new);
|
2005-04-17 00:20:36 +02:00
|
|
|
out0:
|
|
|
|
return error;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void __init init_mount_tree(void)
|
|
|
|
{
|
|
|
|
struct vfsmount *mnt;
|
2006-12-08 11:37:56 +01:00
|
|
|
struct mnt_namespace *ns;
|
2008-02-15 04:34:39 +01:00
|
|
|
struct path root;
|
2012-07-27 06:42:03 +02:00
|
|
|
struct file_system_type *type;
|
2005-04-17 00:20:36 +02:00
|
|
|
|
2012-07-27 06:42:03 +02:00
|
|
|
type = get_fs_type("rootfs");
|
|
|
|
if (!type)
|
|
|
|
panic("Can't find rootfs type");
|
|
|
|
mnt = vfs_kern_mount(type, 0, "rootfs", NULL);
|
|
|
|
put_filesystem(type);
|
2005-04-17 00:20:36 +02:00
|
|
|
if (IS_ERR(mnt))
|
|
|
|
panic("Can't create rootfs");
|
fs: scale mntget/mntput
The problem that this patch aims to fix is vfsmount refcounting scalability.
We need to take a reference on the vfsmount for every successful path lookup,
which often go to the same mount point.
The fundamental difficulty is that a "simple" reference count can never be made
scalable, because any time a reference is dropped, we must check whether that
was the last reference. To do that requires communication with all other CPUs
that may have taken a reference count.
We can make refcounts more scalable in a couple of ways, involving keeping
distributed counters, and checking for the global-zero condition less
frequently.
- check the global sum once every interval (this will delay zero detection
for some interval, so it's probably a showstopper for vfsmounts).
- keep a local count and only taking the global sum when local reaches 0 (this
is difficult for vfsmounts, because we can't hold preempt off for the life of
a reference, so a counter would need to be per-thread or tied strongly to a
particular CPU which requires more locking).
- keep a local difference of increments and decrements, which allows us to sum
the total difference and hence find the refcount when summing all CPUs. Then,
keep a single integer "long" refcount for slow and long lasting references,
and only take the global sum of local counters when the long refcount is 0.
This last scheme is what I implemented here. Attached mounts and process root
and working directory references are "long" references, and everything else is
a short reference.
This allows scalable vfsmount references during path walking over mounted
subtrees and unattached (lazy umounted) mounts with processes still running
in them.
This results in one fewer atomic op in the fastpath: mntget is now just a
per-CPU inc, rather than an atomic inc; and mntput just requires a spinlock
and non-atomic decrement in the common case. However code is otherwise bigger
and heavier, so single threaded performance is basically a wash.
Signed-off-by: Nick Piggin <npiggin@kernel.dk>
2011-01-07 07:50:11 +01:00
|
|
|
|
2009-06-23 23:29:49 +02:00
|
|
|
ns = create_mnt_ns(mnt);
|
|
|
|
if (IS_ERR(ns))
|
2005-04-17 00:20:36 +02:00
|
|
|
panic("Can't allocate initial namespace");
|
2006-12-08 11:37:56 +01:00
|
|
|
|
|
|
|
init_task.nsproxy->mnt_ns = ns;
|
|
|
|
get_mnt_ns(ns);
|
|
|
|
|
2011-12-06 19:32:36 +01:00
|
|
|
root.mnt = mnt;
|
|
|
|
root.dentry = mnt->mnt_root;
|
umount: Do not allow unmounting rootfs.
Andrew Vagin <avagin@parallels.com> writes:
> #define _GNU_SOURCE
> #include <sys/types.h>
> #include <sys/stat.h>
> #include <fcntl.h>
> #include <sched.h>
> #include <unistd.h>
> #include <sys/mount.h>
>
> int main(int argc, char **argv)
> {
> int fd;
>
> fd = open("/proc/self/ns/mnt", O_RDONLY);
> if (fd < 0)
> return 1;
> while (1) {
> if (umount2("/", MNT_DETACH) ||
> setns(fd, CLONE_NEWNS))
> break;
> }
>
> return 0;
> }
>
> root@ubuntu:/home/avagin# gcc -Wall nsenter.c -o nsenter
> root@ubuntu:/home/avagin# strace ./nsenter
> execve("./nsenter", ["./nsenter"], [/* 22 vars */]) = 0
> ...
> open("/proc/self/ns/mnt", O_RDONLY) = 3
> umount("/", MNT_DETACH) = 0
> setns(3, 131072) = 0
> umount("/", MNT_DETACH
>
causes:
> [ 260.548301] ------------[ cut here ]------------
> [ 260.550941] kernel BUG at /build/buildd/linux-3.13.0/fs/pnode.c:372!
> [ 260.552068] invalid opcode: 0000 [#1] SMP
> [ 260.552068] Modules linked in: xt_CHECKSUM iptable_mangle xt_tcpudp xt_addrtype xt_conntrack ipt_MASQUERADE iptable_nat nf_conntrack_ipv4 nf_defrag_ipv4 nf_nat_ipv4 nf_nat nf_conntrack bridge stp llc dm_thin_pool dm_persistent_data dm_bufio dm_bio_prison iptable_filter ip_tables x_tables crct10dif_pclmul crc32_pclmul ghash_clmulni_intel binfmt_misc nfsd auth_rpcgss nfs_acl aesni_intel nfs lockd aes_x86_64 sunrpc fscache lrw gf128mul glue_helper ablk_helper cryptd serio_raw ppdev parport_pc lp parport btrfs xor raid6_pq libcrc32c psmouse floppy
> [ 260.552068] CPU: 0 PID: 1723 Comm: nsenter Not tainted 3.13.0-30-generic #55-Ubuntu
> [ 260.552068] Hardware name: Bochs Bochs, BIOS Bochs 01/01/2011
> [ 260.552068] task: ffff8800376097f0 ti: ffff880074824000 task.ti: ffff880074824000
> [ 260.552068] RIP: 0010:[<ffffffff811e9483>] [<ffffffff811e9483>] propagate_umount+0x123/0x130
> [ 260.552068] RSP: 0018:ffff880074825e98 EFLAGS: 00010246
> [ 260.552068] RAX: ffff88007c741140 RBX: 0000000000000002 RCX: ffff88007c741190
> [ 260.552068] RDX: ffff88007c741190 RSI: ffff880074825ec0 RDI: ffff880074825ec0
> [ 260.552068] RBP: ffff880074825eb0 R08: 00000000000172e0 R09: ffff88007fc172e0
> [ 260.552068] R10: ffffffff811cc642 R11: ffffea0001d59000 R12: ffff88007c741140
> [ 260.552068] R13: ffff88007c741140 R14: ffff88007c741140 R15: 0000000000000000
> [ 260.552068] FS: 00007fd5c7e41740(0000) GS:ffff88007fc00000(0000) knlGS:0000000000000000
> [ 260.552068] CS: 0010 DS: 0000 ES: 0000 CR0: 0000000080050033
> [ 260.552068] CR2: 00007fd5c7968050 CR3: 0000000070124000 CR4: 00000000000406f0
> [ 260.552068] Stack:
> [ 260.552068] 0000000000000002 0000000000000002 ffff88007c631000 ffff880074825ed8
> [ 260.552068] ffffffff811dcfac ffff88007c741140 0000000000000002 ffff88007c741160
> [ 260.552068] ffff880074825f38 ffffffff811dd12b ffffffff811cc642 0000000075640000
> [ 260.552068] Call Trace:
> [ 260.552068] [<ffffffff811dcfac>] umount_tree+0x20c/0x260
> [ 260.552068] [<ffffffff811dd12b>] do_umount+0x12b/0x300
> [ 260.552068] [<ffffffff811cc642>] ? final_putname+0x22/0x50
> [ 260.552068] [<ffffffff811cc849>] ? putname+0x29/0x40
> [ 260.552068] [<ffffffff811dd88c>] SyS_umount+0xdc/0x100
> [ 260.552068] [<ffffffff8172aeff>] tracesys+0xe1/0xe6
> [ 260.552068] Code: 89 50 08 48 8b 50 08 48 89 02 49 89 45 08 e9 72 ff ff ff 0f 1f 44 00 00 4c 89 e6 4c 89 e7 e8 f5 f6 ff ff 48 89 c3 e9 39 ff ff ff <0f> 0b 66 2e 0f 1f 84 00 00 00 00 00 90 66 66 66 66 90 55 b8 01
> [ 260.552068] RIP [<ffffffff811e9483>] propagate_umount+0x123/0x130
> [ 260.552068] RSP <ffff880074825e98>
> [ 260.611451] ---[ end trace 11c33d85f1d4c652 ]--
Which in practice is totally uninteresting. Only the global root user can
do it, and it is just a stupid thing to do.
However that is no excuse to allow a silly way to oops the kernel.
We can avoid this silly problem by setting MNT_LOCKED on the rootfs
mount point and thus avoid needing any special cases in the unmount
code.
Signed-off-by: "Eric W. Biederman" <ebiederm@xmission.com>
2014-10-07 21:19:53 +02:00
|
|
|
mnt->mnt_flags |= MNT_LOCKED;
|
2008-02-15 04:34:39 +01:00
|
|
|
|
|
|
|
set_fs_pwd(current->fs, &root);
|
|
|
|
set_fs_root(current->fs, &root);
|
2005-04-17 00:20:36 +02:00
|
|
|
}
|
|
|
|
|
2007-10-17 08:26:30 +02:00
|
|
|
void __init mnt_init(void)
|
2005-04-17 00:20:36 +02:00
|
|
|
{
|
2006-09-29 10:58:57 +02:00
|
|
|
int err;
|
2005-04-17 00:20:36 +02:00
|
|
|
|
2011-11-23 18:14:10 +01:00
|
|
|
mnt_cache = kmem_cache_create("mnt_cache", sizeof(struct mount),
|
2007-07-20 03:11:58 +02:00
|
|
|
0, SLAB_HWCACHE_ALIGN | SLAB_PANIC, NULL);
|
2005-04-17 00:20:36 +02:00
|
|
|
|
2014-02-28 19:46:44 +01:00
|
|
|
mount_hashtable = alloc_large_system_hash("Mount-cache",
|
2014-03-21 02:10:51 +01:00
|
|
|
sizeof(struct hlist_head),
|
2014-02-28 19:46:44 +01:00
|
|
|
mhash_entries, 19,
|
2017-07-07 00:39:11 +02:00
|
|
|
HASH_ZERO,
|
2014-02-28 19:46:44 +01:00
|
|
|
&m_hash_shift, &m_hash_mask, 0, 0);
|
|
|
|
mountpoint_hashtable = alloc_large_system_hash("Mountpoint-cache",
|
|
|
|
sizeof(struct hlist_head),
|
|
|
|
mphash_entries, 19,
|
2017-07-07 00:39:11 +02:00
|
|
|
HASH_ZERO,
|
2014-02-28 19:46:44 +01:00
|
|
|
&mp_hash_shift, &mp_hash_mask, 0, 0);
|
2005-04-17 00:20:36 +02:00
|
|
|
|
2013-03-15 15:53:28 +01:00
|
|
|
if (!mount_hashtable || !mountpoint_hashtable)
|
2005-04-17 00:20:36 +02:00
|
|
|
panic("Failed to allocate mount hash table\n");
|
|
|
|
|
2013-11-28 20:54:43 +01:00
|
|
|
kernfs_init();
|
|
|
|
|
2006-09-29 10:58:57 +02:00
|
|
|
err = sysfs_init();
|
|
|
|
if (err)
|
|
|
|
printk(KERN_WARNING "%s: sysfs_init error: %d\n",
|
2008-04-30 09:55:09 +02:00
|
|
|
__func__, err);
|
2007-10-29 21:17:23 +01:00
|
|
|
fs_kobj = kobject_create_and_add("fs", NULL);
|
|
|
|
if (!fs_kobj)
|
2008-04-30 09:55:09 +02:00
|
|
|
printk(KERN_WARNING "%s: kobj create error\n", __func__);
|
2005-04-17 00:20:36 +02:00
|
|
|
init_rootfs();
|
|
|
|
init_mount_tree();
|
|
|
|
}
|
|
|
|
|
2009-06-22 21:09:13 +02:00
|
|
|
void put_mnt_ns(struct mnt_namespace *ns)
|
2005-04-17 00:20:36 +02:00
|
|
|
{
|
2010-02-05 08:21:06 +01:00
|
|
|
if (!atomic_dec_and_test(&ns->count))
|
2009-06-22 21:09:13 +02:00
|
|
|
return;
|
2013-09-17 03:19:20 +02:00
|
|
|
drop_collected_mounts(&ns->root->mnt);
|
2012-07-27 06:08:32 +02:00
|
|
|
free_mnt_ns(ns);
|
2005-04-17 00:20:36 +02:00
|
|
|
}
|
2011-03-18 03:08:28 +01:00
|
|
|
|
|
|
|
struct vfsmount *kern_mount_data(struct file_system_type *type, void *data)
|
|
|
|
{
|
2011-07-19 18:32:38 +02:00
|
|
|
struct vfsmount *mnt;
|
2017-07-17 09:45:35 +02:00
|
|
|
mnt = vfs_kern_mount(type, SB_KERNMOUNT, type->name, data);
|
2011-07-19 18:32:38 +02:00
|
|
|
if (!IS_ERR(mnt)) {
|
|
|
|
/*
|
|
|
|
* it is a longterm mount, don't release mnt until
|
|
|
|
* we unmount before file sys is unregistered
|
|
|
|
*/
|
2012-06-09 06:59:08 +02:00
|
|
|
real_mount(mnt)->mnt_ns = MNT_NS_INTERNAL;
|
2011-07-19 18:32:38 +02:00
|
|
|
}
|
|
|
|
return mnt;
|
2011-03-18 03:08:28 +01:00
|
|
|
}
|
|
|
|
EXPORT_SYMBOL_GPL(kern_mount_data);
|
2011-07-19 18:32:38 +02:00
|
|
|
|
|
|
|
void kern_unmount(struct vfsmount *mnt)
|
|
|
|
{
|
|
|
|
/* release long term mount so mount point can be released */
|
|
|
|
if (!IS_ERR_OR_NULL(mnt)) {
|
2012-06-09 06:59:08 +02:00
|
|
|
real_mount(mnt)->mnt_ns = NULL;
|
2013-09-30 04:06:07 +02:00
|
|
|
synchronize_rcu(); /* yecchhh... */
|
2011-07-19 18:32:38 +02:00
|
|
|
mntput(mnt);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL(kern_unmount);
|
fix apparmor dereferencing potentially freed dentry, sanitize __d_path() API
__d_path() API is asking for trouble and in case of apparmor d_namespace_path()
getting just that. The root cause is that when __d_path() misses the root
it had been told to look for, it stores the location of the most remote ancestor
in *root. Without grabbing references. Sure, at the moment of call it had
been pinned down by what we have in *path. And if we raced with umount -l, we
could have very well stopped at vfsmount/dentry that got freed as soon as
prepend_path() dropped vfsmount_lock.
It is safe to compare these pointers with pre-existing (and known to be still
alive) vfsmount and dentry, as long as all we are asking is "is it the same
address?". Dereferencing is not safe and apparmor ended up stepping into
that. d_namespace_path() really wants to examine the place where we stopped,
even if it's not connected to our namespace. As the result, it looked
at ->d_sb->s_magic of a dentry that might've been already freed by that point.
All other callers had been careful enough to avoid that, but it's really
a bad interface - it invites that kind of trouble.
The fix is fairly straightforward, even though it's bigger than I'd like:
* prepend_path() root argument becomes const.
* __d_path() is never called with NULL/NULL root. It was a kludge
to start with. Instead, we have an explicit function - d_absolute_root().
Same as __d_path(), except that it doesn't get root passed and stops where
it stops. apparmor and tomoyo are using it.
* __d_path() returns NULL on path outside of root. The main
caller is show_mountinfo() and that's precisely what we pass root for - to
skip those outside chroot jail. Those who don't want that can (and do)
use d_path().
* __d_path() root argument becomes const. Everyone agrees, I hope.
* apparmor does *NOT* try to use __d_path() or any of its variants
when it sees that path->mnt is an internal vfsmount. In that case it's
definitely not mounted anywhere and dentry_path() is exactly what we want
there. Handling of sysctl()-triggered weirdness is moved to that place.
* if apparmor is asked to do pathname relative to chroot jail
and __d_path() tells it we it's not in that jail, the sucker just calls
d_absolute_path() instead. That's the other remaining caller of __d_path(),
BTW.
* seq_path_root() does _NOT_ return -ENAMETOOLONG (it's stupid anyway -
the normal seq_file logics will take care of growing the buffer and redoing
the call of ->show() just fine). However, if it gets path not reachable
from root, it returns SEQ_SKIP. The only caller adjusted (i.e. stopped
ignoring the return value as it used to do).
Reviewed-by: John Johansen <john.johansen@canonical.com>
ACKed-by: John Johansen <john.johansen@canonical.com>
Signed-off-by: Al Viro <viro@zeniv.linux.org.uk>
Cc: stable@vger.kernel.org
2011-12-05 14:43:34 +01:00
|
|
|
|
|
|
|
bool our_mnt(struct vfsmount *mnt)
|
|
|
|
{
|
2011-11-25 06:46:35 +01:00
|
|
|
return check_mnt(real_mount(mnt));
|
fix apparmor dereferencing potentially freed dentry, sanitize __d_path() API
__d_path() API is asking for trouble and in case of apparmor d_namespace_path()
getting just that. The root cause is that when __d_path() misses the root
it had been told to look for, it stores the location of the most remote ancestor
in *root. Without grabbing references. Sure, at the moment of call it had
been pinned down by what we have in *path. And if we raced with umount -l, we
could have very well stopped at vfsmount/dentry that got freed as soon as
prepend_path() dropped vfsmount_lock.
It is safe to compare these pointers with pre-existing (and known to be still
alive) vfsmount and dentry, as long as all we are asking is "is it the same
address?". Dereferencing is not safe and apparmor ended up stepping into
that. d_namespace_path() really wants to examine the place where we stopped,
even if it's not connected to our namespace. As the result, it looked
at ->d_sb->s_magic of a dentry that might've been already freed by that point.
All other callers had been careful enough to avoid that, but it's really
a bad interface - it invites that kind of trouble.
The fix is fairly straightforward, even though it's bigger than I'd like:
* prepend_path() root argument becomes const.
* __d_path() is never called with NULL/NULL root. It was a kludge
to start with. Instead, we have an explicit function - d_absolute_root().
Same as __d_path(), except that it doesn't get root passed and stops where
it stops. apparmor and tomoyo are using it.
* __d_path() returns NULL on path outside of root. The main
caller is show_mountinfo() and that's precisely what we pass root for - to
skip those outside chroot jail. Those who don't want that can (and do)
use d_path().
* __d_path() root argument becomes const. Everyone agrees, I hope.
* apparmor does *NOT* try to use __d_path() or any of its variants
when it sees that path->mnt is an internal vfsmount. In that case it's
definitely not mounted anywhere and dentry_path() is exactly what we want
there. Handling of sysctl()-triggered weirdness is moved to that place.
* if apparmor is asked to do pathname relative to chroot jail
and __d_path() tells it we it's not in that jail, the sucker just calls
d_absolute_path() instead. That's the other remaining caller of __d_path(),
BTW.
* seq_path_root() does _NOT_ return -ENAMETOOLONG (it's stupid anyway -
the normal seq_file logics will take care of growing the buffer and redoing
the call of ->show() just fine). However, if it gets path not reachable
from root, it returns SEQ_SKIP. The only caller adjusted (i.e. stopped
ignoring the return value as it used to do).
Reviewed-by: John Johansen <john.johansen@canonical.com>
ACKed-by: John Johansen <john.johansen@canonical.com>
Signed-off-by: Al Viro <viro@zeniv.linux.org.uk>
Cc: stable@vger.kernel.org
2011-12-05 14:43:34 +01:00
|
|
|
}
|
2010-03-08 03:49:36 +01:00
|
|
|
|
2013-03-15 09:45:51 +01:00
|
|
|
bool current_chrooted(void)
|
|
|
|
{
|
|
|
|
/* Does the current process have a non-standard root */
|
|
|
|
struct path ns_root;
|
|
|
|
struct path fs_root;
|
|
|
|
bool chrooted;
|
|
|
|
|
|
|
|
/* Find the namespace root */
|
|
|
|
ns_root.mnt = ¤t->nsproxy->mnt_ns->root->mnt;
|
|
|
|
ns_root.dentry = ns_root.mnt->mnt_root;
|
|
|
|
path_get(&ns_root);
|
|
|
|
while (d_mountpoint(ns_root.dentry) && follow_down_one(&ns_root))
|
|
|
|
;
|
|
|
|
|
|
|
|
get_fs_root(current->fs, &fs_root);
|
|
|
|
|
|
|
|
chrooted = !path_equal(&fs_root, &ns_root);
|
|
|
|
|
|
|
|
path_put(&fs_root);
|
|
|
|
path_put(&ns_root);
|
|
|
|
|
|
|
|
return chrooted;
|
|
|
|
}
|
|
|
|
|
2016-06-09 23:06:06 +02:00
|
|
|
static bool mnt_already_visible(struct mnt_namespace *ns, struct vfsmount *new,
|
|
|
|
int *new_mnt_flags)
|
2013-03-24 22:28:27 +01:00
|
|
|
{
|
2015-05-09 06:49:47 +02:00
|
|
|
int new_flags = *new_mnt_flags;
|
2013-03-24 22:28:27 +01:00
|
|
|
struct mount *mnt;
|
2013-03-31 04:57:41 +02:00
|
|
|
bool visible = false;
|
2013-03-24 22:28:27 +01:00
|
|
|
|
2013-09-17 03:37:36 +02:00
|
|
|
down_read(&namespace_sem);
|
2013-03-24 22:28:27 +01:00
|
|
|
list_for_each_entry(mnt, &ns->list, mnt_list) {
|
2013-03-31 04:57:41 +02:00
|
|
|
struct mount *child;
|
2015-06-04 16:43:11 +02:00
|
|
|
int mnt_flags;
|
|
|
|
|
2016-06-09 23:06:06 +02:00
|
|
|
if (mnt->mnt.mnt_sb->s_type != new->mnt_sb->s_type)
|
2013-03-31 04:57:41 +02:00
|
|
|
continue;
|
|
|
|
|
2015-05-08 23:36:50 +02:00
|
|
|
/* This mount is not fully visible if it's root directory
|
|
|
|
* is not the root directory of the filesystem.
|
|
|
|
*/
|
|
|
|
if (mnt->mnt.mnt_root != mnt->mnt.mnt_sb->s_root)
|
|
|
|
continue;
|
|
|
|
|
2016-06-15 13:59:49 +02:00
|
|
|
/* A local view of the mount flags */
|
2015-06-04 16:43:11 +02:00
|
|
|
mnt_flags = mnt->mnt.mnt_flags;
|
|
|
|
|
2016-06-10 19:21:40 +02:00
|
|
|
/* Don't miss readonly hidden in the superblock flags */
|
2017-07-17 09:45:34 +02:00
|
|
|
if (sb_rdonly(mnt->mnt.mnt_sb))
|
2016-06-10 19:21:40 +02:00
|
|
|
mnt_flags |= MNT_LOCK_READONLY;
|
|
|
|
|
2015-05-09 06:49:47 +02:00
|
|
|
/* Verify the mount flags are equal to or more permissive
|
|
|
|
* than the proposed new mount.
|
|
|
|
*/
|
2015-06-04 16:43:11 +02:00
|
|
|
if ((mnt_flags & MNT_LOCK_READONLY) &&
|
2015-05-09 06:49:47 +02:00
|
|
|
!(new_flags & MNT_READONLY))
|
|
|
|
continue;
|
2015-06-04 16:43:11 +02:00
|
|
|
if ((mnt_flags & MNT_LOCK_ATIME) &&
|
|
|
|
((mnt_flags & MNT_ATIME_MASK) != (new_flags & MNT_ATIME_MASK)))
|
2015-05-09 06:49:47 +02:00
|
|
|
continue;
|
|
|
|
|
2015-01-07 15:10:09 +01:00
|
|
|
/* This mount is not fully visible if there are any
|
|
|
|
* locked child mounts that cover anything except for
|
|
|
|
* empty directories.
|
2013-03-31 04:57:41 +02:00
|
|
|
*/
|
|
|
|
list_for_each_entry(child, &mnt->mnt_mounts, mnt_child) {
|
|
|
|
struct inode *inode = child->mnt_mountpoint->d_inode;
|
2015-01-07 15:10:09 +01:00
|
|
|
/* Only worry about locked mounts */
|
2016-05-27 21:50:05 +02:00
|
|
|
if (!(child->mnt.mnt_flags & MNT_LOCKED))
|
2015-01-07 15:10:09 +01:00
|
|
|
continue;
|
2015-05-14 03:51:09 +02:00
|
|
|
/* Is the directory permanetly empty? */
|
|
|
|
if (!is_empty_dir_inode(inode))
|
2013-03-31 04:57:41 +02:00
|
|
|
goto next;
|
2013-03-24 22:28:27 +01:00
|
|
|
}
|
2015-05-09 06:49:47 +02:00
|
|
|
/* Preserve the locked attributes */
|
2015-06-04 16:43:11 +02:00
|
|
|
*new_mnt_flags |= mnt_flags & (MNT_LOCK_READONLY | \
|
|
|
|
MNT_LOCK_ATIME);
|
2013-03-31 04:57:41 +02:00
|
|
|
visible = true;
|
|
|
|
goto found;
|
|
|
|
next: ;
|
2013-03-24 22:28:27 +01:00
|
|
|
}
|
2013-03-31 04:57:41 +02:00
|
|
|
found:
|
2013-09-17 03:37:36 +02:00
|
|
|
up_read(&namespace_sem);
|
2013-03-31 04:57:41 +02:00
|
|
|
return visible;
|
2013-03-24 22:28:27 +01:00
|
|
|
}
|
|
|
|
|
2016-06-09 23:06:06 +02:00
|
|
|
static bool mount_too_revealing(struct vfsmount *mnt, int *new_mnt_flags)
|
|
|
|
{
|
2016-06-15 13:59:49 +02:00
|
|
|
const unsigned long required_iflags = SB_I_NOEXEC | SB_I_NODEV;
|
2016-06-09 23:06:06 +02:00
|
|
|
struct mnt_namespace *ns = current->nsproxy->mnt_ns;
|
|
|
|
unsigned long s_iflags;
|
|
|
|
|
|
|
|
if (ns->user_ns == &init_user_ns)
|
|
|
|
return false;
|
|
|
|
|
|
|
|
/* Can this filesystem be too revealing? */
|
|
|
|
s_iflags = mnt->mnt_sb->s_iflags;
|
|
|
|
if (!(s_iflags & SB_I_USERNS_VISIBLE))
|
|
|
|
return false;
|
|
|
|
|
2016-06-15 13:59:49 +02:00
|
|
|
if ((s_iflags & required_iflags) != required_iflags) {
|
|
|
|
WARN_ONCE(1, "Expected s_iflags to contain 0x%lx\n",
|
|
|
|
required_iflags);
|
|
|
|
return true;
|
|
|
|
}
|
|
|
|
|
2016-06-09 23:06:06 +02:00
|
|
|
return !mnt_already_visible(ns, mnt, new_mnt_flags);
|
|
|
|
}
|
|
|
|
|
fs: Treat foreign mounts as nosuid
If a process gets access to a mount from a different user
namespace, that process should not be able to take advantage of
setuid files or selinux entrypoints from that filesystem. Prevent
this by treating mounts from other mount namespaces and those not
owned by current_user_ns() or an ancestor as nosuid.
This will make it safer to allow more complex filesystems to be
mounted in non-root user namespaces.
This does not remove the need for MNT_LOCK_NOSUID. The setuid,
setgid, and file capability bits can no longer be abused if code in
a user namespace were to clear nosuid on an untrusted filesystem,
but this patch, by itself, is insufficient to protect the system
from abuse of files that, when execed, would increase MAC privilege.
As a more concrete explanation, any task that can manipulate a
vfsmount associated with a given user namespace already has
capabilities in that namespace and all of its descendents. If they
can cause a malicious setuid, setgid, or file-caps executable to
appear in that mount, then that executable will only allow them to
elevate privileges in exactly the set of namespaces in which they
are already privileges.
On the other hand, if they can cause a malicious executable to
appear with a dangerous MAC label, running it could change the
caller's security context in a way that should not have been
possible, even inside the namespace in which the task is confined.
As a hardening measure, this would have made CVE-2014-5207 much
more difficult to exploit.
Signed-off-by: Andy Lutomirski <luto@amacapital.net>
Signed-off-by: Seth Forshee <seth.forshee@canonical.com>
Acked-by: James Morris <james.l.morris@oracle.com>
Acked-by: Serge Hallyn <serge.hallyn@canonical.com>
Signed-off-by: Eric W. Biederman <ebiederm@xmission.com>
2016-06-23 23:41:05 +02:00
|
|
|
bool mnt_may_suid(struct vfsmount *mnt)
|
|
|
|
{
|
|
|
|
/*
|
|
|
|
* Foreign mounts (accessed via fchdir or through /proc
|
|
|
|
* symlinks) are always treated as if they are nosuid. This
|
|
|
|
* prevents namespaces from trusting potentially unsafe
|
|
|
|
* suid/sgid bits, file caps, or security labels that originate
|
|
|
|
* in other namespaces.
|
|
|
|
*/
|
|
|
|
return !(mnt->mnt_flags & MNT_NOSUID) && check_mnt(real_mount(mnt)) &&
|
|
|
|
current_in_userns(mnt->mnt_sb->s_user_ns);
|
|
|
|
}
|
|
|
|
|
2014-11-01 05:37:32 +01:00
|
|
|
static struct ns_common *mntns_get(struct task_struct *task)
|
2010-03-08 03:49:36 +01:00
|
|
|
{
|
2014-11-01 05:00:23 +01:00
|
|
|
struct ns_common *ns = NULL;
|
2010-03-08 03:49:36 +01:00
|
|
|
struct nsproxy *nsproxy;
|
|
|
|
|
2014-02-04 04:13:49 +01:00
|
|
|
task_lock(task);
|
|
|
|
nsproxy = task->nsproxy;
|
2010-03-08 03:49:36 +01:00
|
|
|
if (nsproxy) {
|
2014-11-01 05:00:23 +01:00
|
|
|
ns = &nsproxy->mnt_ns->ns;
|
|
|
|
get_mnt_ns(to_mnt_ns(ns));
|
2010-03-08 03:49:36 +01:00
|
|
|
}
|
2014-02-04 04:13:49 +01:00
|
|
|
task_unlock(task);
|
2010-03-08 03:49:36 +01:00
|
|
|
|
|
|
|
return ns;
|
|
|
|
}
|
|
|
|
|
2014-11-01 05:37:32 +01:00
|
|
|
static void mntns_put(struct ns_common *ns)
|
2010-03-08 03:49:36 +01:00
|
|
|
{
|
2014-11-01 05:00:23 +01:00
|
|
|
put_mnt_ns(to_mnt_ns(ns));
|
2010-03-08 03:49:36 +01:00
|
|
|
}
|
|
|
|
|
2014-11-01 05:37:32 +01:00
|
|
|
static int mntns_install(struct nsproxy *nsproxy, struct ns_common *ns)
|
2010-03-08 03:49:36 +01:00
|
|
|
{
|
|
|
|
struct fs_struct *fs = current->fs;
|
2017-04-15 23:31:22 +02:00
|
|
|
struct mnt_namespace *mnt_ns = to_mnt_ns(ns), *old_mnt_ns;
|
2010-03-08 03:49:36 +01:00
|
|
|
struct path root;
|
2017-04-15 23:31:22 +02:00
|
|
|
int err;
|
2010-03-08 03:49:36 +01:00
|
|
|
|
2012-07-27 06:42:03 +02:00
|
|
|
if (!ns_capable(mnt_ns->user_ns, CAP_SYS_ADMIN) ||
|
2013-03-20 20:49:49 +01:00
|
|
|
!ns_capable(current_user_ns(), CAP_SYS_CHROOT) ||
|
|
|
|
!ns_capable(current_user_ns(), CAP_SYS_ADMIN))
|
2012-09-13 10:38:03 +02:00
|
|
|
return -EPERM;
|
2010-03-08 03:49:36 +01:00
|
|
|
|
|
|
|
if (fs->users != 1)
|
|
|
|
return -EINVAL;
|
|
|
|
|
|
|
|
get_mnt_ns(mnt_ns);
|
2017-04-15 23:31:22 +02:00
|
|
|
old_mnt_ns = nsproxy->mnt_ns;
|
2010-03-08 03:49:36 +01:00
|
|
|
nsproxy->mnt_ns = mnt_ns;
|
|
|
|
|
|
|
|
/* Find the root */
|
2017-04-15 23:31:22 +02:00
|
|
|
err = vfs_path_lookup(mnt_ns->root->mnt.mnt_root, &mnt_ns->root->mnt,
|
|
|
|
"/", LOOKUP_DOWN, &root);
|
|
|
|
if (err) {
|
|
|
|
/* revert to old namespace */
|
|
|
|
nsproxy->mnt_ns = old_mnt_ns;
|
|
|
|
put_mnt_ns(mnt_ns);
|
|
|
|
return err;
|
|
|
|
}
|
2010-03-08 03:49:36 +01:00
|
|
|
|
2017-06-09 02:32:29 +02:00
|
|
|
put_mnt_ns(old_mnt_ns);
|
|
|
|
|
2010-03-08 03:49:36 +01:00
|
|
|
/* Update the pwd and root */
|
|
|
|
set_fs_pwd(fs, &root);
|
|
|
|
set_fs_root(fs, &root);
|
|
|
|
|
|
|
|
path_put(&root);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2016-09-06 09:47:13 +02:00
|
|
|
static struct user_namespace *mntns_owner(struct ns_common *ns)
|
|
|
|
{
|
|
|
|
return to_mnt_ns(ns)->user_ns;
|
|
|
|
}
|
|
|
|
|
2010-03-08 03:49:36 +01:00
|
|
|
const struct proc_ns_operations mntns_operations = {
|
|
|
|
.name = "mnt",
|
|
|
|
.type = CLONE_NEWNS,
|
|
|
|
.get = mntns_get,
|
|
|
|
.put = mntns_put,
|
|
|
|
.install = mntns_install,
|
2016-09-06 09:47:13 +02:00
|
|
|
.owner = mntns_owner,
|
2010-03-08 03:49:36 +01:00
|
|
|
};
|