There are two major features for this merge window. The first is
inline data, which allows small files or directories to be stored in the in-inode extended attribute area. (This requires that the file system use inodes which are at least 256 bytes or larger; 128 byte inodes do not have any room for in-inode xattrs.) The second new feature is SEEK_HOLE/SEEK_DATA support. This is enabled by the extent status tree patches, and this infrastructure will be used to further optimize ext4 in the future. Beyond that, we have the usual collection of code cleanups and bug fixes. -----BEGIN PGP SIGNATURE----- Version: GnuPG v1.4.12 (GNU/Linux) iQIcBAABCAAGBQJQzTaLAAoJENNvdpvBGATwpqEQAM0WO9Kva3R8SoaD6NYOg4lN 8oxRlht6yogSd6wwYZm1c4YF9UrhloS9kHyWcH3Wmr9fhM5vig1ec12eDsDGrjBc Wb+x+YrmczSJzK380JLxmYnVSXQVFl7/hNqaRowffTOJwgySmp8oLrI88ZcaCmVU +qWG2x6eVhCEQrpin9Mv3D6pHkx2hfg9w5sB0K+kpgsdjqLZsmPRmxU9nx0nEJYC gmbpo8Dcsfqra6DJosQGo7eFq7J3fm9v1ql+QOxOjc9/zD2XwdQE1JZImehvno5i Ekwr9771fsw34/QHJebYRC/OkftmOn4OPuQejd+AKNdBR4mO8G/AsLCroD17uLNi NrtMkE6ecJPb3SflarZruNYTUhJfj3H6V9P/8wggpyPzT3l19sqP+2F6GwZspZiV EJb2iTKn0Phc2OD1MqO9gFP0g+IMH0kktYdxEf0V2QOQqhQHnPwxF+2Tp6bVQcQs KCetN37y60qJ+zKH9xukcXmWQJvnjgmWqZqpomoA4lrwgKazTNDJJ+R+N+r5HKMj 5cz2ntAhF8FfPhqVf+8DHgjKNUwm6C++O1+Lb9swZ0FkFi5Ob3OlwWaC75Gf4H+P 2DslBapfM79bX14a9BKaBjly5FsAha7OzR+xo0MZN+fEcMLEk33kcRovcY8DHqxU aadriOatYYixvSZ5lL3m =aNOf -----END PGP SIGNATURE----- Merge tag 'ext4_for_linus' of git://git.kernel.org/pub/scm/linux/kernel/git/tytso/ext4 Pull ext4 update from Ted Ts'o: "There are two major features for this merge window. The first is inline data, which allows small files or directories to be stored in the in-inode extended attribute area. (This requires that the file system use inodes which are at least 256 bytes or larger; 128 byte inodes do not have any room for in-inode xattrs.) The second new feature is SEEK_HOLE/SEEK_DATA support. This is enabled by the extent status tree patches, and this infrastructure will be used to further optimize ext4 in the future. Beyond that, we have the usual collection of code cleanups and bug fixes." * tag 'ext4_for_linus' of git://git.kernel.org/pub/scm/linux/kernel/git/tytso/ext4: (63 commits) ext4: zero out inline data using memset() instead of empty_zero_page ext4: ensure Inode flags consistency are checked at build time ext4: Remove CONFIG_EXT4_FS_XATTR ext4: remove unused variable from ext4_ext_in_cache() ext4: remove redundant initialization in ext4_fill_super() ext4: remove redundant code in ext4_alloc_inode() ext4: use sync_inode_metadata() when syncing inode metadata ext4: enable ext4 inline support ext4: let fallocate handle inline data correctly ext4: let ext4_truncate handle inline data correctly ext4: evict inline data out if we need to strore xattr in inode ext4: let fiemap work with inline data ext4: let ext4_rename handle inline dir ext4: let empty_dir handle inline dir ext4: let ext4_delete_entry() handle inline data ext4: make ext4_delete_entry generic ext4: let ext4_find_entry handle inline data ext4: create a new function search_dir ext4: let ext4_readdir handle inline data ext4: let add_dir_entry handle inline data properly ...
This commit is contained in:
commit
36cd5c19c3
32 changed files with 4176 additions and 1107 deletions
|
@ -200,12 +200,9 @@ inode_readahead_blks=n This tuning parameter controls the maximum
|
|||
table readahead algorithm will pre-read into
|
||||
the buffer cache. The default value is 32 blocks.
|
||||
|
||||
nouser_xattr Disables Extended User Attributes. If you have extended
|
||||
attribute support enabled in the kernel configuration
|
||||
(CONFIG_EXT4_FS_XATTR), extended attribute support
|
||||
is enabled by default on mount. See the attr(5) manual
|
||||
page and http://acl.bestbits.at/ for more information
|
||||
about extended attributes.
|
||||
nouser_xattr Disables Extended User Attributes. See the
|
||||
attr(5) manual page and http://acl.bestbits.at/
|
||||
for more information about extended attributes.
|
||||
|
||||
noacl This option disables POSIX Access Control List
|
||||
support. If ACL support is enabled in the kernel
|
||||
|
|
|
@ -28,8 +28,8 @@ config FS_MBCACHE
|
|||
tristate
|
||||
default y if EXT2_FS=y && EXT2_FS_XATTR
|
||||
default y if EXT3_FS=y && EXT3_FS_XATTR
|
||||
default y if EXT4_FS=y && EXT4_FS_XATTR
|
||||
default m if EXT2_FS_XATTR || EXT3_FS_XATTR || EXT4_FS_XATTR
|
||||
default y if EXT4_FS=y
|
||||
default m if EXT2_FS_XATTR || EXT3_FS_XATTR || EXT4_FS
|
||||
|
||||
source "fs/reiserfs/Kconfig"
|
||||
source "fs/jfs/Kconfig"
|
||||
|
|
|
@ -39,22 +39,8 @@ config EXT4_USE_FOR_EXT23
|
|||
compiled kernel size by using one file system driver for
|
||||
ext2, ext3, and ext4 file systems.
|
||||
|
||||
config EXT4_FS_XATTR
|
||||
bool "Ext4 extended attributes"
|
||||
depends on EXT4_FS
|
||||
default y
|
||||
help
|
||||
Extended attributes are name:value pairs associated with inodes by
|
||||
the kernel or by users (see the attr(5) manual page, or visit
|
||||
<http://acl.bestbits.at/> for details).
|
||||
|
||||
If unsure, say N.
|
||||
|
||||
You need this for POSIX ACL support on ext4.
|
||||
|
||||
config EXT4_FS_POSIX_ACL
|
||||
bool "Ext4 POSIX Access Control Lists"
|
||||
depends on EXT4_FS_XATTR
|
||||
select FS_POSIX_ACL
|
||||
help
|
||||
POSIX Access Control Lists (ACLs) support permissions for users and
|
||||
|
@ -67,7 +53,6 @@ config EXT4_FS_POSIX_ACL
|
|||
|
||||
config EXT4_FS_SECURITY
|
||||
bool "Ext4 Security Labels"
|
||||
depends on EXT4_FS_XATTR
|
||||
help
|
||||
Security labels support alternative access control models
|
||||
implemented by security modules like SELinux. This option
|
||||
|
|
|
@ -7,8 +7,8 @@ obj-$(CONFIG_EXT4_FS) += ext4.o
|
|||
ext4-y := balloc.o bitmap.o dir.o file.o fsync.o ialloc.o inode.o page-io.o \
|
||||
ioctl.o namei.o super.o symlink.o hash.o resize.o extents.o \
|
||||
ext4_jbd2.o migrate.o mballoc.o block_validity.o move_extent.o \
|
||||
mmp.o indirect.o
|
||||
mmp.o indirect.o extents_status.o xattr.o xattr_user.o \
|
||||
xattr_trusted.o inline.o
|
||||
|
||||
ext4-$(CONFIG_EXT4_FS_XATTR) += xattr.o xattr_user.o xattr_trusted.o
|
||||
ext4-$(CONFIG_EXT4_FS_POSIX_ACL) += acl.o
|
||||
ext4-$(CONFIG_EXT4_FS_SECURITY) += xattr_security.o
|
||||
|
|
|
@ -423,8 +423,10 @@ ext4_xattr_set_acl(struct dentry *dentry, const char *name, const void *value,
|
|||
|
||||
retry:
|
||||
handle = ext4_journal_start(inode, EXT4_DATA_TRANS_BLOCKS(inode->i_sb));
|
||||
if (IS_ERR(handle))
|
||||
return PTR_ERR(handle);
|
||||
if (IS_ERR(handle)) {
|
||||
error = PTR_ERR(handle);
|
||||
goto release_and_out;
|
||||
}
|
||||
error = ext4_set_acl(handle, inode, type, acl);
|
||||
ext4_journal_stop(handle);
|
||||
if (error == -ENOSPC && ext4_should_retry_alloc(inode->i_sb, &retries))
|
||||
|
|
|
@ -27,23 +27,11 @@
|
|||
#include <linux/slab.h>
|
||||
#include <linux/rbtree.h>
|
||||
#include "ext4.h"
|
||||
|
||||
static unsigned char ext4_filetype_table[] = {
|
||||
DT_UNKNOWN, DT_REG, DT_DIR, DT_CHR, DT_BLK, DT_FIFO, DT_SOCK, DT_LNK
|
||||
};
|
||||
#include "xattr.h"
|
||||
|
||||
static int ext4_dx_readdir(struct file *filp,
|
||||
void *dirent, filldir_t filldir);
|
||||
|
||||
static unsigned char get_dtype(struct super_block *sb, int filetype)
|
||||
{
|
||||
if (!EXT4_HAS_INCOMPAT_FEATURE(sb, EXT4_FEATURE_INCOMPAT_FILETYPE) ||
|
||||
(filetype >= EXT4_FT_MAX))
|
||||
return DT_UNKNOWN;
|
||||
|
||||
return (ext4_filetype_table[filetype]);
|
||||
}
|
||||
|
||||
/**
|
||||
* Check if the given dir-inode refers to an htree-indexed directory
|
||||
* (or a directory which chould potentially get coverted to use htree
|
||||
|
@ -68,11 +56,14 @@ static int is_dx_dir(struct inode *inode)
|
|||
* Return 0 if the directory entry is OK, and 1 if there is a problem
|
||||
*
|
||||
* Note: this is the opposite of what ext2 and ext3 historically returned...
|
||||
*
|
||||
* bh passed here can be an inode block or a dir data block, depending
|
||||
* on the inode inline data flag.
|
||||
*/
|
||||
int __ext4_check_dir_entry(const char *function, unsigned int line,
|
||||
struct inode *dir, struct file *filp,
|
||||
struct ext4_dir_entry_2 *de,
|
||||
struct buffer_head *bh,
|
||||
struct buffer_head *bh, char *buf, int size,
|
||||
unsigned int offset)
|
||||
{
|
||||
const char *error_msg = NULL;
|
||||
|
@ -85,9 +76,8 @@ int __ext4_check_dir_entry(const char *function, unsigned int line,
|
|||
error_msg = "rec_len % 4 != 0";
|
||||
else if (unlikely(rlen < EXT4_DIR_REC_LEN(de->name_len)))
|
||||
error_msg = "rec_len is too small for name_len";
|
||||
else if (unlikely(((char *) de - bh->b_data) + rlen >
|
||||
dir->i_sb->s_blocksize))
|
||||
error_msg = "directory entry across blocks";
|
||||
else if (unlikely(((char *) de - buf) + rlen > size))
|
||||
error_msg = "directory entry across range";
|
||||
else if (unlikely(le32_to_cpu(de->inode) >
|
||||
le32_to_cpu(EXT4_SB(dir->i_sb)->s_es->s_inodes_count)))
|
||||
error_msg = "inode out of bounds";
|
||||
|
@ -98,14 +88,14 @@ int __ext4_check_dir_entry(const char *function, unsigned int line,
|
|||
ext4_error_file(filp, function, line, bh->b_blocknr,
|
||||
"bad entry in directory: %s - offset=%u(%u), "
|
||||
"inode=%u, rec_len=%d, name_len=%d",
|
||||
error_msg, (unsigned) (offset % bh->b_size),
|
||||
error_msg, (unsigned) (offset % size),
|
||||
offset, le32_to_cpu(de->inode),
|
||||
rlen, de->name_len);
|
||||
else
|
||||
ext4_error_inode(dir, function, line, bh->b_blocknr,
|
||||
"bad entry in directory: %s - offset=%u(%u), "
|
||||
"inode=%u, rec_len=%d, name_len=%d",
|
||||
error_msg, (unsigned) (offset % bh->b_size),
|
||||
error_msg, (unsigned) (offset % size),
|
||||
offset, le32_to_cpu(de->inode),
|
||||
rlen, de->name_len);
|
||||
|
||||
|
@ -125,6 +115,14 @@ static int ext4_readdir(struct file *filp,
|
|||
int ret = 0;
|
||||
int dir_has_error = 0;
|
||||
|
||||
if (ext4_has_inline_data(inode)) {
|
||||
int has_inline_data = 1;
|
||||
ret = ext4_read_inline_dir(filp, dirent, filldir,
|
||||
&has_inline_data);
|
||||
if (has_inline_data)
|
||||
return ret;
|
||||
}
|
||||
|
||||
if (is_dx_dir(inode)) {
|
||||
err = ext4_dx_readdir(filp, dirent, filldir);
|
||||
if (err != ERR_BAD_DX_DIR) {
|
||||
|
@ -221,8 +219,9 @@ revalidate:
|
|||
while (!error && filp->f_pos < inode->i_size
|
||||
&& offset < sb->s_blocksize) {
|
||||
de = (struct ext4_dir_entry_2 *) (bh->b_data + offset);
|
||||
if (ext4_check_dir_entry(inode, filp, de,
|
||||
bh, offset)) {
|
||||
if (ext4_check_dir_entry(inode, filp, de, bh,
|
||||
bh->b_data, bh->b_size,
|
||||
offset)) {
|
||||
/*
|
||||
* On error, skip the f_pos to the next block
|
||||
*/
|
||||
|
|
163
fs/ext4/ext4.h
163
fs/ext4/ext4.h
|
@ -57,6 +57,16 @@
|
|||
#define ext4_debug(fmt, ...) no_printk(fmt, ##__VA_ARGS__)
|
||||
#endif
|
||||
|
||||
/*
|
||||
* Turn on EXT_DEBUG to get lots of info about extents operations.
|
||||
*/
|
||||
#define EXT_DEBUG__
|
||||
#ifdef EXT_DEBUG
|
||||
#define ext_debug(fmt, ...) printk(fmt, ##__VA_ARGS__)
|
||||
#else
|
||||
#define ext_debug(fmt, ...) no_printk(fmt, ##__VA_ARGS__)
|
||||
#endif
|
||||
|
||||
#define EXT4_ERROR_INODE(inode, fmt, a...) \
|
||||
ext4_error_inode((inode), __func__, __LINE__, 0, (fmt), ## a)
|
||||
|
||||
|
@ -392,6 +402,7 @@ struct flex_groups {
|
|||
#define EXT4_EXTENTS_FL 0x00080000 /* Inode uses extents */
|
||||
#define EXT4_EA_INODE_FL 0x00200000 /* Inode used for large EA */
|
||||
#define EXT4_EOFBLOCKS_FL 0x00400000 /* Blocks allocated beyond EOF */
|
||||
#define EXT4_INLINE_DATA_FL 0x10000000 /* Inode has inline data. */
|
||||
#define EXT4_RESERVED_FL 0x80000000 /* reserved for ext4 lib */
|
||||
|
||||
#define EXT4_FL_USER_VISIBLE 0x004BDFFF /* User visible flags */
|
||||
|
@ -448,28 +459,26 @@ enum {
|
|||
EXT4_INODE_EXTENTS = 19, /* Inode uses extents */
|
||||
EXT4_INODE_EA_INODE = 21, /* Inode used for large EA */
|
||||
EXT4_INODE_EOFBLOCKS = 22, /* Blocks allocated beyond EOF */
|
||||
EXT4_INODE_INLINE_DATA = 28, /* Data in inode. */
|
||||
EXT4_INODE_RESERVED = 31, /* reserved for ext4 lib */
|
||||
};
|
||||
|
||||
#define TEST_FLAG_VALUE(FLAG) (EXT4_##FLAG##_FL == (1 << EXT4_INODE_##FLAG))
|
||||
#define CHECK_FLAG_VALUE(FLAG) if (!TEST_FLAG_VALUE(FLAG)) { \
|
||||
printk(KERN_EMERG "EXT4 flag fail: " #FLAG ": %d %d\n", \
|
||||
EXT4_##FLAG##_FL, EXT4_INODE_##FLAG); BUG_ON(1); }
|
||||
|
||||
/*
|
||||
* Since it's pretty easy to mix up bit numbers and hex values, and we
|
||||
* can't do a compile-time test for ENUM values, we use a run-time
|
||||
* test to make sure that EXT4_XXX_FL is consistent with respect to
|
||||
* EXT4_INODE_XXX. If all is well the printk and BUG_ON will all drop
|
||||
* out so it won't cost any extra space in the compiled kernel image.
|
||||
* But it's important that these values are the same, since we are
|
||||
* using EXT4_INODE_XXX to test for the flag values, but EXT4_XX_FL
|
||||
* must be consistent with the values of FS_XXX_FL defined in
|
||||
* include/linux/fs.h and the on-disk values found in ext2, ext3, and
|
||||
* ext4 filesystems, and of course the values defined in e2fsprogs.
|
||||
* Since it's pretty easy to mix up bit numbers and hex values, we use a
|
||||
* build-time check to make sure that EXT4_XXX_FL is consistent with respect to
|
||||
* EXT4_INODE_XXX. If all is well, the macros will be dropped, so, it won't cost
|
||||
* any extra space in the compiled kernel image, otherwise, the build will fail.
|
||||
* It's important that these values are the same, since we are using
|
||||
* EXT4_INODE_XXX to test for flag values, but EXT4_XXX_FL must be consistent
|
||||
* with the values of FS_XXX_FL defined in include/linux/fs.h and the on-disk
|
||||
* values found in ext2, ext3 and ext4 filesystems, and of course the values
|
||||
* defined in e2fsprogs.
|
||||
*
|
||||
* It's not paranoia if the Murphy's Law really *is* out to get you. :-)
|
||||
*/
|
||||
#define TEST_FLAG_VALUE(FLAG) (EXT4_##FLAG##_FL == (1 << EXT4_INODE_##FLAG))
|
||||
#define CHECK_FLAG_VALUE(FLAG) BUILD_BUG_ON(!TEST_FLAG_VALUE(FLAG))
|
||||
|
||||
static inline void ext4_check_flag_values(void)
|
||||
{
|
||||
CHECK_FLAG_VALUE(SECRM);
|
||||
|
@ -494,6 +503,7 @@ static inline void ext4_check_flag_values(void)
|
|||
CHECK_FLAG_VALUE(EXTENTS);
|
||||
CHECK_FLAG_VALUE(EA_INODE);
|
||||
CHECK_FLAG_VALUE(EOFBLOCKS);
|
||||
CHECK_FLAG_VALUE(INLINE_DATA);
|
||||
CHECK_FLAG_VALUE(RESERVED);
|
||||
}
|
||||
|
||||
|
@ -811,6 +821,8 @@ struct ext4_ext_cache {
|
|||
__u32 ec_len; /* must be 32bit to return holes */
|
||||
};
|
||||
|
||||
#include "extents_status.h"
|
||||
|
||||
/*
|
||||
* fourth extended file system inode data in memory
|
||||
*/
|
||||
|
@ -833,7 +845,6 @@ struct ext4_inode_info {
|
|||
#endif
|
||||
unsigned long i_flags;
|
||||
|
||||
#ifdef CONFIG_EXT4_FS_XATTR
|
||||
/*
|
||||
* Extended attributes can be read independently of the main file
|
||||
* data. Taking i_mutex even when reading would cause contention
|
||||
|
@ -842,7 +853,6 @@ struct ext4_inode_info {
|
|||
* EAs.
|
||||
*/
|
||||
struct rw_semaphore xattr_sem;
|
||||
#endif
|
||||
|
||||
struct list_head i_orphan; /* unlinked but open inodes */
|
||||
|
||||
|
@ -888,6 +898,10 @@ struct ext4_inode_info {
|
|||
struct list_head i_prealloc_list;
|
||||
spinlock_t i_prealloc_lock;
|
||||
|
||||
/* extents status tree */
|
||||
struct ext4_es_tree i_es_tree;
|
||||
rwlock_t i_es_lock;
|
||||
|
||||
/* ialloc */
|
||||
ext4_group_t i_last_alloc_group;
|
||||
|
||||
|
@ -902,6 +916,10 @@ struct ext4_inode_info {
|
|||
/* on-disk additional length */
|
||||
__u16 i_extra_isize;
|
||||
|
||||
/* Indicate the inline data space. */
|
||||
u16 i_inline_off;
|
||||
u16 i_inline_size;
|
||||
|
||||
#ifdef CONFIG_QUOTA
|
||||
/* quota space reservation, managed internally by quota code */
|
||||
qsize_t i_reserved_quota;
|
||||
|
@ -1360,6 +1378,7 @@ enum {
|
|||
EXT4_STATE_DELALLOC_RESERVED, /* blks already reserved for delalloc */
|
||||
EXT4_STATE_DIOREAD_LOCK, /* Disable support for dio read
|
||||
nolocking */
|
||||
EXT4_STATE_MAY_INLINE_DATA, /* may have in-inode data */
|
||||
};
|
||||
|
||||
#define EXT4_INODE_BIT_FNS(name, field, offset) \
|
||||
|
@ -1481,7 +1500,7 @@ static inline void ext4_clear_state_flags(struct ext4_inode_info *ei)
|
|||
#define EXT4_FEATURE_INCOMPAT_DIRDATA 0x1000 /* data in dirent */
|
||||
#define EXT4_FEATURE_INCOMPAT_BG_USE_META_CSUM 0x2000 /* use crc32c for bg */
|
||||
#define EXT4_FEATURE_INCOMPAT_LARGEDIR 0x4000 /* >2GB or 3-lvl htree */
|
||||
#define EXT4_FEATURE_INCOMPAT_INLINEDATA 0x8000 /* data in inode */
|
||||
#define EXT4_FEATURE_INCOMPAT_INLINE_DATA 0x8000 /* data in inode */
|
||||
|
||||
#define EXT2_FEATURE_COMPAT_SUPP EXT4_FEATURE_COMPAT_EXT_ATTR
|
||||
#define EXT2_FEATURE_INCOMPAT_SUPP (EXT4_FEATURE_INCOMPAT_FILETYPE| \
|
||||
|
@ -1505,7 +1524,8 @@ static inline void ext4_clear_state_flags(struct ext4_inode_info *ei)
|
|||
EXT4_FEATURE_INCOMPAT_EXTENTS| \
|
||||
EXT4_FEATURE_INCOMPAT_64BIT| \
|
||||
EXT4_FEATURE_INCOMPAT_FLEX_BG| \
|
||||
EXT4_FEATURE_INCOMPAT_MMP)
|
||||
EXT4_FEATURE_INCOMPAT_MMP | \
|
||||
EXT4_FEATURE_INCOMPAT_INLINE_DATA)
|
||||
#define EXT4_FEATURE_RO_COMPAT_SUPP (EXT4_FEATURE_RO_COMPAT_SPARSE_SUPER| \
|
||||
EXT4_FEATURE_RO_COMPAT_LARGE_FILE| \
|
||||
EXT4_FEATURE_RO_COMPAT_GDT_CSUM| \
|
||||
|
@ -1592,6 +1612,11 @@ struct ext4_dir_entry_tail {
|
|||
__le32 det_checksum; /* crc32c(uuid+inum+dirblock) */
|
||||
};
|
||||
|
||||
#define EXT4_DIRENT_TAIL(block, blocksize) \
|
||||
((struct ext4_dir_entry_tail *)(((void *)(block)) + \
|
||||
((blocksize) - \
|
||||
sizeof(struct ext4_dir_entry_tail))))
|
||||
|
||||
/*
|
||||
* Ext4 directory file types. Only the low 3 bits are used. The
|
||||
* other bits are reserved for now.
|
||||
|
@ -1936,14 +1961,42 @@ ext4_fsblk_t ext4_inode_to_goal_block(struct inode *);
|
|||
extern int __ext4_check_dir_entry(const char *, unsigned int, struct inode *,
|
||||
struct file *,
|
||||
struct ext4_dir_entry_2 *,
|
||||
struct buffer_head *, unsigned int);
|
||||
#define ext4_check_dir_entry(dir, filp, de, bh, offset) \
|
||||
struct buffer_head *, char *, int,
|
||||
unsigned int);
|
||||
#define ext4_check_dir_entry(dir, filp, de, bh, buf, size, offset) \
|
||||
unlikely(__ext4_check_dir_entry(__func__, __LINE__, (dir), (filp), \
|
||||
(de), (bh), (offset)))
|
||||
(de), (bh), (buf), (size), (offset)))
|
||||
extern int ext4_htree_store_dirent(struct file *dir_file, __u32 hash,
|
||||
__u32 minor_hash,
|
||||
struct ext4_dir_entry_2 *dirent);
|
||||
extern void ext4_htree_free_dir_info(struct dir_private_info *p);
|
||||
extern int ext4_find_dest_de(struct inode *dir, struct inode *inode,
|
||||
struct buffer_head *bh,
|
||||
void *buf, int buf_size,
|
||||
const char *name, int namelen,
|
||||
struct ext4_dir_entry_2 **dest_de);
|
||||
void ext4_insert_dentry(struct inode *inode,
|
||||
struct ext4_dir_entry_2 *de,
|
||||
int buf_size,
|
||||
const char *name, int namelen);
|
||||
static inline void ext4_update_dx_flag(struct inode *inode)
|
||||
{
|
||||
if (!EXT4_HAS_COMPAT_FEATURE(inode->i_sb,
|
||||
EXT4_FEATURE_COMPAT_DIR_INDEX))
|
||||
ext4_clear_inode_flag(inode, EXT4_INODE_INDEX);
|
||||
}
|
||||
static unsigned char ext4_filetype_table[] = {
|
||||
DT_UNKNOWN, DT_REG, DT_DIR, DT_CHR, DT_BLK, DT_FIFO, DT_SOCK, DT_LNK
|
||||
};
|
||||
|
||||
static inline unsigned char get_dtype(struct super_block *sb, int filetype)
|
||||
{
|
||||
if (!EXT4_HAS_INCOMPAT_FEATURE(sb, EXT4_FEATURE_INCOMPAT_FILETYPE) ||
|
||||
(filetype >= EXT4_FT_MAX))
|
||||
return DT_UNKNOWN;
|
||||
|
||||
return ext4_filetype_table[filetype];
|
||||
}
|
||||
|
||||
/* fsync.c */
|
||||
extern int ext4_sync_file(struct file *, loff_t, loff_t, int);
|
||||
|
@ -1994,8 +2047,23 @@ struct buffer_head *ext4_getblk(handle_t *, struct inode *,
|
|||
ext4_lblk_t, int, int *);
|
||||
struct buffer_head *ext4_bread(handle_t *, struct inode *,
|
||||
ext4_lblk_t, int, int *);
|
||||
int ext4_get_block_write(struct inode *inode, sector_t iblock,
|
||||
struct buffer_head *bh_result, int create);
|
||||
int ext4_get_block(struct inode *inode, sector_t iblock,
|
||||
struct buffer_head *bh_result, int create);
|
||||
int ext4_da_get_block_prep(struct inode *inode, sector_t iblock,
|
||||
struct buffer_head *bh, int create);
|
||||
int ext4_walk_page_buffers(handle_t *handle,
|
||||
struct buffer_head *head,
|
||||
unsigned from,
|
||||
unsigned to,
|
||||
int *partial,
|
||||
int (*fn)(handle_t *handle,
|
||||
struct buffer_head *bh));
|
||||
int do_journal_get_write_access(handle_t *handle,
|
||||
struct buffer_head *bh);
|
||||
#define FALL_BACK_TO_NONDELALLOC 1
|
||||
#define CONVERT_INLINE_DATA 2
|
||||
|
||||
extern struct inode *ext4_iget(struct super_block *, unsigned long);
|
||||
extern int ext4_write_inode(struct inode *, struct writeback_control *);
|
||||
|
@ -2050,6 +2118,20 @@ extern int ext4_orphan_add(handle_t *, struct inode *);
|
|||
extern int ext4_orphan_del(handle_t *, struct inode *);
|
||||
extern int ext4_htree_fill_tree(struct file *dir_file, __u32 start_hash,
|
||||
__u32 start_minor_hash, __u32 *next_hash);
|
||||
extern int search_dir(struct buffer_head *bh,
|
||||
char *search_buf,
|
||||
int buf_size,
|
||||
struct inode *dir,
|
||||
const struct qstr *d_name,
|
||||
unsigned int offset,
|
||||
struct ext4_dir_entry_2 **res_dir);
|
||||
extern int ext4_generic_delete_entry(handle_t *handle,
|
||||
struct inode *dir,
|
||||
struct ext4_dir_entry_2 *de_del,
|
||||
struct buffer_head *bh,
|
||||
void *entry_buf,
|
||||
int buf_size,
|
||||
int csum_size);
|
||||
|
||||
/* resize.c */
|
||||
extern int ext4_group_add(struct super_block *sb,
|
||||
|
@ -2376,6 +2458,15 @@ extern void ext4_unwritten_wait(struct inode *inode);
|
|||
extern const struct inode_operations ext4_dir_inode_operations;
|
||||
extern const struct inode_operations ext4_special_inode_operations;
|
||||
extern struct dentry *ext4_get_parent(struct dentry *child);
|
||||
extern struct ext4_dir_entry_2 *ext4_init_dot_dotdot(struct inode *inode,
|
||||
struct ext4_dir_entry_2 *de,
|
||||
int blocksize, int csum_size,
|
||||
unsigned int parent_ino, int dotdot_real_len);
|
||||
extern void initialize_dirent_tail(struct ext4_dir_entry_tail *t,
|
||||
unsigned int blocksize);
|
||||
extern int ext4_handle_dirty_dirent_node(handle_t *handle,
|
||||
struct inode *inode,
|
||||
struct buffer_head *bh);
|
||||
|
||||
/* symlink.c */
|
||||
extern const struct inode_operations ext4_symlink_inode_operations;
|
||||
|
@ -2393,6 +2484,9 @@ extern int ext4_check_blockref(const char *, unsigned int,
|
|||
struct inode *, __le32 *, unsigned int);
|
||||
|
||||
/* extents.c */
|
||||
struct ext4_ext_path;
|
||||
struct ext4_extent;
|
||||
|
||||
extern int ext4_ext_tree_init(handle_t *handle, struct inode *);
|
||||
extern int ext4_ext_writepage_trans_blocks(struct inode *, int);
|
||||
extern int ext4_ext_index_trans_blocks(struct inode *inode, int nrblocks,
|
||||
|
@ -2410,8 +2504,27 @@ extern int ext4_convert_unwritten_extents(struct inode *inode, loff_t offset,
|
|||
ssize_t len);
|
||||
extern int ext4_map_blocks(handle_t *handle, struct inode *inode,
|
||||
struct ext4_map_blocks *map, int flags);
|
||||
extern int ext4_ext_calc_metadata_amount(struct inode *inode,
|
||||
ext4_lblk_t lblocks);
|
||||
extern int ext4_extent_tree_init(handle_t *, struct inode *);
|
||||
extern int ext4_ext_calc_credits_for_single_extent(struct inode *inode,
|
||||
int num,
|
||||
struct ext4_ext_path *path);
|
||||
extern int ext4_can_extents_be_merged(struct inode *inode,
|
||||
struct ext4_extent *ex1,
|
||||
struct ext4_extent *ex2);
|
||||
extern int ext4_ext_insert_extent(handle_t *, struct inode *,
|
||||
struct ext4_ext_path *,
|
||||
struct ext4_extent *, int);
|
||||
extern struct ext4_ext_path *ext4_ext_find_extent(struct inode *, ext4_lblk_t,
|
||||
struct ext4_ext_path *);
|
||||
extern void ext4_ext_drop_refs(struct ext4_ext_path *);
|
||||
extern int ext4_ext_check_inode(struct inode *inode);
|
||||
extern int ext4_find_delalloc_cluster(struct inode *inode, ext4_lblk_t lblk);
|
||||
extern int ext4_fiemap(struct inode *inode, struct fiemap_extent_info *fieinfo,
|
||||
__u64 start, __u64 len);
|
||||
|
||||
|
||||
/* move_extent.c */
|
||||
extern int ext4_move_extents(struct file *o_filp, struct file *d_filp,
|
||||
__u64 start_orig, __u64 start_donor,
|
||||
|
@ -2445,14 +2558,10 @@ enum ext4_state_bits {
|
|||
* never, ever appear in a buffer_head's state
|
||||
* flag. See EXT4_MAP_FROM_CLUSTER to see where
|
||||
* this is used. */
|
||||
BH_Da_Mapped, /* Delayed allocated block that now has a mapping. This
|
||||
* flag is set when ext4_map_blocks is called on a
|
||||
* delayed allocated block to get its real mapping. */
|
||||
};
|
||||
|
||||
BUFFER_FNS(Uninit, uninit)
|
||||
TAS_BUFFER_FNS(Uninit, uninit)
|
||||
BUFFER_FNS(Da_Mapped, da_mapped)
|
||||
|
||||
/*
|
||||
* Add new method to test whether block and inode bitmaps are properly
|
||||
|
@ -2503,6 +2612,4 @@ extern void ext4_resize_end(struct super_block *sb);
|
|||
|
||||
#endif /* __KERNEL__ */
|
||||
|
||||
#include "ext4_extents.h"
|
||||
|
||||
#endif /* _EXT4_H */
|
||||
|
|
|
@ -42,16 +42,6 @@
|
|||
*/
|
||||
#define CHECK_BINSEARCH__
|
||||
|
||||
/*
|
||||
* Turn on EXT_DEBUG to get lots of info about extents operations.
|
||||
*/
|
||||
#define EXT_DEBUG__
|
||||
#ifdef EXT_DEBUG
|
||||
#define ext_debug(fmt, ...) printk(fmt, ##__VA_ARGS__)
|
||||
#else
|
||||
#define ext_debug(fmt, ...) no_printk(fmt, ##__VA_ARGS__)
|
||||
#endif
|
||||
|
||||
/*
|
||||
* If EXT_STATS is defined then stats numbers are collected.
|
||||
* These number will be displayed at umount time.
|
||||
|
@ -143,20 +133,6 @@ struct ext4_ext_path {
|
|||
* structure for external API
|
||||
*/
|
||||
|
||||
/*
|
||||
* to be called by ext4_ext_walk_space()
|
||||
* negative retcode - error
|
||||
* positive retcode - signal for ext4_ext_walk_space(), see below
|
||||
* callback must return valid extent (passed or newly created)
|
||||
*/
|
||||
typedef int (*ext_prepare_callback)(struct inode *, ext4_lblk_t,
|
||||
struct ext4_ext_cache *,
|
||||
struct ext4_extent *, void *);
|
||||
|
||||
#define EXT_CONTINUE 0
|
||||
#define EXT_BREAK 1
|
||||
#define EXT_REPEAT 2
|
||||
|
||||
/*
|
||||
* Maximum number of logical blocks in a file; ext4_extent's ee_block is
|
||||
* __le32.
|
||||
|
@ -300,21 +276,5 @@ static inline void ext4_idx_store_pblock(struct ext4_extent_idx *ix,
|
|||
0xffff);
|
||||
}
|
||||
|
||||
extern int ext4_ext_calc_metadata_amount(struct inode *inode,
|
||||
ext4_lblk_t lblocks);
|
||||
extern int ext4_extent_tree_init(handle_t *, struct inode *);
|
||||
extern int ext4_ext_calc_credits_for_single_extent(struct inode *inode,
|
||||
int num,
|
||||
struct ext4_ext_path *path);
|
||||
extern int ext4_can_extents_be_merged(struct inode *inode,
|
||||
struct ext4_extent *ex1,
|
||||
struct ext4_extent *ex2);
|
||||
extern int ext4_ext_insert_extent(handle_t *, struct inode *, struct ext4_ext_path *, struct ext4_extent *, int);
|
||||
extern struct ext4_ext_path *ext4_ext_find_extent(struct inode *, ext4_lblk_t,
|
||||
struct ext4_ext_path *);
|
||||
extern void ext4_ext_drop_refs(struct ext4_ext_path *);
|
||||
extern int ext4_ext_check_inode(struct inode *inode);
|
||||
extern int ext4_find_delalloc_cluster(struct inode *inode, ext4_lblk_t lblk,
|
||||
int search_hint_reverse);
|
||||
#endif /* _EXT4_EXTENTS */
|
||||
|
||||
|
|
|
@ -254,13 +254,6 @@ static inline void ext4_handle_sync(handle_t *handle)
|
|||
handle->h_sync = 1;
|
||||
}
|
||||
|
||||
static inline void ext4_handle_release_buffer(handle_t *handle,
|
||||
struct buffer_head *bh)
|
||||
{
|
||||
if (ext4_handle_valid(handle))
|
||||
jbd2_journal_release_buffer(handle, bh);
|
||||
}
|
||||
|
||||
static inline int ext4_handle_is_aborted(handle_t *handle)
|
||||
{
|
||||
if (ext4_handle_valid(handle))
|
||||
|
|
|
@ -41,6 +41,8 @@
|
|||
#include <asm/uaccess.h>
|
||||
#include <linux/fiemap.h>
|
||||
#include "ext4_jbd2.h"
|
||||
#include "ext4_extents.h"
|
||||
#include "xattr.h"
|
||||
|
||||
#include <trace/events/ext4.h>
|
||||
|
||||
|
@ -109,6 +111,9 @@ static int ext4_split_extent_at(handle_t *handle,
|
|||
int split_flag,
|
||||
int flags);
|
||||
|
||||
static int ext4_find_delayed_extent(struct inode *inode,
|
||||
struct ext4_ext_cache *newex);
|
||||
|
||||
static int ext4_ext_truncate_extend_restart(handle_t *handle,
|
||||
struct inode *inode,
|
||||
int needed)
|
||||
|
@ -1959,27 +1964,33 @@ cleanup:
|
|||
return err;
|
||||
}
|
||||
|
||||
static int ext4_ext_walk_space(struct inode *inode, ext4_lblk_t block,
|
||||
ext4_lblk_t num, ext_prepare_callback func,
|
||||
void *cbdata)
|
||||
static int ext4_fill_fiemap_extents(struct inode *inode,
|
||||
ext4_lblk_t block, ext4_lblk_t num,
|
||||
struct fiemap_extent_info *fieinfo)
|
||||
{
|
||||
struct ext4_ext_path *path = NULL;
|
||||
struct ext4_ext_cache cbex;
|
||||
struct ext4_ext_cache newex;
|
||||
struct ext4_extent *ex;
|
||||
ext4_lblk_t next, start = 0, end = 0;
|
||||
ext4_lblk_t next, next_del, start = 0, end = 0;
|
||||
ext4_lblk_t last = block + num;
|
||||
int depth, exists, err = 0;
|
||||
|
||||
BUG_ON(func == NULL);
|
||||
BUG_ON(inode == NULL);
|
||||
int exists, depth = 0, err = 0;
|
||||
unsigned int flags = 0;
|
||||
unsigned char blksize_bits = inode->i_sb->s_blocksize_bits;
|
||||
|
||||
while (block < last && block != EXT_MAX_BLOCKS) {
|
||||
num = last - block;
|
||||
/* find extent for this block */
|
||||
down_read(&EXT4_I(inode)->i_data_sem);
|
||||
|
||||
if (path && ext_depth(inode) != depth) {
|
||||
/* depth was changed. we have to realloc path */
|
||||
kfree(path);
|
||||
path = NULL;
|
||||
}
|
||||
|
||||
path = ext4_ext_find_extent(inode, block, path);
|
||||
up_read(&EXT4_I(inode)->i_data_sem);
|
||||
if (IS_ERR(path)) {
|
||||
up_read(&EXT4_I(inode)->i_data_sem);
|
||||
err = PTR_ERR(path);
|
||||
path = NULL;
|
||||
break;
|
||||
|
@ -1987,13 +1998,16 @@ static int ext4_ext_walk_space(struct inode *inode, ext4_lblk_t block,
|
|||
|
||||
depth = ext_depth(inode);
|
||||
if (unlikely(path[depth].p_hdr == NULL)) {
|
||||
up_read(&EXT4_I(inode)->i_data_sem);
|
||||
EXT4_ERROR_INODE(inode, "path[%d].p_hdr == NULL", depth);
|
||||
err = -EIO;
|
||||
break;
|
||||
}
|
||||
ex = path[depth].p_ext;
|
||||
next = ext4_ext_next_allocated_block(path);
|
||||
ext4_ext_drop_refs(path);
|
||||
|
||||
flags = 0;
|
||||
exists = 0;
|
||||
if (!ex) {
|
||||
/* there is no extent yet, so try to allocate
|
||||
|
@ -2030,40 +2044,64 @@ static int ext4_ext_walk_space(struct inode *inode, ext4_lblk_t block,
|
|||
BUG_ON(end <= start);
|
||||
|
||||
if (!exists) {
|
||||
cbex.ec_block = start;
|
||||
cbex.ec_len = end - start;
|
||||
cbex.ec_start = 0;
|
||||
newex.ec_block = start;
|
||||
newex.ec_len = end - start;
|
||||
newex.ec_start = 0;
|
||||
} else {
|
||||
cbex.ec_block = le32_to_cpu(ex->ee_block);
|
||||
cbex.ec_len = ext4_ext_get_actual_len(ex);
|
||||
cbex.ec_start = ext4_ext_pblock(ex);
|
||||
newex.ec_block = le32_to_cpu(ex->ee_block);
|
||||
newex.ec_len = ext4_ext_get_actual_len(ex);
|
||||
newex.ec_start = ext4_ext_pblock(ex);
|
||||
if (ext4_ext_is_uninitialized(ex))
|
||||
flags |= FIEMAP_EXTENT_UNWRITTEN;
|
||||
}
|
||||
|
||||
if (unlikely(cbex.ec_len == 0)) {
|
||||
EXT4_ERROR_INODE(inode, "cbex.ec_len == 0");
|
||||
/*
|
||||
* Find delayed extent and update newex accordingly. We call
|
||||
* it even in !exists case to find out whether newex is the
|
||||
* last existing extent or not.
|
||||
*/
|
||||
next_del = ext4_find_delayed_extent(inode, &newex);
|
||||
if (!exists && next_del) {
|
||||
exists = 1;
|
||||
flags |= FIEMAP_EXTENT_DELALLOC;
|
||||
}
|
||||
up_read(&EXT4_I(inode)->i_data_sem);
|
||||
|
||||
if (unlikely(newex.ec_len == 0)) {
|
||||
EXT4_ERROR_INODE(inode, "newex.ec_len == 0");
|
||||
err = -EIO;
|
||||
break;
|
||||
}
|
||||
err = func(inode, next, &cbex, ex, cbdata);
|
||||
ext4_ext_drop_refs(path);
|
||||
|
||||
if (err < 0)
|
||||
break;
|
||||
|
||||
if (err == EXT_REPEAT)
|
||||
continue;
|
||||
else if (err == EXT_BREAK) {
|
||||
err = 0;
|
||||
break;
|
||||
/* This is possible iff next == next_del == EXT_MAX_BLOCKS */
|
||||
if (next == next_del) {
|
||||
flags |= FIEMAP_EXTENT_LAST;
|
||||
if (unlikely(next_del != EXT_MAX_BLOCKS ||
|
||||
next != EXT_MAX_BLOCKS)) {
|
||||
EXT4_ERROR_INODE(inode,
|
||||
"next extent == %u, next "
|
||||
"delalloc extent = %u",
|
||||
next, next_del);
|
||||
err = -EIO;
|
||||
break;
|
||||
}
|
||||
}
|
||||
|
||||
if (ext_depth(inode) != depth) {
|
||||
/* depth was changed. we have to realloc path */
|
||||
kfree(path);
|
||||
path = NULL;
|
||||
if (exists) {
|
||||
err = fiemap_fill_next_extent(fieinfo,
|
||||
(__u64)newex.ec_block << blksize_bits,
|
||||
(__u64)newex.ec_start << blksize_bits,
|
||||
(__u64)newex.ec_len << blksize_bits,
|
||||
flags);
|
||||
if (err < 0)
|
||||
break;
|
||||
if (err == 1) {
|
||||
err = 0;
|
||||
break;
|
||||
}
|
||||
}
|
||||
|
||||
block = cbex.ec_block + cbex.ec_len;
|
||||
block = newex.ec_block + newex.ec_len;
|
||||
}
|
||||
|
||||
if (path) {
|
||||
|
@ -2156,7 +2194,6 @@ ext4_ext_in_cache(struct inode *inode, ext4_lblk_t block,
|
|||
struct ext4_extent *ex)
|
||||
{
|
||||
struct ext4_ext_cache *cex;
|
||||
struct ext4_sb_info *sbi;
|
||||
int ret = 0;
|
||||
|
||||
/*
|
||||
|
@ -2164,7 +2201,6 @@ ext4_ext_in_cache(struct inode *inode, ext4_lblk_t block,
|
|||
*/
|
||||
spin_lock(&EXT4_I(inode)->i_block_reservation_lock);
|
||||
cex = &EXT4_I(inode)->i_cached_extent;
|
||||
sbi = EXT4_SB(inode->i_sb);
|
||||
|
||||
/* has cache valid data? */
|
||||
if (cex->ec_len == 0)
|
||||
|
@ -2273,7 +2309,13 @@ int ext4_ext_calc_credits_for_single_extent(struct inode *inode, int nrblocks,
|
|||
int ext4_ext_index_trans_blocks(struct inode *inode, int nrblocks, int chunk)
|
||||
{
|
||||
int index;
|
||||
int depth = ext_depth(inode);
|
||||
int depth;
|
||||
|
||||
/* If we are converting the inline data, only one is needed here. */
|
||||
if (ext4_has_inline_data(inode))
|
||||
return 1;
|
||||
|
||||
depth = ext_depth(inode);
|
||||
|
||||
if (chunk)
|
||||
index = depth * 2;
|
||||
|
@ -3461,115 +3503,34 @@ out:
|
|||
/**
|
||||
* ext4_find_delalloc_range: find delayed allocated block in the given range.
|
||||
*
|
||||
* Goes through the buffer heads in the range [lblk_start, lblk_end] and returns
|
||||
* whether there are any buffers marked for delayed allocation. It returns '1'
|
||||
* on the first delalloc'ed buffer head found. If no buffer head in the given
|
||||
* range is marked for delalloc, it returns 0.
|
||||
* lblk_start should always be <= lblk_end.
|
||||
* search_hint_reverse is to indicate that searching in reverse from lblk_end to
|
||||
* lblk_start might be more efficient (i.e., we will likely hit the delalloc'ed
|
||||
* block sooner). This is useful when blocks are truncated sequentially from
|
||||
* lblk_start towards lblk_end.
|
||||
* Return 1 if there is a delalloc block in the range, otherwise 0.
|
||||
*/
|
||||
static int ext4_find_delalloc_range(struct inode *inode,
|
||||
ext4_lblk_t lblk_start,
|
||||
ext4_lblk_t lblk_end,
|
||||
int search_hint_reverse)
|
||||
ext4_lblk_t lblk_end)
|
||||
{
|
||||
struct address_space *mapping = inode->i_mapping;
|
||||
struct buffer_head *head, *bh = NULL;
|
||||
struct page *page;
|
||||
ext4_lblk_t i, pg_lblk;
|
||||
pgoff_t index;
|
||||
struct extent_status es;
|
||||
|
||||
if (!test_opt(inode->i_sb, DELALLOC))
|
||||
return 0;
|
||||
|
||||
/* reverse search wont work if fs block size is less than page size */
|
||||
if (inode->i_blkbits < PAGE_CACHE_SHIFT)
|
||||
search_hint_reverse = 0;
|
||||
|
||||
if (search_hint_reverse)
|
||||
i = lblk_end;
|
||||
es.start = lblk_start;
|
||||
ext4_es_find_extent(inode, &es);
|
||||
if (es.len == 0)
|
||||
return 0; /* there is no delay extent in this tree */
|
||||
else if (es.start <= lblk_start && lblk_start < es.start + es.len)
|
||||
return 1;
|
||||
else if (lblk_start <= es.start && es.start <= lblk_end)
|
||||
return 1;
|
||||
else
|
||||
i = lblk_start;
|
||||
|
||||
index = i >> (PAGE_CACHE_SHIFT - inode->i_blkbits);
|
||||
|
||||
while ((i >= lblk_start) && (i <= lblk_end)) {
|
||||
page = find_get_page(mapping, index);
|
||||
if (!page)
|
||||
goto nextpage;
|
||||
|
||||
if (!page_has_buffers(page))
|
||||
goto nextpage;
|
||||
|
||||
head = page_buffers(page);
|
||||
if (!head)
|
||||
goto nextpage;
|
||||
|
||||
bh = head;
|
||||
pg_lblk = index << (PAGE_CACHE_SHIFT -
|
||||
inode->i_blkbits);
|
||||
do {
|
||||
if (unlikely(pg_lblk < lblk_start)) {
|
||||
/*
|
||||
* This is possible when fs block size is less
|
||||
* than page size and our cluster starts/ends in
|
||||
* middle of the page. So we need to skip the
|
||||
* initial few blocks till we reach the 'lblk'
|
||||
*/
|
||||
pg_lblk++;
|
||||
continue;
|
||||
}
|
||||
|
||||
/* Check if the buffer is delayed allocated and that it
|
||||
* is not yet mapped. (when da-buffers are mapped during
|
||||
* their writeout, their da_mapped bit is set.)
|
||||
*/
|
||||
if (buffer_delay(bh) && !buffer_da_mapped(bh)) {
|
||||
page_cache_release(page);
|
||||
trace_ext4_find_delalloc_range(inode,
|
||||
lblk_start, lblk_end,
|
||||
search_hint_reverse,
|
||||
1, i);
|
||||
return 1;
|
||||
}
|
||||
if (search_hint_reverse)
|
||||
i--;
|
||||
else
|
||||
i++;
|
||||
} while ((i >= lblk_start) && (i <= lblk_end) &&
|
||||
((bh = bh->b_this_page) != head));
|
||||
nextpage:
|
||||
if (page)
|
||||
page_cache_release(page);
|
||||
/*
|
||||
* Move to next page. 'i' will be the first lblk in the next
|
||||
* page.
|
||||
*/
|
||||
if (search_hint_reverse)
|
||||
index--;
|
||||
else
|
||||
index++;
|
||||
i = index << (PAGE_CACHE_SHIFT - inode->i_blkbits);
|
||||
}
|
||||
|
||||
trace_ext4_find_delalloc_range(inode, lblk_start, lblk_end,
|
||||
search_hint_reverse, 0, 0);
|
||||
return 0;
|
||||
return 0;
|
||||
}
|
||||
|
||||
int ext4_find_delalloc_cluster(struct inode *inode, ext4_lblk_t lblk,
|
||||
int search_hint_reverse)
|
||||
int ext4_find_delalloc_cluster(struct inode *inode, ext4_lblk_t lblk)
|
||||
{
|
||||
struct ext4_sb_info *sbi = EXT4_SB(inode->i_sb);
|
||||
ext4_lblk_t lblk_start, lblk_end;
|
||||
lblk_start = lblk & (~(sbi->s_cluster_ratio - 1));
|
||||
lblk_end = lblk_start + sbi->s_cluster_ratio - 1;
|
||||
|
||||
return ext4_find_delalloc_range(inode, lblk_start, lblk_end,
|
||||
search_hint_reverse);
|
||||
return ext4_find_delalloc_range(inode, lblk_start, lblk_end);
|
||||
}
|
||||
|
||||
/**
|
||||
|
@ -3630,7 +3591,7 @@ get_reserved_cluster_alloc(struct inode *inode, ext4_lblk_t lblk_start,
|
|||
lblk_from = lblk_start & (~(sbi->s_cluster_ratio - 1));
|
||||
lblk_to = lblk_from + c_offset - 1;
|
||||
|
||||
if (ext4_find_delalloc_range(inode, lblk_from, lblk_to, 0))
|
||||
if (ext4_find_delalloc_range(inode, lblk_from, lblk_to))
|
||||
allocated_clusters--;
|
||||
}
|
||||
|
||||
|
@ -3640,7 +3601,7 @@ get_reserved_cluster_alloc(struct inode *inode, ext4_lblk_t lblk_start,
|
|||
lblk_from = lblk_start + num_blks;
|
||||
lblk_to = lblk_from + (sbi->s_cluster_ratio - c_offset) - 1;
|
||||
|
||||
if (ext4_find_delalloc_range(inode, lblk_from, lblk_to, 0))
|
||||
if (ext4_find_delalloc_range(inode, lblk_from, lblk_to))
|
||||
allocated_clusters--;
|
||||
}
|
||||
|
||||
|
@ -3663,8 +3624,8 @@ ext4_ext_handle_uninitialized_extents(handle_t *handle, struct inode *inode,
|
|||
flags, allocated);
|
||||
ext4_ext_show_leaf(inode, path);
|
||||
|
||||
trace_ext4_ext_handle_uninitialized_extents(inode, map, allocated,
|
||||
newblock);
|
||||
trace_ext4_ext_handle_uninitialized_extents(inode, map, flags,
|
||||
allocated, newblock);
|
||||
|
||||
/* get_block() before submit the IO, split the extent */
|
||||
if ((flags & EXT4_GET_BLOCKS_PRE_IO)) {
|
||||
|
@ -3911,7 +3872,7 @@ int ext4_ext_map_blocks(handle_t *handle, struct inode *inode,
|
|||
struct ext4_extent newex, *ex, *ex2;
|
||||
struct ext4_sb_info *sbi = EXT4_SB(inode->i_sb);
|
||||
ext4_fsblk_t newblock = 0;
|
||||
int free_on_err = 0, err = 0, depth, ret;
|
||||
int free_on_err = 0, err = 0, depth;
|
||||
unsigned int allocated = 0, offset = 0;
|
||||
unsigned int allocated_clusters = 0;
|
||||
struct ext4_allocation_request ar;
|
||||
|
@ -3927,7 +3888,7 @@ int ext4_ext_map_blocks(handle_t *handle, struct inode *inode,
|
|||
if (ext4_ext_in_cache(inode, map->m_lblk, &newex)) {
|
||||
if (!newex.ee_start_lo && !newex.ee_start_hi) {
|
||||
if ((sbi->s_cluster_ratio > 1) &&
|
||||
ext4_find_delalloc_cluster(inode, map->m_lblk, 0))
|
||||
ext4_find_delalloc_cluster(inode, map->m_lblk))
|
||||
map->m_flags |= EXT4_MAP_FROM_CLUSTER;
|
||||
|
||||
if ((flags & EXT4_GET_BLOCKS_CREATE) == 0) {
|
||||
|
@ -4007,15 +3968,15 @@ int ext4_ext_map_blocks(handle_t *handle, struct inode *inode,
|
|||
ee_len, ee_start);
|
||||
goto out;
|
||||
}
|
||||
ret = ext4_ext_handle_uninitialized_extents(
|
||||
allocated = ext4_ext_handle_uninitialized_extents(
|
||||
handle, inode, map, path, flags,
|
||||
allocated, newblock);
|
||||
return ret;
|
||||
goto out3;
|
||||
}
|
||||
}
|
||||
|
||||
if ((sbi->s_cluster_ratio > 1) &&
|
||||
ext4_find_delalloc_cluster(inode, map->m_lblk, 0))
|
||||
ext4_find_delalloc_cluster(inode, map->m_lblk))
|
||||
map->m_flags |= EXT4_MAP_FROM_CLUSTER;
|
||||
|
||||
/*
|
||||
|
@ -4284,8 +4245,8 @@ out2:
|
|||
kfree(path);
|
||||
}
|
||||
|
||||
trace_ext4_ext_map_blocks_exit(inode, map->m_lblk,
|
||||
newblock, map->m_len, err ? err : allocated);
|
||||
out3:
|
||||
trace_ext4_ext_map_blocks_exit(inode, map, err ? err : allocated);
|
||||
|
||||
return err ? err : allocated;
|
||||
}
|
||||
|
@ -4344,6 +4305,8 @@ void ext4_ext_truncate(struct inode *inode)
|
|||
|
||||
last_block = (inode->i_size + sb->s_blocksize - 1)
|
||||
>> EXT4_BLOCK_SIZE_BITS(sb);
|
||||
err = ext4_es_remove_extent(inode, last_block,
|
||||
EXT_MAX_BLOCKS - last_block);
|
||||
err = ext4_ext_remove_space(inode, last_block, EXT_MAX_BLOCKS - 1);
|
||||
|
||||
/* In a multi-transaction truncate, we only make the final
|
||||
|
@ -4434,6 +4397,10 @@ long ext4_fallocate(struct file *file, int mode, loff_t offset, loff_t len)
|
|||
if (mode & FALLOC_FL_PUNCH_HOLE)
|
||||
return ext4_punch_hole(file, offset, len);
|
||||
|
||||
ret = ext4_convert_inline_data(inode);
|
||||
if (ret)
|
||||
return ret;
|
||||
|
||||
trace_ext4_fallocate_enter(inode, offset, len, mode);
|
||||
map.m_lblk = offset >> blkbits;
|
||||
/*
|
||||
|
@ -4572,206 +4539,43 @@ int ext4_convert_unwritten_extents(struct inode *inode, loff_t offset,
|
|||
}
|
||||
|
||||
/*
|
||||
* Callback function called for each extent to gather FIEMAP information.
|
||||
* If newex is not existing extent (newex->ec_start equals zero) find
|
||||
* delayed extent at start of newex and update newex accordingly and
|
||||
* return start of the next delayed extent.
|
||||
*
|
||||
* If newex is existing extent (newex->ec_start is not equal zero)
|
||||
* return start of next delayed extent or EXT_MAX_BLOCKS if no delayed
|
||||
* extent found. Leave newex unmodified.
|
||||
*/
|
||||
static int ext4_ext_fiemap_cb(struct inode *inode, ext4_lblk_t next,
|
||||
struct ext4_ext_cache *newex, struct ext4_extent *ex,
|
||||
void *data)
|
||||
static int ext4_find_delayed_extent(struct inode *inode,
|
||||
struct ext4_ext_cache *newex)
|
||||
{
|
||||
__u64 logical;
|
||||
__u64 physical;
|
||||
__u64 length;
|
||||
__u32 flags = 0;
|
||||
int ret = 0;
|
||||
struct fiemap_extent_info *fieinfo = data;
|
||||
unsigned char blksize_bits;
|
||||
struct extent_status es;
|
||||
ext4_lblk_t next_del;
|
||||
|
||||
blksize_bits = inode->i_sb->s_blocksize_bits;
|
||||
logical = (__u64)newex->ec_block << blksize_bits;
|
||||
es.start = newex->ec_block;
|
||||
next_del = ext4_es_find_extent(inode, &es);
|
||||
|
||||
if (newex->ec_start == 0) {
|
||||
/*
|
||||
* No extent in extent-tree contains block @newex->ec_start,
|
||||
* then the block may stay in 1)a hole or 2)delayed-extent.
|
||||
*
|
||||
* Holes or delayed-extents are processed as follows.
|
||||
* 1. lookup dirty pages with specified range in pagecache.
|
||||
* If no page is got, then there is no delayed-extent and
|
||||
* return with EXT_CONTINUE.
|
||||
* 2. find the 1st mapped buffer,
|
||||
* 3. check if the mapped buffer is both in the request range
|
||||
* and a delayed buffer. If not, there is no delayed-extent,
|
||||
* then return.
|
||||
* 4. a delayed-extent is found, the extent will be collected.
|
||||
*/
|
||||
ext4_lblk_t end = 0;
|
||||
pgoff_t last_offset;
|
||||
pgoff_t offset;
|
||||
pgoff_t index;
|
||||
pgoff_t start_index = 0;
|
||||
struct page **pages = NULL;
|
||||
struct buffer_head *bh = NULL;
|
||||
struct buffer_head *head = NULL;
|
||||
unsigned int nr_pages = PAGE_SIZE / sizeof(struct page *);
|
||||
if (es.len == 0)
|
||||
/* A hole found. */
|
||||
return 0;
|
||||
|
||||
pages = kmalloc(PAGE_SIZE, GFP_KERNEL);
|
||||
if (pages == NULL)
|
||||
return -ENOMEM;
|
||||
|
||||
offset = logical >> PAGE_SHIFT;
|
||||
repeat:
|
||||
last_offset = offset;
|
||||
head = NULL;
|
||||
ret = find_get_pages_tag(inode->i_mapping, &offset,
|
||||
PAGECACHE_TAG_DIRTY, nr_pages, pages);
|
||||
|
||||
if (!(flags & FIEMAP_EXTENT_DELALLOC)) {
|
||||
/* First time, try to find a mapped buffer. */
|
||||
if (ret == 0) {
|
||||
out:
|
||||
for (index = 0; index < ret; index++)
|
||||
page_cache_release(pages[index]);
|
||||
/* just a hole. */
|
||||
kfree(pages);
|
||||
return EXT_CONTINUE;
|
||||
}
|
||||
index = 0;
|
||||
|
||||
next_page:
|
||||
/* Try to find the 1st mapped buffer. */
|
||||
end = ((__u64)pages[index]->index << PAGE_SHIFT) >>
|
||||
blksize_bits;
|
||||
if (!page_has_buffers(pages[index]))
|
||||
goto out;
|
||||
head = page_buffers(pages[index]);
|
||||
if (!head)
|
||||
goto out;
|
||||
|
||||
index++;
|
||||
bh = head;
|
||||
do {
|
||||
if (end >= newex->ec_block +
|
||||
newex->ec_len)
|
||||
/* The buffer is out of
|
||||
* the request range.
|
||||
*/
|
||||
goto out;
|
||||
|
||||
if (buffer_mapped(bh) &&
|
||||
end >= newex->ec_block) {
|
||||
start_index = index - 1;
|
||||
/* get the 1st mapped buffer. */
|
||||
goto found_mapped_buffer;
|
||||
}
|
||||
|
||||
bh = bh->b_this_page;
|
||||
end++;
|
||||
} while (bh != head);
|
||||
|
||||
/* No mapped buffer in the range found in this page,
|
||||
* We need to look up next page.
|
||||
*/
|
||||
if (index >= ret) {
|
||||
/* There is no page left, but we need to limit
|
||||
* newex->ec_len.
|
||||
*/
|
||||
newex->ec_len = end - newex->ec_block;
|
||||
goto out;
|
||||
}
|
||||
goto next_page;
|
||||
} else {
|
||||
/*Find contiguous delayed buffers. */
|
||||
if (ret > 0 && pages[0]->index == last_offset)
|
||||
head = page_buffers(pages[0]);
|
||||
bh = head;
|
||||
index = 1;
|
||||
start_index = 0;
|
||||
if (es.start > newex->ec_block) {
|
||||
/* A hole found. */
|
||||
newex->ec_len = min(es.start - newex->ec_block,
|
||||
newex->ec_len);
|
||||
return 0;
|
||||
}
|
||||
|
||||
found_mapped_buffer:
|
||||
if (bh != NULL && buffer_delay(bh)) {
|
||||
/* 1st or contiguous delayed buffer found. */
|
||||
if (!(flags & FIEMAP_EXTENT_DELALLOC)) {
|
||||
/*
|
||||
* 1st delayed buffer found, record
|
||||
* the start of extent.
|
||||
*/
|
||||
flags |= FIEMAP_EXTENT_DELALLOC;
|
||||
newex->ec_block = end;
|
||||
logical = (__u64)end << blksize_bits;
|
||||
}
|
||||
/* Find contiguous delayed buffers. */
|
||||
do {
|
||||
if (!buffer_delay(bh))
|
||||
goto found_delayed_extent;
|
||||
bh = bh->b_this_page;
|
||||
end++;
|
||||
} while (bh != head);
|
||||
|
||||
for (; index < ret; index++) {
|
||||
if (!page_has_buffers(pages[index])) {
|
||||
bh = NULL;
|
||||
break;
|
||||
}
|
||||
head = page_buffers(pages[index]);
|
||||
if (!head) {
|
||||
bh = NULL;
|
||||
break;
|
||||
}
|
||||
|
||||
if (pages[index]->index !=
|
||||
pages[start_index]->index + index
|
||||
- start_index) {
|
||||
/* Blocks are not contiguous. */
|
||||
bh = NULL;
|
||||
break;
|
||||
}
|
||||
bh = head;
|
||||
do {
|
||||
if (!buffer_delay(bh))
|
||||
/* Delayed-extent ends. */
|
||||
goto found_delayed_extent;
|
||||
bh = bh->b_this_page;
|
||||
end++;
|
||||
} while (bh != head);
|
||||
}
|
||||
} else if (!(flags & FIEMAP_EXTENT_DELALLOC))
|
||||
/* a hole found. */
|
||||
goto out;
|
||||
|
||||
found_delayed_extent:
|
||||
newex->ec_len = min(end - newex->ec_block,
|
||||
(ext4_lblk_t)EXT_INIT_MAX_LEN);
|
||||
if (ret == nr_pages && bh != NULL &&
|
||||
newex->ec_len < EXT_INIT_MAX_LEN &&
|
||||
buffer_delay(bh)) {
|
||||
/* Have not collected an extent and continue. */
|
||||
for (index = 0; index < ret; index++)
|
||||
page_cache_release(pages[index]);
|
||||
goto repeat;
|
||||
}
|
||||
|
||||
for (index = 0; index < ret; index++)
|
||||
page_cache_release(pages[index]);
|
||||
kfree(pages);
|
||||
newex->ec_len = es.start + es.len - newex->ec_block;
|
||||
}
|
||||
|
||||
physical = (__u64)newex->ec_start << blksize_bits;
|
||||
length = (__u64)newex->ec_len << blksize_bits;
|
||||
|
||||
if (ex && ext4_ext_is_uninitialized(ex))
|
||||
flags |= FIEMAP_EXTENT_UNWRITTEN;
|
||||
|
||||
if (next == EXT_MAX_BLOCKS)
|
||||
flags |= FIEMAP_EXTENT_LAST;
|
||||
|
||||
ret = fiemap_fill_next_extent(fieinfo, logical, physical,
|
||||
length, flags);
|
||||
if (ret < 0)
|
||||
return ret;
|
||||
if (ret == 1)
|
||||
return EXT_BREAK;
|
||||
return EXT_CONTINUE;
|
||||
return next_del;
|
||||
}
|
||||
/* fiemap flags we can handle specified here */
|
||||
#define EXT4_FIEMAP_FLAGS (FIEMAP_FLAG_SYNC|FIEMAP_FLAG_XATTR)
|
||||
|
@ -4971,6 +4775,8 @@ int ext4_ext_punch_hole(struct file *file, loff_t offset, loff_t length)
|
|||
ext4_ext_invalidate_cache(inode);
|
||||
ext4_discard_preallocations(inode);
|
||||
|
||||
err = ext4_es_remove_extent(inode, first_block,
|
||||
stop_block - first_block);
|
||||
err = ext4_ext_remove_space(inode, first_block, stop_block - 1);
|
||||
|
||||
ext4_ext_invalidate_cache(inode);
|
||||
|
@ -4991,12 +4797,22 @@ out_mutex:
|
|||
mutex_unlock(&inode->i_mutex);
|
||||
return err;
|
||||
}
|
||||
|
||||
int ext4_fiemap(struct inode *inode, struct fiemap_extent_info *fieinfo,
|
||||
__u64 start, __u64 len)
|
||||
{
|
||||
ext4_lblk_t start_blk;
|
||||
int error = 0;
|
||||
|
||||
if (ext4_has_inline_data(inode)) {
|
||||
int has_inline = 1;
|
||||
|
||||
error = ext4_inline_data_fiemap(inode, fieinfo, &has_inline);
|
||||
|
||||
if (has_inline)
|
||||
return error;
|
||||
}
|
||||
|
||||
/* fallback to generic here if not in extents fmt */
|
||||
if (!(ext4_test_inode_flag(inode, EXT4_INODE_EXTENTS)))
|
||||
return generic_block_fiemap(inode, fieinfo, start, len,
|
||||
|
@ -5018,11 +4834,11 @@ int ext4_fiemap(struct inode *inode, struct fiemap_extent_info *fieinfo,
|
|||
len_blks = ((ext4_lblk_t) last_blk) - start_blk + 1;
|
||||
|
||||
/*
|
||||
* Walk the extent tree gathering extent information.
|
||||
* ext4_ext_fiemap_cb will push extents back to user.
|
||||
* Walk the extent tree gathering extent information
|
||||
* and pushing extents back to the user.
|
||||
*/
|
||||
error = ext4_ext_walk_space(inode, start_blk, len_blks,
|
||||
ext4_ext_fiemap_cb, fieinfo);
|
||||
error = ext4_fill_fiemap_extents(inode, start_blk,
|
||||
len_blks, fieinfo);
|
||||
}
|
||||
|
||||
return error;
|
||||
|
|
500
fs/ext4/extents_status.c
Normal file
500
fs/ext4/extents_status.c
Normal file
|
@ -0,0 +1,500 @@
|
|||
/*
|
||||
* fs/ext4/extents_status.c
|
||||
*
|
||||
* Written by Yongqiang Yang <xiaoqiangnk@gmail.com>
|
||||
* Modified by
|
||||
* Allison Henderson <achender@linux.vnet.ibm.com>
|
||||
* Hugh Dickins <hughd@google.com>
|
||||
* Zheng Liu <wenqing.lz@taobao.com>
|
||||
*
|
||||
* Ext4 extents status tree core functions.
|
||||
*/
|
||||
#include <linux/rbtree.h>
|
||||
#include "ext4.h"
|
||||
#include "extents_status.h"
|
||||
#include "ext4_extents.h"
|
||||
|
||||
#include <trace/events/ext4.h>
|
||||
|
||||
/*
|
||||
* According to previous discussion in Ext4 Developer Workshop, we
|
||||
* will introduce a new structure called io tree to track all extent
|
||||
* status in order to solve some problems that we have met
|
||||
* (e.g. Reservation space warning), and provide extent-level locking.
|
||||
* Delay extent tree is the first step to achieve this goal. It is
|
||||
* original built by Yongqiang Yang. At that time it is called delay
|
||||
* extent tree, whose goal is only track delay extent in memory to
|
||||
* simplify the implementation of fiemap and bigalloc, and introduce
|
||||
* lseek SEEK_DATA/SEEK_HOLE support. That is why it is still called
|
||||
* delay extent tree at the following comment. But for better
|
||||
* understand what it does, it has been rename to extent status tree.
|
||||
*
|
||||
* Currently the first step has been done. All delay extents are
|
||||
* tracked in the tree. It maintains the delay extent when a delay
|
||||
* allocation is issued, and the delay extent is written out or
|
||||
* invalidated. Therefore the implementation of fiemap and bigalloc
|
||||
* are simplified, and SEEK_DATA/SEEK_HOLE are introduced.
|
||||
*
|
||||
* The following comment describes the implemenmtation of extent
|
||||
* status tree and future works.
|
||||
*/
|
||||
|
||||
/*
|
||||
* extents status tree implementation for ext4.
|
||||
*
|
||||
*
|
||||
* ==========================================================================
|
||||
* Extents status encompass delayed extents and extent locks
|
||||
*
|
||||
* 1. Why delayed extent implementation ?
|
||||
*
|
||||
* Without delayed extent, ext4 identifies a delayed extent by looking
|
||||
* up page cache, this has several deficiencies - complicated, buggy,
|
||||
* and inefficient code.
|
||||
*
|
||||
* FIEMAP, SEEK_HOLE/DATA, bigalloc, punch hole and writeout all need
|
||||
* to know if a block or a range of blocks are belonged to a delayed
|
||||
* extent.
|
||||
*
|
||||
* Let us have a look at how they do without delayed extents implementation.
|
||||
* -- FIEMAP
|
||||
* FIEMAP looks up page cache to identify delayed allocations from holes.
|
||||
*
|
||||
* -- SEEK_HOLE/DATA
|
||||
* SEEK_HOLE/DATA has the same problem as FIEMAP.
|
||||
*
|
||||
* -- bigalloc
|
||||
* bigalloc looks up page cache to figure out if a block is
|
||||
* already under delayed allocation or not to determine whether
|
||||
* quota reserving is needed for the cluster.
|
||||
*
|
||||
* -- punch hole
|
||||
* punch hole looks up page cache to identify a delayed extent.
|
||||
*
|
||||
* -- writeout
|
||||
* Writeout looks up whole page cache to see if a buffer is
|
||||
* mapped, If there are not very many delayed buffers, then it is
|
||||
* time comsuming.
|
||||
*
|
||||
* With delayed extents implementation, FIEMAP, SEEK_HOLE/DATA,
|
||||
* bigalloc and writeout can figure out if a block or a range of
|
||||
* blocks is under delayed allocation(belonged to a delayed extent) or
|
||||
* not by searching the delayed extent tree.
|
||||
*
|
||||
*
|
||||
* ==========================================================================
|
||||
* 2. ext4 delayed extents impelmentation
|
||||
*
|
||||
* -- delayed extent
|
||||
* A delayed extent is a range of blocks which are contiguous
|
||||
* logically and under delayed allocation. Unlike extent in
|
||||
* ext4, delayed extent in ext4 is a in-memory struct, there is
|
||||
* no corresponding on-disk data. There is no limit on length of
|
||||
* delayed extent, so a delayed extent can contain as many blocks
|
||||
* as they are contiguous logically.
|
||||
*
|
||||
* -- delayed extent tree
|
||||
* Every inode has a delayed extent tree and all under delayed
|
||||
* allocation blocks are added to the tree as delayed extents.
|
||||
* Delayed extents in the tree are ordered by logical block no.
|
||||
*
|
||||
* -- operations on a delayed extent tree
|
||||
* There are three operations on a delayed extent tree: find next
|
||||
* delayed extent, adding a space(a range of blocks) and removing
|
||||
* a space.
|
||||
*
|
||||
* -- race on a delayed extent tree
|
||||
* Delayed extent tree is protected inode->i_es_lock.
|
||||
*
|
||||
*
|
||||
* ==========================================================================
|
||||
* 3. performance analysis
|
||||
* -- overhead
|
||||
* 1. There is a cache extent for write access, so if writes are
|
||||
* not very random, adding space operaions are in O(1) time.
|
||||
*
|
||||
* -- gain
|
||||
* 2. Code is much simpler, more readable, more maintainable and
|
||||
* more efficient.
|
||||
*
|
||||
*
|
||||
* ==========================================================================
|
||||
* 4. TODO list
|
||||
* -- Track all extent status
|
||||
*
|
||||
* -- Improve get block process
|
||||
*
|
||||
* -- Extent-level locking
|
||||
*/
|
||||
|
||||
static struct kmem_cache *ext4_es_cachep;
|
||||
|
||||
int __init ext4_init_es(void)
|
||||
{
|
||||
ext4_es_cachep = KMEM_CACHE(extent_status, SLAB_RECLAIM_ACCOUNT);
|
||||
if (ext4_es_cachep == NULL)
|
||||
return -ENOMEM;
|
||||
return 0;
|
||||
}
|
||||
|
||||
void ext4_exit_es(void)
|
||||
{
|
||||
if (ext4_es_cachep)
|
||||
kmem_cache_destroy(ext4_es_cachep);
|
||||
}
|
||||
|
||||
void ext4_es_init_tree(struct ext4_es_tree *tree)
|
||||
{
|
||||
tree->root = RB_ROOT;
|
||||
tree->cache_es = NULL;
|
||||
}
|
||||
|
||||
#ifdef ES_DEBUG__
|
||||
static void ext4_es_print_tree(struct inode *inode)
|
||||
{
|
||||
struct ext4_es_tree *tree;
|
||||
struct rb_node *node;
|
||||
|
||||
printk(KERN_DEBUG "status extents for inode %lu:", inode->i_ino);
|
||||
tree = &EXT4_I(inode)->i_es_tree;
|
||||
node = rb_first(&tree->root);
|
||||
while (node) {
|
||||
struct extent_status *es;
|
||||
es = rb_entry(node, struct extent_status, rb_node);
|
||||
printk(KERN_DEBUG " [%u/%u)", es->start, es->len);
|
||||
node = rb_next(node);
|
||||
}
|
||||
printk(KERN_DEBUG "\n");
|
||||
}
|
||||
#else
|
||||
#define ext4_es_print_tree(inode)
|
||||
#endif
|
||||
|
||||
static inline ext4_lblk_t extent_status_end(struct extent_status *es)
|
||||
{
|
||||
BUG_ON(es->start + es->len < es->start);
|
||||
return es->start + es->len - 1;
|
||||
}
|
||||
|
||||
/*
|
||||
* search through the tree for an delayed extent with a given offset. If
|
||||
* it can't be found, try to find next extent.
|
||||
*/
|
||||
static struct extent_status *__es_tree_search(struct rb_root *root,
|
||||
ext4_lblk_t offset)
|
||||
{
|
||||
struct rb_node *node = root->rb_node;
|
||||
struct extent_status *es = NULL;
|
||||
|
||||
while (node) {
|
||||
es = rb_entry(node, struct extent_status, rb_node);
|
||||
if (offset < es->start)
|
||||
node = node->rb_left;
|
||||
else if (offset > extent_status_end(es))
|
||||
node = node->rb_right;
|
||||
else
|
||||
return es;
|
||||
}
|
||||
|
||||
if (es && offset < es->start)
|
||||
return es;
|
||||
|
||||
if (es && offset > extent_status_end(es)) {
|
||||
node = rb_next(&es->rb_node);
|
||||
return node ? rb_entry(node, struct extent_status, rb_node) :
|
||||
NULL;
|
||||
}
|
||||
|
||||
return NULL;
|
||||
}
|
||||
|
||||
/*
|
||||
* ext4_es_find_extent: find the 1st delayed extent covering @es->start
|
||||
* if it exists, otherwise, the next extent after @es->start.
|
||||
*
|
||||
* @inode: the inode which owns delayed extents
|
||||
* @es: delayed extent that we found
|
||||
*
|
||||
* Returns the first block of the next extent after es, otherwise
|
||||
* EXT_MAX_BLOCKS if no delay extent is found.
|
||||
* Delayed extent is returned via @es.
|
||||
*/
|
||||
ext4_lblk_t ext4_es_find_extent(struct inode *inode, struct extent_status *es)
|
||||
{
|
||||
struct ext4_es_tree *tree = NULL;
|
||||
struct extent_status *es1 = NULL;
|
||||
struct rb_node *node;
|
||||
ext4_lblk_t ret = EXT_MAX_BLOCKS;
|
||||
|
||||
trace_ext4_es_find_extent_enter(inode, es->start);
|
||||
|
||||
read_lock(&EXT4_I(inode)->i_es_lock);
|
||||
tree = &EXT4_I(inode)->i_es_tree;
|
||||
|
||||
/* find delay extent in cache firstly */
|
||||
if (tree->cache_es) {
|
||||
es1 = tree->cache_es;
|
||||
if (in_range(es->start, es1->start, es1->len)) {
|
||||
es_debug("%u cached by [%u/%u)\n",
|
||||
es->start, es1->start, es1->len);
|
||||
goto out;
|
||||
}
|
||||
}
|
||||
|
||||
es->len = 0;
|
||||
es1 = __es_tree_search(&tree->root, es->start);
|
||||
|
||||
out:
|
||||
if (es1) {
|
||||
tree->cache_es = es1;
|
||||
es->start = es1->start;
|
||||
es->len = es1->len;
|
||||
node = rb_next(&es1->rb_node);
|
||||
if (node) {
|
||||
es1 = rb_entry(node, struct extent_status, rb_node);
|
||||
ret = es1->start;
|
||||
}
|
||||
}
|
||||
|
||||
read_unlock(&EXT4_I(inode)->i_es_lock);
|
||||
|
||||
trace_ext4_es_find_extent_exit(inode, es, ret);
|
||||
return ret;
|
||||
}
|
||||
|
||||
static struct extent_status *
|
||||
ext4_es_alloc_extent(ext4_lblk_t start, ext4_lblk_t len)
|
||||
{
|
||||
struct extent_status *es;
|
||||
es = kmem_cache_alloc(ext4_es_cachep, GFP_ATOMIC);
|
||||
if (es == NULL)
|
||||
return NULL;
|
||||
es->start = start;
|
||||
es->len = len;
|
||||
return es;
|
||||
}
|
||||
|
||||
static void ext4_es_free_extent(struct extent_status *es)
|
||||
{
|
||||
kmem_cache_free(ext4_es_cachep, es);
|
||||
}
|
||||
|
||||
static struct extent_status *
|
||||
ext4_es_try_to_merge_left(struct ext4_es_tree *tree, struct extent_status *es)
|
||||
{
|
||||
struct extent_status *es1;
|
||||
struct rb_node *node;
|
||||
|
||||
node = rb_prev(&es->rb_node);
|
||||
if (!node)
|
||||
return es;
|
||||
|
||||
es1 = rb_entry(node, struct extent_status, rb_node);
|
||||
if (es->start == extent_status_end(es1) + 1) {
|
||||
es1->len += es->len;
|
||||
rb_erase(&es->rb_node, &tree->root);
|
||||
ext4_es_free_extent(es);
|
||||
es = es1;
|
||||
}
|
||||
|
||||
return es;
|
||||
}
|
||||
|
||||
static struct extent_status *
|
||||
ext4_es_try_to_merge_right(struct ext4_es_tree *tree, struct extent_status *es)
|
||||
{
|
||||
struct extent_status *es1;
|
||||
struct rb_node *node;
|
||||
|
||||
node = rb_next(&es->rb_node);
|
||||
if (!node)
|
||||
return es;
|
||||
|
||||
es1 = rb_entry(node, struct extent_status, rb_node);
|
||||
if (es1->start == extent_status_end(es) + 1) {
|
||||
es->len += es1->len;
|
||||
rb_erase(node, &tree->root);
|
||||
ext4_es_free_extent(es1);
|
||||
}
|
||||
|
||||
return es;
|
||||
}
|
||||
|
||||
static int __es_insert_extent(struct ext4_es_tree *tree, ext4_lblk_t offset,
|
||||
ext4_lblk_t len)
|
||||
{
|
||||
struct rb_node **p = &tree->root.rb_node;
|
||||
struct rb_node *parent = NULL;
|
||||
struct extent_status *es;
|
||||
ext4_lblk_t end = offset + len - 1;
|
||||
|
||||
BUG_ON(end < offset);
|
||||
es = tree->cache_es;
|
||||
if (es && offset == (extent_status_end(es) + 1)) {
|
||||
es_debug("cached by [%u/%u)\n", es->start, es->len);
|
||||
es->len += len;
|
||||
es = ext4_es_try_to_merge_right(tree, es);
|
||||
goto out;
|
||||
} else if (es && es->start == end + 1) {
|
||||
es_debug("cached by [%u/%u)\n", es->start, es->len);
|
||||
es->start = offset;
|
||||
es->len += len;
|
||||
es = ext4_es_try_to_merge_left(tree, es);
|
||||
goto out;
|
||||
} else if (es && es->start <= offset &&
|
||||
end <= extent_status_end(es)) {
|
||||
es_debug("cached by [%u/%u)\n", es->start, es->len);
|
||||
goto out;
|
||||
}
|
||||
|
||||
while (*p) {
|
||||
parent = *p;
|
||||
es = rb_entry(parent, struct extent_status, rb_node);
|
||||
|
||||
if (offset < es->start) {
|
||||
if (es->start == end + 1) {
|
||||
es->start = offset;
|
||||
es->len += len;
|
||||
es = ext4_es_try_to_merge_left(tree, es);
|
||||
goto out;
|
||||
}
|
||||
p = &(*p)->rb_left;
|
||||
} else if (offset > extent_status_end(es)) {
|
||||
if (offset == extent_status_end(es) + 1) {
|
||||
es->len += len;
|
||||
es = ext4_es_try_to_merge_right(tree, es);
|
||||
goto out;
|
||||
}
|
||||
p = &(*p)->rb_right;
|
||||
} else {
|
||||
if (extent_status_end(es) <= end)
|
||||
es->len = offset - es->start + len;
|
||||
goto out;
|
||||
}
|
||||
}
|
||||
|
||||
es = ext4_es_alloc_extent(offset, len);
|
||||
if (!es)
|
||||
return -ENOMEM;
|
||||
rb_link_node(&es->rb_node, parent, p);
|
||||
rb_insert_color(&es->rb_node, &tree->root);
|
||||
|
||||
out:
|
||||
tree->cache_es = es;
|
||||
return 0;
|
||||
}
|
||||
|
||||
/*
|
||||
* ext4_es_insert_extent() adds a space to a delayed extent tree.
|
||||
* Caller holds inode->i_es_lock.
|
||||
*
|
||||
* ext4_es_insert_extent is called by ext4_da_write_begin and
|
||||
* ext4_es_remove_extent.
|
||||
*
|
||||
* Return 0 on success, error code on failure.
|
||||
*/
|
||||
int ext4_es_insert_extent(struct inode *inode, ext4_lblk_t offset,
|
||||
ext4_lblk_t len)
|
||||
{
|
||||
struct ext4_es_tree *tree;
|
||||
int err = 0;
|
||||
|
||||
trace_ext4_es_insert_extent(inode, offset, len);
|
||||
es_debug("add [%u/%u) to extent status tree of inode %lu\n",
|
||||
offset, len, inode->i_ino);
|
||||
|
||||
write_lock(&EXT4_I(inode)->i_es_lock);
|
||||
tree = &EXT4_I(inode)->i_es_tree;
|
||||
err = __es_insert_extent(tree, offset, len);
|
||||
write_unlock(&EXT4_I(inode)->i_es_lock);
|
||||
|
||||
ext4_es_print_tree(inode);
|
||||
|
||||
return err;
|
||||
}
|
||||
|
||||
/*
|
||||
* ext4_es_remove_extent() removes a space from a delayed extent tree.
|
||||
* Caller holds inode->i_es_lock.
|
||||
*
|
||||
* Return 0 on success, error code on failure.
|
||||
*/
|
||||
int ext4_es_remove_extent(struct inode *inode, ext4_lblk_t offset,
|
||||
ext4_lblk_t len)
|
||||
{
|
||||
struct rb_node *node;
|
||||
struct ext4_es_tree *tree;
|
||||
struct extent_status *es;
|
||||
struct extent_status orig_es;
|
||||
ext4_lblk_t len1, len2, end;
|
||||
int err = 0;
|
||||
|
||||
trace_ext4_es_remove_extent(inode, offset, len);
|
||||
es_debug("remove [%u/%u) from extent status tree of inode %lu\n",
|
||||
offset, len, inode->i_ino);
|
||||
|
||||
end = offset + len - 1;
|
||||
BUG_ON(end < offset);
|
||||
write_lock(&EXT4_I(inode)->i_es_lock);
|
||||
tree = &EXT4_I(inode)->i_es_tree;
|
||||
es = __es_tree_search(&tree->root, offset);
|
||||
if (!es)
|
||||
goto out;
|
||||
if (es->start > end)
|
||||
goto out;
|
||||
|
||||
/* Simply invalidate cache_es. */
|
||||
tree->cache_es = NULL;
|
||||
|
||||
orig_es.start = es->start;
|
||||
orig_es.len = es->len;
|
||||
len1 = offset > es->start ? offset - es->start : 0;
|
||||
len2 = extent_status_end(es) > end ?
|
||||
extent_status_end(es) - end : 0;
|
||||
if (len1 > 0)
|
||||
es->len = len1;
|
||||
if (len2 > 0) {
|
||||
if (len1 > 0) {
|
||||
err = __es_insert_extent(tree, end + 1, len2);
|
||||
if (err) {
|
||||
es->start = orig_es.start;
|
||||
es->len = orig_es.len;
|
||||
goto out;
|
||||
}
|
||||
} else {
|
||||
es->start = end + 1;
|
||||
es->len = len2;
|
||||
}
|
||||
goto out;
|
||||
}
|
||||
|
||||
if (len1 > 0) {
|
||||
node = rb_next(&es->rb_node);
|
||||
if (node)
|
||||
es = rb_entry(node, struct extent_status, rb_node);
|
||||
else
|
||||
es = NULL;
|
||||
}
|
||||
|
||||
while (es && extent_status_end(es) <= end) {
|
||||
node = rb_next(&es->rb_node);
|
||||
rb_erase(&es->rb_node, &tree->root);
|
||||
ext4_es_free_extent(es);
|
||||
if (!node) {
|
||||
es = NULL;
|
||||
break;
|
||||
}
|
||||
es = rb_entry(node, struct extent_status, rb_node);
|
||||
}
|
||||
|
||||
if (es && es->start < end + 1) {
|
||||
len1 = extent_status_end(es) - end;
|
||||
es->start = end + 1;
|
||||
es->len = len1;
|
||||
}
|
||||
|
||||
out:
|
||||
write_unlock(&EXT4_I(inode)->i_es_lock);
|
||||
ext4_es_print_tree(inode);
|
||||
return err;
|
||||
}
|
45
fs/ext4/extents_status.h
Normal file
45
fs/ext4/extents_status.h
Normal file
|
@ -0,0 +1,45 @@
|
|||
/*
|
||||
* fs/ext4/extents_status.h
|
||||
*
|
||||
* Written by Yongqiang Yang <xiaoqiangnk@gmail.com>
|
||||
* Modified by
|
||||
* Allison Henderson <achender@linux.vnet.ibm.com>
|
||||
* Zheng Liu <wenqing.lz@taobao.com>
|
||||
*
|
||||
*/
|
||||
|
||||
#ifndef _EXT4_EXTENTS_STATUS_H
|
||||
#define _EXT4_EXTENTS_STATUS_H
|
||||
|
||||
/*
|
||||
* Turn on ES_DEBUG__ to get lots of info about extent status operations.
|
||||
*/
|
||||
#ifdef ES_DEBUG__
|
||||
#define es_debug(fmt, ...) printk(fmt, ##__VA_ARGS__)
|
||||
#else
|
||||
#define es_debug(fmt, ...) no_printk(fmt, ##__VA_ARGS__)
|
||||
#endif
|
||||
|
||||
struct extent_status {
|
||||
struct rb_node rb_node;
|
||||
ext4_lblk_t start; /* first block extent covers */
|
||||
ext4_lblk_t len; /* length of extent in block */
|
||||
};
|
||||
|
||||
struct ext4_es_tree {
|
||||
struct rb_root root;
|
||||
struct extent_status *cache_es; /* recently accessed extent */
|
||||
};
|
||||
|
||||
extern int __init ext4_init_es(void);
|
||||
extern void ext4_exit_es(void);
|
||||
extern void ext4_es_init_tree(struct ext4_es_tree *tree);
|
||||
|
||||
extern int ext4_es_insert_extent(struct inode *inode, ext4_lblk_t start,
|
||||
ext4_lblk_t len);
|
||||
extern int ext4_es_remove_extent(struct inode *inode, ext4_lblk_t start,
|
||||
ext4_lblk_t len);
|
||||
extern ext4_lblk_t ext4_es_find_extent(struct inode *inode,
|
||||
struct extent_status *es);
|
||||
|
||||
#endif /* _EXT4_EXTENTS_STATUS_H */
|
336
fs/ext4/file.c
336
fs/ext4/file.c
|
@ -24,6 +24,7 @@
|
|||
#include <linux/mount.h>
|
||||
#include <linux/path.h>
|
||||
#include <linux/quotaops.h>
|
||||
#include <linux/pagevec.h>
|
||||
#include "ext4.h"
|
||||
#include "ext4_jbd2.h"
|
||||
#include "xattr.h"
|
||||
|
@ -285,6 +286,324 @@ static int ext4_file_open(struct inode * inode, struct file * filp)
|
|||
return dquot_file_open(inode, filp);
|
||||
}
|
||||
|
||||
/*
|
||||
* Here we use ext4_map_blocks() to get a block mapping for a extent-based
|
||||
* file rather than ext4_ext_walk_space() because we can introduce
|
||||
* SEEK_DATA/SEEK_HOLE for block-mapped and extent-mapped file at the same
|
||||
* function. When extent status tree has been fully implemented, it will
|
||||
* track all extent status for a file and we can directly use it to
|
||||
* retrieve the offset for SEEK_DATA/SEEK_HOLE.
|
||||
*/
|
||||
|
||||
/*
|
||||
* When we retrieve the offset for SEEK_DATA/SEEK_HOLE, we would need to
|
||||
* lookup page cache to check whether or not there has some data between
|
||||
* [startoff, endoff] because, if this range contains an unwritten extent,
|
||||
* we determine this extent as a data or a hole according to whether the
|
||||
* page cache has data or not.
|
||||
*/
|
||||
static int ext4_find_unwritten_pgoff(struct inode *inode,
|
||||
int origin,
|
||||
struct ext4_map_blocks *map,
|
||||
loff_t *offset)
|
||||
{
|
||||
struct pagevec pvec;
|
||||
unsigned int blkbits;
|
||||
pgoff_t index;
|
||||
pgoff_t end;
|
||||
loff_t endoff;
|
||||
loff_t startoff;
|
||||
loff_t lastoff;
|
||||
int found = 0;
|
||||
|
||||
blkbits = inode->i_sb->s_blocksize_bits;
|
||||
startoff = *offset;
|
||||
lastoff = startoff;
|
||||
endoff = (map->m_lblk + map->m_len) << blkbits;
|
||||
|
||||
index = startoff >> PAGE_CACHE_SHIFT;
|
||||
end = endoff >> PAGE_CACHE_SHIFT;
|
||||
|
||||
pagevec_init(&pvec, 0);
|
||||
do {
|
||||
int i, num;
|
||||
unsigned long nr_pages;
|
||||
|
||||
num = min_t(pgoff_t, end - index, PAGEVEC_SIZE);
|
||||
nr_pages = pagevec_lookup(&pvec, inode->i_mapping, index,
|
||||
(pgoff_t)num);
|
||||
if (nr_pages == 0) {
|
||||
if (origin == SEEK_DATA)
|
||||
break;
|
||||
|
||||
BUG_ON(origin != SEEK_HOLE);
|
||||
/*
|
||||
* If this is the first time to go into the loop and
|
||||
* offset is not beyond the end offset, it will be a
|
||||
* hole at this offset
|
||||
*/
|
||||
if (lastoff == startoff || lastoff < endoff)
|
||||
found = 1;
|
||||
break;
|
||||
}
|
||||
|
||||
/*
|
||||
* If this is the first time to go into the loop and
|
||||
* offset is smaller than the first page offset, it will be a
|
||||
* hole at this offset.
|
||||
*/
|
||||
if (lastoff == startoff && origin == SEEK_HOLE &&
|
||||
lastoff < page_offset(pvec.pages[0])) {
|
||||
found = 1;
|
||||
break;
|
||||
}
|
||||
|
||||
for (i = 0; i < nr_pages; i++) {
|
||||
struct page *page = pvec.pages[i];
|
||||
struct buffer_head *bh, *head;
|
||||
|
||||
/*
|
||||
* If the current offset is not beyond the end of given
|
||||
* range, it will be a hole.
|
||||
*/
|
||||
if (lastoff < endoff && origin == SEEK_HOLE &&
|
||||
page->index > end) {
|
||||
found = 1;
|
||||
*offset = lastoff;
|
||||
goto out;
|
||||
}
|
||||
|
||||
lock_page(page);
|
||||
|
||||
if (unlikely(page->mapping != inode->i_mapping)) {
|
||||
unlock_page(page);
|
||||
continue;
|
||||
}
|
||||
|
||||
if (!page_has_buffers(page)) {
|
||||
unlock_page(page);
|
||||
continue;
|
||||
}
|
||||
|
||||
if (page_has_buffers(page)) {
|
||||
lastoff = page_offset(page);
|
||||
bh = head = page_buffers(page);
|
||||
do {
|
||||
if (buffer_uptodate(bh) ||
|
||||
buffer_unwritten(bh)) {
|
||||
if (origin == SEEK_DATA)
|
||||
found = 1;
|
||||
} else {
|
||||
if (origin == SEEK_HOLE)
|
||||
found = 1;
|
||||
}
|
||||
if (found) {
|
||||
*offset = max_t(loff_t,
|
||||
startoff, lastoff);
|
||||
unlock_page(page);
|
||||
goto out;
|
||||
}
|
||||
lastoff += bh->b_size;
|
||||
bh = bh->b_this_page;
|
||||
} while (bh != head);
|
||||
}
|
||||
|
||||
lastoff = page_offset(page) + PAGE_SIZE;
|
||||
unlock_page(page);
|
||||
}
|
||||
|
||||
/*
|
||||
* The no. of pages is less than our desired, that would be a
|
||||
* hole in there.
|
||||
*/
|
||||
if (nr_pages < num && origin == SEEK_HOLE) {
|
||||
found = 1;
|
||||
*offset = lastoff;
|
||||
break;
|
||||
}
|
||||
|
||||
index = pvec.pages[i - 1]->index + 1;
|
||||
pagevec_release(&pvec);
|
||||
} while (index <= end);
|
||||
|
||||
out:
|
||||
pagevec_release(&pvec);
|
||||
return found;
|
||||
}
|
||||
|
||||
/*
|
||||
* ext4_seek_data() retrieves the offset for SEEK_DATA.
|
||||
*/
|
||||
static loff_t ext4_seek_data(struct file *file, loff_t offset, loff_t maxsize)
|
||||
{
|
||||
struct inode *inode = file->f_mapping->host;
|
||||
struct ext4_map_blocks map;
|
||||
struct extent_status es;
|
||||
ext4_lblk_t start, last, end;
|
||||
loff_t dataoff, isize;
|
||||
int blkbits;
|
||||
int ret = 0;
|
||||
|
||||
mutex_lock(&inode->i_mutex);
|
||||
|
||||
isize = i_size_read(inode);
|
||||
if (offset >= isize) {
|
||||
mutex_unlock(&inode->i_mutex);
|
||||
return -ENXIO;
|
||||
}
|
||||
|
||||
blkbits = inode->i_sb->s_blocksize_bits;
|
||||
start = offset >> blkbits;
|
||||
last = start;
|
||||
end = isize >> blkbits;
|
||||
dataoff = offset;
|
||||
|
||||
do {
|
||||
map.m_lblk = last;
|
||||
map.m_len = end - last + 1;
|
||||
ret = ext4_map_blocks(NULL, inode, &map, 0);
|
||||
if (ret > 0 && !(map.m_flags & EXT4_MAP_UNWRITTEN)) {
|
||||
if (last != start)
|
||||
dataoff = last << blkbits;
|
||||
break;
|
||||
}
|
||||
|
||||
/*
|
||||
* If there is a delay extent at this offset,
|
||||
* it will be as a data.
|
||||
*/
|
||||
es.start = last;
|
||||
(void)ext4_es_find_extent(inode, &es);
|
||||
if (last >= es.start &&
|
||||
last < es.start + es.len) {
|
||||
if (last != start)
|
||||
dataoff = last << blkbits;
|
||||
break;
|
||||
}
|
||||
|
||||
/*
|
||||
* If there is a unwritten extent at this offset,
|
||||
* it will be as a data or a hole according to page
|
||||
* cache that has data or not.
|
||||
*/
|
||||
if (map.m_flags & EXT4_MAP_UNWRITTEN) {
|
||||
int unwritten;
|
||||
unwritten = ext4_find_unwritten_pgoff(inode, SEEK_DATA,
|
||||
&map, &dataoff);
|
||||
if (unwritten)
|
||||
break;
|
||||
}
|
||||
|
||||
last++;
|
||||
dataoff = last << blkbits;
|
||||
} while (last <= end);
|
||||
|
||||
mutex_unlock(&inode->i_mutex);
|
||||
|
||||
if (dataoff > isize)
|
||||
return -ENXIO;
|
||||
|
||||
if (dataoff < 0 && !(file->f_mode & FMODE_UNSIGNED_OFFSET))
|
||||
return -EINVAL;
|
||||
if (dataoff > maxsize)
|
||||
return -EINVAL;
|
||||
|
||||
if (dataoff != file->f_pos) {
|
||||
file->f_pos = dataoff;
|
||||
file->f_version = 0;
|
||||
}
|
||||
|
||||
return dataoff;
|
||||
}
|
||||
|
||||
/*
|
||||
* ext4_seek_hole() retrieves the offset for SEEK_HOLE.
|
||||
*/
|
||||
static loff_t ext4_seek_hole(struct file *file, loff_t offset, loff_t maxsize)
|
||||
{
|
||||
struct inode *inode = file->f_mapping->host;
|
||||
struct ext4_map_blocks map;
|
||||
struct extent_status es;
|
||||
ext4_lblk_t start, last, end;
|
||||
loff_t holeoff, isize;
|
||||
int blkbits;
|
||||
int ret = 0;
|
||||
|
||||
mutex_lock(&inode->i_mutex);
|
||||
|
||||
isize = i_size_read(inode);
|
||||
if (offset >= isize) {
|
||||
mutex_unlock(&inode->i_mutex);
|
||||
return -ENXIO;
|
||||
}
|
||||
|
||||
blkbits = inode->i_sb->s_blocksize_bits;
|
||||
start = offset >> blkbits;
|
||||
last = start;
|
||||
end = isize >> blkbits;
|
||||
holeoff = offset;
|
||||
|
||||
do {
|
||||
map.m_lblk = last;
|
||||
map.m_len = end - last + 1;
|
||||
ret = ext4_map_blocks(NULL, inode, &map, 0);
|
||||
if (ret > 0 && !(map.m_flags & EXT4_MAP_UNWRITTEN)) {
|
||||
last += ret;
|
||||
holeoff = last << blkbits;
|
||||
continue;
|
||||
}
|
||||
|
||||
/*
|
||||
* If there is a delay extent at this offset,
|
||||
* we will skip this extent.
|
||||
*/
|
||||
es.start = last;
|
||||
(void)ext4_es_find_extent(inode, &es);
|
||||
if (last >= es.start &&
|
||||
last < es.start + es.len) {
|
||||
last = es.start + es.len;
|
||||
holeoff = last << blkbits;
|
||||
continue;
|
||||
}
|
||||
|
||||
/*
|
||||
* If there is a unwritten extent at this offset,
|
||||
* it will be as a data or a hole according to page
|
||||
* cache that has data or not.
|
||||
*/
|
||||
if (map.m_flags & EXT4_MAP_UNWRITTEN) {
|
||||
int unwritten;
|
||||
unwritten = ext4_find_unwritten_pgoff(inode, SEEK_HOLE,
|
||||
&map, &holeoff);
|
||||
if (!unwritten) {
|
||||
last += ret;
|
||||
holeoff = last << blkbits;
|
||||
continue;
|
||||
}
|
||||
}
|
||||
|
||||
/* find a hole */
|
||||
break;
|
||||
} while (last <= end);
|
||||
|
||||
mutex_unlock(&inode->i_mutex);
|
||||
|
||||
if (holeoff > isize)
|
||||
holeoff = isize;
|
||||
|
||||
if (holeoff < 0 && !(file->f_mode & FMODE_UNSIGNED_OFFSET))
|
||||
return -EINVAL;
|
||||
if (holeoff > maxsize)
|
||||
return -EINVAL;
|
||||
|
||||
if (holeoff != file->f_pos) {
|
||||
file->f_pos = holeoff;
|
||||
file->f_version = 0;
|
||||
}
|
||||
|
||||
return holeoff;
|
||||
}
|
||||
|
||||
/*
|
||||
* ext4_llseek() handles both block-mapped and extent-mapped maxbytes values
|
||||
* by calling generic_file_llseek_size() with the appropriate maxbytes
|
||||
|
@ -300,8 +619,19 @@ loff_t ext4_llseek(struct file *file, loff_t offset, int origin)
|
|||
else
|
||||
maxbytes = inode->i_sb->s_maxbytes;
|
||||
|
||||
return generic_file_llseek_size(file, offset, origin,
|
||||
maxbytes, i_size_read(inode));
|
||||
switch (origin) {
|
||||
case SEEK_SET:
|
||||
case SEEK_CUR:
|
||||
case SEEK_END:
|
||||
return generic_file_llseek_size(file, offset, origin,
|
||||
maxbytes, i_size_read(inode));
|
||||
case SEEK_DATA:
|
||||
return ext4_seek_data(file, offset, maxbytes);
|
||||
case SEEK_HOLE:
|
||||
return ext4_seek_hole(file, offset, maxbytes);
|
||||
}
|
||||
|
||||
return -EINVAL;
|
||||
}
|
||||
|
||||
const struct file_operations ext4_file_operations = {
|
||||
|
@ -326,12 +656,10 @@ const struct file_operations ext4_file_operations = {
|
|||
const struct inode_operations ext4_file_inode_operations = {
|
||||
.setattr = ext4_setattr,
|
||||
.getattr = ext4_getattr,
|
||||
#ifdef CONFIG_EXT4_FS_XATTR
|
||||
.setxattr = generic_setxattr,
|
||||
.getxattr = generic_getxattr,
|
||||
.listxattr = ext4_listxattr,
|
||||
.removexattr = generic_removexattr,
|
||||
#endif
|
||||
.get_acl = ext4_get_acl,
|
||||
.fiemap = ext4_fiemap,
|
||||
};
|
||||
|
|
|
@ -44,7 +44,6 @@
|
|||
*/
|
||||
static int ext4_sync_parent(struct inode *inode)
|
||||
{
|
||||
struct writeback_control wbc;
|
||||
struct dentry *dentry = NULL;
|
||||
struct inode *next;
|
||||
int ret = 0;
|
||||
|
@ -66,10 +65,7 @@ static int ext4_sync_parent(struct inode *inode)
|
|||
ret = sync_mapping_buffers(inode->i_mapping);
|
||||
if (ret)
|
||||
break;
|
||||
memset(&wbc, 0, sizeof(wbc));
|
||||
wbc.sync_mode = WB_SYNC_ALL;
|
||||
wbc.nr_to_write = 0; /* only write out the inode */
|
||||
ret = sync_inode(inode, &wbc);
|
||||
ret = sync_inode_metadata(inode, 1);
|
||||
if (ret)
|
||||
break;
|
||||
}
|
||||
|
|
|
@ -762,7 +762,6 @@ got:
|
|||
|
||||
BUFFER_TRACE(block_bitmap_bh, "dirty block bitmap");
|
||||
err = ext4_handle_dirty_metadata(handle, NULL, block_bitmap_bh);
|
||||
brelse(block_bitmap_bh);
|
||||
|
||||
/* recheck and clear flag under lock if we still need to */
|
||||
ext4_lock_group(sb, group);
|
||||
|
@ -775,6 +774,7 @@ got:
|
|||
ext4_group_desc_csum_set(sb, group, gdp);
|
||||
}
|
||||
ext4_unlock_group(sb, group);
|
||||
brelse(block_bitmap_bh);
|
||||
|
||||
if (err)
|
||||
goto fail;
|
||||
|
@ -902,6 +902,10 @@ got:
|
|||
|
||||
ei->i_extra_isize = EXT4_SB(sb)->s_want_extra_isize;
|
||||
|
||||
ei->i_inline_off = 0;
|
||||
if (EXT4_HAS_INCOMPAT_FEATURE(sb, EXT4_FEATURE_INCOMPAT_INLINE_DATA))
|
||||
ext4_set_inode_state(inode, EXT4_STATE_MAY_INLINE_DATA);
|
||||
|
||||
ret = inode;
|
||||
dquot_initialize(inode);
|
||||
err = dquot_alloc_inode(inode);
|
||||
|
|
|
@ -22,6 +22,7 @@
|
|||
|
||||
#include "ext4_jbd2.h"
|
||||
#include "truncate.h"
|
||||
#include "ext4_extents.h" /* Needed for EXT_MAX_BLOCKS */
|
||||
|
||||
#include <trace/events/ext4.h>
|
||||
|
||||
|
@ -755,8 +756,7 @@ cleanup:
|
|||
partial--;
|
||||
}
|
||||
out:
|
||||
trace_ext4_ind_map_blocks_exit(inode, map->m_lblk,
|
||||
map->m_pblk, map->m_len, err);
|
||||
trace_ext4_ind_map_blocks_exit(inode, map, err);
|
||||
return err;
|
||||
}
|
||||
|
||||
|
@ -1412,6 +1412,7 @@ void ext4_ind_truncate(struct inode *inode)
|
|||
down_write(&ei->i_data_sem);
|
||||
|
||||
ext4_discard_preallocations(inode);
|
||||
ext4_es_remove_extent(inode, last_block, EXT_MAX_BLOCKS - last_block);
|
||||
|
||||
/*
|
||||
* The orphan list entry will now protect us from any crash which
|
||||
|
|
1884
fs/ext4/inline.c
Normal file
1884
fs/ext4/inline.c
Normal file
File diff suppressed because it is too large
Load diff
639
fs/ext4/inode.c
639
fs/ext4/inode.c
|
@ -483,49 +483,6 @@ static pgoff_t ext4_num_dirty_pages(struct inode *inode, pgoff_t idx,
|
|||
return num;
|
||||
}
|
||||
|
||||
/*
|
||||
* Sets the BH_Da_Mapped bit on the buffer heads corresponding to the given map.
|
||||
*/
|
||||
static void set_buffers_da_mapped(struct inode *inode,
|
||||
struct ext4_map_blocks *map)
|
||||
{
|
||||
struct address_space *mapping = inode->i_mapping;
|
||||
struct pagevec pvec;
|
||||
int i, nr_pages;
|
||||
pgoff_t index, end;
|
||||
|
||||
index = map->m_lblk >> (PAGE_CACHE_SHIFT - inode->i_blkbits);
|
||||
end = (map->m_lblk + map->m_len - 1) >>
|
||||
(PAGE_CACHE_SHIFT - inode->i_blkbits);
|
||||
|
||||
pagevec_init(&pvec, 0);
|
||||
while (index <= end) {
|
||||
nr_pages = pagevec_lookup(&pvec, mapping, index,
|
||||
min(end - index + 1,
|
||||
(pgoff_t)PAGEVEC_SIZE));
|
||||
if (nr_pages == 0)
|
||||
break;
|
||||
for (i = 0; i < nr_pages; i++) {
|
||||
struct page *page = pvec.pages[i];
|
||||
struct buffer_head *bh, *head;
|
||||
|
||||
if (unlikely(page->mapping != mapping) ||
|
||||
!PageDirty(page))
|
||||
break;
|
||||
|
||||
if (page_has_buffers(page)) {
|
||||
bh = head = page_buffers(page);
|
||||
do {
|
||||
set_buffer_da_mapped(bh);
|
||||
bh = bh->b_this_page;
|
||||
} while (bh != head);
|
||||
}
|
||||
index++;
|
||||
}
|
||||
pagevec_release(&pvec);
|
||||
}
|
||||
}
|
||||
|
||||
/*
|
||||
* The ext4_map_blocks() function tries to look up the requested blocks,
|
||||
* and returns if the blocks are already mapped.
|
||||
|
@ -574,7 +531,16 @@ int ext4_map_blocks(handle_t *handle, struct inode *inode,
|
|||
up_read((&EXT4_I(inode)->i_data_sem));
|
||||
|
||||
if (retval > 0 && map->m_flags & EXT4_MAP_MAPPED) {
|
||||
int ret = check_block_validity(inode, map);
|
||||
int ret;
|
||||
if (flags & EXT4_GET_BLOCKS_DELALLOC_RESERVE) {
|
||||
/* delayed alloc may be allocated by fallocate and
|
||||
* coverted to initialized by directIO.
|
||||
* we need to handle delayed extent here.
|
||||
*/
|
||||
down_write((&EXT4_I(inode)->i_data_sem));
|
||||
goto delayed_mapped;
|
||||
}
|
||||
ret = check_block_validity(inode, map);
|
||||
if (ret != 0)
|
||||
return ret;
|
||||
}
|
||||
|
@ -652,12 +618,15 @@ int ext4_map_blocks(handle_t *handle, struct inode *inode,
|
|||
if (flags & EXT4_GET_BLOCKS_DELALLOC_RESERVE) {
|
||||
ext4_clear_inode_state(inode, EXT4_STATE_DELALLOC_RESERVED);
|
||||
|
||||
/* If we have successfully mapped the delayed allocated blocks,
|
||||
* set the BH_Da_Mapped bit on them. Its important to do this
|
||||
* under the protection of i_data_sem.
|
||||
*/
|
||||
if (retval > 0 && map->m_flags & EXT4_MAP_MAPPED)
|
||||
set_buffers_da_mapped(inode, map);
|
||||
if (retval > 0 && map->m_flags & EXT4_MAP_MAPPED) {
|
||||
int ret;
|
||||
delayed_mapped:
|
||||
/* delayed allocation blocks has been allocated */
|
||||
ret = ext4_es_remove_extent(inode, map->m_lblk,
|
||||
map->m_len);
|
||||
if (ret < 0)
|
||||
retval = ret;
|
||||
}
|
||||
}
|
||||
|
||||
up_write((&EXT4_I(inode)->i_data_sem));
|
||||
|
@ -680,10 +649,13 @@ static int _ext4_get_block(struct inode *inode, sector_t iblock,
|
|||
int ret = 0, started = 0;
|
||||
int dio_credits;
|
||||
|
||||
if (ext4_has_inline_data(inode))
|
||||
return -ERANGE;
|
||||
|
||||
map.m_lblk = iblock;
|
||||
map.m_len = bh->b_size >> inode->i_blkbits;
|
||||
|
||||
if (flags && !handle) {
|
||||
if (flags && !(flags & EXT4_GET_BLOCKS_NO_LOCK) && !handle) {
|
||||
/* Direct IO write... */
|
||||
if (map.m_len > DIO_MAX_BLOCKS)
|
||||
map.m_len = DIO_MAX_BLOCKS;
|
||||
|
@ -798,13 +770,13 @@ struct buffer_head *ext4_bread(handle_t *handle, struct inode *inode,
|
|||
return NULL;
|
||||
}
|
||||
|
||||
static int walk_page_buffers(handle_t *handle,
|
||||
struct buffer_head *head,
|
||||
unsigned from,
|
||||
unsigned to,
|
||||
int *partial,
|
||||
int (*fn)(handle_t *handle,
|
||||
struct buffer_head *bh))
|
||||
int ext4_walk_page_buffers(handle_t *handle,
|
||||
struct buffer_head *head,
|
||||
unsigned from,
|
||||
unsigned to,
|
||||
int *partial,
|
||||
int (*fn)(handle_t *handle,
|
||||
struct buffer_head *bh))
|
||||
{
|
||||
struct buffer_head *bh;
|
||||
unsigned block_start, block_end;
|
||||
|
@ -854,8 +826,8 @@ static int walk_page_buffers(handle_t *handle,
|
|||
* is elevated. We'll still have enough credits for the tiny quotafile
|
||||
* write.
|
||||
*/
|
||||
static int do_journal_get_write_access(handle_t *handle,
|
||||
struct buffer_head *bh)
|
||||
int do_journal_get_write_access(handle_t *handle,
|
||||
struct buffer_head *bh)
|
||||
{
|
||||
int dirty = buffer_dirty(bh);
|
||||
int ret;
|
||||
|
@ -878,7 +850,7 @@ static int do_journal_get_write_access(handle_t *handle,
|
|||
return ret;
|
||||
}
|
||||
|
||||
static int ext4_get_block_write(struct inode *inode, sector_t iblock,
|
||||
static int ext4_get_block_write_nolock(struct inode *inode, sector_t iblock,
|
||||
struct buffer_head *bh_result, int create);
|
||||
static int ext4_write_begin(struct file *file, struct address_space *mapping,
|
||||
loff_t pos, unsigned len, unsigned flags,
|
||||
|
@ -902,6 +874,17 @@ static int ext4_write_begin(struct file *file, struct address_space *mapping,
|
|||
from = pos & (PAGE_CACHE_SIZE - 1);
|
||||
to = from + len;
|
||||
|
||||
if (ext4_test_inode_state(inode, EXT4_STATE_MAY_INLINE_DATA)) {
|
||||
ret = ext4_try_to_write_inline_data(mapping, inode, pos, len,
|
||||
flags, pagep);
|
||||
if (ret < 0)
|
||||
goto out;
|
||||
if (ret == 1) {
|
||||
ret = 0;
|
||||
goto out;
|
||||
}
|
||||
}
|
||||
|
||||
retry:
|
||||
handle = ext4_journal_start(inode, needed_blocks);
|
||||
if (IS_ERR(handle)) {
|
||||
|
@ -919,6 +902,7 @@ retry:
|
|||
ret = -ENOMEM;
|
||||
goto out;
|
||||
}
|
||||
|
||||
*pagep = page;
|
||||
|
||||
if (ext4_should_dioread_nolock(inode))
|
||||
|
@ -927,8 +911,9 @@ retry:
|
|||
ret = __block_write_begin(page, pos, len, ext4_get_block);
|
||||
|
||||
if (!ret && ext4_should_journal_data(inode)) {
|
||||
ret = walk_page_buffers(handle, page_buffers(page),
|
||||
from, to, NULL, do_journal_get_write_access);
|
||||
ret = ext4_walk_page_buffers(handle, page_buffers(page),
|
||||
from, to, NULL,
|
||||
do_journal_get_write_access);
|
||||
}
|
||||
|
||||
if (ret) {
|
||||
|
@ -983,7 +968,12 @@ static int ext4_generic_write_end(struct file *file,
|
|||
struct inode *inode = mapping->host;
|
||||
handle_t *handle = ext4_journal_current_handle();
|
||||
|
||||
copied = block_write_end(file, mapping, pos, len, copied, page, fsdata);
|
||||
if (ext4_has_inline_data(inode))
|
||||
copied = ext4_write_inline_data_end(inode, pos, len,
|
||||
copied, page);
|
||||
else
|
||||
copied = block_write_end(file, mapping, pos,
|
||||
len, copied, page, fsdata);
|
||||
|
||||
/*
|
||||
* No need to use i_size_read() here, the i_size
|
||||
|
@ -1134,16 +1124,21 @@ static int ext4_journalled_write_end(struct file *file,
|
|||
|
||||
BUG_ON(!ext4_handle_valid(handle));
|
||||
|
||||
if (copied < len) {
|
||||
if (!PageUptodate(page))
|
||||
copied = 0;
|
||||
page_zero_new_buffers(page, from+copied, to);
|
||||
}
|
||||
if (ext4_has_inline_data(inode))
|
||||
copied = ext4_write_inline_data_end(inode, pos, len,
|
||||
copied, page);
|
||||
else {
|
||||
if (copied < len) {
|
||||
if (!PageUptodate(page))
|
||||
copied = 0;
|
||||
page_zero_new_buffers(page, from+copied, to);
|
||||
}
|
||||
|
||||
ret = walk_page_buffers(handle, page_buffers(page), from,
|
||||
to, &partial, write_end_fn);
|
||||
if (!partial)
|
||||
SetPageUptodate(page);
|
||||
ret = ext4_walk_page_buffers(handle, page_buffers(page), from,
|
||||
to, &partial, write_end_fn);
|
||||
if (!partial)
|
||||
SetPageUptodate(page);
|
||||
}
|
||||
new_i_size = pos + copied;
|
||||
if (new_i_size > inode->i_size)
|
||||
i_size_write(inode, pos+copied);
|
||||
|
@ -1301,6 +1296,7 @@ static void ext4_da_page_release_reservation(struct page *page,
|
|||
struct inode *inode = page->mapping->host;
|
||||
struct ext4_sb_info *sbi = EXT4_SB(inode->i_sb);
|
||||
int num_clusters;
|
||||
ext4_fsblk_t lblk;
|
||||
|
||||
head = page_buffers(page);
|
||||
bh = head;
|
||||
|
@ -1310,20 +1306,23 @@ static void ext4_da_page_release_reservation(struct page *page,
|
|||
if ((offset <= curr_off) && (buffer_delay(bh))) {
|
||||
to_release++;
|
||||
clear_buffer_delay(bh);
|
||||
clear_buffer_da_mapped(bh);
|
||||
}
|
||||
curr_off = next_off;
|
||||
} while ((bh = bh->b_this_page) != head);
|
||||
|
||||
if (to_release) {
|
||||
lblk = page->index << (PAGE_CACHE_SHIFT - inode->i_blkbits);
|
||||
ext4_es_remove_extent(inode, lblk, to_release);
|
||||
}
|
||||
|
||||
/* If we have released all the blocks belonging to a cluster, then we
|
||||
* need to release the reserved space for that cluster. */
|
||||
num_clusters = EXT4_NUM_B2C(sbi, to_release);
|
||||
while (num_clusters > 0) {
|
||||
ext4_fsblk_t lblk;
|
||||
lblk = (page->index << (PAGE_CACHE_SHIFT - inode->i_blkbits)) +
|
||||
((num_clusters - 1) << sbi->s_cluster_bits);
|
||||
if (sbi->s_cluster_ratio == 1 ||
|
||||
!ext4_find_delalloc_cluster(inode, lblk, 1))
|
||||
!ext4_find_delalloc_cluster(inode, lblk))
|
||||
ext4_da_release_space(inode, 1);
|
||||
|
||||
num_clusters--;
|
||||
|
@ -1429,8 +1428,6 @@ static int mpage_da_submit_io(struct mpage_da_data *mpd,
|
|||
clear_buffer_delay(bh);
|
||||
bh->b_blocknr = pblock;
|
||||
}
|
||||
if (buffer_da_mapped(bh))
|
||||
clear_buffer_da_mapped(bh);
|
||||
if (buffer_unwritten(bh) ||
|
||||
buffer_mapped(bh))
|
||||
BUG_ON(bh->b_blocknr != pblock);
|
||||
|
@ -1500,9 +1497,16 @@ static void ext4_da_block_invalidatepages(struct mpage_da_data *mpd)
|
|||
struct pagevec pvec;
|
||||
struct inode *inode = mpd->inode;
|
||||
struct address_space *mapping = inode->i_mapping;
|
||||
ext4_lblk_t start, last;
|
||||
|
||||
index = mpd->first_page;
|
||||
end = mpd->next_page - 1;
|
||||
|
||||
start = index << (PAGE_CACHE_SHIFT - inode->i_blkbits);
|
||||
last = end << (PAGE_CACHE_SHIFT - inode->i_blkbits);
|
||||
ext4_es_remove_extent(inode, start, last - start + 1);
|
||||
|
||||
pagevec_init(&pvec, 0);
|
||||
while (index <= end) {
|
||||
nr_pages = pagevec_lookup(&pvec, mapping, index, PAGEVEC_SIZE);
|
||||
if (nr_pages == 0)
|
||||
|
@ -1656,15 +1660,6 @@ static void mpage_da_map_and_submit(struct mpage_da_data *mpd)
|
|||
|
||||
for (i = 0; i < map.m_len; i++)
|
||||
unmap_underlying_metadata(bdev, map.m_pblk + i);
|
||||
|
||||
if (ext4_should_order_data(mpd->inode)) {
|
||||
err = ext4_jbd2_file_inode(handle, mpd->inode);
|
||||
if (err) {
|
||||
/* Only if the journal is aborted */
|
||||
mpd->retval = err;
|
||||
goto submit_io;
|
||||
}
|
||||
}
|
||||
}
|
||||
|
||||
/*
|
||||
|
@ -1795,7 +1790,19 @@ static int ext4_da_map_blocks(struct inode *inode, sector_t iblock,
|
|||
* file system block.
|
||||
*/
|
||||
down_read((&EXT4_I(inode)->i_data_sem));
|
||||
if (ext4_test_inode_flag(inode, EXT4_INODE_EXTENTS))
|
||||
if (ext4_has_inline_data(inode)) {
|
||||
/*
|
||||
* We will soon create blocks for this page, and let
|
||||
* us pretend as if the blocks aren't allocated yet.
|
||||
* In case of clusters, we have to handle the work
|
||||
* of mapping from cluster so that the reserved space
|
||||
* is calculated properly.
|
||||
*/
|
||||
if ((EXT4_SB(inode->i_sb)->s_cluster_ratio > 1) &&
|
||||
ext4_find_delalloc_cluster(inode, map->m_lblk))
|
||||
map->m_flags |= EXT4_MAP_FROM_CLUSTER;
|
||||
retval = 0;
|
||||
} else if (ext4_test_inode_flag(inode, EXT4_INODE_EXTENTS))
|
||||
retval = ext4_ext_map_blocks(NULL, inode, map, 0);
|
||||
else
|
||||
retval = ext4_ind_map_blocks(NULL, inode, map, 0);
|
||||
|
@ -1814,6 +1821,10 @@ static int ext4_da_map_blocks(struct inode *inode, sector_t iblock,
|
|||
goto out_unlock;
|
||||
}
|
||||
|
||||
retval = ext4_es_insert_extent(inode, map->m_lblk, map->m_len);
|
||||
if (retval)
|
||||
goto out_unlock;
|
||||
|
||||
/* Clear EXT4_MAP_FROM_CLUSTER flag since its purpose is served
|
||||
* and it should not appear on the bh->b_state.
|
||||
*/
|
||||
|
@ -1842,8 +1853,8 @@ out_unlock:
|
|||
* We also have b_blocknr = physicalblock mapping unwritten extent and b_bdev
|
||||
* initialized properly.
|
||||
*/
|
||||
static int ext4_da_get_block_prep(struct inode *inode, sector_t iblock,
|
||||
struct buffer_head *bh, int create)
|
||||
int ext4_da_get_block_prep(struct inode *inode, sector_t iblock,
|
||||
struct buffer_head *bh, int create)
|
||||
{
|
||||
struct ext4_map_blocks map;
|
||||
int ret = 0;
|
||||
|
@ -1917,15 +1928,29 @@ static int __ext4_journalled_writepage(struct page *page,
|
|||
{
|
||||
struct address_space *mapping = page->mapping;
|
||||
struct inode *inode = mapping->host;
|
||||
struct buffer_head *page_bufs;
|
||||
struct buffer_head *page_bufs = NULL;
|
||||
handle_t *handle = NULL;
|
||||
int ret = 0;
|
||||
int err;
|
||||
int ret = 0, err = 0;
|
||||
int inline_data = ext4_has_inline_data(inode);
|
||||
struct buffer_head *inode_bh = NULL;
|
||||
|
||||
ClearPageChecked(page);
|
||||
page_bufs = page_buffers(page);
|
||||
BUG_ON(!page_bufs);
|
||||
walk_page_buffers(handle, page_bufs, 0, len, NULL, bget_one);
|
||||
|
||||
if (inline_data) {
|
||||
BUG_ON(page->index != 0);
|
||||
BUG_ON(len > ext4_get_max_inline_size(inode));
|
||||
inode_bh = ext4_journalled_write_inline_data(inode, len, page);
|
||||
if (inode_bh == NULL)
|
||||
goto out;
|
||||
} else {
|
||||
page_bufs = page_buffers(page);
|
||||
if (!page_bufs) {
|
||||
BUG();
|
||||
goto out;
|
||||
}
|
||||
ext4_walk_page_buffers(handle, page_bufs, 0, len,
|
||||
NULL, bget_one);
|
||||
}
|
||||
/* As soon as we unlock the page, it can go away, but we have
|
||||
* references to buffers so we are safe */
|
||||
unlock_page(page);
|
||||
|
@ -1938,11 +1963,18 @@ static int __ext4_journalled_writepage(struct page *page,
|
|||
|
||||
BUG_ON(!ext4_handle_valid(handle));
|
||||
|
||||
ret = walk_page_buffers(handle, page_bufs, 0, len, NULL,
|
||||
do_journal_get_write_access);
|
||||
if (inline_data) {
|
||||
ret = ext4_journal_get_write_access(handle, inode_bh);
|
||||
|
||||
err = walk_page_buffers(handle, page_bufs, 0, len, NULL,
|
||||
write_end_fn);
|
||||
err = ext4_handle_dirty_metadata(handle, inode, inode_bh);
|
||||
|
||||
} else {
|
||||
ret = ext4_walk_page_buffers(handle, page_bufs, 0, len, NULL,
|
||||
do_journal_get_write_access);
|
||||
|
||||
err = ext4_walk_page_buffers(handle, page_bufs, 0, len, NULL,
|
||||
write_end_fn);
|
||||
}
|
||||
if (ret == 0)
|
||||
ret = err;
|
||||
EXT4_I(inode)->i_datasync_tid = handle->h_transaction->t_tid;
|
||||
|
@ -1950,9 +1982,12 @@ static int __ext4_journalled_writepage(struct page *page,
|
|||
if (!ret)
|
||||
ret = err;
|
||||
|
||||
walk_page_buffers(handle, page_bufs, 0, len, NULL, bput_one);
|
||||
if (!ext4_has_inline_data(inode))
|
||||
ext4_walk_page_buffers(handle, page_bufs, 0, len,
|
||||
NULL, bput_one);
|
||||
ext4_set_inode_state(inode, EXT4_STATE_JDATA);
|
||||
out:
|
||||
brelse(inode_bh);
|
||||
return ret;
|
||||
}
|
||||
|
||||
|
@ -2029,8 +2064,8 @@ static int ext4_writepage(struct page *page,
|
|||
commit_write = 1;
|
||||
}
|
||||
page_bufs = page_buffers(page);
|
||||
if (walk_page_buffers(NULL, page_bufs, 0, len, NULL,
|
||||
ext4_bh_delay_or_unwritten)) {
|
||||
if (ext4_walk_page_buffers(NULL, page_bufs, 0, len, NULL,
|
||||
ext4_bh_delay_or_unwritten)) {
|
||||
/*
|
||||
* We don't want to do block allocation, so redirty
|
||||
* the page and return. We may reach here when we do
|
||||
|
@ -2096,7 +2131,8 @@ static int ext4_da_writepages_trans_blocks(struct inode *inode)
|
|||
* mpage_da_map_and_submit to map a single contiguous memory region
|
||||
* and then write them.
|
||||
*/
|
||||
static int write_cache_pages_da(struct address_space *mapping,
|
||||
static int write_cache_pages_da(handle_t *handle,
|
||||
struct address_space *mapping,
|
||||
struct writeback_control *wbc,
|
||||
struct mpage_da_data *mpd,
|
||||
pgoff_t *done_index)
|
||||
|
@ -2175,6 +2211,17 @@ static int write_cache_pages_da(struct address_space *mapping,
|
|||
wait_on_page_writeback(page);
|
||||
BUG_ON(PageWriteback(page));
|
||||
|
||||
/*
|
||||
* If we have inline data and arrive here, it means that
|
||||
* we will soon create the block for the 1st page, so
|
||||
* we'd better clear the inline data here.
|
||||
*/
|
||||
if (ext4_has_inline_data(inode)) {
|
||||
BUG_ON(ext4_test_inode_state(inode,
|
||||
EXT4_STATE_MAY_INLINE_DATA));
|
||||
ext4_destroy_inline_data(handle, inode);
|
||||
}
|
||||
|
||||
if (mpd->next_page != page->index)
|
||||
mpd->first_page = page->index;
|
||||
mpd->next_page = page->index + 1;
|
||||
|
@ -2381,7 +2428,8 @@ retry:
|
|||
* contiguous region of logical blocks that need
|
||||
* blocks to be allocated by ext4 and submit them.
|
||||
*/
|
||||
ret = write_cache_pages_da(mapping, wbc, &mpd, &done_index);
|
||||
ret = write_cache_pages_da(handle, mapping,
|
||||
wbc, &mpd, &done_index);
|
||||
/*
|
||||
* If we have a contiguous extent of pages and we
|
||||
* haven't done the I/O yet, map the blocks and submit
|
||||
|
@ -2445,7 +2493,6 @@ out_writepages:
|
|||
return ret;
|
||||
}
|
||||
|
||||
#define FALL_BACK_TO_NONDELALLOC 1
|
||||
static int ext4_nonda_switch(struct super_block *sb)
|
||||
{
|
||||
s64 free_blocks, dirty_blocks;
|
||||
|
@ -2502,6 +2549,19 @@ static int ext4_da_write_begin(struct file *file, struct address_space *mapping,
|
|||
}
|
||||
*fsdata = (void *)0;
|
||||
trace_ext4_da_write_begin(inode, pos, len, flags);
|
||||
|
||||
if (ext4_test_inode_state(inode, EXT4_STATE_MAY_INLINE_DATA)) {
|
||||
ret = ext4_da_write_inline_data_begin(mapping, inode,
|
||||
pos, len, flags,
|
||||
pagep, fsdata);
|
||||
if (ret < 0)
|
||||
goto out;
|
||||
if (ret == 1) {
|
||||
ret = 0;
|
||||
goto out;
|
||||
}
|
||||
}
|
||||
|
||||
retry:
|
||||
/*
|
||||
* With delayed allocation, we don't log the i_disksize update
|
||||
|
@ -2603,22 +2663,13 @@ static int ext4_da_write_end(struct file *file,
|
|||
* changes. So let's piggyback the i_disksize mark_inode_dirty
|
||||
* into that.
|
||||
*/
|
||||
|
||||
new_i_size = pos + copied;
|
||||
if (copied && new_i_size > EXT4_I(inode)->i_disksize) {
|
||||
if (ext4_da_should_update_i_disksize(page, end)) {
|
||||
if (ext4_has_inline_data(inode) ||
|
||||
ext4_da_should_update_i_disksize(page, end)) {
|
||||
down_write(&EXT4_I(inode)->i_data_sem);
|
||||
if (new_i_size > EXT4_I(inode)->i_disksize) {
|
||||
/*
|
||||
* Updating i_disksize when extending file
|
||||
* without needing block allocation
|
||||
*/
|
||||
if (ext4_should_order_data(inode))
|
||||
ret = ext4_jbd2_file_inode(handle,
|
||||
inode);
|
||||
|
||||
if (new_i_size > EXT4_I(inode)->i_disksize)
|
||||
EXT4_I(inode)->i_disksize = new_i_size;
|
||||
}
|
||||
up_write(&EXT4_I(inode)->i_data_sem);
|
||||
/* We need to mark inode dirty even if
|
||||
* new_i_size is less that inode->i_size
|
||||
|
@ -2627,8 +2678,16 @@ static int ext4_da_write_end(struct file *file,
|
|||
ext4_mark_inode_dirty(handle, inode);
|
||||
}
|
||||
}
|
||||
ret2 = generic_write_end(file, mapping, pos, len, copied,
|
||||
|
||||
if (write_mode != CONVERT_INLINE_DATA &&
|
||||
ext4_test_inode_state(inode, EXT4_STATE_MAY_INLINE_DATA) &&
|
||||
ext4_has_inline_data(inode))
|
||||
ret2 = ext4_da_write_inline_data_end(inode, pos, len, copied,
|
||||
page);
|
||||
else
|
||||
ret2 = generic_write_end(file, mapping, pos, len, copied,
|
||||
page, fsdata);
|
||||
|
||||
copied = ret2;
|
||||
if (ret2 < 0)
|
||||
ret = ret2;
|
||||
|
@ -2721,6 +2780,12 @@ static sector_t ext4_bmap(struct address_space *mapping, sector_t block)
|
|||
journal_t *journal;
|
||||
int err;
|
||||
|
||||
/*
|
||||
* We can get here for an inline file via the FIBMAP ioctl
|
||||
*/
|
||||
if (ext4_has_inline_data(inode))
|
||||
return 0;
|
||||
|
||||
if (mapping_tagged(mapping, PAGECACHE_TAG_DIRTY) &&
|
||||
test_opt(inode->i_sb, DELALLOC)) {
|
||||
/*
|
||||
|
@ -2766,14 +2831,30 @@ static sector_t ext4_bmap(struct address_space *mapping, sector_t block)
|
|||
|
||||
static int ext4_readpage(struct file *file, struct page *page)
|
||||
{
|
||||
int ret = -EAGAIN;
|
||||
struct inode *inode = page->mapping->host;
|
||||
|
||||
trace_ext4_readpage(page);
|
||||
return mpage_readpage(page, ext4_get_block);
|
||||
|
||||
if (ext4_has_inline_data(inode))
|
||||
ret = ext4_readpage_inline(inode, page);
|
||||
|
||||
if (ret == -EAGAIN)
|
||||
return mpage_readpage(page, ext4_get_block);
|
||||
|
||||
return ret;
|
||||
}
|
||||
|
||||
static int
|
||||
ext4_readpages(struct file *file, struct address_space *mapping,
|
||||
struct list_head *pages, unsigned nr_pages)
|
||||
{
|
||||
struct inode *inode = mapping->host;
|
||||
|
||||
/* If the file has inline data, no need to do readpages. */
|
||||
if (ext4_has_inline_data(inode))
|
||||
return 0;
|
||||
|
||||
return mpage_readpages(mapping, pages, nr_pages, ext4_get_block);
|
||||
}
|
||||
|
||||
|
@ -2840,7 +2921,7 @@ static int ext4_releasepage(struct page *page, gfp_t wait)
|
|||
* We allocate an uinitialized extent if blocks haven't been allocated.
|
||||
* The extent will be converted to initialized after the IO is complete.
|
||||
*/
|
||||
static int ext4_get_block_write(struct inode *inode, sector_t iblock,
|
||||
int ext4_get_block_write(struct inode *inode, sector_t iblock,
|
||||
struct buffer_head *bh_result, int create)
|
||||
{
|
||||
ext4_debug("ext4_get_block_write: inode %lu, create flag %d\n",
|
||||
|
@ -2850,29 +2931,12 @@ static int ext4_get_block_write(struct inode *inode, sector_t iblock,
|
|||
}
|
||||
|
||||
static int ext4_get_block_write_nolock(struct inode *inode, sector_t iblock,
|
||||
struct buffer_head *bh_result, int flags)
|
||||
struct buffer_head *bh_result, int create)
|
||||
{
|
||||
handle_t *handle = ext4_journal_current_handle();
|
||||
struct ext4_map_blocks map;
|
||||
int ret = 0;
|
||||
|
||||
ext4_debug("ext4_get_block_write_nolock: inode %lu, flag %d\n",
|
||||
inode->i_ino, flags);
|
||||
|
||||
flags = EXT4_GET_BLOCKS_NO_LOCK;
|
||||
|
||||
map.m_lblk = iblock;
|
||||
map.m_len = bh_result->b_size >> inode->i_blkbits;
|
||||
|
||||
ret = ext4_map_blocks(handle, inode, &map, flags);
|
||||
if (ret > 0) {
|
||||
map_bh(bh_result, inode->i_sb, map.m_pblk);
|
||||
bh_result->b_state = (bh_result->b_state & ~EXT4_MAP_FLAGS) |
|
||||
map.m_flags;
|
||||
bh_result->b_size = inode->i_sb->s_blocksize * map.m_len;
|
||||
ret = 0;
|
||||
}
|
||||
return ret;
|
||||
ext4_debug("ext4_get_block_write_nolock: inode %lu, create flag %d\n",
|
||||
inode->i_ino, create);
|
||||
return _ext4_get_block(inode, iblock, bh_result,
|
||||
EXT4_GET_BLOCKS_NO_LOCK);
|
||||
}
|
||||
|
||||
static void ext4_end_io_dio(struct kiocb *iocb, loff_t offset,
|
||||
|
@ -2978,10 +3042,10 @@ retry:
|
|||
* fall back to buffered IO.
|
||||
*
|
||||
* For holes, we fallocate those blocks, mark them as uninitialized
|
||||
* If those blocks were preallocated, we mark sure they are splited, but
|
||||
* If those blocks were preallocated, we mark sure they are split, but
|
||||
* still keep the range to write as uninitialized.
|
||||
*
|
||||
* The unwrritten extents will be converted to written when DIO is completed.
|
||||
* The unwritten extents will be converted to written when DIO is completed.
|
||||
* For async direct IO, since the IO may still pending when return, we
|
||||
* set up an end_io call back function, which will do the conversion
|
||||
* when async direct IO completed.
|
||||
|
@ -2999,125 +3063,120 @@ static ssize_t ext4_ext_direct_IO(int rw, struct kiocb *iocb,
|
|||
struct inode *inode = file->f_mapping->host;
|
||||
ssize_t ret;
|
||||
size_t count = iov_length(iov, nr_segs);
|
||||
|
||||
int overwrite = 0;
|
||||
get_block_t *get_block_func = NULL;
|
||||
int dio_flags = 0;
|
||||
loff_t final_size = offset + count;
|
||||
if (rw == WRITE && final_size <= inode->i_size) {
|
||||
int overwrite = 0;
|
||||
|
||||
BUG_ON(iocb->private == NULL);
|
||||
/* Use the old path for reads and writes beyond i_size. */
|
||||
if (rw != WRITE || final_size > inode->i_size)
|
||||
return ext4_ind_direct_IO(rw, iocb, iov, offset, nr_segs);
|
||||
|
||||
/* If we do a overwrite dio, i_mutex locking can be released */
|
||||
overwrite = *((int *)iocb->private);
|
||||
BUG_ON(iocb->private == NULL);
|
||||
|
||||
if (overwrite) {
|
||||
atomic_inc(&inode->i_dio_count);
|
||||
down_read(&EXT4_I(inode)->i_data_sem);
|
||||
mutex_unlock(&inode->i_mutex);
|
||||
}
|
||||
/* If we do a overwrite dio, i_mutex locking can be released */
|
||||
overwrite = *((int *)iocb->private);
|
||||
|
||||
/*
|
||||
* We could direct write to holes and fallocate.
|
||||
*
|
||||
* Allocated blocks to fill the hole are marked as uninitialized
|
||||
* to prevent parallel buffered read to expose the stale data
|
||||
* before DIO complete the data IO.
|
||||
*
|
||||
* As to previously fallocated extents, ext4 get_block
|
||||
* will just simply mark the buffer mapped but still
|
||||
* keep the extents uninitialized.
|
||||
*
|
||||
* for non AIO case, we will convert those unwritten extents
|
||||
* to written after return back from blockdev_direct_IO.
|
||||
*
|
||||
* for async DIO, the conversion needs to be defered when
|
||||
* the IO is completed. The ext4 end_io callback function
|
||||
* will be called to take care of the conversion work.
|
||||
* Here for async case, we allocate an io_end structure to
|
||||
* hook to the iocb.
|
||||
*/
|
||||
iocb->private = NULL;
|
||||
ext4_inode_aio_set(inode, NULL);
|
||||
if (!is_sync_kiocb(iocb)) {
|
||||
ext4_io_end_t *io_end =
|
||||
ext4_init_io_end(inode, GFP_NOFS);
|
||||
if (!io_end) {
|
||||
ret = -ENOMEM;
|
||||
goto retake_lock;
|
||||
}
|
||||
io_end->flag |= EXT4_IO_END_DIRECT;
|
||||
iocb->private = io_end;
|
||||
/*
|
||||
* we save the io structure for current async
|
||||
* direct IO, so that later ext4_map_blocks()
|
||||
* could flag the io structure whether there
|
||||
* is a unwritten extents needs to be converted
|
||||
* when IO is completed.
|
||||
*/
|
||||
ext4_inode_aio_set(inode, io_end);
|
||||
}
|
||||
|
||||
if (overwrite)
|
||||
ret = __blockdev_direct_IO(rw, iocb, inode,
|
||||
inode->i_sb->s_bdev, iov,
|
||||
offset, nr_segs,
|
||||
ext4_get_block_write_nolock,
|
||||
ext4_end_io_dio,
|
||||
NULL,
|
||||
0);
|
||||
else
|
||||
ret = __blockdev_direct_IO(rw, iocb, inode,
|
||||
inode->i_sb->s_bdev, iov,
|
||||
offset, nr_segs,
|
||||
ext4_get_block_write,
|
||||
ext4_end_io_dio,
|
||||
NULL,
|
||||
DIO_LOCKING);
|
||||
if (iocb->private)
|
||||
ext4_inode_aio_set(inode, NULL);
|
||||
/*
|
||||
* The io_end structure takes a reference to the inode,
|
||||
* that structure needs to be destroyed and the
|
||||
* reference to the inode need to be dropped, when IO is
|
||||
* complete, even with 0 byte write, or failed.
|
||||
*
|
||||
* In the successful AIO DIO case, the io_end structure will be
|
||||
* desctroyed and the reference to the inode will be dropped
|
||||
* after the end_io call back function is called.
|
||||
*
|
||||
* In the case there is 0 byte write, or error case, since
|
||||
* VFS direct IO won't invoke the end_io call back function,
|
||||
* we need to free the end_io structure here.
|
||||
*/
|
||||
if (ret != -EIOCBQUEUED && ret <= 0 && iocb->private) {
|
||||
ext4_free_io_end(iocb->private);
|
||||
iocb->private = NULL;
|
||||
} else if (ret > 0 && !overwrite && ext4_test_inode_state(inode,
|
||||
EXT4_STATE_DIO_UNWRITTEN)) {
|
||||
int err;
|
||||
/*
|
||||
* for non AIO case, since the IO is already
|
||||
* completed, we could do the conversion right here
|
||||
*/
|
||||
err = ext4_convert_unwritten_extents(inode,
|
||||
offset, ret);
|
||||
if (err < 0)
|
||||
ret = err;
|
||||
ext4_clear_inode_state(inode, EXT4_STATE_DIO_UNWRITTEN);
|
||||
}
|
||||
|
||||
retake_lock:
|
||||
/* take i_mutex locking again if we do a ovewrite dio */
|
||||
if (overwrite) {
|
||||
inode_dio_done(inode);
|
||||
up_read(&EXT4_I(inode)->i_data_sem);
|
||||
mutex_lock(&inode->i_mutex);
|
||||
}
|
||||
|
||||
return ret;
|
||||
if (overwrite) {
|
||||
atomic_inc(&inode->i_dio_count);
|
||||
down_read(&EXT4_I(inode)->i_data_sem);
|
||||
mutex_unlock(&inode->i_mutex);
|
||||
}
|
||||
|
||||
/* for write the the end of file case, we fall back to old way */
|
||||
return ext4_ind_direct_IO(rw, iocb, iov, offset, nr_segs);
|
||||
/*
|
||||
* We could direct write to holes and fallocate.
|
||||
*
|
||||
* Allocated blocks to fill the hole are marked as
|
||||
* uninitialized to prevent parallel buffered read to expose
|
||||
* the stale data before DIO complete the data IO.
|
||||
*
|
||||
* As to previously fallocated extents, ext4 get_block will
|
||||
* just simply mark the buffer mapped but still keep the
|
||||
* extents uninitialized.
|
||||
*
|
||||
* For non AIO case, we will convert those unwritten extents
|
||||
* to written after return back from blockdev_direct_IO.
|
||||
*
|
||||
* For async DIO, the conversion needs to be deferred when the
|
||||
* IO is completed. The ext4 end_io callback function will be
|
||||
* called to take care of the conversion work. Here for async
|
||||
* case, we allocate an io_end structure to hook to the iocb.
|
||||
*/
|
||||
iocb->private = NULL;
|
||||
ext4_inode_aio_set(inode, NULL);
|
||||
if (!is_sync_kiocb(iocb)) {
|
||||
ext4_io_end_t *io_end = ext4_init_io_end(inode, GFP_NOFS);
|
||||
if (!io_end) {
|
||||
ret = -ENOMEM;
|
||||
goto retake_lock;
|
||||
}
|
||||
io_end->flag |= EXT4_IO_END_DIRECT;
|
||||
iocb->private = io_end;
|
||||
/*
|
||||
* we save the io structure for current async direct
|
||||
* IO, so that later ext4_map_blocks() could flag the
|
||||
* io structure whether there is a unwritten extents
|
||||
* needs to be converted when IO is completed.
|
||||
*/
|
||||
ext4_inode_aio_set(inode, io_end);
|
||||
}
|
||||
|
||||
if (overwrite) {
|
||||
get_block_func = ext4_get_block_write_nolock;
|
||||
} else {
|
||||
get_block_func = ext4_get_block_write;
|
||||
dio_flags = DIO_LOCKING;
|
||||
}
|
||||
ret = __blockdev_direct_IO(rw, iocb, inode,
|
||||
inode->i_sb->s_bdev, iov,
|
||||
offset, nr_segs,
|
||||
get_block_func,
|
||||
ext4_end_io_dio,
|
||||
NULL,
|
||||
dio_flags);
|
||||
|
||||
if (iocb->private)
|
||||
ext4_inode_aio_set(inode, NULL);
|
||||
/*
|
||||
* The io_end structure takes a reference to the inode, that
|
||||
* structure needs to be destroyed and the reference to the
|
||||
* inode need to be dropped, when IO is complete, even with 0
|
||||
* byte write, or failed.
|
||||
*
|
||||
* In the successful AIO DIO case, the io_end structure will
|
||||
* be destroyed and the reference to the inode will be dropped
|
||||
* after the end_io call back function is called.
|
||||
*
|
||||
* In the case there is 0 byte write, or error case, since VFS
|
||||
* direct IO won't invoke the end_io call back function, we
|
||||
* need to free the end_io structure here.
|
||||
*/
|
||||
if (ret != -EIOCBQUEUED && ret <= 0 && iocb->private) {
|
||||
ext4_free_io_end(iocb->private);
|
||||
iocb->private = NULL;
|
||||
} else if (ret > 0 && !overwrite && ext4_test_inode_state(inode,
|
||||
EXT4_STATE_DIO_UNWRITTEN)) {
|
||||
int err;
|
||||
/*
|
||||
* for non AIO case, since the IO is already
|
||||
* completed, we could do the conversion right here
|
||||
*/
|
||||
err = ext4_convert_unwritten_extents(inode,
|
||||
offset, ret);
|
||||
if (err < 0)
|
||||
ret = err;
|
||||
ext4_clear_inode_state(inode, EXT4_STATE_DIO_UNWRITTEN);
|
||||
}
|
||||
|
||||
retake_lock:
|
||||
/* take i_mutex locking again if we do a ovewrite dio */
|
||||
if (overwrite) {
|
||||
inode_dio_done(inode);
|
||||
up_read(&EXT4_I(inode)->i_data_sem);
|
||||
mutex_lock(&inode->i_mutex);
|
||||
}
|
||||
|
||||
return ret;
|
||||
}
|
||||
|
||||
static ssize_t ext4_direct_IO(int rw, struct kiocb *iocb,
|
||||
|
@ -3134,6 +3193,10 @@ static ssize_t ext4_direct_IO(int rw, struct kiocb *iocb,
|
|||
if (ext4_should_journal_data(inode))
|
||||
return 0;
|
||||
|
||||
/* Let buffer I/O handle the inline data case. */
|
||||
if (ext4_has_inline_data(inode))
|
||||
return 0;
|
||||
|
||||
trace_ext4_direct_IO_enter(inode, offset, iov_length(iov, nr_segs), rw);
|
||||
if (ext4_test_inode_flag(inode, EXT4_INODE_EXTENTS))
|
||||
ret = ext4_ext_direct_IO(rw, iocb, iov, offset, nr_segs);
|
||||
|
@ -3531,6 +3594,14 @@ void ext4_truncate(struct inode *inode)
|
|||
if (inode->i_size == 0 && !test_opt(inode->i_sb, NO_AUTO_DA_ALLOC))
|
||||
ext4_set_inode_state(inode, EXT4_STATE_DA_ALLOC_CLOSE);
|
||||
|
||||
if (ext4_has_inline_data(inode)) {
|
||||
int has_inline = 1;
|
||||
|
||||
ext4_inline_data_truncate(inode, &has_inline);
|
||||
if (has_inline)
|
||||
return;
|
||||
}
|
||||
|
||||
if (ext4_test_inode_flag(inode, EXT4_INODE_EXTENTS))
|
||||
ext4_ext_truncate(inode);
|
||||
else
|
||||
|
@ -3756,6 +3827,19 @@ static blkcnt_t ext4_inode_blocks(struct ext4_inode *raw_inode,
|
|||
}
|
||||
}
|
||||
|
||||
static inline void ext4_iget_extra_inode(struct inode *inode,
|
||||
struct ext4_inode *raw_inode,
|
||||
struct ext4_inode_info *ei)
|
||||
{
|
||||
__le32 *magic = (void *)raw_inode +
|
||||
EXT4_GOOD_OLD_INODE_SIZE + ei->i_extra_isize;
|
||||
if (*magic == cpu_to_le32(EXT4_XATTR_MAGIC)) {
|
||||
ext4_set_inode_state(inode, EXT4_STATE_XATTR);
|
||||
ext4_find_inline_data_nolock(inode);
|
||||
} else
|
||||
EXT4_I(inode)->i_inline_off = 0;
|
||||
}
|
||||
|
||||
struct inode *ext4_iget(struct super_block *sb, unsigned long ino)
|
||||
{
|
||||
struct ext4_iloc iloc;
|
||||
|
@ -3826,6 +3910,7 @@ struct inode *ext4_iget(struct super_block *sb, unsigned long ino)
|
|||
set_nlink(inode, le16_to_cpu(raw_inode->i_links_count));
|
||||
|
||||
ext4_clear_state_flags(ei); /* Only relevant on 32-bit archs */
|
||||
ei->i_inline_off = 0;
|
||||
ei->i_dir_start_lookup = 0;
|
||||
ei->i_dtime = le32_to_cpu(raw_inode->i_dtime);
|
||||
/* We now have enough fields to check if the inode was active or not.
|
||||
|
@ -3898,11 +3983,7 @@ struct inode *ext4_iget(struct super_block *sb, unsigned long ino)
|
|||
ei->i_extra_isize = sizeof(struct ext4_inode) -
|
||||
EXT4_GOOD_OLD_INODE_SIZE;
|
||||
} else {
|
||||
__le32 *magic = (void *)raw_inode +
|
||||
EXT4_GOOD_OLD_INODE_SIZE +
|
||||
ei->i_extra_isize;
|
||||
if (*magic == cpu_to_le32(EXT4_XATTR_MAGIC))
|
||||
ext4_set_inode_state(inode, EXT4_STATE_XATTR);
|
||||
ext4_iget_extra_inode(inode, raw_inode, ei);
|
||||
}
|
||||
}
|
||||
|
||||
|
@ -3925,17 +4006,19 @@ struct inode *ext4_iget(struct super_block *sb, unsigned long ino)
|
|||
ei->i_file_acl);
|
||||
ret = -EIO;
|
||||
goto bad_inode;
|
||||
} else if (ext4_test_inode_flag(inode, EXT4_INODE_EXTENTS)) {
|
||||
if (S_ISREG(inode->i_mode) || S_ISDIR(inode->i_mode) ||
|
||||
(S_ISLNK(inode->i_mode) &&
|
||||
!ext4_inode_is_fast_symlink(inode)))
|
||||
/* Validate extent which is part of inode */
|
||||
ret = ext4_ext_check_inode(inode);
|
||||
} else if (S_ISREG(inode->i_mode) || S_ISDIR(inode->i_mode) ||
|
||||
(S_ISLNK(inode->i_mode) &&
|
||||
!ext4_inode_is_fast_symlink(inode))) {
|
||||
/* Validate block references which are part of inode */
|
||||
ret = ext4_ind_check_inode(inode);
|
||||
} else if (!ext4_has_inline_data(inode)) {
|
||||
if (ext4_test_inode_flag(inode, EXT4_INODE_EXTENTS)) {
|
||||
if ((S_ISREG(inode->i_mode) || S_ISDIR(inode->i_mode) ||
|
||||
(S_ISLNK(inode->i_mode) &&
|
||||
!ext4_inode_is_fast_symlink(inode))))
|
||||
/* Validate extent which is part of inode */
|
||||
ret = ext4_ext_check_inode(inode);
|
||||
} else if (S_ISREG(inode->i_mode) || S_ISDIR(inode->i_mode) ||
|
||||
(S_ISLNK(inode->i_mode) &&
|
||||
!ext4_inode_is_fast_symlink(inode))) {
|
||||
/* Validate block references which are part of inode */
|
||||
ret = ext4_ind_check_inode(inode);
|
||||
}
|
||||
}
|
||||
if (ret)
|
||||
goto bad_inode;
|
||||
|
@ -4122,9 +4205,10 @@ static int ext4_do_update_inode(handle_t *handle,
|
|||
cpu_to_le32(new_encode_dev(inode->i_rdev));
|
||||
raw_inode->i_block[2] = 0;
|
||||
}
|
||||
} else
|
||||
} else if (!ext4_has_inline_data(inode)) {
|
||||
for (block = 0; block < EXT4_N_BLOCKS; block++)
|
||||
raw_inode->i_block[block] = ei->i_data[block];
|
||||
}
|
||||
|
||||
raw_inode->i_disk_version = cpu_to_le32(inode->i_version);
|
||||
if (ei->i_extra_isize) {
|
||||
|
@ -4811,8 +4895,9 @@ int ext4_page_mkwrite(struct vm_area_struct *vma, struct vm_fault *vmf)
|
|||
* journal_start/journal_stop which can block and take a long time
|
||||
*/
|
||||
if (page_has_buffers(page)) {
|
||||
if (!walk_page_buffers(NULL, page_buffers(page), 0, len, NULL,
|
||||
ext4_bh_unmapped)) {
|
||||
if (!ext4_walk_page_buffers(NULL, page_buffers(page),
|
||||
0, len, NULL,
|
||||
ext4_bh_unmapped)) {
|
||||
/* Wait so that we don't change page under IO */
|
||||
wait_on_page_writeback(page);
|
||||
ret = VM_FAULT_LOCKED;
|
||||
|
@ -4833,7 +4918,7 @@ retry_alloc:
|
|||
}
|
||||
ret = __block_page_mkwrite(vma, vmf, get_block);
|
||||
if (!ret && ext4_should_journal_data(inode)) {
|
||||
if (walk_page_buffers(handle, page_buffers(page), 0,
|
||||
if (ext4_walk_page_buffers(handle, page_buffers(page), 0,
|
||||
PAGE_CACHE_SIZE, NULL, do_journal_get_write_access)) {
|
||||
unlock_page(page);
|
||||
ret = VM_FAULT_SIGBUS;
|
||||
|
|
|
@ -1373,7 +1373,7 @@ static int mb_find_extent(struct ext4_buddy *e4b, int block,
|
|||
ex->fe_start += next;
|
||||
|
||||
while (needed > ex->fe_len &&
|
||||
(buddy = mb_find_buddy(e4b, order, &max))) {
|
||||
mb_find_buddy(e4b, order, &max)) {
|
||||
|
||||
if (block + 1 >= max)
|
||||
break;
|
||||
|
@ -2607,9 +2607,17 @@ static void ext4_free_data_callback(struct super_block *sb,
|
|||
mb_debug(1, "gonna free %u blocks in group %u (0x%p):",
|
||||
entry->efd_count, entry->efd_group, entry);
|
||||
|
||||
if (test_opt(sb, DISCARD))
|
||||
ext4_issue_discard(sb, entry->efd_group,
|
||||
entry->efd_start_cluster, entry->efd_count);
|
||||
if (test_opt(sb, DISCARD)) {
|
||||
err = ext4_issue_discard(sb, entry->efd_group,
|
||||
entry->efd_start_cluster,
|
||||
entry->efd_count);
|
||||
if (err && err != -EOPNOTSUPP)
|
||||
ext4_msg(sb, KERN_WARNING, "discard request in"
|
||||
" group:%d block:%d count:%d failed"
|
||||
" with %d", entry->efd_group,
|
||||
entry->efd_start_cluster,
|
||||
entry->efd_count, err);
|
||||
}
|
||||
|
||||
err = ext4_mb_load_buddy(sb, entry->efd_group, &e4b);
|
||||
/* we expect to find existing buddy because it's pinned */
|
||||
|
@ -4310,8 +4318,10 @@ ext4_fsblk_t ext4_mb_new_blocks(handle_t *handle,
|
|||
repeat:
|
||||
/* allocate space in core */
|
||||
*errp = ext4_mb_regular_allocator(ac);
|
||||
if (*errp)
|
||||
if (*errp) {
|
||||
ext4_discard_allocated_blocks(ac);
|
||||
goto errout;
|
||||
}
|
||||
|
||||
/* as we've just preallocated more space than
|
||||
* user requested orinally, we store allocated
|
||||
|
@ -4333,10 +4343,10 @@ repeat:
|
|||
ac->ac_b_ex.fe_len = 0;
|
||||
ac->ac_status = AC_STATUS_CONTINUE;
|
||||
goto repeat;
|
||||
} else if (*errp)
|
||||
errout:
|
||||
} else if (*errp) {
|
||||
ext4_discard_allocated_blocks(ac);
|
||||
else {
|
||||
goto errout;
|
||||
} else {
|
||||
block = ext4_grp_offs_to_block(sb, &ac->ac_b_ex);
|
||||
ar->len = ac->ac_b_ex.fe_len;
|
||||
}
|
||||
|
@ -4347,6 +4357,7 @@ repeat:
|
|||
*errp = -ENOSPC;
|
||||
}
|
||||
|
||||
errout:
|
||||
if (*errp) {
|
||||
ac->ac_b_ex.fe_len = 0;
|
||||
ar->len = 0;
|
||||
|
@ -4656,8 +4667,16 @@ do_more:
|
|||
* with group lock held. generate_buddy look at
|
||||
* them with group lock_held
|
||||
*/
|
||||
if (test_opt(sb, DISCARD))
|
||||
ext4_issue_discard(sb, block_group, bit, count);
|
||||
if (test_opt(sb, DISCARD)) {
|
||||
err = ext4_issue_discard(sb, block_group, bit, count);
|
||||
if (err && err != -EOPNOTSUPP)
|
||||
ext4_msg(sb, KERN_WARNING, "discard request in"
|
||||
" group:%d block:%d count:%lu failed"
|
||||
" with %d", block_group, bit, count,
|
||||
err);
|
||||
}
|
||||
|
||||
|
||||
ext4_lock_group(sb, block_group);
|
||||
mb_clear_bits(bitmap_bh->b_data, bit, count_clusters);
|
||||
mb_free_blocks(inode, &e4b, bit, count_clusters);
|
||||
|
@ -4851,10 +4870,11 @@ error_return:
|
|||
* one will allocate those blocks, mark it as used in buddy bitmap. This must
|
||||
* be called with under the group lock.
|
||||
*/
|
||||
static void ext4_trim_extent(struct super_block *sb, int start, int count,
|
||||
static int ext4_trim_extent(struct super_block *sb, int start, int count,
|
||||
ext4_group_t group, struct ext4_buddy *e4b)
|
||||
{
|
||||
struct ext4_free_extent ex;
|
||||
int ret = 0;
|
||||
|
||||
trace_ext4_trim_extent(sb, group, start, count);
|
||||
|
||||
|
@ -4870,9 +4890,10 @@ static void ext4_trim_extent(struct super_block *sb, int start, int count,
|
|||
*/
|
||||
mb_mark_used(e4b, &ex);
|
||||
ext4_unlock_group(sb, group);
|
||||
ext4_issue_discard(sb, group, start, count);
|
||||
ret = ext4_issue_discard(sb, group, start, count);
|
||||
ext4_lock_group(sb, group);
|
||||
mb_free_blocks(NULL, e4b, start, ex.fe_len);
|
||||
return ret;
|
||||
}
|
||||
|
||||
/**
|
||||
|
@ -4901,7 +4922,7 @@ ext4_trim_all_free(struct super_block *sb, ext4_group_t group,
|
|||
void *bitmap;
|
||||
ext4_grpblk_t next, count = 0, free_count = 0;
|
||||
struct ext4_buddy e4b;
|
||||
int ret;
|
||||
int ret = 0;
|
||||
|
||||
trace_ext4_trim_all_free(sb, group, start, max);
|
||||
|
||||
|
@ -4928,8 +4949,11 @@ ext4_trim_all_free(struct super_block *sb, ext4_group_t group,
|
|||
next = mb_find_next_bit(bitmap, max + 1, start);
|
||||
|
||||
if ((next - start) >= minblocks) {
|
||||
ext4_trim_extent(sb, start,
|
||||
next - start, group, &e4b);
|
||||
ret = ext4_trim_extent(sb, start,
|
||||
next - start, group, &e4b);
|
||||
if (ret && ret != -EOPNOTSUPP)
|
||||
break;
|
||||
ret = 0;
|
||||
count += next - start;
|
||||
}
|
||||
free_count += next - start;
|
||||
|
@ -4950,8 +4974,10 @@ ext4_trim_all_free(struct super_block *sb, ext4_group_t group,
|
|||
break;
|
||||
}
|
||||
|
||||
if (!ret)
|
||||
if (!ret) {
|
||||
ret = count;
|
||||
EXT4_MB_GRP_SET_TRIMMED(e4b.bd_info);
|
||||
}
|
||||
out:
|
||||
ext4_unlock_group(sb, group);
|
||||
ext4_mb_unload_buddy(&e4b);
|
||||
|
@ -4959,7 +4985,7 @@ out:
|
|||
ext4_debug("trimmed %d blocks in the group %d\n",
|
||||
count, group);
|
||||
|
||||
return count;
|
||||
return ret;
|
||||
}
|
||||
|
||||
/**
|
||||
|
|
|
@ -14,6 +14,7 @@
|
|||
|
||||
#include <linux/slab.h>
|
||||
#include "ext4_jbd2.h"
|
||||
#include "ext4_extents.h"
|
||||
|
||||
/*
|
||||
* The contiguous blocks details which can be
|
||||
|
|
|
@ -18,6 +18,7 @@
|
|||
#include <linux/slab.h>
|
||||
#include "ext4_jbd2.h"
|
||||
#include "ext4.h"
|
||||
#include "ext4_extents.h"
|
||||
|
||||
/**
|
||||
* get_ext_path - Find an extent path for designated logical block number.
|
||||
|
|
531
fs/ext4/namei.c
531
fs/ext4/namei.c
|
@ -202,13 +202,8 @@ static int ext4_dx_add_entry(handle_t *handle, struct dentry *dentry,
|
|||
struct inode *inode);
|
||||
|
||||
/* checksumming functions */
|
||||
#define EXT4_DIRENT_TAIL(block, blocksize) \
|
||||
((struct ext4_dir_entry_tail *)(((void *)(block)) + \
|
||||
((blocksize) - \
|
||||
sizeof(struct ext4_dir_entry_tail))))
|
||||
|
||||
static void initialize_dirent_tail(struct ext4_dir_entry_tail *t,
|
||||
unsigned int blocksize)
|
||||
void initialize_dirent_tail(struct ext4_dir_entry_tail *t,
|
||||
unsigned int blocksize)
|
||||
{
|
||||
memset(t, 0, sizeof(struct ext4_dir_entry_tail));
|
||||
t->det_rec_len = ext4_rec_len_to_disk(
|
||||
|
@ -261,6 +256,12 @@ static __le32 ext4_dirent_csum(struct inode *inode,
|
|||
return cpu_to_le32(csum);
|
||||
}
|
||||
|
||||
static void warn_no_space_for_csum(struct inode *inode)
|
||||
{
|
||||
ext4_warning(inode->i_sb, "no space in directory inode %lu leaf for "
|
||||
"checksum. Please run e2fsck -D.", inode->i_ino);
|
||||
}
|
||||
|
||||
int ext4_dirent_csum_verify(struct inode *inode, struct ext4_dir_entry *dirent)
|
||||
{
|
||||
struct ext4_dir_entry_tail *t;
|
||||
|
@ -271,8 +272,7 @@ int ext4_dirent_csum_verify(struct inode *inode, struct ext4_dir_entry *dirent)
|
|||
|
||||
t = get_dirent_tail(inode, dirent);
|
||||
if (!t) {
|
||||
EXT4_ERROR_INODE(inode, "metadata_csum set but no space in dir "
|
||||
"leaf for checksum. Please run e2fsck -D.");
|
||||
warn_no_space_for_csum(inode);
|
||||
return 0;
|
||||
}
|
||||
|
||||
|
@ -294,8 +294,7 @@ static void ext4_dirent_csum_set(struct inode *inode,
|
|||
|
||||
t = get_dirent_tail(inode, dirent);
|
||||
if (!t) {
|
||||
EXT4_ERROR_INODE(inode, "metadata_csum set but no space in dir "
|
||||
"leaf for checksum. Please run e2fsck -D.");
|
||||
warn_no_space_for_csum(inode);
|
||||
return;
|
||||
}
|
||||
|
||||
|
@ -303,9 +302,9 @@ static void ext4_dirent_csum_set(struct inode *inode,
|
|||
(void *)t - (void *)dirent);
|
||||
}
|
||||
|
||||
static inline int ext4_handle_dirty_dirent_node(handle_t *handle,
|
||||
struct inode *inode,
|
||||
struct buffer_head *bh)
|
||||
int ext4_handle_dirty_dirent_node(handle_t *handle,
|
||||
struct inode *inode,
|
||||
struct buffer_head *bh)
|
||||
{
|
||||
ext4_dirent_csum_set(inode, (struct ext4_dir_entry *)bh->b_data);
|
||||
return ext4_handle_dirty_metadata(handle, inode, bh);
|
||||
|
@ -377,8 +376,7 @@ static int ext4_dx_csum_verify(struct inode *inode,
|
|||
count = le16_to_cpu(c->count);
|
||||
if (count_offset + (limit * sizeof(struct dx_entry)) >
|
||||
EXT4_BLOCK_SIZE(inode->i_sb) - sizeof(struct dx_tail)) {
|
||||
EXT4_ERROR_INODE(inode, "metadata_csum set but no space for "
|
||||
"tree checksum found. Run e2fsck -D.");
|
||||
warn_no_space_for_csum(inode);
|
||||
return 1;
|
||||
}
|
||||
t = (struct dx_tail *)(((struct dx_entry *)c) + limit);
|
||||
|
@ -408,8 +406,7 @@ static void ext4_dx_csum_set(struct inode *inode, struct ext4_dir_entry *dirent)
|
|||
count = le16_to_cpu(c->count);
|
||||
if (count_offset + (limit * sizeof(struct dx_entry)) >
|
||||
EXT4_BLOCK_SIZE(inode->i_sb) - sizeof(struct dx_tail)) {
|
||||
EXT4_ERROR_INODE(inode, "metadata_csum set but no space for "
|
||||
"tree checksum. Run e2fsck -D.");
|
||||
warn_no_space_for_csum(inode);
|
||||
return;
|
||||
}
|
||||
t = (struct dx_tail *)(((struct dx_entry *)c) + limit);
|
||||
|
@ -890,6 +887,7 @@ static int htree_dirblock_to_tree(struct file *dir_file,
|
|||
EXT4_DIR_REC_LEN(0));
|
||||
for (; de < top; de = ext4_next_entry(de, dir->i_sb->s_blocksize)) {
|
||||
if (ext4_check_dir_entry(dir, NULL, de, bh,
|
||||
bh->b_data, bh->b_size,
|
||||
(block<<EXT4_BLOCK_SIZE_BITS(dir->i_sb))
|
||||
+ ((char *)de - bh->b_data))) {
|
||||
/* On error, skip the f_pos to the next block. */
|
||||
|
@ -1007,6 +1005,15 @@ errout:
|
|||
return (err);
|
||||
}
|
||||
|
||||
static inline int search_dirblock(struct buffer_head *bh,
|
||||
struct inode *dir,
|
||||
const struct qstr *d_name,
|
||||
unsigned int offset,
|
||||
struct ext4_dir_entry_2 **res_dir)
|
||||
{
|
||||
return search_dir(bh, bh->b_data, dir->i_sb->s_blocksize, dir,
|
||||
d_name, offset, res_dir);
|
||||
}
|
||||
|
||||
/*
|
||||
* Directory block splitting, compacting
|
||||
|
@ -1081,13 +1088,6 @@ static void dx_insert_block(struct dx_frame *frame, u32 hash, ext4_lblk_t block)
|
|||
dx_set_count(entries, count + 1);
|
||||
}
|
||||
|
||||
static void ext4_update_dx_flag(struct inode *inode)
|
||||
{
|
||||
if (!EXT4_HAS_COMPAT_FEATURE(inode->i_sb,
|
||||
EXT4_FEATURE_COMPAT_DIR_INDEX))
|
||||
ext4_clear_inode_flag(inode, EXT4_INODE_INDEX);
|
||||
}
|
||||
|
||||
/*
|
||||
* NOTE! unlike strncmp, ext4_match returns 1 for success, 0 for failure.
|
||||
*
|
||||
|
@ -1107,11 +1107,13 @@ static inline int ext4_match (int len, const char * const name,
|
|||
/*
|
||||
* Returns 0 if not found, -1 on failure, and 1 on success
|
||||
*/
|
||||
static inline int search_dirblock(struct buffer_head *bh,
|
||||
struct inode *dir,
|
||||
const struct qstr *d_name,
|
||||
unsigned int offset,
|
||||
struct ext4_dir_entry_2 ** res_dir)
|
||||
int search_dir(struct buffer_head *bh,
|
||||
char *search_buf,
|
||||
int buf_size,
|
||||
struct inode *dir,
|
||||
const struct qstr *d_name,
|
||||
unsigned int offset,
|
||||
struct ext4_dir_entry_2 **res_dir)
|
||||
{
|
||||
struct ext4_dir_entry_2 * de;
|
||||
char * dlimit;
|
||||
|
@ -1119,8 +1121,8 @@ static inline int search_dirblock(struct buffer_head *bh,
|
|||
const char *name = d_name->name;
|
||||
int namelen = d_name->len;
|
||||
|
||||
de = (struct ext4_dir_entry_2 *) bh->b_data;
|
||||
dlimit = bh->b_data + dir->i_sb->s_blocksize;
|
||||
de = (struct ext4_dir_entry_2 *)search_buf;
|
||||
dlimit = search_buf + buf_size;
|
||||
while ((char *) de < dlimit) {
|
||||
/* this code is executed quadratically often */
|
||||
/* do minimal checking `by hand' */
|
||||
|
@ -1128,7 +1130,8 @@ static inline int search_dirblock(struct buffer_head *bh,
|
|||
if ((char *) de + namelen <= dlimit &&
|
||||
ext4_match (namelen, name, de)) {
|
||||
/* found a match - just to be sure, do a full check */
|
||||
if (ext4_check_dir_entry(dir, NULL, de, bh, offset))
|
||||
if (ext4_check_dir_entry(dir, NULL, de, bh, bh->b_data,
|
||||
bh->b_size, offset))
|
||||
return -1;
|
||||
*res_dir = de;
|
||||
return 1;
|
||||
|
@ -1144,6 +1147,21 @@ static inline int search_dirblock(struct buffer_head *bh,
|
|||
return 0;
|
||||
}
|
||||
|
||||
static int is_dx_internal_node(struct inode *dir, ext4_lblk_t block,
|
||||
struct ext4_dir_entry *de)
|
||||
{
|
||||
struct super_block *sb = dir->i_sb;
|
||||
|
||||
if (!is_dx(dir))
|
||||
return 0;
|
||||
if (block == 0)
|
||||
return 1;
|
||||
if (de->inode == 0 &&
|
||||
ext4_rec_len_from_disk(de->rec_len, sb->s_blocksize) ==
|
||||
sb->s_blocksize)
|
||||
return 1;
|
||||
return 0;
|
||||
}
|
||||
|
||||
/*
|
||||
* ext4_find_entry()
|
||||
|
@ -1158,7 +1176,8 @@ static inline int search_dirblock(struct buffer_head *bh,
|
|||
*/
|
||||
static struct buffer_head * ext4_find_entry (struct inode *dir,
|
||||
const struct qstr *d_name,
|
||||
struct ext4_dir_entry_2 ** res_dir)
|
||||
struct ext4_dir_entry_2 **res_dir,
|
||||
int *inlined)
|
||||
{
|
||||
struct super_block *sb;
|
||||
struct buffer_head *bh_use[NAMEI_RA_SIZE];
|
||||
|
@ -1179,6 +1198,18 @@ static struct buffer_head * ext4_find_entry (struct inode *dir,
|
|||
namelen = d_name->len;
|
||||
if (namelen > EXT4_NAME_LEN)
|
||||
return NULL;
|
||||
|
||||
if (ext4_has_inline_data(dir)) {
|
||||
int has_inline_data = 1;
|
||||
ret = ext4_find_inline_entry(dir, d_name, res_dir,
|
||||
&has_inline_data);
|
||||
if (has_inline_data) {
|
||||
if (inlined)
|
||||
*inlined = 1;
|
||||
return ret;
|
||||
}
|
||||
}
|
||||
|
||||
if ((namelen <= 2) && (name[0] == '.') &&
|
||||
(name[1] == '.' || name[1] == '\0')) {
|
||||
/*
|
||||
|
@ -1244,6 +1275,8 @@ restart:
|
|||
goto next;
|
||||
}
|
||||
if (!buffer_verified(bh) &&
|
||||
!is_dx_internal_node(dir, block,
|
||||
(struct ext4_dir_entry *)bh->b_data) &&
|
||||
!ext4_dirent_csum_verify(dir,
|
||||
(struct ext4_dir_entry *)bh->b_data)) {
|
||||
EXT4_ERROR_INODE(dir, "checksumming directory "
|
||||
|
@ -1361,7 +1394,7 @@ static struct dentry *ext4_lookup(struct inode *dir, struct dentry *dentry, unsi
|
|||
if (dentry->d_name.len > EXT4_NAME_LEN)
|
||||
return ERR_PTR(-ENAMETOOLONG);
|
||||
|
||||
bh = ext4_find_entry(dir, &dentry->d_name, &de);
|
||||
bh = ext4_find_entry(dir, &dentry->d_name, &de, NULL);
|
||||
inode = NULL;
|
||||
if (bh) {
|
||||
__u32 ino = le32_to_cpu(de->inode);
|
||||
|
@ -1395,7 +1428,7 @@ struct dentry *ext4_get_parent(struct dentry *child)
|
|||
struct ext4_dir_entry_2 * de;
|
||||
struct buffer_head *bh;
|
||||
|
||||
bh = ext4_find_entry(child->d_inode, &dotdot, &de);
|
||||
bh = ext4_find_entry(child->d_inode, &dotdot, &de, NULL);
|
||||
if (!bh)
|
||||
return ERR_PTR(-ENOENT);
|
||||
ino = le32_to_cpu(de->inode);
|
||||
|
@ -1593,6 +1626,63 @@ errout:
|
|||
return NULL;
|
||||
}
|
||||
|
||||
int ext4_find_dest_de(struct inode *dir, struct inode *inode,
|
||||
struct buffer_head *bh,
|
||||
void *buf, int buf_size,
|
||||
const char *name, int namelen,
|
||||
struct ext4_dir_entry_2 **dest_de)
|
||||
{
|
||||
struct ext4_dir_entry_2 *de;
|
||||
unsigned short reclen = EXT4_DIR_REC_LEN(namelen);
|
||||
int nlen, rlen;
|
||||
unsigned int offset = 0;
|
||||
char *top;
|
||||
|
||||
de = (struct ext4_dir_entry_2 *)buf;
|
||||
top = buf + buf_size - reclen;
|
||||
while ((char *) de <= top) {
|
||||
if (ext4_check_dir_entry(dir, NULL, de, bh,
|
||||
buf, buf_size, offset))
|
||||
return -EIO;
|
||||
if (ext4_match(namelen, name, de))
|
||||
return -EEXIST;
|
||||
nlen = EXT4_DIR_REC_LEN(de->name_len);
|
||||
rlen = ext4_rec_len_from_disk(de->rec_len, buf_size);
|
||||
if ((de->inode ? rlen - nlen : rlen) >= reclen)
|
||||
break;
|
||||
de = (struct ext4_dir_entry_2 *)((char *)de + rlen);
|
||||
offset += rlen;
|
||||
}
|
||||
if ((char *) de > top)
|
||||
return -ENOSPC;
|
||||
|
||||
*dest_de = de;
|
||||
return 0;
|
||||
}
|
||||
|
||||
void ext4_insert_dentry(struct inode *inode,
|
||||
struct ext4_dir_entry_2 *de,
|
||||
int buf_size,
|
||||
const char *name, int namelen)
|
||||
{
|
||||
|
||||
int nlen, rlen;
|
||||
|
||||
nlen = EXT4_DIR_REC_LEN(de->name_len);
|
||||
rlen = ext4_rec_len_from_disk(de->rec_len, buf_size);
|
||||
if (de->inode) {
|
||||
struct ext4_dir_entry_2 *de1 =
|
||||
(struct ext4_dir_entry_2 *)((char *)de + nlen);
|
||||
de1->rec_len = ext4_rec_len_to_disk(rlen - nlen, buf_size);
|
||||
de->rec_len = ext4_rec_len_to_disk(nlen, buf_size);
|
||||
de = de1;
|
||||
}
|
||||
de->file_type = EXT4_FT_UNKNOWN;
|
||||
de->inode = cpu_to_le32(inode->i_ino);
|
||||
ext4_set_de_type(inode->i_sb, de, inode->i_mode);
|
||||
de->name_len = namelen;
|
||||
memcpy(de->name, name, namelen);
|
||||
}
|
||||
/*
|
||||
* Add a new entry into a directory (leaf) block. If de is non-NULL,
|
||||
* it points to a directory entry which is guaranteed to be large
|
||||
|
@ -1608,12 +1698,10 @@ static int add_dirent_to_buf(handle_t *handle, struct dentry *dentry,
|
|||
struct inode *dir = dentry->d_parent->d_inode;
|
||||
const char *name = dentry->d_name.name;
|
||||
int namelen = dentry->d_name.len;
|
||||
unsigned int offset = 0;
|
||||
unsigned int blocksize = dir->i_sb->s_blocksize;
|
||||
unsigned short reclen;
|
||||
int nlen, rlen, err;
|
||||
char *top;
|
||||
int csum_size = 0;
|
||||
int err;
|
||||
|
||||
if (EXT4_HAS_RO_COMPAT_FEATURE(inode->i_sb,
|
||||
EXT4_FEATURE_RO_COMPAT_METADATA_CSUM))
|
||||
|
@ -1621,22 +1709,11 @@ static int add_dirent_to_buf(handle_t *handle, struct dentry *dentry,
|
|||
|
||||
reclen = EXT4_DIR_REC_LEN(namelen);
|
||||
if (!de) {
|
||||
de = (struct ext4_dir_entry_2 *)bh->b_data;
|
||||
top = bh->b_data + (blocksize - csum_size) - reclen;
|
||||
while ((char *) de <= top) {
|
||||
if (ext4_check_dir_entry(dir, NULL, de, bh, offset))
|
||||
return -EIO;
|
||||
if (ext4_match(namelen, name, de))
|
||||
return -EEXIST;
|
||||
nlen = EXT4_DIR_REC_LEN(de->name_len);
|
||||
rlen = ext4_rec_len_from_disk(de->rec_len, blocksize);
|
||||
if ((de->inode? rlen - nlen: rlen) >= reclen)
|
||||
break;
|
||||
de = (struct ext4_dir_entry_2 *)((char *)de + rlen);
|
||||
offset += rlen;
|
||||
}
|
||||
if ((char *) de > top)
|
||||
return -ENOSPC;
|
||||
err = ext4_find_dest_de(dir, inode,
|
||||
bh, bh->b_data, blocksize - csum_size,
|
||||
name, namelen, &de);
|
||||
if (err)
|
||||
return err;
|
||||
}
|
||||
BUFFER_TRACE(bh, "get_write_access");
|
||||
err = ext4_journal_get_write_access(handle, bh);
|
||||
|
@ -1646,19 +1723,8 @@ static int add_dirent_to_buf(handle_t *handle, struct dentry *dentry,
|
|||
}
|
||||
|
||||
/* By now the buffer is marked for journaling */
|
||||
nlen = EXT4_DIR_REC_LEN(de->name_len);
|
||||
rlen = ext4_rec_len_from_disk(de->rec_len, blocksize);
|
||||
if (de->inode) {
|
||||
struct ext4_dir_entry_2 *de1 = (struct ext4_dir_entry_2 *)((char *)de + nlen);
|
||||
de1->rec_len = ext4_rec_len_to_disk(rlen - nlen, blocksize);
|
||||
de->rec_len = ext4_rec_len_to_disk(nlen, blocksize);
|
||||
de = de1;
|
||||
}
|
||||
de->file_type = EXT4_FT_UNKNOWN;
|
||||
de->inode = cpu_to_le32(inode->i_ino);
|
||||
ext4_set_de_type(dir->i_sb, de, inode->i_mode);
|
||||
de->name_len = namelen;
|
||||
memcpy(de->name, name, namelen);
|
||||
ext4_insert_dentry(inode, de, blocksize, name, namelen);
|
||||
|
||||
/*
|
||||
* XXX shouldn't update any times until successful
|
||||
* completion of syscall, but too many callers depend
|
||||
|
@ -1831,6 +1897,17 @@ static int ext4_add_entry(handle_t *handle, struct dentry *dentry,
|
|||
blocksize = sb->s_blocksize;
|
||||
if (!dentry->d_name.len)
|
||||
return -EINVAL;
|
||||
|
||||
if (ext4_has_inline_data(dir)) {
|
||||
retval = ext4_try_add_inline_entry(handle, dentry, inode);
|
||||
if (retval < 0)
|
||||
return retval;
|
||||
if (retval == 1) {
|
||||
retval = 0;
|
||||
return retval;
|
||||
}
|
||||
}
|
||||
|
||||
if (is_dx(dir)) {
|
||||
retval = ext4_dx_add_entry(handle, dentry, inode);
|
||||
if (!retval || (retval != ERR_BAD_DX_DIR))
|
||||
|
@ -2036,36 +2113,29 @@ cleanup:
|
|||
}
|
||||
|
||||
/*
|
||||
* ext4_delete_entry deletes a directory entry by merging it with the
|
||||
* previous entry
|
||||
* ext4_generic_delete_entry deletes a directory entry by merging it
|
||||
* with the previous entry
|
||||
*/
|
||||
static int ext4_delete_entry(handle_t *handle,
|
||||
struct inode *dir,
|
||||
struct ext4_dir_entry_2 *de_del,
|
||||
struct buffer_head *bh)
|
||||
int ext4_generic_delete_entry(handle_t *handle,
|
||||
struct inode *dir,
|
||||
struct ext4_dir_entry_2 *de_del,
|
||||
struct buffer_head *bh,
|
||||
void *entry_buf,
|
||||
int buf_size,
|
||||
int csum_size)
|
||||
{
|
||||
struct ext4_dir_entry_2 *de, *pde;
|
||||
unsigned int blocksize = dir->i_sb->s_blocksize;
|
||||
int csum_size = 0;
|
||||
int i, err;
|
||||
|
||||
if (EXT4_HAS_RO_COMPAT_FEATURE(dir->i_sb,
|
||||
EXT4_FEATURE_RO_COMPAT_METADATA_CSUM))
|
||||
csum_size = sizeof(struct ext4_dir_entry_tail);
|
||||
int i;
|
||||
|
||||
i = 0;
|
||||
pde = NULL;
|
||||
de = (struct ext4_dir_entry_2 *) bh->b_data;
|
||||
while (i < bh->b_size - csum_size) {
|
||||
if (ext4_check_dir_entry(dir, NULL, de, bh, i))
|
||||
de = (struct ext4_dir_entry_2 *)entry_buf;
|
||||
while (i < buf_size - csum_size) {
|
||||
if (ext4_check_dir_entry(dir, NULL, de, bh,
|
||||
bh->b_data, bh->b_size, i))
|
||||
return -EIO;
|
||||
if (de == de_del) {
|
||||
BUFFER_TRACE(bh, "get_write_access");
|
||||
err = ext4_journal_get_write_access(handle, bh);
|
||||
if (unlikely(err)) {
|
||||
ext4_std_error(dir->i_sb, err);
|
||||
return err;
|
||||
}
|
||||
if (pde)
|
||||
pde->rec_len = ext4_rec_len_to_disk(
|
||||
ext4_rec_len_from_disk(pde->rec_len,
|
||||
|
@ -2076,12 +2146,6 @@ static int ext4_delete_entry(handle_t *handle,
|
|||
else
|
||||
de->inode = 0;
|
||||
dir->i_version++;
|
||||
BUFFER_TRACE(bh, "call ext4_handle_dirty_metadata");
|
||||
err = ext4_handle_dirty_dirent_node(handle, dir, bh);
|
||||
if (unlikely(err)) {
|
||||
ext4_std_error(dir->i_sb, err);
|
||||
return err;
|
||||
}
|
||||
return 0;
|
||||
}
|
||||
i += ext4_rec_len_from_disk(de->rec_len, blocksize);
|
||||
|
@ -2091,6 +2155,48 @@ static int ext4_delete_entry(handle_t *handle,
|
|||
return -ENOENT;
|
||||
}
|
||||
|
||||
static int ext4_delete_entry(handle_t *handle,
|
||||
struct inode *dir,
|
||||
struct ext4_dir_entry_2 *de_del,
|
||||
struct buffer_head *bh)
|
||||
{
|
||||
int err, csum_size = 0;
|
||||
|
||||
if (ext4_has_inline_data(dir)) {
|
||||
int has_inline_data = 1;
|
||||
err = ext4_delete_inline_entry(handle, dir, de_del, bh,
|
||||
&has_inline_data);
|
||||
if (has_inline_data)
|
||||
return err;
|
||||
}
|
||||
|
||||
if (EXT4_HAS_RO_COMPAT_FEATURE(dir->i_sb,
|
||||
EXT4_FEATURE_RO_COMPAT_METADATA_CSUM))
|
||||
csum_size = sizeof(struct ext4_dir_entry_tail);
|
||||
|
||||
BUFFER_TRACE(bh, "get_write_access");
|
||||
err = ext4_journal_get_write_access(handle, bh);
|
||||
if (unlikely(err))
|
||||
goto out;
|
||||
|
||||
err = ext4_generic_delete_entry(handle, dir, de_del,
|
||||
bh, bh->b_data,
|
||||
dir->i_sb->s_blocksize, csum_size);
|
||||
if (err)
|
||||
goto out;
|
||||
|
||||
BUFFER_TRACE(bh, "call ext4_handle_dirty_metadata");
|
||||
err = ext4_handle_dirty_dirent_node(handle, dir, bh);
|
||||
if (unlikely(err))
|
||||
goto out;
|
||||
|
||||
return 0;
|
||||
out:
|
||||
if (err != -ENOENT)
|
||||
ext4_std_error(dir->i_sb, err);
|
||||
return err;
|
||||
}
|
||||
|
||||
/*
|
||||
* DIR_NLINK feature is set if 1) nlinks > EXT4_LINK_MAX or 2) nlinks == 2,
|
||||
* since this indicates that nlinks count was previously 1.
|
||||
|
@ -2211,21 +2317,95 @@ retry:
|
|||
return err;
|
||||
}
|
||||
|
||||
static int ext4_mkdir(struct inode *dir, struct dentry *dentry, umode_t mode)
|
||||
struct ext4_dir_entry_2 *ext4_init_dot_dotdot(struct inode *inode,
|
||||
struct ext4_dir_entry_2 *de,
|
||||
int blocksize, int csum_size,
|
||||
unsigned int parent_ino, int dotdot_real_len)
|
||||
{
|
||||
de->inode = cpu_to_le32(inode->i_ino);
|
||||
de->name_len = 1;
|
||||
de->rec_len = ext4_rec_len_to_disk(EXT4_DIR_REC_LEN(de->name_len),
|
||||
blocksize);
|
||||
strcpy(de->name, ".");
|
||||
ext4_set_de_type(inode->i_sb, de, S_IFDIR);
|
||||
|
||||
de = ext4_next_entry(de, blocksize);
|
||||
de->inode = cpu_to_le32(parent_ino);
|
||||
de->name_len = 2;
|
||||
if (!dotdot_real_len)
|
||||
de->rec_len = ext4_rec_len_to_disk(blocksize -
|
||||
(csum_size + EXT4_DIR_REC_LEN(1)),
|
||||
blocksize);
|
||||
else
|
||||
de->rec_len = ext4_rec_len_to_disk(
|
||||
EXT4_DIR_REC_LEN(de->name_len), blocksize);
|
||||
strcpy(de->name, "..");
|
||||
ext4_set_de_type(inode->i_sb, de, S_IFDIR);
|
||||
|
||||
return ext4_next_entry(de, blocksize);
|
||||
}
|
||||
|
||||
static int ext4_init_new_dir(handle_t *handle, struct inode *dir,
|
||||
struct inode *inode)
|
||||
{
|
||||
handle_t *handle;
|
||||
struct inode *inode;
|
||||
struct buffer_head *dir_block = NULL;
|
||||
struct ext4_dir_entry_2 *de;
|
||||
struct ext4_dir_entry_tail *t;
|
||||
unsigned int blocksize = dir->i_sb->s_blocksize;
|
||||
int csum_size = 0;
|
||||
int err, retries = 0;
|
||||
int err;
|
||||
|
||||
if (EXT4_HAS_RO_COMPAT_FEATURE(dir->i_sb,
|
||||
EXT4_FEATURE_RO_COMPAT_METADATA_CSUM))
|
||||
csum_size = sizeof(struct ext4_dir_entry_tail);
|
||||
|
||||
if (ext4_test_inode_state(inode, EXT4_STATE_MAY_INLINE_DATA)) {
|
||||
err = ext4_try_create_inline_dir(handle, dir, inode);
|
||||
if (err < 0 && err != -ENOSPC)
|
||||
goto out;
|
||||
if (!err)
|
||||
goto out;
|
||||
}
|
||||
|
||||
inode->i_size = EXT4_I(inode)->i_disksize = blocksize;
|
||||
dir_block = ext4_bread(handle, inode, 0, 1, &err);
|
||||
if (!(dir_block = ext4_bread(handle, inode, 0, 1, &err))) {
|
||||
if (!err) {
|
||||
err = -EIO;
|
||||
ext4_error(inode->i_sb,
|
||||
"Directory hole detected on inode %lu\n",
|
||||
inode->i_ino);
|
||||
}
|
||||
goto out;
|
||||
}
|
||||
BUFFER_TRACE(dir_block, "get_write_access");
|
||||
err = ext4_journal_get_write_access(handle, dir_block);
|
||||
if (err)
|
||||
goto out;
|
||||
de = (struct ext4_dir_entry_2 *)dir_block->b_data;
|
||||
ext4_init_dot_dotdot(inode, de, blocksize, csum_size, dir->i_ino, 0);
|
||||
set_nlink(inode, 2);
|
||||
if (csum_size) {
|
||||
t = EXT4_DIRENT_TAIL(dir_block->b_data, blocksize);
|
||||
initialize_dirent_tail(t, blocksize);
|
||||
}
|
||||
|
||||
BUFFER_TRACE(dir_block, "call ext4_handle_dirty_metadata");
|
||||
err = ext4_handle_dirty_dirent_node(handle, inode, dir_block);
|
||||
if (err)
|
||||
goto out;
|
||||
set_buffer_verified(dir_block);
|
||||
out:
|
||||
brelse(dir_block);
|
||||
return err;
|
||||
}
|
||||
|
||||
static int ext4_mkdir(struct inode *dir, struct dentry *dentry, umode_t mode)
|
||||
{
|
||||
handle_t *handle;
|
||||
struct inode *inode;
|
||||
int err, retries = 0;
|
||||
|
||||
if (EXT4_DIR_LINK_MAX(dir))
|
||||
return -EMLINK;
|
||||
|
||||
|
@ -2249,47 +2429,9 @@ retry:
|
|||
|
||||
inode->i_op = &ext4_dir_inode_operations;
|
||||
inode->i_fop = &ext4_dir_operations;
|
||||
inode->i_size = EXT4_I(inode)->i_disksize = inode->i_sb->s_blocksize;
|
||||
if (!(dir_block = ext4_bread(handle, inode, 0, 1, &err))) {
|
||||
if (!err) {
|
||||
err = -EIO;
|
||||
ext4_error(inode->i_sb,
|
||||
"Directory hole detected on inode %lu\n",
|
||||
inode->i_ino);
|
||||
}
|
||||
goto out_clear_inode;
|
||||
}
|
||||
BUFFER_TRACE(dir_block, "get_write_access");
|
||||
err = ext4_journal_get_write_access(handle, dir_block);
|
||||
err = ext4_init_new_dir(handle, dir, inode);
|
||||
if (err)
|
||||
goto out_clear_inode;
|
||||
de = (struct ext4_dir_entry_2 *) dir_block->b_data;
|
||||
de->inode = cpu_to_le32(inode->i_ino);
|
||||
de->name_len = 1;
|
||||
de->rec_len = ext4_rec_len_to_disk(EXT4_DIR_REC_LEN(de->name_len),
|
||||
blocksize);
|
||||
strcpy(de->name, ".");
|
||||
ext4_set_de_type(dir->i_sb, de, S_IFDIR);
|
||||
de = ext4_next_entry(de, blocksize);
|
||||
de->inode = cpu_to_le32(dir->i_ino);
|
||||
de->rec_len = ext4_rec_len_to_disk(blocksize -
|
||||
(csum_size + EXT4_DIR_REC_LEN(1)),
|
||||
blocksize);
|
||||
de->name_len = 2;
|
||||
strcpy(de->name, "..");
|
||||
ext4_set_de_type(dir->i_sb, de, S_IFDIR);
|
||||
set_nlink(inode, 2);
|
||||
|
||||
if (csum_size) {
|
||||
t = EXT4_DIRENT_TAIL(dir_block->b_data, blocksize);
|
||||
initialize_dirent_tail(t, blocksize);
|
||||
}
|
||||
|
||||
BUFFER_TRACE(dir_block, "call ext4_handle_dirty_metadata");
|
||||
err = ext4_handle_dirty_dirent_node(handle, inode, dir_block);
|
||||
if (err)
|
||||
goto out_clear_inode;
|
||||
set_buffer_verified(dir_block);
|
||||
err = ext4_mark_inode_dirty(handle, inode);
|
||||
if (!err)
|
||||
err = ext4_add_entry(handle, dentry, inode);
|
||||
|
@ -2309,7 +2451,6 @@ out_clear_inode:
|
|||
unlock_new_inode(inode);
|
||||
d_instantiate(dentry, inode);
|
||||
out_stop:
|
||||
brelse(dir_block);
|
||||
ext4_journal_stop(handle);
|
||||
if (err == -ENOSPC && ext4_should_retry_alloc(dir->i_sb, &retries))
|
||||
goto retry;
|
||||
|
@ -2327,6 +2468,14 @@ static int empty_dir(struct inode *inode)
|
|||
struct super_block *sb;
|
||||
int err = 0;
|
||||
|
||||
if (ext4_has_inline_data(inode)) {
|
||||
int has_inline_data = 1;
|
||||
|
||||
err = empty_inline_dir(inode, &has_inline_data);
|
||||
if (has_inline_data)
|
||||
return err;
|
||||
}
|
||||
|
||||
sb = inode->i_sb;
|
||||
if (inode->i_size < EXT4_DIR_REC_LEN(1) + EXT4_DIR_REC_LEN(2) ||
|
||||
!(bh = ext4_bread(NULL, inode, 0, 0, &err))) {
|
||||
|
@ -2393,7 +2542,8 @@ static int empty_dir(struct inode *inode)
|
|||
set_buffer_verified(bh);
|
||||
de = (struct ext4_dir_entry_2 *) bh->b_data;
|
||||
}
|
||||
if (ext4_check_dir_entry(inode, NULL, de, bh, offset)) {
|
||||
if (ext4_check_dir_entry(inode, NULL, de, bh,
|
||||
bh->b_data, bh->b_size, offset)) {
|
||||
de = (struct ext4_dir_entry_2 *)(bh->b_data +
|
||||
sb->s_blocksize);
|
||||
offset = (offset | (sb->s_blocksize - 1)) + 1;
|
||||
|
@ -2579,7 +2729,7 @@ static int ext4_rmdir(struct inode *dir, struct dentry *dentry)
|
|||
return PTR_ERR(handle);
|
||||
|
||||
retval = -ENOENT;
|
||||
bh = ext4_find_entry(dir, &dentry->d_name, &de);
|
||||
bh = ext4_find_entry(dir, &dentry->d_name, &de, NULL);
|
||||
if (!bh)
|
||||
goto end_rmdir;
|
||||
|
||||
|
@ -2644,7 +2794,7 @@ static int ext4_unlink(struct inode *dir, struct dentry *dentry)
|
|||
ext4_handle_sync(handle);
|
||||
|
||||
retval = -ENOENT;
|
||||
bh = ext4_find_entry(dir, &dentry->d_name, &de);
|
||||
bh = ext4_find_entry(dir, &dentry->d_name, &de, NULL);
|
||||
if (!bh)
|
||||
goto end_unlink;
|
||||
|
||||
|
@ -2826,8 +2976,39 @@ retry:
|
|||
return err;
|
||||
}
|
||||
|
||||
#define PARENT_INO(buffer, size) \
|
||||
(ext4_next_entry((struct ext4_dir_entry_2 *)(buffer), size)->inode)
|
||||
|
||||
/*
|
||||
* Try to find buffer head where contains the parent block.
|
||||
* It should be the inode block if it is inlined or the 1st block
|
||||
* if it is a normal dir.
|
||||
*/
|
||||
static struct buffer_head *ext4_get_first_dir_block(handle_t *handle,
|
||||
struct inode *inode,
|
||||
int *retval,
|
||||
struct ext4_dir_entry_2 **parent_de,
|
||||
int *inlined)
|
||||
{
|
||||
struct buffer_head *bh;
|
||||
|
||||
if (!ext4_has_inline_data(inode)) {
|
||||
if (!(bh = ext4_bread(handle, inode, 0, 0, retval))) {
|
||||
if (!*retval) {
|
||||
*retval = -EIO;
|
||||
ext4_error(inode->i_sb,
|
||||
"Directory hole detected on inode %lu\n",
|
||||
inode->i_ino);
|
||||
}
|
||||
return NULL;
|
||||
}
|
||||
*parent_de = ext4_next_entry(
|
||||
(struct ext4_dir_entry_2 *)bh->b_data,
|
||||
inode->i_sb->s_blocksize);
|
||||
return bh;
|
||||
}
|
||||
|
||||
*inlined = 1;
|
||||
return ext4_get_first_inline_block(inode, parent_de, retval);
|
||||
}
|
||||
|
||||
/*
|
||||
* Anybody can rename anything with this: the permission checks are left to the
|
||||
|
@ -2841,6 +3022,8 @@ static int ext4_rename(struct inode *old_dir, struct dentry *old_dentry,
|
|||
struct buffer_head *old_bh, *new_bh, *dir_bh;
|
||||
struct ext4_dir_entry_2 *old_de, *new_de;
|
||||
int retval, force_da_alloc = 0;
|
||||
int inlined = 0, new_inlined = 0;
|
||||
struct ext4_dir_entry_2 *parent_de;
|
||||
|
||||
dquot_initialize(old_dir);
|
||||
dquot_initialize(new_dir);
|
||||
|
@ -2860,7 +3043,7 @@ static int ext4_rename(struct inode *old_dir, struct dentry *old_dentry,
|
|||
if (IS_DIRSYNC(old_dir) || IS_DIRSYNC(new_dir))
|
||||
ext4_handle_sync(handle);
|
||||
|
||||
old_bh = ext4_find_entry(old_dir, &old_dentry->d_name, &old_de);
|
||||
old_bh = ext4_find_entry(old_dir, &old_dentry->d_name, &old_de, NULL);
|
||||
/*
|
||||
* Check for inode number is _not_ due to possible IO errors.
|
||||
* We might rmdir the source, keep it as pwd of some process
|
||||
|
@ -2873,7 +3056,8 @@ static int ext4_rename(struct inode *old_dir, struct dentry *old_dentry,
|
|||
goto end_rename;
|
||||
|
||||
new_inode = new_dentry->d_inode;
|
||||
new_bh = ext4_find_entry(new_dir, &new_dentry->d_name, &new_de);
|
||||
new_bh = ext4_find_entry(new_dir, &new_dentry->d_name,
|
||||
&new_de, &new_inlined);
|
||||
if (new_bh) {
|
||||
if (!new_inode) {
|
||||
brelse(new_bh);
|
||||
|
@ -2887,22 +3071,17 @@ static int ext4_rename(struct inode *old_dir, struct dentry *old_dentry,
|
|||
goto end_rename;
|
||||
}
|
||||
retval = -EIO;
|
||||
if (!(dir_bh = ext4_bread(handle, old_inode, 0, 0, &retval))) {
|
||||
if (!retval) {
|
||||
retval = -EIO;
|
||||
ext4_error(old_inode->i_sb,
|
||||
"Directory hole detected on inode %lu\n",
|
||||
old_inode->i_ino);
|
||||
}
|
||||
dir_bh = ext4_get_first_dir_block(handle, old_inode,
|
||||
&retval, &parent_de,
|
||||
&inlined);
|
||||
if (!dir_bh)
|
||||
goto end_rename;
|
||||
}
|
||||
if (!buffer_verified(dir_bh) &&
|
||||
if (!inlined && !buffer_verified(dir_bh) &&
|
||||
!ext4_dirent_csum_verify(old_inode,
|
||||
(struct ext4_dir_entry *)dir_bh->b_data))
|
||||
goto end_rename;
|
||||
set_buffer_verified(dir_bh);
|
||||
if (le32_to_cpu(PARENT_INO(dir_bh->b_data,
|
||||
old_dir->i_sb->s_blocksize)) != old_dir->i_ino)
|
||||
if (le32_to_cpu(parent_de->inode) != old_dir->i_ino)
|
||||
goto end_rename;
|
||||
retval = -EMLINK;
|
||||
if (!new_inode && new_dir != old_dir &&
|
||||
|
@ -2931,10 +3110,13 @@ static int ext4_rename(struct inode *old_dir, struct dentry *old_dentry,
|
|||
ext4_current_time(new_dir);
|
||||
ext4_mark_inode_dirty(handle, new_dir);
|
||||
BUFFER_TRACE(new_bh, "call ext4_handle_dirty_metadata");
|
||||
retval = ext4_handle_dirty_dirent_node(handle, new_dir, new_bh);
|
||||
if (unlikely(retval)) {
|
||||
ext4_std_error(new_dir->i_sb, retval);
|
||||
goto end_rename;
|
||||
if (!new_inlined) {
|
||||
retval = ext4_handle_dirty_dirent_node(handle,
|
||||
new_dir, new_bh);
|
||||
if (unlikely(retval)) {
|
||||
ext4_std_error(new_dir->i_sb, retval);
|
||||
goto end_rename;
|
||||
}
|
||||
}
|
||||
brelse(new_bh);
|
||||
new_bh = NULL;
|
||||
|
@ -2962,7 +3144,8 @@ static int ext4_rename(struct inode *old_dir, struct dentry *old_dentry,
|
|||
struct buffer_head *old_bh2;
|
||||
struct ext4_dir_entry_2 *old_de2;
|
||||
|
||||
old_bh2 = ext4_find_entry(old_dir, &old_dentry->d_name, &old_de2);
|
||||
old_bh2 = ext4_find_entry(old_dir, &old_dentry->d_name,
|
||||
&old_de2, NULL);
|
||||
if (old_bh2) {
|
||||
retval = ext4_delete_entry(handle, old_dir,
|
||||
old_de2, old_bh2);
|
||||
|
@ -2982,17 +3165,19 @@ static int ext4_rename(struct inode *old_dir, struct dentry *old_dentry,
|
|||
old_dir->i_ctime = old_dir->i_mtime = ext4_current_time(old_dir);
|
||||
ext4_update_dx_flag(old_dir);
|
||||
if (dir_bh) {
|
||||
PARENT_INO(dir_bh->b_data, new_dir->i_sb->s_blocksize) =
|
||||
cpu_to_le32(new_dir->i_ino);
|
||||
parent_de->inode = cpu_to_le32(new_dir->i_ino);
|
||||
BUFFER_TRACE(dir_bh, "call ext4_handle_dirty_metadata");
|
||||
if (is_dx(old_inode)) {
|
||||
retval = ext4_handle_dirty_dx_node(handle,
|
||||
old_inode,
|
||||
dir_bh);
|
||||
if (!inlined) {
|
||||
if (is_dx(old_inode)) {
|
||||
retval = ext4_handle_dirty_dx_node(handle,
|
||||
old_inode,
|
||||
dir_bh);
|
||||
} else {
|
||||
retval = ext4_handle_dirty_dirent_node(handle,
|
||||
old_inode, dir_bh);
|
||||
}
|
||||
} else {
|
||||
retval = ext4_handle_dirty_dirent_node(handle,
|
||||
old_inode,
|
||||
dir_bh);
|
||||
retval = ext4_mark_inode_dirty(handle, old_inode);
|
||||
}
|
||||
if (retval) {
|
||||
ext4_std_error(old_dir->i_sb, retval);
|
||||
|
@ -3043,23 +3228,19 @@ const struct inode_operations ext4_dir_inode_operations = {
|
|||
.mknod = ext4_mknod,
|
||||
.rename = ext4_rename,
|
||||
.setattr = ext4_setattr,
|
||||
#ifdef CONFIG_EXT4_FS_XATTR
|
||||
.setxattr = generic_setxattr,
|
||||
.getxattr = generic_getxattr,
|
||||
.listxattr = ext4_listxattr,
|
||||
.removexattr = generic_removexattr,
|
||||
#endif
|
||||
.get_acl = ext4_get_acl,
|
||||
.fiemap = ext4_fiemap,
|
||||
};
|
||||
|
||||
const struct inode_operations ext4_special_inode_operations = {
|
||||
.setattr = ext4_setattr,
|
||||
#ifdef CONFIG_EXT4_FS_XATTR
|
||||
.setxattr = generic_setxattr,
|
||||
.getxattr = generic_getxattr,
|
||||
.listxattr = ext4_listxattr,
|
||||
.removexattr = generic_removexattr,
|
||||
#endif
|
||||
.get_acl = ext4_get_acl,
|
||||
};
|
||||
|
|
|
@ -27,7 +27,6 @@
|
|||
#include "ext4_jbd2.h"
|
||||
#include "xattr.h"
|
||||
#include "acl.h"
|
||||
#include "ext4_extents.h"
|
||||
|
||||
static struct kmem_cache *io_page_cachep, *io_end_cachep;
|
||||
|
||||
|
@ -111,7 +110,7 @@ static int ext4_end_io(ext4_io_end_t *io)
|
|||
inode_dio_done(inode);
|
||||
/* Wake up anyone waiting on unwritten extent conversion */
|
||||
if (atomic_dec_and_test(&EXT4_I(inode)->i_unwritten))
|
||||
wake_up_all(ext4_ioend_wq(io->inode));
|
||||
wake_up_all(ext4_ioend_wq(inode));
|
||||
return ret;
|
||||
}
|
||||
|
||||
|
|
|
@ -783,7 +783,7 @@ static int add_new_gdb(handle_t *handle, struct inode *inode,
|
|||
|
||||
err = ext4_journal_get_write_access(handle, gdb_bh);
|
||||
if (unlikely(err))
|
||||
goto exit_sbh;
|
||||
goto exit_dind;
|
||||
|
||||
err = ext4_journal_get_write_access(handle, dind);
|
||||
if (unlikely(err))
|
||||
|
@ -792,7 +792,7 @@ static int add_new_gdb(handle_t *handle, struct inode *inode,
|
|||
/* ext4_reserve_inode_write() gets a reference on the iloc */
|
||||
err = ext4_reserve_inode_write(handle, inode, &iloc);
|
||||
if (unlikely(err))
|
||||
goto exit_dindj;
|
||||
goto exit_dind;
|
||||
|
||||
n_group_desc = ext4_kvmalloc((gdb_num + 1) *
|
||||
sizeof(struct buffer_head *),
|
||||
|
@ -846,12 +846,7 @@ static int add_new_gdb(handle_t *handle, struct inode *inode,
|
|||
|
||||
exit_inode:
|
||||
ext4_kvfree(n_group_desc);
|
||||
/* ext4_handle_release_buffer(handle, iloc.bh); */
|
||||
brelse(iloc.bh);
|
||||
exit_dindj:
|
||||
/* ext4_handle_release_buffer(handle, dind); */
|
||||
exit_sbh:
|
||||
/* ext4_handle_release_buffer(handle, EXT4_SB(sb)->s_sbh); */
|
||||
exit_dind:
|
||||
brelse(dind);
|
||||
exit_bh:
|
||||
|
@ -969,14 +964,8 @@ static int reserve_backup_gdb(handle_t *handle, struct inode *inode,
|
|||
}
|
||||
|
||||
for (i = 0; i < reserved_gdb; i++) {
|
||||
if ((err = ext4_journal_get_write_access(handle, primary[i]))) {
|
||||
/*
|
||||
int j;
|
||||
for (j = 0; j < i; j++)
|
||||
ext4_handle_release_buffer(handle, primary[j]);
|
||||
*/
|
||||
if ((err = ext4_journal_get_write_access(handle, primary[i])))
|
||||
goto exit_bh;
|
||||
}
|
||||
}
|
||||
|
||||
if ((err = ext4_reserve_inode_write(handle, inode, &iloc)))
|
||||
|
|
|
@ -45,7 +45,7 @@
|
|||
#include <linux/freezer.h>
|
||||
|
||||
#include "ext4.h"
|
||||
#include "ext4_extents.h"
|
||||
#include "ext4_extents.h" /* Needed for trace points definition */
|
||||
#include "ext4_jbd2.h"
|
||||
#include "xattr.h"
|
||||
#include "acl.h"
|
||||
|
@ -939,10 +939,11 @@ static struct inode *ext4_alloc_inode(struct super_block *sb)
|
|||
return NULL;
|
||||
|
||||
ei->vfs_inode.i_version = 1;
|
||||
ei->vfs_inode.i_data.writeback_index = 0;
|
||||
memset(&ei->i_cached_extent, 0, sizeof(struct ext4_ext_cache));
|
||||
INIT_LIST_HEAD(&ei->i_prealloc_list);
|
||||
spin_lock_init(&ei->i_prealloc_lock);
|
||||
ext4_es_init_tree(&ei->i_es_tree);
|
||||
rwlock_init(&ei->i_es_lock);
|
||||
ei->i_reserved_data_blocks = 0;
|
||||
ei->i_reserved_meta_blocks = 0;
|
||||
ei->i_allocated_meta_blocks = 0;
|
||||
|
@ -996,9 +997,7 @@ static void init_once(void *foo)
|
|||
struct ext4_inode_info *ei = (struct ext4_inode_info *) foo;
|
||||
|
||||
INIT_LIST_HEAD(&ei->i_orphan);
|
||||
#ifdef CONFIG_EXT4_FS_XATTR
|
||||
init_rwsem(&ei->xattr_sem);
|
||||
#endif
|
||||
init_rwsem(&ei->i_data_sem);
|
||||
inode_init_once(&ei->vfs_inode);
|
||||
}
|
||||
|
@ -1031,6 +1030,7 @@ void ext4_clear_inode(struct inode *inode)
|
|||
clear_inode(inode);
|
||||
dquot_drop(inode);
|
||||
ext4_discard_preallocations(inode);
|
||||
ext4_es_remove_extent(inode, 0, EXT_MAX_BLOCKS);
|
||||
if (EXT4_I(inode)->jinode) {
|
||||
jbd2_journal_release_jbd_inode(EXT4_JOURNAL(inode),
|
||||
EXT4_I(inode)->jinode);
|
||||
|
@ -1447,13 +1447,8 @@ static const struct mount_opts {
|
|||
{Opt_data_journal, EXT4_MOUNT_JOURNAL_DATA, MOPT_DATAJ},
|
||||
{Opt_data_ordered, EXT4_MOUNT_ORDERED_DATA, MOPT_DATAJ},
|
||||
{Opt_data_writeback, EXT4_MOUNT_WRITEBACK_DATA, MOPT_DATAJ},
|
||||
#ifdef CONFIG_EXT4_FS_XATTR
|
||||
{Opt_user_xattr, EXT4_MOUNT_XATTR_USER, MOPT_SET},
|
||||
{Opt_nouser_xattr, EXT4_MOUNT_XATTR_USER, MOPT_CLEAR},
|
||||
#else
|
||||
{Opt_user_xattr, 0, MOPT_NOSUPPORT},
|
||||
{Opt_nouser_xattr, 0, MOPT_NOSUPPORT},
|
||||
#endif
|
||||
#ifdef CONFIG_EXT4_FS_POSIX_ACL
|
||||
{Opt_acl, EXT4_MOUNT_POSIX_ACL, MOPT_SET},
|
||||
{Opt_noacl, EXT4_MOUNT_POSIX_ACL, MOPT_CLEAR},
|
||||
|
@ -3202,7 +3197,6 @@ int ext4_calculate_overhead(struct super_block *sb)
|
|||
ext4_fsblk_t overhead = 0;
|
||||
char *buf = (char *) get_zeroed_page(GFP_KERNEL);
|
||||
|
||||
memset(buf, 0, PAGE_SIZE);
|
||||
if (!buf)
|
||||
return -ENOMEM;
|
||||
|
||||
|
@ -3256,7 +3250,7 @@ static int ext4_fill_super(struct super_block *sb, void *data, int silent)
|
|||
unsigned int i;
|
||||
int needs_recovery, has_huge_files, has_bigalloc;
|
||||
__u64 blocks_count;
|
||||
int err;
|
||||
int err = 0;
|
||||
unsigned int journal_ioprio = DEFAULT_JOURNAL_IOPRIO;
|
||||
ext4_group_t first_not_zeroed;
|
||||
|
||||
|
@ -3272,9 +3266,6 @@ static int ext4_fill_super(struct super_block *sb, void *data, int silent)
|
|||
}
|
||||
sb->s_fs_info = sbi;
|
||||
sbi->s_sb = sb;
|
||||
sbi->s_mount_opt = 0;
|
||||
sbi->s_resuid = make_kuid(&init_user_ns, EXT4_DEF_RESUID);
|
||||
sbi->s_resgid = make_kgid(&init_user_ns, EXT4_DEF_RESGID);
|
||||
sbi->s_inode_readahead_blks = EXT4_DEF_INODE_READAHEAD_BLKS;
|
||||
sbi->s_sb_block = sb_block;
|
||||
if (sb->s_bdev->bd_part)
|
||||
|
@ -3285,6 +3276,7 @@ static int ext4_fill_super(struct super_block *sb, void *data, int silent)
|
|||
for (cp = sb->s_id; (cp = strchr(cp, '/'));)
|
||||
*cp = '!';
|
||||
|
||||
/* -EINVAL is default */
|
||||
ret = -EINVAL;
|
||||
blocksize = sb_min_blocksize(sb, EXT4_MIN_BLOCK_SIZE);
|
||||
if (!blocksize) {
|
||||
|
@ -3369,9 +3361,7 @@ static int ext4_fill_super(struct super_block *sb, void *data, int silent)
|
|||
if (def_mount_opts & EXT4_DEFM_UID16)
|
||||
set_opt(sb, NO_UID32);
|
||||
/* xattr user namespace & acls are now defaulted on */
|
||||
#ifdef CONFIG_EXT4_FS_XATTR
|
||||
set_opt(sb, XATTR_USER);
|
||||
#endif
|
||||
#ifdef CONFIG_EXT4_FS_POSIX_ACL
|
||||
set_opt(sb, POSIX_ACL);
|
||||
#endif
|
||||
|
@ -3662,7 +3652,6 @@ static int ext4_fill_super(struct super_block *sb, void *data, int silent)
|
|||
" too large to mount safely on this system");
|
||||
if (sizeof(sector_t) < 8)
|
||||
ext4_msg(sb, KERN_WARNING, "CONFIG_LBDAF not enabled");
|
||||
ret = err;
|
||||
goto failed_mount;
|
||||
}
|
||||
|
||||
|
@ -3770,7 +3759,6 @@ static int ext4_fill_super(struct super_block *sb, void *data, int silent)
|
|||
}
|
||||
if (err) {
|
||||
ext4_msg(sb, KERN_ERR, "insufficient memory");
|
||||
ret = err;
|
||||
goto failed_mount3;
|
||||
}
|
||||
|
||||
|
@ -3801,7 +3789,6 @@ static int ext4_fill_super(struct super_block *sb, void *data, int silent)
|
|||
|
||||
INIT_LIST_HEAD(&sbi->s_orphan); /* unlinked but open files */
|
||||
mutex_init(&sbi->s_orphan_lock);
|
||||
sbi->s_resize_flags = 0;
|
||||
|
||||
sb->s_root = NULL;
|
||||
|
||||
|
@ -3897,8 +3884,8 @@ no_journal:
|
|||
if (es->s_overhead_clusters)
|
||||
sbi->s_overhead = le32_to_cpu(es->s_overhead_clusters);
|
||||
else {
|
||||
ret = ext4_calculate_overhead(sb);
|
||||
if (ret)
|
||||
err = ext4_calculate_overhead(sb);
|
||||
if (err)
|
||||
goto failed_mount_wq;
|
||||
}
|
||||
|
||||
|
@ -3910,6 +3897,7 @@ no_journal:
|
|||
alloc_workqueue("ext4-dio-unwritten", WQ_MEM_RECLAIM | WQ_UNBOUND, 1);
|
||||
if (!EXT4_SB(sb)->dio_unwritten_wq) {
|
||||
printk(KERN_ERR "EXT4-fs: failed to create DIO workqueue\n");
|
||||
ret = -ENOMEM;
|
||||
goto failed_mount_wq;
|
||||
}
|
||||
|
||||
|
@ -4012,12 +4000,20 @@ no_journal:
|
|||
/* Enable quota usage during mount. */
|
||||
if (EXT4_HAS_RO_COMPAT_FEATURE(sb, EXT4_FEATURE_RO_COMPAT_QUOTA) &&
|
||||
!(sb->s_flags & MS_RDONLY)) {
|
||||
ret = ext4_enable_quotas(sb);
|
||||
if (ret)
|
||||
err = ext4_enable_quotas(sb);
|
||||
if (err)
|
||||
goto failed_mount7;
|
||||
}
|
||||
#endif /* CONFIG_QUOTA */
|
||||
|
||||
if (test_opt(sb, DISCARD)) {
|
||||
struct request_queue *q = bdev_get_queue(sb->s_bdev);
|
||||
if (!blk_queue_discard(q))
|
||||
ext4_msg(sb, KERN_WARNING,
|
||||
"mounting with \"discard\" option, but "
|
||||
"the device does not support discard");
|
||||
}
|
||||
|
||||
ext4_msg(sb, KERN_INFO, "mounted filesystem with%s. "
|
||||
"Opts: %s%s%s", descr, sbi->s_es->s_mount_opts,
|
||||
*sbi->s_es->s_mount_opts ? "; " : "", orig_data);
|
||||
|
@ -4084,7 +4080,7 @@ out_fail:
|
|||
kfree(sbi);
|
||||
out_free_orig:
|
||||
kfree(orig_data);
|
||||
return ret;
|
||||
return err ? err : ret;
|
||||
}
|
||||
|
||||
/*
|
||||
|
@ -4790,7 +4786,7 @@ static int ext4_statfs(struct dentry *dentry, struct kstatfs *buf)
|
|||
|
||||
buf->f_type = EXT4_SUPER_MAGIC;
|
||||
buf->f_bsize = sb->s_blocksize;
|
||||
buf->f_blocks = ext4_blocks_count(es) - EXT4_C2B(sbi, sbi->s_overhead);
|
||||
buf->f_blocks = ext4_blocks_count(es) - EXT4_C2B(sbi, overhead);
|
||||
bfree = percpu_counter_sum_positive(&sbi->s_freeclusters_counter) -
|
||||
percpu_counter_sum_positive(&sbi->s_dirtyclusters_counter);
|
||||
/* prevent underflow in case that few free space is available */
|
||||
|
@ -5282,6 +5278,7 @@ static int __init ext4_init_fs(void)
|
|||
ext4_li_info = NULL;
|
||||
mutex_init(&ext4_li_mtx);
|
||||
|
||||
/* Build-time check for flags consistency */
|
||||
ext4_check_flag_values();
|
||||
|
||||
for (i = 0; i < EXT4_WQ_HASH_SZ; i++) {
|
||||
|
@ -5289,9 +5286,14 @@ static int __init ext4_init_fs(void)
|
|||
init_waitqueue_head(&ext4__ioend_wq[i]);
|
||||
}
|
||||
|
||||
err = ext4_init_pageio();
|
||||
err = ext4_init_es();
|
||||
if (err)
|
||||
return err;
|
||||
|
||||
err = ext4_init_pageio();
|
||||
if (err)
|
||||
goto out7;
|
||||
|
||||
err = ext4_init_system_zone();
|
||||
if (err)
|
||||
goto out6;
|
||||
|
@ -5341,6 +5343,9 @@ out5:
|
|||
ext4_exit_system_zone();
|
||||
out6:
|
||||
ext4_exit_pageio();
|
||||
out7:
|
||||
ext4_exit_es();
|
||||
|
||||
return err;
|
||||
}
|
||||
|
||||
|
|
|
@ -35,22 +35,18 @@ const struct inode_operations ext4_symlink_inode_operations = {
|
|||
.follow_link = page_follow_link_light,
|
||||
.put_link = page_put_link,
|
||||
.setattr = ext4_setattr,
|
||||
#ifdef CONFIG_EXT4_FS_XATTR
|
||||
.setxattr = generic_setxattr,
|
||||
.getxattr = generic_getxattr,
|
||||
.listxattr = ext4_listxattr,
|
||||
.removexattr = generic_removexattr,
|
||||
#endif
|
||||
};
|
||||
|
||||
const struct inode_operations ext4_fast_symlink_inode_operations = {
|
||||
.readlink = generic_readlink,
|
||||
.follow_link = ext4_follow_link,
|
||||
.setattr = ext4_setattr,
|
||||
#ifdef CONFIG_EXT4_FS_XATTR
|
||||
.setxattr = generic_setxattr,
|
||||
.getxattr = generic_getxattr,
|
||||
.listxattr = ext4_listxattr,
|
||||
.removexattr = generic_removexattr,
|
||||
#endif
|
||||
};
|
||||
|
|
110
fs/ext4/xattr.c
110
fs/ext4/xattr.c
|
@ -61,11 +61,6 @@
|
|||
#include "xattr.h"
|
||||
#include "acl.h"
|
||||
|
||||
#define BHDR(bh) ((struct ext4_xattr_header *)((bh)->b_data))
|
||||
#define ENTRY(ptr) ((struct ext4_xattr_entry *)(ptr))
|
||||
#define BFIRST(bh) ENTRY(BHDR(bh)+1)
|
||||
#define IS_LAST_ENTRY(entry) (*(__u32 *)(entry) == 0)
|
||||
|
||||
#ifdef EXT4_XATTR_DEBUG
|
||||
# define ea_idebug(inode, f...) do { \
|
||||
printk(KERN_DEBUG "inode %s:%lu: ", \
|
||||
|
@ -312,7 +307,7 @@ cleanup:
|
|||
return error;
|
||||
}
|
||||
|
||||
static int
|
||||
int
|
||||
ext4_xattr_ibody_get(struct inode *inode, int name_index, const char *name,
|
||||
void *buffer, size_t buffer_size)
|
||||
{
|
||||
|
@ -581,21 +576,6 @@ static size_t ext4_xattr_free_space(struct ext4_xattr_entry *last,
|
|||
return (*min_offs - ((void *)last - base) - sizeof(__u32));
|
||||
}
|
||||
|
||||
struct ext4_xattr_info {
|
||||
int name_index;
|
||||
const char *name;
|
||||
const void *value;
|
||||
size_t value_len;
|
||||
};
|
||||
|
||||
struct ext4_xattr_search {
|
||||
struct ext4_xattr_entry *first;
|
||||
void *base;
|
||||
void *end;
|
||||
struct ext4_xattr_entry *here;
|
||||
int not_found;
|
||||
};
|
||||
|
||||
static int
|
||||
ext4_xattr_set_entry(struct ext4_xattr_info *i, struct ext4_xattr_search *s)
|
||||
{
|
||||
|
@ -648,9 +628,14 @@ ext4_xattr_set_entry(struct ext4_xattr_info *i, struct ext4_xattr_search *s)
|
|||
size. Just replace. */
|
||||
s->here->e_value_size =
|
||||
cpu_to_le32(i->value_len);
|
||||
memset(val + size - EXT4_XATTR_PAD, 0,
|
||||
EXT4_XATTR_PAD); /* Clear pad bytes. */
|
||||
memcpy(val, i->value, i->value_len);
|
||||
if (i->value == EXT4_ZERO_XATTR_VALUE) {
|
||||
memset(val, 0, size);
|
||||
} else {
|
||||
/* Clear pad bytes first. */
|
||||
memset(val + size - EXT4_XATTR_PAD, 0,
|
||||
EXT4_XATTR_PAD);
|
||||
memcpy(val, i->value, i->value_len);
|
||||
}
|
||||
return 0;
|
||||
}
|
||||
|
||||
|
@ -689,9 +674,14 @@ ext4_xattr_set_entry(struct ext4_xattr_info *i, struct ext4_xattr_search *s)
|
|||
size_t size = EXT4_XATTR_SIZE(i->value_len);
|
||||
void *val = s->base + min_offs - size;
|
||||
s->here->e_value_offs = cpu_to_le16(min_offs - size);
|
||||
memset(val + size - EXT4_XATTR_PAD, 0,
|
||||
EXT4_XATTR_PAD); /* Clear the pad bytes. */
|
||||
memcpy(val, i->value, i->value_len);
|
||||
if (i->value == EXT4_ZERO_XATTR_VALUE) {
|
||||
memset(val, 0, size);
|
||||
} else {
|
||||
/* Clear the pad bytes first. */
|
||||
memset(val + size - EXT4_XATTR_PAD, 0,
|
||||
EXT4_XATTR_PAD);
|
||||
memcpy(val, i->value, i->value_len);
|
||||
}
|
||||
}
|
||||
}
|
||||
return 0;
|
||||
|
@ -794,7 +784,6 @@ ext4_xattr_block_set(handle_t *handle, struct inode *inode,
|
|||
int offset = (char *)s->here - bs->bh->b_data;
|
||||
|
||||
unlock_buffer(bs->bh);
|
||||
ext4_handle_release_buffer(handle, bs->bh);
|
||||
if (ce) {
|
||||
mb_cache_entry_release(ce);
|
||||
ce = NULL;
|
||||
|
@ -950,14 +939,8 @@ bad_block:
|
|||
#undef header
|
||||
}
|
||||
|
||||
struct ext4_xattr_ibody_find {
|
||||
struct ext4_xattr_search s;
|
||||
struct ext4_iloc iloc;
|
||||
};
|
||||
|
||||
static int
|
||||
ext4_xattr_ibody_find(struct inode *inode, struct ext4_xattr_info *i,
|
||||
struct ext4_xattr_ibody_find *is)
|
||||
int ext4_xattr_ibody_find(struct inode *inode, struct ext4_xattr_info *i,
|
||||
struct ext4_xattr_ibody_find *is)
|
||||
{
|
||||
struct ext4_xattr_ibody_header *header;
|
||||
struct ext4_inode *raw_inode;
|
||||
|
@ -985,10 +968,47 @@ ext4_xattr_ibody_find(struct inode *inode, struct ext4_xattr_info *i,
|
|||
return 0;
|
||||
}
|
||||
|
||||
static int
|
||||
ext4_xattr_ibody_set(handle_t *handle, struct inode *inode,
|
||||
struct ext4_xattr_info *i,
|
||||
struct ext4_xattr_ibody_find *is)
|
||||
int ext4_xattr_ibody_inline_set(handle_t *handle, struct inode *inode,
|
||||
struct ext4_xattr_info *i,
|
||||
struct ext4_xattr_ibody_find *is)
|
||||
{
|
||||
struct ext4_xattr_ibody_header *header;
|
||||
struct ext4_xattr_search *s = &is->s;
|
||||
int error;
|
||||
|
||||
if (EXT4_I(inode)->i_extra_isize == 0)
|
||||
return -ENOSPC;
|
||||
error = ext4_xattr_set_entry(i, s);
|
||||
if (error) {
|
||||
if (error == -ENOSPC &&
|
||||
ext4_has_inline_data(inode)) {
|
||||
error = ext4_try_to_evict_inline_data(handle, inode,
|
||||
EXT4_XATTR_LEN(strlen(i->name) +
|
||||
EXT4_XATTR_SIZE(i->value_len)));
|
||||
if (error)
|
||||
return error;
|
||||
error = ext4_xattr_ibody_find(inode, i, is);
|
||||
if (error)
|
||||
return error;
|
||||
error = ext4_xattr_set_entry(i, s);
|
||||
}
|
||||
if (error)
|
||||
return error;
|
||||
}
|
||||
header = IHDR(inode, ext4_raw_inode(&is->iloc));
|
||||
if (!IS_LAST_ENTRY(s->first)) {
|
||||
header->h_magic = cpu_to_le32(EXT4_XATTR_MAGIC);
|
||||
ext4_set_inode_state(inode, EXT4_STATE_XATTR);
|
||||
} else {
|
||||
header->h_magic = cpu_to_le32(0);
|
||||
ext4_clear_inode_state(inode, EXT4_STATE_XATTR);
|
||||
}
|
||||
return 0;
|
||||
}
|
||||
|
||||
static int ext4_xattr_ibody_set(handle_t *handle, struct inode *inode,
|
||||
struct ext4_xattr_info *i,
|
||||
struct ext4_xattr_ibody_find *is)
|
||||
{
|
||||
struct ext4_xattr_ibody_header *header;
|
||||
struct ext4_xattr_search *s = &is->s;
|
||||
|
@ -1144,9 +1164,17 @@ ext4_xattr_set(struct inode *inode, int name_index, const char *name,
|
|||
{
|
||||
handle_t *handle;
|
||||
int error, retries = 0;
|
||||
int credits = EXT4_DATA_TRANS_BLOCKS(inode->i_sb);
|
||||
|
||||
retry:
|
||||
handle = ext4_journal_start(inode, EXT4_DATA_TRANS_BLOCKS(inode->i_sb));
|
||||
/*
|
||||
* In case of inline data, we may push out the data to a block,
|
||||
* So reserve the journal space first.
|
||||
*/
|
||||
if (ext4_has_inline_data(inode))
|
||||
credits += ext4_writepage_trans_blocks(inode) + 1;
|
||||
|
||||
handle = ext4_journal_start(inode, credits);
|
||||
if (IS_ERR(handle)) {
|
||||
error = PTR_ERR(handle);
|
||||
} else {
|
||||
|
|
152
fs/ext4/xattr.h
152
fs/ext4/xattr.h
|
@ -21,6 +21,7 @@
|
|||
#define EXT4_XATTR_INDEX_TRUSTED 4
|
||||
#define EXT4_XATTR_INDEX_LUSTRE 5
|
||||
#define EXT4_XATTR_INDEX_SECURITY 6
|
||||
#define EXT4_XATTR_INDEX_SYSTEM 7
|
||||
|
||||
struct ext4_xattr_header {
|
||||
__le32 h_magic; /* magic number for identification */
|
||||
|
@ -65,7 +66,32 @@ struct ext4_xattr_entry {
|
|||
EXT4_I(inode)->i_extra_isize))
|
||||
#define IFIRST(hdr) ((struct ext4_xattr_entry *)((hdr)+1))
|
||||
|
||||
# ifdef CONFIG_EXT4_FS_XATTR
|
||||
#define BHDR(bh) ((struct ext4_xattr_header *)((bh)->b_data))
|
||||
#define ENTRY(ptr) ((struct ext4_xattr_entry *)(ptr))
|
||||
#define BFIRST(bh) ENTRY(BHDR(bh)+1)
|
||||
#define IS_LAST_ENTRY(entry) (*(__u32 *)(entry) == 0)
|
||||
|
||||
#define EXT4_ZERO_XATTR_VALUE ((void *)-1)
|
||||
|
||||
struct ext4_xattr_info {
|
||||
int name_index;
|
||||
const char *name;
|
||||
const void *value;
|
||||
size_t value_len;
|
||||
};
|
||||
|
||||
struct ext4_xattr_search {
|
||||
struct ext4_xattr_entry *first;
|
||||
void *base;
|
||||
void *end;
|
||||
struct ext4_xattr_entry *here;
|
||||
int not_found;
|
||||
};
|
||||
|
||||
struct ext4_xattr_ibody_find {
|
||||
struct ext4_xattr_search s;
|
||||
struct ext4_iloc iloc;
|
||||
};
|
||||
|
||||
extern const struct xattr_handler ext4_xattr_user_handler;
|
||||
extern const struct xattr_handler ext4_xattr_trusted_handler;
|
||||
|
@ -90,60 +116,82 @@ extern void ext4_exit_xattr(void);
|
|||
|
||||
extern const struct xattr_handler *ext4_xattr_handlers[];
|
||||
|
||||
# else /* CONFIG_EXT4_FS_XATTR */
|
||||
extern int ext4_xattr_ibody_find(struct inode *inode, struct ext4_xattr_info *i,
|
||||
struct ext4_xattr_ibody_find *is);
|
||||
extern int ext4_xattr_ibody_get(struct inode *inode, int name_index,
|
||||
const char *name,
|
||||
void *buffer, size_t buffer_size);
|
||||
extern int ext4_xattr_ibody_inline_set(handle_t *handle, struct inode *inode,
|
||||
struct ext4_xattr_info *i,
|
||||
struct ext4_xattr_ibody_find *is);
|
||||
|
||||
static inline int
|
||||
ext4_xattr_get(struct inode *inode, int name_index, const char *name,
|
||||
void *buffer, size_t size, int flags)
|
||||
{
|
||||
return -EOPNOTSUPP;
|
||||
}
|
||||
extern int ext4_has_inline_data(struct inode *inode);
|
||||
extern int ext4_get_inline_size(struct inode *inode);
|
||||
extern int ext4_get_max_inline_size(struct inode *inode);
|
||||
extern int ext4_find_inline_data_nolock(struct inode *inode);
|
||||
extern void ext4_write_inline_data(struct inode *inode,
|
||||
struct ext4_iloc *iloc,
|
||||
void *buffer, loff_t pos,
|
||||
unsigned int len);
|
||||
extern int ext4_prepare_inline_data(handle_t *handle, struct inode *inode,
|
||||
unsigned int len);
|
||||
extern int ext4_init_inline_data(handle_t *handle, struct inode *inode,
|
||||
unsigned int len);
|
||||
extern int ext4_destroy_inline_data(handle_t *handle, struct inode *inode);
|
||||
|
||||
static inline int
|
||||
ext4_xattr_set(struct inode *inode, int name_index, const char *name,
|
||||
const void *value, size_t size, int flags)
|
||||
{
|
||||
return -EOPNOTSUPP;
|
||||
}
|
||||
extern int ext4_readpage_inline(struct inode *inode, struct page *page);
|
||||
extern int ext4_try_to_write_inline_data(struct address_space *mapping,
|
||||
struct inode *inode,
|
||||
loff_t pos, unsigned len,
|
||||
unsigned flags,
|
||||
struct page **pagep);
|
||||
extern int ext4_write_inline_data_end(struct inode *inode,
|
||||
loff_t pos, unsigned len,
|
||||
unsigned copied,
|
||||
struct page *page);
|
||||
extern struct buffer_head *
|
||||
ext4_journalled_write_inline_data(struct inode *inode,
|
||||
unsigned len,
|
||||
struct page *page);
|
||||
extern int ext4_da_write_inline_data_begin(struct address_space *mapping,
|
||||
struct inode *inode,
|
||||
loff_t pos, unsigned len,
|
||||
unsigned flags,
|
||||
struct page **pagep,
|
||||
void **fsdata);
|
||||
extern int ext4_da_write_inline_data_end(struct inode *inode, loff_t pos,
|
||||
unsigned len, unsigned copied,
|
||||
struct page *page);
|
||||
extern int ext4_try_add_inline_entry(handle_t *handle, struct dentry *dentry,
|
||||
struct inode *inode);
|
||||
extern int ext4_try_create_inline_dir(handle_t *handle,
|
||||
struct inode *parent,
|
||||
struct inode *inode);
|
||||
extern int ext4_read_inline_dir(struct file *filp,
|
||||
void *dirent, filldir_t filldir,
|
||||
int *has_inline_data);
|
||||
extern struct buffer_head *ext4_find_inline_entry(struct inode *dir,
|
||||
const struct qstr *d_name,
|
||||
struct ext4_dir_entry_2 **res_dir,
|
||||
int *has_inline_data);
|
||||
extern int ext4_delete_inline_entry(handle_t *handle,
|
||||
struct inode *dir,
|
||||
struct ext4_dir_entry_2 *de_del,
|
||||
struct buffer_head *bh,
|
||||
int *has_inline_data);
|
||||
extern int empty_inline_dir(struct inode *dir, int *has_inline_data);
|
||||
extern struct buffer_head *ext4_get_first_inline_block(struct inode *inode,
|
||||
struct ext4_dir_entry_2 **parent_de,
|
||||
int *retval);
|
||||
extern int ext4_inline_data_fiemap(struct inode *inode,
|
||||
struct fiemap_extent_info *fieinfo,
|
||||
int *has_inline);
|
||||
extern int ext4_try_to_evict_inline_data(handle_t *handle,
|
||||
struct inode *inode,
|
||||
int needed);
|
||||
extern void ext4_inline_data_truncate(struct inode *inode, int *has_inline);
|
||||
|
||||
static inline int
|
||||
ext4_xattr_set_handle(handle_t *handle, struct inode *inode, int name_index,
|
||||
const char *name, const void *value, size_t size, int flags)
|
||||
{
|
||||
return -EOPNOTSUPP;
|
||||
}
|
||||
|
||||
static inline void
|
||||
ext4_xattr_delete_inode(handle_t *handle, struct inode *inode)
|
||||
{
|
||||
}
|
||||
|
||||
static inline void
|
||||
ext4_xattr_put_super(struct super_block *sb)
|
||||
{
|
||||
}
|
||||
|
||||
static __init inline int
|
||||
ext4_init_xattr(void)
|
||||
{
|
||||
return 0;
|
||||
}
|
||||
|
||||
static inline void
|
||||
ext4_exit_xattr(void)
|
||||
{
|
||||
}
|
||||
|
||||
static inline int
|
||||
ext4_expand_extra_isize_ea(struct inode *inode, int new_extra_isize,
|
||||
struct ext4_inode *raw_inode, handle_t *handle)
|
||||
{
|
||||
return -EOPNOTSUPP;
|
||||
}
|
||||
|
||||
#define ext4_xattr_handlers NULL
|
||||
|
||||
# endif /* CONFIG_EXT4_FS_XATTR */
|
||||
extern int ext4_convert_inline_data(struct inode *inode);
|
||||
|
||||
#ifdef CONFIG_EXT4_FS_SECURITY
|
||||
extern int ext4_init_security(handle_t *handle, struct inode *inode,
|
||||
|
|
|
@ -60,7 +60,6 @@ EXPORT_SYMBOL(jbd2_journal_get_create_access);
|
|||
EXPORT_SYMBOL(jbd2_journal_get_undo_access);
|
||||
EXPORT_SYMBOL(jbd2_journal_set_triggers);
|
||||
EXPORT_SYMBOL(jbd2_journal_dirty_metadata);
|
||||
EXPORT_SYMBOL(jbd2_journal_release_buffer);
|
||||
EXPORT_SYMBOL(jbd2_journal_forget);
|
||||
#if 0
|
||||
EXPORT_SYMBOL(journal_sync_buffer);
|
||||
|
|
|
@ -1207,17 +1207,6 @@ out:
|
|||
return ret;
|
||||
}
|
||||
|
||||
/*
|
||||
* jbd2_journal_release_buffer: undo a get_write_access without any buffer
|
||||
* updates, if the update decided in the end that it didn't need access.
|
||||
*
|
||||
*/
|
||||
void
|
||||
jbd2_journal_release_buffer(handle_t *handle, struct buffer_head *bh)
|
||||
{
|
||||
BUFFER_TRACE(bh, "entry");
|
||||
}
|
||||
|
||||
/**
|
||||
* void jbd2_journal_forget() - bforget() for potentially-journaled buffers.
|
||||
* @handle: transaction handle
|
||||
|
|
|
@ -1096,7 +1096,6 @@ extern int jbd2_journal_get_undo_access(handle_t *, struct buffer_head *);
|
|||
void jbd2_journal_set_triggers(struct buffer_head *,
|
||||
struct jbd2_buffer_trigger_type *type);
|
||||
extern int jbd2_journal_dirty_metadata (handle_t *, struct buffer_head *);
|
||||
extern void jbd2_journal_release_buffer (handle_t *, struct buffer_head *);
|
||||
extern int jbd2_journal_forget (handle_t *, struct buffer_head *);
|
||||
extern void journal_sync_buffer (struct buffer_head *);
|
||||
extern void jbd2_journal_invalidatepage(journal_t *,
|
||||
|
@ -1303,15 +1302,21 @@ static inline int jbd_space_needed(journal_t *journal)
|
|||
|
||||
extern int jbd_blocks_per_page(struct inode *inode);
|
||||
|
||||
/* JBD uses a CRC32 checksum */
|
||||
#define JBD_MAX_CHECKSUM_SIZE 4
|
||||
|
||||
static inline u32 jbd2_chksum(journal_t *journal, u32 crc,
|
||||
const void *address, unsigned int length)
|
||||
{
|
||||
struct {
|
||||
struct shash_desc shash;
|
||||
char ctx[crypto_shash_descsize(journal->j_chksum_driver)];
|
||||
char ctx[JBD_MAX_CHECKSUM_SIZE];
|
||||
} desc;
|
||||
int err;
|
||||
|
||||
BUG_ON(crypto_shash_descsize(journal->j_chksum_driver) >
|
||||
JBD_MAX_CHECKSUM_SIZE);
|
||||
|
||||
desc.shash.tfm = journal->j_chksum_driver;
|
||||
desc.shash.flags = 0;
|
||||
*(u32 *)desc.ctx = crc;
|
||||
|
|
|
@ -15,6 +15,7 @@ struct ext4_inode_info;
|
|||
struct mpage_da_data;
|
||||
struct ext4_map_blocks;
|
||||
struct ext4_extent;
|
||||
struct extent_status;
|
||||
|
||||
#define EXT4_I(inode) (container_of(inode, struct ext4_inode_info, vfs_inode))
|
||||
|
||||
|
@ -1519,10 +1520,9 @@ DEFINE_EVENT(ext4__map_blocks_enter, ext4_ind_map_blocks_enter,
|
|||
);
|
||||
|
||||
DECLARE_EVENT_CLASS(ext4__map_blocks_exit,
|
||||
TP_PROTO(struct inode *inode, ext4_lblk_t lblk,
|
||||
ext4_fsblk_t pblk, unsigned int len, int ret),
|
||||
TP_PROTO(struct inode *inode, struct ext4_map_blocks *map, int ret),
|
||||
|
||||
TP_ARGS(inode, lblk, pblk, len, ret),
|
||||
TP_ARGS(inode, map, ret),
|
||||
|
||||
TP_STRUCT__entry(
|
||||
__field( dev_t, dev )
|
||||
|
@ -1530,37 +1530,37 @@ DECLARE_EVENT_CLASS(ext4__map_blocks_exit,
|
|||
__field( ext4_fsblk_t, pblk )
|
||||
__field( ext4_lblk_t, lblk )
|
||||
__field( unsigned int, len )
|
||||
__field( unsigned int, flags )
|
||||
__field( int, ret )
|
||||
),
|
||||
|
||||
TP_fast_assign(
|
||||
__entry->dev = inode->i_sb->s_dev;
|
||||
__entry->ino = inode->i_ino;
|
||||
__entry->pblk = pblk;
|
||||
__entry->lblk = lblk;
|
||||
__entry->len = len;
|
||||
__entry->pblk = map->m_pblk;
|
||||
__entry->lblk = map->m_lblk;
|
||||
__entry->len = map->m_len;
|
||||
__entry->flags = map->m_flags;
|
||||
__entry->ret = ret;
|
||||
),
|
||||
|
||||
TP_printk("dev %d,%d ino %lu lblk %u pblk %llu len %u ret %d",
|
||||
TP_printk("dev %d,%d ino %lu lblk %u pblk %llu len %u flags %x ret %d",
|
||||
MAJOR(__entry->dev), MINOR(__entry->dev),
|
||||
(unsigned long) __entry->ino,
|
||||
__entry->lblk, __entry->pblk,
|
||||
__entry->len, __entry->ret)
|
||||
__entry->len, __entry->flags, __entry->ret)
|
||||
);
|
||||
|
||||
DEFINE_EVENT(ext4__map_blocks_exit, ext4_ext_map_blocks_exit,
|
||||
TP_PROTO(struct inode *inode, ext4_lblk_t lblk,
|
||||
ext4_fsblk_t pblk, unsigned len, int ret),
|
||||
TP_PROTO(struct inode *inode, struct ext4_map_blocks *map, int ret),
|
||||
|
||||
TP_ARGS(inode, lblk, pblk, len, ret)
|
||||
TP_ARGS(inode, map, ret)
|
||||
);
|
||||
|
||||
DEFINE_EVENT(ext4__map_blocks_exit, ext4_ind_map_blocks_exit,
|
||||
TP_PROTO(struct inode *inode, ext4_lblk_t lblk,
|
||||
ext4_fsblk_t pblk, unsigned len, int ret),
|
||||
TP_PROTO(struct inode *inode, struct ext4_map_blocks *map, int ret),
|
||||
|
||||
TP_ARGS(inode, lblk, pblk, len, ret)
|
||||
TP_ARGS(inode, map, ret)
|
||||
);
|
||||
|
||||
TRACE_EVENT(ext4_ext_load_extent,
|
||||
|
@ -1680,10 +1680,10 @@ DEFINE_EVENT(ext4__trim, ext4_trim_all_free,
|
|||
);
|
||||
|
||||
TRACE_EVENT(ext4_ext_handle_uninitialized_extents,
|
||||
TP_PROTO(struct inode *inode, struct ext4_map_blocks *map,
|
||||
TP_PROTO(struct inode *inode, struct ext4_map_blocks *map, int flags,
|
||||
unsigned int allocated, ext4_fsblk_t newblock),
|
||||
|
||||
TP_ARGS(inode, map, allocated, newblock),
|
||||
TP_ARGS(inode, map, flags, allocated, newblock),
|
||||
|
||||
TP_STRUCT__entry(
|
||||
__field( dev_t, dev )
|
||||
|
@ -1699,7 +1699,7 @@ TRACE_EVENT(ext4_ext_handle_uninitialized_extents,
|
|||
TP_fast_assign(
|
||||
__entry->dev = inode->i_sb->s_dev;
|
||||
__entry->ino = inode->i_ino;
|
||||
__entry->flags = map->m_flags;
|
||||
__entry->flags = flags;
|
||||
__entry->lblk = map->m_lblk;
|
||||
__entry->pblk = map->m_pblk;
|
||||
__entry->len = map->m_len;
|
||||
|
@ -1707,7 +1707,7 @@ TRACE_EVENT(ext4_ext_handle_uninitialized_extents,
|
|||
__entry->newblk = newblock;
|
||||
),
|
||||
|
||||
TP_printk("dev %d,%d ino %lu m_lblk %u m_pblk %llu m_len %u flags %d"
|
||||
TP_printk("dev %d,%d ino %lu m_lblk %u m_pblk %llu m_len %u flags %x "
|
||||
"allocated %d newblock %llu",
|
||||
MAJOR(__entry->dev), MINOR(__entry->dev),
|
||||
(unsigned long) __entry->ino,
|
||||
|
@ -2055,6 +2055,106 @@ TRACE_EVENT(ext4_ext_remove_space_done,
|
|||
(unsigned short) __entry->eh_entries)
|
||||
);
|
||||
|
||||
TRACE_EVENT(ext4_es_insert_extent,
|
||||
TP_PROTO(struct inode *inode, ext4_lblk_t start, ext4_lblk_t len),
|
||||
|
||||
TP_ARGS(inode, start, len),
|
||||
|
||||
TP_STRUCT__entry(
|
||||
__field( dev_t, dev )
|
||||
__field( ino_t, ino )
|
||||
__field( loff_t, start )
|
||||
__field( loff_t, len )
|
||||
),
|
||||
|
||||
TP_fast_assign(
|
||||
__entry->dev = inode->i_sb->s_dev;
|
||||
__entry->ino = inode->i_ino;
|
||||
__entry->start = start;
|
||||
__entry->len = len;
|
||||
),
|
||||
|
||||
TP_printk("dev %d,%d ino %lu es [%lld/%lld)",
|
||||
MAJOR(__entry->dev), MINOR(__entry->dev),
|
||||
(unsigned long) __entry->ino,
|
||||
__entry->start, __entry->len)
|
||||
);
|
||||
|
||||
TRACE_EVENT(ext4_es_remove_extent,
|
||||
TP_PROTO(struct inode *inode, ext4_lblk_t start, ext4_lblk_t len),
|
||||
|
||||
TP_ARGS(inode, start, len),
|
||||
|
||||
TP_STRUCT__entry(
|
||||
__field( dev_t, dev )
|
||||
__field( ino_t, ino )
|
||||
__field( loff_t, start )
|
||||
__field( loff_t, len )
|
||||
),
|
||||
|
||||
TP_fast_assign(
|
||||
__entry->dev = inode->i_sb->s_dev;
|
||||
__entry->ino = inode->i_ino;
|
||||
__entry->start = start;
|
||||
__entry->len = len;
|
||||
),
|
||||
|
||||
TP_printk("dev %d,%d ino %lu es [%lld/%lld)",
|
||||
MAJOR(__entry->dev), MINOR(__entry->dev),
|
||||
(unsigned long) __entry->ino,
|
||||
__entry->start, __entry->len)
|
||||
);
|
||||
|
||||
TRACE_EVENT(ext4_es_find_extent_enter,
|
||||
TP_PROTO(struct inode *inode, ext4_lblk_t start),
|
||||
|
||||
TP_ARGS(inode, start),
|
||||
|
||||
TP_STRUCT__entry(
|
||||
__field( dev_t, dev )
|
||||
__field( ino_t, ino )
|
||||
__field( ext4_lblk_t, start )
|
||||
),
|
||||
|
||||
TP_fast_assign(
|
||||
__entry->dev = inode->i_sb->s_dev;
|
||||
__entry->ino = inode->i_ino;
|
||||
__entry->start = start;
|
||||
),
|
||||
|
||||
TP_printk("dev %d,%d ino %lu start %u",
|
||||
MAJOR(__entry->dev), MINOR(__entry->dev),
|
||||
(unsigned long) __entry->ino, __entry->start)
|
||||
);
|
||||
|
||||
TRACE_EVENT(ext4_es_find_extent_exit,
|
||||
TP_PROTO(struct inode *inode, struct extent_status *es,
|
||||
ext4_lblk_t ret),
|
||||
|
||||
TP_ARGS(inode, es, ret),
|
||||
|
||||
TP_STRUCT__entry(
|
||||
__field( dev_t, dev )
|
||||
__field( ino_t, ino )
|
||||
__field( ext4_lblk_t, start )
|
||||
__field( ext4_lblk_t, len )
|
||||
__field( ext4_lblk_t, ret )
|
||||
),
|
||||
|
||||
TP_fast_assign(
|
||||
__entry->dev = inode->i_sb->s_dev;
|
||||
__entry->ino = inode->i_ino;
|
||||
__entry->start = es->start;
|
||||
__entry->len = es->len;
|
||||
__entry->ret = ret;
|
||||
),
|
||||
|
||||
TP_printk("dev %d,%d ino %lu es [%u/%u) ret %u",
|
||||
MAJOR(__entry->dev), MINOR(__entry->dev),
|
||||
(unsigned long) __entry->ino,
|
||||
__entry->start, __entry->len, __entry->ret)
|
||||
);
|
||||
|
||||
#endif /* _TRACE_EXT4_H */
|
||||
|
||||
/* This part must be outside protection */
|
||||
|
|
Loading…
Reference in a new issue