2018-06-06 05:42:14 +03:00
|
|
|
// SPDX-License-Identifier: GPL-2.0
|
2005-04-17 02:20:36 +04:00
|
|
|
/*
|
2006-06-19 02:40:27 +04:00
|
|
|
* Copyright (c) 2000-2006 Silicon Graphics, Inc.
|
2005-11-02 06:58:39 +03:00
|
|
|
* All Rights Reserved.
|
2005-04-17 02:20:36 +04:00
|
|
|
*/
|
2009-12-15 02:14:59 +03:00
|
|
|
|
2005-04-17 02:20:36 +04:00
|
|
|
#include "xfs.h"
|
2013-10-23 03:36:05 +04:00
|
|
|
#include "xfs_shared.h"
|
2013-08-12 14:49:26 +04:00
|
|
|
#include "xfs_format.h"
|
2013-10-23 03:50:10 +04:00
|
|
|
#include "xfs_log_format.h"
|
|
|
|
#include "xfs_trans_resv.h"
|
2005-04-17 02:20:36 +04:00
|
|
|
#include "xfs_sb.h"
|
|
|
|
#include "xfs_mount.h"
|
|
|
|
#include "xfs_inode.h"
|
2005-11-02 06:38:42 +03:00
|
|
|
#include "xfs_btree.h"
|
2005-04-17 02:20:36 +04:00
|
|
|
#include "xfs_bmap.h"
|
2013-10-23 03:51:50 +04:00
|
|
|
#include "xfs_alloc.h"
|
2007-10-11 12:11:14 +04:00
|
|
|
#include "xfs_fsops.h"
|
2013-10-23 03:50:10 +04:00
|
|
|
#include "xfs_trans.h"
|
2005-04-17 02:20:36 +04:00
|
|
|
#include "xfs_buf_item.h"
|
2013-10-23 03:50:10 +04:00
|
|
|
#include "xfs_log.h"
|
2007-11-23 08:29:32 +03:00
|
|
|
#include "xfs_log_priv.h"
|
2013-08-12 14:49:37 +04:00
|
|
|
#include "xfs_dir2.h"
|
2008-07-18 11:11:46 +04:00
|
|
|
#include "xfs_extfree_item.h"
|
|
|
|
#include "xfs_mru_cache.h"
|
|
|
|
#include "xfs_inode_item.h"
|
2012-10-08 14:56:09 +04:00
|
|
|
#include "xfs_icache.h"
|
2009-12-15 02:14:59 +03:00
|
|
|
#include "xfs_trace.h"
|
2013-06-27 10:04:53 +04:00
|
|
|
#include "xfs_icreate_item.h"
|
2013-10-23 03:51:50 +04:00
|
|
|
#include "xfs_filestream.h"
|
|
|
|
#include "xfs_quota.h"
|
2014-09-09 05:52:42 +04:00
|
|
|
#include "xfs_sysfs.h"
|
2016-03-09 00:15:14 +03:00
|
|
|
#include "xfs_ondisk.h"
|
2016-08-03 05:04:45 +03:00
|
|
|
#include "xfs_rmap_item.h"
|
2016-10-03 19:11:20 +03:00
|
|
|
#include "xfs_refcount_item.h"
|
2016-10-03 19:11:25 +03:00
|
|
|
#include "xfs_bmap_item.h"
|
2016-10-03 19:11:38 +03:00
|
|
|
#include "xfs_reflink.h"
|
2005-04-17 02:20:36 +04:00
|
|
|
|
2018-10-18 09:20:19 +03:00
|
|
|
#include <linux/magic.h>
|
2008-07-18 11:12:36 +04:00
|
|
|
#include <linux/parser.h>
|
2005-04-17 02:20:36 +04:00
|
|
|
|
2009-09-22 04:01:09 +04:00
|
|
|
static const struct super_operations xfs_super_operations;
|
2014-09-09 05:52:42 +04:00
|
|
|
|
2014-09-29 04:46:08 +04:00
|
|
|
static struct kset *xfs_kset; /* top-level xfs sysfs dir */
|
2014-09-09 05:52:42 +04:00
|
|
|
#ifdef DEBUG
|
|
|
|
static struct xfs_kobj xfs_dbg_kobj; /* global debug sysfs attrs */
|
|
|
|
#endif
|
2005-04-17 02:20:36 +04:00
|
|
|
|
2008-07-18 11:12:36 +04:00
|
|
|
/*
|
|
|
|
* Table driven mount option parser.
|
|
|
|
*/
|
|
|
|
enum {
|
2019-10-28 18:41:42 +03:00
|
|
|
Opt_logbufs, Opt_logbsize, Opt_logdev, Opt_rtdev,
|
2016-03-02 01:55:38 +03:00
|
|
|
Opt_wsync, Opt_noalign, Opt_swalloc, Opt_sunit, Opt_swidth, Opt_nouuid,
|
2019-04-28 18:32:52 +03:00
|
|
|
Opt_grpid, Opt_nogrpid, Opt_bsdgroups, Opt_sysvgroups,
|
2018-07-26 19:11:27 +03:00
|
|
|
Opt_allocsize, Opt_norecovery, Opt_inode64, Opt_inode32, Opt_ikeep,
|
|
|
|
Opt_noikeep, Opt_largeio, Opt_nolargeio, Opt_attr2, Opt_noattr2,
|
|
|
|
Opt_filestreams, Opt_quota, Opt_noquota, Opt_usrquota, Opt_grpquota,
|
|
|
|
Opt_prjquota, Opt_uquota, Opt_gquota, Opt_pquota,
|
2016-03-02 01:55:38 +03:00
|
|
|
Opt_uqnoenforce, Opt_gqnoenforce, Opt_pqnoenforce, Opt_qnoenforce,
|
|
|
|
Opt_discard, Opt_nodiscard, Opt_dax, Opt_err,
|
2008-07-18 11:12:36 +04:00
|
|
|
};
|
|
|
|
|
2008-10-13 13:46:57 +04:00
|
|
|
static const match_table_t tokens = {
|
2016-03-02 01:55:38 +03:00
|
|
|
{Opt_logbufs, "logbufs=%u"}, /* number of XFS log buffers */
|
|
|
|
{Opt_logbsize, "logbsize=%s"}, /* size of XFS log buffers */
|
|
|
|
{Opt_logdev, "logdev=%s"}, /* log device */
|
|
|
|
{Opt_rtdev, "rtdev=%s"}, /* realtime I/O device */
|
|
|
|
{Opt_wsync, "wsync"}, /* safe-mode nfs compatible mount */
|
|
|
|
{Opt_noalign, "noalign"}, /* turn off stripe alignment */
|
|
|
|
{Opt_swalloc, "swalloc"}, /* turn on stripe width allocation */
|
|
|
|
{Opt_sunit, "sunit=%u"}, /* data volume stripe unit */
|
|
|
|
{Opt_swidth, "swidth=%u"}, /* data volume stripe width */
|
|
|
|
{Opt_nouuid, "nouuid"}, /* ignore filesystem UUID */
|
|
|
|
{Opt_grpid, "grpid"}, /* group-ID from parent directory */
|
|
|
|
{Opt_nogrpid, "nogrpid"}, /* group-ID from current process */
|
|
|
|
{Opt_bsdgroups, "bsdgroups"}, /* group-ID from parent directory */
|
|
|
|
{Opt_sysvgroups,"sysvgroups"}, /* group-ID from current process */
|
|
|
|
{Opt_allocsize, "allocsize=%s"},/* preferred allocation size */
|
|
|
|
{Opt_norecovery,"norecovery"}, /* don't run XFS recovery */
|
|
|
|
{Opt_inode64, "inode64"}, /* inodes can be allocated anywhere */
|
|
|
|
{Opt_inode32, "inode32"}, /* inode allocation limited to
|
|
|
|
* XFS_MAXINUMBER_32 */
|
|
|
|
{Opt_ikeep, "ikeep"}, /* do not free empty inode clusters */
|
|
|
|
{Opt_noikeep, "noikeep"}, /* free empty inode clusters */
|
|
|
|
{Opt_largeio, "largeio"}, /* report large I/O sizes in stat() */
|
|
|
|
{Opt_nolargeio, "nolargeio"}, /* do not report large I/O sizes
|
|
|
|
* in stat(). */
|
|
|
|
{Opt_attr2, "attr2"}, /* do use attr2 attribute format */
|
|
|
|
{Opt_noattr2, "noattr2"}, /* do not use attr2 attribute format */
|
|
|
|
{Opt_filestreams,"filestreams"},/* use filestreams allocator */
|
|
|
|
{Opt_quota, "quota"}, /* disk quotas (user) */
|
|
|
|
{Opt_noquota, "noquota"}, /* no quotas */
|
|
|
|
{Opt_usrquota, "usrquota"}, /* user quota enabled */
|
|
|
|
{Opt_grpquota, "grpquota"}, /* group quota enabled */
|
|
|
|
{Opt_prjquota, "prjquota"}, /* project quota enabled */
|
|
|
|
{Opt_uquota, "uquota"}, /* user quota (IRIX variant) */
|
|
|
|
{Opt_gquota, "gquota"}, /* group quota (IRIX variant) */
|
|
|
|
{Opt_pquota, "pquota"}, /* project quota (IRIX variant) */
|
|
|
|
{Opt_uqnoenforce,"uqnoenforce"},/* user quota limit enforcement */
|
|
|
|
{Opt_gqnoenforce,"gqnoenforce"},/* group quota limit enforcement */
|
|
|
|
{Opt_pqnoenforce,"pqnoenforce"},/* project quota limit enforcement */
|
|
|
|
{Opt_qnoenforce, "qnoenforce"}, /* same as uqnoenforce */
|
|
|
|
{Opt_discard, "discard"}, /* Discard unused blocks */
|
|
|
|
{Opt_nodiscard, "nodiscard"}, /* Do not discard unused blocks */
|
|
|
|
{Opt_dax, "dax"}, /* Enable direct access to bdev pages */
|
|
|
|
{Opt_err, NULL},
|
2008-07-18 11:12:36 +04:00
|
|
|
};
|
|
|
|
|
|
|
|
|
2016-01-04 08:13:21 +03:00
|
|
|
STATIC int
|
2016-03-02 01:55:38 +03:00
|
|
|
suffix_kstrtoint(const substring_t *s, unsigned int base, int *res)
|
2007-11-23 08:29:32 +03:00
|
|
|
{
|
2013-01-09 18:04:42 +04:00
|
|
|
int last, shift_left_factor = 0, _res;
|
2016-03-02 01:55:38 +03:00
|
|
|
char *value;
|
|
|
|
int ret = 0;
|
|
|
|
|
|
|
|
value = match_strdup(s);
|
|
|
|
if (!value)
|
|
|
|
return -ENOMEM;
|
2007-11-23 08:29:32 +03:00
|
|
|
|
|
|
|
last = strlen(value) - 1;
|
|
|
|
if (value[last] == 'K' || value[last] == 'k') {
|
|
|
|
shift_left_factor = 10;
|
|
|
|
value[last] = '\0';
|
|
|
|
}
|
|
|
|
if (value[last] == 'M' || value[last] == 'm') {
|
|
|
|
shift_left_factor = 20;
|
|
|
|
value[last] = '\0';
|
|
|
|
}
|
|
|
|
if (value[last] == 'G' || value[last] == 'g') {
|
|
|
|
shift_left_factor = 30;
|
|
|
|
value[last] = '\0';
|
|
|
|
}
|
|
|
|
|
2016-03-02 01:55:38 +03:00
|
|
|
if (kstrtoint(value, base, &_res))
|
|
|
|
ret = -EINVAL;
|
|
|
|
kfree(value);
|
2013-01-09 18:04:42 +04:00
|
|
|
*res = _res << shift_left_factor;
|
2016-03-02 01:55:38 +03:00
|
|
|
return ret;
|
2007-11-23 08:29:32 +03:00
|
|
|
}
|
|
|
|
|
2008-10-30 09:53:24 +03:00
|
|
|
/*
|
|
|
|
* This function fills in xfs_mount_t fields based on mount args.
|
|
|
|
* Note: the superblock has _not_ yet been read in.
|
|
|
|
*
|
|
|
|
* Note that this function leaks the various device name allocations on
|
|
|
|
* failure. The caller takes care of them.
|
2016-03-02 01:56:31 +03:00
|
|
|
*
|
|
|
|
* *sb is const because this is also used to test options on the remount
|
|
|
|
* path, and we don't want this to have any side effects at remount time.
|
|
|
|
* Today this function does not change *sb, but just to future-proof...
|
2008-10-30 09:53:24 +03:00
|
|
|
*/
|
2007-11-23 08:29:32 +03:00
|
|
|
STATIC int
|
|
|
|
xfs_parseargs(
|
|
|
|
struct xfs_mount *mp,
|
2010-06-23 12:11:15 +04:00
|
|
|
char *options)
|
2007-11-23 08:29:32 +03:00
|
|
|
{
|
2016-03-02 01:56:31 +03:00
|
|
|
const struct super_block *sb = mp->m_super;
|
2016-03-02 01:55:38 +03:00
|
|
|
char *p;
|
|
|
|
substring_t args[MAX_OPT_ARGS];
|
2008-10-30 09:53:24 +03:00
|
|
|
int iosize = 0;
|
2017-06-16 21:00:05 +03:00
|
|
|
uint8_t iosizelog = 0;
|
2008-10-30 09:53:24 +03:00
|
|
|
|
2011-03-07 02:00:35 +03:00
|
|
|
/*
|
|
|
|
* set up the mount name first so all the errors will refer to the
|
|
|
|
* correct device.
|
|
|
|
*/
|
|
|
|
mp->m_fsname = kstrndup(sb->s_id, MAXNAMELEN, GFP_KERNEL);
|
|
|
|
if (!mp->m_fsname)
|
2014-06-25 08:58:08 +04:00
|
|
|
return -ENOMEM;
|
2011-03-07 02:00:35 +03:00
|
|
|
mp->m_fsname_len = strlen(mp->m_fsname) + 1;
|
|
|
|
|
2008-10-30 09:53:24 +03:00
|
|
|
/*
|
|
|
|
* Copy binary VFS mount flags we are interested in.
|
|
|
|
*/
|
2017-07-17 10:45:34 +03:00
|
|
|
if (sb_rdonly(sb))
|
2008-10-30 09:53:24 +03:00
|
|
|
mp->m_flags |= XFS_MOUNT_RDONLY;
|
2017-11-28 00:05:09 +03:00
|
|
|
if (sb->s_flags & SB_DIRSYNC)
|
2008-10-30 09:53:24 +03:00
|
|
|
mp->m_flags |= XFS_MOUNT_DIRSYNC;
|
2017-11-28 00:05:09 +03:00
|
|
|
if (sb->s_flags & SB_SYNCHRONOUS)
|
2008-10-30 09:53:24 +03:00
|
|
|
mp->m_flags |= XFS_MOUNT_WSYNC;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Set some default flags that could be cleared by the mount option
|
|
|
|
* parsing.
|
|
|
|
*/
|
|
|
|
mp->m_flags |= XFS_MOUNT_COMPAT_IOSIZE;
|
2007-11-23 08:29:32 +03:00
|
|
|
|
2008-10-30 09:53:24 +03:00
|
|
|
/*
|
|
|
|
* These can be overridden by the mount option parsing.
|
|
|
|
*/
|
|
|
|
mp->m_logbufs = -1;
|
|
|
|
mp->m_logbsize = -1;
|
2007-11-23 08:29:32 +03:00
|
|
|
|
|
|
|
if (!options)
|
|
|
|
goto done;
|
|
|
|
|
2016-03-02 01:55:38 +03:00
|
|
|
while ((p = strsep(&options, ",")) != NULL) {
|
|
|
|
int token;
|
|
|
|
|
|
|
|
if (!*p)
|
2007-11-23 08:29:32 +03:00
|
|
|
continue;
|
|
|
|
|
2016-03-02 01:55:38 +03:00
|
|
|
token = match_token(p, tokens, args);
|
|
|
|
switch (token) {
|
|
|
|
case Opt_logbufs:
|
|
|
|
if (match_int(args, &mp->m_logbufs))
|
2014-06-25 08:58:08 +04:00
|
|
|
return -EINVAL;
|
2016-03-02 01:55:38 +03:00
|
|
|
break;
|
|
|
|
case Opt_logbsize:
|
|
|
|
if (suffix_kstrtoint(args, 10, &mp->m_logbsize))
|
2014-06-25 08:58:08 +04:00
|
|
|
return -EINVAL;
|
2016-03-02 01:55:38 +03:00
|
|
|
break;
|
|
|
|
case Opt_logdev:
|
2018-02-25 23:17:31 +03:00
|
|
|
kfree(mp->m_logname);
|
2016-03-02 01:55:38 +03:00
|
|
|
mp->m_logname = match_strdup(args);
|
2008-10-30 09:53:24 +03:00
|
|
|
if (!mp->m_logname)
|
2014-06-25 08:58:08 +04:00
|
|
|
return -ENOMEM;
|
2016-03-02 01:55:38 +03:00
|
|
|
break;
|
|
|
|
case Opt_rtdev:
|
2018-02-25 23:17:31 +03:00
|
|
|
kfree(mp->m_rtname);
|
2016-03-02 01:55:38 +03:00
|
|
|
mp->m_rtname = match_strdup(args);
|
2008-10-30 09:53:24 +03:00
|
|
|
if (!mp->m_rtname)
|
2014-06-25 08:58:08 +04:00
|
|
|
return -ENOMEM;
|
2016-03-02 01:55:38 +03:00
|
|
|
break;
|
|
|
|
case Opt_allocsize:
|
|
|
|
if (suffix_kstrtoint(args, 10, &iosize))
|
2014-06-25 08:58:08 +04:00
|
|
|
return -EINVAL;
|
2008-10-30 09:53:24 +03:00
|
|
|
iosizelog = ffs(iosize) - 1;
|
2016-03-02 01:55:38 +03:00
|
|
|
break;
|
|
|
|
case Opt_grpid:
|
|
|
|
case Opt_bsdgroups:
|
2007-11-23 08:29:32 +03:00
|
|
|
mp->m_flags |= XFS_MOUNT_GRPID;
|
2016-03-02 01:55:38 +03:00
|
|
|
break;
|
|
|
|
case Opt_nogrpid:
|
|
|
|
case Opt_sysvgroups:
|
2007-11-23 08:29:32 +03:00
|
|
|
mp->m_flags &= ~XFS_MOUNT_GRPID;
|
2016-03-02 01:55:38 +03:00
|
|
|
break;
|
|
|
|
case Opt_wsync:
|
2008-10-30 09:53:24 +03:00
|
|
|
mp->m_flags |= XFS_MOUNT_WSYNC;
|
2016-03-02 01:55:38 +03:00
|
|
|
break;
|
|
|
|
case Opt_norecovery:
|
2008-10-30 09:53:24 +03:00
|
|
|
mp->m_flags |= XFS_MOUNT_NORECOVERY;
|
2016-03-02 01:55:38 +03:00
|
|
|
break;
|
|
|
|
case Opt_noalign:
|
2008-10-30 09:53:24 +03:00
|
|
|
mp->m_flags |= XFS_MOUNT_NOALIGN;
|
2016-03-02 01:55:38 +03:00
|
|
|
break;
|
|
|
|
case Opt_swalloc:
|
2008-10-30 09:53:24 +03:00
|
|
|
mp->m_flags |= XFS_MOUNT_SWALLOC;
|
2016-03-02 01:55:38 +03:00
|
|
|
break;
|
|
|
|
case Opt_sunit:
|
2019-10-28 18:41:43 +03:00
|
|
|
if (match_int(args, &mp->m_dalign))
|
2014-06-25 08:58:08 +04:00
|
|
|
return -EINVAL;
|
2016-03-02 01:55:38 +03:00
|
|
|
break;
|
|
|
|
case Opt_swidth:
|
2019-10-28 18:41:43 +03:00
|
|
|
if (match_int(args, &mp->m_swidth))
|
2014-06-25 08:58:08 +04:00
|
|
|
return -EINVAL;
|
2016-03-02 01:55:38 +03:00
|
|
|
break;
|
|
|
|
case Opt_inode32:
|
2012-09-20 17:32:37 +04:00
|
|
|
mp->m_flags |= XFS_MOUNT_SMALL_INUMS;
|
2016-03-02 01:55:38 +03:00
|
|
|
break;
|
|
|
|
case Opt_inode64:
|
2008-10-30 09:53:24 +03:00
|
|
|
mp->m_flags &= ~XFS_MOUNT_SMALL_INUMS;
|
2016-03-02 01:55:38 +03:00
|
|
|
break;
|
|
|
|
case Opt_nouuid:
|
2008-10-30 09:53:24 +03:00
|
|
|
mp->m_flags |= XFS_MOUNT_NOUUID;
|
2016-03-02 01:55:38 +03:00
|
|
|
break;
|
|
|
|
case Opt_ikeep:
|
2008-10-30 09:53:24 +03:00
|
|
|
mp->m_flags |= XFS_MOUNT_IKEEP;
|
2016-03-02 01:55:38 +03:00
|
|
|
break;
|
|
|
|
case Opt_noikeep:
|
2008-10-30 09:53:24 +03:00
|
|
|
mp->m_flags &= ~XFS_MOUNT_IKEEP;
|
2016-03-02 01:55:38 +03:00
|
|
|
break;
|
|
|
|
case Opt_largeio:
|
2008-10-30 09:53:24 +03:00
|
|
|
mp->m_flags &= ~XFS_MOUNT_COMPAT_IOSIZE;
|
2016-03-02 01:55:38 +03:00
|
|
|
break;
|
|
|
|
case Opt_nolargeio:
|
2008-10-30 09:53:24 +03:00
|
|
|
mp->m_flags |= XFS_MOUNT_COMPAT_IOSIZE;
|
2016-03-02 01:55:38 +03:00
|
|
|
break;
|
|
|
|
case Opt_attr2:
|
2008-10-30 09:53:24 +03:00
|
|
|
mp->m_flags |= XFS_MOUNT_ATTR2;
|
2016-03-02 01:55:38 +03:00
|
|
|
break;
|
|
|
|
case Opt_noattr2:
|
2008-10-30 09:53:24 +03:00
|
|
|
mp->m_flags &= ~XFS_MOUNT_ATTR2;
|
|
|
|
mp->m_flags |= XFS_MOUNT_NOATTR2;
|
2016-03-02 01:55:38 +03:00
|
|
|
break;
|
|
|
|
case Opt_filestreams:
|
2008-10-30 09:53:24 +03:00
|
|
|
mp->m_flags |= XFS_MOUNT_FILESTREAMS;
|
2016-03-02 01:55:38 +03:00
|
|
|
break;
|
|
|
|
case Opt_noquota:
|
2012-01-23 21:31:43 +04:00
|
|
|
mp->m_qflags &= ~XFS_ALL_QUOTA_ACCT;
|
|
|
|
mp->m_qflags &= ~XFS_ALL_QUOTA_ENFD;
|
|
|
|
mp->m_qflags &= ~XFS_ALL_QUOTA_ACTIVE;
|
2016-03-02 01:55:38 +03:00
|
|
|
break;
|
|
|
|
case Opt_quota:
|
|
|
|
case Opt_uquota:
|
|
|
|
case Opt_usrquota:
|
2008-10-30 09:53:24 +03:00
|
|
|
mp->m_qflags |= (XFS_UQUOTA_ACCT | XFS_UQUOTA_ACTIVE |
|
|
|
|
XFS_UQUOTA_ENFD);
|
2016-03-02 01:55:38 +03:00
|
|
|
break;
|
|
|
|
case Opt_qnoenforce:
|
|
|
|
case Opt_uqnoenforce:
|
2008-10-30 09:53:24 +03:00
|
|
|
mp->m_qflags |= (XFS_UQUOTA_ACCT | XFS_UQUOTA_ACTIVE);
|
|
|
|
mp->m_qflags &= ~XFS_UQUOTA_ENFD;
|
2016-03-02 01:55:38 +03:00
|
|
|
break;
|
|
|
|
case Opt_pquota:
|
|
|
|
case Opt_prjquota:
|
2008-10-30 09:53:24 +03:00
|
|
|
mp->m_qflags |= (XFS_PQUOTA_ACCT | XFS_PQUOTA_ACTIVE |
|
2013-06-28 02:25:10 +04:00
|
|
|
XFS_PQUOTA_ENFD);
|
2016-03-02 01:55:38 +03:00
|
|
|
break;
|
|
|
|
case Opt_pqnoenforce:
|
2008-10-30 09:53:24 +03:00
|
|
|
mp->m_qflags |= (XFS_PQUOTA_ACCT | XFS_PQUOTA_ACTIVE);
|
2013-06-28 02:25:10 +04:00
|
|
|
mp->m_qflags &= ~XFS_PQUOTA_ENFD;
|
2016-04-06 00:19:40 +03:00
|
|
|
break;
|
2016-03-02 01:55:38 +03:00
|
|
|
case Opt_gquota:
|
|
|
|
case Opt_grpquota:
|
2008-10-30 09:53:24 +03:00
|
|
|
mp->m_qflags |= (XFS_GQUOTA_ACCT | XFS_GQUOTA_ACTIVE |
|
2013-06-28 02:25:10 +04:00
|
|
|
XFS_GQUOTA_ENFD);
|
2016-03-02 01:55:38 +03:00
|
|
|
break;
|
|
|
|
case Opt_gqnoenforce:
|
2008-10-30 09:53:24 +03:00
|
|
|
mp->m_qflags |= (XFS_GQUOTA_ACCT | XFS_GQUOTA_ACTIVE);
|
2013-06-28 02:25:10 +04:00
|
|
|
mp->m_qflags &= ~XFS_GQUOTA_ENFD;
|
2016-03-02 01:55:38 +03:00
|
|
|
break;
|
|
|
|
case Opt_discard:
|
2011-05-20 17:45:32 +04:00
|
|
|
mp->m_flags |= XFS_MOUNT_DISCARD;
|
2016-03-02 01:55:38 +03:00
|
|
|
break;
|
|
|
|
case Opt_nodiscard:
|
2011-05-20 17:45:32 +04:00
|
|
|
mp->m_flags &= ~XFS_MOUNT_DISCARD;
|
2016-03-02 01:55:38 +03:00
|
|
|
break;
|
2015-06-04 02:19:18 +03:00
|
|
|
#ifdef CONFIG_FS_DAX
|
2016-03-02 01:55:38 +03:00
|
|
|
case Opt_dax:
|
2015-06-04 02:19:18 +03:00
|
|
|
mp->m_flags |= XFS_MOUNT_DAX;
|
2016-03-02 01:55:38 +03:00
|
|
|
break;
|
2015-06-04 02:19:18 +03:00
|
|
|
#endif
|
2016-03-02 01:55:38 +03:00
|
|
|
default:
|
|
|
|
xfs_warn(mp, "unknown mount option [%s].", p);
|
2014-06-25 08:58:08 +04:00
|
|
|
return -EINVAL;
|
2007-11-23 08:29:32 +03:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2008-10-30 09:53:24 +03:00
|
|
|
/*
|
|
|
|
* no recovery flag requires a read-only mount
|
|
|
|
*/
|
|
|
|
if ((mp->m_flags & XFS_MOUNT_NORECOVERY) &&
|
|
|
|
!(mp->m_flags & XFS_MOUNT_RDONLY)) {
|
2011-03-07 02:00:35 +03:00
|
|
|
xfs_warn(mp, "no-recovery mounts must be read-only.");
|
2014-06-25 08:58:08 +04:00
|
|
|
return -EINVAL;
|
2007-11-23 08:29:32 +03:00
|
|
|
}
|
|
|
|
|
2019-10-28 18:41:43 +03:00
|
|
|
if ((mp->m_flags & XFS_MOUNT_NOALIGN) &&
|
|
|
|
(mp->m_dalign || mp->m_swidth)) {
|
2011-03-07 02:00:35 +03:00
|
|
|
xfs_warn(mp,
|
|
|
|
"sunit and swidth options incompatible with the noalign option");
|
2014-06-25 08:58:08 +04:00
|
|
|
return -EINVAL;
|
2007-11-23 08:29:32 +03:00
|
|
|
}
|
|
|
|
|
2009-06-08 17:33:32 +04:00
|
|
|
#ifndef CONFIG_XFS_QUOTA
|
|
|
|
if (XFS_IS_QUOTA_RUNNING(mp)) {
|
2011-03-07 02:00:35 +03:00
|
|
|
xfs_warn(mp, "quota support not available in this kernel.");
|
2014-06-25 08:58:08 +04:00
|
|
|
return -EINVAL;
|
2009-06-08 17:33:32 +04:00
|
|
|
}
|
|
|
|
#endif
|
|
|
|
|
2019-10-28 18:41:43 +03:00
|
|
|
if ((mp->m_dalign && !mp->m_swidth) ||
|
|
|
|
(!mp->m_dalign && mp->m_swidth)) {
|
2011-03-07 02:00:35 +03:00
|
|
|
xfs_warn(mp, "sunit and swidth must be specified together");
|
2014-06-25 08:58:08 +04:00
|
|
|
return -EINVAL;
|
2007-11-23 08:29:32 +03:00
|
|
|
}
|
|
|
|
|
2019-10-28 18:41:43 +03:00
|
|
|
if (mp->m_dalign && (mp->m_swidth % mp->m_dalign != 0)) {
|
2011-03-07 02:00:35 +03:00
|
|
|
xfs_warn(mp,
|
|
|
|
"stripe width (%d) must be a multiple of the stripe unit (%d)",
|
2019-10-28 18:41:43 +03:00
|
|
|
mp->m_swidth, mp->m_dalign);
|
2014-06-25 08:58:08 +04:00
|
|
|
return -EINVAL;
|
2007-11-23 08:29:32 +03:00
|
|
|
}
|
|
|
|
|
2008-10-30 09:53:24 +03:00
|
|
|
done:
|
|
|
|
if (mp->m_logbufs != -1 &&
|
|
|
|
mp->m_logbufs != 0 &&
|
|
|
|
(mp->m_logbufs < XLOG_MIN_ICLOGS ||
|
|
|
|
mp->m_logbufs > XLOG_MAX_ICLOGS)) {
|
2011-03-07 02:00:35 +03:00
|
|
|
xfs_warn(mp, "invalid logbufs value: %d [not %d-%d]",
|
2008-10-30 09:53:24 +03:00
|
|
|
mp->m_logbufs, XLOG_MIN_ICLOGS, XLOG_MAX_ICLOGS);
|
2014-06-25 08:58:08 +04:00
|
|
|
return -EINVAL;
|
2008-10-30 09:53:24 +03:00
|
|
|
}
|
|
|
|
if (mp->m_logbsize != -1 &&
|
|
|
|
mp->m_logbsize != 0 &&
|
|
|
|
(mp->m_logbsize < XLOG_MIN_RECORD_BSIZE ||
|
|
|
|
mp->m_logbsize > XLOG_MAX_RECORD_BSIZE ||
|
|
|
|
!is_power_of_2(mp->m_logbsize))) {
|
2011-03-07 02:00:35 +03:00
|
|
|
xfs_warn(mp,
|
|
|
|
"invalid logbufsize: %d [not 16k,32k,64k,128k or 256k]",
|
2008-10-30 09:53:24 +03:00
|
|
|
mp->m_logbsize);
|
2014-06-25 08:58:08 +04:00
|
|
|
return -EINVAL;
|
2008-10-30 09:53:24 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
if (iosizelog) {
|
|
|
|
if (iosizelog > XFS_MAX_IO_LOG ||
|
|
|
|
iosizelog < XFS_MIN_IO_LOG) {
|
2011-03-07 02:00:35 +03:00
|
|
|
xfs_warn(mp, "invalid log iosize: %d [not %d-%d]",
|
2008-10-30 09:53:24 +03:00
|
|
|
iosizelog, XFS_MIN_IO_LOG,
|
|
|
|
XFS_MAX_IO_LOG);
|
2014-06-25 08:58:08 +04:00
|
|
|
return -EINVAL;
|
2008-10-30 09:53:24 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
mp->m_flags |= XFS_MOUNT_DFLT_IOSIZE;
|
2019-10-28 18:41:44 +03:00
|
|
|
mp->m_allocsize_log = iosizelog;
|
2007-11-23 08:29:32 +03:00
|
|
|
}
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
struct proc_xfs_info {
|
2015-06-04 02:19:18 +03:00
|
|
|
uint64_t flag;
|
|
|
|
char *str;
|
2007-11-23 08:29:32 +03:00
|
|
|
};
|
|
|
|
|
2019-05-02 06:26:30 +03:00
|
|
|
STATIC void
|
2007-11-23 08:29:32 +03:00
|
|
|
xfs_showargs(
|
|
|
|
struct xfs_mount *mp,
|
|
|
|
struct seq_file *m)
|
|
|
|
{
|
|
|
|
static struct proc_xfs_info xfs_info_set[] = {
|
|
|
|
/* the few simple ones we can get from the mount struct */
|
2016-03-02 01:55:38 +03:00
|
|
|
{ XFS_MOUNT_IKEEP, ",ikeep" },
|
|
|
|
{ XFS_MOUNT_WSYNC, ",wsync" },
|
|
|
|
{ XFS_MOUNT_NOALIGN, ",noalign" },
|
|
|
|
{ XFS_MOUNT_SWALLOC, ",swalloc" },
|
|
|
|
{ XFS_MOUNT_NOUUID, ",nouuid" },
|
|
|
|
{ XFS_MOUNT_NORECOVERY, ",norecovery" },
|
|
|
|
{ XFS_MOUNT_ATTR2, ",attr2" },
|
|
|
|
{ XFS_MOUNT_FILESTREAMS, ",filestreams" },
|
|
|
|
{ XFS_MOUNT_GRPID, ",grpid" },
|
|
|
|
{ XFS_MOUNT_DISCARD, ",discard" },
|
|
|
|
{ XFS_MOUNT_SMALL_INUMS, ",inode32" },
|
|
|
|
{ XFS_MOUNT_DAX, ",dax" },
|
2007-11-23 08:29:32 +03:00
|
|
|
{ 0, NULL }
|
|
|
|
};
|
|
|
|
static struct proc_xfs_info xfs_info_unset[] = {
|
|
|
|
/* the few simple ones we can get from the mount struct */
|
2016-03-02 01:55:38 +03:00
|
|
|
{ XFS_MOUNT_COMPAT_IOSIZE, ",largeio" },
|
|
|
|
{ XFS_MOUNT_SMALL_INUMS, ",inode64" },
|
2007-11-23 08:29:32 +03:00
|
|
|
{ 0, NULL }
|
|
|
|
};
|
|
|
|
struct proc_xfs_info *xfs_infop;
|
|
|
|
|
|
|
|
for (xfs_infop = xfs_info_set; xfs_infop->flag; xfs_infop++) {
|
|
|
|
if (mp->m_flags & xfs_infop->flag)
|
|
|
|
seq_puts(m, xfs_infop->str);
|
|
|
|
}
|
|
|
|
for (xfs_infop = xfs_info_unset; xfs_infop->flag; xfs_infop++) {
|
|
|
|
if (!(mp->m_flags & xfs_infop->flag))
|
|
|
|
seq_puts(m, xfs_infop->str);
|
|
|
|
}
|
|
|
|
|
|
|
|
if (mp->m_flags & XFS_MOUNT_DFLT_IOSIZE)
|
2016-03-02 01:55:38 +03:00
|
|
|
seq_printf(m, ",allocsize=%dk",
|
2019-10-28 18:41:44 +03:00
|
|
|
(int)(1 << mp->m_allocsize_log) >> 10);
|
2007-11-23 08:29:32 +03:00
|
|
|
|
|
|
|
if (mp->m_logbufs > 0)
|
2016-03-02 01:55:38 +03:00
|
|
|
seq_printf(m, ",logbufs=%d", mp->m_logbufs);
|
2007-11-23 08:29:32 +03:00
|
|
|
if (mp->m_logbsize > 0)
|
2016-03-02 01:55:38 +03:00
|
|
|
seq_printf(m, ",logbsize=%dk", mp->m_logbsize >> 10);
|
2007-11-23 08:29:32 +03:00
|
|
|
|
|
|
|
if (mp->m_logname)
|
2016-03-02 01:55:38 +03:00
|
|
|
seq_show_option(m, "logdev", mp->m_logname);
|
2007-11-23 08:29:32 +03:00
|
|
|
if (mp->m_rtname)
|
2016-03-02 01:55:38 +03:00
|
|
|
seq_show_option(m, "rtdev", mp->m_rtname);
|
2007-11-23 08:29:32 +03:00
|
|
|
|
|
|
|
if (mp->m_dalign > 0)
|
2016-03-02 01:55:38 +03:00
|
|
|
seq_printf(m, ",sunit=%d",
|
2007-11-23 08:29:32 +03:00
|
|
|
(int)XFS_FSB_TO_BB(mp, mp->m_dalign));
|
|
|
|
if (mp->m_swidth > 0)
|
2016-03-02 01:55:38 +03:00
|
|
|
seq_printf(m, ",swidth=%d",
|
2007-11-23 08:29:32 +03:00
|
|
|
(int)XFS_FSB_TO_BB(mp, mp->m_swidth));
|
|
|
|
|
|
|
|
if (mp->m_qflags & (XFS_UQUOTA_ACCT|XFS_UQUOTA_ENFD))
|
2016-03-02 01:55:38 +03:00
|
|
|
seq_puts(m, ",usrquota");
|
2007-11-23 08:29:32 +03:00
|
|
|
else if (mp->m_qflags & XFS_UQUOTA_ACCT)
|
2016-03-02 01:55:38 +03:00
|
|
|
seq_puts(m, ",uqnoenforce");
|
2007-11-23 08:29:32 +03:00
|
|
|
|
xfs: xfs_showargs() reports group *and* project quotas enabled
If you enable group or project quotas on an XFS file system, then the
mount table presented through /proc/self/mounts erroneously shows
that both options are in effect for the file system. The root of
the problem is some bad logic in the xfs_showargs() function, which
is used to format the file system type-specific options in effect
for a file system.
The problem originated in this GIT commit:
Move platform specific mount option parse out of core XFS code
Date: 11/22/07
Author: Dave Chinner
SHA1 ID: a67d7c5f5d25d0b13a4dfb182697135b014fa478
For XFS quotas, project and group quota management are mutually
exclusive--only one can be in effect at a time. There are two
parts to managing quotas: aggregating usage information; and
enforcing limits. It is possible to have a quota in effect
(aggregating usage) but not enforced.
These features are recorded on an XFS mount point using these flags:
XFS_PQUOTA_ACCT - Project quotas are aggregated
XFS_GQUOTA_ACCT - Group quotas are aggregated
XFS_OQUOTA_ENFD - Project/group quotas are enforced
The code in error is in fs/xfs/linux-2.6/xfs_super.c:
if (mp->m_qflags & (XFS_PQUOTA_ACCT|XFS_OQUOTA_ENFD))
seq_puts(m, "," MNTOPT_PRJQUOTA);
else if (mp->m_qflags & XFS_PQUOTA_ACCT)
seq_puts(m, "," MNTOPT_PQUOTANOENF);
if (mp->m_qflags & (XFS_GQUOTA_ACCT|XFS_OQUOTA_ENFD))
seq_puts(m, "," MNTOPT_GRPQUOTA);
else if (mp->m_qflags & XFS_GQUOTA_ACCT)
seq_puts(m, "," MNTOPT_GQUOTANOENF);
The problem is that XFS_OQUOTA_ENFD will be set in mp->m_qflags
if either group or project quotas are enforced, and as a result
both MNTOPT_PRJQUOTA and MNTOPT_GRPQUOTA will be shown as mount
options.
Signed-off-by: Alex Elder <aelder@sgi.com>
Reviewed-by: Christoph Hellwig <hch@lst.de>
Reviewed-by: Felix Blyakher <felixb@sgi.com>
2009-09-03 02:02:24 +04:00
|
|
|
if (mp->m_qflags & XFS_PQUOTA_ACCT) {
|
2013-06-28 02:25:10 +04:00
|
|
|
if (mp->m_qflags & XFS_PQUOTA_ENFD)
|
2016-03-02 01:55:38 +03:00
|
|
|
seq_puts(m, ",prjquota");
|
xfs: xfs_showargs() reports group *and* project quotas enabled
If you enable group or project quotas on an XFS file system, then the
mount table presented through /proc/self/mounts erroneously shows
that both options are in effect for the file system. The root of
the problem is some bad logic in the xfs_showargs() function, which
is used to format the file system type-specific options in effect
for a file system.
The problem originated in this GIT commit:
Move platform specific mount option parse out of core XFS code
Date: 11/22/07
Author: Dave Chinner
SHA1 ID: a67d7c5f5d25d0b13a4dfb182697135b014fa478
For XFS quotas, project and group quota management are mutually
exclusive--only one can be in effect at a time. There are two
parts to managing quotas: aggregating usage information; and
enforcing limits. It is possible to have a quota in effect
(aggregating usage) but not enforced.
These features are recorded on an XFS mount point using these flags:
XFS_PQUOTA_ACCT - Project quotas are aggregated
XFS_GQUOTA_ACCT - Group quotas are aggregated
XFS_OQUOTA_ENFD - Project/group quotas are enforced
The code in error is in fs/xfs/linux-2.6/xfs_super.c:
if (mp->m_qflags & (XFS_PQUOTA_ACCT|XFS_OQUOTA_ENFD))
seq_puts(m, "," MNTOPT_PRJQUOTA);
else if (mp->m_qflags & XFS_PQUOTA_ACCT)
seq_puts(m, "," MNTOPT_PQUOTANOENF);
if (mp->m_qflags & (XFS_GQUOTA_ACCT|XFS_OQUOTA_ENFD))
seq_puts(m, "," MNTOPT_GRPQUOTA);
else if (mp->m_qflags & XFS_GQUOTA_ACCT)
seq_puts(m, "," MNTOPT_GQUOTANOENF);
The problem is that XFS_OQUOTA_ENFD will be set in mp->m_qflags
if either group or project quotas are enforced, and as a result
both MNTOPT_PRJQUOTA and MNTOPT_GRPQUOTA will be shown as mount
options.
Signed-off-by: Alex Elder <aelder@sgi.com>
Reviewed-by: Christoph Hellwig <hch@lst.de>
Reviewed-by: Felix Blyakher <felixb@sgi.com>
2009-09-03 02:02:24 +04:00
|
|
|
else
|
2016-03-02 01:55:38 +03:00
|
|
|
seq_puts(m, ",pqnoenforce");
|
2013-07-20 02:36:02 +04:00
|
|
|
}
|
|
|
|
if (mp->m_qflags & XFS_GQUOTA_ACCT) {
|
2013-06-28 02:25:10 +04:00
|
|
|
if (mp->m_qflags & XFS_GQUOTA_ENFD)
|
2016-03-02 01:55:38 +03:00
|
|
|
seq_puts(m, ",grpquota");
|
xfs: xfs_showargs() reports group *and* project quotas enabled
If you enable group or project quotas on an XFS file system, then the
mount table presented through /proc/self/mounts erroneously shows
that both options are in effect for the file system. The root of
the problem is some bad logic in the xfs_showargs() function, which
is used to format the file system type-specific options in effect
for a file system.
The problem originated in this GIT commit:
Move platform specific mount option parse out of core XFS code
Date: 11/22/07
Author: Dave Chinner
SHA1 ID: a67d7c5f5d25d0b13a4dfb182697135b014fa478
For XFS quotas, project and group quota management are mutually
exclusive--only one can be in effect at a time. There are two
parts to managing quotas: aggregating usage information; and
enforcing limits. It is possible to have a quota in effect
(aggregating usage) but not enforced.
These features are recorded on an XFS mount point using these flags:
XFS_PQUOTA_ACCT - Project quotas are aggregated
XFS_GQUOTA_ACCT - Group quotas are aggregated
XFS_OQUOTA_ENFD - Project/group quotas are enforced
The code in error is in fs/xfs/linux-2.6/xfs_super.c:
if (mp->m_qflags & (XFS_PQUOTA_ACCT|XFS_OQUOTA_ENFD))
seq_puts(m, "," MNTOPT_PRJQUOTA);
else if (mp->m_qflags & XFS_PQUOTA_ACCT)
seq_puts(m, "," MNTOPT_PQUOTANOENF);
if (mp->m_qflags & (XFS_GQUOTA_ACCT|XFS_OQUOTA_ENFD))
seq_puts(m, "," MNTOPT_GRPQUOTA);
else if (mp->m_qflags & XFS_GQUOTA_ACCT)
seq_puts(m, "," MNTOPT_GQUOTANOENF);
The problem is that XFS_OQUOTA_ENFD will be set in mp->m_qflags
if either group or project quotas are enforced, and as a result
both MNTOPT_PRJQUOTA and MNTOPT_GRPQUOTA will be shown as mount
options.
Signed-off-by: Alex Elder <aelder@sgi.com>
Reviewed-by: Christoph Hellwig <hch@lst.de>
Reviewed-by: Felix Blyakher <felixb@sgi.com>
2009-09-03 02:02:24 +04:00
|
|
|
else
|
2016-03-02 01:55:38 +03:00
|
|
|
seq_puts(m, ",gqnoenforce");
|
xfs: xfs_showargs() reports group *and* project quotas enabled
If you enable group or project quotas on an XFS file system, then the
mount table presented through /proc/self/mounts erroneously shows
that both options are in effect for the file system. The root of
the problem is some bad logic in the xfs_showargs() function, which
is used to format the file system type-specific options in effect
for a file system.
The problem originated in this GIT commit:
Move platform specific mount option parse out of core XFS code
Date: 11/22/07
Author: Dave Chinner
SHA1 ID: a67d7c5f5d25d0b13a4dfb182697135b014fa478
For XFS quotas, project and group quota management are mutually
exclusive--only one can be in effect at a time. There are two
parts to managing quotas: aggregating usage information; and
enforcing limits. It is possible to have a quota in effect
(aggregating usage) but not enforced.
These features are recorded on an XFS mount point using these flags:
XFS_PQUOTA_ACCT - Project quotas are aggregated
XFS_GQUOTA_ACCT - Group quotas are aggregated
XFS_OQUOTA_ENFD - Project/group quotas are enforced
The code in error is in fs/xfs/linux-2.6/xfs_super.c:
if (mp->m_qflags & (XFS_PQUOTA_ACCT|XFS_OQUOTA_ENFD))
seq_puts(m, "," MNTOPT_PRJQUOTA);
else if (mp->m_qflags & XFS_PQUOTA_ACCT)
seq_puts(m, "," MNTOPT_PQUOTANOENF);
if (mp->m_qflags & (XFS_GQUOTA_ACCT|XFS_OQUOTA_ENFD))
seq_puts(m, "," MNTOPT_GRPQUOTA);
else if (mp->m_qflags & XFS_GQUOTA_ACCT)
seq_puts(m, "," MNTOPT_GQUOTANOENF);
The problem is that XFS_OQUOTA_ENFD will be set in mp->m_qflags
if either group or project quotas are enforced, and as a result
both MNTOPT_PRJQUOTA and MNTOPT_GRPQUOTA will be shown as mount
options.
Signed-off-by: Alex Elder <aelder@sgi.com>
Reviewed-by: Christoph Hellwig <hch@lst.de>
Reviewed-by: Felix Blyakher <felixb@sgi.com>
2009-09-03 02:02:24 +04:00
|
|
|
}
|
2007-11-23 08:29:32 +03:00
|
|
|
|
|
|
|
if (!(mp->m_qflags & XFS_ALL_QUOTA_ACCT))
|
2016-03-02 01:55:38 +03:00
|
|
|
seq_puts(m, ",noquota");
|
2007-11-23 08:29:32 +03:00
|
|
|
}
|
2019-05-02 06:26:30 +03:00
|
|
|
|
2017-06-16 21:00:05 +03:00
|
|
|
static uint64_t
|
2005-04-17 02:20:36 +04:00
|
|
|
xfs_max_file_offset(
|
|
|
|
unsigned int blockshift)
|
|
|
|
{
|
|
|
|
unsigned int pagefactor = 1;
|
|
|
|
unsigned int bitshift = BITS_PER_LONG - 1;
|
|
|
|
|
|
|
|
/* Figure out maximum filesize, on Linux this can depend on
|
|
|
|
* the filesystem blocksize (on 32 bit platforms).
|
2019-04-05 19:08:59 +03:00
|
|
|
* __block_write_begin does this in an [unsigned] long long...
|
2016-04-01 15:29:48 +03:00
|
|
|
* page->index << (PAGE_SHIFT - bbits)
|
2005-04-17 02:20:36 +04:00
|
|
|
* So, for page sized blocks (4K on 32 bit platforms),
|
|
|
|
* this wraps at around 8Tb (hence MAX_LFS_FILESIZE which is
|
2016-04-01 15:29:48 +03:00
|
|
|
* (((u64)PAGE_SIZE << (BITS_PER_LONG-1))-1)
|
2005-04-17 02:20:36 +04:00
|
|
|
* but for smaller blocksizes it is less (bbits = log2 bsize).
|
|
|
|
*/
|
|
|
|
|
|
|
|
#if BITS_PER_LONG == 32
|
|
|
|
ASSERT(sizeof(sector_t) == 8);
|
mm, fs: get rid of PAGE_CACHE_* and page_cache_{get,release} macros
PAGE_CACHE_{SIZE,SHIFT,MASK,ALIGN} macros were introduced *long* time
ago with promise that one day it will be possible to implement page
cache with bigger chunks than PAGE_SIZE.
This promise never materialized. And unlikely will.
We have many places where PAGE_CACHE_SIZE assumed to be equal to
PAGE_SIZE. And it's constant source of confusion on whether
PAGE_CACHE_* or PAGE_* constant should be used in a particular case,
especially on the border between fs and mm.
Global switching to PAGE_CACHE_SIZE != PAGE_SIZE would cause to much
breakage to be doable.
Let's stop pretending that pages in page cache are special. They are
not.
The changes are pretty straight-forward:
- <foo> << (PAGE_CACHE_SHIFT - PAGE_SHIFT) -> <foo>;
- <foo> >> (PAGE_CACHE_SHIFT - PAGE_SHIFT) -> <foo>;
- PAGE_CACHE_{SIZE,SHIFT,MASK,ALIGN} -> PAGE_{SIZE,SHIFT,MASK,ALIGN};
- page_cache_get() -> get_page();
- page_cache_release() -> put_page();
This patch contains automated changes generated with coccinelle using
script below. For some reason, coccinelle doesn't patch header files.
I've called spatch for them manually.
The only adjustment after coccinelle is revert of changes to
PAGE_CAHCE_ALIGN definition: we are going to drop it later.
There are few places in the code where coccinelle didn't reach. I'll
fix them manually in a separate patch. Comments and documentation also
will be addressed with the separate patch.
virtual patch
@@
expression E;
@@
- E << (PAGE_CACHE_SHIFT - PAGE_SHIFT)
+ E
@@
expression E;
@@
- E >> (PAGE_CACHE_SHIFT - PAGE_SHIFT)
+ E
@@
@@
- PAGE_CACHE_SHIFT
+ PAGE_SHIFT
@@
@@
- PAGE_CACHE_SIZE
+ PAGE_SIZE
@@
@@
- PAGE_CACHE_MASK
+ PAGE_MASK
@@
expression E;
@@
- PAGE_CACHE_ALIGN(E)
+ PAGE_ALIGN(E)
@@
expression E;
@@
- page_cache_get(E)
+ get_page(E)
@@
expression E;
@@
- page_cache_release(E)
+ put_page(E)
Signed-off-by: Kirill A. Shutemov <kirill.shutemov@linux.intel.com>
Acked-by: Michal Hocko <mhocko@suse.com>
Signed-off-by: Linus Torvalds <torvalds@linux-foundation.org>
2016-04-01 15:29:47 +03:00
|
|
|
pagefactor = PAGE_SIZE;
|
2005-04-17 02:20:36 +04:00
|
|
|
bitshift = BITS_PER_LONG;
|
|
|
|
#endif
|
|
|
|
|
2017-06-16 21:00:05 +03:00
|
|
|
return (((uint64_t)pagefactor) << bitshift) - 1;
|
2005-04-17 02:20:36 +04:00
|
|
|
}
|
|
|
|
|
xfs: allow inode allocations in post-growfs disk space
Today, if we perform an xfs_growfs which adds allocation groups,
mp->m_maxagi is not properly updated when the growfs is complete.
Therefore inodes will continue to be allocated only in the
AGs which existed prior to the growfs, and the new space
won't be utilized.
This is because of this path in xfs_growfs_data_private():
xfs_growfs_data_private
xfs_initialize_perag(mp, nagcount, &nagimax);
if (mp->m_flags & XFS_MOUNT_32BITINODES)
index = xfs_set_inode32(mp);
else
index = xfs_set_inode64(mp);
if (maxagi)
*maxagi = index;
where xfs_set_inode* iterates over the (old) agcount in
mp->m_sb.sb_agblocks, which has not yet been updated
in the growfs path. So "index" will be returned based on
the old agcount, not the new one, and new AGs are not available
for inode allocation.
Fix this by explicitly passing the proper AG count (which
xfs_initialize_perag() already has) down another level,
so that xfs_set_inode* can make the proper decision about
acceptable AGs for inode allocation in the potentially
newly-added AGs.
This has been broken since 3.7, when these two
xfs_set_inode* functions were added in commit 2d2194f.
Prior to that, we looped over "agcount" not sb_agblocks
in these calculations.
Signed-off-by: Eric Sandeen <sandeen@redhat.com>
Reviewed-by: Brian Foster <bfoster@redhat.com>
Signed-off-by: Dave Chinner <david@fromorbit.com>
2014-07-24 14:51:54 +04:00
|
|
|
/*
|
2016-03-02 01:58:09 +03:00
|
|
|
* Set parameters for inode allocation heuristics, taking into account
|
|
|
|
* filesystem size and inode32/inode64 mount options; i.e. specifically
|
|
|
|
* whether or not XFS_MOUNT_SMALL_INUMS is set.
|
|
|
|
*
|
|
|
|
* Inode allocation patterns are altered only if inode32 is requested
|
|
|
|
* (XFS_MOUNT_SMALL_INUMS), and the filesystem is sufficiently large.
|
|
|
|
* If altered, XFS_MOUNT_32BITINODES is set as well.
|
|
|
|
*
|
|
|
|
* An agcount independent of that in the mount structure is provided
|
|
|
|
* because in the growfs case, mp->m_sb.sb_agcount is not yet updated
|
|
|
|
* to the potentially higher ag count.
|
|
|
|
*
|
|
|
|
* Returns the maximum AG index which may contain inodes.
|
xfs: allow inode allocations in post-growfs disk space
Today, if we perform an xfs_growfs which adds allocation groups,
mp->m_maxagi is not properly updated when the growfs is complete.
Therefore inodes will continue to be allocated only in the
AGs which existed prior to the growfs, and the new space
won't be utilized.
This is because of this path in xfs_growfs_data_private():
xfs_growfs_data_private
xfs_initialize_perag(mp, nagcount, &nagimax);
if (mp->m_flags & XFS_MOUNT_32BITINODES)
index = xfs_set_inode32(mp);
else
index = xfs_set_inode64(mp);
if (maxagi)
*maxagi = index;
where xfs_set_inode* iterates over the (old) agcount in
mp->m_sb.sb_agblocks, which has not yet been updated
in the growfs path. So "index" will be returned based on
the old agcount, not the new one, and new AGs are not available
for inode allocation.
Fix this by explicitly passing the proper AG count (which
xfs_initialize_perag() already has) down another level,
so that xfs_set_inode* can make the proper decision about
acceptable AGs for inode allocation in the potentially
newly-added AGs.
This has been broken since 3.7, when these two
xfs_set_inode* functions were added in commit 2d2194f.
Prior to that, we looped over "agcount" not sb_agblocks
in these calculations.
Signed-off-by: Eric Sandeen <sandeen@redhat.com>
Reviewed-by: Brian Foster <bfoster@redhat.com>
Signed-off-by: Dave Chinner <david@fromorbit.com>
2014-07-24 14:51:54 +04:00
|
|
|
*/
|
2012-09-20 17:32:38 +04:00
|
|
|
xfs_agnumber_t
|
2016-03-02 01:58:09 +03:00
|
|
|
xfs_set_inode_alloc(
|
|
|
|
struct xfs_mount *mp,
|
|
|
|
xfs_agnumber_t agcount)
|
2012-09-20 17:32:38 +04:00
|
|
|
{
|
2016-03-02 01:58:09 +03:00
|
|
|
xfs_agnumber_t index;
|
2012-09-20 17:32:40 +04:00
|
|
|
xfs_agnumber_t maxagi = 0;
|
2012-09-20 17:32:38 +04:00
|
|
|
xfs_sb_t *sbp = &mp->m_sb;
|
|
|
|
xfs_agnumber_t max_metadata;
|
2014-07-24 14:53:10 +04:00
|
|
|
xfs_agino_t agino;
|
|
|
|
xfs_ino_t ino;
|
2012-09-20 17:32:38 +04:00
|
|
|
|
2016-03-02 01:58:09 +03:00
|
|
|
/*
|
|
|
|
* Calculate how much should be reserved for inodes to meet
|
|
|
|
* the max inode percentage. Used only for inode32.
|
2012-09-20 17:32:38 +04:00
|
|
|
*/
|
2019-06-05 21:19:34 +03:00
|
|
|
if (M_IGEO(mp)->maxicount) {
|
2017-06-16 21:00:05 +03:00
|
|
|
uint64_t icount;
|
2012-09-20 17:32:38 +04:00
|
|
|
|
|
|
|
icount = sbp->sb_dblocks * sbp->sb_imax_pct;
|
|
|
|
do_div(icount, 100);
|
|
|
|
icount += sbp->sb_agblocks - 1;
|
|
|
|
do_div(icount, sbp->sb_agblocks);
|
|
|
|
max_metadata = icount;
|
|
|
|
} else {
|
xfs: allow inode allocations in post-growfs disk space
Today, if we perform an xfs_growfs which adds allocation groups,
mp->m_maxagi is not properly updated when the growfs is complete.
Therefore inodes will continue to be allocated only in the
AGs which existed prior to the growfs, and the new space
won't be utilized.
This is because of this path in xfs_growfs_data_private():
xfs_growfs_data_private
xfs_initialize_perag(mp, nagcount, &nagimax);
if (mp->m_flags & XFS_MOUNT_32BITINODES)
index = xfs_set_inode32(mp);
else
index = xfs_set_inode64(mp);
if (maxagi)
*maxagi = index;
where xfs_set_inode* iterates over the (old) agcount in
mp->m_sb.sb_agblocks, which has not yet been updated
in the growfs path. So "index" will be returned based on
the old agcount, not the new one, and new AGs are not available
for inode allocation.
Fix this by explicitly passing the proper AG count (which
xfs_initialize_perag() already has) down another level,
so that xfs_set_inode* can make the proper decision about
acceptable AGs for inode allocation in the potentially
newly-added AGs.
This has been broken since 3.7, when these two
xfs_set_inode* functions were added in commit 2d2194f.
Prior to that, we looped over "agcount" not sb_agblocks
in these calculations.
Signed-off-by: Eric Sandeen <sandeen@redhat.com>
Reviewed-by: Brian Foster <bfoster@redhat.com>
Signed-off-by: Dave Chinner <david@fromorbit.com>
2014-07-24 14:51:54 +04:00
|
|
|
max_metadata = agcount;
|
2012-09-20 17:32:38 +04:00
|
|
|
}
|
|
|
|
|
2016-03-02 01:58:09 +03:00
|
|
|
/* Get the last possible inode in the filesystem */
|
2018-12-12 19:46:24 +03:00
|
|
|
agino = XFS_AGB_TO_AGINO(mp, sbp->sb_agblocks - 1);
|
2016-03-02 01:58:09 +03:00
|
|
|
ino = XFS_AGINO_TO_INO(mp, agcount - 1, agino);
|
|
|
|
|
|
|
|
/*
|
|
|
|
* If user asked for no more than 32-bit inodes, and the fs is
|
|
|
|
* sufficiently large, set XFS_MOUNT_32BITINODES if we must alter
|
|
|
|
* the allocator to accommodate the request.
|
|
|
|
*/
|
|
|
|
if ((mp->m_flags & XFS_MOUNT_SMALL_INUMS) && ino > XFS_MAXINUMBER_32)
|
|
|
|
mp->m_flags |= XFS_MOUNT_32BITINODES;
|
|
|
|
else
|
|
|
|
mp->m_flags &= ~XFS_MOUNT_32BITINODES;
|
2014-07-24 14:53:10 +04:00
|
|
|
|
xfs: allow inode allocations in post-growfs disk space
Today, if we perform an xfs_growfs which adds allocation groups,
mp->m_maxagi is not properly updated when the growfs is complete.
Therefore inodes will continue to be allocated only in the
AGs which existed prior to the growfs, and the new space
won't be utilized.
This is because of this path in xfs_growfs_data_private():
xfs_growfs_data_private
xfs_initialize_perag(mp, nagcount, &nagimax);
if (mp->m_flags & XFS_MOUNT_32BITINODES)
index = xfs_set_inode32(mp);
else
index = xfs_set_inode64(mp);
if (maxagi)
*maxagi = index;
where xfs_set_inode* iterates over the (old) agcount in
mp->m_sb.sb_agblocks, which has not yet been updated
in the growfs path. So "index" will be returned based on
the old agcount, not the new one, and new AGs are not available
for inode allocation.
Fix this by explicitly passing the proper AG count (which
xfs_initialize_perag() already has) down another level,
so that xfs_set_inode* can make the proper decision about
acceptable AGs for inode allocation in the potentially
newly-added AGs.
This has been broken since 3.7, when these two
xfs_set_inode* functions were added in commit 2d2194f.
Prior to that, we looped over "agcount" not sb_agblocks
in these calculations.
Signed-off-by: Eric Sandeen <sandeen@redhat.com>
Reviewed-by: Brian Foster <bfoster@redhat.com>
Signed-off-by: Dave Chinner <david@fromorbit.com>
2014-07-24 14:51:54 +04:00
|
|
|
for (index = 0; index < agcount; index++) {
|
2016-03-02 01:58:09 +03:00
|
|
|
struct xfs_perag *pag;
|
2012-09-20 17:32:40 +04:00
|
|
|
|
2016-03-02 01:58:09 +03:00
|
|
|
ino = XFS_AGINO_TO_INO(mp, index, agino);
|
2012-09-20 17:32:38 +04:00
|
|
|
|
|
|
|
pag = xfs_perag_get(mp, index);
|
|
|
|
|
2016-03-02 01:58:09 +03:00
|
|
|
if (mp->m_flags & XFS_MOUNT_32BITINODES) {
|
|
|
|
if (ino > XFS_MAXINUMBER_32) {
|
|
|
|
pag->pagi_inodeok = 0;
|
|
|
|
pag->pagf_metadata = 0;
|
|
|
|
} else {
|
|
|
|
pag->pagi_inodeok = 1;
|
|
|
|
maxagi++;
|
|
|
|
if (index < max_metadata)
|
|
|
|
pag->pagf_metadata = 1;
|
|
|
|
else
|
|
|
|
pag->pagf_metadata = 0;
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
pag->pagi_inodeok = 1;
|
|
|
|
pag->pagf_metadata = 0;
|
|
|
|
}
|
2012-09-20 17:32:38 +04:00
|
|
|
|
|
|
|
xfs_perag_put(pag);
|
|
|
|
}
|
|
|
|
|
2016-03-02 01:58:09 +03:00
|
|
|
return (mp->m_flags & XFS_MOUNT_32BITINODES) ? maxagi : agcount;
|
2012-09-20 17:32:38 +04:00
|
|
|
}
|
|
|
|
|
2009-03-04 21:34:10 +03:00
|
|
|
STATIC int
|
2005-04-17 02:20:36 +04:00
|
|
|
xfs_blkdev_get(
|
|
|
|
xfs_mount_t *mp,
|
|
|
|
const char *name,
|
|
|
|
struct block_device **bdevp)
|
|
|
|
{
|
|
|
|
int error = 0;
|
|
|
|
|
2010-11-13 13:55:18 +03:00
|
|
|
*bdevp = blkdev_get_by_path(name, FMODE_READ|FMODE_WRITE|FMODE_EXCL,
|
|
|
|
mp);
|
2005-04-17 02:20:36 +04:00
|
|
|
if (IS_ERR(*bdevp)) {
|
|
|
|
error = PTR_ERR(*bdevp);
|
2014-12-24 01:47:27 +03:00
|
|
|
xfs_warn(mp, "Invalid device [%s], error=%d", name, error);
|
2005-04-17 02:20:36 +04:00
|
|
|
}
|
|
|
|
|
2014-06-25 08:58:08 +04:00
|
|
|
return error;
|
2005-04-17 02:20:36 +04:00
|
|
|
}
|
|
|
|
|
2009-03-04 21:34:10 +03:00
|
|
|
STATIC void
|
2005-04-17 02:20:36 +04:00
|
|
|
xfs_blkdev_put(
|
|
|
|
struct block_device *bdev)
|
|
|
|
{
|
|
|
|
if (bdev)
|
block: make blkdev_get/put() handle exclusive access
Over time, block layer has accumulated a set of APIs dealing with bdev
open, close, claim and release.
* blkdev_get/put() are the primary open and close functions.
* bd_claim/release() deal with exclusive open.
* open/close_bdev_exclusive() are combination of open and claim and
the other way around, respectively.
* bd_link/unlink_disk_holder() to create and remove holder/slave
symlinks.
* open_by_devnum() wraps bdget() + blkdev_get().
The interface is a bit confusing and the decoupling of open and claim
makes it impossible to properly guarantee exclusive access as
in-kernel open + claim sequence can disturb the existing exclusive
open even before the block layer knows the current open if for another
exclusive access. Reorganize the interface such that,
* blkdev_get() is extended to include exclusive access management.
@holder argument is added and, if is @FMODE_EXCL specified, it will
gain exclusive access atomically w.r.t. other exclusive accesses.
* blkdev_put() is similarly extended. It now takes @mode argument and
if @FMODE_EXCL is set, it releases an exclusive access. Also, when
the last exclusive claim is released, the holder/slave symlinks are
removed automatically.
* bd_claim/release() and close_bdev_exclusive() are no longer
necessary and either made static or removed.
* bd_link_disk_holder() remains the same but bd_unlink_disk_holder()
is no longer necessary and removed.
* open_bdev_exclusive() becomes a simple wrapper around lookup_bdev()
and blkdev_get(). It also has an unexpected extra bdev_read_only()
test which probably should be moved into blkdev_get().
* open_by_devnum() is modified to take @holder argument and pass it to
blkdev_get().
Most of bdev open/close operations are unified into blkdev_get/put()
and most exclusive accesses are tested atomically at the open time (as
it should). This cleans up code and removes some, both valid and
invalid, but unnecessary all the same, corner cases.
open_bdev_exclusive() and open_by_devnum() can use further cleanup -
rename to blkdev_get_by_path() and blkdev_get_by_devt() and drop
special features. Well, let's leave them for another day.
Most conversions are straight-forward. drbd conversion is a bit more
involved as there was some reordering, but the logic should stay the
same.
Signed-off-by: Tejun Heo <tj@kernel.org>
Acked-by: Neil Brown <neilb@suse.de>
Acked-by: Ryusuke Konishi <konishi.ryusuke@lab.ntt.co.jp>
Acked-by: Mike Snitzer <snitzer@redhat.com>
Acked-by: Philipp Reisner <philipp.reisner@linbit.com>
Cc: Peter Osterlund <petero2@telia.com>
Cc: Martin Schwidefsky <schwidefsky@de.ibm.com>
Cc: Heiko Carstens <heiko.carstens@de.ibm.com>
Cc: Jan Kara <jack@suse.cz>
Cc: Andrew Morton <akpm@linux-foundation.org>
Cc: Andreas Dilger <adilger.kernel@dilger.ca>
Cc: "Theodore Ts'o" <tytso@mit.edu>
Cc: Mark Fasheh <mfasheh@suse.com>
Cc: Joel Becker <joel.becker@oracle.com>
Cc: Alex Elder <aelder@sgi.com>
Cc: Christoph Hellwig <hch@infradead.org>
Cc: dm-devel@redhat.com
Cc: drbd-dev@lists.linbit.com
Cc: Leo Chen <leochen@broadcom.com>
Cc: Scott Branden <sbranden@broadcom.com>
Cc: Chris Mason <chris.mason@oracle.com>
Cc: Steven Whitehouse <swhiteho@redhat.com>
Cc: Dave Kleikamp <shaggy@linux.vnet.ibm.com>
Cc: Joern Engel <joern@logfs.org>
Cc: reiserfs-devel@vger.kernel.org
Cc: Alexander Viro <viro@zeniv.linux.org.uk>
2010-11-13 13:55:17 +03:00
|
|
|
blkdev_put(bdev, FMODE_READ|FMODE_WRITE|FMODE_EXCL);
|
2005-04-17 02:20:36 +04:00
|
|
|
}
|
|
|
|
|
2005-11-02 02:26:59 +03:00
|
|
|
void
|
|
|
|
xfs_blkdev_issue_flush(
|
|
|
|
xfs_buftarg_t *buftarg)
|
|
|
|
{
|
2012-04-24 17:23:46 +04:00
|
|
|
blkdev_issue_flush(buftarg->bt_bdev, GFP_NOFS, NULL);
|
2005-11-02 02:26:59 +03:00
|
|
|
}
|
2005-04-17 02:20:36 +04:00
|
|
|
|
2008-05-20 05:31:13 +04:00
|
|
|
STATIC void
|
|
|
|
xfs_close_devices(
|
|
|
|
struct xfs_mount *mp)
|
|
|
|
{
|
2017-08-25 01:12:50 +03:00
|
|
|
struct dax_device *dax_ddev = mp->m_ddev_targp->bt_daxdev;
|
|
|
|
|
2008-05-20 05:31:13 +04:00
|
|
|
if (mp->m_logdev_targp && mp->m_logdev_targp != mp->m_ddev_targp) {
|
2008-07-18 11:13:12 +04:00
|
|
|
struct block_device *logdev = mp->m_logdev_targp->bt_bdev;
|
2017-08-25 01:12:50 +03:00
|
|
|
struct dax_device *dax_logdev = mp->m_logdev_targp->bt_daxdev;
|
|
|
|
|
2018-04-06 20:09:42 +03:00
|
|
|
xfs_free_buftarg(mp->m_logdev_targp);
|
2008-07-18 11:13:12 +04:00
|
|
|
xfs_blkdev_put(logdev);
|
2017-08-25 01:12:50 +03:00
|
|
|
fs_put_dax(dax_logdev);
|
2008-05-20 05:31:13 +04:00
|
|
|
}
|
|
|
|
if (mp->m_rtdev_targp) {
|
2008-07-18 11:13:12 +04:00
|
|
|
struct block_device *rtdev = mp->m_rtdev_targp->bt_bdev;
|
2017-08-25 01:12:50 +03:00
|
|
|
struct dax_device *dax_rtdev = mp->m_rtdev_targp->bt_daxdev;
|
|
|
|
|
2018-04-06 20:09:42 +03:00
|
|
|
xfs_free_buftarg(mp->m_rtdev_targp);
|
2008-07-18 11:13:12 +04:00
|
|
|
xfs_blkdev_put(rtdev);
|
2017-08-25 01:12:50 +03:00
|
|
|
fs_put_dax(dax_rtdev);
|
2008-05-20 05:31:13 +04:00
|
|
|
}
|
2018-04-06 20:09:42 +03:00
|
|
|
xfs_free_buftarg(mp->m_ddev_targp);
|
2017-08-25 01:12:50 +03:00
|
|
|
fs_put_dax(dax_ddev);
|
2008-05-20 05:31:13 +04:00
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* The file system configurations are:
|
|
|
|
* (1) device (partition) with data and internal log
|
|
|
|
* (2) logical volume with data and log subvolumes.
|
|
|
|
* (3) logical volume with data, log, and realtime subvolumes.
|
|
|
|
*
|
|
|
|
* We only have to handle opening the log and realtime volumes here if
|
|
|
|
* they are present. The data subvolume has already been opened by
|
|
|
|
* get_sb_bdev() and is stored in sb->s_bdev.
|
|
|
|
*/
|
|
|
|
STATIC int
|
|
|
|
xfs_open_devices(
|
2008-10-30 09:53:24 +03:00
|
|
|
struct xfs_mount *mp)
|
2008-05-20 05:31:13 +04:00
|
|
|
{
|
|
|
|
struct block_device *ddev = mp->m_super->s_bdev;
|
2017-08-25 01:12:50 +03:00
|
|
|
struct dax_device *dax_ddev = fs_dax_get_by_bdev(ddev);
|
|
|
|
struct dax_device *dax_logdev = NULL, *dax_rtdev = NULL;
|
2008-05-20 05:31:13 +04:00
|
|
|
struct block_device *logdev = NULL, *rtdev = NULL;
|
|
|
|
int error;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Open real time and log devices - order is important.
|
|
|
|
*/
|
2008-10-30 09:53:24 +03:00
|
|
|
if (mp->m_logname) {
|
|
|
|
error = xfs_blkdev_get(mp, mp->m_logname, &logdev);
|
2008-05-20 05:31:13 +04:00
|
|
|
if (error)
|
|
|
|
goto out;
|
2017-08-25 01:12:50 +03:00
|
|
|
dax_logdev = fs_dax_get_by_bdev(logdev);
|
2008-05-20 05:31:13 +04:00
|
|
|
}
|
|
|
|
|
2008-10-30 09:53:24 +03:00
|
|
|
if (mp->m_rtname) {
|
|
|
|
error = xfs_blkdev_get(mp, mp->m_rtname, &rtdev);
|
2008-05-20 05:31:13 +04:00
|
|
|
if (error)
|
|
|
|
goto out_close_logdev;
|
|
|
|
|
|
|
|
if (rtdev == ddev || rtdev == logdev) {
|
2011-03-07 02:00:35 +03:00
|
|
|
xfs_warn(mp,
|
|
|
|
"Cannot mount filesystem with identical rtdev and ddev/logdev.");
|
2014-06-25 08:58:08 +04:00
|
|
|
error = -EINVAL;
|
2008-05-20 05:31:13 +04:00
|
|
|
goto out_close_rtdev;
|
|
|
|
}
|
2017-08-25 01:12:50 +03:00
|
|
|
dax_rtdev = fs_dax_get_by_bdev(rtdev);
|
2008-05-20 05:31:13 +04:00
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* Setup xfs_mount buffer target pointers
|
|
|
|
*/
|
2014-06-25 08:58:08 +04:00
|
|
|
error = -ENOMEM;
|
2017-08-25 01:12:50 +03:00
|
|
|
mp->m_ddev_targp = xfs_alloc_buftarg(mp, ddev, dax_ddev);
|
2008-05-20 05:31:13 +04:00
|
|
|
if (!mp->m_ddev_targp)
|
|
|
|
goto out_close_rtdev;
|
|
|
|
|
|
|
|
if (rtdev) {
|
2017-08-25 01:12:50 +03:00
|
|
|
mp->m_rtdev_targp = xfs_alloc_buftarg(mp, rtdev, dax_rtdev);
|
2008-05-20 05:31:13 +04:00
|
|
|
if (!mp->m_rtdev_targp)
|
|
|
|
goto out_free_ddev_targ;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (logdev && logdev != ddev) {
|
2017-08-25 01:12:50 +03:00
|
|
|
mp->m_logdev_targp = xfs_alloc_buftarg(mp, logdev, dax_logdev);
|
2008-05-20 05:31:13 +04:00
|
|
|
if (!mp->m_logdev_targp)
|
|
|
|
goto out_free_rtdev_targ;
|
|
|
|
} else {
|
|
|
|
mp->m_logdev_targp = mp->m_ddev_targp;
|
|
|
|
}
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
out_free_rtdev_targ:
|
|
|
|
if (mp->m_rtdev_targp)
|
2018-04-06 20:09:42 +03:00
|
|
|
xfs_free_buftarg(mp->m_rtdev_targp);
|
2008-05-20 05:31:13 +04:00
|
|
|
out_free_ddev_targ:
|
2018-04-06 20:09:42 +03:00
|
|
|
xfs_free_buftarg(mp->m_ddev_targp);
|
2008-05-20 05:31:13 +04:00
|
|
|
out_close_rtdev:
|
2014-12-01 00:24:20 +03:00
|
|
|
xfs_blkdev_put(rtdev);
|
2017-08-25 01:12:50 +03:00
|
|
|
fs_put_dax(dax_rtdev);
|
2008-05-20 05:31:13 +04:00
|
|
|
out_close_logdev:
|
2017-08-25 01:12:50 +03:00
|
|
|
if (logdev && logdev != ddev) {
|
2008-05-20 05:31:13 +04:00
|
|
|
xfs_blkdev_put(logdev);
|
2017-08-25 01:12:50 +03:00
|
|
|
fs_put_dax(dax_logdev);
|
|
|
|
}
|
2008-05-20 05:31:13 +04:00
|
|
|
out:
|
2017-08-25 01:12:50 +03:00
|
|
|
fs_put_dax(dax_ddev);
|
2008-05-20 05:31:13 +04:00
|
|
|
return error;
|
|
|
|
}
|
|
|
|
|
2008-05-20 09:10:36 +04:00
|
|
|
/*
|
|
|
|
* Setup xfs_mount buffer target pointers based on superblock
|
|
|
|
*/
|
|
|
|
STATIC int
|
|
|
|
xfs_setup_devices(
|
|
|
|
struct xfs_mount *mp)
|
|
|
|
{
|
|
|
|
int error;
|
2008-05-20 05:31:13 +04:00
|
|
|
|
2014-04-14 13:00:29 +04:00
|
|
|
error = xfs_setsize_buftarg(mp->m_ddev_targp, mp->m_sb.sb_sectsize);
|
2008-05-20 09:10:36 +04:00
|
|
|
if (error)
|
|
|
|
return error;
|
|
|
|
|
|
|
|
if (mp->m_logdev_targp && mp->m_logdev_targp != mp->m_ddev_targp) {
|
|
|
|
unsigned int log_sector_size = BBSIZE;
|
|
|
|
|
|
|
|
if (xfs_sb_version_hassector(&mp->m_sb))
|
|
|
|
log_sector_size = mp->m_sb.sb_logsectsize;
|
|
|
|
error = xfs_setsize_buftarg(mp->m_logdev_targp,
|
|
|
|
log_sector_size);
|
|
|
|
if (error)
|
|
|
|
return error;
|
|
|
|
}
|
|
|
|
if (mp->m_rtdev_targp) {
|
|
|
|
error = xfs_setsize_buftarg(mp->m_rtdev_targp,
|
|
|
|
mp->m_sb.sb_sectsize);
|
|
|
|
if (error)
|
|
|
|
return error;
|
|
|
|
}
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
2008-05-20 05:31:13 +04:00
|
|
|
|
2012-02-29 13:53:48 +04:00
|
|
|
STATIC int
|
|
|
|
xfs_init_mount_workqueues(
|
|
|
|
struct xfs_mount *mp)
|
|
|
|
{
|
2014-11-28 05:59:58 +03:00
|
|
|
mp->m_buf_workqueue = alloc_workqueue("xfs-buf/%s",
|
2014-12-04 01:43:17 +03:00
|
|
|
WQ_MEM_RECLAIM|WQ_FREEZABLE, 1, mp->m_fsname);
|
2014-11-28 05:59:58 +03:00
|
|
|
if (!mp->m_buf_workqueue)
|
|
|
|
goto out;
|
|
|
|
|
2012-02-29 13:53:48 +04:00
|
|
|
mp->m_unwritten_workqueue = alloc_workqueue("xfs-conv/%s",
|
2014-09-09 05:44:46 +04:00
|
|
|
WQ_MEM_RECLAIM|WQ_FREEZABLE, 0, mp->m_fsname);
|
2012-02-29 13:53:48 +04:00
|
|
|
if (!mp->m_unwritten_workqueue)
|
2019-04-15 23:13:21 +03:00
|
|
|
goto out_destroy_buf;
|
2012-02-29 13:53:48 +04:00
|
|
|
|
2012-04-23 11:54:32 +04:00
|
|
|
mp->m_cil_workqueue = alloc_workqueue("xfs-cil/%s",
|
xfs: prevent CIL push holdoff in log recovery
generic/530 on a machine with enough ram and a non-preemptible
kernel can run the AGI processing phase of log recovery enitrely out
of cache. This means it never blocks on locks, never waits for IO
and runs entirely through the unlinked lists until it either
completes or blocks and hangs because it has run out of log space.
It runs out of log space because the background CIL push is
scheduled but never runs. queue_work() queues the CIL work on the
current CPU that is busy, and the workqueue code will not run it on
any other CPU. Hence if the unlinked list processing never yields
the CPU voluntarily, the push work is delayed indefinitely. This
results in the CIL aggregating changes until all the log space is
consumed.
When the log recoveyr processing evenutally blocks, the CIL flushes
but because the last iclog isn't submitted for IO because it isn't
full, the CIL flush never completes and nothing ever moves the log
head forwards, or indeed inserts anything into the tail of the log,
and hence nothing is able to get the log moving again and recovery
hangs.
There are several problems here, but the two obvious ones from
the trace are that:
a) log recovery does not yield the CPU for over 4 seconds,
b) binding CIL pushes to a single CPU is a really bad idea.
This patch addresses just these two aspects of the problem, and are
suitable for backporting to work around any issues in older kernels.
The more fundamental problem of preventing the CIL from consuming
more than 50% of the log without committing will take more invasive
and complex work, so will be done as followup work.
Signed-off-by: Dave Chinner <dchinner@redhat.com>
Reviewed-by: Darrick J. Wong <darrick.wong@oracle.com>
Signed-off-by: Darrick J. Wong <darrick.wong@oracle.com>
2019-09-06 07:35:39 +03:00
|
|
|
WQ_MEM_RECLAIM | WQ_FREEZABLE | WQ_UNBOUND,
|
|
|
|
0, mp->m_fsname);
|
2012-04-23 11:54:32 +04:00
|
|
|
if (!mp->m_cil_workqueue)
|
|
|
|
goto out_destroy_unwritten;
|
2012-10-08 14:56:05 +04:00
|
|
|
|
|
|
|
mp->m_reclaim_workqueue = alloc_workqueue("xfs-reclaim/%s",
|
2015-11-10 02:10:34 +03:00
|
|
|
WQ_MEM_RECLAIM|WQ_FREEZABLE, 0, mp->m_fsname);
|
2012-10-08 14:56:05 +04:00
|
|
|
if (!mp->m_reclaim_workqueue)
|
|
|
|
goto out_destroy_cil;
|
|
|
|
|
2012-11-06 18:50:47 +04:00
|
|
|
mp->m_eofblocks_workqueue = alloc_workqueue("xfs-eofblocks/%s",
|
2015-11-10 02:10:34 +03:00
|
|
|
WQ_MEM_RECLAIM|WQ_FREEZABLE, 0, mp->m_fsname);
|
2012-11-06 18:50:47 +04:00
|
|
|
if (!mp->m_eofblocks_workqueue)
|
2019-06-29 05:27:25 +03:00
|
|
|
goto out_destroy_reclaim;
|
2012-11-06 18:50:47 +04:00
|
|
|
|
2017-03-29 00:51:44 +03:00
|
|
|
mp->m_sync_workqueue = alloc_workqueue("xfs-sync/%s", WQ_FREEZABLE, 0,
|
|
|
|
mp->m_fsname);
|
|
|
|
if (!mp->m_sync_workqueue)
|
|
|
|
goto out_destroy_eofb;
|
|
|
|
|
2012-02-29 13:53:48 +04:00
|
|
|
return 0;
|
|
|
|
|
2017-03-29 00:51:44 +03:00
|
|
|
out_destroy_eofb:
|
|
|
|
destroy_workqueue(mp->m_eofblocks_workqueue);
|
2012-10-08 14:56:05 +04:00
|
|
|
out_destroy_reclaim:
|
|
|
|
destroy_workqueue(mp->m_reclaim_workqueue);
|
|
|
|
out_destroy_cil:
|
|
|
|
destroy_workqueue(mp->m_cil_workqueue);
|
2012-04-23 11:54:32 +04:00
|
|
|
out_destroy_unwritten:
|
|
|
|
destroy_workqueue(mp->m_unwritten_workqueue);
|
2014-11-28 05:59:58 +03:00
|
|
|
out_destroy_buf:
|
|
|
|
destroy_workqueue(mp->m_buf_workqueue);
|
2012-02-29 13:53:48 +04:00
|
|
|
out:
|
|
|
|
return -ENOMEM;
|
|
|
|
}
|
|
|
|
|
|
|
|
STATIC void
|
|
|
|
xfs_destroy_mount_workqueues(
|
|
|
|
struct xfs_mount *mp)
|
|
|
|
{
|
2017-03-29 00:51:44 +03:00
|
|
|
destroy_workqueue(mp->m_sync_workqueue);
|
2012-11-06 18:50:47 +04:00
|
|
|
destroy_workqueue(mp->m_eofblocks_workqueue);
|
2012-10-08 14:56:05 +04:00
|
|
|
destroy_workqueue(mp->m_reclaim_workqueue);
|
2012-04-23 11:54:32 +04:00
|
|
|
destroy_workqueue(mp->m_cil_workqueue);
|
2012-02-29 13:53:48 +04:00
|
|
|
destroy_workqueue(mp->m_unwritten_workqueue);
|
2014-11-28 05:59:58 +03:00
|
|
|
destroy_workqueue(mp->m_buf_workqueue);
|
2012-02-29 13:53:48 +04:00
|
|
|
}
|
|
|
|
|
2012-10-08 14:56:04 +04:00
|
|
|
/*
|
|
|
|
* Flush all dirty data to disk. Must not be called while holding an XFS_ILOCK
|
|
|
|
* or a page lock. We use sync_inodes_sb() here to ensure we block while waiting
|
|
|
|
* for IO to complete so that we effectively throttle multiple callers to the
|
|
|
|
* rate at which IO is completing.
|
|
|
|
*/
|
|
|
|
void
|
|
|
|
xfs_flush_inodes(
|
|
|
|
struct xfs_mount *mp)
|
|
|
|
{
|
|
|
|
struct super_block *sb = mp->m_super;
|
|
|
|
|
|
|
|
if (down_read_trylock(&sb->s_umount)) {
|
2014-02-21 14:19:04 +04:00
|
|
|
sync_inodes_sb(sb);
|
2012-10-08 14:56:04 +04:00
|
|
|
up_read(&sb->s_umount);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2008-10-30 09:36:14 +03:00
|
|
|
/* Catch misguided souls that try to use this interface on XFS */
|
2005-04-17 02:20:36 +04:00
|
|
|
STATIC struct inode *
|
2006-03-14 06:06:18 +03:00
|
|
|
xfs_fs_alloc_inode(
|
2005-04-17 02:20:36 +04:00
|
|
|
struct super_block *sb)
|
|
|
|
{
|
2008-10-30 09:36:14 +03:00
|
|
|
BUG();
|
2008-10-30 09:36:52 +03:00
|
|
|
return NULL;
|
2005-04-17 02:20:36 +04:00
|
|
|
}
|
|
|
|
|
2018-10-18 09:20:11 +03:00
|
|
|
#ifdef DEBUG
|
|
|
|
static void
|
|
|
|
xfs_check_delalloc(
|
|
|
|
struct xfs_inode *ip,
|
|
|
|
int whichfork)
|
|
|
|
{
|
|
|
|
struct xfs_ifork *ifp = XFS_IFORK_PTR(ip, whichfork);
|
|
|
|
struct xfs_bmbt_irec got;
|
|
|
|
struct xfs_iext_cursor icur;
|
|
|
|
|
|
|
|
if (!ifp || !xfs_iext_lookup_extent(ip, ifp, 0, &icur, &got))
|
|
|
|
return;
|
|
|
|
do {
|
|
|
|
if (isnullstartblock(got.br_startblock)) {
|
|
|
|
xfs_warn(ip->i_mount,
|
|
|
|
"ino %llx %s fork has delalloc extent at [0x%llx:0x%llx]",
|
|
|
|
ip->i_ino,
|
|
|
|
whichfork == XFS_DATA_FORK ? "data" : "cow",
|
|
|
|
got.br_startoff, got.br_blockcount);
|
|
|
|
}
|
|
|
|
} while (xfs_iext_next_extent(ifp, &icur, &got));
|
|
|
|
}
|
|
|
|
#else
|
|
|
|
#define xfs_check_delalloc(ip, whichfork) do { } while (0)
|
|
|
|
#endif
|
|
|
|
|
2008-10-30 09:36:14 +03:00
|
|
|
/*
|
2008-10-30 09:36:40 +03:00
|
|
|
* Now that the generic code is guaranteed not to be accessing
|
2016-05-18 06:52:42 +03:00
|
|
|
* the linux inode, we can inactivate and reclaim the inode.
|
2008-10-30 09:36:14 +03:00
|
|
|
*/
|
2005-04-17 02:20:36 +04:00
|
|
|
STATIC void
|
2006-03-14 06:06:18 +03:00
|
|
|
xfs_fs_destroy_inode(
|
2009-09-29 17:48:56 +04:00
|
|
|
struct inode *inode)
|
2005-04-17 02:20:36 +04:00
|
|
|
{
|
2009-09-29 17:48:56 +04:00
|
|
|
struct xfs_inode *ip = XFS_I(inode);
|
|
|
|
|
2010-06-24 05:57:09 +04:00
|
|
|
trace_xfs_destroy_inode(ip);
|
2008-10-30 09:36:40 +03:00
|
|
|
|
2016-11-30 06:33:25 +03:00
|
|
|
ASSERT(!rwsem_is_locked(&inode->i_rwsem));
|
2016-05-18 06:52:42 +03:00
|
|
|
XFS_STATS_INC(ip->i_mount, vn_rele);
|
|
|
|
XFS_STATS_INC(ip->i_mount, vn_remove);
|
|
|
|
|
|
|
|
xfs_inactive(ip);
|
2009-09-29 17:48:56 +04:00
|
|
|
|
2018-10-18 09:20:11 +03:00
|
|
|
if (!XFS_FORCED_SHUTDOWN(ip->i_mount) && ip->i_delayed_blks) {
|
|
|
|
xfs_check_delalloc(ip, XFS_DATA_FORK);
|
|
|
|
xfs_check_delalloc(ip, XFS_COW_FORK);
|
|
|
|
ASSERT(0);
|
|
|
|
}
|
|
|
|
|
2016-05-18 06:52:42 +03:00
|
|
|
XFS_STATS_INC(ip->i_mount, vn_reclaim);
|
2009-09-29 17:48:56 +04:00
|
|
|
|
|
|
|
/*
|
|
|
|
* We should never get here with one of the reclaim flags already set.
|
|
|
|
*/
|
|
|
|
ASSERT_ALWAYS(!xfs_iflags_test(ip, XFS_IRECLAIMABLE));
|
|
|
|
ASSERT_ALWAYS(!xfs_iflags_test(ip, XFS_IRECLAIM));
|
|
|
|
|
|
|
|
/*
|
2010-01-11 02:51:47 +03:00
|
|
|
* We always use background reclaim here because even if the
|
|
|
|
* inode is clean, it still may be under IO and hence we have
|
|
|
|
* to take the flush lock. The background reclaim path handles
|
|
|
|
* this more efficiently than we can here, so simply let background
|
|
|
|
* reclaim tear down all inodes.
|
2009-09-29 17:48:56 +04:00
|
|
|
*/
|
2010-01-11 02:51:47 +03:00
|
|
|
xfs_inode_set_reclaim_tag(ip);
|
2005-04-17 02:20:36 +04:00
|
|
|
}
|
|
|
|
|
2018-03-07 04:04:00 +03:00
|
|
|
static void
|
|
|
|
xfs_fs_dirty_inode(
|
|
|
|
struct inode *inode,
|
|
|
|
int flag)
|
|
|
|
{
|
|
|
|
struct xfs_inode *ip = XFS_I(inode);
|
|
|
|
struct xfs_mount *mp = ip->i_mount;
|
|
|
|
struct xfs_trans *tp;
|
|
|
|
|
|
|
|
if (!(inode->i_sb->s_flags & SB_LAZYTIME))
|
|
|
|
return;
|
|
|
|
if (flag != I_DIRTY_SYNC || !(inode->i_state & I_DIRTY_TIME))
|
|
|
|
return;
|
|
|
|
|
|
|
|
if (xfs_trans_alloc(mp, &M_RES(mp)->tr_fsyncts, 0, 0, 0, &tp))
|
|
|
|
return;
|
|
|
|
xfs_ilock(ip, XFS_ILOCK_EXCL);
|
|
|
|
xfs_trans_ijoin(tp, ip, XFS_ILOCK_EXCL);
|
|
|
|
xfs_trans_log_inode(tp, ip, XFS_ILOG_TIMESTAMP);
|
|
|
|
xfs_trans_commit(tp);
|
|
|
|
}
|
|
|
|
|
2008-10-30 08:11:59 +03:00
|
|
|
/*
|
|
|
|
* Slab object creation initialisation for the XFS inode.
|
|
|
|
* This covers only the idempotent fields in the XFS inode;
|
|
|
|
* all other fields need to be initialised on allocation
|
tree-wide: fix comment/printk typos
"gadget", "through", "command", "maintain", "maintain", "controller", "address",
"between", "initiali[zs]e", "instead", "function", "select", "already",
"equal", "access", "management", "hierarchy", "registration", "interest",
"relative", "memory", "offset", "already",
Signed-off-by: Uwe Kleine-König <u.kleine-koenig@pengutronix.de>
Signed-off-by: Jiri Kosina <jkosina@suse.cz>
2010-11-01 22:38:34 +03:00
|
|
|
* from the slab. This avoids the need to repeatedly initialise
|
2008-10-30 08:11:59 +03:00
|
|
|
* fields in the xfs inode that left in the initialise state
|
|
|
|
* when freeing the inode.
|
|
|
|
*/
|
2008-10-30 09:36:14 +03:00
|
|
|
STATIC void
|
|
|
|
xfs_fs_inode_init_once(
|
2008-10-30 08:11:59 +03:00
|
|
|
void *inode)
|
|
|
|
{
|
|
|
|
struct xfs_inode *ip = inode;
|
|
|
|
|
|
|
|
memset(ip, 0, sizeof(struct xfs_inode));
|
2008-10-30 09:36:14 +03:00
|
|
|
|
|
|
|
/* vfs inode */
|
|
|
|
inode_init_once(VFS_I(ip));
|
|
|
|
|
|
|
|
/* xfs inode */
|
2008-10-30 08:11:59 +03:00
|
|
|
atomic_set(&ip->i_pincount, 0);
|
|
|
|
spin_lock_init(&ip->i_flags_lock);
|
|
|
|
|
2015-02-23 13:43:37 +03:00
|
|
|
mrlock_init(&ip->i_mmaplock, MRLOCK_ALLOW_EQUAL_PRI|MRLOCK_BARRIER,
|
|
|
|
"xfsino", ip->i_ino);
|
2008-10-30 08:11:59 +03:00
|
|
|
mrlock_init(&ip->i_lock, MRLOCK_ALLOW_EQUAL_PRI|MRLOCK_BARRIER,
|
|
|
|
"xfsino", ip->i_ino);
|
|
|
|
}
|
|
|
|
|
2012-03-22 09:15:10 +04:00
|
|
|
/*
|
|
|
|
* We do an unlocked check for XFS_IDONTCACHE here because we are already
|
|
|
|
* serialised against cache hits here via the inode->i_lock and igrab() in
|
|
|
|
* xfs_iget_cache_hit(). Hence a lookup that might clear this flag will not be
|
|
|
|
* racing with us, and it avoids needing to grab a spinlock here for every inode
|
|
|
|
* we drop the final reference on.
|
|
|
|
*/
|
|
|
|
STATIC int
|
|
|
|
xfs_fs_drop_inode(
|
|
|
|
struct inode *inode)
|
|
|
|
{
|
|
|
|
struct xfs_inode *ip = XFS_I(inode);
|
|
|
|
|
2016-10-03 19:11:29 +03:00
|
|
|
/*
|
|
|
|
* If this unlinked inode is in the middle of recovery, don't
|
|
|
|
* drop the inode just yet; log recovery will take care of
|
|
|
|
* that. See the comment for this inode flag.
|
|
|
|
*/
|
|
|
|
if (ip->i_flags & XFS_IRECOVERY) {
|
|
|
|
ASSERT(ip->i_mount->m_log->l_flags & XLOG_RECOVERY_NEEDED);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2012-03-22 09:15:10 +04:00
|
|
|
return generic_drop_inode(inode) || (ip->i_flags & XFS_IDONTCACHE);
|
|
|
|
}
|
|
|
|
|
2008-08-13 10:04:05 +04:00
|
|
|
STATIC void
|
|
|
|
xfs_free_fsname(
|
|
|
|
struct xfs_mount *mp)
|
|
|
|
{
|
|
|
|
kfree(mp->m_fsname);
|
|
|
|
kfree(mp->m_rtname);
|
|
|
|
kfree(mp->m_logname);
|
|
|
|
}
|
|
|
|
|
2005-04-17 02:20:36 +04:00
|
|
|
STATIC int
|
2009-10-07 00:29:28 +04:00
|
|
|
xfs_fs_sync_fs(
|
2005-04-17 02:20:36 +04:00
|
|
|
struct super_block *sb,
|
|
|
|
int wait)
|
|
|
|
{
|
2007-08-30 11:20:39 +04:00
|
|
|
struct xfs_mount *mp = XFS_M(sb);
|
2005-04-17 02:20:36 +04:00
|
|
|
|
2007-10-12 05:13:35 +04:00
|
|
|
/*
|
2011-12-07 01:58:12 +04:00
|
|
|
* Doing anything during the async pass would be counterproductive.
|
2007-10-12 05:13:35 +04:00
|
|
|
*/
|
2011-12-07 01:58:12 +04:00
|
|
|
if (!wait)
|
2009-10-07 00:29:28 +04:00
|
|
|
return 0;
|
|
|
|
|
xfs: xfs_sync_fsdata is redundant
Why do we need to write the superblock to disk once we've written
all the data? We don't actually - the reasons for doing this are
lost in the mists of time, and go back to the way Irix used to drive
VFS flushing.
On linux, this code is only called from two contexts: remount and
.sync_fs. In the remount case, the call is followed by a metadata
sync, which unpins and writes the superblock. In the sync_fs case,
we only need to force the log to disk to ensure that the superblock
is correctly on disk, so we don't actually need to write it. Hence
the functionality is either redundant or superfluous and thus can be
removed.
Seeing as xfs_quiesce_data is essentially now just a log force,
remove it as well and fold the code back into the two callers.
Neither of them need the log covering check, either, as that is
redundant for the remount case, and unnecessary for the .sync_fs
case.
Signed-off-by: Dave Chinner <dchinner@redhat.com>
Reviewed-by: Christoph Hellwig <hch@lst.de>
Reviewed-by: Mark Tinguely <tinguely@sgi.com>
Signed-off-by: Ben Myers <bpm@sgi.com>
2012-10-08 14:56:06 +04:00
|
|
|
xfs_log_force(mp, XFS_LOG_SYNC);
|
2009-10-07 00:29:28 +04:00
|
|
|
if (laptop_mode) {
|
2005-04-17 02:20:36 +04:00
|
|
|
/*
|
|
|
|
* The disk must be active because we're syncing.
|
2012-10-08 14:56:02 +04:00
|
|
|
* We schedule log work now (now that the disk is
|
2005-04-17 02:20:36 +04:00
|
|
|
* active) instead of later (when it might not be).
|
|
|
|
*/
|
2012-10-08 14:56:02 +04:00
|
|
|
flush_delayed_work(&mp->m_log->l_work);
|
2005-04-17 02:20:36 +04:00
|
|
|
}
|
|
|
|
|
2009-10-07 00:29:28 +04:00
|
|
|
return 0;
|
2005-04-17 02:20:36 +04:00
|
|
|
}
|
|
|
|
|
|
|
|
STATIC int
|
2006-03-14 06:06:18 +03:00
|
|
|
xfs_fs_statfs(
|
2006-06-23 13:02:58 +04:00
|
|
|
struct dentry *dentry,
|
2005-04-17 02:20:36 +04:00
|
|
|
struct kstatfs *statp)
|
|
|
|
{
|
2007-10-11 12:09:40 +04:00
|
|
|
struct xfs_mount *mp = XFS_M(dentry->d_sb);
|
|
|
|
xfs_sb_t *sbp = &mp->m_sb;
|
2015-03-18 01:25:59 +03:00
|
|
|
struct xfs_inode *ip = XFS_I(d_inode(dentry));
|
2017-06-16 21:00:05 +03:00
|
|
|
uint64_t fakeinos, id;
|
|
|
|
uint64_t icount;
|
|
|
|
uint64_t ifree;
|
|
|
|
uint64_t fdblocks;
|
2007-10-11 12:09:40 +04:00
|
|
|
xfs_extlen_t lsize;
|
2017-06-16 21:00:05 +03:00
|
|
|
int64_t ffree;
|
2007-10-11 12:09:40 +04:00
|
|
|
|
2018-10-18 09:20:19 +03:00
|
|
|
statp->f_type = XFS_SUPER_MAGIC;
|
2007-10-11 12:09:40 +04:00
|
|
|
statp->f_namelen = MAXNAMELEN - 1;
|
|
|
|
|
|
|
|
id = huge_encode_dev(mp->m_ddev_targp->bt_dev);
|
|
|
|
statp->f_fsid.val[0] = (u32)id;
|
|
|
|
statp->f_fsid.val[1] = (u32)(id >> 32);
|
|
|
|
|
2015-02-23 13:19:28 +03:00
|
|
|
icount = percpu_counter_sum(&mp->m_icount);
|
2015-02-23 13:19:53 +03:00
|
|
|
ifree = percpu_counter_sum(&mp->m_ifree);
|
2015-02-23 13:22:03 +03:00
|
|
|
fdblocks = percpu_counter_sum(&mp->m_fdblocks);
|
2007-10-11 12:09:40 +04:00
|
|
|
|
|
|
|
spin_lock(&mp->m_sb_lock);
|
|
|
|
statp->f_bsize = sbp->sb_blocksize;
|
|
|
|
lsize = sbp->sb_logstart ? sbp->sb_logblocks : 0;
|
|
|
|
statp->f_blocks = sbp->sb_dblocks - lsize;
|
2015-02-23 13:22:03 +03:00
|
|
|
spin_unlock(&mp->m_sb_lock);
|
|
|
|
|
2016-08-03 04:38:24 +03:00
|
|
|
statp->f_bfree = fdblocks - mp->m_alloc_set_aside;
|
2015-02-23 13:22:03 +03:00
|
|
|
statp->f_bavail = statp->f_bfree;
|
|
|
|
|
2018-12-12 19:46:24 +03:00
|
|
|
fakeinos = XFS_FSB_TO_INO(mp, statp->f_bfree);
|
2018-06-07 17:54:02 +03:00
|
|
|
statp->f_files = min(icount + fakeinos, (uint64_t)XFS_MAXINUMBER);
|
2019-06-05 21:19:34 +03:00
|
|
|
if (M_IGEO(mp)->maxicount)
|
2009-03-29 11:51:08 +04:00
|
|
|
statp->f_files = min_t(typeof(statp->f_files),
|
|
|
|
statp->f_files,
|
2019-06-05 21:19:34 +03:00
|
|
|
M_IGEO(mp)->maxicount);
|
2010-08-24 05:46:05 +04:00
|
|
|
|
2015-02-06 01:53:02 +03:00
|
|
|
/* If sb_icount overshot maxicount, report actual allocation */
|
|
|
|
statp->f_files = max_t(typeof(statp->f_files),
|
|
|
|
statp->f_files,
|
|
|
|
sbp->sb_icount);
|
|
|
|
|
2010-08-24 05:46:05 +04:00
|
|
|
/* make sure statp->f_ffree does not underflow */
|
2015-02-23 13:19:53 +03:00
|
|
|
ffree = statp->f_files - (icount - ifree);
|
2017-06-16 21:00:05 +03:00
|
|
|
statp->f_ffree = max_t(int64_t, ffree, 0);
|
2010-08-24 05:46:05 +04:00
|
|
|
|
2007-10-11 12:09:40 +04:00
|
|
|
|
2012-04-12 07:59:57 +04:00
|
|
|
if ((ip->i_d.di_flags & XFS_DIFLAG_PROJINHERIT) &&
|
2013-06-28 02:25:10 +04:00
|
|
|
((mp->m_qflags & (XFS_PQUOTA_ACCT|XFS_PQUOTA_ENFD))) ==
|
|
|
|
(XFS_PQUOTA_ACCT|XFS_PQUOTA_ENFD))
|
2009-06-08 17:33:32 +04:00
|
|
|
xfs_qm_statvfs(ip, statp);
|
2018-01-08 21:41:33 +03:00
|
|
|
|
|
|
|
if (XFS_IS_REALTIME_MOUNT(mp) &&
|
|
|
|
(ip->i_d.di_flags & (XFS_DIFLAG_RTINHERIT | XFS_DIFLAG_REALTIME))) {
|
|
|
|
statp->f_blocks = sbp->sb_rblocks;
|
|
|
|
statp->f_bavail = statp->f_bfree =
|
|
|
|
sbp->sb_frextents * sbp->sb_rextsize;
|
|
|
|
}
|
|
|
|
|
2007-10-11 12:09:40 +04:00
|
|
|
return 0;
|
2005-04-17 02:20:36 +04:00
|
|
|
}
|
|
|
|
|
2010-02-06 01:59:53 +03:00
|
|
|
STATIC void
|
|
|
|
xfs_save_resvblks(struct xfs_mount *mp)
|
|
|
|
{
|
2017-06-16 21:00:05 +03:00
|
|
|
uint64_t resblks = 0;
|
2010-02-06 01:59:53 +03:00
|
|
|
|
|
|
|
mp->m_resblks_save = mp->m_resblks;
|
|
|
|
xfs_reserve_blocks(mp, &resblks, NULL);
|
|
|
|
}
|
|
|
|
|
|
|
|
STATIC void
|
|
|
|
xfs_restore_resvblks(struct xfs_mount *mp)
|
|
|
|
{
|
2017-06-16 21:00:05 +03:00
|
|
|
uint64_t resblks;
|
2010-02-06 01:59:53 +03:00
|
|
|
|
|
|
|
if (mp->m_resblks_save) {
|
|
|
|
resblks = mp->m_resblks_save;
|
|
|
|
mp->m_resblks_save = 0;
|
|
|
|
} else
|
|
|
|
resblks = xfs_default_resblks(mp);
|
|
|
|
|
|
|
|
xfs_reserve_blocks(mp, &resblks, NULL);
|
|
|
|
}
|
|
|
|
|
2012-10-08 14:56:07 +04:00
|
|
|
/*
|
|
|
|
* Trigger writeback of all the dirty metadata in the file system.
|
|
|
|
*
|
|
|
|
* This ensures that the metadata is written to their location on disk rather
|
|
|
|
* than just existing in transactions in the log. This means after a quiesce
|
2012-10-08 14:56:08 +04:00
|
|
|
* there is no log replay required to write the inodes to disk - this is the
|
|
|
|
* primary difference between a sync and a quiesce.
|
2012-10-08 14:56:07 +04:00
|
|
|
*
|
2012-10-08 14:56:08 +04:00
|
|
|
* Note: xfs_log_quiesce() stops background log work - the callers must ensure
|
|
|
|
* it is started again when appropriate.
|
2012-10-08 14:56:07 +04:00
|
|
|
*/
|
xfs: quiesce the filesystem after recovery on readonly mount
Recently we've had a number of reports where log recovery on a v5
filesystem has reported corruptions that looked to be caused by
recovery being re-run over the top of an already-recovered
metadata. This has uncovered a bug in recovery (fixed elsewhere)
but the vector that caused this was largely unknown.
A kdump test started tripping over this problem - the system
would be crashed, the kdump kernel and environment would boot and
dump the kernel core image, and then the system would reboot. After
reboot, the root filesystem was triggering log recovery and
corruptions were being detected. The metadumps indicated the above
log recovery issue.
What is happening is that the kdump kernel and environment is
mounting the root device read-only to find the binaries needed to do
it's work. The result of this is that it is running log recovery.
However, because there were unlinked files and EFIs to be processed
by recovery, the completion of phase 1 of log recovery could not
mark the log clean. And because it's a read-only mount, the unmount
process does not write records to the log to mark it clean, either.
Hence on the next mount of the filesystem, log recovery was run
again across all the metadata that had already been recovered and
this is what triggered corruption warnings.
To avoid this problem, we need to ensure that a read-only mount
always updates the log when it completes the second phase of
recovery. We already handle this sort of issue with rw->ro remount
transitions, so the solution is as simple as quiescing the
filesystem at the appropriate time during the mount process. This
results in the log being marked clean so the mount behaviour
recorded in the logs on repeated RO mounts will change (i.e. log
recovery will no longer be run on every mount until a RW mount is
done). This is a user visible change in behaviour, but it is
harmless.
Signed-off-by: Dave Chinner <dchinner@redhat.com>
Reviewed-by: Eric Sandeen <sandeen@redhat.com>
Signed-off-by: Dave Chinner <david@fromorbit.com>
2016-09-26 01:21:44 +03:00
|
|
|
void
|
2012-10-08 14:56:07 +04:00
|
|
|
xfs_quiesce_attr(
|
|
|
|
struct xfs_mount *mp)
|
|
|
|
{
|
|
|
|
int error = 0;
|
|
|
|
|
|
|
|
/* wait for all modifications to complete */
|
|
|
|
while (atomic_read(&mp->m_active_trans) > 0)
|
|
|
|
delay(100);
|
|
|
|
|
|
|
|
/* force the log to unpin objects from the now complete transactions */
|
|
|
|
xfs_log_force(mp, XFS_LOG_SYNC);
|
|
|
|
|
|
|
|
/* reclaim inodes to do any IO before the freeze completes */
|
|
|
|
xfs_reclaim_inodes(mp, 0);
|
|
|
|
xfs_reclaim_inodes(mp, SYNC_WAIT);
|
|
|
|
|
|
|
|
/* Push the superblock and write an unmount record */
|
|
|
|
error = xfs_log_sbcount(mp);
|
|
|
|
if (error)
|
|
|
|
xfs_warn(mp, "xfs_attr_quiesce: failed to log sb changes. "
|
|
|
|
"Frozen image may not be consistent.");
|
|
|
|
/*
|
2012-10-08 14:56:08 +04:00
|
|
|
* Just warn here till VFS can correctly support
|
|
|
|
* read-only remount without racing.
|
2012-10-08 14:56:07 +04:00
|
|
|
*/
|
2012-10-08 14:56:08 +04:00
|
|
|
WARN_ON(atomic_read(&mp->m_active_trans) != 0);
|
2012-10-08 14:56:07 +04:00
|
|
|
|
2012-10-08 14:56:08 +04:00
|
|
|
xfs_log_quiesce(mp);
|
2012-10-08 14:56:07 +04:00
|
|
|
}
|
|
|
|
|
2016-03-02 01:56:31 +03:00
|
|
|
STATIC int
|
|
|
|
xfs_test_remount_options(
|
|
|
|
struct super_block *sb,
|
|
|
|
char *options)
|
|
|
|
{
|
|
|
|
int error = 0;
|
|
|
|
struct xfs_mount *tmp_mp;
|
|
|
|
|
|
|
|
tmp_mp = kmem_zalloc(sizeof(*tmp_mp), KM_MAYFAIL);
|
|
|
|
if (!tmp_mp)
|
|
|
|
return -ENOMEM;
|
|
|
|
|
|
|
|
tmp_mp->m_super = sb;
|
|
|
|
error = xfs_parseargs(tmp_mp, options);
|
|
|
|
xfs_free_fsname(tmp_mp);
|
2017-09-02 19:54:23 +03:00
|
|
|
kmem_free(tmp_mp);
|
2016-03-02 01:56:31 +03:00
|
|
|
|
|
|
|
return error;
|
|
|
|
}
|
|
|
|
|
2005-04-17 02:20:36 +04:00
|
|
|
STATIC int
|
2006-03-14 06:06:18 +03:00
|
|
|
xfs_fs_remount(
|
2005-04-17 02:20:36 +04:00
|
|
|
struct super_block *sb,
|
|
|
|
int *flags,
|
|
|
|
char *options)
|
|
|
|
{
|
2007-08-30 11:20:39 +04:00
|
|
|
struct xfs_mount *mp = XFS_M(sb);
|
xfs: allow inode allocations in post-growfs disk space
Today, if we perform an xfs_growfs which adds allocation groups,
mp->m_maxagi is not properly updated when the growfs is complete.
Therefore inodes will continue to be allocated only in the
AGs which existed prior to the growfs, and the new space
won't be utilized.
This is because of this path in xfs_growfs_data_private():
xfs_growfs_data_private
xfs_initialize_perag(mp, nagcount, &nagimax);
if (mp->m_flags & XFS_MOUNT_32BITINODES)
index = xfs_set_inode32(mp);
else
index = xfs_set_inode64(mp);
if (maxagi)
*maxagi = index;
where xfs_set_inode* iterates over the (old) agcount in
mp->m_sb.sb_agblocks, which has not yet been updated
in the growfs path. So "index" will be returned based on
the old agcount, not the new one, and new AGs are not available
for inode allocation.
Fix this by explicitly passing the proper AG count (which
xfs_initialize_perag() already has) down another level,
so that xfs_set_inode* can make the proper decision about
acceptable AGs for inode allocation in the potentially
newly-added AGs.
This has been broken since 3.7, when these two
xfs_set_inode* functions were added in commit 2d2194f.
Prior to that, we looped over "agcount" not sb_agblocks
in these calculations.
Signed-off-by: Eric Sandeen <sandeen@redhat.com>
Reviewed-by: Brian Foster <bfoster@redhat.com>
Signed-off-by: Dave Chinner <david@fromorbit.com>
2014-07-24 14:51:54 +04:00
|
|
|
xfs_sb_t *sbp = &mp->m_sb;
|
2008-07-18 11:12:36 +04:00
|
|
|
substring_t args[MAX_OPT_ARGS];
|
|
|
|
char *p;
|
2009-01-19 04:04:07 +03:00
|
|
|
int error;
|
2005-04-17 02:20:36 +04:00
|
|
|
|
2016-03-02 01:56:31 +03:00
|
|
|
/* First, check for complete junk; i.e. invalid options */
|
2018-04-06 20:09:42 +03:00
|
|
|
error = xfs_test_remount_options(sb, options);
|
2016-03-02 01:56:31 +03:00
|
|
|
if (error)
|
|
|
|
return error;
|
|
|
|
|
2014-03-13 18:14:33 +04:00
|
|
|
sync_filesystem(sb);
|
2008-07-18 11:12:36 +04:00
|
|
|
while ((p = strsep(&options, ",")) != NULL) {
|
|
|
|
int token;
|
2008-05-20 09:10:44 +04:00
|
|
|
|
2008-07-18 11:12:36 +04:00
|
|
|
if (!*p)
|
|
|
|
continue;
|
2008-05-20 05:30:39 +04:00
|
|
|
|
2008-07-18 11:12:36 +04:00
|
|
|
token = match_token(p, tokens, args);
|
|
|
|
switch (token) {
|
2012-08-18 01:19:38 +04:00
|
|
|
case Opt_inode64:
|
2016-03-02 01:58:09 +03:00
|
|
|
mp->m_flags &= ~XFS_MOUNT_SMALL_INUMS;
|
|
|
|
mp->m_maxagi = xfs_set_inode_alloc(mp, sbp->sb_agcount);
|
2012-08-18 01:19:38 +04:00
|
|
|
break;
|
2012-09-20 17:32:41 +04:00
|
|
|
case Opt_inode32:
|
2016-03-02 01:58:09 +03:00
|
|
|
mp->m_flags |= XFS_MOUNT_SMALL_INUMS;
|
|
|
|
mp->m_maxagi = xfs_set_inode_alloc(mp, sbp->sb_agcount);
|
2012-09-20 17:32:41 +04:00
|
|
|
break;
|
2008-07-18 11:12:36 +04:00
|
|
|
default:
|
2008-09-17 10:49:33 +04:00
|
|
|
/*
|
|
|
|
* Logically we would return an error here to prevent
|
|
|
|
* users from believing they might have changed
|
|
|
|
* mount options using remount which can't be changed.
|
|
|
|
*
|
|
|
|
* But unfortunately mount(8) adds all options from
|
|
|
|
* mtab and fstab to the mount arguments in some cases
|
|
|
|
* so we can't blindly reject options, but have to
|
|
|
|
* check for each specified option if it actually
|
|
|
|
* differs from the currently set option and only
|
|
|
|
* reject it if that's the case.
|
|
|
|
*
|
|
|
|
* Until that is implemented we return success for
|
|
|
|
* every remount request, and silently ignore all
|
|
|
|
* options that we can't actually change.
|
|
|
|
*/
|
|
|
|
#if 0
|
2011-03-07 02:00:35 +03:00
|
|
|
xfs_info(mp,
|
2013-10-12 05:59:05 +04:00
|
|
|
"mount option \"%s\" not supported for remount", p);
|
2008-07-18 11:12:36 +04:00
|
|
|
return -EINVAL;
|
2008-09-17 10:49:33 +04:00
|
|
|
#else
|
2008-10-12 16:30:44 +04:00
|
|
|
break;
|
2008-09-17 10:49:33 +04:00
|
|
|
#endif
|
2008-05-20 05:30:39 +04:00
|
|
|
}
|
2008-07-18 11:12:36 +04:00
|
|
|
}
|
|
|
|
|
2009-01-19 04:04:07 +03:00
|
|
|
/* ro -> rw */
|
2017-11-28 00:05:09 +03:00
|
|
|
if ((mp->m_flags & XFS_MOUNT_RDONLY) && !(*flags & SB_RDONLY)) {
|
2015-04-13 04:25:41 +03:00
|
|
|
if (mp->m_flags & XFS_MOUNT_NORECOVERY) {
|
|
|
|
xfs_warn(mp,
|
|
|
|
"ro->rw transition prohibited on norecovery mount");
|
|
|
|
return -EINVAL;
|
|
|
|
}
|
|
|
|
|
2016-04-06 00:05:41 +03:00
|
|
|
if (XFS_SB_VERSION_NUM(sbp) == XFS_SB_VERSION_5 &&
|
|
|
|
xfs_sb_has_ro_compat_feature(sbp,
|
|
|
|
XFS_SB_FEAT_RO_COMPAT_UNKNOWN)) {
|
|
|
|
xfs_warn(mp,
|
|
|
|
"ro->rw transition prohibited on unknown (0x%x) ro-compat filesystem",
|
|
|
|
(sbp->sb_features_ro_compat &
|
|
|
|
XFS_SB_FEAT_RO_COMPAT_UNKNOWN));
|
|
|
|
return -EINVAL;
|
|
|
|
}
|
|
|
|
|
2008-07-18 11:12:36 +04:00
|
|
|
mp->m_flags &= ~XFS_MOUNT_RDONLY;
|
2009-01-19 04:04:07 +03:00
|
|
|
|
|
|
|
/*
|
|
|
|
* If this is the first remount to writeable state we
|
|
|
|
* might have some superblock changes to update.
|
|
|
|
*/
|
2015-01-22 01:10:31 +03:00
|
|
|
if (mp->m_update_sb) {
|
|
|
|
error = xfs_sync_sb(mp, false);
|
2009-01-19 04:04:07 +03:00
|
|
|
if (error) {
|
2011-03-07 02:00:35 +03:00
|
|
|
xfs_warn(mp, "failed to write sb changes");
|
2009-01-19 04:04:07 +03:00
|
|
|
return error;
|
|
|
|
}
|
2015-01-22 01:10:31 +03:00
|
|
|
mp->m_update_sb = false;
|
2009-01-19 04:04:07 +03:00
|
|
|
}
|
xfs: don't hold onto reserved blocks on remount,ro
If we hold onto reserved blocks when doing a remount,ro we end
up writing the blocks used count to disk that includes the reserved
blocks. Reserved blocks are not actually used, so this results in
the values in the superblock being incorrect.
Hence if we run xfs_check or xfs_repair -n while the filesystem is
mounted remount,ro we end up with an inconsistent filesystem being
reported. Also, running xfs_copy on the remount,ro filesystem will
result in an inconsistent image being generated.
To fix this, unreserve the blocks when doing the remount,ro, and
reserved them again on remount,rw. This way a remount,ro filesystem
will appear consistent on disk to all utilities.
Signed-off-by: Dave Chinner <david@fromorbit.com>
Reviewed-by: Christoph Hellwig <hch@lst.de>
2010-01-26 07:08:49 +03:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Fill out the reserve pool if it is empty. Use the stashed
|
|
|
|
* value if it is non-zero, otherwise go with the default.
|
|
|
|
*/
|
2010-02-06 01:59:53 +03:00
|
|
|
xfs_restore_resvblks(mp);
|
2012-10-08 14:56:02 +04:00
|
|
|
xfs_log_work_queue(mp);
|
2016-10-03 19:11:39 +03:00
|
|
|
|
|
|
|
/* Recover any CoW blocks that never got remapped. */
|
|
|
|
error = xfs_reflink_recover_cow(mp);
|
|
|
|
if (error) {
|
|
|
|
xfs_err(mp,
|
|
|
|
"Error %d recovering leftover CoW allocations.", error);
|
|
|
|
xfs_force_shutdown(mp, SHUTDOWN_CORRUPT_INCORE);
|
|
|
|
return error;
|
|
|
|
}
|
2019-04-26 04:26:22 +03:00
|
|
|
xfs_start_block_reaping(mp);
|
2016-10-03 19:11:44 +03:00
|
|
|
|
|
|
|
/* Create the per-AG metadata reservation pool .*/
|
|
|
|
error = xfs_fs_reserve_ag_blocks(mp);
|
|
|
|
if (error && error != -ENOSPC)
|
|
|
|
return error;
|
2008-07-18 11:12:36 +04:00
|
|
|
}
|
|
|
|
|
|
|
|
/* rw -> ro */
|
2017-11-28 00:05:09 +03:00
|
|
|
if (!(mp->m_flags & XFS_MOUNT_RDONLY) && (*flags & SB_RDONLY)) {
|
2018-05-09 20:03:56 +03:00
|
|
|
/*
|
|
|
|
* Cancel background eofb scanning so it cannot race with the
|
|
|
|
* final log force+buftarg wait and deadlock the remount.
|
|
|
|
*/
|
2019-04-26 04:26:22 +03:00
|
|
|
xfs_stop_block_reaping(mp);
|
2018-05-09 20:03:56 +03:00
|
|
|
|
2017-12-15 02:46:05 +03:00
|
|
|
/* Get rid of any leftover CoW reservations... */
|
|
|
|
error = xfs_icache_free_cowblocks(mp, NULL);
|
|
|
|
if (error) {
|
|
|
|
xfs_force_shutdown(mp, SHUTDOWN_CORRUPT_INCORE);
|
|
|
|
return error;
|
|
|
|
}
|
|
|
|
|
2016-10-03 19:11:44 +03:00
|
|
|
/* Free the per-AG metadata reservation pool. */
|
|
|
|
error = xfs_fs_unreserve_ag_blocks(mp);
|
|
|
|
if (error) {
|
|
|
|
xfs_force_shutdown(mp, SHUTDOWN_CORRUPT_INCORE);
|
|
|
|
return error;
|
|
|
|
}
|
|
|
|
|
xfs: don't hold onto reserved blocks on remount,ro
If we hold onto reserved blocks when doing a remount,ro we end
up writing the blocks used count to disk that includes the reserved
blocks. Reserved blocks are not actually used, so this results in
the values in the superblock being incorrect.
Hence if we run xfs_check or xfs_repair -n while the filesystem is
mounted remount,ro we end up with an inconsistent filesystem being
reported. Also, running xfs_copy on the remount,ro filesystem will
result in an inconsistent image being generated.
To fix this, unreserve the blocks when doing the remount,ro, and
reserved them again on remount,rw. This way a remount,ro filesystem
will appear consistent on disk to all utilities.
Signed-off-by: Dave Chinner <david@fromorbit.com>
Reviewed-by: Christoph Hellwig <hch@lst.de>
2010-01-26 07:08:49 +03:00
|
|
|
/*
|
xfs: xfs_sync_fsdata is redundant
Why do we need to write the superblock to disk once we've written
all the data? We don't actually - the reasons for doing this are
lost in the mists of time, and go back to the way Irix used to drive
VFS flushing.
On linux, this code is only called from two contexts: remount and
.sync_fs. In the remount case, the call is followed by a metadata
sync, which unpins and writes the superblock. In the sync_fs case,
we only need to force the log to disk to ensure that the superblock
is correctly on disk, so we don't actually need to write it. Hence
the functionality is either redundant or superfluous and thus can be
removed.
Seeing as xfs_quiesce_data is essentially now just a log force,
remove it as well and fold the code back into the two callers.
Neither of them need the log covering check, either, as that is
redundant for the remount case, and unnecessary for the .sync_fs
case.
Signed-off-by: Dave Chinner <dchinner@redhat.com>
Reviewed-by: Christoph Hellwig <hch@lst.de>
Reviewed-by: Mark Tinguely <tinguely@sgi.com>
Signed-off-by: Ben Myers <bpm@sgi.com>
2012-10-08 14:56:06 +04:00
|
|
|
* Before we sync the metadata, we need to free up the reserve
|
|
|
|
* block pool so that the used block count in the superblock on
|
|
|
|
* disk is correct at the end of the remount. Stash the current
|
|
|
|
* reserve pool size so that if we get remounted rw, we can
|
|
|
|
* return it to the same size.
|
xfs: don't hold onto reserved blocks on remount,ro
If we hold onto reserved blocks when doing a remount,ro we end
up writing the blocks used count to disk that includes the reserved
blocks. Reserved blocks are not actually used, so this results in
the values in the superblock being incorrect.
Hence if we run xfs_check or xfs_repair -n while the filesystem is
mounted remount,ro we end up with an inconsistent filesystem being
reported. Also, running xfs_copy on the remount,ro filesystem will
result in an inconsistent image being generated.
To fix this, unreserve the blocks when doing the remount,ro, and
reserved them again on remount,rw. This way a remount,ro filesystem
will appear consistent on disk to all utilities.
Signed-off-by: Dave Chinner <david@fromorbit.com>
Reviewed-by: Christoph Hellwig <hch@lst.de>
2010-01-26 07:08:49 +03:00
|
|
|
*/
|
2010-02-06 01:59:53 +03:00
|
|
|
xfs_save_resvblks(mp);
|
2016-06-21 04:53:28 +03:00
|
|
|
|
2008-10-30 09:16:21 +03:00
|
|
|
xfs_quiesce_attr(mp);
|
2008-05-20 05:30:39 +04:00
|
|
|
mp->m_flags |= XFS_MOUNT_RDONLY;
|
|
|
|
}
|
|
|
|
|
2008-07-18 11:12:36 +04:00
|
|
|
return 0;
|
2005-04-17 02:20:36 +04:00
|
|
|
}
|
|
|
|
|
2007-10-11 12:11:14 +04:00
|
|
|
/*
|
|
|
|
* Second stage of a freeze. The data is already frozen so we only
|
2015-01-22 01:10:31 +03:00
|
|
|
* need to take care of the metadata. Once that's done sync the superblock
|
|
|
|
* to the log to dirty it in case of a crash while frozen. This ensures that we
|
|
|
|
* will recover the unlinked inode lists on the next mount.
|
2007-10-11 12:11:14 +04:00
|
|
|
*/
|
2009-01-10 03:40:58 +03:00
|
|
|
STATIC int
|
|
|
|
xfs_fs_freeze(
|
2005-04-17 02:20:36 +04:00
|
|
|
struct super_block *sb)
|
|
|
|
{
|
2007-10-11 12:11:14 +04:00
|
|
|
struct xfs_mount *mp = XFS_M(sb);
|
|
|
|
|
2019-04-26 04:26:22 +03:00
|
|
|
xfs_stop_block_reaping(mp);
|
2010-02-06 01:59:53 +03:00
|
|
|
xfs_save_resvblks(mp);
|
2008-10-30 09:16:21 +03:00
|
|
|
xfs_quiesce_attr(mp);
|
2015-01-22 01:10:31 +03:00
|
|
|
return xfs_sync_sb(mp, true);
|
2005-04-17 02:20:36 +04:00
|
|
|
}
|
|
|
|
|
2010-02-06 01:59:53 +03:00
|
|
|
STATIC int
|
|
|
|
xfs_fs_unfreeze(
|
|
|
|
struct super_block *sb)
|
|
|
|
{
|
|
|
|
struct xfs_mount *mp = XFS_M(sb);
|
|
|
|
|
|
|
|
xfs_restore_resvblks(mp);
|
2012-10-08 14:56:02 +04:00
|
|
|
xfs_log_work_queue(mp);
|
2019-04-26 04:26:22 +03:00
|
|
|
xfs_start_block_reaping(mp);
|
2010-02-06 01:59:53 +03:00
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2005-04-17 02:20:36 +04:00
|
|
|
STATIC int
|
2006-03-14 06:06:18 +03:00
|
|
|
xfs_fs_show_options(
|
2005-04-17 02:20:36 +04:00
|
|
|
struct seq_file *m,
|
2011-12-09 06:32:45 +04:00
|
|
|
struct dentry *root)
|
2005-04-17 02:20:36 +04:00
|
|
|
{
|
2019-05-02 06:26:30 +03:00
|
|
|
xfs_showargs(XFS_M(root->d_sb), m);
|
|
|
|
return 0;
|
2005-04-17 02:20:36 +04:00
|
|
|
}
|
|
|
|
|
2008-05-20 05:30:59 +04:00
|
|
|
/*
|
|
|
|
* This function fills in xfs_mount_t fields based on mount args.
|
|
|
|
* Note: the superblock _has_ now been read in.
|
|
|
|
*/
|
|
|
|
STATIC int
|
|
|
|
xfs_finish_flags(
|
|
|
|
struct xfs_mount *mp)
|
|
|
|
{
|
|
|
|
int ronly = (mp->m_flags & XFS_MOUNT_RDONLY);
|
|
|
|
|
2008-10-16 21:02:37 +04:00
|
|
|
/* Fail a mount where the logbuf is smaller than the log stripe */
|
2008-05-20 05:30:59 +04:00
|
|
|
if (xfs_sb_version_haslogv2(&mp->m_sb)) {
|
2008-10-30 09:53:24 +03:00
|
|
|
if (mp->m_logbsize <= 0 &&
|
|
|
|
mp->m_sb.sb_logsunit > XLOG_BIG_RECORD_BSIZE) {
|
2008-05-20 05:30:59 +04:00
|
|
|
mp->m_logbsize = mp->m_sb.sb_logsunit;
|
2008-10-30 09:53:24 +03:00
|
|
|
} else if (mp->m_logbsize > 0 &&
|
|
|
|
mp->m_logbsize < mp->m_sb.sb_logsunit) {
|
2011-03-07 02:00:35 +03:00
|
|
|
xfs_warn(mp,
|
|
|
|
"logbuf size must be greater than or equal to log stripe size");
|
2014-06-25 08:58:08 +04:00
|
|
|
return -EINVAL;
|
2008-05-20 05:30:59 +04:00
|
|
|
}
|
|
|
|
} else {
|
|
|
|
/* Fail a mount if the logbuf is larger than 32K */
|
2008-10-30 09:53:24 +03:00
|
|
|
if (mp->m_logbsize > XLOG_BIG_RECORD_BSIZE) {
|
2011-03-07 02:00:35 +03:00
|
|
|
xfs_warn(mp,
|
|
|
|
"logbuf size for version 1 logs must be 16K or 32K");
|
2014-06-25 08:58:08 +04:00
|
|
|
return -EINVAL;
|
2008-05-20 05:30:59 +04:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2013-06-05 06:09:09 +04:00
|
|
|
/*
|
|
|
|
* V5 filesystems always use attr2 format for attributes.
|
|
|
|
*/
|
|
|
|
if (xfs_sb_version_hascrc(&mp->m_sb) &&
|
|
|
|
(mp->m_flags & XFS_MOUNT_NOATTR2)) {
|
2016-03-02 01:55:38 +03:00
|
|
|
xfs_warn(mp, "Cannot mount a V5 filesystem as noattr2. "
|
|
|
|
"attr2 is always enabled for V5 filesystems.");
|
2014-06-25 08:58:08 +04:00
|
|
|
return -EINVAL;
|
2013-06-05 06:09:09 +04:00
|
|
|
}
|
|
|
|
|
2008-05-20 05:30:59 +04:00
|
|
|
/*
|
|
|
|
* mkfs'ed attr2 will turn on attr2 mount unless explicitly
|
|
|
|
* told by noattr2 to turn it off
|
|
|
|
*/
|
|
|
|
if (xfs_sb_version_hasattr2(&mp->m_sb) &&
|
2008-10-30 09:53:24 +03:00
|
|
|
!(mp->m_flags & XFS_MOUNT_NOATTR2))
|
2008-05-20 05:30:59 +04:00
|
|
|
mp->m_flags |= XFS_MOUNT_ATTR2;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* prohibit r/w mounts of read-only filesystems
|
|
|
|
*/
|
|
|
|
if ((mp->m_sb.sb_flags & XFS_SBF_READONLY) && !ronly) {
|
2011-03-07 02:00:35 +03:00
|
|
|
xfs_warn(mp,
|
|
|
|
"cannot mount a read-only filesystem as read-write");
|
2014-06-25 08:58:08 +04:00
|
|
|
return -EROFS;
|
2008-05-20 05:30:59 +04:00
|
|
|
}
|
|
|
|
|
2013-07-20 02:36:02 +04:00
|
|
|
if ((mp->m_qflags & (XFS_GQUOTA_ACCT | XFS_GQUOTA_ACTIVE)) &&
|
|
|
|
(mp->m_qflags & (XFS_PQUOTA_ACCT | XFS_PQUOTA_ACTIVE)) &&
|
|
|
|
!xfs_sb_version_has_pquotino(&mp->m_sb)) {
|
|
|
|
xfs_warn(mp,
|
|
|
|
"Super block does not support project and group quota together");
|
2014-06-25 08:58:08 +04:00
|
|
|
return -EINVAL;
|
2013-07-20 02:36:02 +04:00
|
|
|
}
|
|
|
|
|
2008-05-20 05:30:59 +04:00
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2015-02-23 13:22:31 +03:00
|
|
|
static int
|
|
|
|
xfs_init_percpu_counters(
|
|
|
|
struct xfs_mount *mp)
|
|
|
|
{
|
|
|
|
int error;
|
|
|
|
|
|
|
|
error = percpu_counter_init(&mp->m_icount, 0, GFP_KERNEL);
|
|
|
|
if (error)
|
2015-03-25 07:00:24 +03:00
|
|
|
return -ENOMEM;
|
2015-02-23 13:22:31 +03:00
|
|
|
|
|
|
|
error = percpu_counter_init(&mp->m_ifree, 0, GFP_KERNEL);
|
|
|
|
if (error)
|
|
|
|
goto free_icount;
|
|
|
|
|
|
|
|
error = percpu_counter_init(&mp->m_fdblocks, 0, GFP_KERNEL);
|
|
|
|
if (error)
|
|
|
|
goto free_ifree;
|
|
|
|
|
2019-04-26 04:26:22 +03:00
|
|
|
error = percpu_counter_init(&mp->m_delalloc_blks, 0, GFP_KERNEL);
|
|
|
|
if (error)
|
|
|
|
goto free_fdblocks;
|
|
|
|
|
2015-02-23 13:22:31 +03:00
|
|
|
return 0;
|
|
|
|
|
2019-04-26 04:26:22 +03:00
|
|
|
free_fdblocks:
|
|
|
|
percpu_counter_destroy(&mp->m_fdblocks);
|
2015-02-23 13:22:31 +03:00
|
|
|
free_ifree:
|
|
|
|
percpu_counter_destroy(&mp->m_ifree);
|
|
|
|
free_icount:
|
|
|
|
percpu_counter_destroy(&mp->m_icount);
|
|
|
|
return -ENOMEM;
|
|
|
|
}
|
|
|
|
|
|
|
|
void
|
|
|
|
xfs_reinit_percpu_counters(
|
|
|
|
struct xfs_mount *mp)
|
|
|
|
{
|
|
|
|
percpu_counter_set(&mp->m_icount, mp->m_sb.sb_icount);
|
|
|
|
percpu_counter_set(&mp->m_ifree, mp->m_sb.sb_ifree);
|
|
|
|
percpu_counter_set(&mp->m_fdblocks, mp->m_sb.sb_fdblocks);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void
|
|
|
|
xfs_destroy_percpu_counters(
|
|
|
|
struct xfs_mount *mp)
|
|
|
|
{
|
|
|
|
percpu_counter_destroy(&mp->m_icount);
|
|
|
|
percpu_counter_destroy(&mp->m_ifree);
|
|
|
|
percpu_counter_destroy(&mp->m_fdblocks);
|
2019-04-26 04:26:22 +03:00
|
|
|
ASSERT(XFS_FORCED_SHUTDOWN(mp) ||
|
|
|
|
percpu_counter_sum(&mp->m_delalloc_blks) == 0);
|
|
|
|
percpu_counter_destroy(&mp->m_delalloc_blks);
|
2015-02-23 13:22:31 +03:00
|
|
|
}
|
|
|
|
|
2018-03-24 03:54:32 +03:00
|
|
|
static struct xfs_mount *
|
|
|
|
xfs_mount_alloc(
|
|
|
|
struct super_block *sb)
|
2005-04-17 02:20:36 +04:00
|
|
|
{
|
2018-03-24 03:54:32 +03:00
|
|
|
struct xfs_mount *mp;
|
2008-05-20 09:10:44 +04:00
|
|
|
|
2008-05-20 09:10:52 +04:00
|
|
|
mp = kzalloc(sizeof(struct xfs_mount), GFP_KERNEL);
|
|
|
|
if (!mp)
|
2018-03-24 03:54:32 +03:00
|
|
|
return NULL;
|
2005-04-17 02:20:36 +04:00
|
|
|
|
2018-03-24 03:54:32 +03:00
|
|
|
mp->m_super = sb;
|
2008-05-20 09:10:52 +04:00
|
|
|
spin_lock_init(&mp->m_sb_lock);
|
2018-03-24 03:54:32 +03:00
|
|
|
spin_lock_init(&mp->m_agirotor_lock);
|
|
|
|
INIT_RADIX_TREE(&mp->m_perag_tree, GFP_ATOMIC);
|
|
|
|
spin_lock_init(&mp->m_perag_lock);
|
2008-05-20 09:10:52 +04:00
|
|
|
mutex_init(&mp->m_growlock);
|
|
|
|
atomic_set(&mp->m_active_trans, 0);
|
2012-10-08 14:56:00 +04:00
|
|
|
INIT_DELAYED_WORK(&mp->m_reclaim_work, xfs_reclaim_worker);
|
2012-11-06 18:50:47 +04:00
|
|
|
INIT_DELAYED_WORK(&mp->m_eofblocks_work, xfs_eofblocks_worker);
|
2016-10-03 19:11:46 +03:00
|
|
|
INIT_DELAYED_WORK(&mp->m_cowblocks_work, xfs_cowblocks_worker);
|
2014-09-29 04:46:08 +04:00
|
|
|
mp->m_kobj.kobject.kset = xfs_kset;
|
2019-02-13 22:46:16 +03:00
|
|
|
/*
|
|
|
|
* We don't create the finobt per-ag space reservation until after log
|
|
|
|
* recovery, so we must set this to true so that an ifree transaction
|
|
|
|
* started during log recovery will not depend on space reservations
|
|
|
|
* for finobt expansion.
|
|
|
|
*/
|
|
|
|
mp->m_finobt_nores = true;
|
2018-03-24 03:54:32 +03:00
|
|
|
return mp;
|
|
|
|
}
|
2007-08-30 11:21:22 +04:00
|
|
|
|
2018-03-24 03:54:32 +03:00
|
|
|
|
|
|
|
STATIC int
|
|
|
|
xfs_fs_fill_super(
|
|
|
|
struct super_block *sb,
|
|
|
|
void *data,
|
|
|
|
int silent)
|
|
|
|
{
|
|
|
|
struct inode *root;
|
|
|
|
struct xfs_mount *mp = NULL;
|
|
|
|
int flags = 0, error = -ENOMEM;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* allocate mp and do all low-level struct initializations before we
|
|
|
|
* attach it to the super
|
|
|
|
*/
|
|
|
|
mp = xfs_mount_alloc(sb);
|
|
|
|
if (!mp)
|
|
|
|
goto out;
|
2007-08-30 11:21:30 +04:00
|
|
|
sb->s_fs_info = mp;
|
2005-04-17 02:20:36 +04:00
|
|
|
|
2010-06-23 12:11:15 +04:00
|
|
|
error = xfs_parseargs(mp, (char *)data);
|
2007-08-30 11:20:39 +04:00
|
|
|
if (error)
|
2008-10-30 09:53:24 +03:00
|
|
|
goto out_free_fsname;
|
2005-04-17 02:20:36 +04:00
|
|
|
|
|
|
|
sb_min_blocksize(sb, BBSIZE);
|
2008-06-23 07:23:01 +04:00
|
|
|
sb->s_xattr = xfs_xattr_handlers;
|
2006-03-14 06:06:18 +03:00
|
|
|
sb->s_export_op = &xfs_export_operations;
|
2009-02-09 10:47:34 +03:00
|
|
|
#ifdef CONFIG_XFS_QUOTA
|
2006-03-14 06:06:18 +03:00
|
|
|
sb->s_qcop = &xfs_quotactl_operations;
|
2014-10-01 00:35:33 +04:00
|
|
|
sb->s_quota_types = QTYPE_MASK_USR | QTYPE_MASK_GRP | QTYPE_MASK_PRJ;
|
2009-02-09 10:47:34 +03:00
|
|
|
#endif
|
2006-03-14 06:06:18 +03:00
|
|
|
sb->s_op = &xfs_super_operations;
|
2005-04-17 02:20:36 +04:00
|
|
|
|
2018-05-11 07:50:23 +03:00
|
|
|
/*
|
|
|
|
* Delay mount work if the debug hook is set. This is debug
|
|
|
|
* instrumention to coordinate simulation of xfs mount failures with
|
|
|
|
* VFS superblock operations
|
|
|
|
*/
|
|
|
|
if (xfs_globals.mount_delay) {
|
|
|
|
xfs_notice(mp, "Delaying mount for %d seconds.",
|
|
|
|
xfs_globals.mount_delay);
|
|
|
|
msleep(xfs_globals.mount_delay * 1000);
|
|
|
|
}
|
|
|
|
|
2008-10-30 09:53:24 +03:00
|
|
|
if (silent)
|
2008-05-20 05:30:59 +04:00
|
|
|
flags |= XFS_MFSI_QUIET;
|
|
|
|
|
2008-10-30 09:53:24 +03:00
|
|
|
error = xfs_open_devices(mp);
|
2008-05-20 05:31:13 +04:00
|
|
|
if (error)
|
2010-06-23 12:11:15 +04:00
|
|
|
goto out_free_fsname;
|
2008-05-20 05:30:59 +04:00
|
|
|
|
2014-06-25 08:58:08 +04:00
|
|
|
error = xfs_init_mount_workqueues(mp);
|
2010-09-30 06:25:54 +04:00
|
|
|
if (error)
|
|
|
|
goto out_close_devices;
|
2008-05-20 09:10:52 +04:00
|
|
|
|
2015-02-23 13:22:31 +03:00
|
|
|
error = xfs_init_percpu_counters(mp);
|
2012-02-29 13:53:48 +04:00
|
|
|
if (error)
|
|
|
|
goto out_destroy_workqueues;
|
|
|
|
|
2015-10-12 10:21:19 +03:00
|
|
|
/* Allocate stats memory before we do operations that might use it */
|
|
|
|
mp->m_stats.xs_stats = alloc_percpu(struct xfsstats);
|
|
|
|
if (!mp->m_stats.xs_stats) {
|
2015-10-19 00:42:47 +03:00
|
|
|
error = -ENOMEM;
|
2015-10-12 10:21:19 +03:00
|
|
|
goto out_destroy_counters;
|
|
|
|
}
|
|
|
|
|
2008-05-20 05:30:59 +04:00
|
|
|
error = xfs_readsb(mp, flags);
|
|
|
|
if (error)
|
2015-10-12 10:21:19 +03:00
|
|
|
goto out_free_stats;
|
2008-10-30 09:53:24 +03:00
|
|
|
|
|
|
|
error = xfs_finish_flags(mp);
|
2008-05-20 05:30:59 +04:00
|
|
|
if (error)
|
2008-05-20 09:11:05 +04:00
|
|
|
goto out_free_sb;
|
2008-05-20 05:30:59 +04:00
|
|
|
|
2008-05-20 09:10:36 +04:00
|
|
|
error = xfs_setup_devices(mp);
|
2008-05-20 05:31:13 +04:00
|
|
|
if (error)
|
2008-05-20 09:11:05 +04:00
|
|
|
goto out_free_sb;
|
2008-05-20 05:30:59 +04:00
|
|
|
|
|
|
|
error = xfs_filestream_mount(mp);
|
|
|
|
if (error)
|
2008-05-20 09:11:05 +04:00
|
|
|
goto out_free_sb;
|
2008-05-20 05:30:59 +04:00
|
|
|
|
2011-03-26 01:14:57 +03:00
|
|
|
/*
|
|
|
|
* we must configure the block size in the superblock before we run the
|
|
|
|
* full mount process as the mount process can lookup and cache inodes.
|
|
|
|
*/
|
2018-10-18 09:20:19 +03:00
|
|
|
sb->s_magic = XFS_SUPER_MAGIC;
|
2007-10-11 12:09:40 +04:00
|
|
|
sb->s_blocksize = mp->m_sb.sb_blocksize;
|
|
|
|
sb->s_blocksize_bits = ffs(sb->s_blocksize) - 1;
|
2005-04-17 02:20:36 +04:00
|
|
|
sb->s_maxbytes = xfs_max_file_offset(sb->s_blocksize_bits);
|
2012-02-06 21:45:27 +04:00
|
|
|
sb->s_max_links = XFS_MAXLINK;
|
2005-04-17 02:20:36 +04:00
|
|
|
sb->s_time_gran = 1;
|
2019-07-30 18:22:29 +03:00
|
|
|
sb->s_time_min = S32_MIN;
|
|
|
|
sb->s_time_max = S32_MAX;
|
2019-06-29 05:30:22 +03:00
|
|
|
sb->s_iflags |= SB_I_CGROUPWB;
|
|
|
|
|
2005-04-17 02:20:36 +04:00
|
|
|
set_posix_acl_flag(sb);
|
|
|
|
|
2013-06-27 10:04:59 +04:00
|
|
|
/* version 5 superblocks support inode version counters. */
|
|
|
|
if (XFS_SB_VERSION_NUM(&mp->m_sb) == XFS_SB_VERSION_5)
|
2017-10-18 23:56:26 +03:00
|
|
|
sb->s_flags |= SB_I_VERSION;
|
2013-06-27 10:04:59 +04:00
|
|
|
|
2015-06-04 02:19:18 +03:00
|
|
|
if (mp->m_flags & XFS_MOUNT_DAX) {
|
2018-05-30 23:03:46 +03:00
|
|
|
bool rtdev_is_dax = false, datadev_is_dax;
|
2018-05-30 23:03:45 +03:00
|
|
|
|
2015-06-04 02:19:18 +03:00
|
|
|
xfs_warn(mp,
|
2016-05-10 19:23:56 +03:00
|
|
|
"DAX enabled. Warning: EXPERIMENTAL, use at your own risk");
|
|
|
|
|
2018-05-30 23:03:46 +03:00
|
|
|
datadev_is_dax = bdev_dax_supported(mp->m_ddev_targp->bt_bdev,
|
|
|
|
sb->s_blocksize);
|
2018-05-30 23:03:45 +03:00
|
|
|
if (mp->m_rtdev_targp)
|
2018-05-30 23:03:46 +03:00
|
|
|
rtdev_is_dax = bdev_dax_supported(
|
|
|
|
mp->m_rtdev_targp->bt_bdev, sb->s_blocksize);
|
|
|
|
if (!rtdev_is_dax && !datadev_is_dax) {
|
2015-06-04 02:19:18 +03:00
|
|
|
xfs_alert(mp,
|
2016-05-10 19:23:56 +03:00
|
|
|
"DAX unsupported by block device. Turning off DAX.");
|
2015-06-04 02:19:18 +03:00
|
|
|
mp->m_flags &= ~XFS_MOUNT_DAX;
|
|
|
|
}
|
2018-02-01 01:21:56 +03:00
|
|
|
if (xfs_sb_version_hasreflink(&mp->m_sb)) {
|
2016-10-03 19:11:52 +03:00
|
|
|
xfs_alert(mp,
|
2018-01-09 00:30:08 +03:00
|
|
|
"DAX and reflink cannot be used together!");
|
2018-02-01 01:21:56 +03:00
|
|
|
error = -EINVAL;
|
|
|
|
goto out_filestream_unmount;
|
|
|
|
}
|
2015-06-04 02:19:18 +03:00
|
|
|
}
|
|
|
|
|
2017-09-18 22:03:56 +03:00
|
|
|
if (mp->m_flags & XFS_MOUNT_DISCARD) {
|
|
|
|
struct request_queue *q = bdev_get_queue(sb->s_bdev);
|
|
|
|
|
|
|
|
if (!blk_queue_discard(q)) {
|
|
|
|
xfs_warn(mp, "mounting with \"discard\" option, but "
|
|
|
|
"the device does not support discard");
|
|
|
|
mp->m_flags &= ~XFS_MOUNT_DISCARD;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2019-02-18 20:38:49 +03:00
|
|
|
if (xfs_sb_version_hasreflink(&mp->m_sb)) {
|
|
|
|
if (mp->m_sb.sb_rblocks) {
|
|
|
|
xfs_alert(mp,
|
2018-02-01 03:38:18 +03:00
|
|
|
"reflink not compatible with realtime device!");
|
2019-02-18 20:38:49 +03:00
|
|
|
error = -EINVAL;
|
|
|
|
goto out_filestream_unmount;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (xfs_globals.always_cow) {
|
|
|
|
xfs_info(mp, "using DEBUG-only always_cow mode.");
|
|
|
|
mp->m_always_cow = true;
|
|
|
|
}
|
2018-02-01 03:38:18 +03:00
|
|
|
}
|
|
|
|
|
2018-01-31 20:47:25 +03:00
|
|
|
if (xfs_sb_version_hasrmapbt(&mp->m_sb) && mp->m_sb.sb_rblocks) {
|
2016-08-03 05:20:57 +03:00
|
|
|
xfs_alert(mp,
|
2018-01-31 20:47:25 +03:00
|
|
|
"reverse mapping btree not compatible with realtime device!");
|
|
|
|
error = -EINVAL;
|
|
|
|
goto out_filestream_unmount;
|
2016-08-26 08:59:19 +03:00
|
|
|
}
|
2016-08-03 05:20:57 +03:00
|
|
|
|
2012-04-13 16:10:44 +04:00
|
|
|
error = xfs_mountfs(mp);
|
2011-07-13 15:43:48 +04:00
|
|
|
if (error)
|
2012-10-08 14:56:00 +04:00
|
|
|
goto out_filestream_unmount;
|
2011-03-26 01:14:57 +03:00
|
|
|
|
2008-08-13 09:45:15 +04:00
|
|
|
root = igrab(VFS_I(mp->m_rootip));
|
2008-03-27 10:00:54 +03:00
|
|
|
if (!root) {
|
2014-06-25 08:58:08 +04:00
|
|
|
error = -ENOENT;
|
2012-04-13 16:10:44 +04:00
|
|
|
goto out_unmount;
|
2008-02-05 04:14:01 +03:00
|
|
|
}
|
2012-01-09 07:15:13 +04:00
|
|
|
sb->s_root = d_make_root(root);
|
2008-03-27 10:00:54 +03:00
|
|
|
if (!sb->s_root) {
|
2014-06-25 08:58:08 +04:00
|
|
|
error = -ENOMEM;
|
2012-04-13 16:10:44 +04:00
|
|
|
goto out_unmount;
|
2005-04-17 02:20:36 +04:00
|
|
|
}
|
2007-08-30 11:21:22 +04:00
|
|
|
|
2005-04-17 02:20:36 +04:00
|
|
|
return 0;
|
2012-10-08 14:55:59 +04:00
|
|
|
|
2012-10-08 14:56:00 +04:00
|
|
|
out_filestream_unmount:
|
2008-05-20 09:11:11 +04:00
|
|
|
xfs_filestream_unmount(mp);
|
2008-05-20 09:11:05 +04:00
|
|
|
out_free_sb:
|
|
|
|
xfs_freesb(mp);
|
2015-10-12 10:21:19 +03:00
|
|
|
out_free_stats:
|
|
|
|
free_percpu(mp->m_stats.xs_stats);
|
2008-10-30 09:53:24 +03:00
|
|
|
out_destroy_counters:
|
2015-02-23 13:22:31 +03:00
|
|
|
xfs_destroy_percpu_counters(mp);
|
2015-10-12 10:21:19 +03:00
|
|
|
out_destroy_workqueues:
|
2012-02-29 13:53:48 +04:00
|
|
|
xfs_destroy_mount_workqueues(mp);
|
2010-09-30 06:25:54 +04:00
|
|
|
out_close_devices:
|
2008-05-20 05:31:13 +04:00
|
|
|
xfs_close_devices(mp);
|
2008-10-30 09:53:24 +03:00
|
|
|
out_free_fsname:
|
2018-05-11 07:50:23 +03:00
|
|
|
sb->s_fs_info = NULL;
|
2008-10-30 09:53:24 +03:00
|
|
|
xfs_free_fsname(mp);
|
2008-05-20 09:10:52 +04:00
|
|
|
kfree(mp);
|
2008-10-30 09:53:24 +03:00
|
|
|
out:
|
2014-06-25 08:58:08 +04:00
|
|
|
return error;
|
2008-05-20 05:30:59 +04:00
|
|
|
|
2011-07-13 15:43:48 +04:00
|
|
|
out_unmount:
|
2008-05-20 05:30:52 +04:00
|
|
|
xfs_filestream_unmount(mp);
|
2008-05-20 05:31:13 +04:00
|
|
|
xfs_unmountfs(mp);
|
2008-08-13 10:50:21 +04:00
|
|
|
goto out_free_sb;
|
2005-04-17 02:20:36 +04:00
|
|
|
}
|
|
|
|
|
2015-02-23 13:22:31 +03:00
|
|
|
STATIC void
|
|
|
|
xfs_fs_put_super(
|
|
|
|
struct super_block *sb)
|
|
|
|
{
|
|
|
|
struct xfs_mount *mp = XFS_M(sb);
|
|
|
|
|
2018-05-11 07:50:23 +03:00
|
|
|
/* if ->fill_super failed, we have no mount to tear down */
|
|
|
|
if (!sb->s_fs_info)
|
|
|
|
return;
|
|
|
|
|
2015-02-24 02:27:28 +03:00
|
|
|
xfs_notice(mp, "Unmounting Filesystem");
|
2015-02-23 13:22:31 +03:00
|
|
|
xfs_filestream_unmount(mp);
|
|
|
|
xfs_unmountfs(mp);
|
|
|
|
|
|
|
|
xfs_freesb(mp);
|
2015-10-12 10:21:19 +03:00
|
|
|
free_percpu(mp->m_stats.xs_stats);
|
2015-02-23 13:22:31 +03:00
|
|
|
xfs_destroy_percpu_counters(mp);
|
|
|
|
xfs_destroy_mount_workqueues(mp);
|
|
|
|
xfs_close_devices(mp);
|
2018-05-11 07:50:23 +03:00
|
|
|
|
|
|
|
sb->s_fs_info = NULL;
|
2015-02-23 13:22:31 +03:00
|
|
|
xfs_free_fsname(mp);
|
|
|
|
kfree(mp);
|
|
|
|
}
|
|
|
|
|
2010-07-25 00:46:55 +04:00
|
|
|
STATIC struct dentry *
|
|
|
|
xfs_fs_mount(
|
2005-04-17 02:20:36 +04:00
|
|
|
struct file_system_type *fs_type,
|
|
|
|
int flags,
|
|
|
|
const char *dev_name,
|
2010-07-25 00:46:55 +04:00
|
|
|
void *data)
|
2005-04-17 02:20:36 +04:00
|
|
|
{
|
2010-07-25 00:46:55 +04:00
|
|
|
return mount_bdev(fs_type, flags, dev_name, data, xfs_fs_fill_super);
|
2006-03-14 06:06:18 +03:00
|
|
|
}
|
|
|
|
|
2013-08-28 04:17:57 +04:00
|
|
|
static long
|
2011-07-08 08:14:46 +04:00
|
|
|
xfs_fs_nr_cached_objects(
|
2013-08-28 04:18:05 +04:00
|
|
|
struct super_block *sb,
|
2015-02-13 01:58:51 +03:00
|
|
|
struct shrink_control *sc)
|
2011-07-08 08:14:46 +04:00
|
|
|
{
|
2018-05-11 07:50:23 +03:00
|
|
|
/* Paranoia: catch incorrect calls during mount setup or teardown */
|
|
|
|
if (WARN_ON_ONCE(!sb->s_fs_info))
|
|
|
|
return 0;
|
2011-07-08 08:14:46 +04:00
|
|
|
return xfs_reclaim_inodes_count(XFS_M(sb));
|
|
|
|
}
|
|
|
|
|
2013-08-28 04:17:57 +04:00
|
|
|
static long
|
2011-07-08 08:14:46 +04:00
|
|
|
xfs_fs_free_cached_objects(
|
|
|
|
struct super_block *sb,
|
2015-02-13 01:58:51 +03:00
|
|
|
struct shrink_control *sc)
|
2011-07-08 08:14:46 +04:00
|
|
|
{
|
2015-02-13 01:58:51 +03:00
|
|
|
return xfs_reclaim_inodes_nr(XFS_M(sb), sc->nr_to_scan);
|
2011-07-08 08:14:46 +04:00
|
|
|
}
|
|
|
|
|
2009-09-22 04:01:09 +04:00
|
|
|
static const struct super_operations xfs_super_operations = {
|
2006-03-14 06:06:18 +03:00
|
|
|
.alloc_inode = xfs_fs_alloc_inode,
|
|
|
|
.destroy_inode = xfs_fs_destroy_inode,
|
2018-03-07 04:04:00 +03:00
|
|
|
.dirty_inode = xfs_fs_dirty_inode,
|
2012-03-22 09:15:10 +04:00
|
|
|
.drop_inode = xfs_fs_drop_inode,
|
2006-03-14 06:06:18 +03:00
|
|
|
.put_super = xfs_fs_put_super,
|
2009-10-07 00:29:28 +04:00
|
|
|
.sync_fs = xfs_fs_sync_fs,
|
2009-01-10 03:40:58 +03:00
|
|
|
.freeze_fs = xfs_fs_freeze,
|
2010-02-06 01:59:53 +03:00
|
|
|
.unfreeze_fs = xfs_fs_unfreeze,
|
2006-03-14 06:06:18 +03:00
|
|
|
.statfs = xfs_fs_statfs,
|
|
|
|
.remount_fs = xfs_fs_remount,
|
|
|
|
.show_options = xfs_fs_show_options,
|
2011-07-08 08:14:46 +04:00
|
|
|
.nr_cached_objects = xfs_fs_nr_cached_objects,
|
|
|
|
.free_cached_objects = xfs_fs_free_cached_objects,
|
2005-04-17 02:20:36 +04:00
|
|
|
};
|
|
|
|
|
2007-02-21 00:57:47 +03:00
|
|
|
static struct file_system_type xfs_fs_type = {
|
2005-04-17 02:20:36 +04:00
|
|
|
.owner = THIS_MODULE,
|
|
|
|
.name = "xfs",
|
2010-07-25 00:46:55 +04:00
|
|
|
.mount = xfs_fs_mount,
|
2005-04-17 02:20:36 +04:00
|
|
|
.kill_sb = kill_block_super,
|
|
|
|
.fs_flags = FS_REQUIRES_DEV,
|
|
|
|
};
|
2013-03-03 07:39:14 +04:00
|
|
|
MODULE_ALIAS_FS("xfs");
|
2005-04-17 02:20:36 +04:00
|
|
|
|
2008-07-18 11:11:46 +04:00
|
|
|
STATIC int __init
|
|
|
|
xfs_init_zones(void)
|
|
|
|
{
|
|
|
|
xfs_log_ticket_zone = kmem_zone_init(sizeof(xlog_ticket_t),
|
|
|
|
"xfs_log_ticket");
|
|
|
|
if (!xfs_log_ticket_zone)
|
2019-10-17 23:12:15 +03:00
|
|
|
goto out;
|
2008-07-18 11:11:46 +04:00
|
|
|
|
2016-06-21 04:53:28 +03:00
|
|
|
xfs_bmap_free_item_zone = kmem_zone_init(
|
2016-08-03 04:18:10 +03:00
|
|
|
sizeof(struct xfs_extent_free_item),
|
2016-06-21 04:53:28 +03:00
|
|
|
"xfs_bmap_free_item");
|
2008-07-18 11:11:46 +04:00
|
|
|
if (!xfs_bmap_free_item_zone)
|
|
|
|
goto out_destroy_log_ticket_zone;
|
2008-10-30 09:36:14 +03:00
|
|
|
|
2008-07-18 11:11:46 +04:00
|
|
|
xfs_btree_cur_zone = kmem_zone_init(sizeof(xfs_btree_cur_t),
|
|
|
|
"xfs_btree_cur");
|
|
|
|
if (!xfs_btree_cur_zone)
|
|
|
|
goto out_destroy_bmap_free_item_zone;
|
|
|
|
|
|
|
|
xfs_da_state_zone = kmem_zone_init(sizeof(xfs_da_state_t),
|
|
|
|
"xfs_da_state");
|
|
|
|
if (!xfs_da_state_zone)
|
|
|
|
goto out_destroy_btree_cur_zone;
|
|
|
|
|
2018-07-18 02:51:50 +03:00
|
|
|
xfs_ifork_zone = kmem_zone_init(sizeof(struct xfs_ifork), "xfs_ifork");
|
2008-07-18 11:11:46 +04:00
|
|
|
if (!xfs_ifork_zone)
|
2012-06-22 12:50:14 +04:00
|
|
|
goto out_destroy_da_state_zone;
|
2008-07-18 11:11:46 +04:00
|
|
|
|
|
|
|
xfs_trans_zone = kmem_zone_init(sizeof(xfs_trans_t), "xfs_trans");
|
|
|
|
if (!xfs_trans_zone)
|
|
|
|
goto out_destroy_ifork_zone;
|
|
|
|
|
2010-06-23 12:11:15 +04:00
|
|
|
|
2008-07-18 11:11:46 +04:00
|
|
|
/*
|
|
|
|
* The size of the zone allocated buf log item is the maximum
|
|
|
|
* size possible under XFS. This wastes a little bit of memory,
|
|
|
|
* but it is much faster.
|
|
|
|
*/
|
2012-06-22 12:50:07 +04:00
|
|
|
xfs_buf_item_zone = kmem_zone_init(sizeof(struct xfs_buf_log_item),
|
|
|
|
"xfs_buf_item");
|
2008-07-18 11:11:46 +04:00
|
|
|
if (!xfs_buf_item_zone)
|
2018-05-09 17:49:37 +03:00
|
|
|
goto out_destroy_trans_zone;
|
2008-07-18 11:11:46 +04:00
|
|
|
|
|
|
|
xfs_efd_zone = kmem_zone_init((sizeof(xfs_efd_log_item_t) +
|
|
|
|
((XFS_EFD_MAX_FAST_EXTENTS - 1) *
|
|
|
|
sizeof(xfs_extent_t))), "xfs_efd_item");
|
|
|
|
if (!xfs_efd_zone)
|
|
|
|
goto out_destroy_buf_item_zone;
|
|
|
|
|
|
|
|
xfs_efi_zone = kmem_zone_init((sizeof(xfs_efi_log_item_t) +
|
|
|
|
((XFS_EFI_MAX_FAST_EXTENTS - 1) *
|
|
|
|
sizeof(xfs_extent_t))), "xfs_efi_item");
|
|
|
|
if (!xfs_efi_zone)
|
|
|
|
goto out_destroy_efd_zone;
|
|
|
|
|
|
|
|
xfs_inode_zone =
|
|
|
|
kmem_zone_init_flags(sizeof(xfs_inode_t), "xfs_inode",
|
2016-01-15 02:18:21 +03:00
|
|
|
KM_ZONE_HWALIGN | KM_ZONE_RECLAIM | KM_ZONE_SPREAD |
|
|
|
|
KM_ZONE_ACCOUNT, xfs_fs_inode_init_once);
|
2008-07-18 11:11:46 +04:00
|
|
|
if (!xfs_inode_zone)
|
|
|
|
goto out_destroy_efi_zone;
|
|
|
|
|
|
|
|
xfs_ili_zone =
|
|
|
|
kmem_zone_init_flags(sizeof(xfs_inode_log_item_t), "xfs_ili",
|
|
|
|
KM_ZONE_SPREAD, NULL);
|
|
|
|
if (!xfs_ili_zone)
|
|
|
|
goto out_destroy_inode_zone;
|
2013-06-27 10:04:53 +04:00
|
|
|
xfs_icreate_zone = kmem_zone_init(sizeof(struct xfs_icreate_item),
|
|
|
|
"xfs_icr");
|
|
|
|
if (!xfs_icreate_zone)
|
|
|
|
goto out_destroy_ili_zone;
|
2008-07-18 11:11:46 +04:00
|
|
|
|
2016-08-03 05:28:43 +03:00
|
|
|
xfs_rud_zone = kmem_zone_init(sizeof(struct xfs_rud_log_item),
|
2016-08-03 05:04:45 +03:00
|
|
|
"xfs_rud_item");
|
|
|
|
if (!xfs_rud_zone)
|
|
|
|
goto out_destroy_icreate_zone;
|
|
|
|
|
2016-09-19 03:24:27 +03:00
|
|
|
xfs_rui_zone = kmem_zone_init(
|
|
|
|
xfs_rui_log_item_sizeof(XFS_RUI_MAX_FAST_EXTENTS),
|
2016-08-03 05:04:45 +03:00
|
|
|
"xfs_rui_item");
|
|
|
|
if (!xfs_rui_zone)
|
|
|
|
goto out_destroy_rud_zone;
|
|
|
|
|
2016-10-03 19:11:20 +03:00
|
|
|
xfs_cud_zone = kmem_zone_init(sizeof(struct xfs_cud_log_item),
|
|
|
|
"xfs_cud_item");
|
|
|
|
if (!xfs_cud_zone)
|
|
|
|
goto out_destroy_rui_zone;
|
|
|
|
|
|
|
|
xfs_cui_zone = kmem_zone_init(
|
|
|
|
xfs_cui_log_item_sizeof(XFS_CUI_MAX_FAST_EXTENTS),
|
|
|
|
"xfs_cui_item");
|
|
|
|
if (!xfs_cui_zone)
|
|
|
|
goto out_destroy_cud_zone;
|
|
|
|
|
2016-10-03 19:11:25 +03:00
|
|
|
xfs_bud_zone = kmem_zone_init(sizeof(struct xfs_bud_log_item),
|
|
|
|
"xfs_bud_item");
|
|
|
|
if (!xfs_bud_zone)
|
|
|
|
goto out_destroy_cui_zone;
|
|
|
|
|
|
|
|
xfs_bui_zone = kmem_zone_init(
|
|
|
|
xfs_bui_log_item_sizeof(XFS_BUI_MAX_FAST_EXTENTS),
|
|
|
|
"xfs_bui_item");
|
|
|
|
if (!xfs_bui_zone)
|
|
|
|
goto out_destroy_bud_zone;
|
|
|
|
|
2008-07-18 11:11:46 +04:00
|
|
|
return 0;
|
|
|
|
|
2016-10-03 19:11:25 +03:00
|
|
|
out_destroy_bud_zone:
|
|
|
|
kmem_zone_destroy(xfs_bud_zone);
|
|
|
|
out_destroy_cui_zone:
|
|
|
|
kmem_zone_destroy(xfs_cui_zone);
|
2016-10-03 19:11:20 +03:00
|
|
|
out_destroy_cud_zone:
|
|
|
|
kmem_zone_destroy(xfs_cud_zone);
|
|
|
|
out_destroy_rui_zone:
|
|
|
|
kmem_zone_destroy(xfs_rui_zone);
|
2016-08-03 05:04:45 +03:00
|
|
|
out_destroy_rud_zone:
|
|
|
|
kmem_zone_destroy(xfs_rud_zone);
|
|
|
|
out_destroy_icreate_zone:
|
|
|
|
kmem_zone_destroy(xfs_icreate_zone);
|
2013-06-27 10:04:53 +04:00
|
|
|
out_destroy_ili_zone:
|
|
|
|
kmem_zone_destroy(xfs_ili_zone);
|
2008-07-18 11:11:46 +04:00
|
|
|
out_destroy_inode_zone:
|
|
|
|
kmem_zone_destroy(xfs_inode_zone);
|
|
|
|
out_destroy_efi_zone:
|
|
|
|
kmem_zone_destroy(xfs_efi_zone);
|
|
|
|
out_destroy_efd_zone:
|
|
|
|
kmem_zone_destroy(xfs_efd_zone);
|
|
|
|
out_destroy_buf_item_zone:
|
|
|
|
kmem_zone_destroy(xfs_buf_item_zone);
|
|
|
|
out_destroy_trans_zone:
|
|
|
|
kmem_zone_destroy(xfs_trans_zone);
|
|
|
|
out_destroy_ifork_zone:
|
|
|
|
kmem_zone_destroy(xfs_ifork_zone);
|
|
|
|
out_destroy_da_state_zone:
|
|
|
|
kmem_zone_destroy(xfs_da_state_zone);
|
|
|
|
out_destroy_btree_cur_zone:
|
|
|
|
kmem_zone_destroy(xfs_btree_cur_zone);
|
|
|
|
out_destroy_bmap_free_item_zone:
|
|
|
|
kmem_zone_destroy(xfs_bmap_free_item_zone);
|
|
|
|
out_destroy_log_ticket_zone:
|
|
|
|
kmem_zone_destroy(xfs_log_ticket_zone);
|
|
|
|
out:
|
|
|
|
return -ENOMEM;
|
|
|
|
}
|
|
|
|
|
|
|
|
STATIC void
|
|
|
|
xfs_destroy_zones(void)
|
|
|
|
{
|
2012-09-26 05:33:07 +04:00
|
|
|
/*
|
|
|
|
* Make sure all delayed rcu free are flushed before we
|
|
|
|
* destroy caches.
|
|
|
|
*/
|
|
|
|
rcu_barrier();
|
2016-10-03 19:11:25 +03:00
|
|
|
kmem_zone_destroy(xfs_bui_zone);
|
|
|
|
kmem_zone_destroy(xfs_bud_zone);
|
2016-10-03 19:11:20 +03:00
|
|
|
kmem_zone_destroy(xfs_cui_zone);
|
|
|
|
kmem_zone_destroy(xfs_cud_zone);
|
2016-08-03 05:04:45 +03:00
|
|
|
kmem_zone_destroy(xfs_rui_zone);
|
|
|
|
kmem_zone_destroy(xfs_rud_zone);
|
2013-06-27 10:04:53 +04:00
|
|
|
kmem_zone_destroy(xfs_icreate_zone);
|
2008-07-18 11:11:46 +04:00
|
|
|
kmem_zone_destroy(xfs_ili_zone);
|
|
|
|
kmem_zone_destroy(xfs_inode_zone);
|
|
|
|
kmem_zone_destroy(xfs_efi_zone);
|
|
|
|
kmem_zone_destroy(xfs_efd_zone);
|
|
|
|
kmem_zone_destroy(xfs_buf_item_zone);
|
|
|
|
kmem_zone_destroy(xfs_trans_zone);
|
|
|
|
kmem_zone_destroy(xfs_ifork_zone);
|
|
|
|
kmem_zone_destroy(xfs_da_state_zone);
|
|
|
|
kmem_zone_destroy(xfs_btree_cur_zone);
|
|
|
|
kmem_zone_destroy(xfs_bmap_free_item_zone);
|
|
|
|
kmem_zone_destroy(xfs_log_ticket_zone);
|
|
|
|
}
|
2005-04-17 02:20:36 +04:00
|
|
|
|
2011-04-08 06:45:07 +04:00
|
|
|
STATIC int __init
|
|
|
|
xfs_init_workqueues(void)
|
|
|
|
{
|
2012-03-22 09:15:07 +04:00
|
|
|
/*
|
|
|
|
* The allocation workqueue can be used in memory reclaim situations
|
|
|
|
* (writepage path), and parallelism is only limited by the number of
|
|
|
|
* AGs in all the filesystems mounted. Hence use the default large
|
|
|
|
* max_active value for this workqueue.
|
|
|
|
*/
|
2014-09-09 05:44:46 +04:00
|
|
|
xfs_alloc_wq = alloc_workqueue("xfsalloc",
|
|
|
|
WQ_MEM_RECLAIM|WQ_FREEZABLE, 0);
|
2012-03-22 09:15:07 +04:00
|
|
|
if (!xfs_alloc_wq)
|
2012-10-08 14:56:05 +04:00
|
|
|
return -ENOMEM;
|
2012-03-22 09:15:07 +04:00
|
|
|
|
2017-02-08 01:07:58 +03:00
|
|
|
xfs_discard_wq = alloc_workqueue("xfsdiscard", WQ_UNBOUND, 0);
|
|
|
|
if (!xfs_discard_wq)
|
|
|
|
goto out_free_alloc_wq;
|
|
|
|
|
2011-04-08 06:45:07 +04:00
|
|
|
return 0;
|
2017-02-08 01:07:58 +03:00
|
|
|
out_free_alloc_wq:
|
|
|
|
destroy_workqueue(xfs_alloc_wq);
|
|
|
|
return -ENOMEM;
|
2011-04-08 06:45:07 +04:00
|
|
|
}
|
|
|
|
|
2011-04-11 23:06:12 +04:00
|
|
|
STATIC void
|
2011-04-08 06:45:07 +04:00
|
|
|
xfs_destroy_workqueues(void)
|
|
|
|
{
|
2017-02-08 01:07:58 +03:00
|
|
|
destroy_workqueue(xfs_discard_wq);
|
2012-03-22 09:15:07 +04:00
|
|
|
destroy_workqueue(xfs_alloc_wq);
|
2011-04-08 06:45:07 +04:00
|
|
|
}
|
|
|
|
|
2005-04-17 02:20:36 +04:00
|
|
|
STATIC int __init
|
2008-07-18 11:11:46 +04:00
|
|
|
init_xfs_fs(void)
|
2005-04-17 02:20:36 +04:00
|
|
|
{
|
|
|
|
int error;
|
|
|
|
|
2016-03-09 00:15:14 +03:00
|
|
|
xfs_check_ondisk_structs();
|
|
|
|
|
2008-11-28 06:23:33 +03:00
|
|
|
printk(KERN_INFO XFS_VERSION_STRING " with "
|
|
|
|
XFS_BUILD_OPTIONS " enabled\n");
|
2005-04-17 02:20:36 +04:00
|
|
|
|
2008-07-18 11:11:46 +04:00
|
|
|
xfs_dir_startup();
|
2005-04-17 02:20:36 +04:00
|
|
|
|
2006-03-14 05:18:19 +03:00
|
|
|
error = xfs_init_zones();
|
2008-07-18 11:11:46 +04:00
|
|
|
if (error)
|
|
|
|
goto out;
|
|
|
|
|
2011-04-08 06:45:07 +04:00
|
|
|
error = xfs_init_workqueues();
|
2008-07-18 11:11:46 +04:00
|
|
|
if (error)
|
2009-12-15 02:14:59 +03:00
|
|
|
goto out_destroy_zones;
|
2008-07-18 11:11:46 +04:00
|
|
|
|
2011-04-08 06:45:07 +04:00
|
|
|
error = xfs_mru_cache_init();
|
|
|
|
if (error)
|
|
|
|
goto out_destroy_wq;
|
|
|
|
|
2006-01-11 07:39:08 +03:00
|
|
|
error = xfs_buf_init();
|
2008-07-18 11:11:46 +04:00
|
|
|
if (error)
|
2014-04-23 01:11:51 +04:00
|
|
|
goto out_mru_cache_uninit;
|
2008-07-18 11:11:46 +04:00
|
|
|
|
|
|
|
error = xfs_init_procfs();
|
|
|
|
if (error)
|
|
|
|
goto out_buf_terminate;
|
|
|
|
|
|
|
|
error = xfs_sysctl_register();
|
|
|
|
if (error)
|
|
|
|
goto out_cleanup_procfs;
|
2005-04-17 02:20:36 +04:00
|
|
|
|
2014-07-15 01:41:37 +04:00
|
|
|
xfs_kset = kset_create_and_add("xfs", NULL, fs_kobj);
|
|
|
|
if (!xfs_kset) {
|
|
|
|
error = -ENOMEM;
|
2015-10-11 21:15:45 +03:00
|
|
|
goto out_sysctl_unregister;
|
2014-07-15 01:41:37 +04:00
|
|
|
}
|
|
|
|
|
2015-10-11 21:19:45 +03:00
|
|
|
xfsstats.xs_kobj.kobject.kset = xfs_kset;
|
|
|
|
|
|
|
|
xfsstats.xs_stats = alloc_percpu(struct xfsstats);
|
|
|
|
if (!xfsstats.xs_stats) {
|
|
|
|
error = -ENOMEM;
|
|
|
|
goto out_kset_unregister;
|
|
|
|
}
|
|
|
|
|
|
|
|
error = xfs_sysfs_init(&xfsstats.xs_kobj, &xfs_stats_ktype, NULL,
|
2015-10-11 21:15:45 +03:00
|
|
|
"stats");
|
|
|
|
if (error)
|
2015-10-11 21:19:45 +03:00
|
|
|
goto out_free_stats;
|
2015-10-11 21:15:45 +03:00
|
|
|
|
2014-09-09 05:52:42 +04:00
|
|
|
#ifdef DEBUG
|
|
|
|
xfs_dbg_kobj.kobject.kset = xfs_kset;
|
|
|
|
error = xfs_sysfs_init(&xfs_dbg_kobj, &xfs_dbg_ktype, NULL, "debug");
|
2012-03-13 12:52:37 +04:00
|
|
|
if (error)
|
2015-10-11 21:15:45 +03:00
|
|
|
goto out_remove_stats_kobj;
|
2014-09-09 05:52:42 +04:00
|
|
|
#endif
|
|
|
|
|
|
|
|
error = xfs_qm_init();
|
|
|
|
if (error)
|
2015-10-11 21:15:45 +03:00
|
|
|
goto out_remove_dbg_kobj;
|
2005-04-17 02:20:36 +04:00
|
|
|
|
|
|
|
error = register_filesystem(&xfs_fs_type);
|
|
|
|
if (error)
|
2012-03-13 12:52:37 +04:00
|
|
|
goto out_qm_exit;
|
2005-04-17 02:20:36 +04:00
|
|
|
return 0;
|
|
|
|
|
2012-03-13 12:52:37 +04:00
|
|
|
out_qm_exit:
|
|
|
|
xfs_qm_exit();
|
2015-10-11 21:15:45 +03:00
|
|
|
out_remove_dbg_kobj:
|
2014-09-09 05:52:42 +04:00
|
|
|
#ifdef DEBUG
|
|
|
|
xfs_sysfs_del(&xfs_dbg_kobj);
|
2015-10-11 21:15:45 +03:00
|
|
|
out_remove_stats_kobj:
|
2014-09-09 05:52:42 +04:00
|
|
|
#endif
|
2015-10-11 21:19:45 +03:00
|
|
|
xfs_sysfs_del(&xfsstats.xs_kobj);
|
|
|
|
out_free_stats:
|
|
|
|
free_percpu(xfsstats.xs_stats);
|
2015-10-11 21:15:45 +03:00
|
|
|
out_kset_unregister:
|
2014-07-15 01:41:37 +04:00
|
|
|
kset_unregister(xfs_kset);
|
2008-07-18 11:11:46 +04:00
|
|
|
out_sysctl_unregister:
|
|
|
|
xfs_sysctl_unregister();
|
|
|
|
out_cleanup_procfs:
|
|
|
|
xfs_cleanup_procfs();
|
|
|
|
out_buf_terminate:
|
2006-01-11 07:39:08 +03:00
|
|
|
xfs_buf_terminate();
|
2008-07-18 11:11:46 +04:00
|
|
|
out_mru_cache_uninit:
|
|
|
|
xfs_mru_cache_uninit();
|
2011-04-08 06:45:07 +04:00
|
|
|
out_destroy_wq:
|
|
|
|
xfs_destroy_workqueues();
|
2008-07-18 11:11:46 +04:00
|
|
|
out_destroy_zones:
|
2006-03-14 05:18:19 +03:00
|
|
|
xfs_destroy_zones();
|
2008-07-18 11:11:46 +04:00
|
|
|
out:
|
2005-04-17 02:20:36 +04:00
|
|
|
return error;
|
|
|
|
}
|
|
|
|
|
|
|
|
STATIC void __exit
|
2008-07-18 11:11:46 +04:00
|
|
|
exit_xfs_fs(void)
|
2005-04-17 02:20:36 +04:00
|
|
|
{
|
2012-03-13 12:52:37 +04:00
|
|
|
xfs_qm_exit();
|
2005-04-17 02:20:36 +04:00
|
|
|
unregister_filesystem(&xfs_fs_type);
|
2014-09-09 05:52:42 +04:00
|
|
|
#ifdef DEBUG
|
|
|
|
xfs_sysfs_del(&xfs_dbg_kobj);
|
|
|
|
#endif
|
2015-10-11 21:19:45 +03:00
|
|
|
xfs_sysfs_del(&xfsstats.xs_kobj);
|
|
|
|
free_percpu(xfsstats.xs_stats);
|
2014-07-15 01:41:37 +04:00
|
|
|
kset_unregister(xfs_kset);
|
2008-07-18 11:11:46 +04:00
|
|
|
xfs_sysctl_unregister();
|
|
|
|
xfs_cleanup_procfs();
|
2006-01-11 07:39:08 +03:00
|
|
|
xfs_buf_terminate();
|
2008-07-18 11:11:46 +04:00
|
|
|
xfs_mru_cache_uninit();
|
2011-04-08 06:45:07 +04:00
|
|
|
xfs_destroy_workqueues();
|
2006-03-14 05:18:19 +03:00
|
|
|
xfs_destroy_zones();
|
2015-11-03 05:06:34 +03:00
|
|
|
xfs_uuid_table_free();
|
2005-04-17 02:20:36 +04:00
|
|
|
}
|
|
|
|
|
|
|
|
module_init(init_xfs_fs);
|
|
|
|
module_exit(exit_xfs_fs);
|
|
|
|
|
|
|
|
MODULE_AUTHOR("Silicon Graphics, Inc.");
|
|
|
|
MODULE_DESCRIPTION(XFS_VERSION_STRING " with " XFS_BUILD_OPTIONS " enabled");
|
|
|
|
MODULE_LICENSE("GPL");
|