mirror of
https://git.kernel.org/pub/scm/linux/kernel/git/next/linux-next.git
synced 2024-12-28 16:52:18 +00:00
c1220522ef
The grant heads in the log track the space reserved in the log for running transactions. They do this by tracking how far ahead of the tail that the reservation has reached, and the units for doing this are {cycle,bytes} for the reserve head rather than {cycle,blocks} which are normal used by LSNs. This is annoyingly complex because we have to split, crack and combined these tuples for any calculation we do to determine log space and targets. This is computationally expensive as well as difficult to do atomically and locklessly, as well as limiting the size of the log to 2^32 bytes. Really, though, all the grant heads are tracking is how much space is currently available for use in the log. We can track this as a simply byte count - we just don't care what the actual physical location in the log the head and tail are at, just how much space we have remaining before the head and tail overlap. So, convert the grant heads to track the byte reservations that are active rather than the current (cycle, offset) tuples. This means an empty log has zero bytes consumed, and a full log is when the reservations reach the size of the log minus the space consumed by the AIL. This greatly simplifies the accounting and checks for whether there is space available. We no longer need to crack or combine LSNs to determine how much space the log has left, nor do we need to look at the head or tail of the log to determine how close to full we are. There is, however, a complexity that needs to be handled. We know how much space is being tracked in the AIL now via log->l_tail_space and the log tickets track active reservations and return the unused portions to the grant heads when ungranted. Unfortunately, we don't track the used portion of the grant, so when we transfer log items from the CIL to the AIL, the space accounted to the grant heads is transferred to the log tail space. Hence when we move the AIL head forwards on item insert, we have to remove that space from the grant heads. We also remove the xlog_verify_grant_tail() debug function as it is no longer useful. The check it performs has been racy since delayed logging was introduced, but now it is clearly only detecting false positives so remove it. The result of this substantially simpler accounting algorithm is an increase in sustained transaction rate from ~1.3 million transactions/s to ~1.9 million transactions/s with no increase in CPU usage. We also remove the 32 bit space limitation on the grant heads, which will allow us to increase the journal size beyond 2GB in future. Note that this renames the sysfs files exposing the log grant space now that the values are exported in bytes. This allows xfstests to auto-detect the old or new ABI. [hch: move xlog_grant_sub_space out of line, update the xlog_grant_{add,sub}_space prototypes, rename the sysfs files to allow auto-detection in xfstests] Signed-off-by: Dave Chinner <dchinner@redhat.com> Reviewed-by: Darrick J. Wong <djwong@kernel.org> Signed-off-by: Christoph Hellwig <hch@lst.de> Reviewed-by: Dave Chinner <dchinner@redhat.com> Signed-off-by: Chandan Babu R <chandanbabu@kernel.org>
782 lines
15 KiB
C
782 lines
15 KiB
C
// SPDX-License-Identifier: GPL-2.0
|
|
/*
|
|
* Copyright (c) 2014 Red Hat, Inc.
|
|
* All Rights Reserved.
|
|
*/
|
|
|
|
#include "xfs.h"
|
|
#include "xfs_shared.h"
|
|
#include "xfs_format.h"
|
|
#include "xfs_log_format.h"
|
|
#include "xfs_trans_resv.h"
|
|
#include "xfs_sysfs.h"
|
|
#include "xfs_log.h"
|
|
#include "xfs_log_priv.h"
|
|
#include "xfs_mount.h"
|
|
|
|
struct xfs_sysfs_attr {
|
|
struct attribute attr;
|
|
ssize_t (*show)(struct kobject *kobject, char *buf);
|
|
ssize_t (*store)(struct kobject *kobject, const char *buf,
|
|
size_t count);
|
|
};
|
|
|
|
static inline struct xfs_sysfs_attr *
|
|
to_attr(struct attribute *attr)
|
|
{
|
|
return container_of(attr, struct xfs_sysfs_attr, attr);
|
|
}
|
|
|
|
#define XFS_SYSFS_ATTR_RW(name) \
|
|
static struct xfs_sysfs_attr xfs_sysfs_attr_##name = __ATTR_RW(name)
|
|
#define XFS_SYSFS_ATTR_RO(name) \
|
|
static struct xfs_sysfs_attr xfs_sysfs_attr_##name = __ATTR_RO(name)
|
|
#define XFS_SYSFS_ATTR_WO(name) \
|
|
static struct xfs_sysfs_attr xfs_sysfs_attr_##name = __ATTR_WO(name)
|
|
|
|
#define ATTR_LIST(name) &xfs_sysfs_attr_##name.attr
|
|
|
|
STATIC ssize_t
|
|
xfs_sysfs_object_show(
|
|
struct kobject *kobject,
|
|
struct attribute *attr,
|
|
char *buf)
|
|
{
|
|
struct xfs_sysfs_attr *xfs_attr = to_attr(attr);
|
|
|
|
return xfs_attr->show ? xfs_attr->show(kobject, buf) : 0;
|
|
}
|
|
|
|
STATIC ssize_t
|
|
xfs_sysfs_object_store(
|
|
struct kobject *kobject,
|
|
struct attribute *attr,
|
|
const char *buf,
|
|
size_t count)
|
|
{
|
|
struct xfs_sysfs_attr *xfs_attr = to_attr(attr);
|
|
|
|
return xfs_attr->store ? xfs_attr->store(kobject, buf, count) : 0;
|
|
}
|
|
|
|
static const struct sysfs_ops xfs_sysfs_ops = {
|
|
.show = xfs_sysfs_object_show,
|
|
.store = xfs_sysfs_object_store,
|
|
};
|
|
|
|
static struct attribute *xfs_mp_attrs[] = {
|
|
NULL,
|
|
};
|
|
ATTRIBUTE_GROUPS(xfs_mp);
|
|
|
|
const struct kobj_type xfs_mp_ktype = {
|
|
.release = xfs_sysfs_release,
|
|
.sysfs_ops = &xfs_sysfs_ops,
|
|
.default_groups = xfs_mp_groups,
|
|
};
|
|
|
|
#ifdef DEBUG
|
|
/* debug */
|
|
|
|
STATIC ssize_t
|
|
bug_on_assert_store(
|
|
struct kobject *kobject,
|
|
const char *buf,
|
|
size_t count)
|
|
{
|
|
int ret;
|
|
int val;
|
|
|
|
ret = kstrtoint(buf, 0, &val);
|
|
if (ret)
|
|
return ret;
|
|
|
|
if (val == 1)
|
|
xfs_globals.bug_on_assert = true;
|
|
else if (val == 0)
|
|
xfs_globals.bug_on_assert = false;
|
|
else
|
|
return -EINVAL;
|
|
|
|
return count;
|
|
}
|
|
|
|
STATIC ssize_t
|
|
bug_on_assert_show(
|
|
struct kobject *kobject,
|
|
char *buf)
|
|
{
|
|
return sysfs_emit(buf, "%d\n", xfs_globals.bug_on_assert);
|
|
}
|
|
XFS_SYSFS_ATTR_RW(bug_on_assert);
|
|
|
|
STATIC ssize_t
|
|
log_recovery_delay_store(
|
|
struct kobject *kobject,
|
|
const char *buf,
|
|
size_t count)
|
|
{
|
|
int ret;
|
|
int val;
|
|
|
|
ret = kstrtoint(buf, 0, &val);
|
|
if (ret)
|
|
return ret;
|
|
|
|
if (val < 0 || val > 60)
|
|
return -EINVAL;
|
|
|
|
xfs_globals.log_recovery_delay = val;
|
|
|
|
return count;
|
|
}
|
|
|
|
STATIC ssize_t
|
|
log_recovery_delay_show(
|
|
struct kobject *kobject,
|
|
char *buf)
|
|
{
|
|
return sysfs_emit(buf, "%d\n", xfs_globals.log_recovery_delay);
|
|
}
|
|
XFS_SYSFS_ATTR_RW(log_recovery_delay);
|
|
|
|
STATIC ssize_t
|
|
mount_delay_store(
|
|
struct kobject *kobject,
|
|
const char *buf,
|
|
size_t count)
|
|
{
|
|
int ret;
|
|
int val;
|
|
|
|
ret = kstrtoint(buf, 0, &val);
|
|
if (ret)
|
|
return ret;
|
|
|
|
if (val < 0 || val > 60)
|
|
return -EINVAL;
|
|
|
|
xfs_globals.mount_delay = val;
|
|
|
|
return count;
|
|
}
|
|
|
|
STATIC ssize_t
|
|
mount_delay_show(
|
|
struct kobject *kobject,
|
|
char *buf)
|
|
{
|
|
return sysfs_emit(buf, "%d\n", xfs_globals.mount_delay);
|
|
}
|
|
XFS_SYSFS_ATTR_RW(mount_delay);
|
|
|
|
static ssize_t
|
|
always_cow_store(
|
|
struct kobject *kobject,
|
|
const char *buf,
|
|
size_t count)
|
|
{
|
|
ssize_t ret;
|
|
|
|
ret = kstrtobool(buf, &xfs_globals.always_cow);
|
|
if (ret < 0)
|
|
return ret;
|
|
return count;
|
|
}
|
|
|
|
static ssize_t
|
|
always_cow_show(
|
|
struct kobject *kobject,
|
|
char *buf)
|
|
{
|
|
return sysfs_emit(buf, "%d\n", xfs_globals.always_cow);
|
|
}
|
|
XFS_SYSFS_ATTR_RW(always_cow);
|
|
|
|
/*
|
|
* Override how many threads the parallel work queue is allowed to create.
|
|
* This has to be a debug-only global (instead of an errortag) because one of
|
|
* the main users of parallel workqueues is mount time quotacheck.
|
|
*/
|
|
STATIC ssize_t
|
|
pwork_threads_store(
|
|
struct kobject *kobject,
|
|
const char *buf,
|
|
size_t count)
|
|
{
|
|
int ret;
|
|
int val;
|
|
|
|
ret = kstrtoint(buf, 0, &val);
|
|
if (ret)
|
|
return ret;
|
|
|
|
if (val < -1 || val > num_possible_cpus())
|
|
return -EINVAL;
|
|
|
|
xfs_globals.pwork_threads = val;
|
|
|
|
return count;
|
|
}
|
|
|
|
STATIC ssize_t
|
|
pwork_threads_show(
|
|
struct kobject *kobject,
|
|
char *buf)
|
|
{
|
|
return sysfs_emit(buf, "%d\n", xfs_globals.pwork_threads);
|
|
}
|
|
XFS_SYSFS_ATTR_RW(pwork_threads);
|
|
|
|
/*
|
|
* The "LARP" (Logged extended Attribute Recovery Persistence) debugging knob
|
|
* sets the XFS_DA_OP_LOGGED flag on all xfs_attr_set operations performed on
|
|
* V5 filesystems. As a result, the intermediate progress of all setxattr and
|
|
* removexattr operations are tracked via the log and can be restarted during
|
|
* recovery. This is useful for testing xattr recovery prior to merging of the
|
|
* parent pointer feature which requires it to maintain consistency, and may be
|
|
* enabled for userspace xattrs in the future.
|
|
*/
|
|
static ssize_t
|
|
larp_store(
|
|
struct kobject *kobject,
|
|
const char *buf,
|
|
size_t count)
|
|
{
|
|
ssize_t ret;
|
|
|
|
ret = kstrtobool(buf, &xfs_globals.larp);
|
|
if (ret < 0)
|
|
return ret;
|
|
return count;
|
|
}
|
|
|
|
STATIC ssize_t
|
|
larp_show(
|
|
struct kobject *kobject,
|
|
char *buf)
|
|
{
|
|
return snprintf(buf, PAGE_SIZE, "%d\n", xfs_globals.larp);
|
|
}
|
|
XFS_SYSFS_ATTR_RW(larp);
|
|
|
|
STATIC ssize_t
|
|
bload_leaf_slack_store(
|
|
struct kobject *kobject,
|
|
const char *buf,
|
|
size_t count)
|
|
{
|
|
int ret;
|
|
int val;
|
|
|
|
ret = kstrtoint(buf, 0, &val);
|
|
if (ret)
|
|
return ret;
|
|
|
|
xfs_globals.bload_leaf_slack = val;
|
|
return count;
|
|
}
|
|
|
|
STATIC ssize_t
|
|
bload_leaf_slack_show(
|
|
struct kobject *kobject,
|
|
char *buf)
|
|
{
|
|
return snprintf(buf, PAGE_SIZE, "%d\n", xfs_globals.bload_leaf_slack);
|
|
}
|
|
XFS_SYSFS_ATTR_RW(bload_leaf_slack);
|
|
|
|
STATIC ssize_t
|
|
bload_node_slack_store(
|
|
struct kobject *kobject,
|
|
const char *buf,
|
|
size_t count)
|
|
{
|
|
int ret;
|
|
int val;
|
|
|
|
ret = kstrtoint(buf, 0, &val);
|
|
if (ret)
|
|
return ret;
|
|
|
|
xfs_globals.bload_node_slack = val;
|
|
return count;
|
|
}
|
|
|
|
STATIC ssize_t
|
|
bload_node_slack_show(
|
|
struct kobject *kobject,
|
|
char *buf)
|
|
{
|
|
return snprintf(buf, PAGE_SIZE, "%d\n", xfs_globals.bload_node_slack);
|
|
}
|
|
XFS_SYSFS_ATTR_RW(bload_node_slack);
|
|
|
|
static struct attribute *xfs_dbg_attrs[] = {
|
|
ATTR_LIST(bug_on_assert),
|
|
ATTR_LIST(log_recovery_delay),
|
|
ATTR_LIST(mount_delay),
|
|
ATTR_LIST(always_cow),
|
|
ATTR_LIST(pwork_threads),
|
|
ATTR_LIST(larp),
|
|
ATTR_LIST(bload_leaf_slack),
|
|
ATTR_LIST(bload_node_slack),
|
|
NULL,
|
|
};
|
|
ATTRIBUTE_GROUPS(xfs_dbg);
|
|
|
|
const struct kobj_type xfs_dbg_ktype = {
|
|
.release = xfs_sysfs_release,
|
|
.sysfs_ops = &xfs_sysfs_ops,
|
|
.default_groups = xfs_dbg_groups,
|
|
};
|
|
|
|
#endif /* DEBUG */
|
|
|
|
/* stats */
|
|
|
|
static inline struct xstats *
|
|
to_xstats(struct kobject *kobject)
|
|
{
|
|
struct xfs_kobj *kobj = to_kobj(kobject);
|
|
|
|
return container_of(kobj, struct xstats, xs_kobj);
|
|
}
|
|
|
|
STATIC ssize_t
|
|
stats_show(
|
|
struct kobject *kobject,
|
|
char *buf)
|
|
{
|
|
struct xstats *stats = to_xstats(kobject);
|
|
|
|
return xfs_stats_format(stats->xs_stats, buf);
|
|
}
|
|
XFS_SYSFS_ATTR_RO(stats);
|
|
|
|
STATIC ssize_t
|
|
stats_clear_store(
|
|
struct kobject *kobject,
|
|
const char *buf,
|
|
size_t count)
|
|
{
|
|
int ret;
|
|
int val;
|
|
struct xstats *stats = to_xstats(kobject);
|
|
|
|
ret = kstrtoint(buf, 0, &val);
|
|
if (ret)
|
|
return ret;
|
|
|
|
if (val != 1)
|
|
return -EINVAL;
|
|
|
|
xfs_stats_clearall(stats->xs_stats);
|
|
return count;
|
|
}
|
|
XFS_SYSFS_ATTR_WO(stats_clear);
|
|
|
|
static struct attribute *xfs_stats_attrs[] = {
|
|
ATTR_LIST(stats),
|
|
ATTR_LIST(stats_clear),
|
|
NULL,
|
|
};
|
|
ATTRIBUTE_GROUPS(xfs_stats);
|
|
|
|
const struct kobj_type xfs_stats_ktype = {
|
|
.release = xfs_sysfs_release,
|
|
.sysfs_ops = &xfs_sysfs_ops,
|
|
.default_groups = xfs_stats_groups,
|
|
};
|
|
|
|
/* xlog */
|
|
|
|
static inline struct xlog *
|
|
to_xlog(struct kobject *kobject)
|
|
{
|
|
struct xfs_kobj *kobj = to_kobj(kobject);
|
|
|
|
return container_of(kobj, struct xlog, l_kobj);
|
|
}
|
|
|
|
STATIC ssize_t
|
|
log_head_lsn_show(
|
|
struct kobject *kobject,
|
|
char *buf)
|
|
{
|
|
int cycle;
|
|
int block;
|
|
struct xlog *log = to_xlog(kobject);
|
|
|
|
spin_lock(&log->l_icloglock);
|
|
cycle = log->l_curr_cycle;
|
|
block = log->l_curr_block;
|
|
spin_unlock(&log->l_icloglock);
|
|
|
|
return sysfs_emit(buf, "%d:%d\n", cycle, block);
|
|
}
|
|
XFS_SYSFS_ATTR_RO(log_head_lsn);
|
|
|
|
STATIC ssize_t
|
|
log_tail_lsn_show(
|
|
struct kobject *kobject,
|
|
char *buf)
|
|
{
|
|
int cycle;
|
|
int block;
|
|
struct xlog *log = to_xlog(kobject);
|
|
|
|
xlog_crack_atomic_lsn(&log->l_tail_lsn, &cycle, &block);
|
|
return sysfs_emit(buf, "%d:%d\n", cycle, block);
|
|
}
|
|
XFS_SYSFS_ATTR_RO(log_tail_lsn);
|
|
|
|
STATIC ssize_t
|
|
reserve_grant_head_bytes_show(
|
|
struct kobject *kobject,
|
|
char *buf)
|
|
{
|
|
return sysfs_emit(buf, "%lld\n",
|
|
atomic64_read(&to_xlog(kobject)->l_reserve_head.grant));
|
|
}
|
|
XFS_SYSFS_ATTR_RO(reserve_grant_head_bytes);
|
|
|
|
STATIC ssize_t
|
|
write_grant_head_bytes_show(
|
|
struct kobject *kobject,
|
|
char *buf)
|
|
{
|
|
return sysfs_emit(buf, "%lld\n",
|
|
atomic64_read(&to_xlog(kobject)->l_write_head.grant));
|
|
}
|
|
XFS_SYSFS_ATTR_RO(write_grant_head_bytes);
|
|
|
|
static struct attribute *xfs_log_attrs[] = {
|
|
ATTR_LIST(log_head_lsn),
|
|
ATTR_LIST(log_tail_lsn),
|
|
ATTR_LIST(reserve_grant_head_bytes),
|
|
ATTR_LIST(write_grant_head_bytes),
|
|
NULL,
|
|
};
|
|
ATTRIBUTE_GROUPS(xfs_log);
|
|
|
|
const struct kobj_type xfs_log_ktype = {
|
|
.release = xfs_sysfs_release,
|
|
.sysfs_ops = &xfs_sysfs_ops,
|
|
.default_groups = xfs_log_groups,
|
|
};
|
|
|
|
/*
|
|
* Metadata IO error configuration
|
|
*
|
|
* The sysfs structure here is:
|
|
* ...xfs/<dev>/error/<class>/<errno>/<error_attrs>
|
|
*
|
|
* where <class> allows us to discriminate between data IO and metadata IO,
|
|
* and any other future type of IO (e.g. special inode or directory error
|
|
* handling) we care to support.
|
|
*/
|
|
static inline struct xfs_error_cfg *
|
|
to_error_cfg(struct kobject *kobject)
|
|
{
|
|
struct xfs_kobj *kobj = to_kobj(kobject);
|
|
return container_of(kobj, struct xfs_error_cfg, kobj);
|
|
}
|
|
|
|
static inline struct xfs_mount *
|
|
err_to_mp(struct kobject *kobject)
|
|
{
|
|
struct xfs_kobj *kobj = to_kobj(kobject);
|
|
return container_of(kobj, struct xfs_mount, m_error_kobj);
|
|
}
|
|
|
|
static ssize_t
|
|
max_retries_show(
|
|
struct kobject *kobject,
|
|
char *buf)
|
|
{
|
|
int retries;
|
|
struct xfs_error_cfg *cfg = to_error_cfg(kobject);
|
|
|
|
if (cfg->max_retries == XFS_ERR_RETRY_FOREVER)
|
|
retries = -1;
|
|
else
|
|
retries = cfg->max_retries;
|
|
|
|
return sysfs_emit(buf, "%d\n", retries);
|
|
}
|
|
|
|
static ssize_t
|
|
max_retries_store(
|
|
struct kobject *kobject,
|
|
const char *buf,
|
|
size_t count)
|
|
{
|
|
struct xfs_error_cfg *cfg = to_error_cfg(kobject);
|
|
int ret;
|
|
int val;
|
|
|
|
ret = kstrtoint(buf, 0, &val);
|
|
if (ret)
|
|
return ret;
|
|
|
|
if (val < -1)
|
|
return -EINVAL;
|
|
|
|
if (val == -1)
|
|
cfg->max_retries = XFS_ERR_RETRY_FOREVER;
|
|
else
|
|
cfg->max_retries = val;
|
|
return count;
|
|
}
|
|
XFS_SYSFS_ATTR_RW(max_retries);
|
|
|
|
static ssize_t
|
|
retry_timeout_seconds_show(
|
|
struct kobject *kobject,
|
|
char *buf)
|
|
{
|
|
int timeout;
|
|
struct xfs_error_cfg *cfg = to_error_cfg(kobject);
|
|
|
|
if (cfg->retry_timeout == XFS_ERR_RETRY_FOREVER)
|
|
timeout = -1;
|
|
else
|
|
timeout = jiffies_to_msecs(cfg->retry_timeout) / MSEC_PER_SEC;
|
|
|
|
return sysfs_emit(buf, "%d\n", timeout);
|
|
}
|
|
|
|
static ssize_t
|
|
retry_timeout_seconds_store(
|
|
struct kobject *kobject,
|
|
const char *buf,
|
|
size_t count)
|
|
{
|
|
struct xfs_error_cfg *cfg = to_error_cfg(kobject);
|
|
int ret;
|
|
int val;
|
|
|
|
ret = kstrtoint(buf, 0, &val);
|
|
if (ret)
|
|
return ret;
|
|
|
|
/* 1 day timeout maximum, -1 means infinite */
|
|
if (val < -1 || val > 86400)
|
|
return -EINVAL;
|
|
|
|
if (val == -1)
|
|
cfg->retry_timeout = XFS_ERR_RETRY_FOREVER;
|
|
else {
|
|
cfg->retry_timeout = msecs_to_jiffies(val * MSEC_PER_SEC);
|
|
ASSERT(msecs_to_jiffies(val * MSEC_PER_SEC) < LONG_MAX);
|
|
}
|
|
return count;
|
|
}
|
|
XFS_SYSFS_ATTR_RW(retry_timeout_seconds);
|
|
|
|
static ssize_t
|
|
fail_at_unmount_show(
|
|
struct kobject *kobject,
|
|
char *buf)
|
|
{
|
|
struct xfs_mount *mp = err_to_mp(kobject);
|
|
|
|
return sysfs_emit(buf, "%d\n", mp->m_fail_unmount);
|
|
}
|
|
|
|
static ssize_t
|
|
fail_at_unmount_store(
|
|
struct kobject *kobject,
|
|
const char *buf,
|
|
size_t count)
|
|
{
|
|
struct xfs_mount *mp = err_to_mp(kobject);
|
|
int ret;
|
|
int val;
|
|
|
|
ret = kstrtoint(buf, 0, &val);
|
|
if (ret)
|
|
return ret;
|
|
|
|
if (val < 0 || val > 1)
|
|
return -EINVAL;
|
|
|
|
mp->m_fail_unmount = val;
|
|
return count;
|
|
}
|
|
XFS_SYSFS_ATTR_RW(fail_at_unmount);
|
|
|
|
static struct attribute *xfs_error_attrs[] = {
|
|
ATTR_LIST(max_retries),
|
|
ATTR_LIST(retry_timeout_seconds),
|
|
NULL,
|
|
};
|
|
ATTRIBUTE_GROUPS(xfs_error);
|
|
|
|
static const struct kobj_type xfs_error_cfg_ktype = {
|
|
.release = xfs_sysfs_release,
|
|
.sysfs_ops = &xfs_sysfs_ops,
|
|
.default_groups = xfs_error_groups,
|
|
};
|
|
|
|
static const struct kobj_type xfs_error_ktype = {
|
|
.release = xfs_sysfs_release,
|
|
.sysfs_ops = &xfs_sysfs_ops,
|
|
};
|
|
|
|
/*
|
|
* Error initialization tables. These need to be ordered in the same
|
|
* order as the enums used to index the array. All class init tables need to
|
|
* define a "default" behaviour as the first entry, all other entries can be
|
|
* empty.
|
|
*/
|
|
struct xfs_error_init {
|
|
char *name;
|
|
int max_retries;
|
|
int retry_timeout; /* in seconds */
|
|
};
|
|
|
|
static const struct xfs_error_init xfs_error_meta_init[XFS_ERR_ERRNO_MAX] = {
|
|
{ .name = "default",
|
|
.max_retries = XFS_ERR_RETRY_FOREVER,
|
|
.retry_timeout = XFS_ERR_RETRY_FOREVER,
|
|
},
|
|
{ .name = "EIO",
|
|
.max_retries = XFS_ERR_RETRY_FOREVER,
|
|
.retry_timeout = XFS_ERR_RETRY_FOREVER,
|
|
},
|
|
{ .name = "ENOSPC",
|
|
.max_retries = XFS_ERR_RETRY_FOREVER,
|
|
.retry_timeout = XFS_ERR_RETRY_FOREVER,
|
|
},
|
|
{ .name = "ENODEV",
|
|
.max_retries = 0, /* We can't recover from devices disappearing */
|
|
.retry_timeout = 0,
|
|
},
|
|
};
|
|
|
|
static int
|
|
xfs_error_sysfs_init_class(
|
|
struct xfs_mount *mp,
|
|
int class,
|
|
const char *parent_name,
|
|
struct xfs_kobj *parent_kobj,
|
|
const struct xfs_error_init init[])
|
|
{
|
|
struct xfs_error_cfg *cfg;
|
|
int error;
|
|
int i;
|
|
|
|
ASSERT(class < XFS_ERR_CLASS_MAX);
|
|
|
|
error = xfs_sysfs_init(parent_kobj, &xfs_error_ktype,
|
|
&mp->m_error_kobj, parent_name);
|
|
if (error)
|
|
return error;
|
|
|
|
for (i = 0; i < XFS_ERR_ERRNO_MAX; i++) {
|
|
cfg = &mp->m_error_cfg[class][i];
|
|
error = xfs_sysfs_init(&cfg->kobj, &xfs_error_cfg_ktype,
|
|
parent_kobj, init[i].name);
|
|
if (error)
|
|
goto out_error;
|
|
|
|
cfg->max_retries = init[i].max_retries;
|
|
if (init[i].retry_timeout == XFS_ERR_RETRY_FOREVER)
|
|
cfg->retry_timeout = XFS_ERR_RETRY_FOREVER;
|
|
else
|
|
cfg->retry_timeout = msecs_to_jiffies(
|
|
init[i].retry_timeout * MSEC_PER_SEC);
|
|
}
|
|
return 0;
|
|
|
|
out_error:
|
|
/* unwind the entries that succeeded */
|
|
for (i--; i >= 0; i--) {
|
|
cfg = &mp->m_error_cfg[class][i];
|
|
xfs_sysfs_del(&cfg->kobj);
|
|
}
|
|
xfs_sysfs_del(parent_kobj);
|
|
return error;
|
|
}
|
|
|
|
int
|
|
xfs_error_sysfs_init(
|
|
struct xfs_mount *mp)
|
|
{
|
|
int error;
|
|
|
|
/* .../xfs/<dev>/error/ */
|
|
error = xfs_sysfs_init(&mp->m_error_kobj, &xfs_error_ktype,
|
|
&mp->m_kobj, "error");
|
|
if (error)
|
|
return error;
|
|
|
|
error = sysfs_create_file(&mp->m_error_kobj.kobject,
|
|
ATTR_LIST(fail_at_unmount));
|
|
|
|
if (error)
|
|
goto out_error;
|
|
|
|
/* .../xfs/<dev>/error/metadata/ */
|
|
error = xfs_error_sysfs_init_class(mp, XFS_ERR_METADATA,
|
|
"metadata", &mp->m_error_meta_kobj,
|
|
xfs_error_meta_init);
|
|
if (error)
|
|
goto out_error;
|
|
|
|
return 0;
|
|
|
|
out_error:
|
|
xfs_sysfs_del(&mp->m_error_kobj);
|
|
return error;
|
|
}
|
|
|
|
void
|
|
xfs_error_sysfs_del(
|
|
struct xfs_mount *mp)
|
|
{
|
|
struct xfs_error_cfg *cfg;
|
|
int i, j;
|
|
|
|
for (i = 0; i < XFS_ERR_CLASS_MAX; i++) {
|
|
for (j = 0; j < XFS_ERR_ERRNO_MAX; j++) {
|
|
cfg = &mp->m_error_cfg[i][j];
|
|
|
|
xfs_sysfs_del(&cfg->kobj);
|
|
}
|
|
}
|
|
xfs_sysfs_del(&mp->m_error_meta_kobj);
|
|
xfs_sysfs_del(&mp->m_error_kobj);
|
|
}
|
|
|
|
struct xfs_error_cfg *
|
|
xfs_error_get_cfg(
|
|
struct xfs_mount *mp,
|
|
int error_class,
|
|
int error)
|
|
{
|
|
struct xfs_error_cfg *cfg;
|
|
|
|
if (error < 0)
|
|
error = -error;
|
|
|
|
switch (error) {
|
|
case EIO:
|
|
cfg = &mp->m_error_cfg[error_class][XFS_ERR_EIO];
|
|
break;
|
|
case ENOSPC:
|
|
cfg = &mp->m_error_cfg[error_class][XFS_ERR_ENOSPC];
|
|
break;
|
|
case ENODEV:
|
|
cfg = &mp->m_error_cfg[error_class][XFS_ERR_ENODEV];
|
|
break;
|
|
default:
|
|
cfg = &mp->m_error_cfg[error_class][XFS_ERR_DEFAULT];
|
|
break;
|
|
}
|
|
|
|
return cfg;
|
|
}
|