2018-01-05 17:47:18 +07:00
|
|
|
// SPDX-License-Identifier: GPL-2.0
|
|
|
|
/*
|
|
|
|
* quota.c - CephFS quota
|
|
|
|
*
|
|
|
|
* Copyright (C) 2017-2018 SUSE
|
|
|
|
*/
|
|
|
|
|
2018-01-31 17:53:13 +07:00
|
|
|
#include <linux/statfs.h>
|
|
|
|
|
2018-01-05 17:47:18 +07:00
|
|
|
#include "super.h"
|
|
|
|
#include "mds_client.h"
|
|
|
|
|
2018-01-13 00:19:29 +07:00
|
|
|
void ceph_adjust_quota_realms_count(struct inode *inode, bool inc)
|
2018-01-05 17:47:20 +07:00
|
|
|
{
|
2020-09-03 20:01:39 +07:00
|
|
|
struct ceph_mds_client *mdsc = ceph_sb_to_mdsc(inode->i_sb);
|
2018-01-13 00:19:29 +07:00
|
|
|
if (inc)
|
|
|
|
atomic64_inc(&mdsc->quotarealms_count);
|
|
|
|
else
|
|
|
|
atomic64_dec(&mdsc->quotarealms_count);
|
|
|
|
}
|
|
|
|
|
|
|
|
static inline bool ceph_has_realms_with_quotas(struct inode *inode)
|
|
|
|
{
|
2020-09-03 20:01:39 +07:00
|
|
|
struct super_block *sb = inode->i_sb;
|
|
|
|
struct ceph_mds_client *mdsc = ceph_sb_to_mdsc(sb);
|
ceph: fix inode number handling on arches with 32-bit ino_t
Tuan and Ulrich mentioned that they were hitting a problem on s390x,
which has a 32-bit ino_t value, even though it's a 64-bit arch (for
historical reasons).
I think the current handling of inode numbers in the ceph driver is
wrong. It tries to use 32-bit inode numbers on 32-bit arches, but that's
actually not a problem. 32-bit arches can deal with 64-bit inode numbers
just fine when userland code is compiled with LFS support (the common
case these days).
What we really want to do is just use 64-bit numbers everywhere, unless
someone has mounted with the ino32 mount option. In that case, we want
to ensure that we hash the inode number down to something that will fit
in 32 bits before presenting the value to userland.
Add new helper functions that do this, and only do the conversion before
presenting these values to userland in getattr and readdir.
The inode table hashvalue is changed to just cast the inode number to
unsigned long, as low-order bits are the most likely to vary anyway.
While it's not strictly required, we do want to put something in
inode->i_ino. Instead of basing it on BITS_PER_LONG, however, base it on
the size of the ino_t type.
NOTE: This is a user-visible change on 32-bit arches:
1/ inode numbers will be seen to have changed between kernel versions.
32-bit arches will see large inode numbers now instead of the hashed
ones they saw before.
2/ any really old software not built with LFS support may start failing
stat() calls with -EOVERFLOW on inode numbers >2^32. Nothing much we
can do about these, but hopefully the intersection of people running
such code on ceph will be very small.
The workaround for both problems is to mount with "-o ino32".
[ idryomov: changelog tweak ]
URL: https://tracker.ceph.com/issues/46828
Reported-by: Ulrich Weigand <Ulrich.Weigand@de.ibm.com>
Reported-and-Tested-by: Tuan Hoang1 <Tuan.Hoang1@ibm.com>
Signed-off-by: Jeff Layton <jlayton@kernel.org>
Reviewed-by: "Yan, Zheng" <zyan@redhat.com>
Signed-off-by: Ilya Dryomov <idryomov@gmail.com>
2020-08-18 19:03:48 +07:00
|
|
|
struct inode *root = d_inode(sb->s_root);
|
2019-03-21 17:20:10 +07:00
|
|
|
|
|
|
|
if (atomic64_read(&mdsc->quotarealms_count) > 0)
|
|
|
|
return true;
|
|
|
|
/* if root is the real CephFS root, we don't have quota realms */
|
ceph: fix inode number handling on arches with 32-bit ino_t
Tuan and Ulrich mentioned that they were hitting a problem on s390x,
which has a 32-bit ino_t value, even though it's a 64-bit arch (for
historical reasons).
I think the current handling of inode numbers in the ceph driver is
wrong. It tries to use 32-bit inode numbers on 32-bit arches, but that's
actually not a problem. 32-bit arches can deal with 64-bit inode numbers
just fine when userland code is compiled with LFS support (the common
case these days).
What we really want to do is just use 64-bit numbers everywhere, unless
someone has mounted with the ino32 mount option. In that case, we want
to ensure that we hash the inode number down to something that will fit
in 32 bits before presenting the value to userland.
Add new helper functions that do this, and only do the conversion before
presenting these values to userland in getattr and readdir.
The inode table hashvalue is changed to just cast the inode number to
unsigned long, as low-order bits are the most likely to vary anyway.
While it's not strictly required, we do want to put something in
inode->i_ino. Instead of basing it on BITS_PER_LONG, however, base it on
the size of the ino_t type.
NOTE: This is a user-visible change on 32-bit arches:
1/ inode numbers will be seen to have changed between kernel versions.
32-bit arches will see large inode numbers now instead of the hashed
ones they saw before.
2/ any really old software not built with LFS support may start failing
stat() calls with -EOVERFLOW on inode numbers >2^32. Nothing much we
can do about these, but hopefully the intersection of people running
such code on ceph will be very small.
The workaround for both problems is to mount with "-o ino32".
[ idryomov: changelog tweak ]
URL: https://tracker.ceph.com/issues/46828
Reported-by: Ulrich Weigand <Ulrich.Weigand@de.ibm.com>
Reported-and-Tested-by: Tuan Hoang1 <Tuan.Hoang1@ibm.com>
Signed-off-by: Jeff Layton <jlayton@kernel.org>
Reviewed-by: "Yan, Zheng" <zyan@redhat.com>
Signed-off-by: Ilya Dryomov <idryomov@gmail.com>
2020-08-18 19:03:48 +07:00
|
|
|
if (root && ceph_ino(root) == CEPH_INO_ROOT)
|
2019-03-21 17:20:10 +07:00
|
|
|
return false;
|
|
|
|
/* otherwise, we can't know for sure */
|
|
|
|
return true;
|
2018-01-05 17:47:20 +07:00
|
|
|
}
|
|
|
|
|
2018-01-05 17:47:18 +07:00
|
|
|
void ceph_handle_quota(struct ceph_mds_client *mdsc,
|
|
|
|
struct ceph_mds_session *session,
|
|
|
|
struct ceph_msg *msg)
|
|
|
|
{
|
|
|
|
struct super_block *sb = mdsc->fsc->sb;
|
|
|
|
struct ceph_mds_quota *h = msg->front.iov_base;
|
|
|
|
struct ceph_vino vino;
|
|
|
|
struct inode *inode;
|
|
|
|
struct ceph_inode_info *ci;
|
|
|
|
|
2018-08-03 15:24:49 +07:00
|
|
|
if (msg->front.iov_len < sizeof(*h)) {
|
2018-01-05 17:47:18 +07:00
|
|
|
pr_err("%s corrupt message mds%d len %d\n", __func__,
|
|
|
|
session->s_mds, (int)msg->front.iov_len);
|
|
|
|
ceph_msg_dump(msg);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* increment msg sequence number */
|
|
|
|
mutex_lock(&session->s_mutex);
|
2020-10-12 20:39:06 +07:00
|
|
|
inc_session_sequence(session);
|
2018-01-05 17:47:18 +07:00
|
|
|
mutex_unlock(&session->s_mutex);
|
|
|
|
|
|
|
|
/* lookup inode */
|
|
|
|
vino.ino = le64_to_cpu(h->ino);
|
|
|
|
vino.snap = CEPH_NOSNAP;
|
|
|
|
inode = ceph_find_inode(sb, vino);
|
|
|
|
if (!inode) {
|
|
|
|
pr_warn("Failed to find inode %llu\n", vino.ino);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
ci = ceph_inode(inode);
|
|
|
|
|
|
|
|
spin_lock(&ci->i_ceph_lock);
|
|
|
|
ci->i_rbytes = le64_to_cpu(h->rbytes);
|
|
|
|
ci->i_rfiles = le64_to_cpu(h->rfiles);
|
|
|
|
ci->i_rsubdirs = le64_to_cpu(h->rsubdirs);
|
2018-01-13 00:19:29 +07:00
|
|
|
__ceph_update_quota(ci, le64_to_cpu(h->max_bytes),
|
|
|
|
le64_to_cpu(h->max_files));
|
2018-01-05 17:47:18 +07:00
|
|
|
spin_unlock(&ci->i_ceph_lock);
|
|
|
|
|
2024-07-05 23:00:04 +07:00
|
|
|
iput(inode);
|
2018-01-05 17:47:18 +07:00
|
|
|
}
|
2018-01-05 17:47:19 +07:00
|
|
|
|
2019-03-21 17:20:10 +07:00
|
|
|
static struct ceph_quotarealm_inode *
|
|
|
|
find_quotarealm_inode(struct ceph_mds_client *mdsc, u64 ino)
|
|
|
|
{
|
|
|
|
struct ceph_quotarealm_inode *qri = NULL;
|
|
|
|
struct rb_node **node, *parent = NULL;
|
|
|
|
|
|
|
|
mutex_lock(&mdsc->quotarealms_inodes_mutex);
|
|
|
|
node = &(mdsc->quotarealms_inodes.rb_node);
|
|
|
|
while (*node) {
|
|
|
|
parent = *node;
|
|
|
|
qri = container_of(*node, struct ceph_quotarealm_inode, node);
|
|
|
|
|
|
|
|
if (ino < qri->ino)
|
|
|
|
node = &((*node)->rb_left);
|
|
|
|
else if (ino > qri->ino)
|
|
|
|
node = &((*node)->rb_right);
|
|
|
|
else
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
if (!qri || (qri->ino != ino)) {
|
|
|
|
/* Not found, create a new one and insert it */
|
|
|
|
qri = kmalloc(sizeof(*qri), GFP_KERNEL);
|
|
|
|
if (qri) {
|
|
|
|
qri->ino = ino;
|
|
|
|
qri->inode = NULL;
|
|
|
|
qri->timeout = 0;
|
|
|
|
mutex_init(&qri->mutex);
|
|
|
|
rb_link_node(&qri->node, parent, node);
|
|
|
|
rb_insert_color(&qri->node, &mdsc->quotarealms_inodes);
|
|
|
|
} else
|
|
|
|
pr_warn("Failed to alloc quotarealms_inode\n");
|
|
|
|
}
|
|
|
|
mutex_unlock(&mdsc->quotarealms_inodes_mutex);
|
|
|
|
|
|
|
|
return qri;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* This function will try to lookup a realm inode which isn't visible in the
|
|
|
|
* filesystem mountpoint. A list of these kind of inodes (not visible) is
|
|
|
|
* maintained in the mdsc and freed only when the filesystem is umounted.
|
|
|
|
*
|
|
|
|
* Note that these inodes are kept in this list even if the lookup fails, which
|
|
|
|
* allows to prevent useless lookup requests.
|
|
|
|
*/
|
|
|
|
static struct inode *lookup_quotarealm_inode(struct ceph_mds_client *mdsc,
|
|
|
|
struct super_block *sb,
|
|
|
|
struct ceph_snap_realm *realm)
|
|
|
|
{
|
|
|
|
struct ceph_quotarealm_inode *qri;
|
|
|
|
struct inode *in;
|
|
|
|
|
|
|
|
qri = find_quotarealm_inode(mdsc, realm->ino);
|
|
|
|
if (!qri)
|
|
|
|
return NULL;
|
|
|
|
|
|
|
|
mutex_lock(&qri->mutex);
|
2019-05-31 19:19:18 +07:00
|
|
|
if (qri->inode && ceph_is_any_caps(qri->inode)) {
|
2019-03-21 17:20:10 +07:00
|
|
|
/* A request has already returned the inode */
|
|
|
|
mutex_unlock(&qri->mutex);
|
|
|
|
return qri->inode;
|
|
|
|
}
|
|
|
|
/* Check if this inode lookup has failed recently */
|
|
|
|
if (qri->timeout &&
|
|
|
|
time_before_eq(jiffies, qri->timeout)) {
|
|
|
|
mutex_unlock(&qri->mutex);
|
|
|
|
return NULL;
|
|
|
|
}
|
2019-05-31 19:19:18 +07:00
|
|
|
if (qri->inode) {
|
|
|
|
/* get caps */
|
|
|
|
int ret = __ceph_do_getattr(qri->inode, NULL,
|
|
|
|
CEPH_STAT_CAP_INODE, true);
|
|
|
|
if (ret >= 0)
|
|
|
|
in = qri->inode;
|
|
|
|
else
|
|
|
|
in = ERR_PTR(ret);
|
|
|
|
} else {
|
|
|
|
in = ceph_lookup_inode(sb, realm->ino);
|
|
|
|
}
|
|
|
|
|
2019-03-21 17:20:10 +07:00
|
|
|
if (IS_ERR(in)) {
|
2020-05-05 19:59:02 +07:00
|
|
|
dout("Can't lookup inode %llx (err: %ld)\n",
|
|
|
|
realm->ino, PTR_ERR(in));
|
2019-03-21 17:20:10 +07:00
|
|
|
qri->timeout = jiffies + msecs_to_jiffies(60 * 1000); /* XXX */
|
|
|
|
} else {
|
|
|
|
qri->timeout = 0;
|
|
|
|
qri->inode = in;
|
|
|
|
}
|
|
|
|
mutex_unlock(&qri->mutex);
|
|
|
|
|
|
|
|
return in;
|
|
|
|
}
|
|
|
|
|
|
|
|
void ceph_cleanup_quotarealms_inodes(struct ceph_mds_client *mdsc)
|
|
|
|
{
|
|
|
|
struct ceph_quotarealm_inode *qri;
|
|
|
|
struct rb_node *node;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* It should now be safe to clean quotarealms_inode tree without holding
|
|
|
|
* mdsc->quotarealms_inodes_mutex...
|
|
|
|
*/
|
|
|
|
mutex_lock(&mdsc->quotarealms_inodes_mutex);
|
|
|
|
while (!RB_EMPTY_ROOT(&mdsc->quotarealms_inodes)) {
|
|
|
|
node = rb_first(&mdsc->quotarealms_inodes);
|
|
|
|
qri = rb_entry(node, struct ceph_quotarealm_inode, node);
|
|
|
|
rb_erase(node, &mdsc->quotarealms_inodes);
|
|
|
|
iput(qri->inode);
|
|
|
|
kfree(qri);
|
|
|
|
}
|
|
|
|
mutex_unlock(&mdsc->quotarealms_inodes_mutex);
|
|
|
|
}
|
|
|
|
|
2018-01-05 17:47:20 +07:00
|
|
|
/*
|
|
|
|
* This function walks through the snaprealm for an inode and returns the
|
|
|
|
* ceph_snap_realm for the first snaprealm that has quotas set (either max_files
|
|
|
|
* or max_bytes). If the root is reached, return the root ceph_snap_realm
|
|
|
|
* instead.
|
|
|
|
*
|
|
|
|
* Note that the caller is responsible for calling ceph_put_snap_realm() on the
|
|
|
|
* returned realm.
|
2019-03-21 17:20:10 +07:00
|
|
|
*
|
|
|
|
* Callers of this function need to hold mdsc->snap_rwsem. However, if there's
|
|
|
|
* a need to do an inode lookup, this rwsem will be temporarily dropped. Hence
|
|
|
|
* the 'retry' argument: if rwsem needs to be dropped and 'retry' is 'false'
|
|
|
|
* this function will return -EAGAIN; otherwise, the snaprealms walk-through
|
|
|
|
* will be restarted.
|
2018-01-05 17:47:20 +07:00
|
|
|
*/
|
|
|
|
static struct ceph_snap_realm *get_quota_realm(struct ceph_mds_client *mdsc,
|
2019-03-21 17:20:10 +07:00
|
|
|
struct inode *inode, bool retry)
|
2018-01-05 17:47:20 +07:00
|
|
|
{
|
|
|
|
struct ceph_inode_info *ci = NULL;
|
|
|
|
struct ceph_snap_realm *realm, *next;
|
|
|
|
struct inode *in;
|
2018-01-12 15:55:31 +07:00
|
|
|
bool has_quota;
|
2018-01-05 17:47:20 +07:00
|
|
|
|
2018-01-12 15:26:17 +07:00
|
|
|
if (ceph_snap(inode) != CEPH_NOSNAP)
|
|
|
|
return NULL;
|
|
|
|
|
2019-03-21 17:20:10 +07:00
|
|
|
restart:
|
2018-01-05 17:47:20 +07:00
|
|
|
realm = ceph_inode(inode)->i_snap_realm;
|
2018-01-12 15:26:17 +07:00
|
|
|
if (realm)
|
|
|
|
ceph_get_snap_realm(mdsc, realm);
|
|
|
|
else
|
|
|
|
pr_err_ratelimited("get_quota_realm: ino (%llx.%llx) "
|
|
|
|
"null i_snap_realm\n", ceph_vinop(inode));
|
2018-01-05 17:47:20 +07:00
|
|
|
while (realm) {
|
2019-03-21 17:20:10 +07:00
|
|
|
bool has_inode;
|
|
|
|
|
2018-01-13 00:19:28 +07:00
|
|
|
spin_lock(&realm->inodes_with_caps_lock);
|
2019-03-21 17:20:10 +07:00
|
|
|
has_inode = realm->inode;
|
|
|
|
in = has_inode ? igrab(realm->inode) : NULL;
|
2018-01-13 00:19:28 +07:00
|
|
|
spin_unlock(&realm->inodes_with_caps_lock);
|
2019-03-21 17:20:10 +07:00
|
|
|
if (has_inode && !in)
|
2018-01-05 17:47:20 +07:00
|
|
|
break;
|
2019-03-21 17:20:10 +07:00
|
|
|
if (!in) {
|
|
|
|
up_read(&mdsc->snap_rwsem);
|
|
|
|
in = lookup_quotarealm_inode(mdsc, inode->i_sb, realm);
|
|
|
|
down_read(&mdsc->snap_rwsem);
|
|
|
|
if (IS_ERR_OR_NULL(in))
|
|
|
|
break;
|
|
|
|
ceph_put_snap_realm(mdsc, realm);
|
|
|
|
if (!retry)
|
|
|
|
return ERR_PTR(-EAGAIN);
|
|
|
|
goto restart;
|
|
|
|
}
|
2018-01-13 00:19:28 +07:00
|
|
|
|
2018-01-05 17:47:20 +07:00
|
|
|
ci = ceph_inode(in);
|
2018-01-13 00:19:29 +07:00
|
|
|
has_quota = __ceph_has_any_quota(ci);
|
2024-07-05 23:00:04 +07:00
|
|
|
iput(in);
|
2018-01-12 15:55:31 +07:00
|
|
|
|
2018-01-05 17:47:20 +07:00
|
|
|
next = realm->parent;
|
2018-01-12 15:55:31 +07:00
|
|
|
if (has_quota || !next)
|
|
|
|
return realm;
|
|
|
|
|
2018-01-05 17:47:20 +07:00
|
|
|
ceph_get_snap_realm(mdsc, next);
|
|
|
|
ceph_put_snap_realm(mdsc, realm);
|
|
|
|
realm = next;
|
|
|
|
}
|
|
|
|
if (realm)
|
|
|
|
ceph_put_snap_realm(mdsc, realm);
|
|
|
|
|
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
2024-07-05 23:00:04 +07:00
|
|
|
bool ceph_quota_is_same_realm(struct inode *old, struct inode *new)
|
2018-01-05 17:47:20 +07:00
|
|
|
{
|
2020-09-03 20:01:39 +07:00
|
|
|
struct ceph_mds_client *mdsc = ceph_sb_to_mdsc(old->i_sb);
|
2018-01-05 17:47:20 +07:00
|
|
|
struct ceph_snap_realm *old_realm, *new_realm;
|
|
|
|
bool is_same;
|
|
|
|
|
2019-03-21 17:20:10 +07:00
|
|
|
restart:
|
|
|
|
/*
|
|
|
|
* We need to lookup 2 quota realms atomically, i.e. with snap_rwsem.
|
|
|
|
* However, get_quota_realm may drop it temporarily. By setting the
|
|
|
|
* 'retry' parameter to 'false', we'll get -EAGAIN if the rwsem was
|
|
|
|
* dropped and we can then restart the whole operation.
|
|
|
|
*/
|
2018-01-05 17:47:20 +07:00
|
|
|
down_read(&mdsc->snap_rwsem);
|
2019-03-21 17:20:10 +07:00
|
|
|
old_realm = get_quota_realm(mdsc, old, true);
|
|
|
|
new_realm = get_quota_realm(mdsc, new, false);
|
|
|
|
if (PTR_ERR(new_realm) == -EAGAIN) {
|
|
|
|
up_read(&mdsc->snap_rwsem);
|
|
|
|
if (old_realm)
|
|
|
|
ceph_put_snap_realm(mdsc, old_realm);
|
|
|
|
goto restart;
|
|
|
|
}
|
2018-01-05 17:47:20 +07:00
|
|
|
is_same = (old_realm == new_realm);
|
|
|
|
up_read(&mdsc->snap_rwsem);
|
|
|
|
|
|
|
|
if (old_realm)
|
|
|
|
ceph_put_snap_realm(mdsc, old_realm);
|
|
|
|
if (new_realm)
|
|
|
|
ceph_put_snap_realm(mdsc, new_realm);
|
|
|
|
|
|
|
|
return is_same;
|
|
|
|
}
|
|
|
|
|
2018-01-05 17:47:19 +07:00
|
|
|
enum quota_check_op {
|
2018-01-05 17:47:21 +07:00
|
|
|
QUOTA_CHECK_MAX_FILES_OP, /* check quota max_files limit */
|
2018-01-05 17:47:22 +07:00
|
|
|
QUOTA_CHECK_MAX_BYTES_OP, /* check quota max_files limit */
|
|
|
|
QUOTA_CHECK_MAX_BYTES_APPROACHING_OP /* check if quota max_files
|
|
|
|
limit is approaching */
|
2018-01-05 17:47:19 +07:00
|
|
|
};
|
|
|
|
|
|
|
|
/*
|
|
|
|
* check_quota_exceeded() will walk up the snaprealm hierarchy and, for each
|
|
|
|
* realm, it will execute quota check operation defined by the 'op' parameter.
|
|
|
|
* The snaprealm walk is interrupted if the quota check detects that the quota
|
|
|
|
* is exceeded or if the root inode is reached.
|
|
|
|
*/
|
|
|
|
static bool check_quota_exceeded(struct inode *inode, enum quota_check_op op,
|
|
|
|
loff_t delta)
|
|
|
|
{
|
2020-09-03 20:01:39 +07:00
|
|
|
struct ceph_mds_client *mdsc = ceph_sb_to_mdsc(inode->i_sb);
|
2018-01-05 17:47:19 +07:00
|
|
|
struct ceph_inode_info *ci;
|
|
|
|
struct ceph_snap_realm *realm, *next;
|
|
|
|
struct inode *in;
|
|
|
|
u64 max, rvalue;
|
|
|
|
bool exceeded = false;
|
|
|
|
|
2018-01-12 15:26:17 +07:00
|
|
|
if (ceph_snap(inode) != CEPH_NOSNAP)
|
|
|
|
return false;
|
|
|
|
|
2018-01-05 17:47:19 +07:00
|
|
|
down_read(&mdsc->snap_rwsem);
|
2019-03-21 17:20:10 +07:00
|
|
|
restart:
|
2018-01-05 17:47:19 +07:00
|
|
|
realm = ceph_inode(inode)->i_snap_realm;
|
2018-01-12 15:26:17 +07:00
|
|
|
if (realm)
|
|
|
|
ceph_get_snap_realm(mdsc, realm);
|
|
|
|
else
|
|
|
|
pr_err_ratelimited("check_quota_exceeded: ino (%llx.%llx) "
|
|
|
|
"null i_snap_realm\n", ceph_vinop(inode));
|
2018-01-05 17:47:19 +07:00
|
|
|
while (realm) {
|
2019-03-21 17:20:10 +07:00
|
|
|
bool has_inode;
|
|
|
|
|
2018-01-13 00:19:28 +07:00
|
|
|
spin_lock(&realm->inodes_with_caps_lock);
|
2019-03-21 17:20:10 +07:00
|
|
|
has_inode = realm->inode;
|
|
|
|
in = has_inode ? igrab(realm->inode) : NULL;
|
2018-01-13 00:19:28 +07:00
|
|
|
spin_unlock(&realm->inodes_with_caps_lock);
|
2019-03-21 17:20:10 +07:00
|
|
|
if (has_inode && !in)
|
2018-01-05 17:47:19 +07:00
|
|
|
break;
|
2019-03-21 17:20:10 +07:00
|
|
|
if (!in) {
|
|
|
|
up_read(&mdsc->snap_rwsem);
|
|
|
|
in = lookup_quotarealm_inode(mdsc, inode->i_sb, realm);
|
|
|
|
down_read(&mdsc->snap_rwsem);
|
|
|
|
if (IS_ERR_OR_NULL(in))
|
|
|
|
break;
|
|
|
|
ceph_put_snap_realm(mdsc, realm);
|
|
|
|
goto restart;
|
|
|
|
}
|
2018-01-05 17:47:19 +07:00
|
|
|
ci = ceph_inode(in);
|
|
|
|
spin_lock(&ci->i_ceph_lock);
|
|
|
|
if (op == QUOTA_CHECK_MAX_FILES_OP) {
|
|
|
|
max = ci->i_max_files;
|
|
|
|
rvalue = ci->i_rfiles + ci->i_rsubdirs;
|
2018-01-05 17:47:21 +07:00
|
|
|
} else {
|
|
|
|
max = ci->i_max_bytes;
|
|
|
|
rvalue = ci->i_rbytes;
|
2018-01-05 17:47:19 +07:00
|
|
|
}
|
|
|
|
spin_unlock(&ci->i_ceph_lock);
|
|
|
|
switch (op) {
|
|
|
|
case QUOTA_CHECK_MAX_FILES_OP:
|
2018-01-05 17:47:21 +07:00
|
|
|
case QUOTA_CHECK_MAX_BYTES_OP:
|
|
|
|
exceeded = (max && (rvalue + delta > max));
|
|
|
|
break;
|
2018-01-05 17:47:22 +07:00
|
|
|
case QUOTA_CHECK_MAX_BYTES_APPROACHING_OP:
|
|
|
|
if (max) {
|
|
|
|
if (rvalue >= max)
|
|
|
|
exceeded = true;
|
|
|
|
else {
|
|
|
|
/*
|
|
|
|
* when we're writing more that 1/16th
|
|
|
|
* of the available space
|
|
|
|
*/
|
|
|
|
exceeded =
|
|
|
|
(((max - rvalue) >> 4) < delta);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
break;
|
2018-01-05 17:47:19 +07:00
|
|
|
default:
|
|
|
|
/* Shouldn't happen */
|
|
|
|
pr_warn("Invalid quota check op (%d)\n", op);
|
|
|
|
exceeded = true; /* Just break the loop */
|
|
|
|
}
|
2024-07-05 23:00:04 +07:00
|
|
|
iput(in);
|
2018-01-05 17:47:19 +07:00
|
|
|
|
|
|
|
next = realm->parent;
|
2018-01-12 15:55:31 +07:00
|
|
|
if (exceeded || !next)
|
|
|
|
break;
|
2018-01-05 17:47:19 +07:00
|
|
|
ceph_get_snap_realm(mdsc, next);
|
|
|
|
ceph_put_snap_realm(mdsc, realm);
|
|
|
|
realm = next;
|
|
|
|
}
|
2018-11-06 02:00:52 +07:00
|
|
|
if (realm)
|
|
|
|
ceph_put_snap_realm(mdsc, realm);
|
2018-01-05 17:47:19 +07:00
|
|
|
up_read(&mdsc->snap_rwsem);
|
|
|
|
|
|
|
|
return exceeded;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* ceph_quota_is_max_files_exceeded - check if we can create a new file
|
|
|
|
* @inode: directory where a new file is being created
|
|
|
|
*
|
|
|
|
* This functions returns true is max_files quota allows a new file to be
|
|
|
|
* created. It is necessary to walk through the snaprealm hierarchy (until the
|
|
|
|
* FS root) to check all realms with quotas set.
|
|
|
|
*/
|
|
|
|
bool ceph_quota_is_max_files_exceeded(struct inode *inode)
|
|
|
|
{
|
2018-01-13 00:19:29 +07:00
|
|
|
if (!ceph_has_realms_with_quotas(inode))
|
|
|
|
return false;
|
|
|
|
|
2018-01-05 17:47:19 +07:00
|
|
|
WARN_ON(!S_ISDIR(inode->i_mode));
|
|
|
|
|
2020-04-07 17:30:19 +07:00
|
|
|
return check_quota_exceeded(inode, QUOTA_CHECK_MAX_FILES_OP, 1);
|
2018-01-05 17:47:19 +07:00
|
|
|
}
|
2018-01-05 17:47:21 +07:00
|
|
|
|
|
|
|
/*
|
|
|
|
* ceph_quota_is_max_bytes_exceeded - check if we can write to a file
|
|
|
|
* @inode: inode being written
|
|
|
|
* @newsize: new size if write succeeds
|
|
|
|
*
|
|
|
|
* This functions returns true is max_bytes quota allows a file size to reach
|
|
|
|
* @newsize; it returns false otherwise.
|
|
|
|
*/
|
|
|
|
bool ceph_quota_is_max_bytes_exceeded(struct inode *inode, loff_t newsize)
|
|
|
|
{
|
|
|
|
loff_t size = i_size_read(inode);
|
|
|
|
|
2018-01-13 00:19:29 +07:00
|
|
|
if (!ceph_has_realms_with_quotas(inode))
|
|
|
|
return false;
|
|
|
|
|
2018-01-05 17:47:21 +07:00
|
|
|
/* return immediately if we're decreasing file size */
|
|
|
|
if (newsize <= size)
|
|
|
|
return false;
|
|
|
|
|
|
|
|
return check_quota_exceeded(inode, QUOTA_CHECK_MAX_BYTES_OP, (newsize - size));
|
|
|
|
}
|
2018-01-05 17:47:22 +07:00
|
|
|
|
|
|
|
/*
|
|
|
|
* ceph_quota_is_max_bytes_approaching - check if we're reaching max_bytes
|
|
|
|
* @inode: inode being written
|
|
|
|
* @newsize: new size if write succeeds
|
|
|
|
*
|
|
|
|
* This function returns true if the new file size @newsize will be consuming
|
|
|
|
* more than 1/16th of the available quota space; it returns false otherwise.
|
|
|
|
*/
|
|
|
|
bool ceph_quota_is_max_bytes_approaching(struct inode *inode, loff_t newsize)
|
|
|
|
{
|
|
|
|
loff_t size = ceph_inode(inode)->i_reported_size;
|
|
|
|
|
2018-01-13 00:19:29 +07:00
|
|
|
if (!ceph_has_realms_with_quotas(inode))
|
|
|
|
return false;
|
|
|
|
|
2018-01-05 17:47:22 +07:00
|
|
|
/* return immediately if we're decreasing file size */
|
|
|
|
if (newsize <= size)
|
|
|
|
return false;
|
|
|
|
|
|
|
|
return check_quota_exceeded(inode, QUOTA_CHECK_MAX_BYTES_APPROACHING_OP,
|
|
|
|
(newsize - size));
|
|
|
|
}
|
2018-01-31 17:53:13 +07:00
|
|
|
|
|
|
|
/*
|
|
|
|
* ceph_quota_update_statfs - if root has quota update statfs with quota status
|
|
|
|
* @fsc: filesystem client instance
|
|
|
|
* @buf: statfs to update
|
|
|
|
*
|
|
|
|
* If the mounted filesystem root has max_bytes quota set, update the filesystem
|
|
|
|
* statistics with the quota status.
|
|
|
|
*
|
|
|
|
* This function returns true if the stats have been updated, false otherwise.
|
|
|
|
*/
|
|
|
|
bool ceph_quota_update_statfs(struct ceph_fs_client *fsc, struct kstatfs *buf)
|
|
|
|
{
|
|
|
|
struct ceph_mds_client *mdsc = fsc->mdsc;
|
|
|
|
struct ceph_inode_info *ci;
|
|
|
|
struct ceph_snap_realm *realm;
|
|
|
|
struct inode *in;
|
|
|
|
u64 total = 0, used, free;
|
|
|
|
bool is_updated = false;
|
|
|
|
|
|
|
|
down_read(&mdsc->snap_rwsem);
|
2019-03-21 17:20:10 +07:00
|
|
|
realm = get_quota_realm(mdsc, d_inode(fsc->sb->s_root), true);
|
2018-01-31 17:53:13 +07:00
|
|
|
up_read(&mdsc->snap_rwsem);
|
|
|
|
if (!realm)
|
|
|
|
return false;
|
|
|
|
|
|
|
|
spin_lock(&realm->inodes_with_caps_lock);
|
|
|
|
in = realm->inode ? igrab(realm->inode) : NULL;
|
|
|
|
spin_unlock(&realm->inodes_with_caps_lock);
|
|
|
|
if (in) {
|
|
|
|
ci = ceph_inode(in);
|
|
|
|
spin_lock(&ci->i_ceph_lock);
|
|
|
|
if (ci->i_max_bytes) {
|
|
|
|
total = ci->i_max_bytes >> CEPH_BLOCK_SHIFT;
|
|
|
|
used = ci->i_rbytes >> CEPH_BLOCK_SHIFT;
|
|
|
|
/* It is possible for a quota to be exceeded.
|
|
|
|
* Report 'zero' in that case
|
|
|
|
*/
|
|
|
|
free = total > used ? total - used : 0;
|
|
|
|
}
|
|
|
|
spin_unlock(&ci->i_ceph_lock);
|
|
|
|
if (total) {
|
|
|
|
buf->f_blocks = total;
|
|
|
|
buf->f_bfree = free;
|
|
|
|
buf->f_bavail = free;
|
|
|
|
is_updated = true;
|
|
|
|
}
|
|
|
|
iput(in);
|
|
|
|
}
|
|
|
|
ceph_put_snap_realm(mdsc, realm);
|
|
|
|
|
|
|
|
return is_updated;
|
|
|
|
}
|
|
|
|
|