mirror of
https://github.com/AuxXxilium/linux_dsm_epyc7002.git
synced 2024-12-23 00:47:11 +07:00
0384892c2d
clang complains about assigning a variable to itself during the declaration: fs/ceph/ioctl.c:187:26: error: variable 'oid' is uninitialized when used within its own initialization [-Werror,-Wuninitialized] CEPH_DEFINE_OID_ONSTACK(oid); ^~~ include/linux/ceph/osdmap.h:122:52: note: expanded from macro 'CEPH_DEFINE_OID_ONSTACK' struct ceph_object_id oid = CEPH_OID_INIT_ONSTACK(oid) ~~~ ^~~ include/linux/ceph/osdmap.h:120:29: note: expanded from macro 'CEPH_OID_INIT_ONSTACK' ({ ceph_oid_init(&oid); oid; }) ^~~ We use this trick in other places, but it is completely unnecessary here, as we can just use a regular struct initializer. Signed-off-by: Arnd Bergmann <arnd@arndb.de> Reviewed-by: Ilya Dryomov <idryomov@gmail.com> Signed-off-by: Ilya Dryomov <idryomov@gmail.com>
309 lines
8.1 KiB
C
309 lines
8.1 KiB
C
/* SPDX-License-Identifier: GPL-2.0 */
|
|
#ifndef _FS_CEPH_OSDMAP_H
|
|
#define _FS_CEPH_OSDMAP_H
|
|
|
|
#include <linux/rbtree.h>
|
|
#include <linux/ceph/types.h>
|
|
#include <linux/ceph/decode.h>
|
|
#include <linux/crush/crush.h>
|
|
|
|
/*
|
|
* The osd map describes the current membership of the osd cluster and
|
|
* specifies the mapping of objects to placement groups and placement
|
|
* groups to (sets of) osds. That is, it completely specifies the
|
|
* (desired) distribution of all data objects in the system at some
|
|
* point in time.
|
|
*
|
|
* Each map version is identified by an epoch, which increases monotonically.
|
|
*
|
|
* The map can be updated either via an incremental map (diff) describing
|
|
* the change between two successive epochs, or as a fully encoded map.
|
|
*/
|
|
struct ceph_pg {
|
|
uint64_t pool;
|
|
uint32_t seed;
|
|
};
|
|
|
|
#define CEPH_SPG_NOSHARD -1
|
|
|
|
struct ceph_spg {
|
|
struct ceph_pg pgid;
|
|
s8 shard;
|
|
};
|
|
|
|
int ceph_pg_compare(const struct ceph_pg *lhs, const struct ceph_pg *rhs);
|
|
int ceph_spg_compare(const struct ceph_spg *lhs, const struct ceph_spg *rhs);
|
|
|
|
#define CEPH_POOL_FLAG_HASHPSPOOL (1ULL << 0) /* hash pg seed and pool id
|
|
together */
|
|
#define CEPH_POOL_FLAG_FULL (1ULL << 1) /* pool is full */
|
|
|
|
struct ceph_pg_pool_info {
|
|
struct rb_node node;
|
|
s64 id;
|
|
u8 type; /* CEPH_POOL_TYPE_* */
|
|
u8 size;
|
|
u8 min_size;
|
|
u8 crush_ruleset;
|
|
u8 object_hash;
|
|
u32 last_force_request_resend;
|
|
u32 pg_num, pgp_num;
|
|
int pg_num_mask, pgp_num_mask;
|
|
s64 read_tier;
|
|
s64 write_tier; /* wins for read+write ops */
|
|
u64 flags; /* CEPH_POOL_FLAG_* */
|
|
char *name;
|
|
|
|
bool was_full; /* for handle_one_map() */
|
|
};
|
|
|
|
static inline bool ceph_can_shift_osds(struct ceph_pg_pool_info *pool)
|
|
{
|
|
switch (pool->type) {
|
|
case CEPH_POOL_TYPE_REP:
|
|
return true;
|
|
case CEPH_POOL_TYPE_EC:
|
|
return false;
|
|
default:
|
|
BUG();
|
|
}
|
|
}
|
|
|
|
struct ceph_object_locator {
|
|
s64 pool;
|
|
struct ceph_string *pool_ns;
|
|
};
|
|
|
|
static inline void ceph_oloc_init(struct ceph_object_locator *oloc)
|
|
{
|
|
oloc->pool = -1;
|
|
oloc->pool_ns = NULL;
|
|
}
|
|
|
|
static inline bool ceph_oloc_empty(const struct ceph_object_locator *oloc)
|
|
{
|
|
return oloc->pool == -1;
|
|
}
|
|
|
|
void ceph_oloc_copy(struct ceph_object_locator *dest,
|
|
const struct ceph_object_locator *src);
|
|
void ceph_oloc_destroy(struct ceph_object_locator *oloc);
|
|
|
|
/*
|
|
* 51-char inline_name is long enough for all cephfs and all but one
|
|
* rbd requests: <imgname> in "<imgname>.rbd"/"rbd_id.<imgname>" can be
|
|
* arbitrarily long (~PAGE_SIZE). It's done once during rbd map; all
|
|
* other rbd requests fit into inline_name.
|
|
*
|
|
* Makes ceph_object_id 64 bytes on 64-bit.
|
|
*/
|
|
#define CEPH_OID_INLINE_LEN 52
|
|
|
|
/*
|
|
* Both inline and external buffers have space for a NUL-terminator,
|
|
* which is carried around. It's not required though - RADOS object
|
|
* names don't have to be NUL-terminated and may contain NULs.
|
|
*/
|
|
struct ceph_object_id {
|
|
char *name;
|
|
char inline_name[CEPH_OID_INLINE_LEN];
|
|
int name_len;
|
|
};
|
|
|
|
#define __CEPH_OID_INITIALIZER(oid) { .name = (oid).inline_name }
|
|
|
|
#define CEPH_DEFINE_OID_ONSTACK(oid) \
|
|
struct ceph_object_id oid = __CEPH_OID_INITIALIZER(oid)
|
|
|
|
static inline void ceph_oid_init(struct ceph_object_id *oid)
|
|
{
|
|
*oid = (struct ceph_object_id) __CEPH_OID_INITIALIZER(*oid);
|
|
}
|
|
|
|
static inline bool ceph_oid_empty(const struct ceph_object_id *oid)
|
|
{
|
|
return oid->name == oid->inline_name && !oid->name_len;
|
|
}
|
|
|
|
void ceph_oid_copy(struct ceph_object_id *dest,
|
|
const struct ceph_object_id *src);
|
|
__printf(2, 3)
|
|
void ceph_oid_printf(struct ceph_object_id *oid, const char *fmt, ...);
|
|
__printf(3, 4)
|
|
int ceph_oid_aprintf(struct ceph_object_id *oid, gfp_t gfp,
|
|
const char *fmt, ...);
|
|
void ceph_oid_destroy(struct ceph_object_id *oid);
|
|
|
|
struct ceph_pg_mapping {
|
|
struct rb_node node;
|
|
struct ceph_pg pgid;
|
|
|
|
union {
|
|
struct {
|
|
int len;
|
|
int osds[];
|
|
} pg_temp, pg_upmap;
|
|
struct {
|
|
int osd;
|
|
} primary_temp;
|
|
struct {
|
|
int len;
|
|
int from_to[][2];
|
|
} pg_upmap_items;
|
|
};
|
|
};
|
|
|
|
struct ceph_osdmap {
|
|
struct ceph_fsid fsid;
|
|
u32 epoch;
|
|
struct ceph_timespec created, modified;
|
|
|
|
u32 flags; /* CEPH_OSDMAP_* */
|
|
|
|
u32 max_osd; /* size of osd_state, _offload, _addr arrays */
|
|
u32 *osd_state; /* CEPH_OSD_* */
|
|
u32 *osd_weight; /* 0 = failed, 0x10000 = 100% normal */
|
|
struct ceph_entity_addr *osd_addr;
|
|
|
|
struct rb_root pg_temp;
|
|
struct rb_root primary_temp;
|
|
|
|
/* remap (post-CRUSH, pre-up) */
|
|
struct rb_root pg_upmap; /* PG := raw set */
|
|
struct rb_root pg_upmap_items; /* from -> to within raw set */
|
|
|
|
u32 *osd_primary_affinity;
|
|
|
|
struct rb_root pg_pools;
|
|
u32 pool_max;
|
|
|
|
/* the CRUSH map specifies the mapping of placement groups to
|
|
* the list of osds that store+replicate them. */
|
|
struct crush_map *crush;
|
|
|
|
struct mutex crush_workspace_mutex;
|
|
void *crush_workspace;
|
|
};
|
|
|
|
static inline bool ceph_osd_exists(struct ceph_osdmap *map, int osd)
|
|
{
|
|
return osd >= 0 && osd < map->max_osd &&
|
|
(map->osd_state[osd] & CEPH_OSD_EXISTS);
|
|
}
|
|
|
|
static inline bool ceph_osd_is_up(struct ceph_osdmap *map, int osd)
|
|
{
|
|
return ceph_osd_exists(map, osd) &&
|
|
(map->osd_state[osd] & CEPH_OSD_UP);
|
|
}
|
|
|
|
static inline bool ceph_osd_is_down(struct ceph_osdmap *map, int osd)
|
|
{
|
|
return !ceph_osd_is_up(map, osd);
|
|
}
|
|
|
|
char *ceph_osdmap_state_str(char *str, int len, u32 state);
|
|
extern u32 ceph_get_primary_affinity(struct ceph_osdmap *map, int osd);
|
|
|
|
static inline struct ceph_entity_addr *ceph_osd_addr(struct ceph_osdmap *map,
|
|
int osd)
|
|
{
|
|
if (osd >= map->max_osd)
|
|
return NULL;
|
|
return &map->osd_addr[osd];
|
|
}
|
|
|
|
#define CEPH_PGID_ENCODING_LEN (1 + 8 + 4 + 4)
|
|
|
|
static inline int ceph_decode_pgid(void **p, void *end, struct ceph_pg *pgid)
|
|
{
|
|
__u8 version;
|
|
|
|
if (!ceph_has_room(p, end, CEPH_PGID_ENCODING_LEN)) {
|
|
pr_warn("incomplete pg encoding\n");
|
|
return -EINVAL;
|
|
}
|
|
version = ceph_decode_8(p);
|
|
if (version > 1) {
|
|
pr_warn("do not understand pg encoding %d > 1\n",
|
|
(int)version);
|
|
return -EINVAL;
|
|
}
|
|
|
|
pgid->pool = ceph_decode_64(p);
|
|
pgid->seed = ceph_decode_32(p);
|
|
*p += 4; /* skip deprecated preferred value */
|
|
|
|
return 0;
|
|
}
|
|
|
|
struct ceph_osdmap *ceph_osdmap_alloc(void);
|
|
extern struct ceph_osdmap *ceph_osdmap_decode(void **p, void *end);
|
|
struct ceph_osdmap *osdmap_apply_incremental(void **p, void *end,
|
|
struct ceph_osdmap *map);
|
|
extern void ceph_osdmap_destroy(struct ceph_osdmap *map);
|
|
|
|
struct ceph_osds {
|
|
int osds[CEPH_PG_MAX_SIZE];
|
|
int size;
|
|
int primary; /* id, NOT index */
|
|
};
|
|
|
|
static inline void ceph_osds_init(struct ceph_osds *set)
|
|
{
|
|
set->size = 0;
|
|
set->primary = -1;
|
|
}
|
|
|
|
void ceph_osds_copy(struct ceph_osds *dest, const struct ceph_osds *src);
|
|
|
|
bool ceph_pg_is_split(const struct ceph_pg *pgid, u32 old_pg_num,
|
|
u32 new_pg_num);
|
|
bool ceph_is_new_interval(const struct ceph_osds *old_acting,
|
|
const struct ceph_osds *new_acting,
|
|
const struct ceph_osds *old_up,
|
|
const struct ceph_osds *new_up,
|
|
int old_size,
|
|
int new_size,
|
|
int old_min_size,
|
|
int new_min_size,
|
|
u32 old_pg_num,
|
|
u32 new_pg_num,
|
|
bool old_sort_bitwise,
|
|
bool new_sort_bitwise,
|
|
bool old_recovery_deletes,
|
|
bool new_recovery_deletes,
|
|
const struct ceph_pg *pgid);
|
|
bool ceph_osds_changed(const struct ceph_osds *old_acting,
|
|
const struct ceph_osds *new_acting,
|
|
bool any_change);
|
|
|
|
void __ceph_object_locator_to_pg(struct ceph_pg_pool_info *pi,
|
|
const struct ceph_object_id *oid,
|
|
const struct ceph_object_locator *oloc,
|
|
struct ceph_pg *raw_pgid);
|
|
int ceph_object_locator_to_pg(struct ceph_osdmap *osdmap,
|
|
const struct ceph_object_id *oid,
|
|
const struct ceph_object_locator *oloc,
|
|
struct ceph_pg *raw_pgid);
|
|
|
|
void ceph_pg_to_up_acting_osds(struct ceph_osdmap *osdmap,
|
|
struct ceph_pg_pool_info *pi,
|
|
const struct ceph_pg *raw_pgid,
|
|
struct ceph_osds *up,
|
|
struct ceph_osds *acting);
|
|
bool ceph_pg_to_primary_shard(struct ceph_osdmap *osdmap,
|
|
struct ceph_pg_pool_info *pi,
|
|
const struct ceph_pg *raw_pgid,
|
|
struct ceph_spg *spgid);
|
|
int ceph_pg_to_acting_primary(struct ceph_osdmap *osdmap,
|
|
const struct ceph_pg *raw_pgid);
|
|
|
|
extern struct ceph_pg_pool_info *ceph_pg_pool_by_id(struct ceph_osdmap *map,
|
|
u64 id);
|
|
|
|
extern const char *ceph_pg_pool_name_by_id(struct ceph_osdmap *map, u64 id);
|
|
extern int ceph_pg_poolid_by_name(struct ceph_osdmap *map, const char *name);
|
|
|
|
#endif
|