mirror of
https://github.com/AuxXxilium/linux_dsm_epyc7002.git
synced 2024-12-24 01:15:59 +07:00
2522fe45a1
Based on 1 normalized pattern(s): this copyrighted material is made available to anyone wishing to use modify copy or redistribute it subject to the terms and conditions of the gnu general public license v 2 extracted by the scancode license scanner the SPDX license identifier GPL-2.0-only has been chosen to replace the boilerplate/reference in 45 file(s). Signed-off-by: Thomas Gleixner <tglx@linutronix.de> Reviewed-by: Richard Fontana <rfontana@redhat.com> Reviewed-by: Allison Randal <allison@lohutok.net> Reviewed-by: Steve Winslow <swinslow@gmail.com> Reviewed-by: Alexios Zavras <alexios.zavras@intel.com> Cc: linux-spdx@vger.kernel.org Link: https://lkml.kernel.org/r/20190528170027.342746075@linutronix.de Signed-off-by: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
719 lines
16 KiB
C
719 lines
16 KiB
C
// SPDX-License-Identifier: GPL-2.0-only
|
|
/******************************************************************************
|
|
*******************************************************************************
|
|
**
|
|
** Copyright (C) 2005-2011 Red Hat, Inc. All rights reserved.
|
|
**
|
|
**
|
|
*******************************************************************************
|
|
******************************************************************************/
|
|
|
|
#include "dlm_internal.h"
|
|
#include "lockspace.h"
|
|
#include "member.h"
|
|
#include "recoverd.h"
|
|
#include "recover.h"
|
|
#include "rcom.h"
|
|
#include "config.h"
|
|
#include "lowcomms.h"
|
|
|
|
int dlm_slots_version(struct dlm_header *h)
|
|
{
|
|
if ((h->h_version & 0x0000FFFF) < DLM_HEADER_SLOTS)
|
|
return 0;
|
|
return 1;
|
|
}
|
|
|
|
void dlm_slot_save(struct dlm_ls *ls, struct dlm_rcom *rc,
|
|
struct dlm_member *memb)
|
|
{
|
|
struct rcom_config *rf = (struct rcom_config *)rc->rc_buf;
|
|
|
|
if (!dlm_slots_version(&rc->rc_header))
|
|
return;
|
|
|
|
memb->slot = le16_to_cpu(rf->rf_our_slot);
|
|
memb->generation = le32_to_cpu(rf->rf_generation);
|
|
}
|
|
|
|
void dlm_slots_copy_out(struct dlm_ls *ls, struct dlm_rcom *rc)
|
|
{
|
|
struct dlm_slot *slot;
|
|
struct rcom_slot *ro;
|
|
int i;
|
|
|
|
ro = (struct rcom_slot *)(rc->rc_buf + sizeof(struct rcom_config));
|
|
|
|
/* ls_slots array is sparse, but not rcom_slots */
|
|
|
|
for (i = 0; i < ls->ls_slots_size; i++) {
|
|
slot = &ls->ls_slots[i];
|
|
if (!slot->nodeid)
|
|
continue;
|
|
ro->ro_nodeid = cpu_to_le32(slot->nodeid);
|
|
ro->ro_slot = cpu_to_le16(slot->slot);
|
|
ro++;
|
|
}
|
|
}
|
|
|
|
#define SLOT_DEBUG_LINE 128
|
|
|
|
static void log_slots(struct dlm_ls *ls, uint32_t gen, int num_slots,
|
|
struct rcom_slot *ro0, struct dlm_slot *array,
|
|
int array_size)
|
|
{
|
|
char line[SLOT_DEBUG_LINE];
|
|
int len = SLOT_DEBUG_LINE - 1;
|
|
int pos = 0;
|
|
int ret, i;
|
|
|
|
memset(line, 0, sizeof(line));
|
|
|
|
if (array) {
|
|
for (i = 0; i < array_size; i++) {
|
|
if (!array[i].nodeid)
|
|
continue;
|
|
|
|
ret = snprintf(line + pos, len - pos, " %d:%d",
|
|
array[i].slot, array[i].nodeid);
|
|
if (ret >= len - pos)
|
|
break;
|
|
pos += ret;
|
|
}
|
|
} else if (ro0) {
|
|
for (i = 0; i < num_slots; i++) {
|
|
ret = snprintf(line + pos, len - pos, " %d:%d",
|
|
ro0[i].ro_slot, ro0[i].ro_nodeid);
|
|
if (ret >= len - pos)
|
|
break;
|
|
pos += ret;
|
|
}
|
|
}
|
|
|
|
log_rinfo(ls, "generation %u slots %d%s", gen, num_slots, line);
|
|
}
|
|
|
|
int dlm_slots_copy_in(struct dlm_ls *ls)
|
|
{
|
|
struct dlm_member *memb;
|
|
struct dlm_rcom *rc = ls->ls_recover_buf;
|
|
struct rcom_config *rf = (struct rcom_config *)rc->rc_buf;
|
|
struct rcom_slot *ro0, *ro;
|
|
int our_nodeid = dlm_our_nodeid();
|
|
int i, num_slots;
|
|
uint32_t gen;
|
|
|
|
if (!dlm_slots_version(&rc->rc_header))
|
|
return -1;
|
|
|
|
gen = le32_to_cpu(rf->rf_generation);
|
|
if (gen <= ls->ls_generation) {
|
|
log_error(ls, "dlm_slots_copy_in gen %u old %u",
|
|
gen, ls->ls_generation);
|
|
}
|
|
ls->ls_generation = gen;
|
|
|
|
num_slots = le16_to_cpu(rf->rf_num_slots);
|
|
if (!num_slots)
|
|
return -1;
|
|
|
|
ro0 = (struct rcom_slot *)(rc->rc_buf + sizeof(struct rcom_config));
|
|
|
|
for (i = 0, ro = ro0; i < num_slots; i++, ro++) {
|
|
ro->ro_nodeid = le32_to_cpu(ro->ro_nodeid);
|
|
ro->ro_slot = le16_to_cpu(ro->ro_slot);
|
|
}
|
|
|
|
log_slots(ls, gen, num_slots, ro0, NULL, 0);
|
|
|
|
list_for_each_entry(memb, &ls->ls_nodes, list) {
|
|
for (i = 0, ro = ro0; i < num_slots; i++, ro++) {
|
|
if (ro->ro_nodeid != memb->nodeid)
|
|
continue;
|
|
memb->slot = ro->ro_slot;
|
|
memb->slot_prev = memb->slot;
|
|
break;
|
|
}
|
|
|
|
if (memb->nodeid == our_nodeid) {
|
|
if (ls->ls_slot && ls->ls_slot != memb->slot) {
|
|
log_error(ls, "dlm_slots_copy_in our slot "
|
|
"changed %d %d", ls->ls_slot,
|
|
memb->slot);
|
|
return -1;
|
|
}
|
|
|
|
if (!ls->ls_slot)
|
|
ls->ls_slot = memb->slot;
|
|
}
|
|
|
|
if (!memb->slot) {
|
|
log_error(ls, "dlm_slots_copy_in nodeid %d no slot",
|
|
memb->nodeid);
|
|
return -1;
|
|
}
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
/* for any nodes that do not support slots, we will not have set memb->slot
|
|
in wait_status_all(), so memb->slot will remain -1, and we will not
|
|
assign slots or set ls_num_slots here */
|
|
|
|
int dlm_slots_assign(struct dlm_ls *ls, int *num_slots, int *slots_size,
|
|
struct dlm_slot **slots_out, uint32_t *gen_out)
|
|
{
|
|
struct dlm_member *memb;
|
|
struct dlm_slot *array;
|
|
int our_nodeid = dlm_our_nodeid();
|
|
int array_size, max_slots, i;
|
|
int need = 0;
|
|
int max = 0;
|
|
int num = 0;
|
|
uint32_t gen = 0;
|
|
|
|
/* our own memb struct will have slot -1 gen 0 */
|
|
|
|
list_for_each_entry(memb, &ls->ls_nodes, list) {
|
|
if (memb->nodeid == our_nodeid) {
|
|
memb->slot = ls->ls_slot;
|
|
memb->generation = ls->ls_generation;
|
|
break;
|
|
}
|
|
}
|
|
|
|
list_for_each_entry(memb, &ls->ls_nodes, list) {
|
|
if (memb->generation > gen)
|
|
gen = memb->generation;
|
|
|
|
/* node doesn't support slots */
|
|
|
|
if (memb->slot == -1)
|
|
return -1;
|
|
|
|
/* node needs a slot assigned */
|
|
|
|
if (!memb->slot)
|
|
need++;
|
|
|
|
/* node has a slot assigned */
|
|
|
|
num++;
|
|
|
|
if (!max || max < memb->slot)
|
|
max = memb->slot;
|
|
|
|
/* sanity check, once slot is assigned it shouldn't change */
|
|
|
|
if (memb->slot_prev && memb->slot && memb->slot_prev != memb->slot) {
|
|
log_error(ls, "nodeid %d slot changed %d %d",
|
|
memb->nodeid, memb->slot_prev, memb->slot);
|
|
return -1;
|
|
}
|
|
memb->slot_prev = memb->slot;
|
|
}
|
|
|
|
array_size = max + need;
|
|
array = kcalloc(array_size, sizeof(*array), GFP_NOFS);
|
|
if (!array)
|
|
return -ENOMEM;
|
|
|
|
num = 0;
|
|
|
|
/* fill in slots (offsets) that are used */
|
|
|
|
list_for_each_entry(memb, &ls->ls_nodes, list) {
|
|
if (!memb->slot)
|
|
continue;
|
|
|
|
if (memb->slot > array_size) {
|
|
log_error(ls, "invalid slot number %d", memb->slot);
|
|
kfree(array);
|
|
return -1;
|
|
}
|
|
|
|
array[memb->slot - 1].nodeid = memb->nodeid;
|
|
array[memb->slot - 1].slot = memb->slot;
|
|
num++;
|
|
}
|
|
|
|
/* assign new slots from unused offsets */
|
|
|
|
list_for_each_entry(memb, &ls->ls_nodes, list) {
|
|
if (memb->slot)
|
|
continue;
|
|
|
|
for (i = 0; i < array_size; i++) {
|
|
if (array[i].nodeid)
|
|
continue;
|
|
|
|
memb->slot = i + 1;
|
|
memb->slot_prev = memb->slot;
|
|
array[i].nodeid = memb->nodeid;
|
|
array[i].slot = memb->slot;
|
|
num++;
|
|
|
|
if (!ls->ls_slot && memb->nodeid == our_nodeid)
|
|
ls->ls_slot = memb->slot;
|
|
break;
|
|
}
|
|
|
|
if (!memb->slot) {
|
|
log_error(ls, "no free slot found");
|
|
kfree(array);
|
|
return -1;
|
|
}
|
|
}
|
|
|
|
gen++;
|
|
|
|
log_slots(ls, gen, num, NULL, array, array_size);
|
|
|
|
max_slots = (dlm_config.ci_buffer_size - sizeof(struct dlm_rcom) -
|
|
sizeof(struct rcom_config)) / sizeof(struct rcom_slot);
|
|
|
|
if (num > max_slots) {
|
|
log_error(ls, "num_slots %d exceeds max_slots %d",
|
|
num, max_slots);
|
|
kfree(array);
|
|
return -1;
|
|
}
|
|
|
|
*gen_out = gen;
|
|
*slots_out = array;
|
|
*slots_size = array_size;
|
|
*num_slots = num;
|
|
return 0;
|
|
}
|
|
|
|
static void add_ordered_member(struct dlm_ls *ls, struct dlm_member *new)
|
|
{
|
|
struct dlm_member *memb = NULL;
|
|
struct list_head *tmp;
|
|
struct list_head *newlist = &new->list;
|
|
struct list_head *head = &ls->ls_nodes;
|
|
|
|
list_for_each(tmp, head) {
|
|
memb = list_entry(tmp, struct dlm_member, list);
|
|
if (new->nodeid < memb->nodeid)
|
|
break;
|
|
}
|
|
|
|
if (!memb)
|
|
list_add_tail(newlist, head);
|
|
else {
|
|
/* FIXME: can use list macro here */
|
|
newlist->prev = tmp->prev;
|
|
newlist->next = tmp;
|
|
tmp->prev->next = newlist;
|
|
tmp->prev = newlist;
|
|
}
|
|
}
|
|
|
|
static int dlm_add_member(struct dlm_ls *ls, struct dlm_config_node *node)
|
|
{
|
|
struct dlm_member *memb;
|
|
int error;
|
|
|
|
memb = kzalloc(sizeof(*memb), GFP_NOFS);
|
|
if (!memb)
|
|
return -ENOMEM;
|
|
|
|
error = dlm_lowcomms_connect_node(node->nodeid);
|
|
if (error < 0) {
|
|
kfree(memb);
|
|
return error;
|
|
}
|
|
|
|
memb->nodeid = node->nodeid;
|
|
memb->weight = node->weight;
|
|
memb->comm_seq = node->comm_seq;
|
|
add_ordered_member(ls, memb);
|
|
ls->ls_num_nodes++;
|
|
return 0;
|
|
}
|
|
|
|
static struct dlm_member *find_memb(struct list_head *head, int nodeid)
|
|
{
|
|
struct dlm_member *memb;
|
|
|
|
list_for_each_entry(memb, head, list) {
|
|
if (memb->nodeid == nodeid)
|
|
return memb;
|
|
}
|
|
return NULL;
|
|
}
|
|
|
|
int dlm_is_member(struct dlm_ls *ls, int nodeid)
|
|
{
|
|
if (find_memb(&ls->ls_nodes, nodeid))
|
|
return 1;
|
|
return 0;
|
|
}
|
|
|
|
int dlm_is_removed(struct dlm_ls *ls, int nodeid)
|
|
{
|
|
if (find_memb(&ls->ls_nodes_gone, nodeid))
|
|
return 1;
|
|
return 0;
|
|
}
|
|
|
|
static void clear_memb_list(struct list_head *head)
|
|
{
|
|
struct dlm_member *memb;
|
|
|
|
while (!list_empty(head)) {
|
|
memb = list_entry(head->next, struct dlm_member, list);
|
|
list_del(&memb->list);
|
|
kfree(memb);
|
|
}
|
|
}
|
|
|
|
void dlm_clear_members(struct dlm_ls *ls)
|
|
{
|
|
clear_memb_list(&ls->ls_nodes);
|
|
ls->ls_num_nodes = 0;
|
|
}
|
|
|
|
void dlm_clear_members_gone(struct dlm_ls *ls)
|
|
{
|
|
clear_memb_list(&ls->ls_nodes_gone);
|
|
}
|
|
|
|
static void make_member_array(struct dlm_ls *ls)
|
|
{
|
|
struct dlm_member *memb;
|
|
int i, w, x = 0, total = 0, all_zero = 0, *array;
|
|
|
|
kfree(ls->ls_node_array);
|
|
ls->ls_node_array = NULL;
|
|
|
|
list_for_each_entry(memb, &ls->ls_nodes, list) {
|
|
if (memb->weight)
|
|
total += memb->weight;
|
|
}
|
|
|
|
/* all nodes revert to weight of 1 if all have weight 0 */
|
|
|
|
if (!total) {
|
|
total = ls->ls_num_nodes;
|
|
all_zero = 1;
|
|
}
|
|
|
|
ls->ls_total_weight = total;
|
|
array = kmalloc_array(total, sizeof(*array), GFP_NOFS);
|
|
if (!array)
|
|
return;
|
|
|
|
list_for_each_entry(memb, &ls->ls_nodes, list) {
|
|
if (!all_zero && !memb->weight)
|
|
continue;
|
|
|
|
if (all_zero)
|
|
w = 1;
|
|
else
|
|
w = memb->weight;
|
|
|
|
DLM_ASSERT(x < total, printk("total %d x %d\n", total, x););
|
|
|
|
for (i = 0; i < w; i++)
|
|
array[x++] = memb->nodeid;
|
|
}
|
|
|
|
ls->ls_node_array = array;
|
|
}
|
|
|
|
/* send a status request to all members just to establish comms connections */
|
|
|
|
static int ping_members(struct dlm_ls *ls)
|
|
{
|
|
struct dlm_member *memb;
|
|
int error = 0;
|
|
|
|
list_for_each_entry(memb, &ls->ls_nodes, list) {
|
|
error = dlm_recovery_stopped(ls);
|
|
if (error)
|
|
break;
|
|
error = dlm_rcom_status(ls, memb->nodeid, 0);
|
|
if (error)
|
|
break;
|
|
}
|
|
if (error)
|
|
log_rinfo(ls, "ping_members aborted %d last nodeid %d",
|
|
error, ls->ls_recover_nodeid);
|
|
return error;
|
|
}
|
|
|
|
static void dlm_lsop_recover_prep(struct dlm_ls *ls)
|
|
{
|
|
if (!ls->ls_ops || !ls->ls_ops->recover_prep)
|
|
return;
|
|
ls->ls_ops->recover_prep(ls->ls_ops_arg);
|
|
}
|
|
|
|
static void dlm_lsop_recover_slot(struct dlm_ls *ls, struct dlm_member *memb)
|
|
{
|
|
struct dlm_slot slot;
|
|
uint32_t seq;
|
|
int error;
|
|
|
|
if (!ls->ls_ops || !ls->ls_ops->recover_slot)
|
|
return;
|
|
|
|
/* if there is no comms connection with this node
|
|
or the present comms connection is newer
|
|
than the one when this member was added, then
|
|
we consider the node to have failed (versus
|
|
being removed due to dlm_release_lockspace) */
|
|
|
|
error = dlm_comm_seq(memb->nodeid, &seq);
|
|
|
|
if (!error && seq == memb->comm_seq)
|
|
return;
|
|
|
|
slot.nodeid = memb->nodeid;
|
|
slot.slot = memb->slot;
|
|
|
|
ls->ls_ops->recover_slot(ls->ls_ops_arg, &slot);
|
|
}
|
|
|
|
void dlm_lsop_recover_done(struct dlm_ls *ls)
|
|
{
|
|
struct dlm_member *memb;
|
|
struct dlm_slot *slots;
|
|
int i, num;
|
|
|
|
if (!ls->ls_ops || !ls->ls_ops->recover_done)
|
|
return;
|
|
|
|
num = ls->ls_num_nodes;
|
|
slots = kcalloc(num, sizeof(*slots), GFP_KERNEL);
|
|
if (!slots)
|
|
return;
|
|
|
|
i = 0;
|
|
list_for_each_entry(memb, &ls->ls_nodes, list) {
|
|
if (i == num) {
|
|
log_error(ls, "dlm_lsop_recover_done bad num %d", num);
|
|
goto out;
|
|
}
|
|
slots[i].nodeid = memb->nodeid;
|
|
slots[i].slot = memb->slot;
|
|
i++;
|
|
}
|
|
|
|
ls->ls_ops->recover_done(ls->ls_ops_arg, slots, num,
|
|
ls->ls_slot, ls->ls_generation);
|
|
out:
|
|
kfree(slots);
|
|
}
|
|
|
|
static struct dlm_config_node *find_config_node(struct dlm_recover *rv,
|
|
int nodeid)
|
|
{
|
|
int i;
|
|
|
|
for (i = 0; i < rv->nodes_count; i++) {
|
|
if (rv->nodes[i].nodeid == nodeid)
|
|
return &rv->nodes[i];
|
|
}
|
|
return NULL;
|
|
}
|
|
|
|
int dlm_recover_members(struct dlm_ls *ls, struct dlm_recover *rv, int *neg_out)
|
|
{
|
|
struct dlm_member *memb, *safe;
|
|
struct dlm_config_node *node;
|
|
int i, error, neg = 0, low = -1;
|
|
|
|
/* previously removed members that we've not finished removing need to
|
|
count as a negative change so the "neg" recovery steps will happen */
|
|
|
|
list_for_each_entry(memb, &ls->ls_nodes_gone, list) {
|
|
log_rinfo(ls, "prev removed member %d", memb->nodeid);
|
|
neg++;
|
|
}
|
|
|
|
/* move departed members from ls_nodes to ls_nodes_gone */
|
|
|
|
list_for_each_entry_safe(memb, safe, &ls->ls_nodes, list) {
|
|
node = find_config_node(rv, memb->nodeid);
|
|
if (node && !node->new)
|
|
continue;
|
|
|
|
if (!node) {
|
|
log_rinfo(ls, "remove member %d", memb->nodeid);
|
|
} else {
|
|
/* removed and re-added */
|
|
log_rinfo(ls, "remove member %d comm_seq %u %u",
|
|
memb->nodeid, memb->comm_seq, node->comm_seq);
|
|
}
|
|
|
|
neg++;
|
|
list_move(&memb->list, &ls->ls_nodes_gone);
|
|
ls->ls_num_nodes--;
|
|
dlm_lsop_recover_slot(ls, memb);
|
|
}
|
|
|
|
/* add new members to ls_nodes */
|
|
|
|
for (i = 0; i < rv->nodes_count; i++) {
|
|
node = &rv->nodes[i];
|
|
if (dlm_is_member(ls, node->nodeid))
|
|
continue;
|
|
dlm_add_member(ls, node);
|
|
log_rinfo(ls, "add member %d", node->nodeid);
|
|
}
|
|
|
|
list_for_each_entry(memb, &ls->ls_nodes, list) {
|
|
if (low == -1 || memb->nodeid < low)
|
|
low = memb->nodeid;
|
|
}
|
|
ls->ls_low_nodeid = low;
|
|
|
|
make_member_array(ls);
|
|
*neg_out = neg;
|
|
|
|
error = ping_members(ls);
|
|
if (!error || error == -EPROTO) {
|
|
/* new_lockspace() may be waiting to know if the config
|
|
is good or bad */
|
|
ls->ls_members_result = error;
|
|
complete(&ls->ls_members_done);
|
|
}
|
|
|
|
log_rinfo(ls, "dlm_recover_members %d nodes", ls->ls_num_nodes);
|
|
return error;
|
|
}
|
|
|
|
/* Userspace guarantees that dlm_ls_stop() has completed on all nodes before
|
|
dlm_ls_start() is called on any of them to start the new recovery. */
|
|
|
|
int dlm_ls_stop(struct dlm_ls *ls)
|
|
{
|
|
int new;
|
|
|
|
/*
|
|
* Prevent dlm_recv from being in the middle of something when we do
|
|
* the stop. This includes ensuring dlm_recv isn't processing a
|
|
* recovery message (rcom), while dlm_recoverd is aborting and
|
|
* resetting things from an in-progress recovery. i.e. we want
|
|
* dlm_recoverd to abort its recovery without worrying about dlm_recv
|
|
* processing an rcom at the same time. Stopping dlm_recv also makes
|
|
* it easy for dlm_receive_message() to check locking stopped and add a
|
|
* message to the requestqueue without races.
|
|
*/
|
|
|
|
down_write(&ls->ls_recv_active);
|
|
|
|
/*
|
|
* Abort any recovery that's in progress (see RECOVER_STOP,
|
|
* dlm_recovery_stopped()) and tell any other threads running in the
|
|
* dlm to quit any processing (see RUNNING, dlm_locking_stopped()).
|
|
*/
|
|
|
|
spin_lock(&ls->ls_recover_lock);
|
|
set_bit(LSFL_RECOVER_STOP, &ls->ls_flags);
|
|
new = test_and_clear_bit(LSFL_RUNNING, &ls->ls_flags);
|
|
ls->ls_recover_seq++;
|
|
spin_unlock(&ls->ls_recover_lock);
|
|
|
|
/*
|
|
* Let dlm_recv run again, now any normal messages will be saved on the
|
|
* requestqueue for later.
|
|
*/
|
|
|
|
up_write(&ls->ls_recv_active);
|
|
|
|
/*
|
|
* This in_recovery lock does two things:
|
|
* 1) Keeps this function from returning until all threads are out
|
|
* of locking routines and locking is truly stopped.
|
|
* 2) Keeps any new requests from being processed until it's unlocked
|
|
* when recovery is complete.
|
|
*/
|
|
|
|
if (new) {
|
|
set_bit(LSFL_RECOVER_DOWN, &ls->ls_flags);
|
|
wake_up_process(ls->ls_recoverd_task);
|
|
wait_event(ls->ls_recover_lock_wait,
|
|
test_bit(LSFL_RECOVER_LOCK, &ls->ls_flags));
|
|
}
|
|
|
|
/*
|
|
* The recoverd suspend/resume makes sure that dlm_recoverd (if
|
|
* running) has noticed RECOVER_STOP above and quit processing the
|
|
* previous recovery.
|
|
*/
|
|
|
|
dlm_recoverd_suspend(ls);
|
|
|
|
spin_lock(&ls->ls_recover_lock);
|
|
kfree(ls->ls_slots);
|
|
ls->ls_slots = NULL;
|
|
ls->ls_num_slots = 0;
|
|
ls->ls_slots_size = 0;
|
|
ls->ls_recover_status = 0;
|
|
spin_unlock(&ls->ls_recover_lock);
|
|
|
|
dlm_recoverd_resume(ls);
|
|
|
|
if (!ls->ls_recover_begin)
|
|
ls->ls_recover_begin = jiffies;
|
|
|
|
dlm_lsop_recover_prep(ls);
|
|
return 0;
|
|
}
|
|
|
|
int dlm_ls_start(struct dlm_ls *ls)
|
|
{
|
|
struct dlm_recover *rv, *rv_old;
|
|
struct dlm_config_node *nodes = NULL;
|
|
int error, count;
|
|
|
|
rv = kzalloc(sizeof(*rv), GFP_NOFS);
|
|
if (!rv)
|
|
return -ENOMEM;
|
|
|
|
error = dlm_config_nodes(ls->ls_name, &nodes, &count);
|
|
if (error < 0)
|
|
goto fail_rv;
|
|
|
|
spin_lock(&ls->ls_recover_lock);
|
|
|
|
/* the lockspace needs to be stopped before it can be started */
|
|
|
|
if (!dlm_locking_stopped(ls)) {
|
|
spin_unlock(&ls->ls_recover_lock);
|
|
log_error(ls, "start ignored: lockspace running");
|
|
error = -EINVAL;
|
|
goto fail;
|
|
}
|
|
|
|
rv->nodes = nodes;
|
|
rv->nodes_count = count;
|
|
rv->seq = ++ls->ls_recover_seq;
|
|
rv_old = ls->ls_recover_args;
|
|
ls->ls_recover_args = rv;
|
|
spin_unlock(&ls->ls_recover_lock);
|
|
|
|
if (rv_old) {
|
|
log_error(ls, "unused recovery %llx %d",
|
|
(unsigned long long)rv_old->seq, rv_old->nodes_count);
|
|
kfree(rv_old->nodes);
|
|
kfree(rv_old);
|
|
}
|
|
|
|
set_bit(LSFL_RECOVER_WORK, &ls->ls_flags);
|
|
wake_up_process(ls->ls_recoverd_task);
|
|
return 0;
|
|
|
|
fail:
|
|
kfree(nodes);
|
|
fail_rv:
|
|
kfree(rv);
|
|
return error;
|
|
}
|
|
|