mirror of
https://github.com/AuxXxilium/linux_dsm_epyc7002.git
synced 2024-12-11 16:06:41 +07:00
5036f0a0ec
This patch fixes sparse warnings related to endian-ness, which were reported by the 0-day kernel build and testing tool. Reported-by: Fengguang Wu <fengguang.wu@intel.com> Signed-off-by: Naresh Kumar Inna <naresh@chelsio.com> Signed-off-by: James Bottomley <JBottomley@Parallels.com>
914 lines
22 KiB
C
914 lines
22 KiB
C
/*
|
|
* This file is part of the Chelsio FCoE driver for Linux.
|
|
*
|
|
* Copyright (c) 2008-2012 Chelsio Communications, Inc. All rights reserved.
|
|
*
|
|
* This software is available to you under a choice of one of two
|
|
* licenses. You may choose to be licensed under the terms of the GNU
|
|
* General Public License (GPL) Version 2, available from the file
|
|
* COPYING in the main directory of this source tree, or the
|
|
* OpenIB.org BSD license below:
|
|
*
|
|
* Redistribution and use in source and binary forms, with or
|
|
* without modification, are permitted provided that the following
|
|
* conditions are met:
|
|
*
|
|
* - Redistributions of source code must retain the above
|
|
* copyright notice, this list of conditions and the following
|
|
* disclaimer.
|
|
*
|
|
* - Redistributions in binary form must reproduce the above
|
|
* copyright notice, this list of conditions and the following
|
|
* disclaimer in the documentation and/or other materials
|
|
* provided with the distribution.
|
|
*
|
|
* THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND,
|
|
* EXPRESS OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF
|
|
* MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND
|
|
* NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT HOLDERS
|
|
* BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, WHETHER IN AN
|
|
* ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, OUT OF OR IN
|
|
* CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN THE
|
|
* SOFTWARE.
|
|
*/
|
|
|
|
#include <linux/string.h>
|
|
#include <scsi/scsi_device.h>
|
|
#include <scsi/scsi_transport_fc.h>
|
|
#include <scsi/fc/fc_els.h>
|
|
#include <scsi/fc/fc_fs.h>
|
|
|
|
#include "csio_hw.h"
|
|
#include "csio_lnode.h"
|
|
#include "csio_rnode.h"
|
|
|
|
static int csio_rnode_init(struct csio_rnode *, struct csio_lnode *);
|
|
static void csio_rnode_exit(struct csio_rnode *);
|
|
|
|
/* Static machine forward declarations */
|
|
static void csio_rns_uninit(struct csio_rnode *, enum csio_rn_ev);
|
|
static void csio_rns_ready(struct csio_rnode *, enum csio_rn_ev);
|
|
static void csio_rns_offline(struct csio_rnode *, enum csio_rn_ev);
|
|
static void csio_rns_disappeared(struct csio_rnode *, enum csio_rn_ev);
|
|
|
|
/* RNF event mapping */
|
|
static enum csio_rn_ev fwevt_to_rnevt[] = {
|
|
CSIO_RNFE_NONE, /* None */
|
|
CSIO_RNFE_LOGGED_IN, /* PLOGI_ACC_RCVD */
|
|
CSIO_RNFE_NONE, /* PLOGI_RJT_RCVD */
|
|
CSIO_RNFE_PLOGI_RECV, /* PLOGI_RCVD */
|
|
CSIO_RNFE_LOGO_RECV, /* PLOGO_RCVD */
|
|
CSIO_RNFE_PRLI_DONE, /* PRLI_ACC_RCVD */
|
|
CSIO_RNFE_NONE, /* PRLI_RJT_RCVD */
|
|
CSIO_RNFE_PRLI_RECV, /* PRLI_RCVD */
|
|
CSIO_RNFE_PRLO_RECV, /* PRLO_RCVD */
|
|
CSIO_RNFE_NONE, /* NPORT_ID_CHGD */
|
|
CSIO_RNFE_LOGO_RECV, /* FLOGO_RCVD */
|
|
CSIO_RNFE_NONE, /* CLR_VIRT_LNK_RCVD */
|
|
CSIO_RNFE_LOGGED_IN, /* FLOGI_ACC_RCVD */
|
|
CSIO_RNFE_NONE, /* FLOGI_RJT_RCVD */
|
|
CSIO_RNFE_LOGGED_IN, /* FDISC_ACC_RCVD */
|
|
CSIO_RNFE_NONE, /* FDISC_RJT_RCVD */
|
|
CSIO_RNFE_NONE, /* FLOGI_TMO_MAX_RETRY */
|
|
CSIO_RNFE_NONE, /* IMPL_LOGO_ADISC_ACC */
|
|
CSIO_RNFE_NONE, /* IMPL_LOGO_ADISC_RJT */
|
|
CSIO_RNFE_NONE, /* IMPL_LOGO_ADISC_CNFLT */
|
|
CSIO_RNFE_NONE, /* PRLI_TMO */
|
|
CSIO_RNFE_NONE, /* ADISC_TMO */
|
|
CSIO_RNFE_NAME_MISSING, /* RSCN_DEV_LOST */
|
|
CSIO_RNFE_NONE, /* SCR_ACC_RCVD */
|
|
CSIO_RNFE_NONE, /* ADISC_RJT_RCVD */
|
|
CSIO_RNFE_NONE, /* LOGO_SNT */
|
|
CSIO_RNFE_LOGO_RECV, /* PROTO_ERR_IMPL_LOGO */
|
|
};
|
|
|
|
#define CSIO_FWE_TO_RNFE(_evt) ((_evt > PROTO_ERR_IMPL_LOGO) ? \
|
|
CSIO_RNFE_NONE : \
|
|
fwevt_to_rnevt[_evt])
|
|
int
|
|
csio_is_rnode_ready(struct csio_rnode *rn)
|
|
{
|
|
return csio_match_state(rn, csio_rns_ready);
|
|
}
|
|
|
|
static int
|
|
csio_is_rnode_uninit(struct csio_rnode *rn)
|
|
{
|
|
return csio_match_state(rn, csio_rns_uninit);
|
|
}
|
|
|
|
static int
|
|
csio_is_rnode_wka(uint8_t rport_type)
|
|
{
|
|
if ((rport_type == FLOGI_VFPORT) ||
|
|
(rport_type == FDISC_VFPORT) ||
|
|
(rport_type == NS_VNPORT) ||
|
|
(rport_type == FDMI_VNPORT))
|
|
return 1;
|
|
|
|
return 0;
|
|
}
|
|
|
|
/*
|
|
* csio_rn_lookup - Finds the rnode with the given flowid
|
|
* @ln - lnode
|
|
* @flowid - flowid.
|
|
*
|
|
* Does the rnode lookup on the given lnode and flowid.If no matching entry
|
|
* found, NULL is returned.
|
|
*/
|
|
static struct csio_rnode *
|
|
csio_rn_lookup(struct csio_lnode *ln, uint32_t flowid)
|
|
{
|
|
struct csio_rnode *rnhead = (struct csio_rnode *) &ln->rnhead;
|
|
struct list_head *tmp;
|
|
struct csio_rnode *rn;
|
|
|
|
list_for_each(tmp, &rnhead->sm.sm_list) {
|
|
rn = (struct csio_rnode *) tmp;
|
|
if (rn->flowid == flowid)
|
|
return rn;
|
|
}
|
|
|
|
return NULL;
|
|
}
|
|
|
|
/*
|
|
* csio_rn_lookup_wwpn - Finds the rnode with the given wwpn
|
|
* @ln: lnode
|
|
* @wwpn: wwpn
|
|
*
|
|
* Does the rnode lookup on the given lnode and wwpn. If no matching entry
|
|
* found, NULL is returned.
|
|
*/
|
|
static struct csio_rnode *
|
|
csio_rn_lookup_wwpn(struct csio_lnode *ln, uint8_t *wwpn)
|
|
{
|
|
struct csio_rnode *rnhead = (struct csio_rnode *) &ln->rnhead;
|
|
struct list_head *tmp;
|
|
struct csio_rnode *rn;
|
|
|
|
list_for_each(tmp, &rnhead->sm.sm_list) {
|
|
rn = (struct csio_rnode *) tmp;
|
|
if (!memcmp(csio_rn_wwpn(rn), wwpn, 8))
|
|
return rn;
|
|
}
|
|
|
|
return NULL;
|
|
}
|
|
|
|
/**
|
|
* csio_rnode_lookup_portid - Finds the rnode with the given portid
|
|
* @ln: lnode
|
|
* @portid: port id
|
|
*
|
|
* Lookup the rnode list for a given portid. If no matching entry
|
|
* found, NULL is returned.
|
|
*/
|
|
struct csio_rnode *
|
|
csio_rnode_lookup_portid(struct csio_lnode *ln, uint32_t portid)
|
|
{
|
|
struct csio_rnode *rnhead = (struct csio_rnode *) &ln->rnhead;
|
|
struct list_head *tmp;
|
|
struct csio_rnode *rn;
|
|
|
|
list_for_each(tmp, &rnhead->sm.sm_list) {
|
|
rn = (struct csio_rnode *) tmp;
|
|
if (rn->nport_id == portid)
|
|
return rn;
|
|
}
|
|
|
|
return NULL;
|
|
}
|
|
|
|
static int
|
|
csio_rn_dup_flowid(struct csio_lnode *ln, uint32_t rdev_flowid,
|
|
uint32_t *vnp_flowid)
|
|
{
|
|
struct csio_rnode *rnhead;
|
|
struct list_head *tmp, *tmp1;
|
|
struct csio_rnode *rn;
|
|
struct csio_lnode *ln_tmp;
|
|
struct csio_hw *hw = csio_lnode_to_hw(ln);
|
|
|
|
list_for_each(tmp1, &hw->sln_head) {
|
|
ln_tmp = (struct csio_lnode *) tmp1;
|
|
if (ln_tmp == ln)
|
|
continue;
|
|
|
|
rnhead = (struct csio_rnode *)&ln_tmp->rnhead;
|
|
list_for_each(tmp, &rnhead->sm.sm_list) {
|
|
|
|
rn = (struct csio_rnode *) tmp;
|
|
if (csio_is_rnode_ready(rn)) {
|
|
if (rn->flowid == rdev_flowid) {
|
|
*vnp_flowid = csio_ln_flowid(ln_tmp);
|
|
return 1;
|
|
}
|
|
}
|
|
}
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static struct csio_rnode *
|
|
csio_alloc_rnode(struct csio_lnode *ln)
|
|
{
|
|
struct csio_hw *hw = csio_lnode_to_hw(ln);
|
|
|
|
struct csio_rnode *rn = mempool_alloc(hw->rnode_mempool, GFP_ATOMIC);
|
|
if (!rn)
|
|
goto err;
|
|
|
|
memset(rn, 0, sizeof(struct csio_rnode));
|
|
if (csio_rnode_init(rn, ln))
|
|
goto err_free;
|
|
|
|
CSIO_INC_STATS(ln, n_rnode_alloc);
|
|
|
|
return rn;
|
|
|
|
err_free:
|
|
mempool_free(rn, hw->rnode_mempool);
|
|
err:
|
|
CSIO_INC_STATS(ln, n_rnode_nomem);
|
|
return NULL;
|
|
}
|
|
|
|
static void
|
|
csio_free_rnode(struct csio_rnode *rn)
|
|
{
|
|
struct csio_hw *hw = csio_lnode_to_hw(csio_rnode_to_lnode(rn));
|
|
|
|
csio_rnode_exit(rn);
|
|
CSIO_INC_STATS(rn->lnp, n_rnode_free);
|
|
mempool_free(rn, hw->rnode_mempool);
|
|
}
|
|
|
|
/*
|
|
* csio_get_rnode - Gets rnode with the given flowid
|
|
* @ln - lnode
|
|
* @flowid - flow id.
|
|
*
|
|
* Does the rnode lookup on the given lnode and flowid. If no matching
|
|
* rnode found, then new rnode with given npid is allocated and returned.
|
|
*/
|
|
static struct csio_rnode *
|
|
csio_get_rnode(struct csio_lnode *ln, uint32_t flowid)
|
|
{
|
|
struct csio_rnode *rn;
|
|
|
|
rn = csio_rn_lookup(ln, flowid);
|
|
if (!rn) {
|
|
rn = csio_alloc_rnode(ln);
|
|
if (!rn)
|
|
return NULL;
|
|
|
|
rn->flowid = flowid;
|
|
}
|
|
|
|
return rn;
|
|
}
|
|
|
|
/*
|
|
* csio_put_rnode - Frees the given rnode
|
|
* @ln - lnode
|
|
* @flowid - flow id.
|
|
*
|
|
* Does the rnode lookup on the given lnode and flowid. If no matching
|
|
* rnode found, then new rnode with given npid is allocated and returned.
|
|
*/
|
|
void
|
|
csio_put_rnode(struct csio_lnode *ln, struct csio_rnode *rn)
|
|
{
|
|
CSIO_DB_ASSERT(csio_is_rnode_uninit(rn) != 0);
|
|
csio_free_rnode(rn);
|
|
}
|
|
|
|
/*
|
|
* csio_confirm_rnode - confirms rnode based on wwpn.
|
|
* @ln: lnode
|
|
* @rdev_flowid: remote device flowid
|
|
* @rdevp: remote device params
|
|
* This routines searches other rnode in list having same wwpn of new rnode.
|
|
* If there is a match, then matched rnode is returned and otherwise new rnode
|
|
* is returned.
|
|
* returns rnode.
|
|
*/
|
|
struct csio_rnode *
|
|
csio_confirm_rnode(struct csio_lnode *ln, uint32_t rdev_flowid,
|
|
struct fcoe_rdev_entry *rdevp)
|
|
{
|
|
uint8_t rport_type;
|
|
struct csio_rnode *rn, *match_rn;
|
|
uint32_t vnp_flowid;
|
|
__be32 *port_id;
|
|
|
|
port_id = (__be32 *)&rdevp->r_id[0];
|
|
rport_type =
|
|
FW_RDEV_WR_RPORT_TYPE_GET(rdevp->rd_xfer_rdy_to_rport_type);
|
|
|
|
/* Drop rdev event for cntrl port */
|
|
if (rport_type == FAB_CTLR_VNPORT) {
|
|
csio_ln_dbg(ln,
|
|
"Unhandled rport_type:%d recv in rdev evt "
|
|
"ssni:x%x\n", rport_type, rdev_flowid);
|
|
return NULL;
|
|
}
|
|
|
|
/* Lookup on flowid */
|
|
rn = csio_rn_lookup(ln, rdev_flowid);
|
|
if (!rn) {
|
|
|
|
/* Drop events with duplicate flowid */
|
|
if (csio_rn_dup_flowid(ln, rdev_flowid, &vnp_flowid)) {
|
|
csio_ln_warn(ln,
|
|
"ssni:%x already active on vnpi:%x",
|
|
rdev_flowid, vnp_flowid);
|
|
return NULL;
|
|
}
|
|
|
|
/* Lookup on wwpn for NPORTs */
|
|
rn = csio_rn_lookup_wwpn(ln, rdevp->wwpn);
|
|
if (!rn)
|
|
goto alloc_rnode;
|
|
|
|
} else {
|
|
/* Lookup well-known ports with nport id */
|
|
if (csio_is_rnode_wka(rport_type)) {
|
|
match_rn = csio_rnode_lookup_portid(ln,
|
|
((ntohl(*port_id) >> 8) & CSIO_DID_MASK));
|
|
if (match_rn == NULL) {
|
|
csio_rn_flowid(rn) = CSIO_INVALID_IDX;
|
|
goto alloc_rnode;
|
|
}
|
|
|
|
/*
|
|
* Now compare the wwpn to confirm that
|
|
* same port relogged in. If so update the matched rn.
|
|
* Else, go ahead and alloc a new rnode.
|
|
*/
|
|
if (!memcmp(csio_rn_wwpn(match_rn), rdevp->wwpn, 8)) {
|
|
if (csio_is_rnode_ready(rn)) {
|
|
csio_ln_warn(ln,
|
|
"rnode is already"
|
|
"active ssni:x%x\n",
|
|
rdev_flowid);
|
|
CSIO_ASSERT(0);
|
|
}
|
|
csio_rn_flowid(rn) = CSIO_INVALID_IDX;
|
|
rn = match_rn;
|
|
|
|
/* Update rn */
|
|
goto found_rnode;
|
|
}
|
|
csio_rn_flowid(rn) = CSIO_INVALID_IDX;
|
|
goto alloc_rnode;
|
|
}
|
|
|
|
/* wwpn match */
|
|
if (!memcmp(csio_rn_wwpn(rn), rdevp->wwpn, 8))
|
|
goto found_rnode;
|
|
|
|
/* Search for rnode that have same wwpn */
|
|
match_rn = csio_rn_lookup_wwpn(ln, rdevp->wwpn);
|
|
if (match_rn != NULL) {
|
|
csio_ln_dbg(ln,
|
|
"ssni:x%x changed for rport name(wwpn):%llx "
|
|
"did:x%x\n", rdev_flowid,
|
|
wwn_to_u64(rdevp->wwpn),
|
|
match_rn->nport_id);
|
|
csio_rn_flowid(rn) = CSIO_INVALID_IDX;
|
|
rn = match_rn;
|
|
} else {
|
|
csio_ln_dbg(ln,
|
|
"rnode wwpn mismatch found ssni:x%x "
|
|
"name(wwpn):%llx\n",
|
|
rdev_flowid,
|
|
wwn_to_u64(csio_rn_wwpn(rn)));
|
|
if (csio_is_rnode_ready(rn)) {
|
|
csio_ln_warn(ln,
|
|
"rnode is already active "
|
|
"wwpn:%llx ssni:x%x\n",
|
|
wwn_to_u64(csio_rn_wwpn(rn)),
|
|
rdev_flowid);
|
|
CSIO_ASSERT(0);
|
|
}
|
|
csio_rn_flowid(rn) = CSIO_INVALID_IDX;
|
|
goto alloc_rnode;
|
|
}
|
|
}
|
|
|
|
found_rnode:
|
|
csio_ln_dbg(ln, "found rnode:%p ssni:x%x name(wwpn):%llx\n",
|
|
rn, rdev_flowid, wwn_to_u64(rdevp->wwpn));
|
|
|
|
/* Update flowid */
|
|
csio_rn_flowid(rn) = rdev_flowid;
|
|
|
|
/* update rdev entry */
|
|
rn->rdev_entry = rdevp;
|
|
CSIO_INC_STATS(ln, n_rnode_match);
|
|
return rn;
|
|
|
|
alloc_rnode:
|
|
rn = csio_get_rnode(ln, rdev_flowid);
|
|
if (!rn)
|
|
return NULL;
|
|
|
|
csio_ln_dbg(ln, "alloc rnode:%p ssni:x%x name(wwpn):%llx\n",
|
|
rn, rdev_flowid, wwn_to_u64(rdevp->wwpn));
|
|
|
|
/* update rdev entry */
|
|
rn->rdev_entry = rdevp;
|
|
return rn;
|
|
}
|
|
|
|
/*
|
|
* csio_rn_verify_rparams - verify rparams.
|
|
* @ln: lnode
|
|
* @rn: rnode
|
|
* @rdevp: remote device params
|
|
* returns success if rparams are verified.
|
|
*/
|
|
static int
|
|
csio_rn_verify_rparams(struct csio_lnode *ln, struct csio_rnode *rn,
|
|
struct fcoe_rdev_entry *rdevp)
|
|
{
|
|
uint8_t null[8];
|
|
uint8_t rport_type;
|
|
uint8_t fc_class;
|
|
__be32 *did;
|
|
|
|
did = (__be32 *) &rdevp->r_id[0];
|
|
rport_type =
|
|
FW_RDEV_WR_RPORT_TYPE_GET(rdevp->rd_xfer_rdy_to_rport_type);
|
|
switch (rport_type) {
|
|
case FLOGI_VFPORT:
|
|
rn->role = CSIO_RNFR_FABRIC;
|
|
if (((ntohl(*did) >> 8) & CSIO_DID_MASK) != FC_FID_FLOGI) {
|
|
csio_ln_err(ln, "ssni:x%x invalid fabric portid\n",
|
|
csio_rn_flowid(rn));
|
|
return -EINVAL;
|
|
}
|
|
/* NPIV support */
|
|
if (FW_RDEV_WR_NPIV_GET(rdevp->vft_to_qos))
|
|
ln->flags |= CSIO_LNF_NPIVSUPP;
|
|
|
|
break;
|
|
|
|
case NS_VNPORT:
|
|
rn->role = CSIO_RNFR_NS;
|
|
if (((ntohl(*did) >> 8) & CSIO_DID_MASK) != FC_FID_DIR_SERV) {
|
|
csio_ln_err(ln, "ssni:x%x invalid fabric portid\n",
|
|
csio_rn_flowid(rn));
|
|
return -EINVAL;
|
|
}
|
|
break;
|
|
|
|
case REG_FC4_VNPORT:
|
|
case REG_VNPORT:
|
|
rn->role = CSIO_RNFR_NPORT;
|
|
if (rdevp->event_cause == PRLI_ACC_RCVD ||
|
|
rdevp->event_cause == PRLI_RCVD) {
|
|
if (FW_RDEV_WR_TASK_RETRY_ID_GET(
|
|
rdevp->enh_disc_to_tgt))
|
|
rn->fcp_flags |= FCP_SPPF_OVLY_ALLOW;
|
|
|
|
if (FW_RDEV_WR_RETRY_GET(rdevp->enh_disc_to_tgt))
|
|
rn->fcp_flags |= FCP_SPPF_RETRY;
|
|
|
|
if (FW_RDEV_WR_CONF_CMPL_GET(rdevp->enh_disc_to_tgt))
|
|
rn->fcp_flags |= FCP_SPPF_CONF_COMPL;
|
|
|
|
if (FW_RDEV_WR_TGT_GET(rdevp->enh_disc_to_tgt))
|
|
rn->role |= CSIO_RNFR_TARGET;
|
|
|
|
if (FW_RDEV_WR_INI_GET(rdevp->enh_disc_to_tgt))
|
|
rn->role |= CSIO_RNFR_INITIATOR;
|
|
}
|
|
|
|
break;
|
|
|
|
case FDMI_VNPORT:
|
|
case FAB_CTLR_VNPORT:
|
|
rn->role = 0;
|
|
break;
|
|
|
|
default:
|
|
csio_ln_err(ln, "ssni:x%x invalid rport type recv x%x\n",
|
|
csio_rn_flowid(rn), rport_type);
|
|
return -EINVAL;
|
|
}
|
|
|
|
/* validate wwpn/wwnn for Name server/remote port */
|
|
if (rport_type == REG_VNPORT || rport_type == NS_VNPORT) {
|
|
memset(null, 0, 8);
|
|
if (!memcmp(rdevp->wwnn, null, 8)) {
|
|
csio_ln_err(ln,
|
|
"ssni:x%x invalid wwnn received from"
|
|
" rport did:x%x\n",
|
|
csio_rn_flowid(rn),
|
|
(ntohl(*did) & CSIO_DID_MASK));
|
|
return -EINVAL;
|
|
}
|
|
|
|
if (!memcmp(rdevp->wwpn, null, 8)) {
|
|
csio_ln_err(ln,
|
|
"ssni:x%x invalid wwpn received from"
|
|
" rport did:x%x\n",
|
|
csio_rn_flowid(rn),
|
|
(ntohl(*did) & CSIO_DID_MASK));
|
|
return -EINVAL;
|
|
}
|
|
|
|
}
|
|
|
|
/* Copy wwnn, wwpn and nport id */
|
|
rn->nport_id = (ntohl(*did) >> 8) & CSIO_DID_MASK;
|
|
memcpy(csio_rn_wwnn(rn), rdevp->wwnn, 8);
|
|
memcpy(csio_rn_wwpn(rn), rdevp->wwpn, 8);
|
|
rn->rn_sparm.csp.sp_bb_data = rdevp->rcv_fr_sz;
|
|
fc_class = FW_RDEV_WR_CLASS_GET(rdevp->vft_to_qos);
|
|
rn->rn_sparm.clsp[fc_class - 1].cp_class = htons(FC_CPC_VALID);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void
|
|
__csio_reg_rnode(struct csio_rnode *rn)
|
|
{
|
|
struct csio_lnode *ln = csio_rnode_to_lnode(rn);
|
|
struct csio_hw *hw = csio_lnode_to_hw(ln);
|
|
|
|
spin_unlock_irq(&hw->lock);
|
|
csio_reg_rnode(rn);
|
|
spin_lock_irq(&hw->lock);
|
|
|
|
if (rn->role & CSIO_RNFR_TARGET)
|
|
ln->n_scsi_tgts++;
|
|
|
|
if (rn->nport_id == FC_FID_MGMT_SERV)
|
|
csio_ln_fdmi_start(ln, (void *) rn);
|
|
}
|
|
|
|
static void
|
|
__csio_unreg_rnode(struct csio_rnode *rn)
|
|
{
|
|
struct csio_lnode *ln = csio_rnode_to_lnode(rn);
|
|
struct csio_hw *hw = csio_lnode_to_hw(ln);
|
|
LIST_HEAD(tmp_q);
|
|
int cmpl = 0;
|
|
|
|
if (!list_empty(&rn->host_cmpl_q)) {
|
|
csio_dbg(hw, "Returning completion queue I/Os\n");
|
|
list_splice_tail_init(&rn->host_cmpl_q, &tmp_q);
|
|
cmpl = 1;
|
|
}
|
|
|
|
if (rn->role & CSIO_RNFR_TARGET) {
|
|
ln->n_scsi_tgts--;
|
|
ln->last_scan_ntgts--;
|
|
}
|
|
|
|
spin_unlock_irq(&hw->lock);
|
|
csio_unreg_rnode(rn);
|
|
spin_lock_irq(&hw->lock);
|
|
|
|
/* Cleanup I/Os that were waiting for rnode to unregister */
|
|
if (cmpl)
|
|
csio_scsi_cleanup_io_q(csio_hw_to_scsim(hw), &tmp_q);
|
|
|
|
}
|
|
|
|
/*****************************************************************************/
|
|
/* START: Rnode SM */
|
|
/*****************************************************************************/
|
|
|
|
/*
|
|
* csio_rns_uninit -
|
|
* @rn - rnode
|
|
* @evt - SM event.
|
|
*
|
|
*/
|
|
static void
|
|
csio_rns_uninit(struct csio_rnode *rn, enum csio_rn_ev evt)
|
|
{
|
|
struct csio_lnode *ln = csio_rnode_to_lnode(rn);
|
|
int ret = 0;
|
|
|
|
CSIO_INC_STATS(rn, n_evt_sm[evt]);
|
|
|
|
switch (evt) {
|
|
case CSIO_RNFE_LOGGED_IN:
|
|
case CSIO_RNFE_PLOGI_RECV:
|
|
ret = csio_rn_verify_rparams(ln, rn, rn->rdev_entry);
|
|
if (!ret) {
|
|
csio_set_state(&rn->sm, csio_rns_ready);
|
|
__csio_reg_rnode(rn);
|
|
} else {
|
|
CSIO_INC_STATS(rn, n_err_inval);
|
|
}
|
|
break;
|
|
case CSIO_RNFE_LOGO_RECV:
|
|
csio_ln_dbg(ln,
|
|
"ssni:x%x Ignoring event %d recv "
|
|
"in rn state[uninit]\n", csio_rn_flowid(rn), evt);
|
|
CSIO_INC_STATS(rn, n_evt_drop);
|
|
break;
|
|
default:
|
|
csio_ln_dbg(ln,
|
|
"ssni:x%x unexp event %d recv "
|
|
"in rn state[uninit]\n", csio_rn_flowid(rn), evt);
|
|
CSIO_INC_STATS(rn, n_evt_unexp);
|
|
break;
|
|
}
|
|
}
|
|
|
|
/*
|
|
* csio_rns_ready -
|
|
* @rn - rnode
|
|
* @evt - SM event.
|
|
*
|
|
*/
|
|
static void
|
|
csio_rns_ready(struct csio_rnode *rn, enum csio_rn_ev evt)
|
|
{
|
|
struct csio_lnode *ln = csio_rnode_to_lnode(rn);
|
|
int ret = 0;
|
|
|
|
CSIO_INC_STATS(rn, n_evt_sm[evt]);
|
|
|
|
switch (evt) {
|
|
case CSIO_RNFE_LOGGED_IN:
|
|
case CSIO_RNFE_PLOGI_RECV:
|
|
csio_ln_dbg(ln,
|
|
"ssni:x%x Ignoring event %d recv from did:x%x "
|
|
"in rn state[ready]\n", csio_rn_flowid(rn), evt,
|
|
rn->nport_id);
|
|
CSIO_INC_STATS(rn, n_evt_drop);
|
|
break;
|
|
|
|
case CSIO_RNFE_PRLI_DONE:
|
|
case CSIO_RNFE_PRLI_RECV:
|
|
ret = csio_rn_verify_rparams(ln, rn, rn->rdev_entry);
|
|
if (!ret)
|
|
__csio_reg_rnode(rn);
|
|
else
|
|
CSIO_INC_STATS(rn, n_err_inval);
|
|
|
|
break;
|
|
case CSIO_RNFE_DOWN:
|
|
csio_set_state(&rn->sm, csio_rns_offline);
|
|
__csio_unreg_rnode(rn);
|
|
|
|
/* FW expected to internally aborted outstanding SCSI WRs
|
|
* and return all SCSI WRs to host with status "ABORTED".
|
|
*/
|
|
break;
|
|
|
|
case CSIO_RNFE_LOGO_RECV:
|
|
csio_set_state(&rn->sm, csio_rns_offline);
|
|
|
|
__csio_unreg_rnode(rn);
|
|
|
|
/* FW expected to internally aborted outstanding SCSI WRs
|
|
* and return all SCSI WRs to host with status "ABORTED".
|
|
*/
|
|
break;
|
|
|
|
case CSIO_RNFE_CLOSE:
|
|
/*
|
|
* Each rnode receives CLOSE event when driver is removed or
|
|
* device is reset
|
|
* Note: All outstanding IOs on remote port need to returned
|
|
* to uppper layer with appropriate error before sending
|
|
* CLOSE event
|
|
*/
|
|
csio_set_state(&rn->sm, csio_rns_uninit);
|
|
__csio_unreg_rnode(rn);
|
|
break;
|
|
|
|
case CSIO_RNFE_NAME_MISSING:
|
|
csio_set_state(&rn->sm, csio_rns_disappeared);
|
|
__csio_unreg_rnode(rn);
|
|
|
|
/*
|
|
* FW expected to internally aborted outstanding SCSI WRs
|
|
* and return all SCSI WRs to host with status "ABORTED".
|
|
*/
|
|
|
|
break;
|
|
|
|
default:
|
|
csio_ln_dbg(ln,
|
|
"ssni:x%x unexp event %d recv from did:x%x "
|
|
"in rn state[uninit]\n", csio_rn_flowid(rn), evt,
|
|
rn->nport_id);
|
|
CSIO_INC_STATS(rn, n_evt_unexp);
|
|
break;
|
|
}
|
|
}
|
|
|
|
/*
|
|
* csio_rns_offline -
|
|
* @rn - rnode
|
|
* @evt - SM event.
|
|
*
|
|
*/
|
|
static void
|
|
csio_rns_offline(struct csio_rnode *rn, enum csio_rn_ev evt)
|
|
{
|
|
struct csio_lnode *ln = csio_rnode_to_lnode(rn);
|
|
int ret = 0;
|
|
|
|
CSIO_INC_STATS(rn, n_evt_sm[evt]);
|
|
|
|
switch (evt) {
|
|
case CSIO_RNFE_LOGGED_IN:
|
|
case CSIO_RNFE_PLOGI_RECV:
|
|
ret = csio_rn_verify_rparams(ln, rn, rn->rdev_entry);
|
|
if (!ret) {
|
|
csio_set_state(&rn->sm, csio_rns_ready);
|
|
__csio_reg_rnode(rn);
|
|
} else {
|
|
CSIO_INC_STATS(rn, n_err_inval);
|
|
csio_post_event(&rn->sm, CSIO_RNFE_CLOSE);
|
|
}
|
|
break;
|
|
|
|
case CSIO_RNFE_DOWN:
|
|
csio_ln_dbg(ln,
|
|
"ssni:x%x Ignoring event %d recv from did:x%x "
|
|
"in rn state[offline]\n", csio_rn_flowid(rn), evt,
|
|
rn->nport_id);
|
|
CSIO_INC_STATS(rn, n_evt_drop);
|
|
break;
|
|
|
|
case CSIO_RNFE_CLOSE:
|
|
/* Each rnode receives CLOSE event when driver is removed or
|
|
* device is reset
|
|
* Note: All outstanding IOs on remote port need to returned
|
|
* to uppper layer with appropriate error before sending
|
|
* CLOSE event
|
|
*/
|
|
csio_set_state(&rn->sm, csio_rns_uninit);
|
|
break;
|
|
|
|
case CSIO_RNFE_NAME_MISSING:
|
|
csio_set_state(&rn->sm, csio_rns_disappeared);
|
|
break;
|
|
|
|
default:
|
|
csio_ln_dbg(ln,
|
|
"ssni:x%x unexp event %d recv from did:x%x "
|
|
"in rn state[offline]\n", csio_rn_flowid(rn), evt,
|
|
rn->nport_id);
|
|
CSIO_INC_STATS(rn, n_evt_unexp);
|
|
break;
|
|
}
|
|
}
|
|
|
|
/*
|
|
* csio_rns_disappeared -
|
|
* @rn - rnode
|
|
* @evt - SM event.
|
|
*
|
|
*/
|
|
static void
|
|
csio_rns_disappeared(struct csio_rnode *rn, enum csio_rn_ev evt)
|
|
{
|
|
struct csio_lnode *ln = csio_rnode_to_lnode(rn);
|
|
int ret = 0;
|
|
|
|
CSIO_INC_STATS(rn, n_evt_sm[evt]);
|
|
|
|
switch (evt) {
|
|
case CSIO_RNFE_LOGGED_IN:
|
|
case CSIO_RNFE_PLOGI_RECV:
|
|
ret = csio_rn_verify_rparams(ln, rn, rn->rdev_entry);
|
|
if (!ret) {
|
|
csio_set_state(&rn->sm, csio_rns_ready);
|
|
__csio_reg_rnode(rn);
|
|
} else {
|
|
CSIO_INC_STATS(rn, n_err_inval);
|
|
csio_post_event(&rn->sm, CSIO_RNFE_CLOSE);
|
|
}
|
|
break;
|
|
|
|
case CSIO_RNFE_CLOSE:
|
|
/* Each rnode receives CLOSE event when driver is removed or
|
|
* device is reset.
|
|
* Note: All outstanding IOs on remote port need to returned
|
|
* to uppper layer with appropriate error before sending
|
|
* CLOSE event
|
|
*/
|
|
csio_set_state(&rn->sm, csio_rns_uninit);
|
|
break;
|
|
|
|
case CSIO_RNFE_DOWN:
|
|
case CSIO_RNFE_NAME_MISSING:
|
|
csio_ln_dbg(ln,
|
|
"ssni:x%x Ignoring event %d recv from did x%x"
|
|
"in rn state[disappeared]\n", csio_rn_flowid(rn),
|
|
evt, rn->nport_id);
|
|
break;
|
|
|
|
default:
|
|
csio_ln_dbg(ln,
|
|
"ssni:x%x unexp event %d recv from did x%x"
|
|
"in rn state[disappeared]\n", csio_rn_flowid(rn),
|
|
evt, rn->nport_id);
|
|
CSIO_INC_STATS(rn, n_evt_unexp);
|
|
break;
|
|
}
|
|
}
|
|
|
|
/*****************************************************************************/
|
|
/* END: Rnode SM */
|
|
/*****************************************************************************/
|
|
|
|
/*
|
|
* csio_rnode_devloss_handler - Device loss event handler
|
|
* @rn: rnode
|
|
*
|
|
* Post event to close rnode SM and free rnode.
|
|
*/
|
|
void
|
|
csio_rnode_devloss_handler(struct csio_rnode *rn)
|
|
{
|
|
struct csio_lnode *ln = csio_rnode_to_lnode(rn);
|
|
|
|
/* ignore if same rnode came back as online */
|
|
if (csio_is_rnode_ready(rn))
|
|
return;
|
|
|
|
csio_post_event(&rn->sm, CSIO_RNFE_CLOSE);
|
|
|
|
/* Free rn if in uninit state */
|
|
if (csio_is_rnode_uninit(rn))
|
|
csio_put_rnode(ln, rn);
|
|
}
|
|
|
|
/**
|
|
* csio_rnode_fwevt_handler - Event handler for firmware rnode events.
|
|
* @rn: rnode
|
|
*
|
|
*/
|
|
void
|
|
csio_rnode_fwevt_handler(struct csio_rnode *rn, uint8_t fwevt)
|
|
{
|
|
struct csio_lnode *ln = csio_rnode_to_lnode(rn);
|
|
enum csio_rn_ev evt;
|
|
|
|
evt = CSIO_FWE_TO_RNFE(fwevt);
|
|
if (!evt) {
|
|
csio_ln_err(ln, "ssni:x%x Unhandled FW Rdev event: %d\n",
|
|
csio_rn_flowid(rn), fwevt);
|
|
CSIO_INC_STATS(rn, n_evt_unexp);
|
|
return;
|
|
}
|
|
CSIO_INC_STATS(rn, n_evt_fw[fwevt]);
|
|
|
|
/* Track previous & current events for debugging */
|
|
rn->prev_evt = rn->cur_evt;
|
|
rn->cur_evt = fwevt;
|
|
|
|
/* Post event to rnode SM */
|
|
csio_post_event(&rn->sm, evt);
|
|
|
|
/* Free rn if in uninit state */
|
|
if (csio_is_rnode_uninit(rn))
|
|
csio_put_rnode(ln, rn);
|
|
}
|
|
|
|
/*
|
|
* csio_rnode_init - Initialize rnode.
|
|
* @rn: RNode
|
|
* @ln: Associated lnode
|
|
*
|
|
* Caller is responsible for holding the lock. The lock is required
|
|
* to be held for inserting the rnode in ln->rnhead list.
|
|
*/
|
|
static int
|
|
csio_rnode_init(struct csio_rnode *rn, struct csio_lnode *ln)
|
|
{
|
|
csio_rnode_to_lnode(rn) = ln;
|
|
csio_init_state(&rn->sm, csio_rns_uninit);
|
|
INIT_LIST_HEAD(&rn->host_cmpl_q);
|
|
csio_rn_flowid(rn) = CSIO_INVALID_IDX;
|
|
|
|
/* Add rnode to list of lnodes->rnhead */
|
|
list_add_tail(&rn->sm.sm_list, &ln->rnhead);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void
|
|
csio_rnode_exit(struct csio_rnode *rn)
|
|
{
|
|
list_del_init(&rn->sm.sm_list);
|
|
CSIO_DB_ASSERT(list_empty(&rn->host_cmpl_q));
|
|
}
|