2009-09-05 09:06:35 +07:00
|
|
|
/**
|
2015-07-04 05:42:33 +07:00
|
|
|
* Copyright (C) 2005 - 2015 Emulex
|
2009-09-05 09:06:35 +07:00
|
|
|
* All rights reserved.
|
|
|
|
*
|
|
|
|
* This program is free software; you can redistribute it and/or
|
|
|
|
* modify it under the terms of the GNU General Public License version 2
|
|
|
|
* as published by the Free Software Foundation. The full GNU General
|
|
|
|
* Public License is included in this distribution in the file called COPYING.
|
|
|
|
*
|
2015-05-15 13:16:17 +07:00
|
|
|
* Written by: Jayamohan Kallickal (jayamohan.kallickal@avagotech.com)
|
2009-09-05 09:06:35 +07:00
|
|
|
*
|
|
|
|
* Contact Information:
|
2015-05-15 13:16:17 +07:00
|
|
|
* linux-drivers@avagotech.com
|
2009-09-05 09:06:35 +07:00
|
|
|
*
|
2015-07-04 05:42:33 +07:00
|
|
|
* Emulex
|
2011-03-26 04:23:57 +07:00
|
|
|
* 3333 Susan Street
|
|
|
|
* Costa Mesa, CA 92626
|
2009-09-05 09:06:35 +07:00
|
|
|
*/
|
2011-03-26 04:23:57 +07:00
|
|
|
|
2009-09-05 09:06:35 +07:00
|
|
|
#include <linux/reboot.h>
|
|
|
|
#include <linux/delay.h>
|
include cleanup: Update gfp.h and slab.h includes to prepare for breaking implicit slab.h inclusion from percpu.h
percpu.h is included by sched.h and module.h and thus ends up being
included when building most .c files. percpu.h includes slab.h which
in turn includes gfp.h making everything defined by the two files
universally available and complicating inclusion dependencies.
percpu.h -> slab.h dependency is about to be removed. Prepare for
this change by updating users of gfp and slab facilities include those
headers directly instead of assuming availability. As this conversion
needs to touch large number of source files, the following script is
used as the basis of conversion.
http://userweb.kernel.org/~tj/misc/slabh-sweep.py
The script does the followings.
* Scan files for gfp and slab usages and update includes such that
only the necessary includes are there. ie. if only gfp is used,
gfp.h, if slab is used, slab.h.
* When the script inserts a new include, it looks at the include
blocks and try to put the new include such that its order conforms
to its surrounding. It's put in the include block which contains
core kernel includes, in the same order that the rest are ordered -
alphabetical, Christmas tree, rev-Xmas-tree or at the end if there
doesn't seem to be any matching order.
* If the script can't find a place to put a new include (mostly
because the file doesn't have fitting include block), it prints out
an error message indicating which .h file needs to be added to the
file.
The conversion was done in the following steps.
1. The initial automatic conversion of all .c files updated slightly
over 4000 files, deleting around 700 includes and adding ~480 gfp.h
and ~3000 slab.h inclusions. The script emitted errors for ~400
files.
2. Each error was manually checked. Some didn't need the inclusion,
some needed manual addition while adding it to implementation .h or
embedding .c file was more appropriate for others. This step added
inclusions to around 150 files.
3. The script was run again and the output was compared to the edits
from #2 to make sure no file was left behind.
4. Several build tests were done and a couple of problems were fixed.
e.g. lib/decompress_*.c used malloc/free() wrappers around slab
APIs requiring slab.h to be added manually.
5. The script was run on all .h files but without automatically
editing them as sprinkling gfp.h and slab.h inclusions around .h
files could easily lead to inclusion dependency hell. Most gfp.h
inclusion directives were ignored as stuff from gfp.h was usually
wildly available and often used in preprocessor macros. Each
slab.h inclusion directive was examined and added manually as
necessary.
6. percpu.h was updated not to include slab.h.
7. Build test were done on the following configurations and failures
were fixed. CONFIG_GCOV_KERNEL was turned off for all tests (as my
distributed build env didn't work with gcov compiles) and a few
more options had to be turned off depending on archs to make things
build (like ipr on powerpc/64 which failed due to missing writeq).
* x86 and x86_64 UP and SMP allmodconfig and a custom test config.
* powerpc and powerpc64 SMP allmodconfig
* sparc and sparc64 SMP allmodconfig
* ia64 SMP allmodconfig
* s390 SMP allmodconfig
* alpha SMP allmodconfig
* um on x86_64 SMP allmodconfig
8. percpu.h modifications were reverted so that it could be applied as
a separate patch and serve as bisection point.
Given the fact that I had only a couple of failures from tests on step
6, I'm fairly confident about the coverage of this conversion patch.
If there is a breakage, it's likely to be something in one of the arch
headers which should be easily discoverable easily on most builds of
the specific arch.
Signed-off-by: Tejun Heo <tj@kernel.org>
Guess-its-ok-by: Christoph Lameter <cl@linux-foundation.org>
Cc: Ingo Molnar <mingo@redhat.com>
Cc: Lee Schermerhorn <Lee.Schermerhorn@hp.com>
2010-03-24 15:04:11 +07:00
|
|
|
#include <linux/slab.h>
|
2009-09-05 09:06:35 +07:00
|
|
|
#include <linux/interrupt.h>
|
|
|
|
#include <linux/blkdev.h>
|
|
|
|
#include <linux/pci.h>
|
|
|
|
#include <linux/string.h>
|
|
|
|
#include <linux/kernel.h>
|
|
|
|
#include <linux/semaphore.h>
|
2010-07-22 05:59:18 +07:00
|
|
|
#include <linux/iscsi_boot_sysfs.h>
|
2011-05-27 20:47:43 +07:00
|
|
|
#include <linux/module.h>
|
2012-04-04 11:41:50 +07:00
|
|
|
#include <linux/bsg-lib.h>
|
2016-01-20 15:40:56 +07:00
|
|
|
#include <linux/irq_poll.h>
|
2009-09-05 09:06:35 +07:00
|
|
|
|
|
|
|
#include <scsi/libiscsi.h>
|
2012-04-04 11:41:50 +07:00
|
|
|
#include <scsi/scsi_bsg_iscsi.h>
|
|
|
|
#include <scsi/scsi_netlink.h>
|
2009-09-05 09:06:35 +07:00
|
|
|
#include <scsi/scsi_transport_iscsi.h>
|
|
|
|
#include <scsi/scsi_transport.h>
|
|
|
|
#include <scsi/scsi_cmnd.h>
|
|
|
|
#include <scsi/scsi_device.h>
|
|
|
|
#include <scsi/scsi_host.h>
|
|
|
|
#include <scsi/scsi.h>
|
|
|
|
#include "be_main.h"
|
|
|
|
#include "be_iscsi.h"
|
|
|
|
#include "be_mgmt.h"
|
2012-08-21 00:30:55 +07:00
|
|
|
#include "be_cmds.h"
|
2009-09-05 09:06:35 +07:00
|
|
|
|
|
|
|
static unsigned int be_iopoll_budget = 10;
|
|
|
|
static unsigned int be_max_phys_size = 64;
|
2009-10-23 13:22:33 +07:00
|
|
|
static unsigned int enable_msix = 1;
|
2009-09-05 09:06:35 +07:00
|
|
|
|
|
|
|
MODULE_DESCRIPTION(DRV_DESC " " BUILD_STR);
|
2012-04-04 11:41:46 +07:00
|
|
|
MODULE_VERSION(BUILD_STR);
|
2015-07-04 05:42:33 +07:00
|
|
|
MODULE_AUTHOR("Emulex Corporation");
|
2009-09-05 09:06:35 +07:00
|
|
|
MODULE_LICENSE("GPL");
|
|
|
|
module_param(be_iopoll_budget, int, 0);
|
|
|
|
module_param(enable_msix, int, 0);
|
|
|
|
module_param(be_max_phys_size, uint, S_IRUGO);
|
2012-08-21 00:30:18 +07:00
|
|
|
MODULE_PARM_DESC(be_max_phys_size,
|
|
|
|
"Maximum Size (In Kilobytes) of physically contiguous "
|
|
|
|
"memory that can be allocated. Range is 16 - 128");
|
|
|
|
|
|
|
|
#define beiscsi_disp_param(_name)\
|
|
|
|
ssize_t \
|
|
|
|
beiscsi_##_name##_disp(struct device *dev,\
|
|
|
|
struct device_attribute *attrib, char *buf) \
|
|
|
|
{ \
|
|
|
|
struct Scsi_Host *shost = class_to_shost(dev);\
|
|
|
|
struct beiscsi_hba *phba = iscsi_host_priv(shost); \
|
|
|
|
uint32_t param_val = 0; \
|
|
|
|
param_val = phba->attr_##_name;\
|
|
|
|
return snprintf(buf, PAGE_SIZE, "%d\n",\
|
|
|
|
phba->attr_##_name);\
|
|
|
|
}
|
|
|
|
|
|
|
|
#define beiscsi_change_param(_name, _minval, _maxval, _defaval)\
|
|
|
|
int \
|
|
|
|
beiscsi_##_name##_change(struct beiscsi_hba *phba, uint32_t val)\
|
|
|
|
{\
|
|
|
|
if (val >= _minval && val <= _maxval) {\
|
|
|
|
beiscsi_log(phba, KERN_ERR, BEISCSI_LOG_INIT,\
|
|
|
|
"BA_%d : beiscsi_"#_name" updated "\
|
|
|
|
"from 0x%x ==> 0x%x\n",\
|
|
|
|
phba->attr_##_name, val); \
|
|
|
|
phba->attr_##_name = val;\
|
|
|
|
return 0;\
|
|
|
|
} \
|
|
|
|
beiscsi_log(phba, KERN_ERR, BEISCSI_LOG_INIT, \
|
|
|
|
"BA_%d beiscsi_"#_name" attribute "\
|
|
|
|
"cannot be updated to 0x%x, "\
|
|
|
|
"range allowed is ["#_minval" - "#_maxval"]\n", val);\
|
|
|
|
return -EINVAL;\
|
|
|
|
}
|
|
|
|
|
|
|
|
#define beiscsi_store_param(_name) \
|
|
|
|
ssize_t \
|
|
|
|
beiscsi_##_name##_store(struct device *dev,\
|
|
|
|
struct device_attribute *attr, const char *buf,\
|
|
|
|
size_t count) \
|
|
|
|
{ \
|
|
|
|
struct Scsi_Host *shost = class_to_shost(dev);\
|
|
|
|
struct beiscsi_hba *phba = iscsi_host_priv(shost);\
|
|
|
|
uint32_t param_val = 0;\
|
|
|
|
if (!isdigit(buf[0]))\
|
|
|
|
return -EINVAL;\
|
|
|
|
if (sscanf(buf, "%i", ¶m_val) != 1)\
|
|
|
|
return -EINVAL;\
|
|
|
|
if (beiscsi_##_name##_change(phba, param_val) == 0) \
|
|
|
|
return strlen(buf);\
|
|
|
|
else \
|
|
|
|
return -EINVAL;\
|
|
|
|
}
|
|
|
|
|
|
|
|
#define beiscsi_init_param(_name, _minval, _maxval, _defval) \
|
|
|
|
int \
|
|
|
|
beiscsi_##_name##_init(struct beiscsi_hba *phba, uint32_t val) \
|
|
|
|
{ \
|
|
|
|
if (val >= _minval && val <= _maxval) {\
|
|
|
|
phba->attr_##_name = val;\
|
|
|
|
return 0;\
|
|
|
|
} \
|
|
|
|
beiscsi_log(phba, KERN_ERR, BEISCSI_LOG_INIT,\
|
|
|
|
"BA_%d beiscsi_"#_name" attribute " \
|
|
|
|
"cannot be updated to 0x%x, "\
|
|
|
|
"range allowed is ["#_minval" - "#_maxval"]\n", val);\
|
|
|
|
phba->attr_##_name = _defval;\
|
|
|
|
return -EINVAL;\
|
|
|
|
}
|
|
|
|
|
|
|
|
#define BEISCSI_RW_ATTR(_name, _minval, _maxval, _defval, _descp) \
|
|
|
|
static uint beiscsi_##_name = _defval;\
|
|
|
|
module_param(beiscsi_##_name, uint, S_IRUGO);\
|
|
|
|
MODULE_PARM_DESC(beiscsi_##_name, _descp);\
|
|
|
|
beiscsi_disp_param(_name)\
|
|
|
|
beiscsi_change_param(_name, _minval, _maxval, _defval)\
|
|
|
|
beiscsi_store_param(_name)\
|
|
|
|
beiscsi_init_param(_name, _minval, _maxval, _defval)\
|
|
|
|
DEVICE_ATTR(beiscsi_##_name, S_IRUGO | S_IWUSR,\
|
|
|
|
beiscsi_##_name##_disp, beiscsi_##_name##_store)
|
|
|
|
|
|
|
|
/*
|
|
|
|
* When new log level added update the
|
|
|
|
* the MAX allowed value for log_enable
|
|
|
|
*/
|
|
|
|
BEISCSI_RW_ATTR(log_enable, 0x00,
|
|
|
|
0xFF, 0x00, "Enable logging Bit Mask\n"
|
|
|
|
"\t\t\t\tInitialization Events : 0x01\n"
|
|
|
|
"\t\t\t\tMailbox Events : 0x02\n"
|
|
|
|
"\t\t\t\tMiscellaneous Events : 0x04\n"
|
|
|
|
"\t\t\t\tError Handling : 0x08\n"
|
|
|
|
"\t\t\t\tIO Path Events : 0x10\n"
|
2013-09-29 05:35:55 +07:00
|
|
|
"\t\t\t\tConfiguration Path : 0x20\n"
|
|
|
|
"\t\t\t\tiSCSI Protocol : 0x40\n");
|
2012-08-21 00:30:18 +07:00
|
|
|
|
2012-10-20 06:12:25 +07:00
|
|
|
DEVICE_ATTR(beiscsi_drvr_ver, S_IRUGO, beiscsi_drvr_ver_disp, NULL);
|
2012-10-20 06:15:06 +07:00
|
|
|
DEVICE_ATTR(beiscsi_adapter_family, S_IRUGO, beiscsi_adap_family_disp, NULL);
|
2013-04-06 10:38:28 +07:00
|
|
|
DEVICE_ATTR(beiscsi_fw_ver, S_IRUGO, beiscsi_fw_ver_disp, NULL);
|
2013-09-29 05:35:53 +07:00
|
|
|
DEVICE_ATTR(beiscsi_phys_port, S_IRUGO, beiscsi_phys_port_disp, NULL);
|
2013-09-29 05:35:52 +07:00
|
|
|
DEVICE_ATTR(beiscsi_active_session_count, S_IRUGO,
|
|
|
|
beiscsi_active_session_disp, NULL);
|
|
|
|
DEVICE_ATTR(beiscsi_free_session_count, S_IRUGO,
|
|
|
|
beiscsi_free_session_disp, NULL);
|
2012-08-21 00:30:18 +07:00
|
|
|
struct device_attribute *beiscsi_attrs[] = {
|
|
|
|
&dev_attr_beiscsi_log_enable,
|
2012-10-20 06:12:25 +07:00
|
|
|
&dev_attr_beiscsi_drvr_ver,
|
2012-10-20 06:15:06 +07:00
|
|
|
&dev_attr_beiscsi_adapter_family,
|
2013-04-06 10:38:28 +07:00
|
|
|
&dev_attr_beiscsi_fw_ver,
|
2013-09-29 05:35:52 +07:00
|
|
|
&dev_attr_beiscsi_active_session_count,
|
|
|
|
&dev_attr_beiscsi_free_session_count,
|
2013-09-29 05:35:53 +07:00
|
|
|
&dev_attr_beiscsi_phys_port,
|
2012-08-21 00:30:18 +07:00
|
|
|
NULL,
|
|
|
|
};
|
2009-09-05 09:06:35 +07:00
|
|
|
|
2012-10-20 06:11:45 +07:00
|
|
|
static char const *cqe_desc[] = {
|
|
|
|
"RESERVED_DESC",
|
|
|
|
"SOL_CMD_COMPLETE",
|
|
|
|
"SOL_CMD_KILLED_DATA_DIGEST_ERR",
|
|
|
|
"CXN_KILLED_PDU_SIZE_EXCEEDS_DSL",
|
|
|
|
"CXN_KILLED_BURST_LEN_MISMATCH",
|
|
|
|
"CXN_KILLED_AHS_RCVD",
|
|
|
|
"CXN_KILLED_HDR_DIGEST_ERR",
|
|
|
|
"CXN_KILLED_UNKNOWN_HDR",
|
|
|
|
"CXN_KILLED_STALE_ITT_TTT_RCVD",
|
|
|
|
"CXN_KILLED_INVALID_ITT_TTT_RCVD",
|
|
|
|
"CXN_KILLED_RST_RCVD",
|
|
|
|
"CXN_KILLED_TIMED_OUT",
|
|
|
|
"CXN_KILLED_RST_SENT",
|
|
|
|
"CXN_KILLED_FIN_RCVD",
|
|
|
|
"CXN_KILLED_BAD_UNSOL_PDU_RCVD",
|
|
|
|
"CXN_KILLED_BAD_WRB_INDEX_ERROR",
|
|
|
|
"CXN_KILLED_OVER_RUN_RESIDUAL",
|
|
|
|
"CXN_KILLED_UNDER_RUN_RESIDUAL",
|
|
|
|
"CMD_KILLED_INVALID_STATSN_RCVD",
|
|
|
|
"CMD_KILLED_INVALID_R2T_RCVD",
|
|
|
|
"CMD_CXN_KILLED_LUN_INVALID",
|
|
|
|
"CMD_CXN_KILLED_ICD_INVALID",
|
|
|
|
"CMD_CXN_KILLED_ITT_INVALID",
|
|
|
|
"CMD_CXN_KILLED_SEQ_OUTOFORDER",
|
|
|
|
"CMD_CXN_KILLED_INVALID_DATASN_RCVD",
|
|
|
|
"CXN_INVALIDATE_NOTIFY",
|
|
|
|
"CXN_INVALIDATE_INDEX_NOTIFY",
|
|
|
|
"CMD_INVALIDATED_NOTIFY",
|
|
|
|
"UNSOL_HDR_NOTIFY",
|
|
|
|
"UNSOL_DATA_NOTIFY",
|
|
|
|
"UNSOL_DATA_DIGEST_ERROR_NOTIFY",
|
|
|
|
"DRIVERMSG_NOTIFY",
|
|
|
|
"CXN_KILLED_CMND_DATA_NOT_ON_SAME_CONN",
|
|
|
|
"SOL_CMD_KILLED_DIF_ERR",
|
|
|
|
"CXN_KILLED_SYN_RCVD",
|
|
|
|
"CXN_KILLED_IMM_DATA_RCVD"
|
|
|
|
};
|
|
|
|
|
2009-09-05 09:06:35 +07:00
|
|
|
static int beiscsi_slave_configure(struct scsi_device *sdev)
|
|
|
|
{
|
|
|
|
blk_queue_max_segment_size(sdev->request_queue, 65536);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2010-02-20 09:32:39 +07:00
|
|
|
static int beiscsi_eh_abort(struct scsi_cmnd *sc)
|
|
|
|
{
|
|
|
|
struct iscsi_cls_session *cls_session;
|
|
|
|
struct iscsi_task *aborted_task = (struct iscsi_task *)sc->SCp.ptr;
|
|
|
|
struct beiscsi_io_task *aborted_io_task;
|
|
|
|
struct iscsi_conn *conn;
|
|
|
|
struct beiscsi_conn *beiscsi_conn;
|
|
|
|
struct beiscsi_hba *phba;
|
|
|
|
struct iscsi_session *session;
|
|
|
|
struct invalidate_command_table *inv_tbl;
|
2010-07-22 05:57:47 +07:00
|
|
|
struct be_dma_mem nonemb_cmd;
|
2010-02-20 09:32:39 +07:00
|
|
|
unsigned int cid, tag, num_invalidate;
|
2014-01-29 14:16:39 +07:00
|
|
|
int rc;
|
2010-02-20 09:32:39 +07:00
|
|
|
|
|
|
|
cls_session = starget_to_session(scsi_target(sc->device));
|
|
|
|
session = cls_session->dd_data;
|
|
|
|
|
[SCSI] libiscsi: Reduce locking contention in fast path
Replace the session lock with two locks, a forward lock and
a backwards lock named frwd_lock and back_lock respectively.
The forward lock protects resources that change while sending a
request to the target, such as cmdsn, queued_cmdsn, and allocating
task from the commands' pool with kfifo_out.
The backward lock protects resources that change while processing
a response or in error path, such as cmdsn_exp, cmdsn_max, and
returning tasks to the commands' pool with kfifo_in.
Under a steady state fast-path situation, that is when one
or more processes/threads submit IO to an iscsi device and
a single kernel upcall (e.g softirq) is dealing with processing
of responses without errors, this patch eliminates the contention
between the queuecommand()/request response/scsi_done() flows
associated with iscsi sessions.
Between the forward and the backward locks exists a strict locking
hierarchy. The mutual exclusion zone protected by the forward lock can
enclose the mutual exclusion zone protected by the backward lock but not
vice versa.
For example, in iscsi_conn_teardown or in iscsi_xmit_data when there is
a failure and __iscsi_put_task is called, the backward lock is taken while
the forward lock is still taken. On the other hand, if in the RX path a nop
is to be sent, for example in iscsi_handle_reject or __iscsi_complete_pdu
than the forward lock is released and the backward lock is taken for the
duration of iscsi_send_nopout, later the backward lock is released and the
forward lock is retaken.
libiscsi_tcp uses two kernel fifos the r2t pool and the r2t queue.
The insertion and deletion from these queues didn't corespond to the
assumption taken by the new forward/backwards session locking paradigm.
That is, in iscsi_tcp_clenup_task which belongs to the RX (backwards)
path, r2t is taken out from r2t queue and inserted to the r2t pool.
In iscsi_tcp_get_curr_r2t which belong to the TX (forward) path, r2t
is also inserted to the r2t pool and another r2t is pulled from r2t
queue.
Only in iscsi_tcp_r2t_rsp which is called in the RX path but can requeue
to the TX path, r2t is taken from the r2t pool and inserted to the r2t
queue.
In order to cope with this situation, two spin locks were added,
pool2queue and queue2pool. The former protects extracting from the
r2t pool and inserting to the r2t queue, and the later protects the
extracing from the r2t queue and inserting to the r2t pool.
Signed-off-by: Shlomo Pongratz <shlomop@mellanox.com>
Signed-off-by: Or Gerlitz <ogerlitz@mellanox.com>
[minor fix up to apply cleanly and compile fix]
Signed-off-by: Mike Christie <michaelc@cs.wisc.edu>
Signed-off-by: James Bottomley <JBottomley@Parallels.com>
2014-02-07 13:41:38 +07:00
|
|
|
spin_lock_bh(&session->frwd_lock);
|
2010-02-20 09:32:39 +07:00
|
|
|
if (!aborted_task || !aborted_task->sc) {
|
|
|
|
/* we raced */
|
[SCSI] libiscsi: Reduce locking contention in fast path
Replace the session lock with two locks, a forward lock and
a backwards lock named frwd_lock and back_lock respectively.
The forward lock protects resources that change while sending a
request to the target, such as cmdsn, queued_cmdsn, and allocating
task from the commands' pool with kfifo_out.
The backward lock protects resources that change while processing
a response or in error path, such as cmdsn_exp, cmdsn_max, and
returning tasks to the commands' pool with kfifo_in.
Under a steady state fast-path situation, that is when one
or more processes/threads submit IO to an iscsi device and
a single kernel upcall (e.g softirq) is dealing with processing
of responses without errors, this patch eliminates the contention
between the queuecommand()/request response/scsi_done() flows
associated with iscsi sessions.
Between the forward and the backward locks exists a strict locking
hierarchy. The mutual exclusion zone protected by the forward lock can
enclose the mutual exclusion zone protected by the backward lock but not
vice versa.
For example, in iscsi_conn_teardown or in iscsi_xmit_data when there is
a failure and __iscsi_put_task is called, the backward lock is taken while
the forward lock is still taken. On the other hand, if in the RX path a nop
is to be sent, for example in iscsi_handle_reject or __iscsi_complete_pdu
than the forward lock is released and the backward lock is taken for the
duration of iscsi_send_nopout, later the backward lock is released and the
forward lock is retaken.
libiscsi_tcp uses two kernel fifos the r2t pool and the r2t queue.
The insertion and deletion from these queues didn't corespond to the
assumption taken by the new forward/backwards session locking paradigm.
That is, in iscsi_tcp_clenup_task which belongs to the RX (backwards)
path, r2t is taken out from r2t queue and inserted to the r2t pool.
In iscsi_tcp_get_curr_r2t which belong to the TX (forward) path, r2t
is also inserted to the r2t pool and another r2t is pulled from r2t
queue.
Only in iscsi_tcp_r2t_rsp which is called in the RX path but can requeue
to the TX path, r2t is taken from the r2t pool and inserted to the r2t
queue.
In order to cope with this situation, two spin locks were added,
pool2queue and queue2pool. The former protects extracting from the
r2t pool and inserting to the r2t queue, and the later protects the
extracing from the r2t queue and inserting to the r2t pool.
Signed-off-by: Shlomo Pongratz <shlomop@mellanox.com>
Signed-off-by: Or Gerlitz <ogerlitz@mellanox.com>
[minor fix up to apply cleanly and compile fix]
Signed-off-by: Mike Christie <michaelc@cs.wisc.edu>
Signed-off-by: James Bottomley <JBottomley@Parallels.com>
2014-02-07 13:41:38 +07:00
|
|
|
spin_unlock_bh(&session->frwd_lock);
|
2010-02-20 09:32:39 +07:00
|
|
|
return SUCCESS;
|
|
|
|
}
|
|
|
|
|
|
|
|
aborted_io_task = aborted_task->dd_data;
|
|
|
|
if (!aborted_io_task->scsi_cmnd) {
|
|
|
|
/* raced or invalid command */
|
[SCSI] libiscsi: Reduce locking contention in fast path
Replace the session lock with two locks, a forward lock and
a backwards lock named frwd_lock and back_lock respectively.
The forward lock protects resources that change while sending a
request to the target, such as cmdsn, queued_cmdsn, and allocating
task from the commands' pool with kfifo_out.
The backward lock protects resources that change while processing
a response or in error path, such as cmdsn_exp, cmdsn_max, and
returning tasks to the commands' pool with kfifo_in.
Under a steady state fast-path situation, that is when one
or more processes/threads submit IO to an iscsi device and
a single kernel upcall (e.g softirq) is dealing with processing
of responses without errors, this patch eliminates the contention
between the queuecommand()/request response/scsi_done() flows
associated with iscsi sessions.
Between the forward and the backward locks exists a strict locking
hierarchy. The mutual exclusion zone protected by the forward lock can
enclose the mutual exclusion zone protected by the backward lock but not
vice versa.
For example, in iscsi_conn_teardown or in iscsi_xmit_data when there is
a failure and __iscsi_put_task is called, the backward lock is taken while
the forward lock is still taken. On the other hand, if in the RX path a nop
is to be sent, for example in iscsi_handle_reject or __iscsi_complete_pdu
than the forward lock is released and the backward lock is taken for the
duration of iscsi_send_nopout, later the backward lock is released and the
forward lock is retaken.
libiscsi_tcp uses two kernel fifos the r2t pool and the r2t queue.
The insertion and deletion from these queues didn't corespond to the
assumption taken by the new forward/backwards session locking paradigm.
That is, in iscsi_tcp_clenup_task which belongs to the RX (backwards)
path, r2t is taken out from r2t queue and inserted to the r2t pool.
In iscsi_tcp_get_curr_r2t which belong to the TX (forward) path, r2t
is also inserted to the r2t pool and another r2t is pulled from r2t
queue.
Only in iscsi_tcp_r2t_rsp which is called in the RX path but can requeue
to the TX path, r2t is taken from the r2t pool and inserted to the r2t
queue.
In order to cope with this situation, two spin locks were added,
pool2queue and queue2pool. The former protects extracting from the
r2t pool and inserting to the r2t queue, and the later protects the
extracing from the r2t queue and inserting to the r2t pool.
Signed-off-by: Shlomo Pongratz <shlomop@mellanox.com>
Signed-off-by: Or Gerlitz <ogerlitz@mellanox.com>
[minor fix up to apply cleanly and compile fix]
Signed-off-by: Mike Christie <michaelc@cs.wisc.edu>
Signed-off-by: James Bottomley <JBottomley@Parallels.com>
2014-02-07 13:41:38 +07:00
|
|
|
spin_unlock_bh(&session->frwd_lock);
|
2010-02-20 09:32:39 +07:00
|
|
|
return SUCCESS;
|
|
|
|
}
|
[SCSI] libiscsi: Reduce locking contention in fast path
Replace the session lock with two locks, a forward lock and
a backwards lock named frwd_lock and back_lock respectively.
The forward lock protects resources that change while sending a
request to the target, such as cmdsn, queued_cmdsn, and allocating
task from the commands' pool with kfifo_out.
The backward lock protects resources that change while processing
a response or in error path, such as cmdsn_exp, cmdsn_max, and
returning tasks to the commands' pool with kfifo_in.
Under a steady state fast-path situation, that is when one
or more processes/threads submit IO to an iscsi device and
a single kernel upcall (e.g softirq) is dealing with processing
of responses without errors, this patch eliminates the contention
between the queuecommand()/request response/scsi_done() flows
associated with iscsi sessions.
Between the forward and the backward locks exists a strict locking
hierarchy. The mutual exclusion zone protected by the forward lock can
enclose the mutual exclusion zone protected by the backward lock but not
vice versa.
For example, in iscsi_conn_teardown or in iscsi_xmit_data when there is
a failure and __iscsi_put_task is called, the backward lock is taken while
the forward lock is still taken. On the other hand, if in the RX path a nop
is to be sent, for example in iscsi_handle_reject or __iscsi_complete_pdu
than the forward lock is released and the backward lock is taken for the
duration of iscsi_send_nopout, later the backward lock is released and the
forward lock is retaken.
libiscsi_tcp uses two kernel fifos the r2t pool and the r2t queue.
The insertion and deletion from these queues didn't corespond to the
assumption taken by the new forward/backwards session locking paradigm.
That is, in iscsi_tcp_clenup_task which belongs to the RX (backwards)
path, r2t is taken out from r2t queue and inserted to the r2t pool.
In iscsi_tcp_get_curr_r2t which belong to the TX (forward) path, r2t
is also inserted to the r2t pool and another r2t is pulled from r2t
queue.
Only in iscsi_tcp_r2t_rsp which is called in the RX path but can requeue
to the TX path, r2t is taken from the r2t pool and inserted to the r2t
queue.
In order to cope with this situation, two spin locks were added,
pool2queue and queue2pool. The former protects extracting from the
r2t pool and inserting to the r2t queue, and the later protects the
extracing from the r2t queue and inserting to the r2t pool.
Signed-off-by: Shlomo Pongratz <shlomop@mellanox.com>
Signed-off-by: Or Gerlitz <ogerlitz@mellanox.com>
[minor fix up to apply cleanly and compile fix]
Signed-off-by: Mike Christie <michaelc@cs.wisc.edu>
Signed-off-by: James Bottomley <JBottomley@Parallels.com>
2014-02-07 13:41:38 +07:00
|
|
|
spin_unlock_bh(&session->frwd_lock);
|
2013-09-29 05:35:57 +07:00
|
|
|
/* Invalidate WRB Posted for this Task */
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_wrb, invld,
|
|
|
|
aborted_io_task->pwrb_handle->pwrb,
|
|
|
|
1);
|
|
|
|
|
2010-02-20 09:32:39 +07:00
|
|
|
conn = aborted_task->conn;
|
|
|
|
beiscsi_conn = conn->dd_data;
|
|
|
|
phba = beiscsi_conn->phba;
|
|
|
|
|
|
|
|
/* invalidate iocb */
|
|
|
|
cid = beiscsi_conn->beiscsi_conn_cid;
|
|
|
|
inv_tbl = phba->inv_tbl;
|
|
|
|
memset(inv_tbl, 0x0, sizeof(*inv_tbl));
|
|
|
|
inv_tbl->cid = cid;
|
|
|
|
inv_tbl->icd = aborted_io_task->psgl_handle->sgl_index;
|
|
|
|
num_invalidate = 1;
|
2010-07-22 05:57:47 +07:00
|
|
|
nonemb_cmd.va = pci_alloc_consistent(phba->ctrl.pdev,
|
|
|
|
sizeof(struct invalidate_commands_params_in),
|
|
|
|
&nonemb_cmd.dma);
|
|
|
|
if (nonemb_cmd.va == NULL) {
|
2012-08-21 00:30:18 +07:00
|
|
|
beiscsi_log(phba, KERN_ERR, BEISCSI_LOG_EH,
|
|
|
|
"BM_%d : Failed to allocate memory for"
|
|
|
|
"mgmt_invalidate_icds\n");
|
2010-07-22 05:57:47 +07:00
|
|
|
return FAILED;
|
|
|
|
}
|
|
|
|
nonemb_cmd.size = sizeof(struct invalidate_commands_params_in);
|
|
|
|
|
|
|
|
tag = mgmt_invalidate_icds(phba, inv_tbl, num_invalidate,
|
|
|
|
cid, &nonemb_cmd);
|
2010-02-20 09:32:39 +07:00
|
|
|
if (!tag) {
|
2012-08-21 00:30:18 +07:00
|
|
|
beiscsi_log(phba, KERN_WARNING, BEISCSI_LOG_EH,
|
|
|
|
"BM_%d : mgmt_invalidate_icds could not be"
|
|
|
|
"submitted\n");
|
2010-07-22 05:57:47 +07:00
|
|
|
pci_free_consistent(phba->ctrl.pdev, nonemb_cmd.size,
|
|
|
|
nonemb_cmd.va, nonemb_cmd.dma);
|
|
|
|
|
2010-02-20 09:32:39 +07:00
|
|
|
return FAILED;
|
|
|
|
}
|
2012-10-20 06:15:40 +07:00
|
|
|
|
2016-02-04 17:19:12 +07:00
|
|
|
rc = beiscsi_mccq_compl_wait(phba, tag, NULL, &nonemb_cmd);
|
2014-01-29 14:16:39 +07:00
|
|
|
if (rc != -EBUSY)
|
|
|
|
pci_free_consistent(phba->ctrl.pdev, nonemb_cmd.size,
|
|
|
|
nonemb_cmd.va, nonemb_cmd.dma);
|
|
|
|
|
2010-02-20 09:32:39 +07:00
|
|
|
return iscsi_eh_abort(sc);
|
|
|
|
}
|
|
|
|
|
|
|
|
static int beiscsi_eh_device_reset(struct scsi_cmnd *sc)
|
|
|
|
{
|
|
|
|
struct iscsi_task *abrt_task;
|
|
|
|
struct beiscsi_io_task *abrt_io_task;
|
|
|
|
struct iscsi_conn *conn;
|
|
|
|
struct beiscsi_conn *beiscsi_conn;
|
|
|
|
struct beiscsi_hba *phba;
|
|
|
|
struct iscsi_session *session;
|
|
|
|
struct iscsi_cls_session *cls_session;
|
|
|
|
struct invalidate_command_table *inv_tbl;
|
2010-07-22 05:57:47 +07:00
|
|
|
struct be_dma_mem nonemb_cmd;
|
2010-02-20 09:32:39 +07:00
|
|
|
unsigned int cid, tag, i, num_invalidate;
|
2014-01-29 14:16:39 +07:00
|
|
|
int rc;
|
2010-02-20 09:32:39 +07:00
|
|
|
|
|
|
|
/* invalidate iocbs */
|
|
|
|
cls_session = starget_to_session(scsi_target(sc->device));
|
|
|
|
session = cls_session->dd_data;
|
[SCSI] libiscsi: Reduce locking contention in fast path
Replace the session lock with two locks, a forward lock and
a backwards lock named frwd_lock and back_lock respectively.
The forward lock protects resources that change while sending a
request to the target, such as cmdsn, queued_cmdsn, and allocating
task from the commands' pool with kfifo_out.
The backward lock protects resources that change while processing
a response or in error path, such as cmdsn_exp, cmdsn_max, and
returning tasks to the commands' pool with kfifo_in.
Under a steady state fast-path situation, that is when one
or more processes/threads submit IO to an iscsi device and
a single kernel upcall (e.g softirq) is dealing with processing
of responses without errors, this patch eliminates the contention
between the queuecommand()/request response/scsi_done() flows
associated with iscsi sessions.
Between the forward and the backward locks exists a strict locking
hierarchy. The mutual exclusion zone protected by the forward lock can
enclose the mutual exclusion zone protected by the backward lock but not
vice versa.
For example, in iscsi_conn_teardown or in iscsi_xmit_data when there is
a failure and __iscsi_put_task is called, the backward lock is taken while
the forward lock is still taken. On the other hand, if in the RX path a nop
is to be sent, for example in iscsi_handle_reject or __iscsi_complete_pdu
than the forward lock is released and the backward lock is taken for the
duration of iscsi_send_nopout, later the backward lock is released and the
forward lock is retaken.
libiscsi_tcp uses two kernel fifos the r2t pool and the r2t queue.
The insertion and deletion from these queues didn't corespond to the
assumption taken by the new forward/backwards session locking paradigm.
That is, in iscsi_tcp_clenup_task which belongs to the RX (backwards)
path, r2t is taken out from r2t queue and inserted to the r2t pool.
In iscsi_tcp_get_curr_r2t which belong to the TX (forward) path, r2t
is also inserted to the r2t pool and another r2t is pulled from r2t
queue.
Only in iscsi_tcp_r2t_rsp which is called in the RX path but can requeue
to the TX path, r2t is taken from the r2t pool and inserted to the r2t
queue.
In order to cope with this situation, two spin locks were added,
pool2queue and queue2pool. The former protects extracting from the
r2t pool and inserting to the r2t queue, and the later protects the
extracing from the r2t queue and inserting to the r2t pool.
Signed-off-by: Shlomo Pongratz <shlomop@mellanox.com>
Signed-off-by: Or Gerlitz <ogerlitz@mellanox.com>
[minor fix up to apply cleanly and compile fix]
Signed-off-by: Mike Christie <michaelc@cs.wisc.edu>
Signed-off-by: James Bottomley <JBottomley@Parallels.com>
2014-02-07 13:41:38 +07:00
|
|
|
spin_lock_bh(&session->frwd_lock);
|
2012-04-04 11:41:43 +07:00
|
|
|
if (!session->leadconn || session->state != ISCSI_STATE_LOGGED_IN) {
|
[SCSI] libiscsi: Reduce locking contention in fast path
Replace the session lock with two locks, a forward lock and
a backwards lock named frwd_lock and back_lock respectively.
The forward lock protects resources that change while sending a
request to the target, such as cmdsn, queued_cmdsn, and allocating
task from the commands' pool with kfifo_out.
The backward lock protects resources that change while processing
a response or in error path, such as cmdsn_exp, cmdsn_max, and
returning tasks to the commands' pool with kfifo_in.
Under a steady state fast-path situation, that is when one
or more processes/threads submit IO to an iscsi device and
a single kernel upcall (e.g softirq) is dealing with processing
of responses without errors, this patch eliminates the contention
between the queuecommand()/request response/scsi_done() flows
associated with iscsi sessions.
Between the forward and the backward locks exists a strict locking
hierarchy. The mutual exclusion zone protected by the forward lock can
enclose the mutual exclusion zone protected by the backward lock but not
vice versa.
For example, in iscsi_conn_teardown or in iscsi_xmit_data when there is
a failure and __iscsi_put_task is called, the backward lock is taken while
the forward lock is still taken. On the other hand, if in the RX path a nop
is to be sent, for example in iscsi_handle_reject or __iscsi_complete_pdu
than the forward lock is released and the backward lock is taken for the
duration of iscsi_send_nopout, later the backward lock is released and the
forward lock is retaken.
libiscsi_tcp uses two kernel fifos the r2t pool and the r2t queue.
The insertion and deletion from these queues didn't corespond to the
assumption taken by the new forward/backwards session locking paradigm.
That is, in iscsi_tcp_clenup_task which belongs to the RX (backwards)
path, r2t is taken out from r2t queue and inserted to the r2t pool.
In iscsi_tcp_get_curr_r2t which belong to the TX (forward) path, r2t
is also inserted to the r2t pool and another r2t is pulled from r2t
queue.
Only in iscsi_tcp_r2t_rsp which is called in the RX path but can requeue
to the TX path, r2t is taken from the r2t pool and inserted to the r2t
queue.
In order to cope with this situation, two spin locks were added,
pool2queue and queue2pool. The former protects extracting from the
r2t pool and inserting to the r2t queue, and the later protects the
extracing from the r2t queue and inserting to the r2t pool.
Signed-off-by: Shlomo Pongratz <shlomop@mellanox.com>
Signed-off-by: Or Gerlitz <ogerlitz@mellanox.com>
[minor fix up to apply cleanly and compile fix]
Signed-off-by: Mike Christie <michaelc@cs.wisc.edu>
Signed-off-by: James Bottomley <JBottomley@Parallels.com>
2014-02-07 13:41:38 +07:00
|
|
|
spin_unlock_bh(&session->frwd_lock);
|
2012-04-04 11:41:43 +07:00
|
|
|
return FAILED;
|
|
|
|
}
|
2010-02-20 09:32:39 +07:00
|
|
|
conn = session->leadconn;
|
|
|
|
beiscsi_conn = conn->dd_data;
|
|
|
|
phba = beiscsi_conn->phba;
|
|
|
|
cid = beiscsi_conn->beiscsi_conn_cid;
|
|
|
|
inv_tbl = phba->inv_tbl;
|
|
|
|
memset(inv_tbl, 0x0, sizeof(*inv_tbl) * BE2_CMDS_PER_CXN);
|
|
|
|
num_invalidate = 0;
|
|
|
|
for (i = 0; i < conn->session->cmds_max; i++) {
|
|
|
|
abrt_task = conn->session->cmds[i];
|
|
|
|
abrt_io_task = abrt_task->dd_data;
|
|
|
|
if (!abrt_task->sc || abrt_task->state == ISCSI_TASK_FREE)
|
|
|
|
continue;
|
|
|
|
|
2013-12-19 14:16:21 +07:00
|
|
|
if (sc->device->lun != abrt_task->sc->device->lun)
|
2010-02-20 09:32:39 +07:00
|
|
|
continue;
|
|
|
|
|
2013-09-29 05:35:57 +07:00
|
|
|
/* Invalidate WRB Posted for this Task */
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_wrb, invld,
|
|
|
|
abrt_io_task->pwrb_handle->pwrb,
|
|
|
|
1);
|
|
|
|
|
2010-02-20 09:32:39 +07:00
|
|
|
inv_tbl->cid = cid;
|
|
|
|
inv_tbl->icd = abrt_io_task->psgl_handle->sgl_index;
|
|
|
|
num_invalidate++;
|
|
|
|
inv_tbl++;
|
|
|
|
}
|
[SCSI] libiscsi: Reduce locking contention in fast path
Replace the session lock with two locks, a forward lock and
a backwards lock named frwd_lock and back_lock respectively.
The forward lock protects resources that change while sending a
request to the target, such as cmdsn, queued_cmdsn, and allocating
task from the commands' pool with kfifo_out.
The backward lock protects resources that change while processing
a response or in error path, such as cmdsn_exp, cmdsn_max, and
returning tasks to the commands' pool with kfifo_in.
Under a steady state fast-path situation, that is when one
or more processes/threads submit IO to an iscsi device and
a single kernel upcall (e.g softirq) is dealing with processing
of responses without errors, this patch eliminates the contention
between the queuecommand()/request response/scsi_done() flows
associated with iscsi sessions.
Between the forward and the backward locks exists a strict locking
hierarchy. The mutual exclusion zone protected by the forward lock can
enclose the mutual exclusion zone protected by the backward lock but not
vice versa.
For example, in iscsi_conn_teardown or in iscsi_xmit_data when there is
a failure and __iscsi_put_task is called, the backward lock is taken while
the forward lock is still taken. On the other hand, if in the RX path a nop
is to be sent, for example in iscsi_handle_reject or __iscsi_complete_pdu
than the forward lock is released and the backward lock is taken for the
duration of iscsi_send_nopout, later the backward lock is released and the
forward lock is retaken.
libiscsi_tcp uses two kernel fifos the r2t pool and the r2t queue.
The insertion and deletion from these queues didn't corespond to the
assumption taken by the new forward/backwards session locking paradigm.
That is, in iscsi_tcp_clenup_task which belongs to the RX (backwards)
path, r2t is taken out from r2t queue and inserted to the r2t pool.
In iscsi_tcp_get_curr_r2t which belong to the TX (forward) path, r2t
is also inserted to the r2t pool and another r2t is pulled from r2t
queue.
Only in iscsi_tcp_r2t_rsp which is called in the RX path but can requeue
to the TX path, r2t is taken from the r2t pool and inserted to the r2t
queue.
In order to cope with this situation, two spin locks were added,
pool2queue and queue2pool. The former protects extracting from the
r2t pool and inserting to the r2t queue, and the later protects the
extracing from the r2t queue and inserting to the r2t pool.
Signed-off-by: Shlomo Pongratz <shlomop@mellanox.com>
Signed-off-by: Or Gerlitz <ogerlitz@mellanox.com>
[minor fix up to apply cleanly and compile fix]
Signed-off-by: Mike Christie <michaelc@cs.wisc.edu>
Signed-off-by: James Bottomley <JBottomley@Parallels.com>
2014-02-07 13:41:38 +07:00
|
|
|
spin_unlock_bh(&session->frwd_lock);
|
2010-02-20 09:32:39 +07:00
|
|
|
inv_tbl = phba->inv_tbl;
|
|
|
|
|
2010-07-22 05:57:47 +07:00
|
|
|
nonemb_cmd.va = pci_alloc_consistent(phba->ctrl.pdev,
|
|
|
|
sizeof(struct invalidate_commands_params_in),
|
|
|
|
&nonemb_cmd.dma);
|
|
|
|
if (nonemb_cmd.va == NULL) {
|
2012-08-21 00:30:18 +07:00
|
|
|
beiscsi_log(phba, KERN_ERR, BEISCSI_LOG_EH,
|
|
|
|
"BM_%d : Failed to allocate memory for"
|
|
|
|
"mgmt_invalidate_icds\n");
|
2010-07-22 05:57:47 +07:00
|
|
|
return FAILED;
|
|
|
|
}
|
|
|
|
nonemb_cmd.size = sizeof(struct invalidate_commands_params_in);
|
|
|
|
memset(nonemb_cmd.va, 0, nonemb_cmd.size);
|
|
|
|
tag = mgmt_invalidate_icds(phba, inv_tbl, num_invalidate,
|
|
|
|
cid, &nonemb_cmd);
|
2010-02-20 09:32:39 +07:00
|
|
|
if (!tag) {
|
2012-08-21 00:30:18 +07:00
|
|
|
beiscsi_log(phba, KERN_WARNING, BEISCSI_LOG_EH,
|
|
|
|
"BM_%d : mgmt_invalidate_icds could not be"
|
|
|
|
" submitted\n");
|
2010-07-22 05:57:47 +07:00
|
|
|
pci_free_consistent(phba->ctrl.pdev, nonemb_cmd.size,
|
|
|
|
nonemb_cmd.va, nonemb_cmd.dma);
|
2010-02-20 09:32:39 +07:00
|
|
|
return FAILED;
|
|
|
|
}
|
2012-10-20 06:15:40 +07:00
|
|
|
|
2016-02-04 17:19:12 +07:00
|
|
|
rc = beiscsi_mccq_compl_wait(phba, tag, NULL, &nonemb_cmd);
|
2014-01-29 14:16:39 +07:00
|
|
|
if (rc != -EBUSY)
|
|
|
|
pci_free_consistent(phba->ctrl.pdev, nonemb_cmd.size,
|
|
|
|
nonemb_cmd.va, nonemb_cmd.dma);
|
2010-02-20 09:32:39 +07:00
|
|
|
return iscsi_eh_device_reset(sc);
|
|
|
|
}
|
|
|
|
|
2009-10-23 13:22:33 +07:00
|
|
|
/*------------------- PCI Driver operations and data ----------------- */
|
2014-08-08 20:56:03 +07:00
|
|
|
static const struct pci_device_id beiscsi_pci_id_table[] = {
|
2009-10-23 13:22:33 +07:00
|
|
|
{ PCI_DEVICE(BE_VENDOR_ID, BE_DEVICE_ID1) },
|
2010-02-11 06:41:15 +07:00
|
|
|
{ PCI_DEVICE(BE_VENDOR_ID, BE_DEVICE_ID2) },
|
2009-10-23 13:22:33 +07:00
|
|
|
{ PCI_DEVICE(BE_VENDOR_ID, OC_DEVICE_ID1) },
|
|
|
|
{ PCI_DEVICE(BE_VENDOR_ID, OC_DEVICE_ID2) },
|
|
|
|
{ PCI_DEVICE(BE_VENDOR_ID, OC_DEVICE_ID3) },
|
2012-10-20 06:13:20 +07:00
|
|
|
{ PCI_DEVICE(ELX_VENDOR_ID, OC_SKH_ID1) },
|
2009-10-23 13:22:33 +07:00
|
|
|
{ 0 }
|
|
|
|
};
|
|
|
|
MODULE_DEVICE_TABLE(pci, beiscsi_pci_id_table);
|
|
|
|
|
2012-08-21 00:30:18 +07:00
|
|
|
|
2009-09-05 09:06:35 +07:00
|
|
|
static struct scsi_host_template beiscsi_sht = {
|
|
|
|
.module = THIS_MODULE,
|
2015-07-04 05:42:33 +07:00
|
|
|
.name = "Emulex 10Gbe open-iscsi Initiator Driver",
|
2009-09-05 09:06:35 +07:00
|
|
|
.proc_name = DRV_NAME,
|
|
|
|
.queuecommand = iscsi_queuecommand,
|
2014-11-13 21:08:42 +07:00
|
|
|
.change_queue_depth = scsi_change_queue_depth,
|
2009-09-05 09:06:35 +07:00
|
|
|
.slave_configure = beiscsi_slave_configure,
|
|
|
|
.target_alloc = iscsi_target_alloc,
|
2010-02-20 09:32:39 +07:00
|
|
|
.eh_abort_handler = beiscsi_eh_abort,
|
|
|
|
.eh_device_reset_handler = beiscsi_eh_device_reset,
|
2010-02-20 09:32:10 +07:00
|
|
|
.eh_target_reset_handler = iscsi_eh_session_reset,
|
2012-08-21 00:30:18 +07:00
|
|
|
.shost_attrs = beiscsi_attrs,
|
2009-09-05 09:06:35 +07:00
|
|
|
.sg_tablesize = BEISCSI_SGLIST_ELEMENTS,
|
|
|
|
.can_queue = BE2_IO_DEPTH,
|
|
|
|
.this_id = -1,
|
|
|
|
.max_sectors = BEISCSI_MAX_SECTORS,
|
|
|
|
.cmd_per_lun = BEISCSI_CMD_PER_LUN,
|
|
|
|
.use_clustering = ENABLE_CLUSTERING,
|
2012-04-04 11:41:50 +07:00
|
|
|
.vendor_id = SCSI_NL_VID_TYPE_PCI | BE_VENDOR_ID,
|
2014-11-13 20:25:11 +07:00
|
|
|
.track_queue_depth = 1,
|
2009-09-05 09:06:35 +07:00
|
|
|
};
|
|
|
|
|
2009-10-23 13:22:33 +07:00
|
|
|
static struct scsi_transport_template *beiscsi_scsi_transport;
|
2009-09-05 09:06:35 +07:00
|
|
|
|
|
|
|
static struct beiscsi_hba *beiscsi_hba_alloc(struct pci_dev *pcidev)
|
|
|
|
{
|
|
|
|
struct beiscsi_hba *phba;
|
|
|
|
struct Scsi_Host *shost;
|
|
|
|
|
|
|
|
shost = iscsi_host_alloc(&beiscsi_sht, sizeof(*phba), 0);
|
|
|
|
if (!shost) {
|
2012-08-21 00:30:18 +07:00
|
|
|
dev_err(&pcidev->dev,
|
|
|
|
"beiscsi_hba_alloc - iscsi_host_alloc failed\n");
|
2009-09-05 09:06:35 +07:00
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
shost->max_id = BE2_MAX_SESSIONS;
|
|
|
|
shost->max_channel = 0;
|
|
|
|
shost->max_cmd_len = BEISCSI_MAX_CMD_LEN;
|
|
|
|
shost->max_lun = BEISCSI_NUM_MAX_LUN;
|
|
|
|
shost->transportt = beiscsi_scsi_transport;
|
|
|
|
phba = iscsi_host_priv(shost);
|
|
|
|
memset(phba, 0, sizeof(*phba));
|
|
|
|
phba->shost = shost;
|
|
|
|
phba->pcidev = pci_dev_get(pcidev);
|
2010-01-05 06:37:49 +07:00
|
|
|
pci_set_drvdata(pcidev, phba);
|
2012-04-04 11:41:51 +07:00
|
|
|
phba->interface_handle = 0xFFFFFFFF;
|
2009-09-05 09:06:35 +07:00
|
|
|
|
|
|
|
return phba;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void beiscsi_unmap_pci_function(struct beiscsi_hba *phba)
|
|
|
|
{
|
|
|
|
if (phba->csr_va) {
|
|
|
|
iounmap(phba->csr_va);
|
|
|
|
phba->csr_va = NULL;
|
|
|
|
}
|
|
|
|
if (phba->db_va) {
|
|
|
|
iounmap(phba->db_va);
|
|
|
|
phba->db_va = NULL;
|
|
|
|
}
|
|
|
|
if (phba->pci_va) {
|
|
|
|
iounmap(phba->pci_va);
|
|
|
|
phba->pci_va = NULL;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
static int beiscsi_map_pci_bars(struct beiscsi_hba *phba,
|
|
|
|
struct pci_dev *pcidev)
|
|
|
|
{
|
|
|
|
u8 __iomem *addr;
|
2010-02-11 06:41:15 +07:00
|
|
|
int pcicfg_reg;
|
2009-09-05 09:06:35 +07:00
|
|
|
|
|
|
|
addr = ioremap_nocache(pci_resource_start(pcidev, 2),
|
|
|
|
pci_resource_len(pcidev, 2));
|
|
|
|
if (addr == NULL)
|
|
|
|
return -ENOMEM;
|
|
|
|
phba->ctrl.csr = addr;
|
|
|
|
phba->csr_va = addr;
|
|
|
|
phba->csr_pa.u.a64.address = pci_resource_start(pcidev, 2);
|
|
|
|
|
|
|
|
addr = ioremap_nocache(pci_resource_start(pcidev, 4), 128 * 1024);
|
|
|
|
if (addr == NULL)
|
|
|
|
goto pci_map_err;
|
|
|
|
phba->ctrl.db = addr;
|
|
|
|
phba->db_va = addr;
|
|
|
|
phba->db_pa.u.a64.address = pci_resource_start(pcidev, 4);
|
|
|
|
|
2010-02-11 06:41:15 +07:00
|
|
|
if (phba->generation == BE_GEN2)
|
|
|
|
pcicfg_reg = 1;
|
|
|
|
else
|
|
|
|
pcicfg_reg = 0;
|
|
|
|
|
|
|
|
addr = ioremap_nocache(pci_resource_start(pcidev, pcicfg_reg),
|
|
|
|
pci_resource_len(pcidev, pcicfg_reg));
|
|
|
|
|
2009-09-05 09:06:35 +07:00
|
|
|
if (addr == NULL)
|
|
|
|
goto pci_map_err;
|
|
|
|
phba->ctrl.pcicfg = addr;
|
|
|
|
phba->pci_va = addr;
|
2010-02-11 06:41:15 +07:00
|
|
|
phba->pci_pa.u.a64.address = pci_resource_start(pcidev, pcicfg_reg);
|
2009-09-05 09:06:35 +07:00
|
|
|
return 0;
|
|
|
|
|
|
|
|
pci_map_err:
|
|
|
|
beiscsi_unmap_pci_function(phba);
|
|
|
|
return -ENOMEM;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int beiscsi_enable_pci(struct pci_dev *pcidev)
|
|
|
|
{
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
ret = pci_enable_device(pcidev);
|
|
|
|
if (ret) {
|
2012-08-21 00:30:18 +07:00
|
|
|
dev_err(&pcidev->dev,
|
|
|
|
"beiscsi_enable_pci - enable device failed\n");
|
2009-09-05 09:06:35 +07:00
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
2015-04-25 09:47:19 +07:00
|
|
|
ret = pci_request_regions(pcidev, DRV_NAME);
|
|
|
|
if (ret) {
|
|
|
|
dev_err(&pcidev->dev,
|
|
|
|
"beiscsi_enable_pci - request region failed\n");
|
|
|
|
goto pci_dev_disable;
|
|
|
|
}
|
|
|
|
|
2009-10-23 13:22:33 +07:00
|
|
|
pci_set_master(pcidev);
|
2014-01-29 14:16:45 +07:00
|
|
|
ret = pci_set_dma_mask(pcidev, DMA_BIT_MASK(64));
|
|
|
|
if (ret) {
|
|
|
|
ret = pci_set_dma_mask(pcidev, DMA_BIT_MASK(32));
|
|
|
|
if (ret) {
|
|
|
|
dev_err(&pcidev->dev, "Could not set PCI DMA Mask\n");
|
2015-04-25 09:47:19 +07:00
|
|
|
goto pci_region_release;
|
2014-01-29 14:16:45 +07:00
|
|
|
} else {
|
|
|
|
ret = pci_set_consistent_dma_mask(pcidev,
|
|
|
|
DMA_BIT_MASK(32));
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
ret = pci_set_consistent_dma_mask(pcidev, DMA_BIT_MASK(64));
|
2009-09-05 09:06:35 +07:00
|
|
|
if (ret) {
|
|
|
|
dev_err(&pcidev->dev, "Could not set PCI DMA Mask\n");
|
2015-04-25 09:47:19 +07:00
|
|
|
goto pci_region_release;
|
2009-09-05 09:06:35 +07:00
|
|
|
}
|
|
|
|
}
|
|
|
|
return 0;
|
2015-04-25 09:47:19 +07:00
|
|
|
|
|
|
|
pci_region_release:
|
|
|
|
pci_release_regions(pcidev);
|
|
|
|
pci_dev_disable:
|
|
|
|
pci_disable_device(pcidev);
|
|
|
|
|
|
|
|
return ret;
|
2009-09-05 09:06:35 +07:00
|
|
|
}
|
|
|
|
|
|
|
|
static int be_ctrl_init(struct beiscsi_hba *phba, struct pci_dev *pdev)
|
|
|
|
{
|
|
|
|
struct be_ctrl_info *ctrl = &phba->ctrl;
|
|
|
|
struct be_dma_mem *mbox_mem_alloc = &ctrl->mbox_mem_alloced;
|
|
|
|
struct be_dma_mem *mbox_mem_align = &ctrl->mbox_mem;
|
|
|
|
int status = 0;
|
|
|
|
|
|
|
|
ctrl->pdev = pdev;
|
|
|
|
status = beiscsi_map_pci_bars(phba, pdev);
|
|
|
|
if (status)
|
|
|
|
return status;
|
|
|
|
mbox_mem_alloc->size = sizeof(struct be_mcc_mailbox) + 16;
|
|
|
|
mbox_mem_alloc->va = pci_alloc_consistent(pdev,
|
|
|
|
mbox_mem_alloc->size,
|
|
|
|
&mbox_mem_alloc->dma);
|
|
|
|
if (!mbox_mem_alloc->va) {
|
|
|
|
beiscsi_unmap_pci_function(phba);
|
2012-04-04 11:41:44 +07:00
|
|
|
return -ENOMEM;
|
2009-09-05 09:06:35 +07:00
|
|
|
}
|
|
|
|
|
|
|
|
mbox_mem_align->size = sizeof(struct be_mcc_mailbox);
|
|
|
|
mbox_mem_align->va = PTR_ALIGN(mbox_mem_alloc->va, 16);
|
|
|
|
mbox_mem_align->dma = PTR_ALIGN(mbox_mem_alloc->dma, 16);
|
|
|
|
memset(mbox_mem_align->va, 0, sizeof(struct be_mcc_mailbox));
|
2016-01-20 15:40:46 +07:00
|
|
|
mutex_init(&ctrl->mbox_lock);
|
2009-10-23 13:22:33 +07:00
|
|
|
spin_lock_init(&phba->ctrl.mcc_lock);
|
|
|
|
|
2009-09-05 09:06:35 +07:00
|
|
|
return status;
|
|
|
|
}
|
|
|
|
|
2013-09-29 05:35:44 +07:00
|
|
|
/**
|
|
|
|
* beiscsi_get_params()- Set the config paramters
|
|
|
|
* @phba: ptr device priv structure
|
|
|
|
**/
|
2009-09-05 09:06:35 +07:00
|
|
|
static void beiscsi_get_params(struct beiscsi_hba *phba)
|
|
|
|
{
|
2013-09-29 05:35:44 +07:00
|
|
|
uint32_t total_cid_count = 0;
|
|
|
|
uint32_t total_icd_count = 0;
|
|
|
|
uint8_t ulp_num = 0;
|
|
|
|
|
|
|
|
total_cid_count = BEISCSI_GET_CID_COUNT(phba, BEISCSI_ULP0) +
|
|
|
|
BEISCSI_GET_CID_COUNT(phba, BEISCSI_ULP1);
|
|
|
|
|
2013-09-29 05:35:59 +07:00
|
|
|
for (ulp_num = 0; ulp_num < BEISCSI_ULP_COUNT; ulp_num++) {
|
|
|
|
uint32_t align_mask = 0;
|
|
|
|
uint32_t icd_post_per_page = 0;
|
|
|
|
uint32_t icd_count_unavailable = 0;
|
|
|
|
uint32_t icd_start = 0, icd_count = 0;
|
|
|
|
uint32_t icd_start_align = 0, icd_count_align = 0;
|
|
|
|
|
2013-09-29 05:35:44 +07:00
|
|
|
if (test_bit(ulp_num, &phba->fw_config.ulp_supported)) {
|
2013-09-29 05:35:59 +07:00
|
|
|
icd_start = phba->fw_config.iscsi_icd_start[ulp_num];
|
|
|
|
icd_count = phba->fw_config.iscsi_icd_count[ulp_num];
|
|
|
|
|
|
|
|
/* Get ICD count that can be posted on each page */
|
|
|
|
icd_post_per_page = (PAGE_SIZE / (BE2_SGE *
|
|
|
|
sizeof(struct iscsi_sge)));
|
|
|
|
align_mask = (icd_post_per_page - 1);
|
|
|
|
|
|
|
|
/* Check if icd_start is aligned ICD per page posting */
|
|
|
|
if (icd_start % icd_post_per_page) {
|
|
|
|
icd_start_align = ((icd_start +
|
|
|
|
icd_post_per_page) &
|
|
|
|
~(align_mask));
|
|
|
|
phba->fw_config.
|
|
|
|
iscsi_icd_start[ulp_num] =
|
|
|
|
icd_start_align;
|
|
|
|
}
|
|
|
|
|
|
|
|
icd_count_align = (icd_count & ~align_mask);
|
|
|
|
|
|
|
|
/* ICD discarded in the process of alignment */
|
|
|
|
if (icd_start_align)
|
|
|
|
icd_count_unavailable = ((icd_start_align -
|
|
|
|
icd_start) +
|
|
|
|
(icd_count -
|
|
|
|
icd_count_align));
|
|
|
|
|
|
|
|
/* Updated ICD count available */
|
|
|
|
phba->fw_config.iscsi_icd_count[ulp_num] = (icd_count -
|
|
|
|
icd_count_unavailable);
|
|
|
|
|
|
|
|
beiscsi_log(phba, KERN_INFO, BEISCSI_LOG_INIT,
|
|
|
|
"BM_%d : Aligned ICD values\n"
|
|
|
|
"\t ICD Start : %d\n"
|
|
|
|
"\t ICD Count : %d\n"
|
|
|
|
"\t ICD Discarded : %d\n",
|
|
|
|
phba->fw_config.
|
|
|
|
iscsi_icd_start[ulp_num],
|
|
|
|
phba->fw_config.
|
|
|
|
iscsi_icd_count[ulp_num],
|
|
|
|
icd_count_unavailable);
|
2013-09-29 05:35:44 +07:00
|
|
|
break;
|
|
|
|
}
|
2013-09-29 05:35:59 +07:00
|
|
|
}
|
2013-09-29 05:35:44 +07:00
|
|
|
|
2013-09-29 05:35:59 +07:00
|
|
|
total_icd_count = phba->fw_config.iscsi_icd_count[ulp_num];
|
2013-09-29 05:35:44 +07:00
|
|
|
phba->params.ios_per_ctrl = (total_icd_count -
|
|
|
|
(total_cid_count +
|
|
|
|
BE2_TMFS + BE2_NOPOUT_REQ));
|
|
|
|
phba->params.cxns_per_ctrl = total_cid_count;
|
|
|
|
phba->params.asyncpdus_per_ctrl = total_cid_count;
|
|
|
|
phba->params.icds_per_ctrl = total_icd_count;
|
2009-09-05 09:06:35 +07:00
|
|
|
phba->params.num_sge_per_io = BE2_SGE;
|
|
|
|
phba->params.defpdu_hdr_sz = BE2_DEFPDU_HDR_SZ;
|
|
|
|
phba->params.defpdu_data_sz = BE2_DEFPDU_DATA_SZ;
|
|
|
|
phba->params.eq_timer = 64;
|
2013-09-29 05:35:44 +07:00
|
|
|
phba->params.num_eq_entries = 1024;
|
|
|
|
phba->params.num_cq_entries = 1024;
|
2009-09-05 09:06:35 +07:00
|
|
|
phba->params.wrbs_per_cxn = 256;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void hwi_ring_eq_db(struct beiscsi_hba *phba,
|
|
|
|
unsigned int id, unsigned int clr_interrupt,
|
|
|
|
unsigned int num_processed,
|
|
|
|
unsigned char rearm, unsigned char event)
|
|
|
|
{
|
|
|
|
u32 val = 0;
|
2014-01-29 14:16:42 +07:00
|
|
|
|
2009-09-05 09:06:35 +07:00
|
|
|
if (rearm)
|
|
|
|
val |= 1 << DB_EQ_REARM_SHIFT;
|
|
|
|
if (clr_interrupt)
|
|
|
|
val |= 1 << DB_EQ_CLR_SHIFT;
|
|
|
|
if (event)
|
|
|
|
val |= 1 << DB_EQ_EVNT_SHIFT;
|
2014-01-29 14:16:42 +07:00
|
|
|
|
2009-09-05 09:06:35 +07:00
|
|
|
val |= num_processed << DB_EQ_NUM_POPPED_SHIFT;
|
2014-01-29 14:16:42 +07:00
|
|
|
/* Setting lower order EQ_ID Bits */
|
|
|
|
val |= (id & DB_EQ_RING_ID_LOW_MASK);
|
|
|
|
|
|
|
|
/* Setting Higher order EQ_ID Bits */
|
|
|
|
val |= (((id >> DB_EQ_HIGH_FEILD_SHIFT) &
|
|
|
|
DB_EQ_RING_ID_HIGH_MASK)
|
|
|
|
<< DB_EQ_HIGH_SET_SHIFT);
|
|
|
|
|
2009-09-05 09:06:35 +07:00
|
|
|
iowrite32(val, phba->db_va + DB_EQ_OFFSET);
|
|
|
|
}
|
|
|
|
|
2009-10-23 13:22:33 +07:00
|
|
|
/**
|
|
|
|
* be_isr_mcc - The isr routine of the driver.
|
|
|
|
* @irq: Not used
|
|
|
|
* @dev_id: Pointer to host adapter structure
|
|
|
|
*/
|
|
|
|
static irqreturn_t be_isr_mcc(int irq, void *dev_id)
|
|
|
|
{
|
|
|
|
struct beiscsi_hba *phba;
|
2016-08-19 16:50:10 +07:00
|
|
|
struct be_eq_entry *eqe;
|
2009-10-23 13:22:33 +07:00
|
|
|
struct be_queue_info *eq;
|
|
|
|
struct be_queue_info *mcc;
|
2016-08-19 16:50:10 +07:00
|
|
|
unsigned int mcc_events;
|
2009-10-23 13:22:33 +07:00
|
|
|
struct be_eq_obj *pbe_eq;
|
|
|
|
|
|
|
|
pbe_eq = dev_id;
|
|
|
|
eq = &pbe_eq->q;
|
|
|
|
phba = pbe_eq->phba;
|
|
|
|
mcc = &phba->ctrl.mcc_obj.cq;
|
|
|
|
eqe = queue_tail_node(eq);
|
|
|
|
|
2016-08-19 16:50:10 +07:00
|
|
|
mcc_events = 0;
|
2009-10-23 13:22:33 +07:00
|
|
|
while (eqe->dw[offsetof(struct amap_eq_entry, valid) / 32]
|
|
|
|
& EQE_VALID_MASK) {
|
|
|
|
if (((eqe->dw[offsetof(struct amap_eq_entry,
|
|
|
|
resource_id) / 32] &
|
|
|
|
EQE_RESID_MASK) >> 16) == mcc->id) {
|
2016-08-19 16:50:10 +07:00
|
|
|
mcc_events++;
|
2009-10-23 13:22:33 +07:00
|
|
|
}
|
|
|
|
AMAP_SET_BITS(struct amap_eq_entry, valid, eqe, 0);
|
|
|
|
queue_tail_inc(eq);
|
|
|
|
eqe = queue_tail_node(eq);
|
|
|
|
}
|
|
|
|
|
2016-08-19 16:50:10 +07:00
|
|
|
if (mcc_events) {
|
|
|
|
queue_work(phba->wq, &pbe_eq->mcc_work);
|
|
|
|
hwi_ring_eq_db(phba, eq->id, 1, mcc_events, 1, 1);
|
|
|
|
}
|
2009-10-23 13:22:33 +07:00
|
|
|
return IRQ_HANDLED;
|
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
|
|
|
* be_isr_msix - The isr routine of the driver.
|
|
|
|
* @irq: Not used
|
|
|
|
* @dev_id: Pointer to host adapter structure
|
|
|
|
*/
|
|
|
|
static irqreturn_t be_isr_msix(int irq, void *dev_id)
|
|
|
|
{
|
|
|
|
struct beiscsi_hba *phba;
|
|
|
|
struct be_queue_info *eq;
|
|
|
|
struct be_eq_obj *pbe_eq;
|
|
|
|
|
|
|
|
pbe_eq = dev_id;
|
|
|
|
eq = &pbe_eq->q;
|
|
|
|
|
|
|
|
phba = pbe_eq->phba;
|
2016-01-20 15:40:56 +07:00
|
|
|
/* disable interrupt till iopoll completes */
|
|
|
|
hwi_ring_eq_db(phba, eq->id, 1, 0, 0, 1);
|
|
|
|
irq_poll_sched(&pbe_eq->iopoll);
|
2012-10-20 06:12:49 +07:00
|
|
|
|
|
|
|
return IRQ_HANDLED;
|
2009-10-23 13:22:33 +07:00
|
|
|
}
|
|
|
|
|
2009-09-05 09:06:35 +07:00
|
|
|
/**
|
|
|
|
* be_isr - The isr routine of the driver.
|
|
|
|
* @irq: Not used
|
|
|
|
* @dev_id: Pointer to host adapter structure
|
|
|
|
*/
|
|
|
|
static irqreturn_t be_isr(int irq, void *dev_id)
|
|
|
|
{
|
|
|
|
struct beiscsi_hba *phba;
|
|
|
|
struct hwi_controller *phwi_ctrlr;
|
|
|
|
struct hwi_context_memory *phwi_context;
|
2016-08-19 16:50:10 +07:00
|
|
|
struct be_eq_entry *eqe;
|
2009-09-05 09:06:35 +07:00
|
|
|
struct be_queue_info *eq;
|
2009-10-23 13:22:33 +07:00
|
|
|
struct be_queue_info *mcc;
|
2016-08-19 16:50:10 +07:00
|
|
|
unsigned int mcc_events, io_events;
|
2009-09-05 09:06:35 +07:00
|
|
|
struct be_ctrl_info *ctrl;
|
2009-10-23 13:22:33 +07:00
|
|
|
struct be_eq_obj *pbe_eq;
|
2016-08-19 16:50:10 +07:00
|
|
|
int isr, rearm;
|
2009-09-05 09:06:35 +07:00
|
|
|
|
|
|
|
phba = dev_id;
|
2011-04-09 09:49:08 +07:00
|
|
|
ctrl = &phba->ctrl;
|
2009-10-23 13:22:33 +07:00
|
|
|
isr = ioread32(ctrl->csr + CEV_ISR0_OFFSET +
|
|
|
|
(PCI_FUNC(ctrl->pdev->devfn) * CEV_ISR_SIZE));
|
|
|
|
if (!isr)
|
|
|
|
return IRQ_NONE;
|
2009-09-05 09:06:35 +07:00
|
|
|
|
|
|
|
phwi_ctrlr = phba->phwi_ctrlr;
|
|
|
|
phwi_context = phwi_ctrlr->phwi_ctxt;
|
2009-10-23 13:22:33 +07:00
|
|
|
pbe_eq = &phwi_context->be_eq[0];
|
|
|
|
|
|
|
|
eq = &phwi_context->be_eq[0].q;
|
|
|
|
mcc = &phba->ctrl.mcc_obj.cq;
|
2009-09-05 09:06:35 +07:00
|
|
|
eqe = queue_tail_node(eq);
|
|
|
|
|
2016-08-19 16:50:10 +07:00
|
|
|
io_events = 0;
|
|
|
|
mcc_events = 0;
|
2014-03-13 22:38:42 +07:00
|
|
|
while (eqe->dw[offsetof(struct amap_eq_entry, valid) / 32]
|
|
|
|
& EQE_VALID_MASK) {
|
|
|
|
if (((eqe->dw[offsetof(struct amap_eq_entry,
|
2016-08-19 16:50:10 +07:00
|
|
|
resource_id) / 32] & EQE_RESID_MASK) >> 16) == mcc->id)
|
|
|
|
mcc_events++;
|
|
|
|
else
|
|
|
|
io_events++;
|
2014-03-13 22:38:42 +07:00
|
|
|
AMAP_SET_BITS(struct amap_eq_entry, valid, eqe, 0);
|
|
|
|
queue_tail_inc(eq);
|
|
|
|
eqe = queue_tail_node(eq);
|
|
|
|
}
|
2016-08-19 16:50:10 +07:00
|
|
|
if (!io_events && !mcc_events)
|
2014-03-13 22:38:42 +07:00
|
|
|
return IRQ_NONE;
|
2016-08-19 16:50:10 +07:00
|
|
|
|
|
|
|
/* no need to rearm if interrupt is only for IOs */
|
|
|
|
rearm = 0;
|
|
|
|
if (mcc_events) {
|
|
|
|
queue_work(phba->wq, &pbe_eq->mcc_work);
|
|
|
|
/* rearm for MCCQ */
|
|
|
|
rearm = 1;
|
|
|
|
}
|
|
|
|
if (io_events)
|
|
|
|
irq_poll_sched(&pbe_eq->iopoll);
|
|
|
|
hwi_ring_eq_db(phba, eq->id, 0, (io_events + mcc_events), rearm, 1);
|
|
|
|
return IRQ_HANDLED;
|
2009-09-05 09:06:35 +07:00
|
|
|
}
|
|
|
|
|
2016-01-20 15:40:56 +07:00
|
|
|
|
2009-09-05 09:06:35 +07:00
|
|
|
static int beiscsi_init_irqs(struct beiscsi_hba *phba)
|
|
|
|
{
|
|
|
|
struct pci_dev *pcidev = phba->pcidev;
|
2009-10-23 13:22:33 +07:00
|
|
|
struct hwi_controller *phwi_ctrlr;
|
|
|
|
struct hwi_context_memory *phwi_context;
|
2010-07-22 05:53:55 +07:00
|
|
|
int ret, msix_vec, i, j;
|
2009-09-05 09:06:35 +07:00
|
|
|
|
2009-10-23 13:22:33 +07:00
|
|
|
phwi_ctrlr = phba->phwi_ctrlr;
|
|
|
|
phwi_context = phwi_ctrlr->phwi_ctxt;
|
|
|
|
|
|
|
|
if (phba->msix_enabled) {
|
|
|
|
for (i = 0; i < phba->num_cpus; i++) {
|
2011-08-25 06:05:30 +07:00
|
|
|
phba->msi_name[i] = kzalloc(BEISCSI_MSI_NAME,
|
|
|
|
GFP_KERNEL);
|
|
|
|
if (!phba->msi_name[i]) {
|
|
|
|
ret = -ENOMEM;
|
|
|
|
goto free_msix_irqs;
|
|
|
|
}
|
|
|
|
|
|
|
|
sprintf(phba->msi_name[i], "beiscsi_%02x_%02x",
|
|
|
|
phba->shost->host_no, i);
|
2009-10-23 13:22:33 +07:00
|
|
|
msix_vec = phba->msix_entries[i].vector;
|
2011-08-25 06:05:30 +07:00
|
|
|
ret = request_irq(msix_vec, be_isr_msix, 0,
|
|
|
|
phba->msi_name[i],
|
2009-10-23 13:22:33 +07:00
|
|
|
&phwi_context->be_eq[i]);
|
2010-07-22 05:53:55 +07:00
|
|
|
if (ret) {
|
2012-08-21 00:30:18 +07:00
|
|
|
beiscsi_log(phba, KERN_ERR, BEISCSI_LOG_INIT,
|
|
|
|
"BM_%d : beiscsi_init_irqs-Failed to"
|
|
|
|
"register msix for i = %d\n",
|
|
|
|
i);
|
2011-08-25 06:05:30 +07:00
|
|
|
kfree(phba->msi_name[i]);
|
2010-07-22 05:53:55 +07:00
|
|
|
goto free_msix_irqs;
|
|
|
|
}
|
2009-10-23 13:22:33 +07:00
|
|
|
}
|
2011-08-25 06:05:30 +07:00
|
|
|
phba->msi_name[i] = kzalloc(BEISCSI_MSI_NAME, GFP_KERNEL);
|
|
|
|
if (!phba->msi_name[i]) {
|
|
|
|
ret = -ENOMEM;
|
|
|
|
goto free_msix_irqs;
|
|
|
|
}
|
|
|
|
sprintf(phba->msi_name[i], "beiscsi_mcc_%02x",
|
|
|
|
phba->shost->host_no);
|
2009-10-23 13:22:33 +07:00
|
|
|
msix_vec = phba->msix_entries[i].vector;
|
2011-08-25 06:05:30 +07:00
|
|
|
ret = request_irq(msix_vec, be_isr_mcc, 0, phba->msi_name[i],
|
2009-10-23 13:22:33 +07:00
|
|
|
&phwi_context->be_eq[i]);
|
2010-07-22 05:53:55 +07:00
|
|
|
if (ret) {
|
2012-08-21 00:30:18 +07:00
|
|
|
beiscsi_log(phba, KERN_ERR, BEISCSI_LOG_INIT ,
|
|
|
|
"BM_%d : beiscsi_init_irqs-"
|
|
|
|
"Failed to register beiscsi_msix_mcc\n");
|
2011-08-25 06:05:30 +07:00
|
|
|
kfree(phba->msi_name[i]);
|
2010-07-22 05:53:55 +07:00
|
|
|
goto free_msix_irqs;
|
|
|
|
}
|
|
|
|
|
2009-10-23 13:22:33 +07:00
|
|
|
} else {
|
|
|
|
ret = request_irq(pcidev->irq, be_isr, IRQF_SHARED,
|
|
|
|
"beiscsi", phba);
|
|
|
|
if (ret) {
|
2012-08-21 00:30:18 +07:00
|
|
|
beiscsi_log(phba, KERN_ERR, BEISCSI_LOG_INIT,
|
|
|
|
"BM_%d : beiscsi_init_irqs-"
|
|
|
|
"Failed to register irq\\n");
|
2009-10-23 13:22:33 +07:00
|
|
|
return ret;
|
|
|
|
}
|
2009-09-05 09:06:35 +07:00
|
|
|
}
|
|
|
|
return 0;
|
2010-07-22 05:53:55 +07:00
|
|
|
free_msix_irqs:
|
2011-08-25 06:05:30 +07:00
|
|
|
for (j = i - 1; j >= 0; j--) {
|
|
|
|
kfree(phba->msi_name[j]);
|
|
|
|
msix_vec = phba->msix_entries[j].vector;
|
2010-07-22 05:53:55 +07:00
|
|
|
free_irq(msix_vec, &phwi_context->be_eq[j]);
|
2011-08-25 06:05:30 +07:00
|
|
|
}
|
2010-07-22 05:53:55 +07:00
|
|
|
return ret;
|
2009-09-05 09:06:35 +07:00
|
|
|
}
|
|
|
|
|
2014-01-29 14:16:42 +07:00
|
|
|
void hwi_ring_cq_db(struct beiscsi_hba *phba,
|
2009-09-05 09:06:35 +07:00
|
|
|
unsigned int id, unsigned int num_processed,
|
2016-01-20 15:40:56 +07:00
|
|
|
unsigned char rearm)
|
2009-09-05 09:06:35 +07:00
|
|
|
{
|
|
|
|
u32 val = 0;
|
2014-01-29 14:16:42 +07:00
|
|
|
|
2009-09-05 09:06:35 +07:00
|
|
|
if (rearm)
|
|
|
|
val |= 1 << DB_CQ_REARM_SHIFT;
|
2014-01-29 14:16:42 +07:00
|
|
|
|
2009-09-05 09:06:35 +07:00
|
|
|
val |= num_processed << DB_CQ_NUM_POPPED_SHIFT;
|
2014-01-29 14:16:42 +07:00
|
|
|
|
|
|
|
/* Setting lower order CQ_ID Bits */
|
|
|
|
val |= (id & DB_CQ_RING_ID_LOW_MASK);
|
|
|
|
|
|
|
|
/* Setting Higher order CQ_ID Bits */
|
|
|
|
val |= (((id >> DB_CQ_HIGH_FEILD_SHIFT) &
|
|
|
|
DB_CQ_RING_ID_HIGH_MASK)
|
|
|
|
<< DB_CQ_HIGH_SET_SHIFT);
|
|
|
|
|
2009-09-05 09:06:35 +07:00
|
|
|
iowrite32(val, phba->db_va + DB_CQ_OFFSET);
|
|
|
|
}
|
|
|
|
|
|
|
|
static unsigned int
|
|
|
|
beiscsi_process_async_pdu(struct beiscsi_conn *beiscsi_conn,
|
|
|
|
struct beiscsi_hba *phba,
|
|
|
|
struct pdu_base *ppdu,
|
|
|
|
unsigned long pdu_len,
|
|
|
|
void *pbuffer, unsigned long buf_len)
|
|
|
|
{
|
|
|
|
struct iscsi_conn *conn = beiscsi_conn->conn;
|
|
|
|
struct iscsi_session *session = conn->session;
|
2009-10-23 13:22:33 +07:00
|
|
|
struct iscsi_task *task;
|
|
|
|
struct beiscsi_io_task *io_task;
|
|
|
|
struct iscsi_hdr *login_hdr;
|
2009-09-05 09:06:35 +07:00
|
|
|
|
|
|
|
switch (ppdu->dw[offsetof(struct amap_pdu_base, opcode) / 32] &
|
|
|
|
PDUBASE_OPCODE_MASK) {
|
|
|
|
case ISCSI_OP_NOOP_IN:
|
|
|
|
pbuffer = NULL;
|
|
|
|
buf_len = 0;
|
|
|
|
break;
|
|
|
|
case ISCSI_OP_ASYNC_EVENT:
|
|
|
|
break;
|
|
|
|
case ISCSI_OP_REJECT:
|
|
|
|
WARN_ON(!pbuffer);
|
|
|
|
WARN_ON(!(buf_len == 48));
|
2012-08-21 00:30:18 +07:00
|
|
|
beiscsi_log(phba, KERN_ERR,
|
|
|
|
BEISCSI_LOG_CONFIG | BEISCSI_LOG_IO,
|
|
|
|
"BM_%d : In ISCSI_OP_REJECT\n");
|
2009-09-05 09:06:35 +07:00
|
|
|
break;
|
|
|
|
case ISCSI_OP_LOGIN_RSP:
|
2010-01-05 06:37:02 +07:00
|
|
|
case ISCSI_OP_TEXT_RSP:
|
2009-10-23 13:22:33 +07:00
|
|
|
task = conn->login_task;
|
|
|
|
io_task = task->dd_data;
|
|
|
|
login_hdr = (struct iscsi_hdr *)ppdu;
|
|
|
|
login_hdr->itt = io_task->libiscsi_itt;
|
2009-09-05 09:06:35 +07:00
|
|
|
break;
|
|
|
|
default:
|
2012-08-21 00:30:18 +07:00
|
|
|
beiscsi_log(phba, KERN_WARNING,
|
|
|
|
BEISCSI_LOG_IO | BEISCSI_LOG_CONFIG,
|
|
|
|
"BM_%d : Unrecognized opcode 0x%x in async msg\n",
|
|
|
|
(ppdu->
|
2009-09-05 09:06:35 +07:00
|
|
|
dw[offsetof(struct amap_pdu_base, opcode) / 32]
|
2012-08-21 00:30:18 +07:00
|
|
|
& PDUBASE_OPCODE_MASK));
|
2009-09-05 09:06:35 +07:00
|
|
|
return 1;
|
|
|
|
}
|
|
|
|
|
[SCSI] libiscsi: Reduce locking contention in fast path
Replace the session lock with two locks, a forward lock and
a backwards lock named frwd_lock and back_lock respectively.
The forward lock protects resources that change while sending a
request to the target, such as cmdsn, queued_cmdsn, and allocating
task from the commands' pool with kfifo_out.
The backward lock protects resources that change while processing
a response or in error path, such as cmdsn_exp, cmdsn_max, and
returning tasks to the commands' pool with kfifo_in.
Under a steady state fast-path situation, that is when one
or more processes/threads submit IO to an iscsi device and
a single kernel upcall (e.g softirq) is dealing with processing
of responses without errors, this patch eliminates the contention
between the queuecommand()/request response/scsi_done() flows
associated with iscsi sessions.
Between the forward and the backward locks exists a strict locking
hierarchy. The mutual exclusion zone protected by the forward lock can
enclose the mutual exclusion zone protected by the backward lock but not
vice versa.
For example, in iscsi_conn_teardown or in iscsi_xmit_data when there is
a failure and __iscsi_put_task is called, the backward lock is taken while
the forward lock is still taken. On the other hand, if in the RX path a nop
is to be sent, for example in iscsi_handle_reject or __iscsi_complete_pdu
than the forward lock is released and the backward lock is taken for the
duration of iscsi_send_nopout, later the backward lock is released and the
forward lock is retaken.
libiscsi_tcp uses two kernel fifos the r2t pool and the r2t queue.
The insertion and deletion from these queues didn't corespond to the
assumption taken by the new forward/backwards session locking paradigm.
That is, in iscsi_tcp_clenup_task which belongs to the RX (backwards)
path, r2t is taken out from r2t queue and inserted to the r2t pool.
In iscsi_tcp_get_curr_r2t which belong to the TX (forward) path, r2t
is also inserted to the r2t pool and another r2t is pulled from r2t
queue.
Only in iscsi_tcp_r2t_rsp which is called in the RX path but can requeue
to the TX path, r2t is taken from the r2t pool and inserted to the r2t
queue.
In order to cope with this situation, two spin locks were added,
pool2queue and queue2pool. The former protects extracting from the
r2t pool and inserting to the r2t queue, and the later protects the
extracing from the r2t queue and inserting to the r2t pool.
Signed-off-by: Shlomo Pongratz <shlomop@mellanox.com>
Signed-off-by: Or Gerlitz <ogerlitz@mellanox.com>
[minor fix up to apply cleanly and compile fix]
Signed-off-by: Mike Christie <michaelc@cs.wisc.edu>
Signed-off-by: James Bottomley <JBottomley@Parallels.com>
2014-02-07 13:41:38 +07:00
|
|
|
spin_lock_bh(&session->back_lock);
|
2009-09-05 09:06:35 +07:00
|
|
|
__iscsi_complete_pdu(conn, (struct iscsi_hdr *)ppdu, pbuffer, buf_len);
|
[SCSI] libiscsi: Reduce locking contention in fast path
Replace the session lock with two locks, a forward lock and
a backwards lock named frwd_lock and back_lock respectively.
The forward lock protects resources that change while sending a
request to the target, such as cmdsn, queued_cmdsn, and allocating
task from the commands' pool with kfifo_out.
The backward lock protects resources that change while processing
a response or in error path, such as cmdsn_exp, cmdsn_max, and
returning tasks to the commands' pool with kfifo_in.
Under a steady state fast-path situation, that is when one
or more processes/threads submit IO to an iscsi device and
a single kernel upcall (e.g softirq) is dealing with processing
of responses without errors, this patch eliminates the contention
between the queuecommand()/request response/scsi_done() flows
associated with iscsi sessions.
Between the forward and the backward locks exists a strict locking
hierarchy. The mutual exclusion zone protected by the forward lock can
enclose the mutual exclusion zone protected by the backward lock but not
vice versa.
For example, in iscsi_conn_teardown or in iscsi_xmit_data when there is
a failure and __iscsi_put_task is called, the backward lock is taken while
the forward lock is still taken. On the other hand, if in the RX path a nop
is to be sent, for example in iscsi_handle_reject or __iscsi_complete_pdu
than the forward lock is released and the backward lock is taken for the
duration of iscsi_send_nopout, later the backward lock is released and the
forward lock is retaken.
libiscsi_tcp uses two kernel fifos the r2t pool and the r2t queue.
The insertion and deletion from these queues didn't corespond to the
assumption taken by the new forward/backwards session locking paradigm.
That is, in iscsi_tcp_clenup_task which belongs to the RX (backwards)
path, r2t is taken out from r2t queue and inserted to the r2t pool.
In iscsi_tcp_get_curr_r2t which belong to the TX (forward) path, r2t
is also inserted to the r2t pool and another r2t is pulled from r2t
queue.
Only in iscsi_tcp_r2t_rsp which is called in the RX path but can requeue
to the TX path, r2t is taken from the r2t pool and inserted to the r2t
queue.
In order to cope with this situation, two spin locks were added,
pool2queue and queue2pool. The former protects extracting from the
r2t pool and inserting to the r2t queue, and the later protects the
extracing from the r2t queue and inserting to the r2t pool.
Signed-off-by: Shlomo Pongratz <shlomop@mellanox.com>
Signed-off-by: Or Gerlitz <ogerlitz@mellanox.com>
[minor fix up to apply cleanly and compile fix]
Signed-off-by: Mike Christie <michaelc@cs.wisc.edu>
Signed-off-by: James Bottomley <JBottomley@Parallels.com>
2014-02-07 13:41:38 +07:00
|
|
|
spin_unlock_bh(&session->back_lock);
|
2009-09-05 09:06:35 +07:00
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static struct sgl_handle *alloc_io_sgl_handle(struct beiscsi_hba *phba)
|
|
|
|
{
|
|
|
|
struct sgl_handle *psgl_handle;
|
|
|
|
|
2016-02-04 17:19:19 +07:00
|
|
|
spin_lock_bh(&phba->io_sgl_lock);
|
2009-09-05 09:06:35 +07:00
|
|
|
if (phba->io_sgl_hndl_avbl) {
|
2012-08-21 00:30:18 +07:00
|
|
|
beiscsi_log(phba, KERN_INFO, BEISCSI_LOG_IO,
|
|
|
|
"BM_%d : In alloc_io_sgl_handle,"
|
|
|
|
" io_sgl_alloc_index=%d\n",
|
|
|
|
phba->io_sgl_alloc_index);
|
|
|
|
|
2009-09-05 09:06:35 +07:00
|
|
|
psgl_handle = phba->io_sgl_hndl_base[phba->
|
|
|
|
io_sgl_alloc_index];
|
|
|
|
phba->io_sgl_hndl_base[phba->io_sgl_alloc_index] = NULL;
|
|
|
|
phba->io_sgl_hndl_avbl--;
|
2009-10-23 13:22:33 +07:00
|
|
|
if (phba->io_sgl_alloc_index == (phba->params.
|
|
|
|
ios_per_ctrl - 1))
|
2009-09-05 09:06:35 +07:00
|
|
|
phba->io_sgl_alloc_index = 0;
|
|
|
|
else
|
|
|
|
phba->io_sgl_alloc_index++;
|
|
|
|
} else
|
|
|
|
psgl_handle = NULL;
|
2016-02-04 17:19:19 +07:00
|
|
|
spin_unlock_bh(&phba->io_sgl_lock);
|
2009-09-05 09:06:35 +07:00
|
|
|
return psgl_handle;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void
|
|
|
|
free_io_sgl_handle(struct beiscsi_hba *phba, struct sgl_handle *psgl_handle)
|
|
|
|
{
|
2016-02-04 17:19:19 +07:00
|
|
|
spin_lock_bh(&phba->io_sgl_lock);
|
2012-08-21 00:30:18 +07:00
|
|
|
beiscsi_log(phba, KERN_INFO, BEISCSI_LOG_IO,
|
|
|
|
"BM_%d : In free_,io_sgl_free_index=%d\n",
|
|
|
|
phba->io_sgl_free_index);
|
|
|
|
|
2009-09-05 09:06:35 +07:00
|
|
|
if (phba->io_sgl_hndl_base[phba->io_sgl_free_index]) {
|
|
|
|
/*
|
|
|
|
* this can happen if clean_task is called on a task that
|
|
|
|
* failed in xmit_task or alloc_pdu.
|
|
|
|
*/
|
2012-08-21 00:30:18 +07:00
|
|
|
beiscsi_log(phba, KERN_INFO, BEISCSI_LOG_IO,
|
|
|
|
"BM_%d : Double Free in IO SGL io_sgl_free_index=%d,"
|
|
|
|
"value there=%p\n", phba->io_sgl_free_index,
|
|
|
|
phba->io_sgl_hndl_base
|
|
|
|
[phba->io_sgl_free_index]);
|
2016-02-04 17:19:19 +07:00
|
|
|
spin_unlock_bh(&phba->io_sgl_lock);
|
2009-09-05 09:06:35 +07:00
|
|
|
return;
|
|
|
|
}
|
|
|
|
phba->io_sgl_hndl_base[phba->io_sgl_free_index] = psgl_handle;
|
|
|
|
phba->io_sgl_hndl_avbl++;
|
|
|
|
if (phba->io_sgl_free_index == (phba->params.ios_per_ctrl - 1))
|
|
|
|
phba->io_sgl_free_index = 0;
|
|
|
|
else
|
|
|
|
phba->io_sgl_free_index++;
|
2016-02-04 17:19:19 +07:00
|
|
|
spin_unlock_bh(&phba->io_sgl_lock);
|
2009-09-05 09:06:35 +07:00
|
|
|
}
|
|
|
|
|
2016-01-20 15:41:00 +07:00
|
|
|
static inline struct wrb_handle *
|
|
|
|
beiscsi_get_wrb_handle(struct hwi_wrb_context *pwrb_context,
|
|
|
|
unsigned int wrbs_per_cxn)
|
|
|
|
{
|
|
|
|
struct wrb_handle *pwrb_handle;
|
|
|
|
|
2016-02-04 17:19:20 +07:00
|
|
|
spin_lock_bh(&pwrb_context->wrb_lock);
|
2016-01-20 15:41:00 +07:00
|
|
|
pwrb_handle = pwrb_context->pwrb_handle_base[pwrb_context->alloc_index];
|
|
|
|
pwrb_context->wrb_handles_available--;
|
|
|
|
if (pwrb_context->alloc_index == (wrbs_per_cxn - 1))
|
|
|
|
pwrb_context->alloc_index = 0;
|
|
|
|
else
|
|
|
|
pwrb_context->alloc_index++;
|
2016-02-04 17:19:20 +07:00
|
|
|
spin_unlock_bh(&pwrb_context->wrb_lock);
|
2016-01-20 15:41:00 +07:00
|
|
|
|
|
|
|
return pwrb_handle;
|
|
|
|
}
|
|
|
|
|
2009-09-05 09:06:35 +07:00
|
|
|
/**
|
|
|
|
* alloc_wrb_handle - To allocate a wrb handle
|
|
|
|
* @phba: The hba pointer
|
|
|
|
* @cid: The cid to use for allocation
|
2015-08-20 06:14:30 +07:00
|
|
|
* @pwrb_context: ptr to ptr to wrb context
|
2009-09-05 09:06:35 +07:00
|
|
|
*
|
|
|
|
* This happens under session_lock until submission to chip
|
|
|
|
*/
|
2015-08-20 06:14:30 +07:00
|
|
|
struct wrb_handle *alloc_wrb_handle(struct beiscsi_hba *phba, unsigned int cid,
|
2016-01-20 15:41:00 +07:00
|
|
|
struct hwi_wrb_context **pcontext)
|
2009-09-05 09:06:35 +07:00
|
|
|
{
|
|
|
|
struct hwi_wrb_context *pwrb_context;
|
|
|
|
struct hwi_controller *phwi_ctrlr;
|
2013-04-06 10:38:32 +07:00
|
|
|
uint16_t cri_index = BE_GET_CRI_FROM_CID(cid);
|
2009-09-05 09:06:35 +07:00
|
|
|
|
|
|
|
phwi_ctrlr = phba->phwi_ctrlr;
|
2013-04-06 10:38:32 +07:00
|
|
|
pwrb_context = &phwi_ctrlr->wrb_context[cri_index];
|
2016-01-20 15:41:00 +07:00
|
|
|
/* return the context address */
|
|
|
|
*pcontext = pwrb_context;
|
|
|
|
return beiscsi_get_wrb_handle(pwrb_context, phba->params.wrbs_per_cxn);
|
|
|
|
}
|
2015-08-20 06:14:30 +07:00
|
|
|
|
2016-01-20 15:41:00 +07:00
|
|
|
static inline void
|
|
|
|
beiscsi_put_wrb_handle(struct hwi_wrb_context *pwrb_context,
|
|
|
|
struct wrb_handle *pwrb_handle,
|
|
|
|
unsigned int wrbs_per_cxn)
|
|
|
|
{
|
2016-02-04 17:19:20 +07:00
|
|
|
spin_lock_bh(&pwrb_context->wrb_lock);
|
2016-01-20 15:41:00 +07:00
|
|
|
pwrb_context->pwrb_handle_base[pwrb_context->free_index] = pwrb_handle;
|
|
|
|
pwrb_context->wrb_handles_available++;
|
|
|
|
if (pwrb_context->free_index == (wrbs_per_cxn - 1))
|
|
|
|
pwrb_context->free_index = 0;
|
|
|
|
else
|
|
|
|
pwrb_context->free_index++;
|
2016-02-04 17:19:20 +07:00
|
|
|
spin_unlock_bh(&pwrb_context->wrb_lock);
|
2009-09-05 09:06:35 +07:00
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
|
|
|
* free_wrb_handle - To free the wrb handle back to pool
|
|
|
|
* @phba: The hba pointer
|
|
|
|
* @pwrb_context: The context to free from
|
|
|
|
* @pwrb_handle: The wrb_handle to free
|
|
|
|
*
|
|
|
|
* This happens under session_lock until submission to chip
|
|
|
|
*/
|
|
|
|
static void
|
|
|
|
free_wrb_handle(struct beiscsi_hba *phba, struct hwi_wrb_context *pwrb_context,
|
|
|
|
struct wrb_handle *pwrb_handle)
|
|
|
|
{
|
2016-01-20 15:41:00 +07:00
|
|
|
beiscsi_put_wrb_handle(pwrb_context,
|
|
|
|
pwrb_handle,
|
|
|
|
phba->params.wrbs_per_cxn);
|
2012-08-21 00:30:18 +07:00
|
|
|
beiscsi_log(phba, KERN_INFO,
|
|
|
|
BEISCSI_LOG_IO | BEISCSI_LOG_CONFIG,
|
|
|
|
"BM_%d : FREE WRB: pwrb_handle=%p free_index=0x%x"
|
|
|
|
"wrb_handles_available=%d\n",
|
|
|
|
pwrb_handle, pwrb_context->free_index,
|
|
|
|
pwrb_context->wrb_handles_available);
|
2009-09-05 09:06:35 +07:00
|
|
|
}
|
|
|
|
|
|
|
|
static struct sgl_handle *alloc_mgmt_sgl_handle(struct beiscsi_hba *phba)
|
|
|
|
{
|
|
|
|
struct sgl_handle *psgl_handle;
|
|
|
|
|
2016-02-04 17:19:19 +07:00
|
|
|
spin_lock_bh(&phba->mgmt_sgl_lock);
|
2009-09-05 09:06:35 +07:00
|
|
|
if (phba->eh_sgl_hndl_avbl) {
|
|
|
|
psgl_handle = phba->eh_sgl_hndl_base[phba->eh_sgl_alloc_index];
|
|
|
|
phba->eh_sgl_hndl_base[phba->eh_sgl_alloc_index] = NULL;
|
2012-08-21 00:30:18 +07:00
|
|
|
beiscsi_log(phba, KERN_INFO, BEISCSI_LOG_CONFIG,
|
|
|
|
"BM_%d : mgmt_sgl_alloc_index=%d=0x%x\n",
|
|
|
|
phba->eh_sgl_alloc_index,
|
|
|
|
phba->eh_sgl_alloc_index);
|
|
|
|
|
2009-09-05 09:06:35 +07:00
|
|
|
phba->eh_sgl_hndl_avbl--;
|
|
|
|
if (phba->eh_sgl_alloc_index ==
|
|
|
|
(phba->params.icds_per_ctrl - phba->params.ios_per_ctrl -
|
|
|
|
1))
|
|
|
|
phba->eh_sgl_alloc_index = 0;
|
|
|
|
else
|
|
|
|
phba->eh_sgl_alloc_index++;
|
|
|
|
} else
|
|
|
|
psgl_handle = NULL;
|
2016-02-04 17:19:19 +07:00
|
|
|
spin_unlock_bh(&phba->mgmt_sgl_lock);
|
2009-09-05 09:06:35 +07:00
|
|
|
return psgl_handle;
|
|
|
|
}
|
|
|
|
|
|
|
|
void
|
|
|
|
free_mgmt_sgl_handle(struct beiscsi_hba *phba, struct sgl_handle *psgl_handle)
|
|
|
|
{
|
2016-02-04 17:19:19 +07:00
|
|
|
spin_lock_bh(&phba->mgmt_sgl_lock);
|
2012-08-21 00:30:18 +07:00
|
|
|
beiscsi_log(phba, KERN_INFO, BEISCSI_LOG_CONFIG,
|
|
|
|
"BM_%d : In free_mgmt_sgl_handle,"
|
|
|
|
"eh_sgl_free_index=%d\n",
|
|
|
|
phba->eh_sgl_free_index);
|
|
|
|
|
2009-09-05 09:06:35 +07:00
|
|
|
if (phba->eh_sgl_hndl_base[phba->eh_sgl_free_index]) {
|
|
|
|
/*
|
|
|
|
* this can happen if clean_task is called on a task that
|
|
|
|
* failed in xmit_task or alloc_pdu.
|
|
|
|
*/
|
2012-08-21 00:30:18 +07:00
|
|
|
beiscsi_log(phba, KERN_WARNING, BEISCSI_LOG_CONFIG,
|
|
|
|
"BM_%d : Double Free in eh SGL ,"
|
|
|
|
"eh_sgl_free_index=%d\n",
|
|
|
|
phba->eh_sgl_free_index);
|
2016-02-04 17:19:19 +07:00
|
|
|
spin_unlock_bh(&phba->mgmt_sgl_lock);
|
2009-09-05 09:06:35 +07:00
|
|
|
return;
|
|
|
|
}
|
|
|
|
phba->eh_sgl_hndl_base[phba->eh_sgl_free_index] = psgl_handle;
|
|
|
|
phba->eh_sgl_hndl_avbl++;
|
|
|
|
if (phba->eh_sgl_free_index ==
|
|
|
|
(phba->params.icds_per_ctrl - phba->params.ios_per_ctrl - 1))
|
|
|
|
phba->eh_sgl_free_index = 0;
|
|
|
|
else
|
|
|
|
phba->eh_sgl_free_index++;
|
2016-02-04 17:19:19 +07:00
|
|
|
spin_unlock_bh(&phba->mgmt_sgl_lock);
|
2009-09-05 09:06:35 +07:00
|
|
|
}
|
|
|
|
|
|
|
|
static void
|
|
|
|
be_complete_io(struct beiscsi_conn *beiscsi_conn,
|
2012-10-20 06:14:49 +07:00
|
|
|
struct iscsi_task *task,
|
|
|
|
struct common_sol_cqe *csol_cqe)
|
2009-09-05 09:06:35 +07:00
|
|
|
{
|
|
|
|
struct beiscsi_io_task *io_task = task->dd_data;
|
|
|
|
struct be_status_bhs *sts_bhs =
|
|
|
|
(struct be_status_bhs *)io_task->cmd_bhs;
|
|
|
|
struct iscsi_conn *conn = beiscsi_conn->conn;
|
|
|
|
unsigned char *sense;
|
|
|
|
u32 resid = 0, exp_cmdsn, max_cmdsn;
|
|
|
|
u8 rsp, status, flags;
|
|
|
|
|
2012-10-20 06:14:49 +07:00
|
|
|
exp_cmdsn = csol_cqe->exp_cmdsn;
|
|
|
|
max_cmdsn = (csol_cqe->exp_cmdsn +
|
|
|
|
csol_cqe->cmd_wnd - 1);
|
|
|
|
rsp = csol_cqe->i_resp;
|
|
|
|
status = csol_cqe->i_sts;
|
|
|
|
flags = csol_cqe->i_flags;
|
|
|
|
resid = csol_cqe->res_cnt;
|
|
|
|
|
2011-10-08 07:31:10 +07:00
|
|
|
if (!task->sc) {
|
2014-01-29 14:16:44 +07:00
|
|
|
if (io_task->scsi_cmnd) {
|
2011-10-08 07:31:10 +07:00
|
|
|
scsi_dma_unmap(io_task->scsi_cmnd);
|
2014-01-29 14:16:44 +07:00
|
|
|
io_task->scsi_cmnd = NULL;
|
|
|
|
}
|
2009-09-05 09:06:35 +07:00
|
|
|
|
2011-10-08 07:31:10 +07:00
|
|
|
return;
|
|
|
|
}
|
2009-09-05 09:06:35 +07:00
|
|
|
task->sc->result = (DID_OK << 16) | status;
|
|
|
|
if (rsp != ISCSI_STATUS_CMD_COMPLETED) {
|
|
|
|
task->sc->result = DID_ERROR << 16;
|
|
|
|
goto unmap;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* bidi not initially supported */
|
|
|
|
if (flags & (ISCSI_FLAG_CMD_UNDERFLOW | ISCSI_FLAG_CMD_OVERFLOW)) {
|
|
|
|
if (!status && (flags & ISCSI_FLAG_CMD_OVERFLOW))
|
|
|
|
task->sc->result = DID_ERROR << 16;
|
|
|
|
|
|
|
|
if (flags & ISCSI_FLAG_CMD_UNDERFLOW) {
|
|
|
|
scsi_set_resid(task->sc, resid);
|
|
|
|
if (!status && (scsi_bufflen(task->sc) - resid <
|
|
|
|
task->sc->underflow))
|
|
|
|
task->sc->result = DID_ERROR << 16;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
if (status == SAM_STAT_CHECK_CONDITION) {
|
2011-09-26 13:23:37 +07:00
|
|
|
u16 sense_len;
|
2009-10-23 13:22:33 +07:00
|
|
|
unsigned short *slen = (unsigned short *)sts_bhs->sense_info;
|
2011-09-26 13:23:37 +07:00
|
|
|
|
2009-09-05 09:06:35 +07:00
|
|
|
sense = sts_bhs->sense_info + sizeof(unsigned short);
|
2011-09-26 13:23:37 +07:00
|
|
|
sense_len = be16_to_cpu(*slen);
|
2009-09-05 09:06:35 +07:00
|
|
|
memcpy(task->sc->sense_buffer, sense,
|
|
|
|
min_t(u16, sense_len, SCSI_SENSE_BUFFERSIZE));
|
|
|
|
}
|
2010-01-05 06:40:46 +07:00
|
|
|
|
2012-10-20 06:14:49 +07:00
|
|
|
if (io_task->cmd_bhs->iscsi_hdr.flags & ISCSI_FLAG_CMD_READ)
|
|
|
|
conn->rxdata_octets += resid;
|
2009-09-05 09:06:35 +07:00
|
|
|
unmap:
|
2015-04-25 09:47:45 +07:00
|
|
|
if (io_task->scsi_cmnd) {
|
|
|
|
scsi_dma_unmap(io_task->scsi_cmnd);
|
|
|
|
io_task->scsi_cmnd = NULL;
|
|
|
|
}
|
2009-09-05 09:06:35 +07:00
|
|
|
iscsi_complete_scsi_task(task, exp_cmdsn, max_cmdsn);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void
|
|
|
|
be_complete_logout(struct beiscsi_conn *beiscsi_conn,
|
2012-10-20 06:14:49 +07:00
|
|
|
struct iscsi_task *task,
|
|
|
|
struct common_sol_cqe *csol_cqe)
|
2009-09-05 09:06:35 +07:00
|
|
|
{
|
|
|
|
struct iscsi_logout_rsp *hdr;
|
2009-10-23 13:22:33 +07:00
|
|
|
struct beiscsi_io_task *io_task = task->dd_data;
|
2009-09-05 09:06:35 +07:00
|
|
|
struct iscsi_conn *conn = beiscsi_conn->conn;
|
|
|
|
|
|
|
|
hdr = (struct iscsi_logout_rsp *)task->hdr;
|
2010-01-05 06:37:02 +07:00
|
|
|
hdr->opcode = ISCSI_OP_LOGOUT_RSP;
|
2009-09-05 09:06:35 +07:00
|
|
|
hdr->t2wait = 5;
|
|
|
|
hdr->t2retain = 0;
|
2012-10-20 06:14:49 +07:00
|
|
|
hdr->flags = csol_cqe->i_flags;
|
|
|
|
hdr->response = csol_cqe->i_resp;
|
2013-04-06 10:38:37 +07:00
|
|
|
hdr->exp_cmdsn = cpu_to_be32(csol_cqe->exp_cmdsn);
|
|
|
|
hdr->max_cmdsn = cpu_to_be32(csol_cqe->exp_cmdsn +
|
|
|
|
csol_cqe->cmd_wnd - 1);
|
2012-10-20 06:14:49 +07:00
|
|
|
|
2010-01-05 06:37:02 +07:00
|
|
|
hdr->dlength[0] = 0;
|
|
|
|
hdr->dlength[1] = 0;
|
|
|
|
hdr->dlength[2] = 0;
|
2009-09-05 09:06:35 +07:00
|
|
|
hdr->hlength = 0;
|
2009-10-23 13:22:33 +07:00
|
|
|
hdr->itt = io_task->libiscsi_itt;
|
2009-09-05 09:06:35 +07:00
|
|
|
__iscsi_complete_pdu(conn, (struct iscsi_hdr *)hdr, NULL, 0);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void
|
|
|
|
be_complete_tmf(struct beiscsi_conn *beiscsi_conn,
|
2012-10-20 06:14:49 +07:00
|
|
|
struct iscsi_task *task,
|
|
|
|
struct common_sol_cqe *csol_cqe)
|
2009-09-05 09:06:35 +07:00
|
|
|
{
|
|
|
|
struct iscsi_tm_rsp *hdr;
|
|
|
|
struct iscsi_conn *conn = beiscsi_conn->conn;
|
2009-10-23 13:22:33 +07:00
|
|
|
struct beiscsi_io_task *io_task = task->dd_data;
|
2009-09-05 09:06:35 +07:00
|
|
|
|
|
|
|
hdr = (struct iscsi_tm_rsp *)task->hdr;
|
2010-01-05 06:37:02 +07:00
|
|
|
hdr->opcode = ISCSI_OP_SCSI_TMFUNC_RSP;
|
2012-10-20 06:14:49 +07:00
|
|
|
hdr->flags = csol_cqe->i_flags;
|
|
|
|
hdr->response = csol_cqe->i_resp;
|
2013-04-06 10:38:37 +07:00
|
|
|
hdr->exp_cmdsn = cpu_to_be32(csol_cqe->exp_cmdsn);
|
|
|
|
hdr->max_cmdsn = cpu_to_be32(csol_cqe->exp_cmdsn +
|
|
|
|
csol_cqe->cmd_wnd - 1);
|
2012-10-20 06:14:49 +07:00
|
|
|
|
2009-10-23 13:22:33 +07:00
|
|
|
hdr->itt = io_task->libiscsi_itt;
|
2009-09-05 09:06:35 +07:00
|
|
|
__iscsi_complete_pdu(conn, (struct iscsi_hdr *)hdr, NULL, 0);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void
|
|
|
|
hwi_complete_drvr_msgs(struct beiscsi_conn *beiscsi_conn,
|
|
|
|
struct beiscsi_hba *phba, struct sol_cqe *psol)
|
|
|
|
{
|
|
|
|
struct hwi_wrb_context *pwrb_context;
|
2009-10-23 13:22:33 +07:00
|
|
|
struct wrb_handle *pwrb_handle = NULL;
|
2009-09-05 09:06:35 +07:00
|
|
|
struct hwi_controller *phwi_ctrlr;
|
2009-10-23 13:22:33 +07:00
|
|
|
struct iscsi_task *task;
|
|
|
|
struct beiscsi_io_task *io_task;
|
2013-04-06 10:38:32 +07:00
|
|
|
uint16_t wrb_index, cid, cri_index;
|
2009-09-05 09:06:35 +07:00
|
|
|
|
|
|
|
phwi_ctrlr = phba->phwi_ctrlr;
|
2013-04-06 10:38:26 +07:00
|
|
|
if (is_chip_be2_be3r(phba)) {
|
|
|
|
wrb_index = AMAP_GET_BITS(struct amap_it_dmsg_cqe,
|
2012-10-20 06:14:49 +07:00
|
|
|
wrb_idx, psol);
|
2013-04-06 10:38:26 +07:00
|
|
|
cid = AMAP_GET_BITS(struct amap_it_dmsg_cqe,
|
2012-10-20 06:14:49 +07:00
|
|
|
cid, psol);
|
|
|
|
} else {
|
2013-04-06 10:38:26 +07:00
|
|
|
wrb_index = AMAP_GET_BITS(struct amap_it_dmsg_cqe_v2,
|
2012-10-20 06:14:49 +07:00
|
|
|
wrb_idx, psol);
|
2013-04-06 10:38:26 +07:00
|
|
|
cid = AMAP_GET_BITS(struct amap_it_dmsg_cqe_v2,
|
2012-10-20 06:14:49 +07:00
|
|
|
cid, psol);
|
|
|
|
}
|
|
|
|
|
2013-04-06 10:38:32 +07:00
|
|
|
cri_index = BE_GET_CRI_FROM_CID(cid);
|
|
|
|
pwrb_context = &phwi_ctrlr->wrb_context[cri_index];
|
2012-10-20 06:14:49 +07:00
|
|
|
pwrb_handle = pwrb_context->pwrb_handle_basestd[wrb_index];
|
2010-01-23 07:04:24 +07:00
|
|
|
task = pwrb_handle->pio_handle;
|
2009-10-23 13:23:49 +07:00
|
|
|
|
2009-10-23 13:22:33 +07:00
|
|
|
io_task = task->dd_data;
|
2013-04-06 10:38:31 +07:00
|
|
|
memset(io_task->pwrb_handle->pwrb, 0, sizeof(struct iscsi_wrb));
|
|
|
|
iscsi_put_task(task);
|
2009-09-05 09:06:35 +07:00
|
|
|
}
|
|
|
|
|
|
|
|
static void
|
|
|
|
be_complete_nopin_resp(struct beiscsi_conn *beiscsi_conn,
|
2012-10-20 06:14:49 +07:00
|
|
|
struct iscsi_task *task,
|
|
|
|
struct common_sol_cqe *csol_cqe)
|
2009-09-05 09:06:35 +07:00
|
|
|
{
|
|
|
|
struct iscsi_nopin *hdr;
|
|
|
|
struct iscsi_conn *conn = beiscsi_conn->conn;
|
2009-10-23 13:22:33 +07:00
|
|
|
struct beiscsi_io_task *io_task = task->dd_data;
|
2009-09-05 09:06:35 +07:00
|
|
|
|
|
|
|
hdr = (struct iscsi_nopin *)task->hdr;
|
2012-10-20 06:14:49 +07:00
|
|
|
hdr->flags = csol_cqe->i_flags;
|
|
|
|
hdr->exp_cmdsn = cpu_to_be32(csol_cqe->exp_cmdsn);
|
2013-04-06 10:38:37 +07:00
|
|
|
hdr->max_cmdsn = cpu_to_be32(csol_cqe->exp_cmdsn +
|
|
|
|
csol_cqe->cmd_wnd - 1);
|
2012-10-20 06:14:49 +07:00
|
|
|
|
2009-09-05 09:06:35 +07:00
|
|
|
hdr->opcode = ISCSI_OP_NOOP_IN;
|
2009-10-23 13:22:33 +07:00
|
|
|
hdr->itt = io_task->libiscsi_itt;
|
2009-09-05 09:06:35 +07:00
|
|
|
__iscsi_complete_pdu(conn, (struct iscsi_hdr *)hdr, NULL, 0);
|
|
|
|
}
|
|
|
|
|
2012-10-20 06:14:49 +07:00
|
|
|
static void adapter_get_sol_cqe(struct beiscsi_hba *phba,
|
|
|
|
struct sol_cqe *psol,
|
|
|
|
struct common_sol_cqe *csol_cqe)
|
|
|
|
{
|
2013-04-06 10:38:26 +07:00
|
|
|
if (is_chip_be2_be3r(phba)) {
|
|
|
|
csol_cqe->exp_cmdsn = AMAP_GET_BITS(struct amap_sol_cqe,
|
|
|
|
i_exp_cmd_sn, psol);
|
|
|
|
csol_cqe->res_cnt = AMAP_GET_BITS(struct amap_sol_cqe,
|
|
|
|
i_res_cnt, psol);
|
|
|
|
csol_cqe->cmd_wnd = AMAP_GET_BITS(struct amap_sol_cqe,
|
|
|
|
i_cmd_wnd, psol);
|
|
|
|
csol_cqe->wrb_index = AMAP_GET_BITS(struct amap_sol_cqe,
|
|
|
|
wrb_index, psol);
|
|
|
|
csol_cqe->cid = AMAP_GET_BITS(struct amap_sol_cqe,
|
|
|
|
cid, psol);
|
|
|
|
csol_cqe->hw_sts = AMAP_GET_BITS(struct amap_sol_cqe,
|
|
|
|
hw_sts, psol);
|
|
|
|
csol_cqe->i_resp = AMAP_GET_BITS(struct amap_sol_cqe,
|
|
|
|
i_resp, psol);
|
|
|
|
csol_cqe->i_sts = AMAP_GET_BITS(struct amap_sol_cqe,
|
|
|
|
i_sts, psol);
|
|
|
|
csol_cqe->i_flags = AMAP_GET_BITS(struct amap_sol_cqe,
|
|
|
|
i_flags, psol);
|
|
|
|
} else {
|
2012-10-20 06:14:49 +07:00
|
|
|
csol_cqe->exp_cmdsn = AMAP_GET_BITS(struct amap_sol_cqe_v2,
|
|
|
|
i_exp_cmd_sn, psol);
|
|
|
|
csol_cqe->res_cnt = AMAP_GET_BITS(struct amap_sol_cqe_v2,
|
|
|
|
i_res_cnt, psol);
|
|
|
|
csol_cqe->wrb_index = AMAP_GET_BITS(struct amap_sol_cqe_v2,
|
|
|
|
wrb_index, psol);
|
|
|
|
csol_cqe->cid = AMAP_GET_BITS(struct amap_sol_cqe_v2,
|
|
|
|
cid, psol);
|
|
|
|
csol_cqe->hw_sts = AMAP_GET_BITS(struct amap_sol_cqe_v2,
|
|
|
|
hw_sts, psol);
|
2013-04-06 10:38:37 +07:00
|
|
|
csol_cqe->cmd_wnd = AMAP_GET_BITS(struct amap_sol_cqe_v2,
|
2012-10-20 06:14:49 +07:00
|
|
|
i_cmd_wnd, psol);
|
|
|
|
if (AMAP_GET_BITS(struct amap_sol_cqe_v2,
|
|
|
|
cmd_cmpl, psol))
|
|
|
|
csol_cqe->i_sts = AMAP_GET_BITS(struct amap_sol_cqe_v2,
|
|
|
|
i_sts, psol);
|
|
|
|
else
|
|
|
|
csol_cqe->i_resp = AMAP_GET_BITS(struct amap_sol_cqe_v2,
|
|
|
|
i_sts, psol);
|
|
|
|
if (AMAP_GET_BITS(struct amap_sol_cqe_v2,
|
|
|
|
u, psol))
|
|
|
|
csol_cqe->i_flags = ISCSI_FLAG_CMD_UNDERFLOW;
|
|
|
|
|
|
|
|
if (AMAP_GET_BITS(struct amap_sol_cqe_v2,
|
|
|
|
o, psol))
|
|
|
|
csol_cqe->i_flags |= ISCSI_FLAG_CMD_OVERFLOW;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
|
2009-09-05 09:06:35 +07:00
|
|
|
static void hwi_complete_cmd(struct beiscsi_conn *beiscsi_conn,
|
|
|
|
struct beiscsi_hba *phba, struct sol_cqe *psol)
|
|
|
|
{
|
|
|
|
struct hwi_wrb_context *pwrb_context;
|
|
|
|
struct wrb_handle *pwrb_handle;
|
|
|
|
struct iscsi_wrb *pwrb = NULL;
|
|
|
|
struct hwi_controller *phwi_ctrlr;
|
|
|
|
struct iscsi_task *task;
|
2009-10-23 13:22:33 +07:00
|
|
|
unsigned int type;
|
2009-09-05 09:06:35 +07:00
|
|
|
struct iscsi_conn *conn = beiscsi_conn->conn;
|
|
|
|
struct iscsi_session *session = conn->session;
|
2012-10-20 06:14:49 +07:00
|
|
|
struct common_sol_cqe csol_cqe = {0};
|
2013-04-06 10:38:32 +07:00
|
|
|
uint16_t cri_index = 0;
|
2009-09-05 09:06:35 +07:00
|
|
|
|
|
|
|
phwi_ctrlr = phba->phwi_ctrlr;
|
2012-10-20 06:14:49 +07:00
|
|
|
|
|
|
|
/* Copy the elements to a common structure */
|
|
|
|
adapter_get_sol_cqe(phba, psol, &csol_cqe);
|
|
|
|
|
2013-04-06 10:38:32 +07:00
|
|
|
cri_index = BE_GET_CRI_FROM_CID(csol_cqe.cid);
|
|
|
|
pwrb_context = &phwi_ctrlr->wrb_context[cri_index];
|
2012-10-20 06:14:49 +07:00
|
|
|
|
|
|
|
pwrb_handle = pwrb_context->pwrb_handle_basestd[
|
|
|
|
csol_cqe.wrb_index];
|
|
|
|
|
2010-01-23 07:04:24 +07:00
|
|
|
task = pwrb_handle->pio_handle;
|
|
|
|
pwrb = pwrb_handle->pwrb;
|
2012-10-20 06:14:49 +07:00
|
|
|
type = ((struct beiscsi_io_task *)task->dd_data)->wrb_type;
|
2010-01-23 07:04:24 +07:00
|
|
|
|
[SCSI] libiscsi: Reduce locking contention in fast path
Replace the session lock with two locks, a forward lock and
a backwards lock named frwd_lock and back_lock respectively.
The forward lock protects resources that change while sending a
request to the target, such as cmdsn, queued_cmdsn, and allocating
task from the commands' pool with kfifo_out.
The backward lock protects resources that change while processing
a response or in error path, such as cmdsn_exp, cmdsn_max, and
returning tasks to the commands' pool with kfifo_in.
Under a steady state fast-path situation, that is when one
or more processes/threads submit IO to an iscsi device and
a single kernel upcall (e.g softirq) is dealing with processing
of responses without errors, this patch eliminates the contention
between the queuecommand()/request response/scsi_done() flows
associated with iscsi sessions.
Between the forward and the backward locks exists a strict locking
hierarchy. The mutual exclusion zone protected by the forward lock can
enclose the mutual exclusion zone protected by the backward lock but not
vice versa.
For example, in iscsi_conn_teardown or in iscsi_xmit_data when there is
a failure and __iscsi_put_task is called, the backward lock is taken while
the forward lock is still taken. On the other hand, if in the RX path a nop
is to be sent, for example in iscsi_handle_reject or __iscsi_complete_pdu
than the forward lock is released and the backward lock is taken for the
duration of iscsi_send_nopout, later the backward lock is released and the
forward lock is retaken.
libiscsi_tcp uses two kernel fifos the r2t pool and the r2t queue.
The insertion and deletion from these queues didn't corespond to the
assumption taken by the new forward/backwards session locking paradigm.
That is, in iscsi_tcp_clenup_task which belongs to the RX (backwards)
path, r2t is taken out from r2t queue and inserted to the r2t pool.
In iscsi_tcp_get_curr_r2t which belong to the TX (forward) path, r2t
is also inserted to the r2t pool and another r2t is pulled from r2t
queue.
Only in iscsi_tcp_r2t_rsp which is called in the RX path but can requeue
to the TX path, r2t is taken from the r2t pool and inserted to the r2t
queue.
In order to cope with this situation, two spin locks were added,
pool2queue and queue2pool. The former protects extracting from the
r2t pool and inserting to the r2t queue, and the later protects the
extracing from the r2t queue and inserting to the r2t pool.
Signed-off-by: Shlomo Pongratz <shlomop@mellanox.com>
Signed-off-by: Or Gerlitz <ogerlitz@mellanox.com>
[minor fix up to apply cleanly and compile fix]
Signed-off-by: Mike Christie <michaelc@cs.wisc.edu>
Signed-off-by: James Bottomley <JBottomley@Parallels.com>
2014-02-07 13:41:38 +07:00
|
|
|
spin_lock_bh(&session->back_lock);
|
2009-10-23 13:22:33 +07:00
|
|
|
switch (type) {
|
2009-09-05 09:06:35 +07:00
|
|
|
case HWH_TYPE_IO:
|
|
|
|
case HWH_TYPE_IO_RD:
|
|
|
|
if ((task->hdr->opcode & ISCSI_OPCODE_MASK) ==
|
2010-02-20 09:33:56 +07:00
|
|
|
ISCSI_OP_NOOP_OUT)
|
2012-10-20 06:14:49 +07:00
|
|
|
be_complete_nopin_resp(beiscsi_conn, task, &csol_cqe);
|
2010-02-20 09:33:56 +07:00
|
|
|
else
|
2012-10-20 06:14:49 +07:00
|
|
|
be_complete_io(beiscsi_conn, task, &csol_cqe);
|
2009-09-05 09:06:35 +07:00
|
|
|
break;
|
|
|
|
|
|
|
|
case HWH_TYPE_LOGOUT:
|
2010-02-20 09:33:56 +07:00
|
|
|
if ((task->hdr->opcode & ISCSI_OPCODE_MASK) == ISCSI_OP_LOGOUT)
|
2012-10-20 06:14:49 +07:00
|
|
|
be_complete_logout(beiscsi_conn, task, &csol_cqe);
|
2010-02-20 09:33:56 +07:00
|
|
|
else
|
2012-10-20 06:14:49 +07:00
|
|
|
be_complete_tmf(beiscsi_conn, task, &csol_cqe);
|
2009-09-05 09:06:35 +07:00
|
|
|
break;
|
|
|
|
|
|
|
|
case HWH_TYPE_LOGIN:
|
2012-08-21 00:30:18 +07:00
|
|
|
beiscsi_log(phba, KERN_ERR,
|
|
|
|
BEISCSI_LOG_CONFIG | BEISCSI_LOG_IO,
|
|
|
|
"BM_%d :\t\t No HWH_TYPE_LOGIN Expected in"
|
|
|
|
" hwi_complete_cmd- Solicited path\n");
|
2009-09-05 09:06:35 +07:00
|
|
|
break;
|
|
|
|
|
|
|
|
case HWH_TYPE_NOP:
|
2012-10-20 06:14:49 +07:00
|
|
|
be_complete_nopin_resp(beiscsi_conn, task, &csol_cqe);
|
2009-09-05 09:06:35 +07:00
|
|
|
break;
|
|
|
|
|
|
|
|
default:
|
2012-08-21 00:30:18 +07:00
|
|
|
beiscsi_log(phba, KERN_WARNING,
|
|
|
|
BEISCSI_LOG_CONFIG | BEISCSI_LOG_IO,
|
|
|
|
"BM_%d : In hwi_complete_cmd, unknown type = %d"
|
|
|
|
"wrb_index 0x%x CID 0x%x\n", type,
|
2012-10-20 06:14:49 +07:00
|
|
|
csol_cqe.wrb_index,
|
|
|
|
csol_cqe.cid);
|
2009-09-05 09:06:35 +07:00
|
|
|
break;
|
|
|
|
}
|
2009-10-23 13:23:49 +07:00
|
|
|
|
[SCSI] libiscsi: Reduce locking contention in fast path
Replace the session lock with two locks, a forward lock and
a backwards lock named frwd_lock and back_lock respectively.
The forward lock protects resources that change while sending a
request to the target, such as cmdsn, queued_cmdsn, and allocating
task from the commands' pool with kfifo_out.
The backward lock protects resources that change while processing
a response or in error path, such as cmdsn_exp, cmdsn_max, and
returning tasks to the commands' pool with kfifo_in.
Under a steady state fast-path situation, that is when one
or more processes/threads submit IO to an iscsi device and
a single kernel upcall (e.g softirq) is dealing with processing
of responses without errors, this patch eliminates the contention
between the queuecommand()/request response/scsi_done() flows
associated with iscsi sessions.
Between the forward and the backward locks exists a strict locking
hierarchy. The mutual exclusion zone protected by the forward lock can
enclose the mutual exclusion zone protected by the backward lock but not
vice versa.
For example, in iscsi_conn_teardown or in iscsi_xmit_data when there is
a failure and __iscsi_put_task is called, the backward lock is taken while
the forward lock is still taken. On the other hand, if in the RX path a nop
is to be sent, for example in iscsi_handle_reject or __iscsi_complete_pdu
than the forward lock is released and the backward lock is taken for the
duration of iscsi_send_nopout, later the backward lock is released and the
forward lock is retaken.
libiscsi_tcp uses two kernel fifos the r2t pool and the r2t queue.
The insertion and deletion from these queues didn't corespond to the
assumption taken by the new forward/backwards session locking paradigm.
That is, in iscsi_tcp_clenup_task which belongs to the RX (backwards)
path, r2t is taken out from r2t queue and inserted to the r2t pool.
In iscsi_tcp_get_curr_r2t which belong to the TX (forward) path, r2t
is also inserted to the r2t pool and another r2t is pulled from r2t
queue.
Only in iscsi_tcp_r2t_rsp which is called in the RX path but can requeue
to the TX path, r2t is taken from the r2t pool and inserted to the r2t
queue.
In order to cope with this situation, two spin locks were added,
pool2queue and queue2pool. The former protects extracting from the
r2t pool and inserting to the r2t queue, and the later protects the
extracing from the r2t queue and inserting to the r2t pool.
Signed-off-by: Shlomo Pongratz <shlomop@mellanox.com>
Signed-off-by: Or Gerlitz <ogerlitz@mellanox.com>
[minor fix up to apply cleanly and compile fix]
Signed-off-by: Mike Christie <michaelc@cs.wisc.edu>
Signed-off-by: James Bottomley <JBottomley@Parallels.com>
2014-02-07 13:41:38 +07:00
|
|
|
spin_unlock_bh(&session->back_lock);
|
2009-09-05 09:06:35 +07:00
|
|
|
}
|
|
|
|
|
|
|
|
static struct list_head *hwi_get_async_busy_list(struct hwi_async_pdu_context
|
|
|
|
*pasync_ctx, unsigned int is_header,
|
|
|
|
unsigned int host_write_ptr)
|
|
|
|
{
|
|
|
|
if (is_header)
|
|
|
|
return &pasync_ctx->async_entry[host_write_ptr].
|
|
|
|
header_busy_list;
|
|
|
|
else
|
|
|
|
return &pasync_ctx->async_entry[host_write_ptr].data_busy_list;
|
|
|
|
}
|
|
|
|
|
|
|
|
static struct async_pdu_handle *
|
|
|
|
hwi_get_async_handle(struct beiscsi_hba *phba,
|
|
|
|
struct beiscsi_conn *beiscsi_conn,
|
|
|
|
struct hwi_async_pdu_context *pasync_ctx,
|
|
|
|
struct i_t_dpdu_cqe *pdpdu_cqe, unsigned int *pcq_index)
|
|
|
|
{
|
|
|
|
struct be_bus_address phys_addr;
|
|
|
|
struct list_head *pbusy_list;
|
|
|
|
struct async_pdu_handle *pasync_handle = NULL;
|
|
|
|
unsigned char is_header = 0;
|
2012-10-20 06:14:49 +07:00
|
|
|
unsigned int index, dpl;
|
|
|
|
|
2013-04-06 10:38:26 +07:00
|
|
|
if (is_chip_be2_be3r(phba)) {
|
|
|
|
dpl = AMAP_GET_BITS(struct amap_i_t_dpdu_cqe,
|
2012-10-20 06:14:49 +07:00
|
|
|
dpl, pdpdu_cqe);
|
2013-04-06 10:38:26 +07:00
|
|
|
index = AMAP_GET_BITS(struct amap_i_t_dpdu_cqe,
|
2012-10-20 06:14:49 +07:00
|
|
|
index, pdpdu_cqe);
|
|
|
|
} else {
|
2013-04-06 10:38:26 +07:00
|
|
|
dpl = AMAP_GET_BITS(struct amap_i_t_dpdu_cqe_v2,
|
2012-10-20 06:14:49 +07:00
|
|
|
dpl, pdpdu_cqe);
|
2013-04-06 10:38:26 +07:00
|
|
|
index = AMAP_GET_BITS(struct amap_i_t_dpdu_cqe_v2,
|
2012-10-20 06:14:49 +07:00
|
|
|
index, pdpdu_cqe);
|
|
|
|
}
|
2009-09-05 09:06:35 +07:00
|
|
|
|
|
|
|
phys_addr.u.a32.address_lo =
|
2012-10-20 06:14:49 +07:00
|
|
|
(pdpdu_cqe->dw[offsetof(struct amap_i_t_dpdu_cqe,
|
|
|
|
db_addr_lo) / 32] - dpl);
|
2009-09-05 09:06:35 +07:00
|
|
|
phys_addr.u.a32.address_hi =
|
2012-10-20 06:14:49 +07:00
|
|
|
pdpdu_cqe->dw[offsetof(struct amap_i_t_dpdu_cqe,
|
|
|
|
db_addr_hi) / 32];
|
2009-09-05 09:06:35 +07:00
|
|
|
|
|
|
|
phys_addr.u.a64.address =
|
|
|
|
*((unsigned long long *)(&phys_addr.u.a64.address));
|
|
|
|
|
|
|
|
switch (pdpdu_cqe->dw[offsetof(struct amap_i_t_dpdu_cqe, code) / 32]
|
|
|
|
& PDUCQE_CODE_MASK) {
|
|
|
|
case UNSOL_HDR_NOTIFY:
|
|
|
|
is_header = 1;
|
|
|
|
|
2012-10-20 06:14:49 +07:00
|
|
|
pbusy_list = hwi_get_async_busy_list(pasync_ctx,
|
|
|
|
is_header, index);
|
2009-09-05 09:06:35 +07:00
|
|
|
break;
|
|
|
|
case UNSOL_DATA_NOTIFY:
|
2012-10-20 06:14:49 +07:00
|
|
|
pbusy_list = hwi_get_async_busy_list(pasync_ctx,
|
|
|
|
is_header, index);
|
2009-09-05 09:06:35 +07:00
|
|
|
break;
|
|
|
|
default:
|
|
|
|
pbusy_list = NULL;
|
2012-08-21 00:30:18 +07:00
|
|
|
beiscsi_log(phba, KERN_WARNING,
|
|
|
|
BEISCSI_LOG_IO | BEISCSI_LOG_CONFIG,
|
|
|
|
"BM_%d : Unexpected code=%d\n",
|
|
|
|
pdpdu_cqe->dw[offsetof(struct amap_i_t_dpdu_cqe,
|
|
|
|
code) / 32] & PDUCQE_CODE_MASK);
|
2009-09-05 09:06:35 +07:00
|
|
|
return NULL;
|
|
|
|
}
|
|
|
|
|
|
|
|
WARN_ON(list_empty(pbusy_list));
|
|
|
|
list_for_each_entry(pasync_handle, pbusy_list, link) {
|
2012-04-04 11:41:36 +07:00
|
|
|
if (pasync_handle->pa.u.a64.address == phys_addr.u.a64.address)
|
2009-09-05 09:06:35 +07:00
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
WARN_ON(!pasync_handle);
|
|
|
|
|
2013-09-29 05:35:45 +07:00
|
|
|
pasync_handle->cri = BE_GET_ASYNC_CRI_FROM_CID(
|
|
|
|
beiscsi_conn->beiscsi_conn_cid);
|
2009-09-05 09:06:35 +07:00
|
|
|
pasync_handle->is_header = is_header;
|
2012-10-20 06:14:49 +07:00
|
|
|
pasync_handle->buffer_len = dpl;
|
|
|
|
*pcq_index = index;
|
2009-09-05 09:06:35 +07:00
|
|
|
|
|
|
|
return pasync_handle;
|
|
|
|
}
|
|
|
|
|
|
|
|
static unsigned int
|
2012-08-21 00:30:18 +07:00
|
|
|
hwi_update_async_writables(struct beiscsi_hba *phba,
|
|
|
|
struct hwi_async_pdu_context *pasync_ctx,
|
|
|
|
unsigned int is_header, unsigned int cq_index)
|
2009-09-05 09:06:35 +07:00
|
|
|
{
|
|
|
|
struct list_head *pbusy_list;
|
|
|
|
struct async_pdu_handle *pasync_handle;
|
|
|
|
unsigned int num_entries, writables = 0;
|
|
|
|
unsigned int *pep_read_ptr, *pwritables;
|
|
|
|
|
2012-04-04 11:41:36 +07:00
|
|
|
num_entries = pasync_ctx->num_entries;
|
2009-09-05 09:06:35 +07:00
|
|
|
if (is_header) {
|
|
|
|
pep_read_ptr = &pasync_ctx->async_header.ep_read_ptr;
|
|
|
|
pwritables = &pasync_ctx->async_header.writables;
|
|
|
|
} else {
|
|
|
|
pep_read_ptr = &pasync_ctx->async_data.ep_read_ptr;
|
|
|
|
pwritables = &pasync_ctx->async_data.writables;
|
|
|
|
}
|
|
|
|
|
|
|
|
while ((*pep_read_ptr) != cq_index) {
|
|
|
|
(*pep_read_ptr)++;
|
|
|
|
*pep_read_ptr = (*pep_read_ptr) % num_entries;
|
|
|
|
|
|
|
|
pbusy_list = hwi_get_async_busy_list(pasync_ctx, is_header,
|
|
|
|
*pep_read_ptr);
|
|
|
|
if (writables == 0)
|
|
|
|
WARN_ON(list_empty(pbusy_list));
|
|
|
|
|
|
|
|
if (!list_empty(pbusy_list)) {
|
|
|
|
pasync_handle = list_entry(pbusy_list->next,
|
|
|
|
struct async_pdu_handle,
|
|
|
|
link);
|
|
|
|
WARN_ON(!pasync_handle);
|
|
|
|
pasync_handle->consumed = 1;
|
|
|
|
}
|
|
|
|
|
|
|
|
writables++;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (!writables) {
|
2012-08-21 00:30:18 +07:00
|
|
|
beiscsi_log(phba, KERN_ERR,
|
|
|
|
BEISCSI_LOG_CONFIG | BEISCSI_LOG_IO,
|
|
|
|
"BM_%d : Duplicate notification received - index 0x%x!!\n",
|
|
|
|
cq_index);
|
2009-09-05 09:06:35 +07:00
|
|
|
WARN_ON(1);
|
|
|
|
}
|
|
|
|
|
|
|
|
*pwritables = *pwritables + writables;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2012-04-04 11:41:47 +07:00
|
|
|
static void hwi_free_async_msg(struct beiscsi_hba *phba,
|
2013-09-29 05:35:45 +07:00
|
|
|
struct hwi_async_pdu_context *pasync_ctx,
|
|
|
|
unsigned int cri)
|
2009-09-05 09:06:35 +07:00
|
|
|
{
|
|
|
|
struct async_pdu_handle *pasync_handle, *tmp_handle;
|
|
|
|
struct list_head *plist;
|
|
|
|
|
|
|
|
plist = &pasync_ctx->async_entry[cri].wait_queue.list;
|
|
|
|
list_for_each_entry_safe(pasync_handle, tmp_handle, plist, link) {
|
|
|
|
list_del(&pasync_handle->link);
|
|
|
|
|
2012-04-04 11:41:47 +07:00
|
|
|
if (pasync_handle->is_header) {
|
2009-09-05 09:06:35 +07:00
|
|
|
list_add_tail(&pasync_handle->link,
|
|
|
|
&pasync_ctx->async_header.free_list);
|
|
|
|
pasync_ctx->async_header.free_entries++;
|
|
|
|
} else {
|
|
|
|
list_add_tail(&pasync_handle->link,
|
|
|
|
&pasync_ctx->async_data.free_list);
|
|
|
|
pasync_ctx->async_data.free_entries++;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
INIT_LIST_HEAD(&pasync_ctx->async_entry[cri].wait_queue.list);
|
|
|
|
pasync_ctx->async_entry[cri].wait_queue.hdr_received = 0;
|
|
|
|
pasync_ctx->async_entry[cri].wait_queue.bytes_received = 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static struct phys_addr *
|
|
|
|
hwi_get_ring_address(struct hwi_async_pdu_context *pasync_ctx,
|
|
|
|
unsigned int is_header, unsigned int host_write_ptr)
|
|
|
|
{
|
|
|
|
struct phys_addr *pasync_sge = NULL;
|
|
|
|
|
|
|
|
if (is_header)
|
|
|
|
pasync_sge = pasync_ctx->async_header.ring_base;
|
|
|
|
else
|
|
|
|
pasync_sge = pasync_ctx->async_data.ring_base;
|
|
|
|
|
|
|
|
return pasync_sge + host_write_ptr;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void hwi_post_async_buffers(struct beiscsi_hba *phba,
|
2013-09-29 05:35:45 +07:00
|
|
|
unsigned int is_header, uint8_t ulp_num)
|
2009-09-05 09:06:35 +07:00
|
|
|
{
|
|
|
|
struct hwi_controller *phwi_ctrlr;
|
|
|
|
struct hwi_async_pdu_context *pasync_ctx;
|
|
|
|
struct async_pdu_handle *pasync_handle;
|
|
|
|
struct list_head *pfree_link, *pbusy_list;
|
|
|
|
struct phys_addr *pasync_sge;
|
|
|
|
unsigned int ring_id, num_entries;
|
2013-09-29 05:35:45 +07:00
|
|
|
unsigned int host_write_num, doorbell_offset;
|
2009-09-05 09:06:35 +07:00
|
|
|
unsigned int writables;
|
|
|
|
unsigned int i = 0;
|
|
|
|
u32 doorbell = 0;
|
|
|
|
|
|
|
|
phwi_ctrlr = phba->phwi_ctrlr;
|
2013-09-29 05:35:45 +07:00
|
|
|
pasync_ctx = HWI_GET_ASYNC_PDU_CTX(phwi_ctrlr, ulp_num);
|
2012-04-04 11:41:36 +07:00
|
|
|
num_entries = pasync_ctx->num_entries;
|
2009-09-05 09:06:35 +07:00
|
|
|
|
|
|
|
if (is_header) {
|
|
|
|
writables = min(pasync_ctx->async_header.writables,
|
|
|
|
pasync_ctx->async_header.free_entries);
|
|
|
|
pfree_link = pasync_ctx->async_header.free_list.next;
|
|
|
|
host_write_num = pasync_ctx->async_header.host_write_ptr;
|
2013-09-29 05:35:45 +07:00
|
|
|
ring_id = phwi_ctrlr->default_pdu_hdr[ulp_num].id;
|
|
|
|
doorbell_offset = phwi_ctrlr->default_pdu_hdr[ulp_num].
|
|
|
|
doorbell_offset;
|
2009-09-05 09:06:35 +07:00
|
|
|
} else {
|
|
|
|
writables = min(pasync_ctx->async_data.writables,
|
|
|
|
pasync_ctx->async_data.free_entries);
|
|
|
|
pfree_link = pasync_ctx->async_data.free_list.next;
|
|
|
|
host_write_num = pasync_ctx->async_data.host_write_ptr;
|
2013-09-29 05:35:45 +07:00
|
|
|
ring_id = phwi_ctrlr->default_pdu_data[ulp_num].id;
|
|
|
|
doorbell_offset = phwi_ctrlr->default_pdu_data[ulp_num].
|
|
|
|
doorbell_offset;
|
2009-09-05 09:06:35 +07:00
|
|
|
}
|
|
|
|
|
|
|
|
writables = (writables / 8) * 8;
|
|
|
|
if (writables) {
|
|
|
|
for (i = 0; i < writables; i++) {
|
|
|
|
pbusy_list =
|
|
|
|
hwi_get_async_busy_list(pasync_ctx, is_header,
|
|
|
|
host_write_num);
|
|
|
|
pasync_handle =
|
|
|
|
list_entry(pfree_link, struct async_pdu_handle,
|
|
|
|
link);
|
|
|
|
WARN_ON(!pasync_handle);
|
|
|
|
pasync_handle->consumed = 0;
|
|
|
|
|
|
|
|
pfree_link = pfree_link->next;
|
|
|
|
|
|
|
|
pasync_sge = hwi_get_ring_address(pasync_ctx,
|
|
|
|
is_header, host_write_num);
|
|
|
|
|
|
|
|
pasync_sge->hi = pasync_handle->pa.u.a32.address_lo;
|
|
|
|
pasync_sge->lo = pasync_handle->pa.u.a32.address_hi;
|
|
|
|
|
|
|
|
list_move(&pasync_handle->link, pbusy_list);
|
|
|
|
|
|
|
|
host_write_num++;
|
|
|
|
host_write_num = host_write_num % num_entries;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (is_header) {
|
|
|
|
pasync_ctx->async_header.host_write_ptr =
|
|
|
|
host_write_num;
|
|
|
|
pasync_ctx->async_header.free_entries -= writables;
|
|
|
|
pasync_ctx->async_header.writables -= writables;
|
|
|
|
pasync_ctx->async_header.busy_entries += writables;
|
|
|
|
} else {
|
|
|
|
pasync_ctx->async_data.host_write_ptr = host_write_num;
|
|
|
|
pasync_ctx->async_data.free_entries -= writables;
|
|
|
|
pasync_ctx->async_data.writables -= writables;
|
|
|
|
pasync_ctx->async_data.busy_entries += writables;
|
|
|
|
}
|
|
|
|
|
|
|
|
doorbell |= ring_id & DB_DEF_PDU_RING_ID_MASK;
|
|
|
|
doorbell |= 1 << DB_DEF_PDU_REARM_SHIFT;
|
|
|
|
doorbell |= 0 << DB_DEF_PDU_EVENT_SHIFT;
|
|
|
|
doorbell |= (writables & DB_DEF_PDU_CQPROC_MASK)
|
|
|
|
<< DB_DEF_PDU_CQPROC_SHIFT;
|
|
|
|
|
2013-09-29 05:35:45 +07:00
|
|
|
iowrite32(doorbell, phba->db_va + doorbell_offset);
|
2009-09-05 09:06:35 +07:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
static void hwi_flush_default_pdu_buffer(struct beiscsi_hba *phba,
|
|
|
|
struct beiscsi_conn *beiscsi_conn,
|
|
|
|
struct i_t_dpdu_cqe *pdpdu_cqe)
|
|
|
|
{
|
|
|
|
struct hwi_controller *phwi_ctrlr;
|
|
|
|
struct hwi_async_pdu_context *pasync_ctx;
|
|
|
|
struct async_pdu_handle *pasync_handle = NULL;
|
|
|
|
unsigned int cq_index = -1;
|
2013-09-29 05:35:45 +07:00
|
|
|
uint16_t cri_index = BE_GET_CRI_FROM_CID(
|
|
|
|
beiscsi_conn->beiscsi_conn_cid);
|
2009-09-05 09:06:35 +07:00
|
|
|
|
|
|
|
phwi_ctrlr = phba->phwi_ctrlr;
|
2013-09-29 05:35:45 +07:00
|
|
|
pasync_ctx = HWI_GET_ASYNC_PDU_CTX(phwi_ctrlr,
|
|
|
|
BEISCSI_GET_ULP_FROM_CRI(phwi_ctrlr,
|
|
|
|
cri_index));
|
2009-09-05 09:06:35 +07:00
|
|
|
|
|
|
|
pasync_handle = hwi_get_async_handle(phba, beiscsi_conn, pasync_ctx,
|
|
|
|
pdpdu_cqe, &cq_index);
|
|
|
|
BUG_ON(pasync_handle->is_header != 0);
|
|
|
|
if (pasync_handle->consumed == 0)
|
2012-08-21 00:30:18 +07:00
|
|
|
hwi_update_async_writables(phba, pasync_ctx,
|
|
|
|
pasync_handle->is_header, cq_index);
|
2009-09-05 09:06:35 +07:00
|
|
|
|
2013-09-29 05:35:45 +07:00
|
|
|
hwi_free_async_msg(phba, pasync_ctx, pasync_handle->cri);
|
|
|
|
hwi_post_async_buffers(phba, pasync_handle->is_header,
|
|
|
|
BEISCSI_GET_ULP_FROM_CRI(phwi_ctrlr,
|
|
|
|
cri_index));
|
2009-09-05 09:06:35 +07:00
|
|
|
}
|
|
|
|
|
|
|
|
static unsigned int
|
|
|
|
hwi_fwd_async_msg(struct beiscsi_conn *beiscsi_conn,
|
|
|
|
struct beiscsi_hba *phba,
|
|
|
|
struct hwi_async_pdu_context *pasync_ctx, unsigned short cri)
|
|
|
|
{
|
|
|
|
struct list_head *plist;
|
|
|
|
struct async_pdu_handle *pasync_handle;
|
|
|
|
void *phdr = NULL;
|
|
|
|
unsigned int hdr_len = 0, buf_len = 0;
|
|
|
|
unsigned int status, index = 0, offset = 0;
|
|
|
|
void *pfirst_buffer = NULL;
|
|
|
|
unsigned int num_buf = 0;
|
|
|
|
|
|
|
|
plist = &pasync_ctx->async_entry[cri].wait_queue.list;
|
|
|
|
|
|
|
|
list_for_each_entry(pasync_handle, plist, link) {
|
|
|
|
if (index == 0) {
|
|
|
|
phdr = pasync_handle->pbuffer;
|
|
|
|
hdr_len = pasync_handle->buffer_len;
|
|
|
|
} else {
|
|
|
|
buf_len = pasync_handle->buffer_len;
|
|
|
|
if (!num_buf) {
|
|
|
|
pfirst_buffer = pasync_handle->pbuffer;
|
|
|
|
num_buf++;
|
|
|
|
}
|
|
|
|
memcpy(pfirst_buffer + offset,
|
|
|
|
pasync_handle->pbuffer, buf_len);
|
2012-04-04 11:41:37 +07:00
|
|
|
offset += buf_len;
|
2009-09-05 09:06:35 +07:00
|
|
|
}
|
|
|
|
index++;
|
|
|
|
}
|
|
|
|
|
|
|
|
status = beiscsi_process_async_pdu(beiscsi_conn, phba,
|
2010-01-05 06:34:12 +07:00
|
|
|
phdr, hdr_len, pfirst_buffer,
|
2012-04-04 11:41:37 +07:00
|
|
|
offset);
|
2009-09-05 09:06:35 +07:00
|
|
|
|
2013-09-29 05:35:45 +07:00
|
|
|
hwi_free_async_msg(phba, pasync_ctx, cri);
|
2009-09-05 09:06:35 +07:00
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static unsigned int
|
|
|
|
hwi_gather_async_pdu(struct beiscsi_conn *beiscsi_conn,
|
|
|
|
struct beiscsi_hba *phba,
|
|
|
|
struct async_pdu_handle *pasync_handle)
|
|
|
|
{
|
|
|
|
struct hwi_async_pdu_context *pasync_ctx;
|
|
|
|
struct hwi_controller *phwi_ctrlr;
|
|
|
|
unsigned int bytes_needed = 0, status = 0;
|
|
|
|
unsigned short cri = pasync_handle->cri;
|
|
|
|
struct pdu_base *ppdu;
|
|
|
|
|
|
|
|
phwi_ctrlr = phba->phwi_ctrlr;
|
2013-09-29 05:35:45 +07:00
|
|
|
pasync_ctx = HWI_GET_ASYNC_PDU_CTX(phwi_ctrlr,
|
|
|
|
BEISCSI_GET_ULP_FROM_CRI(phwi_ctrlr,
|
|
|
|
BE_GET_CRI_FROM_CID(beiscsi_conn->
|
|
|
|
beiscsi_conn_cid)));
|
2009-09-05 09:06:35 +07:00
|
|
|
|
|
|
|
list_del(&pasync_handle->link);
|
|
|
|
if (pasync_handle->is_header) {
|
|
|
|
pasync_ctx->async_header.busy_entries--;
|
|
|
|
if (pasync_ctx->async_entry[cri].wait_queue.hdr_received) {
|
2013-09-29 05:35:45 +07:00
|
|
|
hwi_free_async_msg(phba, pasync_ctx, cri);
|
2009-09-05 09:06:35 +07:00
|
|
|
BUG();
|
|
|
|
}
|
|
|
|
|
|
|
|
pasync_ctx->async_entry[cri].wait_queue.bytes_received = 0;
|
|
|
|
pasync_ctx->async_entry[cri].wait_queue.hdr_received = 1;
|
|
|
|
pasync_ctx->async_entry[cri].wait_queue.hdr_len =
|
|
|
|
(unsigned short)pasync_handle->buffer_len;
|
|
|
|
list_add_tail(&pasync_handle->link,
|
|
|
|
&pasync_ctx->async_entry[cri].wait_queue.list);
|
|
|
|
|
|
|
|
ppdu = pasync_handle->pbuffer;
|
|
|
|
bytes_needed = ((((ppdu->dw[offsetof(struct amap_pdu_base,
|
|
|
|
data_len_hi) / 32] & PDUBASE_DATALENHI_MASK) << 8) &
|
|
|
|
0xFFFF0000) | ((be16_to_cpu((ppdu->
|
|
|
|
dw[offsetof(struct amap_pdu_base, data_len_lo) / 32]
|
|
|
|
& PDUBASE_DATALENLO_MASK) >> 16)) & 0x0000FFFF));
|
|
|
|
|
|
|
|
if (status == 0) {
|
|
|
|
pasync_ctx->async_entry[cri].wait_queue.bytes_needed =
|
|
|
|
bytes_needed;
|
|
|
|
|
|
|
|
if (bytes_needed == 0)
|
|
|
|
status = hwi_fwd_async_msg(beiscsi_conn, phba,
|
|
|
|
pasync_ctx, cri);
|
|
|
|
}
|
|
|
|
} else {
|
|
|
|
pasync_ctx->async_data.busy_entries--;
|
|
|
|
if (pasync_ctx->async_entry[cri].wait_queue.hdr_received) {
|
|
|
|
list_add_tail(&pasync_handle->link,
|
|
|
|
&pasync_ctx->async_entry[cri].wait_queue.
|
|
|
|
list);
|
|
|
|
pasync_ctx->async_entry[cri].wait_queue.
|
|
|
|
bytes_received +=
|
|
|
|
(unsigned short)pasync_handle->buffer_len;
|
|
|
|
|
|
|
|
if (pasync_ctx->async_entry[cri].wait_queue.
|
|
|
|
bytes_received >=
|
|
|
|
pasync_ctx->async_entry[cri].wait_queue.
|
|
|
|
bytes_needed)
|
|
|
|
status = hwi_fwd_async_msg(beiscsi_conn, phba,
|
|
|
|
pasync_ctx, cri);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
return status;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void hwi_process_default_pdu_ring(struct beiscsi_conn *beiscsi_conn,
|
|
|
|
struct beiscsi_hba *phba,
|
|
|
|
struct i_t_dpdu_cqe *pdpdu_cqe)
|
|
|
|
{
|
|
|
|
struct hwi_controller *phwi_ctrlr;
|
|
|
|
struct hwi_async_pdu_context *pasync_ctx;
|
|
|
|
struct async_pdu_handle *pasync_handle = NULL;
|
|
|
|
unsigned int cq_index = -1;
|
2013-09-29 05:35:45 +07:00
|
|
|
uint16_t cri_index = BE_GET_CRI_FROM_CID(
|
|
|
|
beiscsi_conn->beiscsi_conn_cid);
|
2009-09-05 09:06:35 +07:00
|
|
|
|
|
|
|
phwi_ctrlr = phba->phwi_ctrlr;
|
2013-09-29 05:35:45 +07:00
|
|
|
pasync_ctx = HWI_GET_ASYNC_PDU_CTX(phwi_ctrlr,
|
|
|
|
BEISCSI_GET_ULP_FROM_CRI(phwi_ctrlr,
|
|
|
|
cri_index));
|
|
|
|
|
2009-09-05 09:06:35 +07:00
|
|
|
pasync_handle = hwi_get_async_handle(phba, beiscsi_conn, pasync_ctx,
|
|
|
|
pdpdu_cqe, &cq_index);
|
|
|
|
|
|
|
|
if (pasync_handle->consumed == 0)
|
2012-08-21 00:30:18 +07:00
|
|
|
hwi_update_async_writables(phba, pasync_ctx,
|
|
|
|
pasync_handle->is_header, cq_index);
|
|
|
|
|
2009-09-05 09:06:35 +07:00
|
|
|
hwi_gather_async_pdu(beiscsi_conn, phba, pasync_handle);
|
2013-09-29 05:35:45 +07:00
|
|
|
hwi_post_async_buffers(phba, pasync_handle->is_header,
|
|
|
|
BEISCSI_GET_ULP_FROM_CRI(
|
|
|
|
phwi_ctrlr, cri_index));
|
2009-09-05 09:06:35 +07:00
|
|
|
}
|
|
|
|
|
2016-02-04 17:19:11 +07:00
|
|
|
void beiscsi_process_mcc_cq(struct beiscsi_hba *phba)
|
2010-01-05 06:40:46 +07:00
|
|
|
{
|
|
|
|
struct be_queue_info *mcc_cq;
|
|
|
|
struct be_mcc_compl *mcc_compl;
|
|
|
|
unsigned int num_processed = 0;
|
|
|
|
|
|
|
|
mcc_cq = &phba->ctrl.mcc_obj.cq;
|
|
|
|
mcc_compl = queue_tail_node(mcc_cq);
|
|
|
|
mcc_compl->flags = le32_to_cpu(mcc_compl->flags);
|
|
|
|
while (mcc_compl->flags & CQE_FLAGS_VALID_MASK) {
|
2016-08-19 16:50:11 +07:00
|
|
|
if (beiscsi_hba_in_error(phba))
|
|
|
|
return;
|
|
|
|
|
2010-01-05 06:40:46 +07:00
|
|
|
if (num_processed >= 32) {
|
|
|
|
hwi_ring_cq_db(phba, mcc_cq->id,
|
2016-01-20 15:40:56 +07:00
|
|
|
num_processed, 0);
|
2010-01-05 06:40:46 +07:00
|
|
|
num_processed = 0;
|
|
|
|
}
|
|
|
|
if (mcc_compl->flags & CQE_FLAGS_ASYNC_MASK) {
|
2016-01-20 15:40:53 +07:00
|
|
|
beiscsi_process_async_event(phba, mcc_compl);
|
2010-01-05 06:40:46 +07:00
|
|
|
} else if (mcc_compl->flags & CQE_FLAGS_COMPLETED_MASK) {
|
2016-02-04 17:19:11 +07:00
|
|
|
beiscsi_process_mcc_compl(&phba->ctrl, mcc_compl);
|
2010-01-05 06:40:46 +07:00
|
|
|
}
|
|
|
|
|
|
|
|
mcc_compl->flags = 0;
|
|
|
|
queue_tail_inc(mcc_cq);
|
|
|
|
mcc_compl = queue_tail_node(mcc_cq);
|
|
|
|
mcc_compl->flags = le32_to_cpu(mcc_compl->flags);
|
|
|
|
num_processed++;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (num_processed > 0)
|
2016-01-20 15:40:56 +07:00
|
|
|
hwi_ring_cq_db(phba, mcc_cq->id, num_processed, 1);
|
2010-01-05 06:40:46 +07:00
|
|
|
}
|
2009-10-23 13:22:33 +07:00
|
|
|
|
2016-08-19 16:50:10 +07:00
|
|
|
static void beiscsi_mcc_work(struct work_struct *work)
|
|
|
|
{
|
|
|
|
struct be_eq_obj *pbe_eq;
|
|
|
|
struct beiscsi_hba *phba;
|
|
|
|
|
|
|
|
pbe_eq = container_of(work, struct be_eq_obj, mcc_work);
|
|
|
|
phba = pbe_eq->phba;
|
|
|
|
beiscsi_process_mcc_cq(phba);
|
|
|
|
/* rearm EQ for further interrupts */
|
2016-08-19 16:50:11 +07:00
|
|
|
if (!beiscsi_hba_in_error(phba))
|
|
|
|
hwi_ring_eq_db(phba, pbe_eq->q.id, 0, 0, 1, 1);
|
2016-08-19 16:50:10 +07:00
|
|
|
}
|
|
|
|
|
2012-10-20 06:11:45 +07:00
|
|
|
/**
|
|
|
|
* beiscsi_process_cq()- Process the Completion Queue
|
|
|
|
* @pbe_eq: Event Q on which the Completion has come
|
2016-01-20 15:40:56 +07:00
|
|
|
* @budget: Max number of events to processed
|
2012-10-20 06:11:45 +07:00
|
|
|
*
|
|
|
|
* return
|
|
|
|
* Number of Completion Entries processed.
|
|
|
|
**/
|
2016-01-20 15:40:56 +07:00
|
|
|
unsigned int beiscsi_process_cq(struct be_eq_obj *pbe_eq, int budget)
|
2009-09-05 09:06:35 +07:00
|
|
|
{
|
|
|
|
struct be_queue_info *cq;
|
|
|
|
struct sol_cqe *sol;
|
|
|
|
struct dmsg_cqe *dmsg;
|
2016-01-20 15:40:56 +07:00
|
|
|
unsigned int total = 0;
|
2009-09-05 09:06:35 +07:00
|
|
|
unsigned int num_processed = 0;
|
2012-08-21 00:30:55 +07:00
|
|
|
unsigned short code = 0, cid = 0;
|
2013-04-06 10:38:32 +07:00
|
|
|
uint16_t cri_index = 0;
|
2009-09-05 09:06:35 +07:00
|
|
|
struct beiscsi_conn *beiscsi_conn;
|
2010-01-05 06:35:34 +07:00
|
|
|
struct beiscsi_endpoint *beiscsi_ep;
|
|
|
|
struct iscsi_endpoint *ep;
|
2009-10-23 13:22:33 +07:00
|
|
|
struct beiscsi_hba *phba;
|
2009-09-05 09:06:35 +07:00
|
|
|
|
2009-10-23 13:22:33 +07:00
|
|
|
cq = pbe_eq->cq;
|
2009-09-05 09:06:35 +07:00
|
|
|
sol = queue_tail_node(cq);
|
2009-10-23 13:22:33 +07:00
|
|
|
phba = pbe_eq->phba;
|
2009-09-05 09:06:35 +07:00
|
|
|
|
|
|
|
while (sol->dw[offsetof(struct amap_sol_cqe, valid) / 32] &
|
|
|
|
CQE_VALID_MASK) {
|
2016-08-19 16:50:11 +07:00
|
|
|
if (beiscsi_hba_in_error(phba))
|
|
|
|
return 0;
|
|
|
|
|
2009-09-05 09:06:35 +07:00
|
|
|
be_dws_le_to_cpu(sol, sizeof(struct sol_cqe));
|
|
|
|
|
2012-10-20 06:14:49 +07:00
|
|
|
code = (sol->dw[offsetof(struct amap_sol_cqe, code) /
|
|
|
|
32] & CQE_CODE_MASK);
|
|
|
|
|
|
|
|
/* Get the CID */
|
2013-04-06 10:38:26 +07:00
|
|
|
if (is_chip_be2_be3r(phba)) {
|
|
|
|
cid = AMAP_GET_BITS(struct amap_sol_cqe, cid, sol);
|
|
|
|
} else {
|
2012-10-20 06:14:49 +07:00
|
|
|
if ((code == DRIVERMSG_NOTIFY) ||
|
|
|
|
(code == UNSOL_HDR_NOTIFY) ||
|
|
|
|
(code == UNSOL_DATA_NOTIFY))
|
|
|
|
cid = AMAP_GET_BITS(
|
|
|
|
struct amap_i_t_dpdu_cqe_v2,
|
|
|
|
cid, sol);
|
|
|
|
else
|
|
|
|
cid = AMAP_GET_BITS(struct amap_sol_cqe_v2,
|
|
|
|
cid, sol);
|
2013-04-06 10:38:26 +07:00
|
|
|
}
|
2010-01-23 07:04:24 +07:00
|
|
|
|
2013-04-06 10:38:32 +07:00
|
|
|
cri_index = BE_GET_CRI_FROM_CID(cid);
|
|
|
|
ep = phba->ep_array[cri_index];
|
2014-08-08 12:00:01 +07:00
|
|
|
|
|
|
|
if (ep == NULL) {
|
|
|
|
/* connection has already been freed
|
|
|
|
* just move on to next one
|
|
|
|
*/
|
|
|
|
beiscsi_log(phba, KERN_WARNING,
|
|
|
|
BEISCSI_LOG_INIT,
|
|
|
|
"BM_%d : proc cqe of disconn ep: cid %d\n",
|
|
|
|
cid);
|
|
|
|
goto proc_next_cqe;
|
|
|
|
}
|
|
|
|
|
2010-01-05 06:35:34 +07:00
|
|
|
beiscsi_ep = ep->dd_data;
|
|
|
|
beiscsi_conn = beiscsi_ep->conn;
|
2010-01-05 06:40:46 +07:00
|
|
|
|
2016-01-20 15:40:56 +07:00
|
|
|
/* replenish cq */
|
|
|
|
if (num_processed == 32) {
|
|
|
|
hwi_ring_cq_db(phba, cq->id, 32, 0);
|
2009-09-05 09:06:35 +07:00
|
|
|
num_processed = 0;
|
|
|
|
}
|
2016-01-20 15:40:56 +07:00
|
|
|
total++;
|
2009-09-05 09:06:35 +07:00
|
|
|
|
2012-08-21 00:30:55 +07:00
|
|
|
switch (code) {
|
2009-09-05 09:06:35 +07:00
|
|
|
case SOL_CMD_COMPLETE:
|
|
|
|
hwi_complete_cmd(beiscsi_conn, phba, sol);
|
|
|
|
break;
|
|
|
|
case DRIVERMSG_NOTIFY:
|
2012-08-21 00:30:18 +07:00
|
|
|
beiscsi_log(phba, KERN_INFO,
|
|
|
|
BEISCSI_LOG_IO | BEISCSI_LOG_CONFIG,
|
2012-10-20 06:11:45 +07:00
|
|
|
"BM_%d : Received %s[%d] on CID : %d\n",
|
|
|
|
cqe_desc[code], code, cid);
|
2012-08-21 00:30:18 +07:00
|
|
|
|
2009-09-05 09:06:35 +07:00
|
|
|
dmsg = (struct dmsg_cqe *)sol;
|
|
|
|
hwi_complete_drvr_msgs(beiscsi_conn, phba, sol);
|
|
|
|
break;
|
|
|
|
case UNSOL_HDR_NOTIFY:
|
2012-08-21 00:30:18 +07:00
|
|
|
beiscsi_log(phba, KERN_INFO,
|
|
|
|
BEISCSI_LOG_IO | BEISCSI_LOG_CONFIG,
|
2012-10-20 06:11:45 +07:00
|
|
|
"BM_%d : Received %s[%d] on CID : %d\n",
|
|
|
|
cqe_desc[code], code, cid);
|
2012-08-21 00:30:18 +07:00
|
|
|
|
2013-09-29 05:35:42 +07:00
|
|
|
spin_lock_bh(&phba->async_pdu_lock);
|
2009-10-23 13:22:33 +07:00
|
|
|
hwi_process_default_pdu_ring(beiscsi_conn, phba,
|
|
|
|
(struct i_t_dpdu_cqe *)sol);
|
2013-09-29 05:35:42 +07:00
|
|
|
spin_unlock_bh(&phba->async_pdu_lock);
|
2009-10-23 13:22:33 +07:00
|
|
|
break;
|
2009-09-05 09:06:35 +07:00
|
|
|
case UNSOL_DATA_NOTIFY:
|
2012-08-21 00:30:18 +07:00
|
|
|
beiscsi_log(phba, KERN_INFO,
|
|
|
|
BEISCSI_LOG_CONFIG | BEISCSI_LOG_IO,
|
2012-10-20 06:11:45 +07:00
|
|
|
"BM_%d : Received %s[%d] on CID : %d\n",
|
|
|
|
cqe_desc[code], code, cid);
|
2012-08-21 00:30:18 +07:00
|
|
|
|
2013-09-29 05:35:42 +07:00
|
|
|
spin_lock_bh(&phba->async_pdu_lock);
|
2009-09-05 09:06:35 +07:00
|
|
|
hwi_process_default_pdu_ring(beiscsi_conn, phba,
|
|
|
|
(struct i_t_dpdu_cqe *)sol);
|
2013-09-29 05:35:42 +07:00
|
|
|
spin_unlock_bh(&phba->async_pdu_lock);
|
2009-09-05 09:06:35 +07:00
|
|
|
break;
|
|
|
|
case CXN_INVALIDATE_INDEX_NOTIFY:
|
|
|
|
case CMD_INVALIDATED_NOTIFY:
|
|
|
|
case CXN_INVALIDATE_NOTIFY:
|
2012-08-21 00:30:18 +07:00
|
|
|
beiscsi_log(phba, KERN_ERR,
|
|
|
|
BEISCSI_LOG_IO | BEISCSI_LOG_CONFIG,
|
2012-10-20 06:11:45 +07:00
|
|
|
"BM_%d : Ignoring %s[%d] on CID : %d\n",
|
|
|
|
cqe_desc[code], code, cid);
|
2009-09-05 09:06:35 +07:00
|
|
|
break;
|
2016-01-20 15:40:56 +07:00
|
|
|
case CXN_KILLED_HDR_DIGEST_ERR:
|
2009-09-05 09:06:35 +07:00
|
|
|
case SOL_CMD_KILLED_DATA_DIGEST_ERR:
|
2016-01-20 15:40:56 +07:00
|
|
|
beiscsi_log(phba, KERN_ERR,
|
|
|
|
BEISCSI_LOG_CONFIG | BEISCSI_LOG_IO,
|
|
|
|
"BM_%d : Cmd Notification %s[%d] on CID : %d\n",
|
|
|
|
cqe_desc[code], code, cid);
|
|
|
|
break;
|
2009-09-05 09:06:35 +07:00
|
|
|
case CMD_KILLED_INVALID_STATSN_RCVD:
|
|
|
|
case CMD_KILLED_INVALID_R2T_RCVD:
|
|
|
|
case CMD_CXN_KILLED_LUN_INVALID:
|
|
|
|
case CMD_CXN_KILLED_ICD_INVALID:
|
|
|
|
case CMD_CXN_KILLED_ITT_INVALID:
|
|
|
|
case CMD_CXN_KILLED_SEQ_OUTOFORDER:
|
|
|
|
case CMD_CXN_KILLED_INVALID_DATASN_RCVD:
|
2012-08-21 00:30:18 +07:00
|
|
|
beiscsi_log(phba, KERN_ERR,
|
|
|
|
BEISCSI_LOG_CONFIG | BEISCSI_LOG_IO,
|
2012-10-20 06:11:45 +07:00
|
|
|
"BM_%d : Cmd Notification %s[%d] on CID : %d\n",
|
|
|
|
cqe_desc[code], code, cid);
|
2009-09-05 09:06:35 +07:00
|
|
|
break;
|
|
|
|
case UNSOL_DATA_DIGEST_ERROR_NOTIFY:
|
2012-08-21 00:30:18 +07:00
|
|
|
beiscsi_log(phba, KERN_ERR,
|
|
|
|
BEISCSI_LOG_IO | BEISCSI_LOG_CONFIG,
|
2012-10-20 06:11:45 +07:00
|
|
|
"BM_%d : Dropping %s[%d] on DPDU ring on CID : %d\n",
|
|
|
|
cqe_desc[code], code, cid);
|
2013-09-29 05:35:42 +07:00
|
|
|
spin_lock_bh(&phba->async_pdu_lock);
|
2009-09-05 09:06:35 +07:00
|
|
|
hwi_flush_default_pdu_buffer(phba, beiscsi_conn,
|
|
|
|
(struct i_t_dpdu_cqe *) sol);
|
2013-09-29 05:35:42 +07:00
|
|
|
spin_unlock_bh(&phba->async_pdu_lock);
|
2009-09-05 09:06:35 +07:00
|
|
|
break;
|
|
|
|
case CXN_KILLED_PDU_SIZE_EXCEEDS_DSL:
|
|
|
|
case CXN_KILLED_BURST_LEN_MISMATCH:
|
|
|
|
case CXN_KILLED_AHS_RCVD:
|
|
|
|
case CXN_KILLED_UNKNOWN_HDR:
|
|
|
|
case CXN_KILLED_STALE_ITT_TTT_RCVD:
|
|
|
|
case CXN_KILLED_INVALID_ITT_TTT_RCVD:
|
|
|
|
case CXN_KILLED_TIMED_OUT:
|
|
|
|
case CXN_KILLED_FIN_RCVD:
|
2012-10-20 06:11:45 +07:00
|
|
|
case CXN_KILLED_RST_SENT:
|
|
|
|
case CXN_KILLED_RST_RCVD:
|
2009-09-05 09:06:35 +07:00
|
|
|
case CXN_KILLED_BAD_UNSOL_PDU_RCVD:
|
|
|
|
case CXN_KILLED_BAD_WRB_INDEX_ERROR:
|
|
|
|
case CXN_KILLED_OVER_RUN_RESIDUAL:
|
|
|
|
case CXN_KILLED_UNDER_RUN_RESIDUAL:
|
|
|
|
case CXN_KILLED_CMND_DATA_NOT_ON_SAME_CONN:
|
2012-08-21 00:30:18 +07:00
|
|
|
beiscsi_log(phba, KERN_ERR,
|
|
|
|
BEISCSI_LOG_IO | BEISCSI_LOG_CONFIG,
|
2012-10-20 06:11:45 +07:00
|
|
|
"BM_%d : Event %s[%d] received on CID : %d\n",
|
|
|
|
cqe_desc[code], code, cid);
|
2012-08-21 00:30:55 +07:00
|
|
|
if (beiscsi_conn)
|
|
|
|
iscsi_conn_failure(beiscsi_conn->conn,
|
|
|
|
ISCSI_ERR_CONN_FAILED);
|
2009-09-05 09:06:35 +07:00
|
|
|
break;
|
|
|
|
default:
|
2012-08-21 00:30:18 +07:00
|
|
|
beiscsi_log(phba, KERN_ERR,
|
|
|
|
BEISCSI_LOG_IO | BEISCSI_LOG_CONFIG,
|
2012-10-20 06:11:45 +07:00
|
|
|
"BM_%d : Invalid CQE Event Received Code : %d"
|
|
|
|
"CID 0x%x...\n",
|
2012-08-21 00:30:55 +07:00
|
|
|
code, cid);
|
2009-09-05 09:06:35 +07:00
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
2014-08-08 12:00:01 +07:00
|
|
|
proc_next_cqe:
|
2009-09-05 09:06:35 +07:00
|
|
|
AMAP_SET_BITS(struct amap_sol_cqe, valid, sol, 0);
|
|
|
|
queue_tail_inc(cq);
|
|
|
|
sol = queue_tail_node(cq);
|
|
|
|
num_processed++;
|
2016-01-20 15:40:56 +07:00
|
|
|
if (total == budget)
|
|
|
|
break;
|
2009-09-05 09:06:35 +07:00
|
|
|
}
|
|
|
|
|
2016-01-20 15:40:56 +07:00
|
|
|
hwi_ring_cq_db(phba, cq->id, num_processed, 1);
|
|
|
|
return total;
|
2009-09-05 09:06:35 +07:00
|
|
|
}
|
|
|
|
|
2015-11-10 20:56:14 +07:00
|
|
|
static int be_iopoll(struct irq_poll *iop, int budget)
|
2009-09-05 09:06:35 +07:00
|
|
|
{
|
2016-08-19 16:50:10 +07:00
|
|
|
unsigned int ret, io_events;
|
2009-09-05 09:06:35 +07:00
|
|
|
struct beiscsi_hba *phba;
|
2009-10-23 13:22:33 +07:00
|
|
|
struct be_eq_obj *pbe_eq;
|
2016-01-20 15:40:56 +07:00
|
|
|
struct be_eq_entry *eqe = NULL;
|
|
|
|
struct be_queue_info *eq;
|
2009-09-05 09:06:35 +07:00
|
|
|
|
2009-10-23 13:22:33 +07:00
|
|
|
pbe_eq = container_of(iop, struct be_eq_obj, iopoll);
|
2016-01-20 15:40:56 +07:00
|
|
|
phba = pbe_eq->phba;
|
2016-08-19 16:50:11 +07:00
|
|
|
if (beiscsi_hba_in_error(phba)) {
|
|
|
|
irq_poll_complete(iop);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
io_events = 0;
|
2016-01-20 15:40:56 +07:00
|
|
|
eq = &pbe_eq->q;
|
|
|
|
eqe = queue_tail_node(eq);
|
|
|
|
while (eqe->dw[offsetof(struct amap_eq_entry, valid) / 32] &
|
|
|
|
EQE_VALID_MASK) {
|
|
|
|
AMAP_SET_BITS(struct amap_eq_entry, valid, eqe, 0);
|
|
|
|
queue_tail_inc(eq);
|
|
|
|
eqe = queue_tail_node(eq);
|
2016-08-19 16:50:10 +07:00
|
|
|
io_events++;
|
2016-01-20 15:40:56 +07:00
|
|
|
}
|
2016-08-19 16:50:10 +07:00
|
|
|
hwi_ring_eq_db(phba, eq->id, 1, io_events, 0, 1);
|
2016-01-20 15:40:56 +07:00
|
|
|
|
|
|
|
ret = beiscsi_process_cq(pbe_eq, budget);
|
2014-05-06 08:41:26 +07:00
|
|
|
pbe_eq->cq_count += ret;
|
2009-09-05 09:06:35 +07:00
|
|
|
if (ret < budget) {
|
2015-11-10 20:56:14 +07:00
|
|
|
irq_poll_complete(iop);
|
2012-08-21 00:30:18 +07:00
|
|
|
beiscsi_log(phba, KERN_INFO,
|
|
|
|
BEISCSI_LOG_CONFIG | BEISCSI_LOG_IO,
|
2016-01-20 15:40:56 +07:00
|
|
|
"BM_%d : rearm pbe_eq->q.id =%d ret %d\n",
|
|
|
|
pbe_eq->q.id, ret);
|
2016-08-19 16:50:11 +07:00
|
|
|
if (!beiscsi_hba_in_error(phba))
|
|
|
|
hwi_ring_eq_db(phba, pbe_eq->q.id, 0, 0, 1, 1);
|
2009-09-05 09:06:35 +07:00
|
|
|
}
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
2012-10-20 06:14:23 +07:00
|
|
|
static void
|
|
|
|
hwi_write_sgl_v2(struct iscsi_wrb *pwrb, struct scatterlist *sg,
|
|
|
|
unsigned int num_sg, struct beiscsi_io_task *io_task)
|
|
|
|
{
|
|
|
|
struct iscsi_sge *psgl;
|
|
|
|
unsigned int sg_len, index;
|
|
|
|
unsigned int sge_len = 0;
|
|
|
|
unsigned long long addr;
|
|
|
|
struct scatterlist *l_sg;
|
|
|
|
unsigned int offset;
|
|
|
|
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_wrb_v2, iscsi_bhs_addr_lo, pwrb,
|
|
|
|
io_task->bhs_pa.u.a32.address_lo);
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_wrb_v2, iscsi_bhs_addr_hi, pwrb,
|
|
|
|
io_task->bhs_pa.u.a32.address_hi);
|
|
|
|
|
|
|
|
l_sg = sg;
|
|
|
|
for (index = 0; (index < num_sg) && (index < 2); index++,
|
|
|
|
sg = sg_next(sg)) {
|
|
|
|
if (index == 0) {
|
|
|
|
sg_len = sg_dma_len(sg);
|
|
|
|
addr = (u64) sg_dma_address(sg);
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_wrb_v2,
|
|
|
|
sge0_addr_lo, pwrb,
|
|
|
|
lower_32_bits(addr));
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_wrb_v2,
|
|
|
|
sge0_addr_hi, pwrb,
|
|
|
|
upper_32_bits(addr));
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_wrb_v2,
|
|
|
|
sge0_len, pwrb,
|
|
|
|
sg_len);
|
|
|
|
sge_len = sg_len;
|
|
|
|
} else {
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_wrb_v2, sge1_r2t_offset,
|
|
|
|
pwrb, sge_len);
|
|
|
|
sg_len = sg_dma_len(sg);
|
|
|
|
addr = (u64) sg_dma_address(sg);
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_wrb_v2,
|
|
|
|
sge1_addr_lo, pwrb,
|
|
|
|
lower_32_bits(addr));
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_wrb_v2,
|
|
|
|
sge1_addr_hi, pwrb,
|
|
|
|
upper_32_bits(addr));
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_wrb_v2,
|
|
|
|
sge1_len, pwrb,
|
|
|
|
sg_len);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
psgl = (struct iscsi_sge *)io_task->psgl_handle->pfrag;
|
|
|
|
memset(psgl, 0, sizeof(*psgl) * BE2_SGE);
|
|
|
|
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_sge, len, psgl, io_task->bhs_len - 2);
|
|
|
|
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_sge, addr_hi, psgl,
|
|
|
|
io_task->bhs_pa.u.a32.address_hi);
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_sge, addr_lo, psgl,
|
|
|
|
io_task->bhs_pa.u.a32.address_lo);
|
|
|
|
|
|
|
|
if (num_sg == 1) {
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_wrb_v2, sge0_last, pwrb,
|
|
|
|
1);
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_wrb_v2, sge1_last, pwrb,
|
|
|
|
0);
|
|
|
|
} else if (num_sg == 2) {
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_wrb_v2, sge0_last, pwrb,
|
|
|
|
0);
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_wrb_v2, sge1_last, pwrb,
|
|
|
|
1);
|
|
|
|
} else {
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_wrb_v2, sge0_last, pwrb,
|
|
|
|
0);
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_wrb_v2, sge1_last, pwrb,
|
|
|
|
0);
|
|
|
|
}
|
|
|
|
|
|
|
|
sg = l_sg;
|
|
|
|
psgl++;
|
|
|
|
psgl++;
|
|
|
|
offset = 0;
|
|
|
|
for (index = 0; index < num_sg; index++, sg = sg_next(sg), psgl++) {
|
|
|
|
sg_len = sg_dma_len(sg);
|
|
|
|
addr = (u64) sg_dma_address(sg);
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_sge, addr_lo, psgl,
|
|
|
|
lower_32_bits(addr));
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_sge, addr_hi, psgl,
|
|
|
|
upper_32_bits(addr));
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_sge, len, psgl, sg_len);
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_sge, sge_offset, psgl, offset);
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_sge, last_sge, psgl, 0);
|
|
|
|
offset += sg_len;
|
|
|
|
}
|
|
|
|
psgl--;
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_sge, last_sge, psgl, 1);
|
|
|
|
}
|
|
|
|
|
2009-09-05 09:06:35 +07:00
|
|
|
static void
|
|
|
|
hwi_write_sgl(struct iscsi_wrb *pwrb, struct scatterlist *sg,
|
|
|
|
unsigned int num_sg, struct beiscsi_io_task *io_task)
|
|
|
|
{
|
|
|
|
struct iscsi_sge *psgl;
|
2010-10-07 01:16:47 +07:00
|
|
|
unsigned int sg_len, index;
|
2009-09-05 09:06:35 +07:00
|
|
|
unsigned int sge_len = 0;
|
|
|
|
unsigned long long addr;
|
|
|
|
struct scatterlist *l_sg;
|
|
|
|
unsigned int offset;
|
|
|
|
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_wrb, iscsi_bhs_addr_lo, pwrb,
|
|
|
|
io_task->bhs_pa.u.a32.address_lo);
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_wrb, iscsi_bhs_addr_hi, pwrb,
|
|
|
|
io_task->bhs_pa.u.a32.address_hi);
|
|
|
|
|
|
|
|
l_sg = sg;
|
2010-01-07 03:20:19 +07:00
|
|
|
for (index = 0; (index < num_sg) && (index < 2); index++,
|
|
|
|
sg = sg_next(sg)) {
|
2009-09-05 09:06:35 +07:00
|
|
|
if (index == 0) {
|
|
|
|
sg_len = sg_dma_len(sg);
|
|
|
|
addr = (u64) sg_dma_address(sg);
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_wrb, sge0_addr_lo, pwrb,
|
2010-07-22 05:46:00 +07:00
|
|
|
((u32)(addr & 0xFFFFFFFF)));
|
2009-09-05 09:06:35 +07:00
|
|
|
AMAP_SET_BITS(struct amap_iscsi_wrb, sge0_addr_hi, pwrb,
|
2010-07-22 05:46:00 +07:00
|
|
|
((u32)(addr >> 32)));
|
2009-09-05 09:06:35 +07:00
|
|
|
AMAP_SET_BITS(struct amap_iscsi_wrb, sge0_len, pwrb,
|
|
|
|
sg_len);
|
|
|
|
sge_len = sg_len;
|
|
|
|
} else {
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_wrb, sge1_r2t_offset,
|
|
|
|
pwrb, sge_len);
|
|
|
|
sg_len = sg_dma_len(sg);
|
|
|
|
addr = (u64) sg_dma_address(sg);
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_wrb, sge1_addr_lo, pwrb,
|
2010-07-22 05:46:00 +07:00
|
|
|
((u32)(addr & 0xFFFFFFFF)));
|
2009-09-05 09:06:35 +07:00
|
|
|
AMAP_SET_BITS(struct amap_iscsi_wrb, sge1_addr_hi, pwrb,
|
2010-07-22 05:46:00 +07:00
|
|
|
((u32)(addr >> 32)));
|
2009-09-05 09:06:35 +07:00
|
|
|
AMAP_SET_BITS(struct amap_iscsi_wrb, sge1_len, pwrb,
|
|
|
|
sg_len);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
psgl = (struct iscsi_sge *)io_task->psgl_handle->pfrag;
|
|
|
|
memset(psgl, 0, sizeof(*psgl) * BE2_SGE);
|
|
|
|
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_sge, len, psgl, io_task->bhs_len - 2);
|
|
|
|
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_sge, addr_hi, psgl,
|
|
|
|
io_task->bhs_pa.u.a32.address_hi);
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_sge, addr_lo, psgl,
|
|
|
|
io_task->bhs_pa.u.a32.address_lo);
|
|
|
|
|
2010-01-23 07:08:18 +07:00
|
|
|
if (num_sg == 1) {
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_wrb, sge0_last, pwrb,
|
|
|
|
1);
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_wrb, sge1_last, pwrb,
|
|
|
|
0);
|
|
|
|
} else if (num_sg == 2) {
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_wrb, sge0_last, pwrb,
|
|
|
|
0);
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_wrb, sge1_last, pwrb,
|
|
|
|
1);
|
|
|
|
} else {
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_wrb, sge0_last, pwrb,
|
|
|
|
0);
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_wrb, sge1_last, pwrb,
|
|
|
|
0);
|
|
|
|
}
|
2009-09-05 09:06:35 +07:00
|
|
|
sg = l_sg;
|
|
|
|
psgl++;
|
|
|
|
psgl++;
|
|
|
|
offset = 0;
|
2010-01-07 03:20:19 +07:00
|
|
|
for (index = 0; index < num_sg; index++, sg = sg_next(sg), psgl++) {
|
2009-09-05 09:06:35 +07:00
|
|
|
sg_len = sg_dma_len(sg);
|
|
|
|
addr = (u64) sg_dma_address(sg);
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_sge, addr_lo, psgl,
|
|
|
|
(addr & 0xFFFFFFFF));
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_sge, addr_hi, psgl,
|
|
|
|
(addr >> 32));
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_sge, len, psgl, sg_len);
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_sge, sge_offset, psgl, offset);
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_sge, last_sge, psgl, 0);
|
|
|
|
offset += sg_len;
|
|
|
|
}
|
|
|
|
psgl--;
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_sge, last_sge, psgl, 1);
|
|
|
|
}
|
|
|
|
|
2012-10-20 06:12:00 +07:00
|
|
|
/**
|
|
|
|
* hwi_write_buffer()- Populate the WRB with task info
|
|
|
|
* @pwrb: ptr to the WRB entry
|
|
|
|
* @task: iscsi task which is to be executed
|
|
|
|
**/
|
2016-01-16 04:45:34 +07:00
|
|
|
static int hwi_write_buffer(struct iscsi_wrb *pwrb, struct iscsi_task *task)
|
2009-09-05 09:06:35 +07:00
|
|
|
{
|
|
|
|
struct iscsi_sge *psgl;
|
|
|
|
struct beiscsi_io_task *io_task = task->dd_data;
|
|
|
|
struct beiscsi_conn *beiscsi_conn = io_task->conn;
|
|
|
|
struct beiscsi_hba *phba = beiscsi_conn->phba;
|
2012-10-20 06:14:23 +07:00
|
|
|
uint8_t dsp_value = 0;
|
2009-09-05 09:06:35 +07:00
|
|
|
|
|
|
|
io_task->bhs_len = sizeof(struct be_nonio_bhs) - 2;
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_wrb, iscsi_bhs_addr_lo, pwrb,
|
|
|
|
io_task->bhs_pa.u.a32.address_lo);
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_wrb, iscsi_bhs_addr_hi, pwrb,
|
|
|
|
io_task->bhs_pa.u.a32.address_hi);
|
|
|
|
|
|
|
|
if (task->data) {
|
2012-10-20 06:14:23 +07:00
|
|
|
|
|
|
|
/* Check for the data_count */
|
|
|
|
dsp_value = (task->data_count) ? 1 : 0;
|
|
|
|
|
2013-04-06 10:38:26 +07:00
|
|
|
if (is_chip_be2_be3r(phba))
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_wrb, dsp,
|
2012-10-20 06:14:23 +07:00
|
|
|
pwrb, dsp_value);
|
|
|
|
else
|
2013-04-06 10:38:26 +07:00
|
|
|
AMAP_SET_BITS(struct amap_iscsi_wrb_v2, dsp,
|
2012-10-20 06:14:23 +07:00
|
|
|
pwrb, dsp_value);
|
|
|
|
|
|
|
|
/* Map addr only if there is data_count */
|
|
|
|
if (dsp_value) {
|
2012-10-20 06:12:00 +07:00
|
|
|
io_task->mtask_addr = pci_map_single(phba->pcidev,
|
|
|
|
task->data,
|
|
|
|
task->data_count,
|
|
|
|
PCI_DMA_TODEVICE);
|
2016-01-16 04:45:34 +07:00
|
|
|
if (pci_dma_mapping_error(phba->pcidev,
|
|
|
|
io_task->mtask_addr))
|
|
|
|
return -ENOMEM;
|
2012-10-20 06:12:00 +07:00
|
|
|
io_task->mtask_data_count = task->data_count;
|
2012-10-20 06:14:23 +07:00
|
|
|
} else
|
2012-10-20 06:12:00 +07:00
|
|
|
io_task->mtask_addr = 0;
|
2012-10-20 06:14:23 +07:00
|
|
|
|
2009-09-05 09:06:35 +07:00
|
|
|
AMAP_SET_BITS(struct amap_iscsi_wrb, sge0_addr_lo, pwrb,
|
2012-10-20 06:12:00 +07:00
|
|
|
lower_32_bits(io_task->mtask_addr));
|
2009-09-05 09:06:35 +07:00
|
|
|
AMAP_SET_BITS(struct amap_iscsi_wrb, sge0_addr_hi, pwrb,
|
2012-10-20 06:12:00 +07:00
|
|
|
upper_32_bits(io_task->mtask_addr));
|
2009-09-05 09:06:35 +07:00
|
|
|
AMAP_SET_BITS(struct amap_iscsi_wrb, sge0_len, pwrb,
|
|
|
|
task->data_count);
|
|
|
|
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_wrb, sge0_last, pwrb, 1);
|
|
|
|
} else {
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_wrb, dsp, pwrb, 0);
|
2012-10-20 06:12:00 +07:00
|
|
|
io_task->mtask_addr = 0;
|
2009-09-05 09:06:35 +07:00
|
|
|
}
|
|
|
|
|
|
|
|
psgl = (struct iscsi_sge *)io_task->psgl_handle->pfrag;
|
|
|
|
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_sge, len, psgl, io_task->bhs_len);
|
|
|
|
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_sge, addr_hi, psgl,
|
|
|
|
io_task->bhs_pa.u.a32.address_hi);
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_sge, addr_lo, psgl,
|
|
|
|
io_task->bhs_pa.u.a32.address_lo);
|
|
|
|
if (task->data) {
|
|
|
|
psgl++;
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_sge, addr_hi, psgl, 0);
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_sge, addr_lo, psgl, 0);
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_sge, len, psgl, 0);
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_sge, sge_offset, psgl, 0);
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_sge, rsvd0, psgl, 0);
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_sge, last_sge, psgl, 0);
|
|
|
|
|
|
|
|
psgl++;
|
|
|
|
if (task->data) {
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_sge, addr_lo, psgl,
|
2012-10-20 06:12:00 +07:00
|
|
|
lower_32_bits(io_task->mtask_addr));
|
2009-09-05 09:06:35 +07:00
|
|
|
AMAP_SET_BITS(struct amap_iscsi_sge, addr_hi, psgl,
|
2012-10-20 06:12:00 +07:00
|
|
|
upper_32_bits(io_task->mtask_addr));
|
2009-09-05 09:06:35 +07:00
|
|
|
}
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_sge, len, psgl, 0x106);
|
|
|
|
}
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_sge, last_sge, psgl, 1);
|
2016-01-16 04:45:34 +07:00
|
|
|
return 0;
|
2009-09-05 09:06:35 +07:00
|
|
|
}
|
|
|
|
|
2013-09-29 05:35:44 +07:00
|
|
|
/**
|
|
|
|
* beiscsi_find_mem_req()- Find mem needed
|
|
|
|
* @phba: ptr to HBA struct
|
|
|
|
**/
|
2009-09-05 09:06:35 +07:00
|
|
|
static void beiscsi_find_mem_req(struct beiscsi_hba *phba)
|
|
|
|
{
|
2013-09-29 05:35:45 +07:00
|
|
|
uint8_t mem_descr_index, ulp_num;
|
2009-10-23 13:22:33 +07:00
|
|
|
unsigned int num_cq_pages, num_async_pdu_buf_pages;
|
2009-09-05 09:06:35 +07:00
|
|
|
unsigned int num_async_pdu_data_pages, wrb_sz_per_cxn;
|
|
|
|
unsigned int num_async_pdu_buf_sgl_pages, num_async_pdu_data_sgl_pages;
|
|
|
|
|
|
|
|
num_cq_pages = PAGES_REQUIRED(phba->params.num_cq_entries * \
|
|
|
|
sizeof(struct sol_cqe));
|
|
|
|
|
|
|
|
phba->params.hwi_ws_sz = sizeof(struct hwi_controller);
|
|
|
|
|
|
|
|
phba->mem_req[ISCSI_MEM_GLOBAL_HEADER] = 2 *
|
|
|
|
BE_ISCSI_PDU_HEADER_SIZE;
|
|
|
|
phba->mem_req[HWI_MEM_ADDN_CONTEXT] =
|
|
|
|
sizeof(struct hwi_context_memory);
|
|
|
|
|
|
|
|
|
|
|
|
phba->mem_req[HWI_MEM_WRB] = sizeof(struct iscsi_wrb)
|
|
|
|
* (phba->params.wrbs_per_cxn)
|
|
|
|
* phba->params.cxns_per_ctrl;
|
|
|
|
wrb_sz_per_cxn = sizeof(struct wrb_handle) *
|
|
|
|
(phba->params.wrbs_per_cxn);
|
|
|
|
phba->mem_req[HWI_MEM_WRBH] = roundup_pow_of_two((wrb_sz_per_cxn) *
|
|
|
|
phba->params.cxns_per_ctrl);
|
|
|
|
|
|
|
|
phba->mem_req[HWI_MEM_SGLH] = sizeof(struct sgl_handle) *
|
|
|
|
phba->params.icds_per_ctrl;
|
|
|
|
phba->mem_req[HWI_MEM_SGE] = sizeof(struct iscsi_sge) *
|
|
|
|
phba->params.num_sge_per_io * phba->params.icds_per_ctrl;
|
2013-09-29 05:35:45 +07:00
|
|
|
for (ulp_num = 0; ulp_num < BEISCSI_ULP_COUNT; ulp_num++) {
|
|
|
|
if (test_bit(ulp_num, &phba->fw_config.ulp_supported)) {
|
2009-09-05 09:06:35 +07:00
|
|
|
|
2013-09-29 05:35:45 +07:00
|
|
|
num_async_pdu_buf_sgl_pages =
|
|
|
|
PAGES_REQUIRED(BEISCSI_GET_CID_COUNT(
|
|
|
|
phba, ulp_num) *
|
|
|
|
sizeof(struct phys_addr));
|
|
|
|
|
|
|
|
num_async_pdu_buf_pages =
|
|
|
|
PAGES_REQUIRED(BEISCSI_GET_CID_COUNT(
|
|
|
|
phba, ulp_num) *
|
|
|
|
phba->params.defpdu_hdr_sz);
|
|
|
|
|
|
|
|
num_async_pdu_data_pages =
|
|
|
|
PAGES_REQUIRED(BEISCSI_GET_CID_COUNT(
|
|
|
|
phba, ulp_num) *
|
|
|
|
phba->params.defpdu_data_sz);
|
|
|
|
|
|
|
|
num_async_pdu_data_sgl_pages =
|
|
|
|
PAGES_REQUIRED(BEISCSI_GET_CID_COUNT(
|
|
|
|
phba, ulp_num) *
|
|
|
|
sizeof(struct phys_addr));
|
|
|
|
|
2013-09-29 05:35:46 +07:00
|
|
|
mem_descr_index = (HWI_MEM_TEMPLATE_HDR_ULP0 +
|
|
|
|
(ulp_num * MEM_DESCR_OFFSET));
|
|
|
|
phba->mem_req[mem_descr_index] =
|
|
|
|
BEISCSI_GET_CID_COUNT(phba, ulp_num) *
|
|
|
|
BEISCSI_TEMPLATE_HDR_PER_CXN_SIZE;
|
|
|
|
|
2013-09-29 05:35:45 +07:00
|
|
|
mem_descr_index = (HWI_MEM_ASYNC_HEADER_BUF_ULP0 +
|
|
|
|
(ulp_num * MEM_DESCR_OFFSET));
|
|
|
|
phba->mem_req[mem_descr_index] =
|
|
|
|
num_async_pdu_buf_pages *
|
|
|
|
PAGE_SIZE;
|
|
|
|
|
|
|
|
mem_descr_index = (HWI_MEM_ASYNC_DATA_BUF_ULP0 +
|
|
|
|
(ulp_num * MEM_DESCR_OFFSET));
|
|
|
|
phba->mem_req[mem_descr_index] =
|
|
|
|
num_async_pdu_data_pages *
|
|
|
|
PAGE_SIZE;
|
|
|
|
|
|
|
|
mem_descr_index = (HWI_MEM_ASYNC_HEADER_RING_ULP0 +
|
|
|
|
(ulp_num * MEM_DESCR_OFFSET));
|
|
|
|
phba->mem_req[mem_descr_index] =
|
|
|
|
num_async_pdu_buf_sgl_pages *
|
|
|
|
PAGE_SIZE;
|
|
|
|
|
|
|
|
mem_descr_index = (HWI_MEM_ASYNC_DATA_RING_ULP0 +
|
|
|
|
(ulp_num * MEM_DESCR_OFFSET));
|
|
|
|
phba->mem_req[mem_descr_index] =
|
|
|
|
num_async_pdu_data_sgl_pages *
|
|
|
|
PAGE_SIZE;
|
|
|
|
|
|
|
|
mem_descr_index = (HWI_MEM_ASYNC_HEADER_HANDLE_ULP0 +
|
|
|
|
(ulp_num * MEM_DESCR_OFFSET));
|
|
|
|
phba->mem_req[mem_descr_index] =
|
|
|
|
BEISCSI_GET_CID_COUNT(phba, ulp_num) *
|
|
|
|
sizeof(struct async_pdu_handle);
|
|
|
|
|
|
|
|
mem_descr_index = (HWI_MEM_ASYNC_DATA_HANDLE_ULP0 +
|
|
|
|
(ulp_num * MEM_DESCR_OFFSET));
|
|
|
|
phba->mem_req[mem_descr_index] =
|
|
|
|
BEISCSI_GET_CID_COUNT(phba, ulp_num) *
|
|
|
|
sizeof(struct async_pdu_handle);
|
|
|
|
|
|
|
|
mem_descr_index = (HWI_MEM_ASYNC_PDU_CONTEXT_ULP0 +
|
|
|
|
(ulp_num * MEM_DESCR_OFFSET));
|
|
|
|
phba->mem_req[mem_descr_index] =
|
|
|
|
sizeof(struct hwi_async_pdu_context) +
|
|
|
|
(BEISCSI_GET_CID_COUNT(phba, ulp_num) *
|
|
|
|
sizeof(struct hwi_async_entry));
|
|
|
|
}
|
|
|
|
}
|
2009-09-05 09:06:35 +07:00
|
|
|
}
|
|
|
|
|
|
|
|
static int beiscsi_alloc_mem(struct beiscsi_hba *phba)
|
|
|
|
{
|
|
|
|
dma_addr_t bus_add;
|
2013-04-06 10:38:32 +07:00
|
|
|
struct hwi_controller *phwi_ctrlr;
|
|
|
|
struct be_mem_descriptor *mem_descr;
|
2009-09-05 09:06:35 +07:00
|
|
|
struct mem_array *mem_arr, *mem_arr_orig;
|
|
|
|
unsigned int i, j, alloc_size, curr_alloc_size;
|
|
|
|
|
2012-04-04 11:41:38 +07:00
|
|
|
phba->phwi_ctrlr = kzalloc(phba->params.hwi_ws_sz, GFP_KERNEL);
|
2009-09-05 09:06:35 +07:00
|
|
|
if (!phba->phwi_ctrlr)
|
|
|
|
return -ENOMEM;
|
|
|
|
|
2013-04-06 10:38:32 +07:00
|
|
|
/* Allocate memory for wrb_context */
|
|
|
|
phwi_ctrlr = phba->phwi_ctrlr;
|
|
|
|
phwi_ctrlr->wrb_context = kzalloc(sizeof(struct hwi_wrb_context) *
|
|
|
|
phba->params.cxns_per_ctrl,
|
|
|
|
GFP_KERNEL);
|
2015-10-01 15:56:25 +07:00
|
|
|
if (!phwi_ctrlr->wrb_context) {
|
|
|
|
kfree(phba->phwi_ctrlr);
|
2013-04-06 10:38:32 +07:00
|
|
|
return -ENOMEM;
|
2015-10-01 15:56:25 +07:00
|
|
|
}
|
2013-04-06 10:38:32 +07:00
|
|
|
|
2009-09-05 09:06:35 +07:00
|
|
|
phba->init_mem = kcalloc(SE_MEM_MAX, sizeof(*mem_descr),
|
|
|
|
GFP_KERNEL);
|
|
|
|
if (!phba->init_mem) {
|
2013-04-06 10:38:32 +07:00
|
|
|
kfree(phwi_ctrlr->wrb_context);
|
2009-09-05 09:06:35 +07:00
|
|
|
kfree(phba->phwi_ctrlr);
|
|
|
|
return -ENOMEM;
|
|
|
|
}
|
|
|
|
|
|
|
|
mem_arr_orig = kmalloc(sizeof(*mem_arr_orig) * BEISCSI_MAX_FRAGS_INIT,
|
|
|
|
GFP_KERNEL);
|
|
|
|
if (!mem_arr_orig) {
|
|
|
|
kfree(phba->init_mem);
|
2013-04-06 10:38:32 +07:00
|
|
|
kfree(phwi_ctrlr->wrb_context);
|
2009-09-05 09:06:35 +07:00
|
|
|
kfree(phba->phwi_ctrlr);
|
|
|
|
return -ENOMEM;
|
|
|
|
}
|
|
|
|
|
|
|
|
mem_descr = phba->init_mem;
|
|
|
|
for (i = 0; i < SE_MEM_MAX; i++) {
|
2013-09-29 05:35:45 +07:00
|
|
|
if (!phba->mem_req[i]) {
|
|
|
|
mem_descr->mem_array = NULL;
|
|
|
|
mem_descr++;
|
|
|
|
continue;
|
|
|
|
}
|
|
|
|
|
2009-09-05 09:06:35 +07:00
|
|
|
j = 0;
|
|
|
|
mem_arr = mem_arr_orig;
|
|
|
|
alloc_size = phba->mem_req[i];
|
|
|
|
memset(mem_arr, 0, sizeof(struct mem_array) *
|
|
|
|
BEISCSI_MAX_FRAGS_INIT);
|
|
|
|
curr_alloc_size = min(be_max_phys_size * 1024, alloc_size);
|
|
|
|
do {
|
|
|
|
mem_arr->virtual_address = pci_alloc_consistent(
|
|
|
|
phba->pcidev,
|
|
|
|
curr_alloc_size,
|
|
|
|
&bus_add);
|
|
|
|
if (!mem_arr->virtual_address) {
|
|
|
|
if (curr_alloc_size <= BE_MIN_MEM_SIZE)
|
|
|
|
goto free_mem;
|
|
|
|
if (curr_alloc_size -
|
|
|
|
rounddown_pow_of_two(curr_alloc_size))
|
|
|
|
curr_alloc_size = rounddown_pow_of_two
|
|
|
|
(curr_alloc_size);
|
|
|
|
else
|
|
|
|
curr_alloc_size = curr_alloc_size / 2;
|
|
|
|
} else {
|
|
|
|
mem_arr->bus_address.u.
|
|
|
|
a64.address = (__u64) bus_add;
|
|
|
|
mem_arr->size = curr_alloc_size;
|
|
|
|
alloc_size -= curr_alloc_size;
|
|
|
|
curr_alloc_size = min(be_max_phys_size *
|
|
|
|
1024, alloc_size);
|
|
|
|
j++;
|
|
|
|
mem_arr++;
|
|
|
|
}
|
|
|
|
} while (alloc_size);
|
|
|
|
mem_descr->num_elements = j;
|
|
|
|
mem_descr->size_in_bytes = phba->mem_req[i];
|
|
|
|
mem_descr->mem_array = kmalloc(sizeof(*mem_arr) * j,
|
|
|
|
GFP_KERNEL);
|
|
|
|
if (!mem_descr->mem_array)
|
|
|
|
goto free_mem;
|
|
|
|
|
|
|
|
memcpy(mem_descr->mem_array, mem_arr_orig,
|
|
|
|
sizeof(struct mem_array) * j);
|
|
|
|
mem_descr++;
|
|
|
|
}
|
|
|
|
kfree(mem_arr_orig);
|
|
|
|
return 0;
|
|
|
|
free_mem:
|
|
|
|
mem_descr->num_elements = j;
|
|
|
|
while ((i) || (j)) {
|
|
|
|
for (j = mem_descr->num_elements; j > 0; j--) {
|
|
|
|
pci_free_consistent(phba->pcidev,
|
|
|
|
mem_descr->mem_array[j - 1].size,
|
|
|
|
mem_descr->mem_array[j - 1].
|
|
|
|
virtual_address,
|
2010-07-22 05:46:00 +07:00
|
|
|
(unsigned long)mem_descr->
|
|
|
|
mem_array[j - 1].
|
2009-09-05 09:06:35 +07:00
|
|
|
bus_address.u.a64.address);
|
|
|
|
}
|
|
|
|
if (i) {
|
|
|
|
i--;
|
|
|
|
kfree(mem_descr->mem_array);
|
|
|
|
mem_descr--;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
kfree(mem_arr_orig);
|
|
|
|
kfree(phba->init_mem);
|
2013-04-06 10:38:32 +07:00
|
|
|
kfree(phba->phwi_ctrlr->wrb_context);
|
2009-09-05 09:06:35 +07:00
|
|
|
kfree(phba->phwi_ctrlr);
|
|
|
|
return -ENOMEM;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int beiscsi_get_memory(struct beiscsi_hba *phba)
|
|
|
|
{
|
|
|
|
beiscsi_find_mem_req(phba);
|
|
|
|
return beiscsi_alloc_mem(phba);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void iscsi_init_global_templates(struct beiscsi_hba *phba)
|
|
|
|
{
|
|
|
|
struct pdu_data_out *pdata_out;
|
|
|
|
struct pdu_nop_out *pnop_out;
|
|
|
|
struct be_mem_descriptor *mem_descr;
|
|
|
|
|
|
|
|
mem_descr = phba->init_mem;
|
|
|
|
mem_descr += ISCSI_MEM_GLOBAL_HEADER;
|
|
|
|
pdata_out =
|
|
|
|
(struct pdu_data_out *)mem_descr->mem_array[0].virtual_address;
|
|
|
|
memset(pdata_out, 0, BE_ISCSI_PDU_HEADER_SIZE);
|
|
|
|
|
|
|
|
AMAP_SET_BITS(struct amap_pdu_data_out, opcode, pdata_out,
|
|
|
|
IIOC_SCSI_DATA);
|
|
|
|
|
|
|
|
pnop_out =
|
|
|
|
(struct pdu_nop_out *)((unsigned char *)mem_descr->mem_array[0].
|
|
|
|
virtual_address + BE_ISCSI_PDU_HEADER_SIZE);
|
|
|
|
|
|
|
|
memset(pnop_out, 0, BE_ISCSI_PDU_HEADER_SIZE);
|
|
|
|
AMAP_SET_BITS(struct amap_pdu_nop_out, ttt, pnop_out, 0xFFFFFFFF);
|
|
|
|
AMAP_SET_BITS(struct amap_pdu_nop_out, f_bit, pnop_out, 1);
|
|
|
|
AMAP_SET_BITS(struct amap_pdu_nop_out, i_bit, pnop_out, 0);
|
|
|
|
}
|
|
|
|
|
2012-04-04 11:41:38 +07:00
|
|
|
static int beiscsi_init_wrb_handle(struct beiscsi_hba *phba)
|
2009-09-05 09:06:35 +07:00
|
|
|
{
|
|
|
|
struct be_mem_descriptor *mem_descr_wrbh, *mem_descr_wrb;
|
2013-04-06 10:38:32 +07:00
|
|
|
struct hwi_context_memory *phwi_ctxt;
|
2012-04-04 11:41:38 +07:00
|
|
|
struct wrb_handle *pwrb_handle = NULL;
|
2009-09-05 09:06:35 +07:00
|
|
|
struct hwi_controller *phwi_ctrlr;
|
|
|
|
struct hwi_wrb_context *pwrb_context;
|
2012-04-04 11:41:38 +07:00
|
|
|
struct iscsi_wrb *pwrb = NULL;
|
|
|
|
unsigned int num_cxn_wrbh = 0;
|
|
|
|
unsigned int num_cxn_wrb = 0, j, idx = 0, index;
|
2009-09-05 09:06:35 +07:00
|
|
|
|
|
|
|
mem_descr_wrbh = phba->init_mem;
|
|
|
|
mem_descr_wrbh += HWI_MEM_WRBH;
|
|
|
|
|
|
|
|
mem_descr_wrb = phba->init_mem;
|
|
|
|
mem_descr_wrb += HWI_MEM_WRB;
|
|
|
|
phwi_ctrlr = phba->phwi_ctrlr;
|
|
|
|
|
2013-04-06 10:38:32 +07:00
|
|
|
/* Allocate memory for WRBQ */
|
|
|
|
phwi_ctxt = phwi_ctrlr->phwi_ctxt;
|
|
|
|
phwi_ctxt->be_wrbq = kzalloc(sizeof(struct be_queue_info) *
|
2013-09-29 05:35:44 +07:00
|
|
|
phba->params.cxns_per_ctrl,
|
2013-04-06 10:38:32 +07:00
|
|
|
GFP_KERNEL);
|
|
|
|
if (!phwi_ctxt->be_wrbq) {
|
|
|
|
beiscsi_log(phba, KERN_ERR, BEISCSI_LOG_INIT,
|
|
|
|
"BM_%d : WRBQ Mem Alloc Failed\n");
|
|
|
|
return -ENOMEM;
|
|
|
|
}
|
|
|
|
|
|
|
|
for (index = 0; index < phba->params.cxns_per_ctrl; index++) {
|
2009-09-05 09:06:35 +07:00
|
|
|
pwrb_context = &phwi_ctrlr->wrb_context[index];
|
|
|
|
pwrb_context->pwrb_handle_base =
|
|
|
|
kzalloc(sizeof(struct wrb_handle *) *
|
|
|
|
phba->params.wrbs_per_cxn, GFP_KERNEL);
|
2012-04-04 11:41:38 +07:00
|
|
|
if (!pwrb_context->pwrb_handle_base) {
|
2012-08-21 00:30:18 +07:00
|
|
|
beiscsi_log(phba, KERN_ERR, BEISCSI_LOG_INIT,
|
|
|
|
"BM_%d : Mem Alloc Failed. Failing to load\n");
|
2012-04-04 11:41:38 +07:00
|
|
|
goto init_wrb_hndl_failed;
|
|
|
|
}
|
2009-09-05 09:06:35 +07:00
|
|
|
pwrb_context->pwrb_handle_basestd =
|
|
|
|
kzalloc(sizeof(struct wrb_handle *) *
|
|
|
|
phba->params.wrbs_per_cxn, GFP_KERNEL);
|
2012-04-04 11:41:38 +07:00
|
|
|
if (!pwrb_context->pwrb_handle_basestd) {
|
2012-08-21 00:30:18 +07:00
|
|
|
beiscsi_log(phba, KERN_ERR, BEISCSI_LOG_INIT,
|
|
|
|
"BM_%d : Mem Alloc Failed. Failing to load\n");
|
2012-04-04 11:41:38 +07:00
|
|
|
goto init_wrb_hndl_failed;
|
|
|
|
}
|
|
|
|
if (!num_cxn_wrbh) {
|
|
|
|
pwrb_handle =
|
|
|
|
mem_descr_wrbh->mem_array[idx].virtual_address;
|
|
|
|
num_cxn_wrbh = ((mem_descr_wrbh->mem_array[idx].size) /
|
|
|
|
((sizeof(struct wrb_handle)) *
|
|
|
|
phba->params.wrbs_per_cxn));
|
|
|
|
idx++;
|
|
|
|
}
|
|
|
|
pwrb_context->alloc_index = 0;
|
|
|
|
pwrb_context->wrb_handles_available = 0;
|
|
|
|
pwrb_context->free_index = 0;
|
|
|
|
|
2009-09-05 09:06:35 +07:00
|
|
|
if (num_cxn_wrbh) {
|
|
|
|
for (j = 0; j < phba->params.wrbs_per_cxn; j++) {
|
|
|
|
pwrb_context->pwrb_handle_base[j] = pwrb_handle;
|
|
|
|
pwrb_context->pwrb_handle_basestd[j] =
|
|
|
|
pwrb_handle;
|
|
|
|
pwrb_context->wrb_handles_available++;
|
2009-10-23 13:22:33 +07:00
|
|
|
pwrb_handle->wrb_index = j;
|
2009-09-05 09:06:35 +07:00
|
|
|
pwrb_handle++;
|
|
|
|
}
|
|
|
|
num_cxn_wrbh--;
|
|
|
|
}
|
2016-02-04 17:19:20 +07:00
|
|
|
spin_lock_init(&pwrb_context->wrb_lock);
|
2009-09-05 09:06:35 +07:00
|
|
|
}
|
|
|
|
idx = 0;
|
2013-04-06 10:38:32 +07:00
|
|
|
for (index = 0; index < phba->params.cxns_per_ctrl; index++) {
|
2009-09-05 09:06:35 +07:00
|
|
|
pwrb_context = &phwi_ctrlr->wrb_context[index];
|
2012-04-04 11:41:38 +07:00
|
|
|
if (!num_cxn_wrb) {
|
2009-09-05 09:06:35 +07:00
|
|
|
pwrb = mem_descr_wrb->mem_array[idx].virtual_address;
|
2010-01-23 07:09:37 +07:00
|
|
|
num_cxn_wrb = (mem_descr_wrb->mem_array[idx].size) /
|
2012-04-04 11:41:38 +07:00
|
|
|
((sizeof(struct iscsi_wrb) *
|
|
|
|
phba->params.wrbs_per_cxn));
|
|
|
|
idx++;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (num_cxn_wrb) {
|
2009-09-05 09:06:35 +07:00
|
|
|
for (j = 0; j < phba->params.wrbs_per_cxn; j++) {
|
|
|
|
pwrb_handle = pwrb_context->pwrb_handle_base[j];
|
|
|
|
pwrb_handle->pwrb = pwrb;
|
|
|
|
pwrb++;
|
|
|
|
}
|
|
|
|
num_cxn_wrb--;
|
|
|
|
}
|
|
|
|
}
|
2012-04-04 11:41:38 +07:00
|
|
|
return 0;
|
|
|
|
init_wrb_hndl_failed:
|
|
|
|
for (j = index; j > 0; j--) {
|
|
|
|
pwrb_context = &phwi_ctrlr->wrb_context[j];
|
|
|
|
kfree(pwrb_context->pwrb_handle_base);
|
|
|
|
kfree(pwrb_context->pwrb_handle_basestd);
|
|
|
|
}
|
|
|
|
return -ENOMEM;
|
2009-09-05 09:06:35 +07:00
|
|
|
}
|
|
|
|
|
2013-04-06 10:38:32 +07:00
|
|
|
static int hwi_init_async_pdu_ctx(struct beiscsi_hba *phba)
|
2009-09-05 09:06:35 +07:00
|
|
|
{
|
2013-09-29 05:35:45 +07:00
|
|
|
uint8_t ulp_num;
|
2009-09-05 09:06:35 +07:00
|
|
|
struct hwi_controller *phwi_ctrlr;
|
|
|
|
struct hba_parameters *p = &phba->params;
|
|
|
|
struct hwi_async_pdu_context *pasync_ctx;
|
|
|
|
struct async_pdu_handle *pasync_header_h, *pasync_data_h;
|
2012-04-04 11:41:36 +07:00
|
|
|
unsigned int index, idx, num_per_mem, num_async_data;
|
2009-09-05 09:06:35 +07:00
|
|
|
struct be_mem_descriptor *mem_descr;
|
|
|
|
|
2013-09-29 05:35:45 +07:00
|
|
|
for (ulp_num = 0; ulp_num < BEISCSI_ULP_COUNT; ulp_num++) {
|
|
|
|
if (test_bit(ulp_num, &phba->fw_config.ulp_supported)) {
|
2009-09-05 09:06:35 +07:00
|
|
|
|
2013-09-29 05:35:45 +07:00
|
|
|
mem_descr = (struct be_mem_descriptor *)phba->init_mem;
|
|
|
|
mem_descr += (HWI_MEM_ASYNC_PDU_CONTEXT_ULP0 +
|
|
|
|
(ulp_num * MEM_DESCR_OFFSET));
|
|
|
|
|
|
|
|
phwi_ctrlr = phba->phwi_ctrlr;
|
|
|
|
phwi_ctrlr->phwi_ctxt->pasync_ctx[ulp_num] =
|
|
|
|
(struct hwi_async_pdu_context *)
|
|
|
|
mem_descr->mem_array[0].virtual_address;
|
|
|
|
|
|
|
|
pasync_ctx = phwi_ctrlr->phwi_ctxt->pasync_ctx[ulp_num];
|
|
|
|
memset(pasync_ctx, 0, sizeof(*pasync_ctx));
|
|
|
|
|
|
|
|
pasync_ctx->async_entry =
|
|
|
|
(struct hwi_async_entry *)
|
|
|
|
((long unsigned int)pasync_ctx +
|
|
|
|
sizeof(struct hwi_async_pdu_context));
|
|
|
|
|
|
|
|
pasync_ctx->num_entries = BEISCSI_GET_CID_COUNT(phba,
|
|
|
|
ulp_num);
|
|
|
|
pasync_ctx->buffer_size = p->defpdu_hdr_sz;
|
|
|
|
|
|
|
|
mem_descr = (struct be_mem_descriptor *)phba->init_mem;
|
|
|
|
mem_descr += HWI_MEM_ASYNC_HEADER_BUF_ULP0 +
|
|
|
|
(ulp_num * MEM_DESCR_OFFSET);
|
|
|
|
if (mem_descr->mem_array[0].virtual_address) {
|
|
|
|
beiscsi_log(phba, KERN_INFO, BEISCSI_LOG_INIT,
|
|
|
|
"BM_%d : hwi_init_async_pdu_ctx"
|
|
|
|
" HWI_MEM_ASYNC_HEADER_BUF_ULP%d va=%p\n",
|
|
|
|
ulp_num,
|
|
|
|
mem_descr->mem_array[0].
|
|
|
|
virtual_address);
|
|
|
|
} else
|
|
|
|
beiscsi_log(phba, KERN_WARNING,
|
|
|
|
BEISCSI_LOG_INIT,
|
|
|
|
"BM_%d : No Virtual address for ULP : %d\n",
|
|
|
|
ulp_num);
|
|
|
|
|
|
|
|
pasync_ctx->async_header.va_base =
|
2009-09-05 09:06:35 +07:00
|
|
|
mem_descr->mem_array[0].virtual_address;
|
|
|
|
|
2013-09-29 05:35:45 +07:00
|
|
|
pasync_ctx->async_header.pa_base.u.a64.address =
|
|
|
|
mem_descr->mem_array[0].
|
|
|
|
bus_address.u.a64.address;
|
2009-09-05 09:06:35 +07:00
|
|
|
|
2013-09-29 05:35:45 +07:00
|
|
|
mem_descr = (struct be_mem_descriptor *)phba->init_mem;
|
|
|
|
mem_descr += HWI_MEM_ASYNC_HEADER_RING_ULP0 +
|
|
|
|
(ulp_num * MEM_DESCR_OFFSET);
|
|
|
|
if (mem_descr->mem_array[0].virtual_address) {
|
|
|
|
beiscsi_log(phba, KERN_INFO, BEISCSI_LOG_INIT,
|
|
|
|
"BM_%d : hwi_init_async_pdu_ctx"
|
|
|
|
" HWI_MEM_ASYNC_HEADER_RING_ULP%d va=%p\n",
|
|
|
|
ulp_num,
|
|
|
|
mem_descr->mem_array[0].
|
|
|
|
virtual_address);
|
|
|
|
} else
|
|
|
|
beiscsi_log(phba, KERN_WARNING,
|
|
|
|
BEISCSI_LOG_INIT,
|
|
|
|
"BM_%d : No Virtual address for ULP : %d\n",
|
|
|
|
ulp_num);
|
|
|
|
|
|
|
|
pasync_ctx->async_header.ring_base =
|
|
|
|
mem_descr->mem_array[0].virtual_address;
|
2009-09-05 09:06:35 +07:00
|
|
|
|
2013-09-29 05:35:45 +07:00
|
|
|
mem_descr = (struct be_mem_descriptor *)phba->init_mem;
|
|
|
|
mem_descr += HWI_MEM_ASYNC_HEADER_HANDLE_ULP0 +
|
|
|
|
(ulp_num * MEM_DESCR_OFFSET);
|
|
|
|
if (mem_descr->mem_array[0].virtual_address) {
|
|
|
|
beiscsi_log(phba, KERN_INFO, BEISCSI_LOG_INIT,
|
|
|
|
"BM_%d : hwi_init_async_pdu_ctx"
|
|
|
|
" HWI_MEM_ASYNC_HEADER_HANDLE_ULP%d va=%p\n",
|
|
|
|
ulp_num,
|
|
|
|
mem_descr->mem_array[0].
|
|
|
|
virtual_address);
|
|
|
|
} else
|
|
|
|
beiscsi_log(phba, KERN_WARNING,
|
|
|
|
BEISCSI_LOG_INIT,
|
|
|
|
"BM_%d : No Virtual address for ULP : %d\n",
|
|
|
|
ulp_num);
|
|
|
|
|
|
|
|
pasync_ctx->async_header.handle_base =
|
|
|
|
mem_descr->mem_array[0].virtual_address;
|
|
|
|
pasync_ctx->async_header.writables = 0;
|
|
|
|
INIT_LIST_HEAD(&pasync_ctx->async_header.free_list);
|
|
|
|
|
|
|
|
mem_descr = (struct be_mem_descriptor *)phba->init_mem;
|
|
|
|
mem_descr += HWI_MEM_ASYNC_DATA_RING_ULP0 +
|
|
|
|
(ulp_num * MEM_DESCR_OFFSET);
|
|
|
|
if (mem_descr->mem_array[0].virtual_address) {
|
|
|
|
beiscsi_log(phba, KERN_INFO, BEISCSI_LOG_INIT,
|
|
|
|
"BM_%d : hwi_init_async_pdu_ctx"
|
|
|
|
" HWI_MEM_ASYNC_DATA_RING_ULP%d va=%p\n",
|
|
|
|
ulp_num,
|
|
|
|
mem_descr->mem_array[0].
|
|
|
|
virtual_address);
|
|
|
|
} else
|
|
|
|
beiscsi_log(phba, KERN_WARNING,
|
|
|
|
BEISCSI_LOG_INIT,
|
|
|
|
"BM_%d : No Virtual address for ULP : %d\n",
|
|
|
|
ulp_num);
|
|
|
|
|
|
|
|
pasync_ctx->async_data.ring_base =
|
|
|
|
mem_descr->mem_array[0].virtual_address;
|
2009-09-05 09:06:35 +07:00
|
|
|
|
2013-09-29 05:35:45 +07:00
|
|
|
mem_descr = (struct be_mem_descriptor *)phba->init_mem;
|
|
|
|
mem_descr += HWI_MEM_ASYNC_DATA_HANDLE_ULP0 +
|
|
|
|
(ulp_num * MEM_DESCR_OFFSET);
|
|
|
|
if (!mem_descr->mem_array[0].virtual_address)
|
|
|
|
beiscsi_log(phba, KERN_WARNING,
|
|
|
|
BEISCSI_LOG_INIT,
|
|
|
|
"BM_%d : No Virtual address for ULP : %d\n",
|
|
|
|
ulp_num);
|
2012-08-21 00:30:18 +07:00
|
|
|
|
2013-09-29 05:35:45 +07:00
|
|
|
pasync_ctx->async_data.handle_base =
|
|
|
|
mem_descr->mem_array[0].virtual_address;
|
|
|
|
pasync_ctx->async_data.writables = 0;
|
|
|
|
INIT_LIST_HEAD(&pasync_ctx->async_data.free_list);
|
|
|
|
|
|
|
|
pasync_header_h =
|
|
|
|
(struct async_pdu_handle *)
|
|
|
|
pasync_ctx->async_header.handle_base;
|
|
|
|
pasync_data_h =
|
|
|
|
(struct async_pdu_handle *)
|
|
|
|
pasync_ctx->async_data.handle_base;
|
|
|
|
|
|
|
|
mem_descr = (struct be_mem_descriptor *)phba->init_mem;
|
|
|
|
mem_descr += HWI_MEM_ASYNC_DATA_BUF_ULP0 +
|
|
|
|
(ulp_num * MEM_DESCR_OFFSET);
|
|
|
|
if (mem_descr->mem_array[0].virtual_address) {
|
|
|
|
beiscsi_log(phba, KERN_INFO, BEISCSI_LOG_INIT,
|
|
|
|
"BM_%d : hwi_init_async_pdu_ctx"
|
|
|
|
" HWI_MEM_ASYNC_DATA_BUF_ULP%d va=%p\n",
|
|
|
|
ulp_num,
|
|
|
|
mem_descr->mem_array[0].
|
|
|
|
virtual_address);
|
|
|
|
} else
|
|
|
|
beiscsi_log(phba, KERN_WARNING,
|
|
|
|
BEISCSI_LOG_INIT,
|
|
|
|
"BM_%d : No Virtual address for ULP : %d\n",
|
|
|
|
ulp_num);
|
|
|
|
|
|
|
|
idx = 0;
|
2012-04-04 11:41:36 +07:00
|
|
|
pasync_ctx->async_data.va_base =
|
|
|
|
mem_descr->mem_array[idx].virtual_address;
|
|
|
|
pasync_ctx->async_data.pa_base.u.a64.address =
|
|
|
|
mem_descr->mem_array[idx].
|
|
|
|
bus_address.u.a64.address;
|
|
|
|
|
|
|
|
num_async_data = ((mem_descr->mem_array[idx].size) /
|
|
|
|
phba->params.defpdu_data_sz);
|
2013-09-29 05:35:45 +07:00
|
|
|
num_per_mem = 0;
|
2009-09-05 09:06:35 +07:00
|
|
|
|
2013-09-29 05:35:45 +07:00
|
|
|
for (index = 0; index < BEISCSI_GET_CID_COUNT
|
|
|
|
(phba, ulp_num); index++) {
|
|
|
|
pasync_header_h->cri = -1;
|
|
|
|
pasync_header_h->index = (char)index;
|
|
|
|
INIT_LIST_HEAD(&pasync_header_h->link);
|
|
|
|
pasync_header_h->pbuffer =
|
|
|
|
(void *)((unsigned long)
|
|
|
|
(pasync_ctx->
|
|
|
|
async_header.va_base) +
|
|
|
|
(p->defpdu_hdr_sz * index));
|
|
|
|
|
|
|
|
pasync_header_h->pa.u.a64.address =
|
|
|
|
pasync_ctx->async_header.pa_base.u.a64.
|
|
|
|
address + (p->defpdu_hdr_sz * index);
|
|
|
|
|
|
|
|
list_add_tail(&pasync_header_h->link,
|
|
|
|
&pasync_ctx->async_header.
|
|
|
|
free_list);
|
|
|
|
pasync_header_h++;
|
|
|
|
pasync_ctx->async_header.free_entries++;
|
|
|
|
pasync_ctx->async_header.writables++;
|
|
|
|
|
|
|
|
INIT_LIST_HEAD(&pasync_ctx->async_entry[index].
|
|
|
|
wait_queue.list);
|
|
|
|
INIT_LIST_HEAD(&pasync_ctx->async_entry[index].
|
|
|
|
header_busy_list);
|
|
|
|
pasync_data_h->cri = -1;
|
|
|
|
pasync_data_h->index = (char)index;
|
|
|
|
INIT_LIST_HEAD(&pasync_data_h->link);
|
|
|
|
|
|
|
|
if (!num_async_data) {
|
|
|
|
num_per_mem = 0;
|
|
|
|
idx++;
|
|
|
|
pasync_ctx->async_data.va_base =
|
|
|
|
mem_descr->mem_array[idx].
|
|
|
|
virtual_address;
|
|
|
|
pasync_ctx->async_data.pa_base.u.
|
|
|
|
a64.address =
|
|
|
|
mem_descr->mem_array[idx].
|
|
|
|
bus_address.u.a64.address;
|
|
|
|
num_async_data =
|
|
|
|
((mem_descr->mem_array[idx].
|
|
|
|
size) /
|
|
|
|
phba->params.defpdu_data_sz);
|
|
|
|
}
|
|
|
|
pasync_data_h->pbuffer =
|
|
|
|
(void *)((unsigned long)
|
|
|
|
(pasync_ctx->async_data.va_base) +
|
|
|
|
(p->defpdu_data_sz * num_per_mem));
|
|
|
|
|
|
|
|
pasync_data_h->pa.u.a64.address =
|
|
|
|
pasync_ctx->async_data.pa_base.u.a64.
|
|
|
|
address + (p->defpdu_data_sz *
|
|
|
|
num_per_mem);
|
|
|
|
num_per_mem++;
|
|
|
|
num_async_data--;
|
|
|
|
|
|
|
|
list_add_tail(&pasync_data_h->link,
|
|
|
|
&pasync_ctx->async_data.
|
|
|
|
free_list);
|
|
|
|
pasync_data_h++;
|
|
|
|
pasync_ctx->async_data.free_entries++;
|
|
|
|
pasync_ctx->async_data.writables++;
|
|
|
|
|
|
|
|
INIT_LIST_HEAD(&pasync_ctx->async_entry[index].
|
|
|
|
data_busy_list);
|
|
|
|
}
|
2009-09-05 09:06:35 +07:00
|
|
|
|
2013-09-29 05:35:45 +07:00
|
|
|
pasync_ctx->async_header.host_write_ptr = 0;
|
|
|
|
pasync_ctx->async_header.ep_read_ptr = -1;
|
|
|
|
pasync_ctx->async_data.host_write_ptr = 0;
|
|
|
|
pasync_ctx->async_data.ep_read_ptr = -1;
|
|
|
|
}
|
2009-09-05 09:06:35 +07:00
|
|
|
}
|
|
|
|
|
2013-04-06 10:38:32 +07:00
|
|
|
return 0;
|
2009-09-05 09:06:35 +07:00
|
|
|
}
|
|
|
|
|
|
|
|
static int
|
|
|
|
be_sgl_create_contiguous(void *virtual_address,
|
|
|
|
u64 physical_address, u32 length,
|
|
|
|
struct be_dma_mem *sgl)
|
|
|
|
{
|
|
|
|
WARN_ON(!virtual_address);
|
|
|
|
WARN_ON(!physical_address);
|
2015-10-31 01:22:58 +07:00
|
|
|
WARN_ON(!length);
|
2009-09-05 09:06:35 +07:00
|
|
|
WARN_ON(!sgl);
|
|
|
|
|
|
|
|
sgl->va = virtual_address;
|
2010-07-22 05:46:00 +07:00
|
|
|
sgl->dma = (unsigned long)physical_address;
|
2009-09-05 09:06:35 +07:00
|
|
|
sgl->size = length;
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void be_sgl_destroy_contiguous(struct be_dma_mem *sgl)
|
|
|
|
{
|
|
|
|
memset(sgl, 0, sizeof(*sgl));
|
|
|
|
}
|
|
|
|
|
|
|
|
static void
|
|
|
|
hwi_build_be_sgl_arr(struct beiscsi_hba *phba,
|
|
|
|
struct mem_array *pmem, struct be_dma_mem *sgl)
|
|
|
|
{
|
|
|
|
if (sgl->va)
|
|
|
|
be_sgl_destroy_contiguous(sgl);
|
|
|
|
|
|
|
|
be_sgl_create_contiguous(pmem->virtual_address,
|
|
|
|
pmem->bus_address.u.a64.address,
|
|
|
|
pmem->size, sgl);
|
|
|
|
}
|
|
|
|
|
|
|
|
static void
|
|
|
|
hwi_build_be_sgl_by_offset(struct beiscsi_hba *phba,
|
|
|
|
struct mem_array *pmem, struct be_dma_mem *sgl)
|
|
|
|
{
|
|
|
|
if (sgl->va)
|
|
|
|
be_sgl_destroy_contiguous(sgl);
|
|
|
|
|
|
|
|
be_sgl_create_contiguous((unsigned char *)pmem->virtual_address,
|
|
|
|
pmem->bus_address.u.a64.address,
|
|
|
|
pmem->size, sgl);
|
|
|
|
}
|
|
|
|
|
|
|
|
static int be_fill_queue(struct be_queue_info *q,
|
|
|
|
u16 len, u16 entry_size, void *vaddress)
|
|
|
|
{
|
|
|
|
struct be_dma_mem *mem = &q->dma_mem;
|
|
|
|
|
|
|
|
memset(q, 0, sizeof(*q));
|
|
|
|
q->len = len;
|
|
|
|
q->entry_size = entry_size;
|
|
|
|
mem->size = len * entry_size;
|
|
|
|
mem->va = vaddress;
|
|
|
|
if (!mem->va)
|
|
|
|
return -ENOMEM;
|
|
|
|
memset(mem->va, 0, mem->size);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2009-10-23 13:22:33 +07:00
|
|
|
static int beiscsi_create_eqs(struct beiscsi_hba *phba,
|
2009-09-05 09:06:35 +07:00
|
|
|
struct hwi_context_memory *phwi_context)
|
|
|
|
{
|
2009-10-23 13:22:33 +07:00
|
|
|
unsigned int i, num_eq_pages;
|
2012-08-21 00:30:18 +07:00
|
|
|
int ret = 0, eq_for_mcc;
|
2009-09-05 09:06:35 +07:00
|
|
|
struct be_queue_info *eq;
|
|
|
|
struct be_dma_mem *mem;
|
|
|
|
void *eq_vaddress;
|
2009-10-23 13:22:33 +07:00
|
|
|
dma_addr_t paddr;
|
2009-09-05 09:06:35 +07:00
|
|
|
|
2009-10-23 13:22:33 +07:00
|
|
|
num_eq_pages = PAGES_REQUIRED(phba->params.num_eq_entries * \
|
|
|
|
sizeof(struct be_eq_entry));
|
2009-09-05 09:06:35 +07:00
|
|
|
|
2009-10-23 13:22:33 +07:00
|
|
|
if (phba->msix_enabled)
|
|
|
|
eq_for_mcc = 1;
|
|
|
|
else
|
|
|
|
eq_for_mcc = 0;
|
|
|
|
for (i = 0; i < (phba->num_cpus + eq_for_mcc); i++) {
|
|
|
|
eq = &phwi_context->be_eq[i].q;
|
|
|
|
mem = &eq->dma_mem;
|
|
|
|
phwi_context->be_eq[i].phba = phba;
|
|
|
|
eq_vaddress = pci_alloc_consistent(phba->pcidev,
|
|
|
|
num_eq_pages * PAGE_SIZE,
|
|
|
|
&paddr);
|
|
|
|
if (!eq_vaddress)
|
|
|
|
goto create_eq_error;
|
|
|
|
|
|
|
|
mem->va = eq_vaddress;
|
|
|
|
ret = be_fill_queue(eq, phba->params.num_eq_entries,
|
|
|
|
sizeof(struct be_eq_entry), eq_vaddress);
|
|
|
|
if (ret) {
|
2012-08-21 00:30:18 +07:00
|
|
|
beiscsi_log(phba, KERN_ERR, BEISCSI_LOG_INIT,
|
|
|
|
"BM_%d : be_fill_queue Failed for EQ\n");
|
2009-10-23 13:22:33 +07:00
|
|
|
goto create_eq_error;
|
|
|
|
}
|
2009-09-05 09:06:35 +07:00
|
|
|
|
2009-10-23 13:22:33 +07:00
|
|
|
mem->dma = paddr;
|
|
|
|
ret = beiscsi_cmd_eq_create(&phba->ctrl, eq,
|
|
|
|
phwi_context->cur_eqd);
|
|
|
|
if (ret) {
|
2012-08-21 00:30:18 +07:00
|
|
|
beiscsi_log(phba, KERN_ERR, BEISCSI_LOG_INIT,
|
|
|
|
"BM_%d : beiscsi_cmd_eq_create"
|
|
|
|
"Failed for EQ\n");
|
2009-10-23 13:22:33 +07:00
|
|
|
goto create_eq_error;
|
|
|
|
}
|
2012-08-21 00:30:18 +07:00
|
|
|
|
|
|
|
beiscsi_log(phba, KERN_INFO, BEISCSI_LOG_INIT,
|
|
|
|
"BM_%d : eqid = %d\n",
|
|
|
|
phwi_context->be_eq[i].q.id);
|
2009-09-05 09:06:35 +07:00
|
|
|
}
|
|
|
|
return 0;
|
2009-10-23 13:22:33 +07:00
|
|
|
create_eq_error:
|
2012-10-20 06:12:13 +07:00
|
|
|
for (i = 0; i < (phba->num_cpus + eq_for_mcc); i++) {
|
2009-10-23 13:22:33 +07:00
|
|
|
eq = &phwi_context->be_eq[i].q;
|
|
|
|
mem = &eq->dma_mem;
|
|
|
|
if (mem->va)
|
|
|
|
pci_free_consistent(phba->pcidev, num_eq_pages
|
|
|
|
* PAGE_SIZE,
|
|
|
|
mem->va, mem->dma);
|
|
|
|
}
|
|
|
|
return ret;
|
2009-09-05 09:06:35 +07:00
|
|
|
}
|
|
|
|
|
2009-10-23 13:22:33 +07:00
|
|
|
static int beiscsi_create_cqs(struct beiscsi_hba *phba,
|
2009-09-05 09:06:35 +07:00
|
|
|
struct hwi_context_memory *phwi_context)
|
|
|
|
{
|
2009-10-23 13:22:33 +07:00
|
|
|
unsigned int i, num_cq_pages;
|
2012-08-21 00:30:18 +07:00
|
|
|
int ret = 0;
|
2009-09-05 09:06:35 +07:00
|
|
|
struct be_queue_info *cq, *eq;
|
|
|
|
struct be_dma_mem *mem;
|
2009-10-23 13:22:33 +07:00
|
|
|
struct be_eq_obj *pbe_eq;
|
2009-09-05 09:06:35 +07:00
|
|
|
void *cq_vaddress;
|
2009-10-23 13:22:33 +07:00
|
|
|
dma_addr_t paddr;
|
2009-09-05 09:06:35 +07:00
|
|
|
|
2009-10-23 13:22:33 +07:00
|
|
|
num_cq_pages = PAGES_REQUIRED(phba->params.num_cq_entries * \
|
|
|
|
sizeof(struct sol_cqe));
|
2009-09-05 09:06:35 +07:00
|
|
|
|
2009-10-23 13:22:33 +07:00
|
|
|
for (i = 0; i < phba->num_cpus; i++) {
|
|
|
|
cq = &phwi_context->be_cq[i];
|
|
|
|
eq = &phwi_context->be_eq[i].q;
|
|
|
|
pbe_eq = &phwi_context->be_eq[i];
|
|
|
|
pbe_eq->cq = cq;
|
|
|
|
pbe_eq->phba = phba;
|
|
|
|
mem = &cq->dma_mem;
|
|
|
|
cq_vaddress = pci_alloc_consistent(phba->pcidev,
|
|
|
|
num_cq_pages * PAGE_SIZE,
|
|
|
|
&paddr);
|
|
|
|
if (!cq_vaddress)
|
|
|
|
goto create_cq_error;
|
2010-01-05 06:34:12 +07:00
|
|
|
ret = be_fill_queue(cq, phba->params.num_cq_entries,
|
2009-10-23 13:22:33 +07:00
|
|
|
sizeof(struct sol_cqe), cq_vaddress);
|
|
|
|
if (ret) {
|
2012-08-21 00:30:18 +07:00
|
|
|
beiscsi_log(phba, KERN_ERR, BEISCSI_LOG_INIT,
|
|
|
|
"BM_%d : be_fill_queue Failed "
|
|
|
|
"for ISCSI CQ\n");
|
2009-10-23 13:22:33 +07:00
|
|
|
goto create_cq_error;
|
|
|
|
}
|
|
|
|
|
|
|
|
mem->dma = paddr;
|
|
|
|
ret = beiscsi_cmd_cq_create(&phba->ctrl, cq, eq, false,
|
|
|
|
false, 0);
|
|
|
|
if (ret) {
|
2012-08-21 00:30:18 +07:00
|
|
|
beiscsi_log(phba, KERN_ERR, BEISCSI_LOG_INIT,
|
|
|
|
"BM_%d : beiscsi_cmd_eq_create"
|
|
|
|
"Failed for ISCSI CQ\n");
|
2009-10-23 13:22:33 +07:00
|
|
|
goto create_cq_error;
|
|
|
|
}
|
2012-08-21 00:30:18 +07:00
|
|
|
beiscsi_log(phba, KERN_INFO, BEISCSI_LOG_INIT,
|
|
|
|
"BM_%d : iscsi cq_id is %d for eq_id %d\n"
|
|
|
|
"iSCSI CQ CREATED\n", cq->id, eq->id);
|
2009-09-05 09:06:35 +07:00
|
|
|
}
|
|
|
|
return 0;
|
2009-10-23 13:22:33 +07:00
|
|
|
|
|
|
|
create_cq_error:
|
|
|
|
for (i = 0; i < phba->num_cpus; i++) {
|
|
|
|
cq = &phwi_context->be_cq[i];
|
|
|
|
mem = &cq->dma_mem;
|
|
|
|
if (mem->va)
|
|
|
|
pci_free_consistent(phba->pcidev, num_cq_pages
|
|
|
|
* PAGE_SIZE,
|
|
|
|
mem->va, mem->dma);
|
|
|
|
}
|
|
|
|
return ret;
|
|
|
|
|
2009-09-05 09:06:35 +07:00
|
|
|
}
|
|
|
|
|
|
|
|
static int
|
|
|
|
beiscsi_create_def_hdr(struct beiscsi_hba *phba,
|
|
|
|
struct hwi_context_memory *phwi_context,
|
|
|
|
struct hwi_controller *phwi_ctrlr,
|
2013-09-29 05:35:45 +07:00
|
|
|
unsigned int def_pdu_ring_sz, uint8_t ulp_num)
|
2009-09-05 09:06:35 +07:00
|
|
|
{
|
|
|
|
unsigned int idx;
|
|
|
|
int ret;
|
|
|
|
struct be_queue_info *dq, *cq;
|
|
|
|
struct be_dma_mem *mem;
|
|
|
|
struct be_mem_descriptor *mem_descr;
|
|
|
|
void *dq_vaddress;
|
|
|
|
|
|
|
|
idx = 0;
|
2013-09-29 05:35:45 +07:00
|
|
|
dq = &phwi_context->be_def_hdrq[ulp_num];
|
2009-10-23 13:22:33 +07:00
|
|
|
cq = &phwi_context->be_cq[0];
|
2009-09-05 09:06:35 +07:00
|
|
|
mem = &dq->dma_mem;
|
|
|
|
mem_descr = phba->init_mem;
|
2013-09-29 05:35:45 +07:00
|
|
|
mem_descr += HWI_MEM_ASYNC_HEADER_RING_ULP0 +
|
|
|
|
(ulp_num * MEM_DESCR_OFFSET);
|
2009-09-05 09:06:35 +07:00
|
|
|
dq_vaddress = mem_descr->mem_array[idx].virtual_address;
|
|
|
|
ret = be_fill_queue(dq, mem_descr->mem_array[0].size /
|
|
|
|
sizeof(struct phys_addr),
|
|
|
|
sizeof(struct phys_addr), dq_vaddress);
|
|
|
|
if (ret) {
|
2012-08-21 00:30:18 +07:00
|
|
|
beiscsi_log(phba, KERN_ERR, BEISCSI_LOG_INIT,
|
2013-09-29 05:35:45 +07:00
|
|
|
"BM_%d : be_fill_queue Failed for DEF PDU HDR on ULP : %d\n",
|
|
|
|
ulp_num);
|
|
|
|
|
2009-09-05 09:06:35 +07:00
|
|
|
return ret;
|
|
|
|
}
|
2010-07-22 05:46:00 +07:00
|
|
|
mem->dma = (unsigned long)mem_descr->mem_array[idx].
|
|
|
|
bus_address.u.a64.address;
|
2009-09-05 09:06:35 +07:00
|
|
|
ret = be_cmd_create_default_pdu_queue(&phba->ctrl, cq, dq,
|
|
|
|
def_pdu_ring_sz,
|
2013-09-29 05:35:45 +07:00
|
|
|
phba->params.defpdu_hdr_sz,
|
|
|
|
BEISCSI_DEFQ_HDR, ulp_num);
|
2009-09-05 09:06:35 +07:00
|
|
|
if (ret) {
|
2012-08-21 00:30:18 +07:00
|
|
|
beiscsi_log(phba, KERN_ERR, BEISCSI_LOG_INIT,
|
2013-09-29 05:35:45 +07:00
|
|
|
"BM_%d : be_cmd_create_default_pdu_queue Failed DEFHDR on ULP : %d\n",
|
|
|
|
ulp_num);
|
|
|
|
|
2009-09-05 09:06:35 +07:00
|
|
|
return ret;
|
|
|
|
}
|
2012-08-21 00:30:18 +07:00
|
|
|
|
2013-09-29 05:35:45 +07:00
|
|
|
beiscsi_log(phba, KERN_INFO, BEISCSI_LOG_INIT,
|
|
|
|
"BM_%d : iscsi hdr def pdu id for ULP : %d is %d\n",
|
|
|
|
ulp_num,
|
|
|
|
phwi_context->be_def_hdrq[ulp_num].id);
|
|
|
|
hwi_post_async_buffers(phba, BEISCSI_DEFQ_HDR, ulp_num);
|
2009-09-05 09:06:35 +07:00
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int
|
|
|
|
beiscsi_create_def_data(struct beiscsi_hba *phba,
|
|
|
|
struct hwi_context_memory *phwi_context,
|
|
|
|
struct hwi_controller *phwi_ctrlr,
|
2013-09-29 05:35:45 +07:00
|
|
|
unsigned int def_pdu_ring_sz, uint8_t ulp_num)
|
2009-09-05 09:06:35 +07:00
|
|
|
{
|
|
|
|
unsigned int idx;
|
|
|
|
int ret;
|
|
|
|
struct be_queue_info *dataq, *cq;
|
|
|
|
struct be_dma_mem *mem;
|
|
|
|
struct be_mem_descriptor *mem_descr;
|
|
|
|
void *dq_vaddress;
|
|
|
|
|
|
|
|
idx = 0;
|
2013-09-29 05:35:45 +07:00
|
|
|
dataq = &phwi_context->be_def_dataq[ulp_num];
|
2009-10-23 13:22:33 +07:00
|
|
|
cq = &phwi_context->be_cq[0];
|
2009-09-05 09:06:35 +07:00
|
|
|
mem = &dataq->dma_mem;
|
|
|
|
mem_descr = phba->init_mem;
|
2013-09-29 05:35:45 +07:00
|
|
|
mem_descr += HWI_MEM_ASYNC_DATA_RING_ULP0 +
|
|
|
|
(ulp_num * MEM_DESCR_OFFSET);
|
2009-09-05 09:06:35 +07:00
|
|
|
dq_vaddress = mem_descr->mem_array[idx].virtual_address;
|
|
|
|
ret = be_fill_queue(dataq, mem_descr->mem_array[0].size /
|
|
|
|
sizeof(struct phys_addr),
|
|
|
|
sizeof(struct phys_addr), dq_vaddress);
|
|
|
|
if (ret) {
|
2012-08-21 00:30:18 +07:00
|
|
|
beiscsi_log(phba, KERN_ERR, BEISCSI_LOG_INIT,
|
2013-09-29 05:35:45 +07:00
|
|
|
"BM_%d : be_fill_queue Failed for DEF PDU "
|
|
|
|
"DATA on ULP : %d\n",
|
|
|
|
ulp_num);
|
|
|
|
|
2009-09-05 09:06:35 +07:00
|
|
|
return ret;
|
|
|
|
}
|
2010-07-22 05:46:00 +07:00
|
|
|
mem->dma = (unsigned long)mem_descr->mem_array[idx].
|
|
|
|
bus_address.u.a64.address;
|
2009-09-05 09:06:35 +07:00
|
|
|
ret = be_cmd_create_default_pdu_queue(&phba->ctrl, cq, dataq,
|
|
|
|
def_pdu_ring_sz,
|
2013-09-29 05:35:45 +07:00
|
|
|
phba->params.defpdu_data_sz,
|
|
|
|
BEISCSI_DEFQ_DATA, ulp_num);
|
2009-09-05 09:06:35 +07:00
|
|
|
if (ret) {
|
2012-08-21 00:30:18 +07:00
|
|
|
beiscsi_log(phba, KERN_ERR, BEISCSI_LOG_INIT,
|
|
|
|
"BM_%d be_cmd_create_default_pdu_queue"
|
2013-09-29 05:35:45 +07:00
|
|
|
" Failed for DEF PDU DATA on ULP : %d\n",
|
|
|
|
ulp_num);
|
2009-09-05 09:06:35 +07:00
|
|
|
return ret;
|
|
|
|
}
|
2013-09-29 05:35:45 +07:00
|
|
|
|
2012-08-21 00:30:18 +07:00
|
|
|
beiscsi_log(phba, KERN_INFO, BEISCSI_LOG_INIT,
|
2013-09-29 05:35:45 +07:00
|
|
|
"BM_%d : iscsi def data id on ULP : %d is %d\n",
|
|
|
|
ulp_num,
|
|
|
|
phwi_context->be_def_dataq[ulp_num].id);
|
2012-08-21 00:30:18 +07:00
|
|
|
|
2013-09-29 05:35:45 +07:00
|
|
|
hwi_post_async_buffers(phba, BEISCSI_DEFQ_DATA, ulp_num);
|
2012-08-21 00:30:18 +07:00
|
|
|
beiscsi_log(phba, KERN_INFO, BEISCSI_LOG_INIT,
|
2013-09-29 05:35:45 +07:00
|
|
|
"BM_%d : DEFAULT PDU DATA RING CREATED"
|
|
|
|
"on ULP : %d\n", ulp_num);
|
2012-08-21 00:30:18 +07:00
|
|
|
|
2009-09-05 09:06:35 +07:00
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2013-09-29 05:35:38 +07:00
|
|
|
|
|
|
|
static int
|
|
|
|
beiscsi_post_template_hdr(struct beiscsi_hba *phba)
|
|
|
|
{
|
|
|
|
struct be_mem_descriptor *mem_descr;
|
|
|
|
struct mem_array *pm_arr;
|
|
|
|
struct be_dma_mem sgl;
|
2013-09-29 05:35:46 +07:00
|
|
|
int status, ulp_num;
|
2013-09-29 05:35:38 +07:00
|
|
|
|
2013-09-29 05:35:46 +07:00
|
|
|
for (ulp_num = 0; ulp_num < BEISCSI_ULP_COUNT; ulp_num++) {
|
|
|
|
if (test_bit(ulp_num, &phba->fw_config.ulp_supported)) {
|
|
|
|
mem_descr = (struct be_mem_descriptor *)phba->init_mem;
|
|
|
|
mem_descr += HWI_MEM_TEMPLATE_HDR_ULP0 +
|
|
|
|
(ulp_num * MEM_DESCR_OFFSET);
|
|
|
|
pm_arr = mem_descr->mem_array;
|
2013-09-29 05:35:38 +07:00
|
|
|
|
2013-09-29 05:35:46 +07:00
|
|
|
hwi_build_be_sgl_arr(phba, pm_arr, &sgl);
|
|
|
|
status = be_cmd_iscsi_post_template_hdr(
|
|
|
|
&phba->ctrl, &sgl);
|
2013-09-29 05:35:38 +07:00
|
|
|
|
2013-09-29 05:35:46 +07:00
|
|
|
if (status != 0) {
|
|
|
|
beiscsi_log(phba, KERN_ERR, BEISCSI_LOG_INIT,
|
|
|
|
"BM_%d : Post Template HDR Failed for"
|
|
|
|
"ULP_%d\n", ulp_num);
|
|
|
|
return status;
|
|
|
|
}
|
|
|
|
|
|
|
|
beiscsi_log(phba, KERN_INFO, BEISCSI_LOG_INIT,
|
|
|
|
"BM_%d : Template HDR Pages Posted for"
|
|
|
|
"ULP_%d\n", ulp_num);
|
2013-09-29 05:35:38 +07:00
|
|
|
}
|
|
|
|
}
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2009-09-05 09:06:35 +07:00
|
|
|
static int
|
|
|
|
beiscsi_post_pages(struct beiscsi_hba *phba)
|
|
|
|
{
|
|
|
|
struct be_mem_descriptor *mem_descr;
|
|
|
|
struct mem_array *pm_arr;
|
|
|
|
unsigned int page_offset, i;
|
|
|
|
struct be_dma_mem sgl;
|
2013-09-29 05:35:44 +07:00
|
|
|
int status, ulp_num = 0;
|
2009-09-05 09:06:35 +07:00
|
|
|
|
|
|
|
mem_descr = phba->init_mem;
|
|
|
|
mem_descr += HWI_MEM_SGE;
|
|
|
|
pm_arr = mem_descr->mem_array;
|
|
|
|
|
2013-09-29 05:35:47 +07:00
|
|
|
for (ulp_num = 0; ulp_num < BEISCSI_ULP_COUNT; ulp_num++)
|
|
|
|
if (test_bit(ulp_num, &phba->fw_config.ulp_supported))
|
|
|
|
break;
|
|
|
|
|
2009-09-05 09:06:35 +07:00
|
|
|
page_offset = (sizeof(struct iscsi_sge) * phba->params.num_sge_per_io *
|
2013-09-29 05:35:44 +07:00
|
|
|
phba->fw_config.iscsi_icd_start[ulp_num]) / PAGE_SIZE;
|
2009-09-05 09:06:35 +07:00
|
|
|
for (i = 0; i < mem_descr->num_elements; i++) {
|
|
|
|
hwi_build_be_sgl_arr(phba, pm_arr, &sgl);
|
|
|
|
status = be_cmd_iscsi_post_sgl_pages(&phba->ctrl, &sgl,
|
|
|
|
page_offset,
|
|
|
|
(pm_arr->size / PAGE_SIZE));
|
|
|
|
page_offset += pm_arr->size / PAGE_SIZE;
|
|
|
|
if (status != 0) {
|
2012-08-21 00:30:18 +07:00
|
|
|
beiscsi_log(phba, KERN_ERR, BEISCSI_LOG_INIT,
|
|
|
|
"BM_%d : post sgl failed.\n");
|
2009-09-05 09:06:35 +07:00
|
|
|
return status;
|
|
|
|
}
|
|
|
|
pm_arr++;
|
|
|
|
}
|
2012-08-21 00:30:18 +07:00
|
|
|
beiscsi_log(phba, KERN_INFO, BEISCSI_LOG_INIT,
|
|
|
|
"BM_%d : POSTED PAGES\n");
|
2009-09-05 09:06:35 +07:00
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2009-10-23 13:22:33 +07:00
|
|
|
static void be_queue_free(struct beiscsi_hba *phba, struct be_queue_info *q)
|
|
|
|
{
|
|
|
|
struct be_dma_mem *mem = &q->dma_mem;
|
2012-04-04 11:41:42 +07:00
|
|
|
if (mem->va) {
|
2009-10-23 13:22:33 +07:00
|
|
|
pci_free_consistent(phba->pcidev, mem->size,
|
|
|
|
mem->va, mem->dma);
|
2012-04-04 11:41:42 +07:00
|
|
|
mem->va = NULL;
|
|
|
|
}
|
2009-10-23 13:22:33 +07:00
|
|
|
}
|
|
|
|
|
|
|
|
static int be_queue_alloc(struct beiscsi_hba *phba, struct be_queue_info *q,
|
|
|
|
u16 len, u16 entry_size)
|
|
|
|
{
|
|
|
|
struct be_dma_mem *mem = &q->dma_mem;
|
|
|
|
|
|
|
|
memset(q, 0, sizeof(*q));
|
|
|
|
q->len = len;
|
|
|
|
q->entry_size = entry_size;
|
|
|
|
mem->size = len * entry_size;
|
2014-08-09 04:24:46 +07:00
|
|
|
mem->va = pci_zalloc_consistent(phba->pcidev, mem->size, &mem->dma);
|
2009-10-23 13:22:33 +07:00
|
|
|
if (!mem->va)
|
2010-07-22 05:46:38 +07:00
|
|
|
return -ENOMEM;
|
2009-10-23 13:22:33 +07:00
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
2009-09-05 09:06:35 +07:00
|
|
|
static int
|
|
|
|
beiscsi_create_wrb_rings(struct beiscsi_hba *phba,
|
|
|
|
struct hwi_context_memory *phwi_context,
|
|
|
|
struct hwi_controller *phwi_ctrlr)
|
|
|
|
{
|
|
|
|
unsigned int wrb_mem_index, offset, size, num_wrb_rings;
|
|
|
|
u64 pa_addr_lo;
|
2013-09-29 05:35:48 +07:00
|
|
|
unsigned int idx, num, i, ulp_num;
|
2009-09-05 09:06:35 +07:00
|
|
|
struct mem_array *pwrb_arr;
|
|
|
|
void *wrb_vaddr;
|
|
|
|
struct be_dma_mem sgl;
|
|
|
|
struct be_mem_descriptor *mem_descr;
|
2013-04-06 10:38:32 +07:00
|
|
|
struct hwi_wrb_context *pwrb_context;
|
2009-09-05 09:06:35 +07:00
|
|
|
int status;
|
2013-09-29 05:35:48 +07:00
|
|
|
uint8_t ulp_count = 0, ulp_base_num = 0;
|
|
|
|
uint16_t cid_count_ulp[BEISCSI_ULP_COUNT] = { 0 };
|
2009-09-05 09:06:35 +07:00
|
|
|
|
|
|
|
idx = 0;
|
|
|
|
mem_descr = phba->init_mem;
|
|
|
|
mem_descr += HWI_MEM_WRB;
|
|
|
|
pwrb_arr = kmalloc(sizeof(*pwrb_arr) * phba->params.cxns_per_ctrl,
|
|
|
|
GFP_KERNEL);
|
|
|
|
if (!pwrb_arr) {
|
2012-08-21 00:30:18 +07:00
|
|
|
beiscsi_log(phba, KERN_ERR, BEISCSI_LOG_INIT,
|
|
|
|
"BM_%d : Memory alloc failed in create wrb ring.\n");
|
2009-09-05 09:06:35 +07:00
|
|
|
return -ENOMEM;
|
|
|
|
}
|
|
|
|
wrb_vaddr = mem_descr->mem_array[idx].virtual_address;
|
|
|
|
pa_addr_lo = mem_descr->mem_array[idx].bus_address.u.a64.address;
|
|
|
|
num_wrb_rings = mem_descr->mem_array[idx].size /
|
|
|
|
(phba->params.wrbs_per_cxn * sizeof(struct iscsi_wrb));
|
|
|
|
|
|
|
|
for (num = 0; num < phba->params.cxns_per_ctrl; num++) {
|
|
|
|
if (num_wrb_rings) {
|
|
|
|
pwrb_arr[num].virtual_address = wrb_vaddr;
|
|
|
|
pwrb_arr[num].bus_address.u.a64.address = pa_addr_lo;
|
|
|
|
pwrb_arr[num].size = phba->params.wrbs_per_cxn *
|
|
|
|
sizeof(struct iscsi_wrb);
|
|
|
|
wrb_vaddr += pwrb_arr[num].size;
|
|
|
|
pa_addr_lo += pwrb_arr[num].size;
|
|
|
|
num_wrb_rings--;
|
|
|
|
} else {
|
|
|
|
idx++;
|
|
|
|
wrb_vaddr = mem_descr->mem_array[idx].virtual_address;
|
|
|
|
pa_addr_lo = mem_descr->mem_array[idx].\
|
|
|
|
bus_address.u.a64.address;
|
|
|
|
num_wrb_rings = mem_descr->mem_array[idx].size /
|
|
|
|
(phba->params.wrbs_per_cxn *
|
|
|
|
sizeof(struct iscsi_wrb));
|
|
|
|
pwrb_arr[num].virtual_address = wrb_vaddr;
|
|
|
|
pwrb_arr[num].bus_address.u.a64.address\
|
|
|
|
= pa_addr_lo;
|
|
|
|
pwrb_arr[num].size = phba->params.wrbs_per_cxn *
|
|
|
|
sizeof(struct iscsi_wrb);
|
|
|
|
wrb_vaddr += pwrb_arr[num].size;
|
|
|
|
pa_addr_lo += pwrb_arr[num].size;
|
|
|
|
num_wrb_rings--;
|
|
|
|
}
|
|
|
|
}
|
2013-09-29 05:35:48 +07:00
|
|
|
|
|
|
|
/* Get the ULP Count */
|
|
|
|
for (ulp_num = 0; ulp_num < BEISCSI_ULP_COUNT; ulp_num++)
|
|
|
|
if (test_bit(ulp_num, &phba->fw_config.ulp_supported)) {
|
|
|
|
ulp_count++;
|
|
|
|
ulp_base_num = ulp_num;
|
|
|
|
cid_count_ulp[ulp_num] =
|
|
|
|
BEISCSI_GET_CID_COUNT(phba, ulp_num);
|
|
|
|
}
|
|
|
|
|
2009-09-05 09:06:35 +07:00
|
|
|
for (i = 0; i < phba->params.cxns_per_ctrl; i++) {
|
|
|
|
wrb_mem_index = 0;
|
|
|
|
offset = 0;
|
|
|
|
size = 0;
|
|
|
|
|
2013-09-29 05:35:48 +07:00
|
|
|
if (ulp_count > 1) {
|
|
|
|
ulp_base_num = (ulp_base_num + 1) % BEISCSI_ULP_COUNT;
|
|
|
|
|
|
|
|
if (!cid_count_ulp[ulp_base_num])
|
|
|
|
ulp_base_num = (ulp_base_num + 1) %
|
|
|
|
BEISCSI_ULP_COUNT;
|
|
|
|
|
|
|
|
cid_count_ulp[ulp_base_num]--;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
2009-09-05 09:06:35 +07:00
|
|
|
hwi_build_be_sgl_by_offset(phba, &pwrb_arr[i], &sgl);
|
|
|
|
status = be_cmd_wrbq_create(&phba->ctrl, &sgl,
|
2013-09-29 05:35:48 +07:00
|
|
|
&phwi_context->be_wrbq[i],
|
|
|
|
&phwi_ctrlr->wrb_context[i],
|
|
|
|
ulp_base_num);
|
2009-09-05 09:06:35 +07:00
|
|
|
if (status != 0) {
|
2012-08-21 00:30:18 +07:00
|
|
|
beiscsi_log(phba, KERN_ERR, BEISCSI_LOG_INIT,
|
|
|
|
"BM_%d : wrbq create failed.");
|
2010-06-10 14:52:21 +07:00
|
|
|
kfree(pwrb_arr);
|
2009-09-05 09:06:35 +07:00
|
|
|
return status;
|
|
|
|
}
|
2013-04-06 10:38:32 +07:00
|
|
|
pwrb_context = &phwi_ctrlr->wrb_context[i];
|
|
|
|
BE_SET_CID_TO_CRI(i, pwrb_context->cid);
|
2009-09-05 09:06:35 +07:00
|
|
|
}
|
|
|
|
kfree(pwrb_arr);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void free_wrb_handles(struct beiscsi_hba *phba)
|
|
|
|
{
|
|
|
|
unsigned int index;
|
|
|
|
struct hwi_controller *phwi_ctrlr;
|
|
|
|
struct hwi_wrb_context *pwrb_context;
|
|
|
|
|
|
|
|
phwi_ctrlr = phba->phwi_ctrlr;
|
2013-04-06 10:38:32 +07:00
|
|
|
for (index = 0; index < phba->params.cxns_per_ctrl; index++) {
|
2009-09-05 09:06:35 +07:00
|
|
|
pwrb_context = &phwi_ctrlr->wrb_context[index];
|
|
|
|
kfree(pwrb_context->pwrb_handle_base);
|
|
|
|
kfree(pwrb_context->pwrb_handle_basestd);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2009-10-23 13:22:33 +07:00
|
|
|
static void be_mcc_queues_destroy(struct beiscsi_hba *phba)
|
|
|
|
{
|
|
|
|
struct be_queue_info *q;
|
|
|
|
struct be_ctrl_info *ctrl = &phba->ctrl;
|
|
|
|
|
|
|
|
q = &phba->ctrl.mcc_obj.q;
|
2015-04-25 09:47:31 +07:00
|
|
|
if (q->created) {
|
2009-10-23 13:22:33 +07:00
|
|
|
beiscsi_cmd_q_destroy(ctrl, q, QTYPE_MCCQ);
|
2015-04-25 09:47:31 +07:00
|
|
|
be_queue_free(phba, q);
|
|
|
|
}
|
2009-10-23 13:22:33 +07:00
|
|
|
|
|
|
|
q = &phba->ctrl.mcc_obj.cq;
|
2015-04-25 09:47:31 +07:00
|
|
|
if (q->created) {
|
2009-10-23 13:22:33 +07:00
|
|
|
beiscsi_cmd_q_destroy(ctrl, q, QTYPE_CQ);
|
2015-04-25 09:47:31 +07:00
|
|
|
be_queue_free(phba, q);
|
|
|
|
}
|
2009-10-23 13:22:33 +07:00
|
|
|
}
|
|
|
|
|
2009-09-05 09:06:35 +07:00
|
|
|
static void hwi_cleanup(struct beiscsi_hba *phba)
|
|
|
|
{
|
|
|
|
struct be_queue_info *q;
|
|
|
|
struct be_ctrl_info *ctrl = &phba->ctrl;
|
|
|
|
struct hwi_controller *phwi_ctrlr;
|
|
|
|
struct hwi_context_memory *phwi_context;
|
2013-04-06 10:38:32 +07:00
|
|
|
struct hwi_async_pdu_context *pasync_ctx;
|
2014-05-06 08:41:29 +07:00
|
|
|
int i, eq_for_mcc, ulp_num;
|
2009-09-05 09:06:35 +07:00
|
|
|
|
|
|
|
phwi_ctrlr = phba->phwi_ctrlr;
|
|
|
|
phwi_context = phwi_ctrlr->phwi_ctxt;
|
2013-09-29 05:35:38 +07:00
|
|
|
|
|
|
|
be_cmd_iscsi_remove_template_hdr(ctrl);
|
|
|
|
|
2009-09-05 09:06:35 +07:00
|
|
|
for (i = 0; i < phba->params.cxns_per_ctrl; i++) {
|
|
|
|
q = &phwi_context->be_wrbq[i];
|
|
|
|
if (q->created)
|
|
|
|
beiscsi_cmd_q_destroy(ctrl, q, QTYPE_WRBQ);
|
|
|
|
}
|
2013-04-06 10:38:32 +07:00
|
|
|
kfree(phwi_context->be_wrbq);
|
2009-09-05 09:06:35 +07:00
|
|
|
free_wrb_handles(phba);
|
|
|
|
|
2013-09-29 05:35:45 +07:00
|
|
|
for (ulp_num = 0; ulp_num < BEISCSI_ULP_COUNT; ulp_num++) {
|
|
|
|
if (test_bit(ulp_num, &phba->fw_config.ulp_supported)) {
|
2009-09-05 09:06:35 +07:00
|
|
|
|
2013-09-29 05:35:45 +07:00
|
|
|
q = &phwi_context->be_def_hdrq[ulp_num];
|
|
|
|
if (q->created)
|
|
|
|
beiscsi_cmd_q_destroy(ctrl, q, QTYPE_DPDUQ);
|
|
|
|
|
|
|
|
q = &phwi_context->be_def_dataq[ulp_num];
|
|
|
|
if (q->created)
|
|
|
|
beiscsi_cmd_q_destroy(ctrl, q, QTYPE_DPDUQ);
|
|
|
|
|
|
|
|
pasync_ctx = phwi_ctrlr->phwi_ctxt->pasync_ctx[ulp_num];
|
|
|
|
}
|
|
|
|
}
|
2009-09-05 09:06:35 +07:00
|
|
|
|
|
|
|
beiscsi_cmd_q_destroy(ctrl, NULL, QTYPE_SGL);
|
|
|
|
|
2009-10-23 13:22:33 +07:00
|
|
|
for (i = 0; i < (phba->num_cpus); i++) {
|
|
|
|
q = &phwi_context->be_cq[i];
|
2015-04-25 09:47:31 +07:00
|
|
|
if (q->created) {
|
|
|
|
be_queue_free(phba, q);
|
2009-10-23 13:22:33 +07:00
|
|
|
beiscsi_cmd_q_destroy(ctrl, q, QTYPE_CQ);
|
2015-04-25 09:47:31 +07:00
|
|
|
}
|
2009-10-23 13:22:33 +07:00
|
|
|
}
|
2014-05-06 08:41:29 +07:00
|
|
|
|
|
|
|
be_mcc_queues_destroy(phba);
|
2009-10-23 13:22:33 +07:00
|
|
|
if (phba->msix_enabled)
|
2014-05-06 08:41:29 +07:00
|
|
|
eq_for_mcc = 1;
|
2009-10-23 13:22:33 +07:00
|
|
|
else
|
2014-05-06 08:41:29 +07:00
|
|
|
eq_for_mcc = 0;
|
|
|
|
for (i = 0; i < (phba->num_cpus + eq_for_mcc); i++) {
|
2009-10-23 13:22:33 +07:00
|
|
|
q = &phwi_context->be_eq[i].q;
|
2015-04-25 09:47:31 +07:00
|
|
|
if (q->created) {
|
|
|
|
be_queue_free(phba, q);
|
2009-10-23 13:22:33 +07:00
|
|
|
beiscsi_cmd_q_destroy(ctrl, q, QTYPE_EQ);
|
2015-04-25 09:47:31 +07:00
|
|
|
}
|
2009-10-23 13:22:33 +07:00
|
|
|
}
|
2013-04-06 10:38:21 +07:00
|
|
|
be_cmd_fw_uninit(ctrl);
|
2009-10-23 13:22:33 +07:00
|
|
|
}
|
2009-09-05 09:06:35 +07:00
|
|
|
|
2009-10-23 13:22:33 +07:00
|
|
|
static int be_mcc_queues_create(struct beiscsi_hba *phba,
|
|
|
|
struct hwi_context_memory *phwi_context)
|
|
|
|
{
|
|
|
|
struct be_queue_info *q, *cq;
|
|
|
|
struct be_ctrl_info *ctrl = &phba->ctrl;
|
|
|
|
|
|
|
|
/* Alloc MCC compl queue */
|
|
|
|
cq = &phba->ctrl.mcc_obj.cq;
|
|
|
|
if (be_queue_alloc(phba, cq, MCC_CQ_LEN,
|
|
|
|
sizeof(struct be_mcc_compl)))
|
|
|
|
goto err;
|
|
|
|
/* Ask BE to create MCC compl queue; */
|
|
|
|
if (phba->msix_enabled) {
|
|
|
|
if (beiscsi_cmd_cq_create(ctrl, cq, &phwi_context->be_eq
|
|
|
|
[phba->num_cpus].q, false, true, 0))
|
|
|
|
goto mcc_cq_free;
|
|
|
|
} else {
|
|
|
|
if (beiscsi_cmd_cq_create(ctrl, cq, &phwi_context->be_eq[0].q,
|
|
|
|
false, true, 0))
|
|
|
|
goto mcc_cq_free;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Alloc MCC queue */
|
|
|
|
q = &phba->ctrl.mcc_obj.q;
|
|
|
|
if (be_queue_alloc(phba, q, MCC_Q_LEN, sizeof(struct be_mcc_wrb)))
|
|
|
|
goto mcc_cq_destroy;
|
|
|
|
|
|
|
|
/* Ask BE to create MCC queue */
|
2009-10-23 13:23:49 +07:00
|
|
|
if (beiscsi_cmd_mccq_create(phba, q, cq))
|
2009-10-23 13:22:33 +07:00
|
|
|
goto mcc_q_free;
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
mcc_q_free:
|
|
|
|
be_queue_free(phba, q);
|
|
|
|
mcc_cq_destroy:
|
|
|
|
beiscsi_cmd_q_destroy(ctrl, cq, QTYPE_CQ);
|
|
|
|
mcc_cq_free:
|
|
|
|
be_queue_free(phba, cq);
|
|
|
|
err:
|
2010-07-22 05:46:38 +07:00
|
|
|
return -ENOMEM;
|
2009-10-23 13:22:33 +07:00
|
|
|
}
|
|
|
|
|
2012-10-20 06:12:13 +07:00
|
|
|
/**
|
|
|
|
* find_num_cpus()- Get the CPU online count
|
|
|
|
* @phba: ptr to priv structure
|
|
|
|
*
|
|
|
|
* CPU count is used for creating EQ.
|
|
|
|
**/
|
|
|
|
static void find_num_cpus(struct beiscsi_hba *phba)
|
2009-10-23 13:22:33 +07:00
|
|
|
{
|
|
|
|
int num_cpus = 0;
|
|
|
|
|
|
|
|
num_cpus = num_online_cpus();
|
|
|
|
|
2012-10-20 06:13:32 +07:00
|
|
|
switch (phba->generation) {
|
|
|
|
case BE_GEN2:
|
|
|
|
case BE_GEN3:
|
|
|
|
phba->num_cpus = (num_cpus > BEISCSI_MAX_NUM_CPUS) ?
|
|
|
|
BEISCSI_MAX_NUM_CPUS : num_cpus;
|
|
|
|
break;
|
|
|
|
case BE_GEN4:
|
2013-09-29 05:35:54 +07:00
|
|
|
/*
|
|
|
|
* If eqid_count == 1 fall back to
|
|
|
|
* INTX mechanism
|
|
|
|
**/
|
|
|
|
if (phba->fw_config.eqid_count == 1) {
|
|
|
|
enable_msix = 0;
|
|
|
|
phba->num_cpus = 1;
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
|
|
|
phba->num_cpus =
|
|
|
|
(num_cpus > (phba->fw_config.eqid_count - 1)) ?
|
|
|
|
(phba->fw_config.eqid_count - 1) : num_cpus;
|
2012-10-20 06:13:32 +07:00
|
|
|
break;
|
|
|
|
default:
|
|
|
|
phba->num_cpus = 1;
|
|
|
|
}
|
2009-09-05 09:06:35 +07:00
|
|
|
}
|
|
|
|
|
|
|
|
static int hwi_init_port(struct beiscsi_hba *phba)
|
|
|
|
{
|
|
|
|
struct hwi_controller *phwi_ctrlr;
|
|
|
|
struct hwi_context_memory *phwi_context;
|
|
|
|
unsigned int def_pdu_ring_sz;
|
|
|
|
struct be_ctrl_info *ctrl = &phba->ctrl;
|
2013-09-29 05:35:45 +07:00
|
|
|
int status, ulp_num;
|
2009-09-05 09:06:35 +07:00
|
|
|
|
|
|
|
phwi_ctrlr = phba->phwi_ctrlr;
|
|
|
|
phwi_context = phwi_ctrlr->phwi_ctxt;
|
2014-05-06 08:41:26 +07:00
|
|
|
phwi_context->max_eqd = 128;
|
2009-10-23 13:22:33 +07:00
|
|
|
phwi_context->min_eqd = 0;
|
2014-05-06 08:41:26 +07:00
|
|
|
phwi_context->cur_eqd = 0;
|
2009-09-05 09:06:35 +07:00
|
|
|
be_cmd_fw_initialize(&phba->ctrl);
|
2016-01-20 15:40:53 +07:00
|
|
|
/* set optic state to unknown */
|
|
|
|
phba->optic_state = 0xff;
|
2009-10-23 13:22:33 +07:00
|
|
|
|
|
|
|
status = beiscsi_create_eqs(phba, phwi_context);
|
2009-09-05 09:06:35 +07:00
|
|
|
if (status != 0) {
|
2012-08-21 00:30:18 +07:00
|
|
|
beiscsi_log(phba, KERN_ERR, BEISCSI_LOG_INIT,
|
|
|
|
"BM_%d : EQ not created\n");
|
2009-09-05 09:06:35 +07:00
|
|
|
goto error;
|
|
|
|
}
|
|
|
|
|
2009-10-23 13:22:33 +07:00
|
|
|
status = be_mcc_queues_create(phba, phwi_context);
|
|
|
|
if (status != 0)
|
|
|
|
goto error;
|
|
|
|
|
|
|
|
status = mgmt_check_supported_fw(ctrl, phba);
|
2009-09-05 09:06:35 +07:00
|
|
|
if (status != 0) {
|
2012-08-21 00:30:18 +07:00
|
|
|
beiscsi_log(phba, KERN_ERR, BEISCSI_LOG_INIT,
|
|
|
|
"BM_%d : Unsupported fw version\n");
|
2009-09-05 09:06:35 +07:00
|
|
|
goto error;
|
|
|
|
}
|
|
|
|
|
2009-10-23 13:22:33 +07:00
|
|
|
status = beiscsi_create_cqs(phba, phwi_context);
|
2009-09-05 09:06:35 +07:00
|
|
|
if (status != 0) {
|
2012-08-21 00:30:18 +07:00
|
|
|
beiscsi_log(phba, KERN_ERR, BEISCSI_LOG_INIT,
|
|
|
|
"BM_%d : CQ not created\n");
|
2009-09-05 09:06:35 +07:00
|
|
|
goto error;
|
|
|
|
}
|
|
|
|
|
2013-09-29 05:35:45 +07:00
|
|
|
for (ulp_num = 0; ulp_num < BEISCSI_ULP_COUNT; ulp_num++) {
|
|
|
|
if (test_bit(ulp_num, &phba->fw_config.ulp_supported)) {
|
2009-09-05 09:06:35 +07:00
|
|
|
|
2013-09-29 05:35:45 +07:00
|
|
|
def_pdu_ring_sz =
|
|
|
|
BEISCSI_GET_CID_COUNT(phba, ulp_num) *
|
|
|
|
sizeof(struct phys_addr);
|
|
|
|
|
|
|
|
status = beiscsi_create_def_hdr(phba, phwi_context,
|
|
|
|
phwi_ctrlr,
|
|
|
|
def_pdu_ring_sz,
|
|
|
|
ulp_num);
|
|
|
|
if (status != 0) {
|
|
|
|
beiscsi_log(phba, KERN_ERR, BEISCSI_LOG_INIT,
|
|
|
|
"BM_%d : Default Header not created for ULP : %d\n",
|
|
|
|
ulp_num);
|
|
|
|
goto error;
|
|
|
|
}
|
|
|
|
|
|
|
|
status = beiscsi_create_def_data(phba, phwi_context,
|
|
|
|
phwi_ctrlr,
|
|
|
|
def_pdu_ring_sz,
|
|
|
|
ulp_num);
|
|
|
|
if (status != 0) {
|
|
|
|
beiscsi_log(phba, KERN_ERR, BEISCSI_LOG_INIT,
|
|
|
|
"BM_%d : Default Data not created for ULP : %d\n",
|
|
|
|
ulp_num);
|
|
|
|
goto error;
|
|
|
|
}
|
|
|
|
}
|
2009-09-05 09:06:35 +07:00
|
|
|
}
|
|
|
|
|
|
|
|
status = beiscsi_post_pages(phba);
|
|
|
|
if (status != 0) {
|
2012-08-21 00:30:18 +07:00
|
|
|
beiscsi_log(phba, KERN_ERR, BEISCSI_LOG_INIT,
|
|
|
|
"BM_%d : Post SGL Pages Failed\n");
|
2009-09-05 09:06:35 +07:00
|
|
|
goto error;
|
|
|
|
}
|
|
|
|
|
2013-09-29 05:35:38 +07:00
|
|
|
status = beiscsi_post_template_hdr(phba);
|
|
|
|
if (status != 0) {
|
|
|
|
beiscsi_log(phba, KERN_ERR, BEISCSI_LOG_INIT,
|
|
|
|
"BM_%d : Template HDR Posting for CXN Failed\n");
|
|
|
|
}
|
|
|
|
|
2009-09-05 09:06:35 +07:00
|
|
|
status = beiscsi_create_wrb_rings(phba, phwi_context, phwi_ctrlr);
|
|
|
|
if (status != 0) {
|
2012-08-21 00:30:18 +07:00
|
|
|
beiscsi_log(phba, KERN_ERR, BEISCSI_LOG_INIT,
|
|
|
|
"BM_%d : WRB Rings not created\n");
|
2009-09-05 09:06:35 +07:00
|
|
|
goto error;
|
|
|
|
}
|
|
|
|
|
2013-09-29 05:35:45 +07:00
|
|
|
for (ulp_num = 0; ulp_num < BEISCSI_ULP_COUNT; ulp_num++) {
|
|
|
|
uint16_t async_arr_idx = 0;
|
|
|
|
|
|
|
|
if (test_bit(ulp_num, &phba->fw_config.ulp_supported)) {
|
|
|
|
uint16_t cri = 0;
|
|
|
|
struct hwi_async_pdu_context *pasync_ctx;
|
|
|
|
|
|
|
|
pasync_ctx = HWI_GET_ASYNC_PDU_CTX(
|
|
|
|
phwi_ctrlr, ulp_num);
|
|
|
|
for (cri = 0; cri <
|
|
|
|
phba->params.cxns_per_ctrl; cri++) {
|
|
|
|
if (ulp_num == BEISCSI_GET_ULP_FROM_CRI
|
|
|
|
(phwi_ctrlr, cri))
|
|
|
|
pasync_ctx->cid_to_async_cri_map[
|
|
|
|
phwi_ctrlr->wrb_context[cri].cid] =
|
|
|
|
async_arr_idx++;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2012-08-21 00:30:18 +07:00
|
|
|
beiscsi_log(phba, KERN_INFO, BEISCSI_LOG_INIT,
|
|
|
|
"BM_%d : hwi_init_port success\n");
|
2009-09-05 09:06:35 +07:00
|
|
|
return 0;
|
|
|
|
|
|
|
|
error:
|
2012-08-21 00:30:18 +07:00
|
|
|
beiscsi_log(phba, KERN_ERR, BEISCSI_LOG_INIT,
|
|
|
|
"BM_%d : hwi_init_port failed");
|
2009-09-05 09:06:35 +07:00
|
|
|
hwi_cleanup(phba);
|
2012-04-04 11:41:44 +07:00
|
|
|
return status;
|
2009-09-05 09:06:35 +07:00
|
|
|
}
|
|
|
|
|
|
|
|
static int hwi_init_controller(struct beiscsi_hba *phba)
|
|
|
|
{
|
|
|
|
struct hwi_controller *phwi_ctrlr;
|
|
|
|
|
|
|
|
phwi_ctrlr = phba->phwi_ctrlr;
|
|
|
|
if (1 == phba->init_mem[HWI_MEM_ADDN_CONTEXT].num_elements) {
|
|
|
|
phwi_ctrlr->phwi_ctxt = (struct hwi_context_memory *)phba->
|
|
|
|
init_mem[HWI_MEM_ADDN_CONTEXT].mem_array[0].virtual_address;
|
2012-08-21 00:30:18 +07:00
|
|
|
beiscsi_log(phba, KERN_INFO, BEISCSI_LOG_INIT,
|
|
|
|
"BM_%d : phwi_ctrlr->phwi_ctxt=%p\n",
|
|
|
|
phwi_ctrlr->phwi_ctxt);
|
2009-09-05 09:06:35 +07:00
|
|
|
} else {
|
2012-08-21 00:30:18 +07:00
|
|
|
beiscsi_log(phba, KERN_ERR, BEISCSI_LOG_INIT,
|
|
|
|
"BM_%d : HWI_MEM_ADDN_CONTEXT is more "
|
|
|
|
"than one element.Failing to load\n");
|
2009-09-05 09:06:35 +07:00
|
|
|
return -ENOMEM;
|
|
|
|
}
|
|
|
|
|
|
|
|
iscsi_init_global_templates(phba);
|
2012-04-04 11:41:38 +07:00
|
|
|
if (beiscsi_init_wrb_handle(phba))
|
|
|
|
return -ENOMEM;
|
|
|
|
|
2013-04-06 10:38:32 +07:00
|
|
|
if (hwi_init_async_pdu_ctx(phba)) {
|
|
|
|
beiscsi_log(phba, KERN_ERR, BEISCSI_LOG_INIT,
|
|
|
|
"BM_%d : hwi_init_async_pdu_ctx failed\n");
|
|
|
|
return -ENOMEM;
|
|
|
|
}
|
|
|
|
|
2009-09-05 09:06:35 +07:00
|
|
|
if (hwi_init_port(phba) != 0) {
|
2012-08-21 00:30:18 +07:00
|
|
|
beiscsi_log(phba, KERN_ERR, BEISCSI_LOG_INIT,
|
|
|
|
"BM_%d : hwi_init_controller failed\n");
|
|
|
|
|
2009-09-05 09:06:35 +07:00
|
|
|
return -ENOMEM;
|
|
|
|
}
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void beiscsi_free_mem(struct beiscsi_hba *phba)
|
|
|
|
{
|
|
|
|
struct be_mem_descriptor *mem_descr;
|
|
|
|
int i, j;
|
|
|
|
|
|
|
|
mem_descr = phba->init_mem;
|
|
|
|
i = 0;
|
|
|
|
j = 0;
|
|
|
|
for (i = 0; i < SE_MEM_MAX; i++) {
|
|
|
|
for (j = mem_descr->num_elements; j > 0; j--) {
|
|
|
|
pci_free_consistent(phba->pcidev,
|
|
|
|
mem_descr->mem_array[j - 1].size,
|
|
|
|
mem_descr->mem_array[j - 1].virtual_address,
|
2010-07-22 05:46:00 +07:00
|
|
|
(unsigned long)mem_descr->mem_array[j - 1].
|
|
|
|
bus_address.u.a64.address);
|
2009-09-05 09:06:35 +07:00
|
|
|
}
|
2013-09-29 05:35:45 +07:00
|
|
|
|
2009-09-05 09:06:35 +07:00
|
|
|
kfree(mem_descr->mem_array);
|
|
|
|
mem_descr++;
|
|
|
|
}
|
|
|
|
kfree(phba->init_mem);
|
2013-04-06 10:38:32 +07:00
|
|
|
kfree(phba->phwi_ctrlr->wrb_context);
|
2009-09-05 09:06:35 +07:00
|
|
|
kfree(phba->phwi_ctrlr);
|
|
|
|
}
|
|
|
|
|
|
|
|
static int beiscsi_init_controller(struct beiscsi_hba *phba)
|
|
|
|
{
|
|
|
|
int ret = -ENOMEM;
|
|
|
|
|
|
|
|
ret = beiscsi_get_memory(phba);
|
|
|
|
if (ret < 0) {
|
2012-08-21 00:30:18 +07:00
|
|
|
beiscsi_log(phba, KERN_ERR, BEISCSI_LOG_INIT,
|
|
|
|
"BM_%d : beiscsi_dev_probe -"
|
|
|
|
"Failed in beiscsi_alloc_memory\n");
|
2009-09-05 09:06:35 +07:00
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
|
|
|
ret = hwi_init_controller(phba);
|
|
|
|
if (ret)
|
|
|
|
goto free_init;
|
2012-08-21 00:30:18 +07:00
|
|
|
beiscsi_log(phba, KERN_INFO, BEISCSI_LOG_INIT,
|
|
|
|
"BM_%d : Return success from beiscsi_init_controller");
|
|
|
|
|
2009-09-05 09:06:35 +07:00
|
|
|
return 0;
|
|
|
|
|
|
|
|
free_init:
|
|
|
|
beiscsi_free_mem(phba);
|
2012-04-04 11:41:44 +07:00
|
|
|
return ret;
|
2009-09-05 09:06:35 +07:00
|
|
|
}
|
|
|
|
|
|
|
|
static int beiscsi_init_sgl_handle(struct beiscsi_hba *phba)
|
|
|
|
{
|
|
|
|
struct be_mem_descriptor *mem_descr_sglh, *mem_descr_sg;
|
|
|
|
struct sgl_handle *psgl_handle;
|
|
|
|
struct iscsi_sge *pfrag;
|
2013-09-29 05:35:47 +07:00
|
|
|
unsigned int arr_index, i, idx;
|
|
|
|
unsigned int ulp_icd_start, ulp_num = 0;
|
2009-09-05 09:06:35 +07:00
|
|
|
|
|
|
|
phba->io_sgl_hndl_avbl = 0;
|
|
|
|
phba->eh_sgl_hndl_avbl = 0;
|
2009-10-23 13:22:33 +07:00
|
|
|
|
2009-09-05 09:06:35 +07:00
|
|
|
mem_descr_sglh = phba->init_mem;
|
|
|
|
mem_descr_sglh += HWI_MEM_SGLH;
|
|
|
|
if (1 == mem_descr_sglh->num_elements) {
|
|
|
|
phba->io_sgl_hndl_base = kzalloc(sizeof(struct sgl_handle *) *
|
|
|
|
phba->params.ios_per_ctrl,
|
|
|
|
GFP_KERNEL);
|
|
|
|
if (!phba->io_sgl_hndl_base) {
|
2012-08-21 00:30:18 +07:00
|
|
|
beiscsi_log(phba, KERN_ERR, BEISCSI_LOG_INIT,
|
|
|
|
"BM_%d : Mem Alloc Failed. Failing to load\n");
|
2009-09-05 09:06:35 +07:00
|
|
|
return -ENOMEM;
|
|
|
|
}
|
|
|
|
phba->eh_sgl_hndl_base = kzalloc(sizeof(struct sgl_handle *) *
|
|
|
|
(phba->params.icds_per_ctrl -
|
|
|
|
phba->params.ios_per_ctrl),
|
|
|
|
GFP_KERNEL);
|
|
|
|
if (!phba->eh_sgl_hndl_base) {
|
|
|
|
kfree(phba->io_sgl_hndl_base);
|
2012-08-21 00:30:18 +07:00
|
|
|
beiscsi_log(phba, KERN_ERR, BEISCSI_LOG_INIT,
|
|
|
|
"BM_%d : Mem Alloc Failed. Failing to load\n");
|
2009-09-05 09:06:35 +07:00
|
|
|
return -ENOMEM;
|
|
|
|
}
|
|
|
|
} else {
|
2012-08-21 00:30:18 +07:00
|
|
|
beiscsi_log(phba, KERN_ERR, BEISCSI_LOG_INIT,
|
|
|
|
"BM_%d : HWI_MEM_SGLH is more than one element."
|
|
|
|
"Failing to load\n");
|
2009-09-05 09:06:35 +07:00
|
|
|
return -ENOMEM;
|
|
|
|
}
|
|
|
|
|
|
|
|
arr_index = 0;
|
|
|
|
idx = 0;
|
|
|
|
while (idx < mem_descr_sglh->num_elements) {
|
|
|
|
psgl_handle = mem_descr_sglh->mem_array[idx].virtual_address;
|
|
|
|
|
|
|
|
for (i = 0; i < (mem_descr_sglh->mem_array[idx].size /
|
|
|
|
sizeof(struct sgl_handle)); i++) {
|
|
|
|
if (arr_index < phba->params.ios_per_ctrl) {
|
|
|
|
phba->io_sgl_hndl_base[arr_index] = psgl_handle;
|
|
|
|
phba->io_sgl_hndl_avbl++;
|
|
|
|
arr_index++;
|
|
|
|
} else {
|
|
|
|
phba->eh_sgl_hndl_base[arr_index -
|
|
|
|
phba->params.ios_per_ctrl] =
|
|
|
|
psgl_handle;
|
|
|
|
arr_index++;
|
|
|
|
phba->eh_sgl_hndl_avbl++;
|
|
|
|
}
|
|
|
|
psgl_handle++;
|
|
|
|
}
|
|
|
|
idx++;
|
|
|
|
}
|
2012-08-21 00:30:18 +07:00
|
|
|
beiscsi_log(phba, KERN_INFO, BEISCSI_LOG_INIT,
|
|
|
|
"BM_%d : phba->io_sgl_hndl_avbl=%d"
|
|
|
|
"phba->eh_sgl_hndl_avbl=%d\n",
|
|
|
|
phba->io_sgl_hndl_avbl,
|
|
|
|
phba->eh_sgl_hndl_avbl);
|
|
|
|
|
2009-09-05 09:06:35 +07:00
|
|
|
mem_descr_sg = phba->init_mem;
|
|
|
|
mem_descr_sg += HWI_MEM_SGE;
|
2012-08-21 00:30:18 +07:00
|
|
|
beiscsi_log(phba, KERN_INFO, BEISCSI_LOG_INIT,
|
|
|
|
"\n BM_%d : mem_descr_sg->num_elements=%d\n",
|
|
|
|
mem_descr_sg->num_elements);
|
|
|
|
|
2013-09-29 05:35:47 +07:00
|
|
|
for (ulp_num = 0; ulp_num < BEISCSI_ULP_COUNT; ulp_num++)
|
|
|
|
if (test_bit(ulp_num, &phba->fw_config.ulp_supported))
|
|
|
|
break;
|
|
|
|
|
|
|
|
ulp_icd_start = phba->fw_config.iscsi_icd_start[ulp_num];
|
|
|
|
|
2009-09-05 09:06:35 +07:00
|
|
|
arr_index = 0;
|
|
|
|
idx = 0;
|
|
|
|
while (idx < mem_descr_sg->num_elements) {
|
|
|
|
pfrag = mem_descr_sg->mem_array[idx].virtual_address;
|
|
|
|
|
|
|
|
for (i = 0;
|
|
|
|
i < (mem_descr_sg->mem_array[idx].size) /
|
|
|
|
(sizeof(struct iscsi_sge) * phba->params.num_sge_per_io);
|
|
|
|
i++) {
|
|
|
|
if (arr_index < phba->params.ios_per_ctrl)
|
|
|
|
psgl_handle = phba->io_sgl_hndl_base[arr_index];
|
|
|
|
else
|
|
|
|
psgl_handle = phba->eh_sgl_hndl_base[arr_index -
|
|
|
|
phba->params.ios_per_ctrl];
|
|
|
|
psgl_handle->pfrag = pfrag;
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_sge, addr_hi, pfrag, 0);
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_sge, addr_lo, pfrag, 0);
|
|
|
|
pfrag += phba->params.num_sge_per_io;
|
2013-09-29 05:35:47 +07:00
|
|
|
psgl_handle->sgl_index = ulp_icd_start + arr_index++;
|
2009-09-05 09:06:35 +07:00
|
|
|
}
|
|
|
|
idx++;
|
|
|
|
}
|
|
|
|
phba->io_sgl_free_index = 0;
|
|
|
|
phba->io_sgl_alloc_index = 0;
|
|
|
|
phba->eh_sgl_free_index = 0;
|
|
|
|
phba->eh_sgl_alloc_index = 0;
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int hba_setup_cid_tbls(struct beiscsi_hba *phba)
|
|
|
|
{
|
2013-09-29 05:35:49 +07:00
|
|
|
int ret;
|
|
|
|
uint16_t i, ulp_num;
|
|
|
|
struct ulp_cid_info *ptr_cid_info = NULL;
|
2009-09-05 09:06:35 +07:00
|
|
|
|
2013-09-29 05:35:49 +07:00
|
|
|
for (ulp_num = 0; ulp_num < BEISCSI_ULP_COUNT; ulp_num++) {
|
|
|
|
if (test_bit(ulp_num, (void *)&phba->fw_config.ulp_supported)) {
|
|
|
|
ptr_cid_info = kzalloc(sizeof(struct ulp_cid_info),
|
|
|
|
GFP_KERNEL);
|
|
|
|
|
|
|
|
if (!ptr_cid_info) {
|
|
|
|
beiscsi_log(phba, KERN_ERR, BEISCSI_LOG_INIT,
|
|
|
|
"BM_%d : Failed to allocate memory"
|
|
|
|
"for ULP_CID_INFO for ULP : %d\n",
|
|
|
|
ulp_num);
|
|
|
|
ret = -ENOMEM;
|
|
|
|
goto free_memory;
|
|
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Allocate memory for CID array */
|
|
|
|
ptr_cid_info->cid_array = kzalloc(sizeof(void *) *
|
|
|
|
BEISCSI_GET_CID_COUNT(phba,
|
|
|
|
ulp_num), GFP_KERNEL);
|
|
|
|
if (!ptr_cid_info->cid_array) {
|
|
|
|
beiscsi_log(phba, KERN_ERR, BEISCSI_LOG_INIT,
|
|
|
|
"BM_%d : Failed to allocate memory"
|
|
|
|
"for CID_ARRAY for ULP : %d\n",
|
|
|
|
ulp_num);
|
|
|
|
kfree(ptr_cid_info);
|
|
|
|
ptr_cid_info = NULL;
|
|
|
|
ret = -ENOMEM;
|
|
|
|
|
|
|
|
goto free_memory;
|
|
|
|
}
|
|
|
|
ptr_cid_info->avlbl_cids = BEISCSI_GET_CID_COUNT(
|
|
|
|
phba, ulp_num);
|
|
|
|
|
|
|
|
/* Save the cid_info_array ptr */
|
|
|
|
phba->cid_array_info[ulp_num] = ptr_cid_info;
|
|
|
|
}
|
2009-09-05 09:06:35 +07:00
|
|
|
}
|
2010-01-05 06:35:34 +07:00
|
|
|
phba->ep_array = kzalloc(sizeof(struct iscsi_endpoint *) *
|
2013-04-06 10:38:32 +07:00
|
|
|
phba->params.cxns_per_ctrl, GFP_KERNEL);
|
2009-09-05 09:06:35 +07:00
|
|
|
if (!phba->ep_array) {
|
2012-08-21 00:30:18 +07:00
|
|
|
beiscsi_log(phba, KERN_ERR, BEISCSI_LOG_INIT,
|
|
|
|
"BM_%d : Failed to allocate memory in "
|
|
|
|
"hba_setup_cid_tbls\n");
|
2013-09-29 05:35:49 +07:00
|
|
|
ret = -ENOMEM;
|
|
|
|
|
|
|
|
goto free_memory;
|
2009-09-05 09:06:35 +07:00
|
|
|
}
|
2013-04-06 10:38:32 +07:00
|
|
|
|
|
|
|
phba->conn_table = kzalloc(sizeof(struct beiscsi_conn *) *
|
|
|
|
phba->params.cxns_per_ctrl, GFP_KERNEL);
|
|
|
|
if (!phba->conn_table) {
|
|
|
|
beiscsi_log(phba, KERN_ERR, BEISCSI_LOG_INIT,
|
|
|
|
"BM_%d : Failed to allocate memory in"
|
|
|
|
"hba_setup_cid_tbls\n");
|
|
|
|
|
|
|
|
kfree(phba->ep_array);
|
|
|
|
phba->ep_array = NULL;
|
2013-09-29 05:35:49 +07:00
|
|
|
ret = -ENOMEM;
|
2014-06-06 19:06:30 +07:00
|
|
|
|
|
|
|
goto free_memory;
|
2009-09-05 09:06:35 +07:00
|
|
|
}
|
2013-04-06 10:38:32 +07:00
|
|
|
|
2013-09-29 05:35:49 +07:00
|
|
|
for (i = 0; i < phba->params.cxns_per_ctrl; i++) {
|
|
|
|
ulp_num = phba->phwi_ctrlr->wrb_context[i].ulp_num;
|
|
|
|
|
|
|
|
ptr_cid_info = phba->cid_array_info[ulp_num];
|
|
|
|
ptr_cid_info->cid_array[ptr_cid_info->cid_alloc++] =
|
|
|
|
phba->phwi_ctrlr->wrb_context[i].cid;
|
|
|
|
|
|
|
|
}
|
|
|
|
|
|
|
|
for (ulp_num = 0; ulp_num < BEISCSI_ULP_COUNT; ulp_num++) {
|
|
|
|
if (test_bit(ulp_num, (void *)&phba->fw_config.ulp_supported)) {
|
|
|
|
ptr_cid_info = phba->cid_array_info[ulp_num];
|
2013-04-06 10:38:32 +07:00
|
|
|
|
2013-09-29 05:35:49 +07:00
|
|
|
ptr_cid_info->cid_alloc = 0;
|
|
|
|
ptr_cid_info->cid_free = 0;
|
|
|
|
}
|
|
|
|
}
|
2009-09-05 09:06:35 +07:00
|
|
|
return 0;
|
2013-09-29 05:35:49 +07:00
|
|
|
|
|
|
|
free_memory:
|
|
|
|
for (ulp_num = 0; ulp_num < BEISCSI_ULP_COUNT; ulp_num++) {
|
|
|
|
if (test_bit(ulp_num, (void *)&phba->fw_config.ulp_supported)) {
|
|
|
|
ptr_cid_info = phba->cid_array_info[ulp_num];
|
|
|
|
|
|
|
|
if (ptr_cid_info) {
|
|
|
|
kfree(ptr_cid_info->cid_array);
|
|
|
|
kfree(ptr_cid_info);
|
|
|
|
phba->cid_array_info[ulp_num] = NULL;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
return ret;
|
2009-09-05 09:06:35 +07:00
|
|
|
}
|
|
|
|
|
2010-07-22 05:53:22 +07:00
|
|
|
static void hwi_enable_intr(struct beiscsi_hba *phba)
|
2009-09-05 09:06:35 +07:00
|
|
|
{
|
|
|
|
struct be_ctrl_info *ctrl = &phba->ctrl;
|
|
|
|
struct hwi_controller *phwi_ctrlr;
|
|
|
|
struct hwi_context_memory *phwi_context;
|
|
|
|
struct be_queue_info *eq;
|
|
|
|
u8 __iomem *addr;
|
2009-10-23 13:22:33 +07:00
|
|
|
u32 reg, i;
|
2009-09-05 09:06:35 +07:00
|
|
|
u32 enabled;
|
|
|
|
|
|
|
|
phwi_ctrlr = phba->phwi_ctrlr;
|
|
|
|
phwi_context = phwi_ctrlr->phwi_ctxt;
|
|
|
|
|
|
|
|
addr = (u8 __iomem *) ((u8 __iomem *) ctrl->pcicfg +
|
|
|
|
PCICFG_MEMBAR_CTRL_INT_CTRL_OFFSET);
|
|
|
|
reg = ioread32(addr);
|
|
|
|
|
|
|
|
enabled = reg & MEMBAR_CTRL_INT_CTRL_HOSTINTR_MASK;
|
|
|
|
if (!enabled) {
|
|
|
|
reg |= MEMBAR_CTRL_INT_CTRL_HOSTINTR_MASK;
|
2012-08-21 00:30:18 +07:00
|
|
|
beiscsi_log(phba, KERN_INFO, BEISCSI_LOG_INIT,
|
|
|
|
"BM_%d : reg =x%08x addr=%p\n", reg, addr);
|
2009-09-05 09:06:35 +07:00
|
|
|
iowrite32(reg, addr);
|
2011-04-30 02:30:06 +07:00
|
|
|
}
|
|
|
|
|
|
|
|
if (!phba->msix_enabled) {
|
|
|
|
eq = &phwi_context->be_eq[0].q;
|
2012-08-21 00:30:18 +07:00
|
|
|
beiscsi_log(phba, KERN_INFO, BEISCSI_LOG_INIT,
|
|
|
|
"BM_%d : eq->id=%d\n", eq->id);
|
|
|
|
|
2011-04-30 02:30:06 +07:00
|
|
|
hwi_ring_eq_db(phba, eq->id, 0, 0, 1, 1);
|
|
|
|
} else {
|
|
|
|
for (i = 0; i <= phba->num_cpus; i++) {
|
|
|
|
eq = &phwi_context->be_eq[i].q;
|
2012-08-21 00:30:18 +07:00
|
|
|
beiscsi_log(phba, KERN_INFO, BEISCSI_LOG_INIT,
|
|
|
|
"BM_%d : eq->id=%d\n", eq->id);
|
2009-10-23 13:22:33 +07:00
|
|
|
hwi_ring_eq_db(phba, eq->id, 0, 0, 1, 1);
|
|
|
|
}
|
2010-02-20 09:35:43 +07:00
|
|
|
}
|
2009-09-05 09:06:35 +07:00
|
|
|
}
|
|
|
|
|
|
|
|
static void hwi_disable_intr(struct beiscsi_hba *phba)
|
|
|
|
{
|
|
|
|
struct be_ctrl_info *ctrl = &phba->ctrl;
|
|
|
|
|
|
|
|
u8 __iomem *addr = ctrl->pcicfg + PCICFG_MEMBAR_CTRL_INT_CTRL_OFFSET;
|
|
|
|
u32 reg = ioread32(addr);
|
|
|
|
|
|
|
|
u32 enabled = reg & MEMBAR_CTRL_INT_CTRL_HOSTINTR_MASK;
|
|
|
|
if (enabled) {
|
|
|
|
reg &= ~MEMBAR_CTRL_INT_CTRL_HOSTINTR_MASK;
|
|
|
|
iowrite32(reg, addr);
|
|
|
|
} else
|
2012-08-21 00:30:18 +07:00
|
|
|
beiscsi_log(phba, KERN_WARNING, BEISCSI_LOG_INIT,
|
|
|
|
"BM_%d : In hwi_disable_intr, Already Disabled\n");
|
2009-09-05 09:06:35 +07:00
|
|
|
}
|
|
|
|
|
|
|
|
static int beiscsi_init_port(struct beiscsi_hba *phba)
|
|
|
|
{
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
ret = beiscsi_init_controller(phba);
|
|
|
|
if (ret < 0) {
|
2012-08-21 00:30:18 +07:00
|
|
|
beiscsi_log(phba, KERN_ERR, BEISCSI_LOG_INIT,
|
|
|
|
"BM_%d : beiscsi_dev_probe - Failed in"
|
|
|
|
"beiscsi_init_controller\n");
|
2009-09-05 09:06:35 +07:00
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
ret = beiscsi_init_sgl_handle(phba);
|
|
|
|
if (ret < 0) {
|
2012-08-21 00:30:18 +07:00
|
|
|
beiscsi_log(phba, KERN_ERR, BEISCSI_LOG_INIT,
|
|
|
|
"BM_%d : beiscsi_dev_probe - Failed in"
|
|
|
|
"beiscsi_init_sgl_handle\n");
|
2009-09-05 09:06:35 +07:00
|
|
|
goto do_cleanup_ctrlr;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (hba_setup_cid_tbls(phba)) {
|
2012-08-21 00:30:18 +07:00
|
|
|
beiscsi_log(phba, KERN_ERR, BEISCSI_LOG_INIT,
|
|
|
|
"BM_%d : Failed in hba_setup_cid_tbls\n");
|
2009-09-05 09:06:35 +07:00
|
|
|
kfree(phba->io_sgl_hndl_base);
|
|
|
|
kfree(phba->eh_sgl_hndl_base);
|
|
|
|
goto do_cleanup_ctrlr;
|
|
|
|
}
|
|
|
|
|
|
|
|
return ret;
|
|
|
|
|
|
|
|
do_cleanup_ctrlr:
|
|
|
|
hwi_cleanup(phba);
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void hwi_purge_eq(struct beiscsi_hba *phba)
|
|
|
|
{
|
|
|
|
struct hwi_controller *phwi_ctrlr;
|
|
|
|
struct hwi_context_memory *phwi_context;
|
|
|
|
struct be_queue_info *eq;
|
|
|
|
struct be_eq_entry *eqe = NULL;
|
2009-10-23 13:22:33 +07:00
|
|
|
int i, eq_msix;
|
2010-01-05 06:40:46 +07:00
|
|
|
unsigned int num_processed;
|
2009-09-05 09:06:35 +07:00
|
|
|
|
|
|
|
phwi_ctrlr = phba->phwi_ctrlr;
|
|
|
|
phwi_context = phwi_ctrlr->phwi_ctxt;
|
2009-10-23 13:22:33 +07:00
|
|
|
if (phba->msix_enabled)
|
|
|
|
eq_msix = 1;
|
|
|
|
else
|
|
|
|
eq_msix = 0;
|
2009-09-05 09:06:35 +07:00
|
|
|
|
2009-10-23 13:22:33 +07:00
|
|
|
for (i = 0; i < (phba->num_cpus + eq_msix); i++) {
|
|
|
|
eq = &phwi_context->be_eq[i].q;
|
2009-09-05 09:06:35 +07:00
|
|
|
eqe = queue_tail_node(eq);
|
2010-01-05 06:40:46 +07:00
|
|
|
num_processed = 0;
|
2009-10-23 13:22:33 +07:00
|
|
|
while (eqe->dw[offsetof(struct amap_eq_entry, valid) / 32]
|
|
|
|
& EQE_VALID_MASK) {
|
|
|
|
AMAP_SET_BITS(struct amap_eq_entry, valid, eqe, 0);
|
|
|
|
queue_tail_inc(eq);
|
|
|
|
eqe = queue_tail_node(eq);
|
2010-01-05 06:40:46 +07:00
|
|
|
num_processed++;
|
2009-10-23 13:22:33 +07:00
|
|
|
}
|
2010-01-05 06:40:46 +07:00
|
|
|
|
|
|
|
if (num_processed)
|
|
|
|
hwi_ring_eq_db(phba, eq->id, 1, num_processed, 1, 1);
|
2009-09-05 09:06:35 +07:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
|
|
|
static void beiscsi_clean_port(struct beiscsi_hba *phba)
|
|
|
|
{
|
2013-09-29 05:35:49 +07:00
|
|
|
int mgmt_status, ulp_num;
|
|
|
|
struct ulp_cid_info *ptr_cid_info = NULL;
|
2009-09-05 09:06:35 +07:00
|
|
|
|
2013-09-29 05:35:51 +07:00
|
|
|
for (ulp_num = 0; ulp_num < BEISCSI_ULP_COUNT; ulp_num++) {
|
|
|
|
if (test_bit(ulp_num, (void *)&phba->fw_config.ulp_supported)) {
|
|
|
|
mgmt_status = mgmt_epfw_cleanup(phba, ulp_num);
|
|
|
|
if (mgmt_status)
|
|
|
|
beiscsi_log(phba, KERN_WARNING,
|
|
|
|
BEISCSI_LOG_INIT,
|
|
|
|
"BM_%d : mgmt_epfw_cleanup FAILED"
|
|
|
|
" for ULP_%d\n", ulp_num);
|
|
|
|
}
|
|
|
|
}
|
2010-01-05 06:40:46 +07:00
|
|
|
|
2009-09-05 09:06:35 +07:00
|
|
|
hwi_purge_eq(phba);
|
2010-01-05 06:40:46 +07:00
|
|
|
hwi_cleanup(phba);
|
2009-09-05 09:06:35 +07:00
|
|
|
kfree(phba->io_sgl_hndl_base);
|
|
|
|
kfree(phba->eh_sgl_hndl_base);
|
|
|
|
kfree(phba->ep_array);
|
2013-04-06 10:38:32 +07:00
|
|
|
kfree(phba->conn_table);
|
2013-09-29 05:35:49 +07:00
|
|
|
|
|
|
|
for (ulp_num = 0; ulp_num < BEISCSI_ULP_COUNT; ulp_num++) {
|
|
|
|
if (test_bit(ulp_num, (void *)&phba->fw_config.ulp_supported)) {
|
|
|
|
ptr_cid_info = phba->cid_array_info[ulp_num];
|
|
|
|
|
|
|
|
if (ptr_cid_info) {
|
|
|
|
kfree(ptr_cid_info->cid_array);
|
|
|
|
kfree(ptr_cid_info);
|
|
|
|
phba->cid_array_info[ulp_num] = NULL;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2009-09-05 09:06:35 +07:00
|
|
|
}
|
|
|
|
|
2013-04-06 10:38:25 +07:00
|
|
|
/**
|
|
|
|
* beiscsi_free_mgmt_task_handles()- Free driver CXN resources
|
|
|
|
* @beiscsi_conn: ptr to the conn to be cleaned up
|
2013-04-06 10:38:31 +07:00
|
|
|
* @task: ptr to iscsi_task resource to be freed.
|
2013-04-06 10:38:25 +07:00
|
|
|
*
|
|
|
|
* Free driver mgmt resources binded to CXN.
|
|
|
|
**/
|
|
|
|
void
|
2013-04-06 10:38:31 +07:00
|
|
|
beiscsi_free_mgmt_task_handles(struct beiscsi_conn *beiscsi_conn,
|
|
|
|
struct iscsi_task *task)
|
2013-04-06 10:38:25 +07:00
|
|
|
{
|
|
|
|
struct beiscsi_io_task *io_task;
|
|
|
|
struct beiscsi_hba *phba = beiscsi_conn->phba;
|
|
|
|
struct hwi_wrb_context *pwrb_context;
|
|
|
|
struct hwi_controller *phwi_ctrlr;
|
2013-04-06 10:38:32 +07:00
|
|
|
uint16_t cri_index = BE_GET_CRI_FROM_CID(
|
|
|
|
beiscsi_conn->beiscsi_conn_cid);
|
2013-04-06 10:38:25 +07:00
|
|
|
|
|
|
|
phwi_ctrlr = phba->phwi_ctrlr;
|
2013-04-06 10:38:32 +07:00
|
|
|
pwrb_context = &phwi_ctrlr->wrb_context[cri_index];
|
|
|
|
|
2013-04-06 10:38:31 +07:00
|
|
|
io_task = task->dd_data;
|
2013-04-06 10:38:25 +07:00
|
|
|
|
|
|
|
if (io_task->pwrb_handle) {
|
|
|
|
memset(io_task->pwrb_handle->pwrb, 0,
|
|
|
|
sizeof(struct iscsi_wrb));
|
|
|
|
free_wrb_handle(phba, pwrb_context,
|
|
|
|
io_task->pwrb_handle);
|
|
|
|
io_task->pwrb_handle = NULL;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (io_task->psgl_handle) {
|
|
|
|
free_mgmt_sgl_handle(phba,
|
|
|
|
io_task->psgl_handle);
|
|
|
|
io_task->psgl_handle = NULL;
|
|
|
|
}
|
|
|
|
|
2015-04-25 09:47:45 +07:00
|
|
|
if (io_task->mtask_addr) {
|
2013-04-06 10:38:25 +07:00
|
|
|
pci_unmap_single(phba->pcidev,
|
|
|
|
io_task->mtask_addr,
|
|
|
|
io_task->mtask_data_count,
|
|
|
|
PCI_DMA_TODEVICE);
|
2015-04-25 09:47:45 +07:00
|
|
|
io_task->mtask_addr = 0;
|
|
|
|
}
|
2013-04-06 10:38:25 +07:00
|
|
|
}
|
|
|
|
|
2012-10-20 06:12:00 +07:00
|
|
|
/**
|
|
|
|
* beiscsi_cleanup_task()- Free driver resources of the task
|
|
|
|
* @task: ptr to the iscsi task
|
|
|
|
*
|
|
|
|
**/
|
2012-04-18 15:06:00 +07:00
|
|
|
static void beiscsi_cleanup_task(struct iscsi_task *task)
|
|
|
|
{
|
|
|
|
struct beiscsi_io_task *io_task = task->dd_data;
|
|
|
|
struct iscsi_conn *conn = task->conn;
|
|
|
|
struct beiscsi_conn *beiscsi_conn = conn->dd_data;
|
|
|
|
struct beiscsi_hba *phba = beiscsi_conn->phba;
|
|
|
|
struct beiscsi_session *beiscsi_sess = beiscsi_conn->beiscsi_sess;
|
|
|
|
struct hwi_wrb_context *pwrb_context;
|
|
|
|
struct hwi_controller *phwi_ctrlr;
|
2013-04-06 10:38:32 +07:00
|
|
|
uint16_t cri_index = BE_GET_CRI_FROM_CID(
|
|
|
|
beiscsi_conn->beiscsi_conn_cid);
|
2012-04-18 15:06:00 +07:00
|
|
|
|
|
|
|
phwi_ctrlr = phba->phwi_ctrlr;
|
2013-04-06 10:38:32 +07:00
|
|
|
pwrb_context = &phwi_ctrlr->wrb_context[cri_index];
|
2012-04-18 15:06:00 +07:00
|
|
|
|
|
|
|
if (io_task->cmd_bhs) {
|
|
|
|
pci_pool_free(beiscsi_sess->bhs_pool, io_task->cmd_bhs,
|
|
|
|
io_task->bhs_pa.u.a64.address);
|
|
|
|
io_task->cmd_bhs = NULL;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (task->sc) {
|
|
|
|
if (io_task->pwrb_handle) {
|
|
|
|
free_wrb_handle(phba, pwrb_context,
|
|
|
|
io_task->pwrb_handle);
|
|
|
|
io_task->pwrb_handle = NULL;
|
|
|
|
}
|
|
|
|
|
|
|
|
if (io_task->psgl_handle) {
|
|
|
|
free_io_sgl_handle(phba, io_task->psgl_handle);
|
|
|
|
io_task->psgl_handle = NULL;
|
|
|
|
}
|
2014-01-29 14:16:44 +07:00
|
|
|
|
|
|
|
if (io_task->scsi_cmnd) {
|
2016-08-19 16:50:11 +07:00
|
|
|
if (io_task->num_sg)
|
|
|
|
scsi_dma_unmap(io_task->scsi_cmnd);
|
2014-01-29 14:16:44 +07:00
|
|
|
io_task->scsi_cmnd = NULL;
|
|
|
|
}
|
2012-04-18 15:06:00 +07:00
|
|
|
} else {
|
2013-04-06 10:38:25 +07:00
|
|
|
if (!beiscsi_conn->login_in_progress)
|
2013-04-06 10:38:31 +07:00
|
|
|
beiscsi_free_mgmt_task_handles(beiscsi_conn, task);
|
2012-04-18 15:06:00 +07:00
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2009-09-05 09:06:35 +07:00
|
|
|
void
|
|
|
|
beiscsi_offload_connection(struct beiscsi_conn *beiscsi_conn,
|
|
|
|
struct beiscsi_offload_params *params)
|
|
|
|
{
|
|
|
|
struct wrb_handle *pwrb_handle;
|
2015-08-20 06:14:30 +07:00
|
|
|
struct hwi_wrb_context *pwrb_context = NULL;
|
2009-09-05 09:06:35 +07:00
|
|
|
struct beiscsi_hba *phba = beiscsi_conn->phba;
|
2012-04-18 15:06:00 +07:00
|
|
|
struct iscsi_task *task = beiscsi_conn->task;
|
|
|
|
struct iscsi_session *session = task->conn->session;
|
2009-09-05 09:06:35 +07:00
|
|
|
u32 doorbell = 0;
|
|
|
|
|
|
|
|
/*
|
|
|
|
* We can always use 0 here because it is reserved by libiscsi for
|
|
|
|
* login/startup related tasks.
|
|
|
|
*/
|
2012-04-18 15:06:00 +07:00
|
|
|
beiscsi_conn->login_in_progress = 0;
|
[SCSI] libiscsi: Reduce locking contention in fast path
Replace the session lock with two locks, a forward lock and
a backwards lock named frwd_lock and back_lock respectively.
The forward lock protects resources that change while sending a
request to the target, such as cmdsn, queued_cmdsn, and allocating
task from the commands' pool with kfifo_out.
The backward lock protects resources that change while processing
a response or in error path, such as cmdsn_exp, cmdsn_max, and
returning tasks to the commands' pool with kfifo_in.
Under a steady state fast-path situation, that is when one
or more processes/threads submit IO to an iscsi device and
a single kernel upcall (e.g softirq) is dealing with processing
of responses without errors, this patch eliminates the contention
between the queuecommand()/request response/scsi_done() flows
associated with iscsi sessions.
Between the forward and the backward locks exists a strict locking
hierarchy. The mutual exclusion zone protected by the forward lock can
enclose the mutual exclusion zone protected by the backward lock but not
vice versa.
For example, in iscsi_conn_teardown or in iscsi_xmit_data when there is
a failure and __iscsi_put_task is called, the backward lock is taken while
the forward lock is still taken. On the other hand, if in the RX path a nop
is to be sent, for example in iscsi_handle_reject or __iscsi_complete_pdu
than the forward lock is released and the backward lock is taken for the
duration of iscsi_send_nopout, later the backward lock is released and the
forward lock is retaken.
libiscsi_tcp uses two kernel fifos the r2t pool and the r2t queue.
The insertion and deletion from these queues didn't corespond to the
assumption taken by the new forward/backwards session locking paradigm.
That is, in iscsi_tcp_clenup_task which belongs to the RX (backwards)
path, r2t is taken out from r2t queue and inserted to the r2t pool.
In iscsi_tcp_get_curr_r2t which belong to the TX (forward) path, r2t
is also inserted to the r2t pool and another r2t is pulled from r2t
queue.
Only in iscsi_tcp_r2t_rsp which is called in the RX path but can requeue
to the TX path, r2t is taken from the r2t pool and inserted to the r2t
queue.
In order to cope with this situation, two spin locks were added,
pool2queue and queue2pool. The former protects extracting from the
r2t pool and inserting to the r2t queue, and the later protects the
extracing from the r2t queue and inserting to the r2t pool.
Signed-off-by: Shlomo Pongratz <shlomop@mellanox.com>
Signed-off-by: Or Gerlitz <ogerlitz@mellanox.com>
[minor fix up to apply cleanly and compile fix]
Signed-off-by: Mike Christie <michaelc@cs.wisc.edu>
Signed-off-by: James Bottomley <JBottomley@Parallels.com>
2014-02-07 13:41:38 +07:00
|
|
|
spin_lock_bh(&session->back_lock);
|
2012-04-18 15:06:00 +07:00
|
|
|
beiscsi_cleanup_task(task);
|
[SCSI] libiscsi: Reduce locking contention in fast path
Replace the session lock with two locks, a forward lock and
a backwards lock named frwd_lock and back_lock respectively.
The forward lock protects resources that change while sending a
request to the target, such as cmdsn, queued_cmdsn, and allocating
task from the commands' pool with kfifo_out.
The backward lock protects resources that change while processing
a response or in error path, such as cmdsn_exp, cmdsn_max, and
returning tasks to the commands' pool with kfifo_in.
Under a steady state fast-path situation, that is when one
or more processes/threads submit IO to an iscsi device and
a single kernel upcall (e.g softirq) is dealing with processing
of responses without errors, this patch eliminates the contention
between the queuecommand()/request response/scsi_done() flows
associated with iscsi sessions.
Between the forward and the backward locks exists a strict locking
hierarchy. The mutual exclusion zone protected by the forward lock can
enclose the mutual exclusion zone protected by the backward lock but not
vice versa.
For example, in iscsi_conn_teardown or in iscsi_xmit_data when there is
a failure and __iscsi_put_task is called, the backward lock is taken while
the forward lock is still taken. On the other hand, if in the RX path a nop
is to be sent, for example in iscsi_handle_reject or __iscsi_complete_pdu
than the forward lock is released and the backward lock is taken for the
duration of iscsi_send_nopout, later the backward lock is released and the
forward lock is retaken.
libiscsi_tcp uses two kernel fifos the r2t pool and the r2t queue.
The insertion and deletion from these queues didn't corespond to the
assumption taken by the new forward/backwards session locking paradigm.
That is, in iscsi_tcp_clenup_task which belongs to the RX (backwards)
path, r2t is taken out from r2t queue and inserted to the r2t pool.
In iscsi_tcp_get_curr_r2t which belong to the TX (forward) path, r2t
is also inserted to the r2t pool and another r2t is pulled from r2t
queue.
Only in iscsi_tcp_r2t_rsp which is called in the RX path but can requeue
to the TX path, r2t is taken from the r2t pool and inserted to the r2t
queue.
In order to cope with this situation, two spin locks were added,
pool2queue and queue2pool. The former protects extracting from the
r2t pool and inserting to the r2t queue, and the later protects the
extracing from the r2t queue and inserting to the r2t pool.
Signed-off-by: Shlomo Pongratz <shlomop@mellanox.com>
Signed-off-by: Or Gerlitz <ogerlitz@mellanox.com>
[minor fix up to apply cleanly and compile fix]
Signed-off-by: Mike Christie <michaelc@cs.wisc.edu>
Signed-off-by: James Bottomley <JBottomley@Parallels.com>
2014-02-07 13:41:38 +07:00
|
|
|
spin_unlock_bh(&session->back_lock);
|
2012-04-18 15:06:00 +07:00
|
|
|
|
2015-08-20 06:14:30 +07:00
|
|
|
pwrb_handle = alloc_wrb_handle(phba, beiscsi_conn->beiscsi_conn_cid,
|
|
|
|
&pwrb_context);
|
2009-09-05 09:06:35 +07:00
|
|
|
|
2012-10-20 06:14:35 +07:00
|
|
|
/* Check for the adapter family */
|
2013-04-06 10:38:26 +07:00
|
|
|
if (is_chip_be2_be3r(phba))
|
2012-10-20 06:14:35 +07:00
|
|
|
beiscsi_offload_cxn_v0(params, pwrb_handle,
|
2015-08-20 06:14:30 +07:00
|
|
|
phba->init_mem,
|
|
|
|
pwrb_context);
|
2013-04-06 10:38:26 +07:00
|
|
|
else
|
2015-08-20 06:14:30 +07:00
|
|
|
beiscsi_offload_cxn_v2(params, pwrb_handle,
|
|
|
|
pwrb_context);
|
2009-09-05 09:06:35 +07:00
|
|
|
|
2012-10-20 06:14:35 +07:00
|
|
|
be_dws_le_to_cpu(pwrb_handle->pwrb,
|
|
|
|
sizeof(struct iscsi_target_context_update_wrb));
|
2009-09-05 09:06:35 +07:00
|
|
|
|
|
|
|
doorbell |= beiscsi_conn->beiscsi_conn_cid & DB_WRB_POST_CID_MASK;
|
2010-01-23 07:04:24 +07:00
|
|
|
doorbell |= (pwrb_handle->wrb_index & DB_DEF_PDU_WRB_INDEX_MASK)
|
2009-10-23 13:22:33 +07:00
|
|
|
<< DB_DEF_PDU_WRB_INDEX_SHIFT;
|
2009-09-05 09:06:35 +07:00
|
|
|
doorbell |= 1 << DB_DEF_PDU_NUM_POSTED_SHIFT;
|
2013-09-29 05:35:50 +07:00
|
|
|
iowrite32(doorbell, phba->db_va +
|
|
|
|
beiscsi_conn->doorbell_offset);
|
2016-01-20 15:41:00 +07:00
|
|
|
|
|
|
|
/*
|
|
|
|
* There is no completion for CONTEXT_UPDATE. The completion of next
|
|
|
|
* WRB posted guarantees FW's processing and DMA'ing of it.
|
|
|
|
* Use beiscsi_put_wrb_handle to put it back in the pool which makes
|
|
|
|
* sure zero'ing or reuse of the WRB only after wrbs_per_cxn.
|
|
|
|
*/
|
|
|
|
beiscsi_put_wrb_handle(pwrb_context, pwrb_handle,
|
|
|
|
phba->params.wrbs_per_cxn);
|
|
|
|
beiscsi_log(phba, KERN_INFO,
|
|
|
|
BEISCSI_LOG_IO | BEISCSI_LOG_CONFIG,
|
|
|
|
"BM_%d : put CONTEXT_UPDATE pwrb_handle=%p free_index=0x%x wrb_handles_available=%d\n",
|
|
|
|
pwrb_handle, pwrb_context->free_index,
|
|
|
|
pwrb_context->wrb_handles_available);
|
2009-09-05 09:06:35 +07:00
|
|
|
}
|
|
|
|
|
|
|
|
static void beiscsi_parse_pdu(struct iscsi_conn *conn, itt_t itt,
|
|
|
|
int *index, int *age)
|
|
|
|
{
|
2009-10-23 13:22:33 +07:00
|
|
|
*index = (int)itt;
|
2009-09-05 09:06:35 +07:00
|
|
|
if (age)
|
|
|
|
*age = conn->session->age;
|
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
|
|
|
* beiscsi_alloc_pdu - allocates pdu and related resources
|
|
|
|
* @task: libiscsi task
|
|
|
|
* @opcode: opcode of pdu for task
|
|
|
|
*
|
|
|
|
* This is called with the session lock held. It will allocate
|
|
|
|
* the wrb and sgl if needed for the command. And it will prep
|
|
|
|
* the pdu's itt. beiscsi_parse_pdu will later translate
|
|
|
|
* the pdu itt to the libiscsi task itt.
|
|
|
|
*/
|
|
|
|
static int beiscsi_alloc_pdu(struct iscsi_task *task, uint8_t opcode)
|
|
|
|
{
|
|
|
|
struct beiscsi_io_task *io_task = task->dd_data;
|
|
|
|
struct iscsi_conn *conn = task->conn;
|
|
|
|
struct beiscsi_conn *beiscsi_conn = conn->dd_data;
|
|
|
|
struct beiscsi_hba *phba = beiscsi_conn->phba;
|
|
|
|
struct hwi_wrb_context *pwrb_context;
|
|
|
|
struct hwi_controller *phwi_ctrlr;
|
|
|
|
itt_t itt;
|
2013-04-06 10:38:32 +07:00
|
|
|
uint16_t cri_index = 0;
|
2009-09-22 09:52:26 +07:00
|
|
|
struct beiscsi_session *beiscsi_sess = beiscsi_conn->beiscsi_sess;
|
|
|
|
dma_addr_t paddr;
|
2009-09-05 09:06:35 +07:00
|
|
|
|
2009-09-22 09:52:26 +07:00
|
|
|
io_task->cmd_bhs = pci_pool_alloc(beiscsi_sess->bhs_pool,
|
2010-12-31 15:22:19 +07:00
|
|
|
GFP_ATOMIC, &paddr);
|
2009-09-22 09:52:26 +07:00
|
|
|
if (!io_task->cmd_bhs)
|
|
|
|
return -ENOMEM;
|
|
|
|
io_task->bhs_pa.u.a64.address = paddr;
|
2009-10-23 13:22:33 +07:00
|
|
|
io_task->libiscsi_itt = (itt_t)task->itt;
|
2009-09-05 09:06:35 +07:00
|
|
|
io_task->conn = beiscsi_conn;
|
|
|
|
|
|
|
|
task->hdr = (struct iscsi_hdr *)&io_task->cmd_bhs->iscsi_hdr;
|
|
|
|
task->hdr_max = sizeof(struct be_cmd_bhs);
|
2010-07-22 05:55:40 +07:00
|
|
|
io_task->psgl_handle = NULL;
|
2012-04-04 11:41:38 +07:00
|
|
|
io_task->pwrb_handle = NULL;
|
2009-09-05 09:06:35 +07:00
|
|
|
|
|
|
|
if (task->sc) {
|
|
|
|
io_task->psgl_handle = alloc_io_sgl_handle(phba);
|
2012-10-20 06:13:03 +07:00
|
|
|
if (!io_task->psgl_handle) {
|
|
|
|
beiscsi_log(phba, KERN_ERR,
|
|
|
|
BEISCSI_LOG_IO | BEISCSI_LOG_CONFIG,
|
|
|
|
"BM_%d : Alloc of IO_SGL_ICD Failed"
|
|
|
|
"for the CID : %d\n",
|
|
|
|
beiscsi_conn->beiscsi_conn_cid);
|
2009-09-22 09:52:26 +07:00
|
|
|
goto free_hndls;
|
2012-10-20 06:13:03 +07:00
|
|
|
}
|
2010-07-22 05:55:40 +07:00
|
|
|
io_task->pwrb_handle = alloc_wrb_handle(phba,
|
2015-08-20 06:14:30 +07:00
|
|
|
beiscsi_conn->beiscsi_conn_cid,
|
|
|
|
&io_task->pwrb_context);
|
2012-10-20 06:13:03 +07:00
|
|
|
if (!io_task->pwrb_handle) {
|
|
|
|
beiscsi_log(phba, KERN_ERR,
|
|
|
|
BEISCSI_LOG_IO | BEISCSI_LOG_CONFIG,
|
|
|
|
"BM_%d : Alloc of WRB_HANDLE Failed"
|
|
|
|
"for the CID : %d\n",
|
|
|
|
beiscsi_conn->beiscsi_conn_cid);
|
2010-07-22 05:55:40 +07:00
|
|
|
goto free_io_hndls;
|
2012-10-20 06:13:03 +07:00
|
|
|
}
|
2009-09-05 09:06:35 +07:00
|
|
|
} else {
|
|
|
|
io_task->scsi_cmnd = NULL;
|
2010-01-05 06:38:39 +07:00
|
|
|
if ((opcode & ISCSI_OPCODE_MASK) == ISCSI_OP_LOGIN) {
|
2013-04-06 10:38:25 +07:00
|
|
|
beiscsi_conn->task = task;
|
2009-09-05 09:06:35 +07:00
|
|
|
if (!beiscsi_conn->login_in_progress) {
|
|
|
|
io_task->psgl_handle = (struct sgl_handle *)
|
|
|
|
alloc_mgmt_sgl_handle(phba);
|
2012-10-20 06:13:03 +07:00
|
|
|
if (!io_task->psgl_handle) {
|
|
|
|
beiscsi_log(phba, KERN_ERR,
|
|
|
|
BEISCSI_LOG_IO |
|
|
|
|
BEISCSI_LOG_CONFIG,
|
|
|
|
"BM_%d : Alloc of MGMT_SGL_ICD Failed"
|
|
|
|
"for the CID : %d\n",
|
|
|
|
beiscsi_conn->
|
|
|
|
beiscsi_conn_cid);
|
2009-09-22 09:52:26 +07:00
|
|
|
goto free_hndls;
|
2012-10-20 06:13:03 +07:00
|
|
|
}
|
2009-09-22 09:52:26 +07:00
|
|
|
|
2009-09-05 09:06:35 +07:00
|
|
|
beiscsi_conn->login_in_progress = 1;
|
|
|
|
beiscsi_conn->plogin_sgl_handle =
|
|
|
|
io_task->psgl_handle;
|
2010-07-22 05:55:40 +07:00
|
|
|
io_task->pwrb_handle =
|
|
|
|
alloc_wrb_handle(phba,
|
2015-08-20 06:14:30 +07:00
|
|
|
beiscsi_conn->beiscsi_conn_cid,
|
|
|
|
&io_task->pwrb_context);
|
2012-10-20 06:13:03 +07:00
|
|
|
if (!io_task->pwrb_handle) {
|
|
|
|
beiscsi_log(phba, KERN_ERR,
|
|
|
|
BEISCSI_LOG_IO |
|
|
|
|
BEISCSI_LOG_CONFIG,
|
|
|
|
"BM_%d : Alloc of WRB_HANDLE Failed"
|
|
|
|
"for the CID : %d\n",
|
|
|
|
beiscsi_conn->
|
|
|
|
beiscsi_conn_cid);
|
|
|
|
goto free_mgmt_hndls;
|
|
|
|
}
|
2010-07-22 05:55:40 +07:00
|
|
|
beiscsi_conn->plogin_wrb_handle =
|
|
|
|
io_task->pwrb_handle;
|
|
|
|
|
2009-09-05 09:06:35 +07:00
|
|
|
} else {
|
|
|
|
io_task->psgl_handle =
|
|
|
|
beiscsi_conn->plogin_sgl_handle;
|
2010-07-22 05:55:40 +07:00
|
|
|
io_task->pwrb_handle =
|
|
|
|
beiscsi_conn->plogin_wrb_handle;
|
2009-09-05 09:06:35 +07:00
|
|
|
}
|
|
|
|
} else {
|
|
|
|
io_task->psgl_handle = alloc_mgmt_sgl_handle(phba);
|
2012-10-20 06:13:03 +07:00
|
|
|
if (!io_task->psgl_handle) {
|
|
|
|
beiscsi_log(phba, KERN_ERR,
|
|
|
|
BEISCSI_LOG_IO |
|
|
|
|
BEISCSI_LOG_CONFIG,
|
|
|
|
"BM_%d : Alloc of MGMT_SGL_ICD Failed"
|
|
|
|
"for the CID : %d\n",
|
|
|
|
beiscsi_conn->
|
|
|
|
beiscsi_conn_cid);
|
2009-09-22 09:52:26 +07:00
|
|
|
goto free_hndls;
|
2012-10-20 06:13:03 +07:00
|
|
|
}
|
2010-07-22 05:55:40 +07:00
|
|
|
io_task->pwrb_handle =
|
|
|
|
alloc_wrb_handle(phba,
|
2015-08-20 06:14:30 +07:00
|
|
|
beiscsi_conn->beiscsi_conn_cid,
|
|
|
|
&io_task->pwrb_context);
|
2012-10-20 06:13:03 +07:00
|
|
|
if (!io_task->pwrb_handle) {
|
|
|
|
beiscsi_log(phba, KERN_ERR,
|
|
|
|
BEISCSI_LOG_IO | BEISCSI_LOG_CONFIG,
|
|
|
|
"BM_%d : Alloc of WRB_HANDLE Failed"
|
|
|
|
"for the CID : %d\n",
|
|
|
|
beiscsi_conn->beiscsi_conn_cid);
|
2010-07-22 05:55:40 +07:00
|
|
|
goto free_mgmt_hndls;
|
2012-10-20 06:13:03 +07:00
|
|
|
}
|
2010-07-22 05:55:40 +07:00
|
|
|
|
2009-09-05 09:06:35 +07:00
|
|
|
}
|
|
|
|
}
|
2009-10-23 13:22:33 +07:00
|
|
|
itt = (itt_t) cpu_to_be32(((unsigned int)io_task->pwrb_handle->
|
|
|
|
wrb_index << 16) | (unsigned int)
|
|
|
|
(io_task->psgl_handle->sgl_index));
|
2010-01-23 07:04:24 +07:00
|
|
|
io_task->pwrb_handle->pio_handle = task;
|
2009-10-23 13:22:33 +07:00
|
|
|
|
2009-09-05 09:06:35 +07:00
|
|
|
io_task->cmd_bhs->iscsi_hdr.itt = itt;
|
|
|
|
return 0;
|
2009-09-22 09:52:26 +07:00
|
|
|
|
2010-07-22 05:55:40 +07:00
|
|
|
free_io_hndls:
|
|
|
|
free_io_sgl_handle(phba, io_task->psgl_handle);
|
|
|
|
goto free_hndls;
|
|
|
|
free_mgmt_hndls:
|
|
|
|
free_mgmt_sgl_handle(phba, io_task->psgl_handle);
|
2013-04-06 10:38:32 +07:00
|
|
|
io_task->psgl_handle = NULL;
|
2009-09-22 09:52:26 +07:00
|
|
|
free_hndls:
|
|
|
|
phwi_ctrlr = phba->phwi_ctrlr;
|
2013-04-06 10:38:32 +07:00
|
|
|
cri_index = BE_GET_CRI_FROM_CID(
|
|
|
|
beiscsi_conn->beiscsi_conn_cid);
|
|
|
|
pwrb_context = &phwi_ctrlr->wrb_context[cri_index];
|
2010-07-22 05:55:40 +07:00
|
|
|
if (io_task->pwrb_handle)
|
|
|
|
free_wrb_handle(phba, pwrb_context, io_task->pwrb_handle);
|
2009-09-22 09:52:26 +07:00
|
|
|
io_task->pwrb_handle = NULL;
|
|
|
|
pci_pool_free(beiscsi_sess->bhs_pool, io_task->cmd_bhs,
|
|
|
|
io_task->bhs_pa.u.a64.address);
|
2012-04-18 15:06:00 +07:00
|
|
|
io_task->cmd_bhs = NULL;
|
2009-09-22 09:52:26 +07:00
|
|
|
return -ENOMEM;
|
2009-09-05 09:06:35 +07:00
|
|
|
}
|
2012-10-20 06:14:23 +07:00
|
|
|
int beiscsi_iotask_v2(struct iscsi_task *task, struct scatterlist *sg,
|
|
|
|
unsigned int num_sg, unsigned int xferlen,
|
|
|
|
unsigned int writedir)
|
|
|
|
{
|
|
|
|
|
|
|
|
struct beiscsi_io_task *io_task = task->dd_data;
|
|
|
|
struct iscsi_conn *conn = task->conn;
|
|
|
|
struct beiscsi_conn *beiscsi_conn = conn->dd_data;
|
|
|
|
struct beiscsi_hba *phba = beiscsi_conn->phba;
|
|
|
|
struct iscsi_wrb *pwrb = NULL;
|
|
|
|
unsigned int doorbell = 0;
|
|
|
|
|
|
|
|
pwrb = io_task->pwrb_handle->pwrb;
|
|
|
|
|
|
|
|
io_task->bhs_len = sizeof(struct be_cmd_bhs);
|
|
|
|
|
|
|
|
if (writedir) {
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_wrb_v2, type, pwrb,
|
|
|
|
INI_WR_CMD);
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_wrb_v2, dsp, pwrb, 1);
|
|
|
|
} else {
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_wrb_v2, type, pwrb,
|
|
|
|
INI_RD_CMD);
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_wrb_v2, dsp, pwrb, 0);
|
|
|
|
}
|
|
|
|
|
|
|
|
io_task->wrb_type = AMAP_GET_BITS(struct amap_iscsi_wrb_v2,
|
|
|
|
type, pwrb);
|
|
|
|
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_wrb_v2, lun, pwrb,
|
|
|
|
cpu_to_be16(*(unsigned short *)
|
|
|
|
&io_task->cmd_bhs->iscsi_hdr.lun));
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_wrb_v2, r2t_exp_dtl, pwrb, xferlen);
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_wrb_v2, wrb_idx, pwrb,
|
|
|
|
io_task->pwrb_handle->wrb_index);
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_wrb_v2, cmdsn_itt, pwrb,
|
|
|
|
be32_to_cpu(task->cmdsn));
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_wrb_v2, sgl_idx, pwrb,
|
|
|
|
io_task->psgl_handle->sgl_index);
|
|
|
|
|
|
|
|
hwi_write_sgl_v2(pwrb, sg, num_sg, io_task);
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_wrb_v2, ptr2nextwrb, pwrb,
|
2015-08-20 06:14:30 +07:00
|
|
|
io_task->pwrb_handle->wrb_index);
|
|
|
|
if (io_task->pwrb_context->plast_wrb)
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_wrb_v2, ptr2nextwrb,
|
|
|
|
io_task->pwrb_context->plast_wrb,
|
|
|
|
io_task->pwrb_handle->wrb_index);
|
|
|
|
io_task->pwrb_context->plast_wrb = pwrb;
|
2012-10-20 06:14:23 +07:00
|
|
|
|
|
|
|
be_dws_le_to_cpu(pwrb, sizeof(struct iscsi_wrb));
|
|
|
|
|
|
|
|
doorbell |= beiscsi_conn->beiscsi_conn_cid & DB_WRB_POST_CID_MASK;
|
|
|
|
doorbell |= (io_task->pwrb_handle->wrb_index &
|
|
|
|
DB_DEF_PDU_WRB_INDEX_MASK) <<
|
|
|
|
DB_DEF_PDU_WRB_INDEX_SHIFT;
|
|
|
|
doorbell |= 1 << DB_DEF_PDU_NUM_POSTED_SHIFT;
|
2013-09-29 05:35:50 +07:00
|
|
|
iowrite32(doorbell, phba->db_va +
|
|
|
|
beiscsi_conn->doorbell_offset);
|
2012-10-20 06:14:23 +07:00
|
|
|
return 0;
|
|
|
|
}
|
2009-09-05 09:06:35 +07:00
|
|
|
|
|
|
|
static int beiscsi_iotask(struct iscsi_task *task, struct scatterlist *sg,
|
|
|
|
unsigned int num_sg, unsigned int xferlen,
|
|
|
|
unsigned int writedir)
|
|
|
|
{
|
|
|
|
|
|
|
|
struct beiscsi_io_task *io_task = task->dd_data;
|
|
|
|
struct iscsi_conn *conn = task->conn;
|
|
|
|
struct beiscsi_conn *beiscsi_conn = conn->dd_data;
|
|
|
|
struct beiscsi_hba *phba = beiscsi_conn->phba;
|
|
|
|
struct iscsi_wrb *pwrb = NULL;
|
|
|
|
unsigned int doorbell = 0;
|
|
|
|
|
|
|
|
pwrb = io_task->pwrb_handle->pwrb;
|
|
|
|
io_task->bhs_len = sizeof(struct be_cmd_bhs);
|
|
|
|
|
|
|
|
if (writedir) {
|
2010-01-23 07:04:24 +07:00
|
|
|
AMAP_SET_BITS(struct amap_iscsi_wrb, type, pwrb,
|
|
|
|
INI_WR_CMD);
|
2009-09-05 09:06:35 +07:00
|
|
|
AMAP_SET_BITS(struct amap_iscsi_wrb, dsp, pwrb, 1);
|
|
|
|
} else {
|
2010-01-23 07:04:24 +07:00
|
|
|
AMAP_SET_BITS(struct amap_iscsi_wrb, type, pwrb,
|
|
|
|
INI_RD_CMD);
|
2009-09-05 09:06:35 +07:00
|
|
|
AMAP_SET_BITS(struct amap_iscsi_wrb, dsp, pwrb, 0);
|
|
|
|
}
|
|
|
|
|
2012-10-20 06:14:23 +07:00
|
|
|
io_task->wrb_type = AMAP_GET_BITS(struct amap_iscsi_wrb,
|
|
|
|
type, pwrb);
|
|
|
|
|
2009-09-05 09:06:35 +07:00
|
|
|
AMAP_SET_BITS(struct amap_iscsi_wrb, lun, pwrb,
|
2012-04-04 11:41:36 +07:00
|
|
|
cpu_to_be16(*(unsigned short *)
|
|
|
|
&io_task->cmd_bhs->iscsi_hdr.lun));
|
2009-09-05 09:06:35 +07:00
|
|
|
AMAP_SET_BITS(struct amap_iscsi_wrb, r2t_exp_dtl, pwrb, xferlen);
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_wrb, wrb_idx, pwrb,
|
|
|
|
io_task->pwrb_handle->wrb_index);
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_wrb, cmdsn_itt, pwrb,
|
|
|
|
be32_to_cpu(task->cmdsn));
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_wrb, sgl_icd_idx, pwrb,
|
|
|
|
io_task->psgl_handle->sgl_index);
|
|
|
|
|
|
|
|
hwi_write_sgl(pwrb, sg, num_sg, io_task);
|
|
|
|
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_wrb, ptr2nextwrb, pwrb,
|
2015-08-20 06:14:30 +07:00
|
|
|
io_task->pwrb_handle->wrb_index);
|
|
|
|
if (io_task->pwrb_context->plast_wrb)
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_wrb, ptr2nextwrb,
|
|
|
|
io_task->pwrb_context->plast_wrb,
|
|
|
|
io_task->pwrb_handle->wrb_index);
|
|
|
|
io_task->pwrb_context->plast_wrb = pwrb;
|
|
|
|
|
2009-09-05 09:06:35 +07:00
|
|
|
be_dws_le_to_cpu(pwrb, sizeof(struct iscsi_wrb));
|
|
|
|
|
|
|
|
doorbell |= beiscsi_conn->beiscsi_conn_cid & DB_WRB_POST_CID_MASK;
|
2010-01-23 07:04:24 +07:00
|
|
|
doorbell |= (io_task->pwrb_handle->wrb_index &
|
2009-09-05 09:06:35 +07:00
|
|
|
DB_DEF_PDU_WRB_INDEX_MASK) << DB_DEF_PDU_WRB_INDEX_SHIFT;
|
|
|
|
doorbell |= 1 << DB_DEF_PDU_NUM_POSTED_SHIFT;
|
|
|
|
|
2013-09-29 05:35:50 +07:00
|
|
|
iowrite32(doorbell, phba->db_va +
|
|
|
|
beiscsi_conn->doorbell_offset);
|
2009-09-05 09:06:35 +07:00
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int beiscsi_mtask(struct iscsi_task *task)
|
|
|
|
{
|
2010-02-20 09:33:56 +07:00
|
|
|
struct beiscsi_io_task *io_task = task->dd_data;
|
2009-09-05 09:06:35 +07:00
|
|
|
struct iscsi_conn *conn = task->conn;
|
|
|
|
struct beiscsi_conn *beiscsi_conn = conn->dd_data;
|
|
|
|
struct beiscsi_hba *phba = beiscsi_conn->phba;
|
|
|
|
struct iscsi_wrb *pwrb = NULL;
|
|
|
|
unsigned int doorbell = 0;
|
2010-02-20 09:33:56 +07:00
|
|
|
unsigned int cid;
|
2012-10-20 06:14:23 +07:00
|
|
|
unsigned int pwrb_typeoffset = 0;
|
2016-01-16 04:45:34 +07:00
|
|
|
int ret = 0;
|
2009-09-05 09:06:35 +07:00
|
|
|
|
2009-10-23 13:22:33 +07:00
|
|
|
cid = beiscsi_conn->beiscsi_conn_cid;
|
2009-09-05 09:06:35 +07:00
|
|
|
pwrb = io_task->pwrb_handle->pwrb;
|
2010-01-23 07:08:18 +07:00
|
|
|
memset(pwrb, 0, sizeof(*pwrb));
|
2012-10-20 06:14:23 +07:00
|
|
|
|
2013-04-06 10:38:26 +07:00
|
|
|
if (is_chip_be2_be3r(phba)) {
|
2012-10-20 06:14:23 +07:00
|
|
|
AMAP_SET_BITS(struct amap_iscsi_wrb, cmdsn_itt, pwrb,
|
|
|
|
be32_to_cpu(task->cmdsn));
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_wrb, wrb_idx, pwrb,
|
|
|
|
io_task->pwrb_handle->wrb_index);
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_wrb, sgl_icd_idx, pwrb,
|
|
|
|
io_task->psgl_handle->sgl_index);
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_wrb, r2t_exp_dtl, pwrb,
|
|
|
|
task->data_count);
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_wrb, ptr2nextwrb, pwrb,
|
2015-08-20 06:14:30 +07:00
|
|
|
io_task->pwrb_handle->wrb_index);
|
|
|
|
if (io_task->pwrb_context->plast_wrb)
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_wrb, ptr2nextwrb,
|
|
|
|
io_task->pwrb_context->plast_wrb,
|
|
|
|
io_task->pwrb_handle->wrb_index);
|
|
|
|
io_task->pwrb_context->plast_wrb = pwrb;
|
|
|
|
|
2012-10-20 06:14:23 +07:00
|
|
|
pwrb_typeoffset = BE_WRB_TYPE_OFFSET;
|
2013-04-06 10:38:26 +07:00
|
|
|
} else {
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_wrb_v2, cmdsn_itt, pwrb,
|
|
|
|
be32_to_cpu(task->cmdsn));
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_wrb_v2, wrb_idx, pwrb,
|
|
|
|
io_task->pwrb_handle->wrb_index);
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_wrb_v2, sgl_idx, pwrb,
|
|
|
|
io_task->psgl_handle->sgl_index);
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_wrb_v2, r2t_exp_dtl, pwrb,
|
|
|
|
task->data_count);
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_wrb_v2, ptr2nextwrb, pwrb,
|
2015-08-20 06:14:30 +07:00
|
|
|
io_task->pwrb_handle->wrb_index);
|
|
|
|
if (io_task->pwrb_context->plast_wrb)
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_wrb_v2, ptr2nextwrb,
|
|
|
|
io_task->pwrb_context->plast_wrb,
|
|
|
|
io_task->pwrb_handle->wrb_index);
|
|
|
|
io_task->pwrb_context->plast_wrb = pwrb;
|
|
|
|
|
2013-04-06 10:38:26 +07:00
|
|
|
pwrb_typeoffset = SKH_WRB_TYPE_OFFSET;
|
2012-10-20 06:14:23 +07:00
|
|
|
}
|
|
|
|
|
2010-02-20 09:33:56 +07:00
|
|
|
|
2009-09-05 09:06:35 +07:00
|
|
|
switch (task->hdr->opcode & ISCSI_OPCODE_MASK) {
|
|
|
|
case ISCSI_OP_LOGIN:
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_wrb, cmdsn_itt, pwrb, 1);
|
2012-10-20 06:14:23 +07:00
|
|
|
ADAPTER_SET_WRB_TYPE(pwrb, TGT_DM_CMD, pwrb_typeoffset);
|
2016-01-16 04:45:34 +07:00
|
|
|
ret = hwi_write_buffer(pwrb, task);
|
2009-09-05 09:06:35 +07:00
|
|
|
break;
|
|
|
|
case ISCSI_OP_NOOP_OUT:
|
2011-03-26 04:24:01 +07:00
|
|
|
if (task->hdr->ttt != ISCSI_RESERVED_TAG) {
|
2012-10-20 06:14:23 +07:00
|
|
|
ADAPTER_SET_WRB_TYPE(pwrb, TGT_DM_CMD, pwrb_typeoffset);
|
2013-04-06 10:38:26 +07:00
|
|
|
if (is_chip_be2_be3r(phba))
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_wrb,
|
2012-10-20 06:14:23 +07:00
|
|
|
dmsg, pwrb, 1);
|
|
|
|
else
|
2013-04-06 10:38:26 +07:00
|
|
|
AMAP_SET_BITS(struct amap_iscsi_wrb_v2,
|
2012-10-20 06:14:23 +07:00
|
|
|
dmsg, pwrb, 1);
|
2011-03-26 04:24:01 +07:00
|
|
|
} else {
|
2012-10-20 06:14:23 +07:00
|
|
|
ADAPTER_SET_WRB_TYPE(pwrb, INI_RD_CMD, pwrb_typeoffset);
|
2013-04-06 10:38:26 +07:00
|
|
|
if (is_chip_be2_be3r(phba))
|
|
|
|
AMAP_SET_BITS(struct amap_iscsi_wrb,
|
2012-10-20 06:14:23 +07:00
|
|
|
dmsg, pwrb, 0);
|
|
|
|
else
|
2013-04-06 10:38:26 +07:00
|
|
|
AMAP_SET_BITS(struct amap_iscsi_wrb_v2,
|
2012-10-20 06:14:23 +07:00
|
|
|
dmsg, pwrb, 0);
|
2011-03-26 04:24:01 +07:00
|
|
|
}
|
2016-01-16 04:45:34 +07:00
|
|
|
ret = hwi_write_buffer(pwrb, task);
|
2009-09-05 09:06:35 +07:00
|
|
|
break;
|
|
|
|
case ISCSI_OP_TEXT:
|
2012-10-20 06:14:23 +07:00
|
|
|
ADAPTER_SET_WRB_TYPE(pwrb, TGT_DM_CMD, pwrb_typeoffset);
|
2016-01-16 04:45:34 +07:00
|
|
|
ret = hwi_write_buffer(pwrb, task);
|
2009-09-05 09:06:35 +07:00
|
|
|
break;
|
|
|
|
case ISCSI_OP_SCSI_TMFUNC:
|
2012-10-20 06:14:23 +07:00
|
|
|
ADAPTER_SET_WRB_TYPE(pwrb, INI_TMF_CMD, pwrb_typeoffset);
|
2016-01-16 04:45:34 +07:00
|
|
|
ret = hwi_write_buffer(pwrb, task);
|
2009-09-05 09:06:35 +07:00
|
|
|
break;
|
|
|
|
case ISCSI_OP_LOGOUT:
|
2012-10-20 06:14:23 +07:00
|
|
|
ADAPTER_SET_WRB_TYPE(pwrb, HWH_TYPE_LOGOUT, pwrb_typeoffset);
|
2016-01-16 04:45:34 +07:00
|
|
|
ret = hwi_write_buffer(pwrb, task);
|
2009-09-05 09:06:35 +07:00
|
|
|
break;
|
|
|
|
|
|
|
|
default:
|
2012-08-21 00:30:18 +07:00
|
|
|
beiscsi_log(phba, KERN_ERR, BEISCSI_LOG_CONFIG,
|
|
|
|
"BM_%d : opcode =%d Not supported\n",
|
|
|
|
task->hdr->opcode & ISCSI_OPCODE_MASK);
|
|
|
|
|
2009-09-05 09:06:35 +07:00
|
|
|
return -EINVAL;
|
|
|
|
}
|
|
|
|
|
2016-01-16 04:45:34 +07:00
|
|
|
if (ret)
|
|
|
|
return ret;
|
|
|
|
|
2012-10-20 06:14:23 +07:00
|
|
|
/* Set the task type */
|
2013-04-06 10:38:26 +07:00
|
|
|
io_task->wrb_type = (is_chip_be2_be3r(phba)) ?
|
|
|
|
AMAP_GET_BITS(struct amap_iscsi_wrb, type, pwrb) :
|
|
|
|
AMAP_GET_BITS(struct amap_iscsi_wrb_v2, type, pwrb);
|
2009-09-05 09:06:35 +07:00
|
|
|
|
2009-10-23 13:22:33 +07:00
|
|
|
doorbell |= cid & DB_WRB_POST_CID_MASK;
|
2010-01-23 07:04:24 +07:00
|
|
|
doorbell |= (io_task->pwrb_handle->wrb_index &
|
2009-09-05 09:06:35 +07:00
|
|
|
DB_DEF_PDU_WRB_INDEX_MASK) << DB_DEF_PDU_WRB_INDEX_SHIFT;
|
|
|
|
doorbell |= 1 << DB_DEF_PDU_NUM_POSTED_SHIFT;
|
2013-09-29 05:35:50 +07:00
|
|
|
iowrite32(doorbell, phba->db_va +
|
|
|
|
beiscsi_conn->doorbell_offset);
|
2009-09-05 09:06:35 +07:00
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
static int beiscsi_task_xmit(struct iscsi_task *task)
|
|
|
|
{
|
|
|
|
struct beiscsi_io_task *io_task = task->dd_data;
|
|
|
|
struct scsi_cmnd *sc = task->sc;
|
2016-02-04 17:19:18 +07:00
|
|
|
struct beiscsi_hba *phba;
|
2009-09-05 09:06:35 +07:00
|
|
|
struct scatterlist *sg;
|
|
|
|
int num_sg;
|
|
|
|
unsigned int writedir = 0, xferlen = 0;
|
|
|
|
|
2016-08-19 16:50:11 +07:00
|
|
|
phba = io_task->conn->phba;
|
|
|
|
/**
|
|
|
|
* HBA in error includes BEISCSI_HBA_FW_TIMEOUT. IO path might be
|
|
|
|
* operational if FW still gets heartbeat from EP FW. Is management
|
|
|
|
* path really needed to continue further?
|
|
|
|
*/
|
|
|
|
if (beiscsi_hba_in_error(phba))
|
|
|
|
return -EIO;
|
|
|
|
|
2016-02-04 17:19:18 +07:00
|
|
|
if (!io_task->conn->login_in_progress)
|
|
|
|
task->hdr->exp_statsn = 0;
|
2012-10-20 06:14:23 +07:00
|
|
|
|
2009-09-05 09:06:35 +07:00
|
|
|
if (!sc)
|
|
|
|
return beiscsi_mtask(task);
|
|
|
|
|
|
|
|
io_task->scsi_cmnd = sc;
|
2016-08-19 16:50:11 +07:00
|
|
|
io_task->num_sg = 0;
|
2009-09-05 09:06:35 +07:00
|
|
|
num_sg = scsi_dma_map(sc);
|
|
|
|
if (num_sg < 0) {
|
2013-09-29 05:35:55 +07:00
|
|
|
beiscsi_log(phba, KERN_ERR,
|
|
|
|
BEISCSI_LOG_IO | BEISCSI_LOG_ISCSI,
|
|
|
|
"BM_%d : scsi_dma_map Failed "
|
|
|
|
"Driver_ITT : 0x%x ITT : 0x%x Xferlen : 0x%x\n",
|
|
|
|
be32_to_cpu(io_task->cmd_bhs->iscsi_hdr.itt),
|
|
|
|
io_task->libiscsi_itt, scsi_bufflen(sc));
|
2012-08-21 00:30:18 +07:00
|
|
|
|
2009-09-05 09:06:35 +07:00
|
|
|
return num_sg;
|
|
|
|
}
|
2016-08-19 16:50:11 +07:00
|
|
|
/**
|
|
|
|
* For scsi cmd task, check num_sg before unmapping in cleanup_task.
|
|
|
|
* For management task, cleanup_task checks mtask_addr before unmapping.
|
|
|
|
*/
|
|
|
|
io_task->num_sg = num_sg;
|
2009-09-05 09:06:35 +07:00
|
|
|
xferlen = scsi_bufflen(sc);
|
|
|
|
sg = scsi_sglist(sc);
|
2012-08-21 00:30:18 +07:00
|
|
|
if (sc->sc_data_direction == DMA_TO_DEVICE)
|
2009-09-05 09:06:35 +07:00
|
|
|
writedir = 1;
|
2012-08-21 00:30:18 +07:00
|
|
|
else
|
2009-09-05 09:06:35 +07:00
|
|
|
writedir = 0;
|
2012-08-21 00:30:18 +07:00
|
|
|
|
2012-10-20 06:14:23 +07:00
|
|
|
return phba->iotask_fn(task, sg, num_sg, xferlen, writedir);
|
2009-09-05 09:06:35 +07:00
|
|
|
}
|
|
|
|
|
2012-04-04 11:41:50 +07:00
|
|
|
/**
|
|
|
|
* beiscsi_bsg_request - handle bsg request from ISCSI transport
|
|
|
|
* @job: job to handle
|
|
|
|
*/
|
|
|
|
static int beiscsi_bsg_request(struct bsg_job *job)
|
|
|
|
{
|
|
|
|
struct Scsi_Host *shost;
|
|
|
|
struct beiscsi_hba *phba;
|
|
|
|
struct iscsi_bsg_request *bsg_req = job->request;
|
|
|
|
int rc = -EINVAL;
|
|
|
|
unsigned int tag;
|
|
|
|
struct be_dma_mem nonemb_cmd;
|
|
|
|
struct be_cmd_resp_hdr *resp;
|
|
|
|
struct iscsi_bsg_reply *bsg_reply = job->reply;
|
|
|
|
unsigned short status, extd_status;
|
|
|
|
|
|
|
|
shost = iscsi_job_to_shost(job);
|
|
|
|
phba = iscsi_host_priv(shost);
|
|
|
|
|
2016-08-19 16:50:11 +07:00
|
|
|
if (beiscsi_hba_in_error(phba)) {
|
|
|
|
beiscsi_log(phba, KERN_INFO, BEISCSI_LOG_CONFIG,
|
|
|
|
"BM_%d : HBA in error 0x%lx\n", phba->state);
|
|
|
|
return -ENXIO;
|
|
|
|
}
|
|
|
|
|
2012-04-04 11:41:50 +07:00
|
|
|
switch (bsg_req->msgcode) {
|
|
|
|
case ISCSI_BSG_HST_VENDOR:
|
|
|
|
nonemb_cmd.va = pci_alloc_consistent(phba->ctrl.pdev,
|
|
|
|
job->request_payload.payload_len,
|
|
|
|
&nonemb_cmd.dma);
|
|
|
|
if (nonemb_cmd.va == NULL) {
|
2012-08-21 00:30:18 +07:00
|
|
|
beiscsi_log(phba, KERN_ERR, BEISCSI_LOG_CONFIG,
|
|
|
|
"BM_%d : Failed to allocate memory for "
|
|
|
|
"beiscsi_bsg_request\n");
|
2012-10-20 06:13:03 +07:00
|
|
|
return -ENOMEM;
|
2012-04-04 11:41:50 +07:00
|
|
|
}
|
|
|
|
tag = mgmt_vendor_specific_fw_cmd(&phba->ctrl, phba, job,
|
|
|
|
&nonemb_cmd);
|
|
|
|
if (!tag) {
|
2012-08-21 00:30:18 +07:00
|
|
|
beiscsi_log(phba, KERN_ERR, BEISCSI_LOG_CONFIG,
|
2012-10-20 06:13:03 +07:00
|
|
|
"BM_%d : MBX Tag Allocation Failed\n");
|
2012-08-21 00:30:18 +07:00
|
|
|
|
2012-04-04 11:41:50 +07:00
|
|
|
pci_free_consistent(phba->ctrl.pdev, nonemb_cmd.size,
|
|
|
|
nonemb_cmd.va, nonemb_cmd.dma);
|
|
|
|
return -EAGAIN;
|
2012-10-20 06:15:40 +07:00
|
|
|
}
|
|
|
|
|
|
|
|
rc = wait_event_interruptible_timeout(
|
|
|
|
phba->ctrl.mcc_wait[tag],
|
2016-02-04 17:19:10 +07:00
|
|
|
phba->ctrl.mcc_tag_status[tag],
|
2012-10-20 06:15:40 +07:00
|
|
|
msecs_to_jiffies(
|
|
|
|
BEISCSI_HOST_MBX_TIMEOUT));
|
2016-02-04 17:19:10 +07:00
|
|
|
extd_status = (phba->ctrl.mcc_tag_status[tag] &
|
|
|
|
CQE_STATUS_ADDL_MASK) >> CQE_STATUS_ADDL_SHIFT;
|
|
|
|
status = phba->ctrl.mcc_tag_status[tag] & CQE_STATUS_MASK;
|
2016-02-04 17:19:17 +07:00
|
|
|
free_mcc_wrb(&phba->ctrl, tag);
|
2012-04-04 11:41:50 +07:00
|
|
|
resp = (struct be_cmd_resp_hdr *)nonemb_cmd.va;
|
|
|
|
sg_copy_from_buffer(job->reply_payload.sg_list,
|
|
|
|
job->reply_payload.sg_cnt,
|
|
|
|
nonemb_cmd.va, (resp->response_length
|
|
|
|
+ sizeof(*resp)));
|
|
|
|
bsg_reply->reply_payload_rcv_len = resp->response_length;
|
|
|
|
bsg_reply->result = status;
|
|
|
|
bsg_job_done(job, bsg_reply->result,
|
|
|
|
bsg_reply->reply_payload_rcv_len);
|
|
|
|
pci_free_consistent(phba->ctrl.pdev, nonemb_cmd.size,
|
|
|
|
nonemb_cmd.va, nonemb_cmd.dma);
|
|
|
|
if (status || extd_status) {
|
2012-08-21 00:30:18 +07:00
|
|
|
beiscsi_log(phba, KERN_ERR, BEISCSI_LOG_CONFIG,
|
2012-10-20 06:13:03 +07:00
|
|
|
"BM_%d : MBX Cmd Failed"
|
2012-08-21 00:30:18 +07:00
|
|
|
" status = %d extd_status = %d\n",
|
|
|
|
status, extd_status);
|
|
|
|
|
2012-04-04 11:41:50 +07:00
|
|
|
return -EIO;
|
2012-10-20 06:13:03 +07:00
|
|
|
} else {
|
|
|
|
rc = 0;
|
2012-04-04 11:41:50 +07:00
|
|
|
}
|
|
|
|
break;
|
|
|
|
|
|
|
|
default:
|
2012-08-21 00:30:18 +07:00
|
|
|
beiscsi_log(phba, KERN_ERR, BEISCSI_LOG_CONFIG,
|
|
|
|
"BM_%d : Unsupported bsg command: 0x%x\n",
|
|
|
|
bsg_req->msgcode);
|
2012-04-04 11:41:50 +07:00
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
return rc;
|
|
|
|
}
|
|
|
|
|
2012-08-21 00:30:18 +07:00
|
|
|
void beiscsi_hba_attrs_init(struct beiscsi_hba *phba)
|
|
|
|
{
|
|
|
|
/* Set the logging parameter */
|
|
|
|
beiscsi_log_enable_init(phba, beiscsi_log_enable);
|
|
|
|
}
|
|
|
|
|
2012-10-20 06:12:37 +07:00
|
|
|
/*
|
|
|
|
* beiscsi_quiesce()- Cleanup Driver resources
|
|
|
|
* @phba: Instance Priv structure
|
|
|
|
*
|
|
|
|
* Free the OS and HW resources held by the driver
|
|
|
|
**/
|
2016-08-19 16:50:11 +07:00
|
|
|
static void beiscsi_quiesce(struct beiscsi_hba *phba)
|
2009-09-05 09:06:35 +07:00
|
|
|
{
|
2009-10-23 13:22:33 +07:00
|
|
|
struct hwi_controller *phwi_ctrlr;
|
|
|
|
struct hwi_context_memory *phwi_context;
|
|
|
|
struct be_eq_obj *pbe_eq;
|
|
|
|
unsigned int i, msix_vec;
|
2009-09-05 09:06:35 +07:00
|
|
|
|
2009-10-23 13:22:33 +07:00
|
|
|
phwi_ctrlr = phba->phwi_ctrlr;
|
|
|
|
phwi_context = phwi_ctrlr->phwi_ctxt;
|
2009-09-05 09:06:35 +07:00
|
|
|
hwi_disable_intr(phba);
|
2009-10-23 13:22:33 +07:00
|
|
|
if (phba->msix_enabled) {
|
|
|
|
for (i = 0; i <= phba->num_cpus; i++) {
|
|
|
|
msix_vec = phba->msix_entries[i].vector;
|
|
|
|
free_irq(msix_vec, &phwi_context->be_eq[i]);
|
2011-08-25 06:05:30 +07:00
|
|
|
kfree(phba->msi_name[i]);
|
2009-10-23 13:22:33 +07:00
|
|
|
}
|
|
|
|
} else
|
2016-03-04 17:15:06 +07:00
|
|
|
if (phba->pcidev->irq)
|
2009-10-23 13:22:33 +07:00
|
|
|
free_irq(phba->pcidev->irq, phba);
|
|
|
|
pci_disable_msix(phba->pcidev);
|
2016-08-19 16:50:13 +07:00
|
|
|
cancel_delayed_work_sync(&phba->eqd_update);
|
2016-08-19 16:50:12 +07:00
|
|
|
cancel_work_sync(&phba->boot_work);
|
2016-08-19 16:50:13 +07:00
|
|
|
del_timer_sync(&phba->hw_check);
|
2013-09-29 05:35:58 +07:00
|
|
|
|
2014-03-13 22:38:42 +07:00
|
|
|
for (i = 0; i < phba->num_cpus; i++) {
|
|
|
|
pbe_eq = &phwi_context->be_eq[i];
|
2015-11-10 20:56:14 +07:00
|
|
|
irq_poll_disable(&pbe_eq->iopoll);
|
2014-03-13 22:38:42 +07:00
|
|
|
}
|
2009-09-05 09:06:35 +07:00
|
|
|
|
2016-08-19 16:50:11 +07:00
|
|
|
/* PCI_ERR is set then check if driver is not unloading */
|
|
|
|
if (test_bit(BEISCSI_HBA_RUNNING, &phba->state) &&
|
|
|
|
test_bit(BEISCSI_HBA_PCI_ERR, &phba->state)) {
|
2013-09-29 05:35:58 +07:00
|
|
|
hwi_cleanup(phba);
|
2016-08-19 16:50:11 +07:00
|
|
|
return;
|
2013-09-29 05:35:58 +07:00
|
|
|
}
|
2012-10-20 06:15:51 +07:00
|
|
|
|
2016-08-19 16:50:11 +07:00
|
|
|
destroy_workqueue(phba->wq);
|
|
|
|
beiscsi_clean_port(phba);
|
|
|
|
beiscsi_free_mem(phba);
|
|
|
|
|
|
|
|
beiscsi_unmap_pci_function(phba);
|
|
|
|
pci_free_consistent(phba->pcidev,
|
|
|
|
phba->ctrl.mbox_mem_alloced.size,
|
|
|
|
phba->ctrl.mbox_mem_alloced.va,
|
|
|
|
phba->ctrl.mbox_mem_alloced.dma);
|
2011-08-23 00:08:28 +07:00
|
|
|
}
|
|
|
|
|
|
|
|
static void beiscsi_remove(struct pci_dev *pcidev)
|
|
|
|
{
|
|
|
|
struct beiscsi_hba *phba = NULL;
|
|
|
|
|
|
|
|
phba = pci_get_drvdata(pcidev);
|
|
|
|
if (!phba) {
|
|
|
|
dev_err(&pcidev->dev, "beiscsi_remove called with no phba\n");
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
2016-08-19 16:50:11 +07:00
|
|
|
clear_bit(BEISCSI_HBA_RUNNING, &phba->state);
|
2016-08-19 16:50:06 +07:00
|
|
|
beiscsi_iface_destroy_default(phba);
|
2009-09-05 09:06:35 +07:00
|
|
|
iscsi_host_remove(phba->shost);
|
2016-08-19 16:50:11 +07:00
|
|
|
beiscsi_quiesce(phba);
|
2016-08-19 16:50:12 +07:00
|
|
|
/* after cancelling boot_work */
|
|
|
|
iscsi_boot_destroy_kset(phba->boot_struct.boot_kset);
|
2009-09-05 09:06:35 +07:00
|
|
|
pci_dev_put(phba->pcidev);
|
|
|
|
iscsi_host_free(phba->shost);
|
2013-09-29 05:35:58 +07:00
|
|
|
pci_disable_pcie_error_reporting(pcidev);
|
|
|
|
pci_set_drvdata(pcidev, NULL);
|
2015-04-25 09:47:19 +07:00
|
|
|
pci_release_regions(pcidev);
|
2011-08-23 00:08:29 +07:00
|
|
|
pci_disable_device(pcidev);
|
2009-09-05 09:06:35 +07:00
|
|
|
}
|
|
|
|
|
2009-10-23 13:22:33 +07:00
|
|
|
static void beiscsi_msix_enable(struct beiscsi_hba *phba)
|
|
|
|
{
|
|
|
|
int i, status;
|
|
|
|
|
|
|
|
for (i = 0; i <= phba->num_cpus; i++)
|
|
|
|
phba->msix_entries[i].entry = i;
|
|
|
|
|
2014-08-18 13:01:48 +07:00
|
|
|
status = pci_enable_msix_range(phba->pcidev, phba->msix_entries,
|
|
|
|
phba->num_cpus + 1, phba->num_cpus + 1);
|
|
|
|
if (status > 0)
|
2009-10-23 13:22:33 +07:00
|
|
|
phba->msix_enabled = true;
|
|
|
|
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
|
2016-08-19 16:50:12 +07:00
|
|
|
void beiscsi_start_boot_work(struct beiscsi_hba *phba, unsigned int s_handle)
|
|
|
|
{
|
|
|
|
if (phba->boot_struct.boot_kset)
|
|
|
|
return;
|
|
|
|
|
|
|
|
/* skip if boot work is already in progress */
|
|
|
|
if (test_and_set_bit(BEISCSI_HBA_BOOT_WORK, &phba->state))
|
|
|
|
return;
|
|
|
|
|
|
|
|
phba->boot_struct.retry = 3;
|
|
|
|
phba->boot_struct.tag = 0;
|
|
|
|
phba->boot_struct.s_handle = s_handle;
|
|
|
|
phba->boot_struct.action = BEISCSI_BOOT_GET_SHANDLE;
|
|
|
|
schedule_work(&phba->boot_work);
|
|
|
|
}
|
|
|
|
|
|
|
|
static ssize_t beiscsi_show_boot_tgt_info(void *data, int type, char *buf)
|
|
|
|
{
|
|
|
|
struct beiscsi_hba *phba = data;
|
|
|
|
struct mgmt_session_info *boot_sess = &phba->boot_struct.boot_sess;
|
|
|
|
struct mgmt_conn_info *boot_conn = &boot_sess->conn_list[0];
|
|
|
|
char *str = buf;
|
|
|
|
int rc = -EPERM;
|
|
|
|
|
|
|
|
switch (type) {
|
|
|
|
case ISCSI_BOOT_TGT_NAME:
|
|
|
|
rc = sprintf(buf, "%.*s\n",
|
|
|
|
(int)strlen(boot_sess->target_name),
|
|
|
|
(char *)&boot_sess->target_name);
|
|
|
|
break;
|
|
|
|
case ISCSI_BOOT_TGT_IP_ADDR:
|
|
|
|
if (boot_conn->dest_ipaddr.ip_type == BEISCSI_IP_TYPE_V4)
|
|
|
|
rc = sprintf(buf, "%pI4\n",
|
|
|
|
(char *)&boot_conn->dest_ipaddr.addr);
|
|
|
|
else
|
|
|
|
rc = sprintf(str, "%pI6\n",
|
|
|
|
(char *)&boot_conn->dest_ipaddr.addr);
|
|
|
|
break;
|
|
|
|
case ISCSI_BOOT_TGT_PORT:
|
|
|
|
rc = sprintf(str, "%d\n", boot_conn->dest_port);
|
|
|
|
break;
|
|
|
|
|
|
|
|
case ISCSI_BOOT_TGT_CHAP_NAME:
|
|
|
|
rc = sprintf(str, "%.*s\n",
|
|
|
|
boot_conn->negotiated_login_options.auth_data.chap.
|
|
|
|
target_chap_name_length,
|
|
|
|
(char *)&boot_conn->negotiated_login_options.
|
|
|
|
auth_data.chap.target_chap_name);
|
|
|
|
break;
|
|
|
|
case ISCSI_BOOT_TGT_CHAP_SECRET:
|
|
|
|
rc = sprintf(str, "%.*s\n",
|
|
|
|
boot_conn->negotiated_login_options.auth_data.chap.
|
|
|
|
target_secret_length,
|
|
|
|
(char *)&boot_conn->negotiated_login_options.
|
|
|
|
auth_data.chap.target_secret);
|
|
|
|
break;
|
|
|
|
case ISCSI_BOOT_TGT_REV_CHAP_NAME:
|
|
|
|
rc = sprintf(str, "%.*s\n",
|
|
|
|
boot_conn->negotiated_login_options.auth_data.chap.
|
|
|
|
intr_chap_name_length,
|
|
|
|
(char *)&boot_conn->negotiated_login_options.
|
|
|
|
auth_data.chap.intr_chap_name);
|
|
|
|
break;
|
|
|
|
case ISCSI_BOOT_TGT_REV_CHAP_SECRET:
|
|
|
|
rc = sprintf(str, "%.*s\n",
|
|
|
|
boot_conn->negotiated_login_options.auth_data.chap.
|
|
|
|
intr_secret_length,
|
|
|
|
(char *)&boot_conn->negotiated_login_options.
|
|
|
|
auth_data.chap.intr_secret);
|
|
|
|
break;
|
|
|
|
case ISCSI_BOOT_TGT_FLAGS:
|
|
|
|
rc = sprintf(str, "2\n");
|
|
|
|
break;
|
|
|
|
case ISCSI_BOOT_TGT_NIC_ASSOC:
|
|
|
|
rc = sprintf(str, "0\n");
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
return rc;
|
|
|
|
}
|
|
|
|
|
|
|
|
static ssize_t beiscsi_show_boot_ini_info(void *data, int type, char *buf)
|
|
|
|
{
|
|
|
|
struct beiscsi_hba *phba = data;
|
|
|
|
char *str = buf;
|
|
|
|
int rc = -EPERM;
|
|
|
|
|
|
|
|
switch (type) {
|
|
|
|
case ISCSI_BOOT_INI_INITIATOR_NAME:
|
|
|
|
rc = sprintf(str, "%s\n",
|
|
|
|
phba->boot_struct.boot_sess.initiator_iscsiname);
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
return rc;
|
|
|
|
}
|
|
|
|
|
|
|
|
static ssize_t beiscsi_show_boot_eth_info(void *data, int type, char *buf)
|
|
|
|
{
|
|
|
|
struct beiscsi_hba *phba = data;
|
|
|
|
char *str = buf;
|
|
|
|
int rc = -EPERM;
|
|
|
|
|
|
|
|
switch (type) {
|
|
|
|
case ISCSI_BOOT_ETH_FLAGS:
|
|
|
|
rc = sprintf(str, "2\n");
|
|
|
|
break;
|
|
|
|
case ISCSI_BOOT_ETH_INDEX:
|
|
|
|
rc = sprintf(str, "0\n");
|
|
|
|
break;
|
|
|
|
case ISCSI_BOOT_ETH_MAC:
|
|
|
|
rc = beiscsi_get_macaddr(str, phba);
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
return rc;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
static umode_t beiscsi_tgt_get_attr_visibility(void *data, int type)
|
|
|
|
{
|
|
|
|
umode_t rc = 0;
|
|
|
|
|
|
|
|
switch (type) {
|
|
|
|
case ISCSI_BOOT_TGT_NAME:
|
|
|
|
case ISCSI_BOOT_TGT_IP_ADDR:
|
|
|
|
case ISCSI_BOOT_TGT_PORT:
|
|
|
|
case ISCSI_BOOT_TGT_CHAP_NAME:
|
|
|
|
case ISCSI_BOOT_TGT_CHAP_SECRET:
|
|
|
|
case ISCSI_BOOT_TGT_REV_CHAP_NAME:
|
|
|
|
case ISCSI_BOOT_TGT_REV_CHAP_SECRET:
|
|
|
|
case ISCSI_BOOT_TGT_NIC_ASSOC:
|
|
|
|
case ISCSI_BOOT_TGT_FLAGS:
|
|
|
|
rc = S_IRUGO;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
return rc;
|
|
|
|
}
|
|
|
|
|
|
|
|
static umode_t beiscsi_ini_get_attr_visibility(void *data, int type)
|
|
|
|
{
|
|
|
|
umode_t rc = 0;
|
|
|
|
|
|
|
|
switch (type) {
|
|
|
|
case ISCSI_BOOT_INI_INITIATOR_NAME:
|
|
|
|
rc = S_IRUGO;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
return rc;
|
|
|
|
}
|
|
|
|
|
|
|
|
|
|
|
|
static umode_t beiscsi_eth_get_attr_visibility(void *data, int type)
|
|
|
|
{
|
|
|
|
umode_t rc = 0;
|
|
|
|
|
|
|
|
switch (type) {
|
|
|
|
case ISCSI_BOOT_ETH_FLAGS:
|
|
|
|
case ISCSI_BOOT_ETH_MAC:
|
|
|
|
case ISCSI_BOOT_ETH_INDEX:
|
|
|
|
rc = S_IRUGO;
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
return rc;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void beiscsi_boot_kobj_release(void *data)
|
|
|
|
{
|
|
|
|
struct beiscsi_hba *phba = data;
|
|
|
|
|
|
|
|
scsi_host_put(phba->shost);
|
|
|
|
}
|
|
|
|
|
|
|
|
static int beiscsi_boot_create_kset(struct beiscsi_hba *phba)
|
|
|
|
{
|
|
|
|
struct boot_struct *bs = &phba->boot_struct;
|
|
|
|
struct iscsi_boot_kobj *boot_kobj;
|
|
|
|
|
|
|
|
if (bs->boot_kset) {
|
|
|
|
__beiscsi_log(phba, KERN_ERR,
|
|
|
|
"BM_%d: boot_kset already created\n");
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
|
|
|
|
bs->boot_kset = iscsi_boot_create_host_kset(phba->shost->host_no);
|
|
|
|
if (!bs->boot_kset) {
|
|
|
|
__beiscsi_log(phba, KERN_ERR,
|
|
|
|
"BM_%d: boot_kset alloc failed\n");
|
|
|
|
return -ENOMEM;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* get shost ref because the show function will refer phba */
|
|
|
|
if (!scsi_host_get(phba->shost))
|
|
|
|
goto free_kset;
|
|
|
|
|
|
|
|
boot_kobj = iscsi_boot_create_target(bs->boot_kset, 0, phba,
|
|
|
|
beiscsi_show_boot_tgt_info,
|
|
|
|
beiscsi_tgt_get_attr_visibility,
|
|
|
|
beiscsi_boot_kobj_release);
|
|
|
|
if (!boot_kobj)
|
|
|
|
goto put_shost;
|
|
|
|
|
|
|
|
if (!scsi_host_get(phba->shost))
|
|
|
|
goto free_kset;
|
|
|
|
|
|
|
|
boot_kobj = iscsi_boot_create_initiator(bs->boot_kset, 0, phba,
|
|
|
|
beiscsi_show_boot_ini_info,
|
|
|
|
beiscsi_ini_get_attr_visibility,
|
|
|
|
beiscsi_boot_kobj_release);
|
|
|
|
if (!boot_kobj)
|
|
|
|
goto put_shost;
|
|
|
|
|
|
|
|
if (!scsi_host_get(phba->shost))
|
|
|
|
goto free_kset;
|
|
|
|
|
|
|
|
boot_kobj = iscsi_boot_create_ethernet(bs->boot_kset, 0, phba,
|
|
|
|
beiscsi_show_boot_eth_info,
|
|
|
|
beiscsi_eth_get_attr_visibility,
|
|
|
|
beiscsi_boot_kobj_release);
|
|
|
|
if (!boot_kobj)
|
|
|
|
goto put_shost;
|
|
|
|
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
put_shost:
|
|
|
|
scsi_host_put(phba->shost);
|
|
|
|
free_kset:
|
|
|
|
iscsi_boot_destroy_kset(bs->boot_kset);
|
|
|
|
bs->boot_kset = NULL;
|
|
|
|
return -ENOMEM;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void beiscsi_boot_work(struct work_struct *work)
|
|
|
|
{
|
|
|
|
struct beiscsi_hba *phba =
|
|
|
|
container_of(work, struct beiscsi_hba, boot_work);
|
|
|
|
struct boot_struct *bs = &phba->boot_struct;
|
|
|
|
unsigned int tag = 0;
|
|
|
|
|
|
|
|
if (beiscsi_hba_in_error(phba))
|
|
|
|
return;
|
|
|
|
|
|
|
|
beiscsi_log(phba, KERN_INFO,
|
|
|
|
BEISCSI_LOG_CONFIG | BEISCSI_LOG_MBOX,
|
|
|
|
"BM_%d : %s action %d\n",
|
|
|
|
__func__, phba->boot_struct.action);
|
|
|
|
|
|
|
|
switch (phba->boot_struct.action) {
|
|
|
|
case BEISCSI_BOOT_REOPEN_SESS:
|
|
|
|
tag = beiscsi_boot_reopen_sess(phba);
|
|
|
|
break;
|
|
|
|
case BEISCSI_BOOT_GET_SHANDLE:
|
|
|
|
tag = __beiscsi_boot_get_shandle(phba, 1);
|
|
|
|
break;
|
|
|
|
case BEISCSI_BOOT_GET_SINFO:
|
|
|
|
tag = beiscsi_boot_get_sinfo(phba);
|
|
|
|
break;
|
|
|
|
case BEISCSI_BOOT_LOGOUT_SESS:
|
|
|
|
tag = beiscsi_boot_logout_sess(phba);
|
|
|
|
break;
|
|
|
|
case BEISCSI_BOOT_CREATE_KSET:
|
|
|
|
beiscsi_boot_create_kset(phba);
|
|
|
|
/**
|
|
|
|
* updated boot_kset is made visible to all before
|
|
|
|
* ending the boot work.
|
|
|
|
*/
|
|
|
|
mb();
|
|
|
|
clear_bit(BEISCSI_HBA_BOOT_WORK, &phba->state);
|
|
|
|
return;
|
|
|
|
}
|
|
|
|
if (!tag) {
|
|
|
|
if (bs->retry--)
|
|
|
|
schedule_work(&phba->boot_work);
|
|
|
|
else
|
|
|
|
clear_bit(BEISCSI_HBA_BOOT_WORK, &phba->state);
|
|
|
|
}
|
|
|
|
}
|
|
|
|
|
2016-08-19 16:50:13 +07:00
|
|
|
static void beiscsi_hw_health_check(unsigned long ptr)
|
2014-05-06 08:41:26 +07:00
|
|
|
{
|
2016-08-19 16:50:13 +07:00
|
|
|
struct beiscsi_hba *phba;
|
|
|
|
|
|
|
|
phba = (struct beiscsi_hba *)ptr;
|
|
|
|
beiscsi_ue_detect(phba);
|
|
|
|
if (test_bit(BEISCSI_HBA_IN_UE, &phba->state))
|
|
|
|
return;
|
|
|
|
|
|
|
|
mod_timer(&phba->hw_check,
|
|
|
|
jiffies + msecs_to_jiffies(BEISCSI_UE_DETECT_INTERVAL));
|
|
|
|
}
|
|
|
|
|
|
|
|
static void beiscsi_eqd_update_work(struct work_struct *work)
|
|
|
|
{
|
|
|
|
struct hwi_context_memory *phwi_context;
|
2014-05-06 08:41:26 +07:00
|
|
|
struct be_set_eqd set_eqd[MAX_CPUS];
|
|
|
|
struct hwi_controller *phwi_ctrlr;
|
2016-08-19 16:50:13 +07:00
|
|
|
struct be_eq_obj *pbe_eq;
|
|
|
|
struct beiscsi_hba *phba;
|
|
|
|
unsigned int pps, delta;
|
|
|
|
struct be_aic_obj *aic;
|
2014-05-06 08:41:26 +07:00
|
|
|
int eqd, i, num = 0;
|
2016-08-19 16:50:13 +07:00
|
|
|
unsigned long now;
|
2014-05-06 08:41:26 +07:00
|
|
|
|
2016-08-19 16:50:13 +07:00
|
|
|
phba = container_of(work, struct beiscsi_hba, eqd_update.work);
|
2016-08-19 16:50:11 +07:00
|
|
|
if (beiscsi_hba_in_error(phba))
|
|
|
|
return;
|
|
|
|
|
2014-05-06 08:41:26 +07:00
|
|
|
phwi_ctrlr = phba->phwi_ctrlr;
|
|
|
|
phwi_context = phwi_ctrlr->phwi_ctxt;
|
|
|
|
|
|
|
|
for (i = 0; i <= phba->num_cpus; i++) {
|
|
|
|
aic = &phba->aic_obj[i];
|
|
|
|
pbe_eq = &phwi_context->be_eq[i];
|
|
|
|
now = jiffies;
|
2016-08-19 16:50:13 +07:00
|
|
|
if (!aic->jiffies || time_before(now, aic->jiffies) ||
|
2014-05-06 08:41:26 +07:00
|
|
|
pbe_eq->cq_count < aic->eq_prev) {
|
2016-08-19 16:50:13 +07:00
|
|
|
aic->jiffies = now;
|
2014-05-06 08:41:26 +07:00
|
|
|
aic->eq_prev = pbe_eq->cq_count;
|
|
|
|
continue;
|
|
|
|
}
|
2016-08-19 16:50:13 +07:00
|
|
|
delta = jiffies_to_msecs(now - aic->jiffies);
|
2014-05-06 08:41:26 +07:00
|
|
|
pps = (((u32)(pbe_eq->cq_count - aic->eq_prev) * 1000) / delta);
|
|
|
|
eqd = (pps / 1500) << 2;
|
|
|
|
|
|
|
|
if (eqd < 8)
|
|
|
|
eqd = 0;
|
|
|
|
eqd = min_t(u32, eqd, phwi_context->max_eqd);
|
|
|
|
eqd = max_t(u32, eqd, phwi_context->min_eqd);
|
|
|
|
|
2016-08-19 16:50:13 +07:00
|
|
|
aic->jiffies = now;
|
2014-05-06 08:41:26 +07:00
|
|
|
aic->eq_prev = pbe_eq->cq_count;
|
|
|
|
|
|
|
|
if (eqd != aic->prev_eqd) {
|
|
|
|
set_eqd[num].delay_multiplier = (eqd * 65)/100;
|
|
|
|
set_eqd[num].eq_id = pbe_eq->q.id;
|
|
|
|
aic->prev_eqd = eqd;
|
|
|
|
num++;
|
|
|
|
}
|
|
|
|
}
|
2016-08-19 16:50:13 +07:00
|
|
|
if (num)
|
|
|
|
/* completion of this is ignored */
|
|
|
|
beiscsi_modify_eq_delay(phba, set_eqd, num);
|
2014-05-06 08:41:26 +07:00
|
|
|
|
2016-08-19 16:50:13 +07:00
|
|
|
schedule_delayed_work(&phba->eqd_update,
|
|
|
|
msecs_to_jiffies(BEISCSI_EQD_UPDATE_INTERVAL));
|
2012-10-20 06:15:51 +07:00
|
|
|
}
|
|
|
|
|
2013-09-29 05:35:58 +07:00
|
|
|
|
|
|
|
static pci_ers_result_t beiscsi_eeh_err_detected(struct pci_dev *pdev,
|
|
|
|
pci_channel_state_t state)
|
|
|
|
{
|
|
|
|
struct beiscsi_hba *phba = NULL;
|
|
|
|
|
|
|
|
phba = (struct beiscsi_hba *)pci_get_drvdata(pdev);
|
2016-08-19 16:50:11 +07:00
|
|
|
set_bit(BEISCSI_HBA_PCI_ERR, &phba->state);
|
2013-09-29 05:35:58 +07:00
|
|
|
|
|
|
|
beiscsi_log(phba, KERN_ERR, BEISCSI_LOG_INIT,
|
|
|
|
"BM_%d : EEH error detected\n");
|
|
|
|
|
2016-08-19 16:50:11 +07:00
|
|
|
beiscsi_quiesce(phba);
|
2013-09-29 05:35:58 +07:00
|
|
|
|
|
|
|
if (state == pci_channel_io_perm_failure) {
|
|
|
|
beiscsi_log(phba, KERN_ERR, BEISCSI_LOG_INIT,
|
|
|
|
"BM_%d : EEH : State PERM Failure");
|
|
|
|
return PCI_ERS_RESULT_DISCONNECT;
|
|
|
|
}
|
|
|
|
|
|
|
|
pci_disable_device(pdev);
|
|
|
|
|
|
|
|
/* The error could cause the FW to trigger a flash debug dump.
|
|
|
|
* Resetting the card while flash dump is in progress
|
|
|
|
* can cause it not to recover; wait for it to finish.
|
|
|
|
* Wait only for first function as it is needed only once per
|
|
|
|
* adapter.
|
|
|
|
**/
|
|
|
|
if (pdev->devfn == 0)
|
|
|
|
ssleep(30);
|
|
|
|
|
|
|
|
return PCI_ERS_RESULT_NEED_RESET;
|
|
|
|
}
|
|
|
|
|
|
|
|
static pci_ers_result_t beiscsi_eeh_reset(struct pci_dev *pdev)
|
|
|
|
{
|
|
|
|
struct beiscsi_hba *phba = NULL;
|
|
|
|
int status = 0;
|
|
|
|
|
|
|
|
phba = (struct beiscsi_hba *)pci_get_drvdata(pdev);
|
|
|
|
|
|
|
|
beiscsi_log(phba, KERN_ERR, BEISCSI_LOG_INIT,
|
|
|
|
"BM_%d : EEH Reset\n");
|
|
|
|
|
|
|
|
status = pci_enable_device(pdev);
|
|
|
|
if (status)
|
|
|
|
return PCI_ERS_RESULT_DISCONNECT;
|
|
|
|
|
|
|
|
pci_set_master(pdev);
|
|
|
|
pci_set_power_state(pdev, PCI_D0);
|
|
|
|
pci_restore_state(pdev);
|
|
|
|
|
|
|
|
/* Wait for the CHIP Reset to complete */
|
|
|
|
status = be_chk_reset_complete(phba);
|
|
|
|
if (!status) {
|
|
|
|
beiscsi_log(phba, KERN_WARNING, BEISCSI_LOG_INIT,
|
|
|
|
"BM_%d : EEH Reset Completed\n");
|
|
|
|
} else {
|
|
|
|
beiscsi_log(phba, KERN_WARNING, BEISCSI_LOG_INIT,
|
|
|
|
"BM_%d : EEH Reset Completion Failure\n");
|
|
|
|
return PCI_ERS_RESULT_DISCONNECT;
|
|
|
|
}
|
|
|
|
|
|
|
|
pci_cleanup_aer_uncorrect_error_status(pdev);
|
|
|
|
return PCI_ERS_RESULT_RECOVERED;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void beiscsi_eeh_resume(struct pci_dev *pdev)
|
|
|
|
{
|
|
|
|
int ret = 0, i;
|
|
|
|
struct be_eq_obj *pbe_eq;
|
|
|
|
struct beiscsi_hba *phba = NULL;
|
|
|
|
struct hwi_controller *phwi_ctrlr;
|
|
|
|
struct hwi_context_memory *phwi_context;
|
|
|
|
|
|
|
|
phba = (struct beiscsi_hba *)pci_get_drvdata(pdev);
|
|
|
|
pci_save_state(pdev);
|
|
|
|
|
|
|
|
if (enable_msix)
|
|
|
|
find_num_cpus(phba);
|
|
|
|
else
|
|
|
|
phba->num_cpus = 1;
|
|
|
|
|
|
|
|
if (enable_msix) {
|
|
|
|
beiscsi_msix_enable(phba);
|
|
|
|
if (!phba->msix_enabled)
|
|
|
|
phba->num_cpus = 1;
|
|
|
|
}
|
|
|
|
|
|
|
|
ret = beiscsi_cmd_reset_function(phba);
|
|
|
|
if (ret) {
|
|
|
|
beiscsi_log(phba, KERN_ERR, BEISCSI_LOG_INIT,
|
|
|
|
"BM_%d : Reset Failed\n");
|
|
|
|
goto ret_err;
|
|
|
|
}
|
|
|
|
|
|
|
|
ret = be_chk_reset_complete(phba);
|
|
|
|
if (ret) {
|
|
|
|
beiscsi_log(phba, KERN_ERR, BEISCSI_LOG_INIT,
|
|
|
|
"BM_%d : Failed to get out of reset.\n");
|
|
|
|
goto ret_err;
|
|
|
|
}
|
|
|
|
|
|
|
|
beiscsi_get_params(phba);
|
|
|
|
phba->shost->max_id = phba->params.cxns_per_ctrl;
|
|
|
|
phba->shost->can_queue = phba->params.ios_per_ctrl;
|
|
|
|
ret = hwi_init_controller(phba);
|
2016-02-28 00:43:25 +07:00
|
|
|
if (ret) {
|
|
|
|
beiscsi_log(phba, KERN_ERR, BEISCSI_LOG_INIT,
|
|
|
|
"BM_%d : beiscsi_eeh_resume -"
|
|
|
|
"Failed to initialize beiscsi_hba.\n");
|
|
|
|
goto ret_err;
|
|
|
|
}
|
2013-09-29 05:35:58 +07:00
|
|
|
|
|
|
|
for (i = 0; i < MAX_MCC_CMD; i++) {
|
|
|
|
init_waitqueue_head(&phba->ctrl.mcc_wait[i + 1]);
|
|
|
|
phba->ctrl.mcc_tag[i] = i + 1;
|
2016-02-04 17:19:10 +07:00
|
|
|
phba->ctrl.mcc_tag_status[i + 1] = 0;
|
2013-09-29 05:35:58 +07:00
|
|
|
phba->ctrl.mcc_tag_available++;
|
|
|
|
}
|
|
|
|
|
|
|
|
phwi_ctrlr = phba->phwi_ctrlr;
|
|
|
|
phwi_context = phwi_ctrlr->phwi_ctxt;
|
|
|
|
|
2014-03-13 22:38:42 +07:00
|
|
|
for (i = 0; i < phba->num_cpus; i++) {
|
2013-09-29 05:35:58 +07:00
|
|
|
pbe_eq = &phwi_context->be_eq[i];
|
2015-11-10 20:56:14 +07:00
|
|
|
irq_poll_init(&pbe_eq->iopoll, be_iopoll_budget,
|
2014-03-13 22:38:42 +07:00
|
|
|
be_iopoll);
|
2013-09-29 05:35:58 +07:00
|
|
|
}
|
|
|
|
|
2014-03-13 22:38:42 +07:00
|
|
|
i = (phba->msix_enabled) ? i : 0;
|
|
|
|
/* Work item for MCC handling */
|
|
|
|
pbe_eq = &phwi_context->be_eq[i];
|
2016-08-19 16:50:10 +07:00
|
|
|
INIT_WORK(&pbe_eq->mcc_work, beiscsi_mcc_work);
|
2014-03-13 22:38:42 +07:00
|
|
|
|
2013-09-29 05:35:58 +07:00
|
|
|
ret = beiscsi_init_irqs(phba);
|
|
|
|
if (ret < 0) {
|
|
|
|
beiscsi_log(phba, KERN_ERR, BEISCSI_LOG_INIT,
|
|
|
|
"BM_%d : beiscsi_eeh_resume - "
|
|
|
|
"Failed to beiscsi_init_irqs\n");
|
|
|
|
goto ret_err;
|
|
|
|
}
|
|
|
|
|
|
|
|
hwi_enable_intr(phba);
|
2016-08-19 16:50:11 +07:00
|
|
|
clear_bit(BEISCSI_HBA_PCI_ERR, &phba->state);
|
2013-09-29 05:35:58 +07:00
|
|
|
|
2016-08-19 16:50:13 +07:00
|
|
|
/* start hw_check timer and eqd_update work */
|
|
|
|
schedule_delayed_work(&phba->eqd_update,
|
|
|
|
msecs_to_jiffies(BEISCSI_EQD_UPDATE_INTERVAL));
|
|
|
|
mod_timer(&phba->hw_check,
|
|
|
|
jiffies + msecs_to_jiffies(BEISCSI_UE_DETECT_INTERVAL));
|
2013-09-29 05:35:58 +07:00
|
|
|
return;
|
|
|
|
ret_err:
|
|
|
|
beiscsi_log(phba, KERN_ERR, BEISCSI_LOG_INIT,
|
|
|
|
"BM_%d : AER EEH Resume Failed\n");
|
|
|
|
}
|
|
|
|
|
2012-12-22 04:08:55 +07:00
|
|
|
static int beiscsi_dev_probe(struct pci_dev *pcidev,
|
|
|
|
const struct pci_device_id *id)
|
2009-09-05 09:06:35 +07:00
|
|
|
{
|
|
|
|
struct beiscsi_hba *phba = NULL;
|
2009-10-23 13:22:33 +07:00
|
|
|
struct hwi_controller *phwi_ctrlr;
|
|
|
|
struct hwi_context_memory *phwi_context;
|
|
|
|
struct be_eq_obj *pbe_eq;
|
2016-08-19 16:50:12 +07:00
|
|
|
unsigned int s_handle;
|
2013-09-29 05:35:58 +07:00
|
|
|
int ret = 0, i;
|
2009-09-05 09:06:35 +07:00
|
|
|
|
|
|
|
ret = beiscsi_enable_pci(pcidev);
|
|
|
|
if (ret < 0) {
|
2012-08-21 00:30:18 +07:00
|
|
|
dev_err(&pcidev->dev,
|
|
|
|
"beiscsi_dev_probe - Failed to enable pci device\n");
|
2009-09-05 09:06:35 +07:00
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
|
|
|
phba = beiscsi_hba_alloc(pcidev);
|
|
|
|
if (!phba) {
|
2012-08-21 00:30:18 +07:00
|
|
|
dev_err(&pcidev->dev,
|
|
|
|
"beiscsi_dev_probe - Failed in beiscsi_hba_alloc\n");
|
2009-09-05 09:06:35 +07:00
|
|
|
goto disable_pci;
|
|
|
|
}
|
|
|
|
|
2013-09-29 05:35:58 +07:00
|
|
|
/* Enable EEH reporting */
|
|
|
|
ret = pci_enable_pcie_error_reporting(pcidev);
|
|
|
|
if (ret)
|
|
|
|
beiscsi_log(phba, KERN_WARNING, BEISCSI_LOG_INIT,
|
|
|
|
"BM_%d : PCIe Error Reporting "
|
|
|
|
"Enabling Failed\n");
|
|
|
|
|
|
|
|
pci_save_state(pcidev);
|
|
|
|
|
2012-08-21 00:30:18 +07:00
|
|
|
/* Initialize Driver configuration Paramters */
|
|
|
|
beiscsi_hba_attrs_init(phba);
|
|
|
|
|
2013-09-29 05:35:40 +07:00
|
|
|
phba->mac_addr_set = false;
|
2012-10-20 06:15:40 +07:00
|
|
|
|
2010-02-11 06:41:15 +07:00
|
|
|
switch (pcidev->device) {
|
|
|
|
case BE_DEVICE_ID1:
|
|
|
|
case OC_DEVICE_ID1:
|
|
|
|
case OC_DEVICE_ID2:
|
|
|
|
phba->generation = BE_GEN2;
|
2012-10-20 06:14:23 +07:00
|
|
|
phba->iotask_fn = beiscsi_iotask;
|
2010-02-11 06:41:15 +07:00
|
|
|
break;
|
|
|
|
case BE_DEVICE_ID2:
|
|
|
|
case OC_DEVICE_ID3:
|
|
|
|
phba->generation = BE_GEN3;
|
2012-10-20 06:14:23 +07:00
|
|
|
phba->iotask_fn = beiscsi_iotask;
|
2010-02-11 06:41:15 +07:00
|
|
|
break;
|
2012-10-20 06:13:20 +07:00
|
|
|
case OC_SKH_ID1:
|
|
|
|
phba->generation = BE_GEN4;
|
2012-10-20 06:14:23 +07:00
|
|
|
phba->iotask_fn = beiscsi_iotask_v2;
|
2013-04-06 10:38:24 +07:00
|
|
|
break;
|
2010-02-11 06:41:15 +07:00
|
|
|
default:
|
|
|
|
phba->generation = 0;
|
|
|
|
}
|
|
|
|
|
2009-09-05 09:06:35 +07:00
|
|
|
ret = be_ctrl_init(phba, pcidev);
|
|
|
|
if (ret) {
|
2012-08-21 00:30:18 +07:00
|
|
|
beiscsi_log(phba, KERN_ERR, BEISCSI_LOG_INIT,
|
|
|
|
"BM_%d : beiscsi_dev_probe-"
|
|
|
|
"Failed in be_ctrl_init\n");
|
2009-09-05 09:06:35 +07:00
|
|
|
goto hba_free;
|
|
|
|
}
|
|
|
|
|
2016-08-19 16:50:11 +07:00
|
|
|
set_bit(BEISCSI_HBA_RUNNING, &phba->state);
|
2016-01-20 15:40:51 +07:00
|
|
|
/*
|
|
|
|
* FUNCTION_RESET should clean up any stale info in FW for this fn
|
|
|
|
*/
|
2012-10-20 06:12:37 +07:00
|
|
|
ret = beiscsi_cmd_reset_function(phba);
|
|
|
|
if (ret) {
|
|
|
|
beiscsi_log(phba, KERN_ERR, BEISCSI_LOG_INIT,
|
2013-09-29 05:35:43 +07:00
|
|
|
"BM_%d : Reset Failed\n");
|
2012-10-20 06:12:37 +07:00
|
|
|
goto hba_free;
|
|
|
|
}
|
|
|
|
ret = be_chk_reset_complete(phba);
|
|
|
|
if (ret) {
|
|
|
|
beiscsi_log(phba, KERN_ERR, BEISCSI_LOG_INIT,
|
2013-09-29 05:35:43 +07:00
|
|
|
"BM_%d : Failed to get out of reset.\n");
|
2012-10-20 06:12:37 +07:00
|
|
|
goto hba_free;
|
2010-07-22 05:54:53 +07:00
|
|
|
}
|
|
|
|
|
2009-09-05 09:06:35 +07:00
|
|
|
spin_lock_init(&phba->io_sgl_lock);
|
|
|
|
spin_lock_init(&phba->mgmt_sgl_lock);
|
2013-09-29 05:35:42 +07:00
|
|
|
spin_lock_init(&phba->async_pdu_lock);
|
2010-01-05 06:34:12 +07:00
|
|
|
ret = mgmt_get_fw_config(&phba->ctrl, phba);
|
|
|
|
if (ret != 0) {
|
2012-08-21 00:30:18 +07:00
|
|
|
beiscsi_log(phba, KERN_ERR, BEISCSI_LOG_INIT,
|
|
|
|
"BM_%d : Error getting fw config\n");
|
2010-01-05 06:34:12 +07:00
|
|
|
goto free_port;
|
|
|
|
}
|
2016-01-20 15:40:53 +07:00
|
|
|
mgmt_get_port_name(&phba->ctrl, phba);
|
2016-01-20 15:40:54 +07:00
|
|
|
beiscsi_get_params(phba);
|
2016-08-19 16:50:14 +07:00
|
|
|
beiscsi_set_uer_feature(phba);
|
2013-09-29 05:35:54 +07:00
|
|
|
|
|
|
|
if (enable_msix)
|
|
|
|
find_num_cpus(phba);
|
|
|
|
else
|
|
|
|
phba->num_cpus = 1;
|
|
|
|
|
|
|
|
beiscsi_log(phba, KERN_INFO, BEISCSI_LOG_INIT,
|
|
|
|
"BM_%d : num_cpus = %d\n",
|
|
|
|
phba->num_cpus);
|
|
|
|
|
|
|
|
if (enable_msix) {
|
|
|
|
beiscsi_msix_enable(phba);
|
|
|
|
if (!phba->msix_enabled)
|
|
|
|
phba->num_cpus = 1;
|
|
|
|
}
|
|
|
|
|
2013-09-29 05:35:44 +07:00
|
|
|
phba->shost->max_id = phba->params.cxns_per_ctrl;
|
2010-01-05 06:42:03 +07:00
|
|
|
phba->shost->can_queue = phba->params.ios_per_ctrl;
|
2009-09-05 09:06:35 +07:00
|
|
|
ret = beiscsi_init_port(phba);
|
|
|
|
if (ret < 0) {
|
2012-08-21 00:30:18 +07:00
|
|
|
beiscsi_log(phba, KERN_ERR, BEISCSI_LOG_INIT,
|
|
|
|
"BM_%d : beiscsi_dev_probe-"
|
|
|
|
"Failed in beiscsi_init_port\n");
|
2009-09-05 09:06:35 +07:00
|
|
|
goto free_port;
|
|
|
|
}
|
|
|
|
|
2013-09-29 05:35:58 +07:00
|
|
|
for (i = 0; i < MAX_MCC_CMD; i++) {
|
2010-01-05 06:40:46 +07:00
|
|
|
init_waitqueue_head(&phba->ctrl.mcc_wait[i + 1]);
|
|
|
|
phba->ctrl.mcc_tag[i] = i + 1;
|
2016-02-04 17:19:10 +07:00
|
|
|
phba->ctrl.mcc_tag_status[i + 1] = 0;
|
2010-01-05 06:40:46 +07:00
|
|
|
phba->ctrl.mcc_tag_available++;
|
2014-01-29 14:16:39 +07:00
|
|
|
memset(&phba->ctrl.ptag_state[i].tag_mem_state, 0,
|
2014-05-06 08:41:28 +07:00
|
|
|
sizeof(struct be_dma_mem));
|
2010-01-05 06:40:46 +07:00
|
|
|
}
|
|
|
|
|
|
|
|
phba->ctrl.mcc_alloc_index = phba->ctrl.mcc_free_index = 0;
|
|
|
|
|
2012-10-20 06:12:49 +07:00
|
|
|
snprintf(phba->wq_name, sizeof(phba->wq_name), "beiscsi_%02x_wq",
|
2009-09-05 09:06:35 +07:00
|
|
|
phba->shost->host_no);
|
2013-07-04 05:04:57 +07:00
|
|
|
phba->wq = alloc_workqueue("%s", WQ_MEM_RECLAIM, 1, phba->wq_name);
|
2009-09-05 09:06:35 +07:00
|
|
|
if (!phba->wq) {
|
2012-08-21 00:30:18 +07:00
|
|
|
beiscsi_log(phba, KERN_ERR, BEISCSI_LOG_INIT,
|
|
|
|
"BM_%d : beiscsi_dev_probe-"
|
|
|
|
"Failed to allocate work queue\n");
|
2009-09-05 09:06:35 +07:00
|
|
|
goto free_twq;
|
|
|
|
}
|
|
|
|
|
2016-08-19 16:50:13 +07:00
|
|
|
INIT_DELAYED_WORK(&phba->eqd_update, beiscsi_eqd_update_work);
|
2009-09-05 09:06:35 +07:00
|
|
|
|
2009-10-23 13:22:33 +07:00
|
|
|
phwi_ctrlr = phba->phwi_ctrlr;
|
|
|
|
phwi_context = phwi_ctrlr->phwi_ctxt;
|
2012-10-20 06:12:49 +07:00
|
|
|
|
2014-03-13 22:38:42 +07:00
|
|
|
for (i = 0; i < phba->num_cpus; i++) {
|
2012-10-20 06:12:49 +07:00
|
|
|
pbe_eq = &phwi_context->be_eq[i];
|
2015-11-10 20:56:14 +07:00
|
|
|
irq_poll_init(&pbe_eq->iopoll, be_iopoll_budget,
|
2014-03-13 22:38:42 +07:00
|
|
|
be_iopoll);
|
2009-09-05 09:06:35 +07:00
|
|
|
}
|
2012-10-20 06:12:49 +07:00
|
|
|
|
2014-03-13 22:38:42 +07:00
|
|
|
i = (phba->msix_enabled) ? i : 0;
|
|
|
|
/* Work item for MCC handling */
|
|
|
|
pbe_eq = &phwi_context->be_eq[i];
|
2016-08-19 16:50:10 +07:00
|
|
|
INIT_WORK(&pbe_eq->mcc_work, beiscsi_mcc_work);
|
2014-03-13 22:38:42 +07:00
|
|
|
|
2009-09-05 09:06:35 +07:00
|
|
|
ret = beiscsi_init_irqs(phba);
|
|
|
|
if (ret < 0) {
|
2012-08-21 00:30:18 +07:00
|
|
|
beiscsi_log(phba, KERN_ERR, BEISCSI_LOG_INIT,
|
|
|
|
"BM_%d : beiscsi_dev_probe-"
|
|
|
|
"Failed to beiscsi_init_irqs\n");
|
2009-09-05 09:06:35 +07:00
|
|
|
goto free_blkenbld;
|
|
|
|
}
|
2010-07-22 05:53:22 +07:00
|
|
|
hwi_enable_intr(phba);
|
2011-06-25 03:11:53 +07:00
|
|
|
|
2014-05-06 08:41:25 +07:00
|
|
|
if (iscsi_host_add(phba->shost, &phba->pcidev->dev))
|
|
|
|
goto free_blkenbld;
|
|
|
|
|
2016-08-19 16:50:12 +07:00
|
|
|
INIT_WORK(&phba->boot_work, beiscsi_boot_work);
|
|
|
|
ret = beiscsi_boot_get_shandle(phba, &s_handle);
|
|
|
|
if (ret > 0) {
|
|
|
|
beiscsi_start_boot_work(phba, s_handle);
|
|
|
|
/**
|
|
|
|
* Set this bit after starting the work to let
|
|
|
|
* probe handle it first.
|
|
|
|
* ASYNC event can too schedule this work.
|
2011-06-25 03:11:53 +07:00
|
|
|
*/
|
2016-08-19 16:50:12 +07:00
|
|
|
set_bit(BEISCSI_HBA_BOOT_FOUND, &phba->state);
|
|
|
|
}
|
2011-06-25 03:11:53 +07:00
|
|
|
|
2016-08-19 16:50:06 +07:00
|
|
|
beiscsi_iface_create_default(phba);
|
2016-08-19 16:50:13 +07:00
|
|
|
schedule_delayed_work(&phba->eqd_update,
|
|
|
|
msecs_to_jiffies(BEISCSI_EQD_UPDATE_INTERVAL));
|
|
|
|
/**
|
|
|
|
* Start UE detection here. UE before this will cause stall in probe
|
|
|
|
* and eventually fail the probe.
|
|
|
|
*/
|
|
|
|
init_timer(&phba->hw_check);
|
|
|
|
phba->hw_check.function = beiscsi_hw_health_check;
|
|
|
|
phba->hw_check.data = (unsigned long)phba;
|
|
|
|
mod_timer(&phba->hw_check,
|
|
|
|
jiffies + msecs_to_jiffies(BEISCSI_UE_DETECT_INTERVAL));
|
2012-10-20 06:15:51 +07:00
|
|
|
|
2012-08-21 00:30:18 +07:00
|
|
|
beiscsi_log(phba, KERN_INFO, BEISCSI_LOG_INIT,
|
|
|
|
"\n\n\n BM_%d : SUCCESS - DRIVER LOADED\n\n\n");
|
2009-09-05 09:06:35 +07:00
|
|
|
return 0;
|
|
|
|
|
|
|
|
free_blkenbld:
|
|
|
|
destroy_workqueue(phba->wq);
|
2014-03-13 22:38:42 +07:00
|
|
|
for (i = 0; i < phba->num_cpus; i++) {
|
|
|
|
pbe_eq = &phwi_context->be_eq[i];
|
2015-11-10 20:56:14 +07:00
|
|
|
irq_poll_disable(&pbe_eq->iopoll);
|
2014-03-13 22:38:42 +07:00
|
|
|
}
|
2009-09-05 09:06:35 +07:00
|
|
|
free_twq:
|
|
|
|
beiscsi_clean_port(phba);
|
|
|
|
beiscsi_free_mem(phba);
|
|
|
|
free_port:
|
|
|
|
pci_free_consistent(phba->pcidev,
|
|
|
|
phba->ctrl.mbox_mem_alloced.size,
|
|
|
|
phba->ctrl.mbox_mem_alloced.va,
|
|
|
|
phba->ctrl.mbox_mem_alloced.dma);
|
|
|
|
beiscsi_unmap_pci_function(phba);
|
|
|
|
hba_free:
|
2010-07-22 05:53:22 +07:00
|
|
|
if (phba->msix_enabled)
|
|
|
|
pci_disable_msix(phba->pcidev);
|
2009-09-05 09:06:35 +07:00
|
|
|
pci_dev_put(phba->pcidev);
|
|
|
|
iscsi_host_free(phba->shost);
|
2015-02-12 08:15:47 +07:00
|
|
|
pci_set_drvdata(pcidev, NULL);
|
2009-09-05 09:06:35 +07:00
|
|
|
disable_pci:
|
2015-04-25 09:47:19 +07:00
|
|
|
pci_release_regions(pcidev);
|
2009-09-05 09:06:35 +07:00
|
|
|
pci_disable_device(pcidev);
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
2013-09-29 05:35:58 +07:00
|
|
|
static struct pci_error_handlers beiscsi_eeh_handlers = {
|
|
|
|
.error_detected = beiscsi_eeh_err_detected,
|
|
|
|
.slot_reset = beiscsi_eeh_reset,
|
|
|
|
.resume = beiscsi_eeh_resume,
|
|
|
|
};
|
|
|
|
|
2009-09-05 09:06:35 +07:00
|
|
|
struct iscsi_transport beiscsi_iscsi_transport = {
|
|
|
|
.owner = THIS_MODULE,
|
|
|
|
.name = DRV_NAME,
|
2010-01-05 06:42:43 +07:00
|
|
|
.caps = CAP_RECOVERY_L0 | CAP_HDRDGST | CAP_TEXT_NEGO |
|
2009-09-05 09:06:35 +07:00
|
|
|
CAP_MULTI_R2T | CAP_DATADGST | CAP_DATA_PATH_OFFLOAD,
|
|
|
|
.create_session = beiscsi_session_create,
|
|
|
|
.destroy_session = beiscsi_session_destroy,
|
|
|
|
.create_conn = beiscsi_conn_create,
|
|
|
|
.bind_conn = beiscsi_conn_bind,
|
|
|
|
.destroy_conn = iscsi_conn_teardown,
|
2016-08-19 16:50:06 +07:00
|
|
|
.attr_is_visible = beiscsi_attr_is_visible,
|
|
|
|
.set_iface_param = beiscsi_iface_set_param,
|
|
|
|
.get_iface_param = beiscsi_iface_get_param,
|
2009-09-05 09:06:35 +07:00
|
|
|
.set_param = beiscsi_set_param,
|
2011-02-17 04:04:41 +07:00
|
|
|
.get_conn_param = iscsi_conn_get_param,
|
2009-09-05 09:06:35 +07:00
|
|
|
.get_session_param = iscsi_session_get_param,
|
|
|
|
.get_host_param = beiscsi_get_host_param,
|
|
|
|
.start_conn = beiscsi_conn_start,
|
2010-06-09 15:30:08 +07:00
|
|
|
.stop_conn = iscsi_conn_stop,
|
2009-09-05 09:06:35 +07:00
|
|
|
.send_pdu = iscsi_conn_send_pdu,
|
|
|
|
.xmit_task = beiscsi_task_xmit,
|
|
|
|
.cleanup_task = beiscsi_cleanup_task,
|
|
|
|
.alloc_pdu = beiscsi_alloc_pdu,
|
|
|
|
.parse_pdu_itt = beiscsi_parse_pdu,
|
|
|
|
.get_stats = beiscsi_conn_get_stats,
|
2011-02-17 04:04:41 +07:00
|
|
|
.get_ep_param = beiscsi_ep_get_param,
|
2009-09-05 09:06:35 +07:00
|
|
|
.ep_connect = beiscsi_ep_connect,
|
|
|
|
.ep_poll = beiscsi_ep_poll,
|
|
|
|
.ep_disconnect = beiscsi_ep_disconnect,
|
|
|
|
.session_recovery_timedout = iscsi_session_recovery_timedout,
|
2012-04-04 11:41:50 +07:00
|
|
|
.bsg_request = beiscsi_bsg_request,
|
2009-09-05 09:06:35 +07:00
|
|
|
};
|
|
|
|
|
|
|
|
static struct pci_driver beiscsi_pci_driver = {
|
|
|
|
.name = DRV_NAME,
|
|
|
|
.probe = beiscsi_dev_probe,
|
|
|
|
.remove = beiscsi_remove,
|
2013-09-29 05:35:58 +07:00
|
|
|
.id_table = beiscsi_pci_id_table,
|
|
|
|
.err_handler = &beiscsi_eeh_handlers
|
2009-09-05 09:06:35 +07:00
|
|
|
};
|
|
|
|
|
2009-10-23 13:22:33 +07:00
|
|
|
|
2009-09-05 09:06:35 +07:00
|
|
|
static int __init beiscsi_module_init(void)
|
|
|
|
{
|
|
|
|
int ret;
|
|
|
|
|
|
|
|
beiscsi_scsi_transport =
|
|
|
|
iscsi_register_transport(&beiscsi_iscsi_transport);
|
|
|
|
if (!beiscsi_scsi_transport) {
|
2012-08-21 00:30:18 +07:00
|
|
|
printk(KERN_ERR
|
|
|
|
"beiscsi_module_init - Unable to register beiscsi transport.\n");
|
2010-01-23 07:07:40 +07:00
|
|
|
return -ENOMEM;
|
2009-09-05 09:06:35 +07:00
|
|
|
}
|
2012-08-21 00:30:18 +07:00
|
|
|
printk(KERN_INFO "In beiscsi_module_init, tt=%p\n",
|
|
|
|
&beiscsi_iscsi_transport);
|
2009-09-05 09:06:35 +07:00
|
|
|
|
|
|
|
ret = pci_register_driver(&beiscsi_pci_driver);
|
|
|
|
if (ret) {
|
2012-08-21 00:30:18 +07:00
|
|
|
printk(KERN_ERR
|
|
|
|
"beiscsi_module_init - Unable to register beiscsi pci driver.\n");
|
2009-09-05 09:06:35 +07:00
|
|
|
goto unregister_iscsi_transport;
|
|
|
|
}
|
|
|
|
return 0;
|
|
|
|
|
|
|
|
unregister_iscsi_transport:
|
|
|
|
iscsi_unregister_transport(&beiscsi_iscsi_transport);
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void __exit beiscsi_module_exit(void)
|
|
|
|
{
|
|
|
|
pci_unregister_driver(&beiscsi_pci_driver);
|
|
|
|
iscsi_unregister_transport(&beiscsi_iscsi_transport);
|
|
|
|
}
|
|
|
|
|
|
|
|
module_init(beiscsi_module_init);
|
|
|
|
module_exit(beiscsi_module_exit);
|