mirror of
https://github.com/AuxXxilium/linux_dsm_epyc7002.git
synced 2024-11-24 17:20:50 +07:00
72e39ea7e0
srp transport works for target drivers without supported_mode attribute but it would be better to use it explicitly. Signed-off-by: FUJITA Tomonori <fujita.tomonori@lab.ntt.co.jp> Signed-off-by: James Bottomley <James.Bottomley@SteelEye.com>
998 lines
24 KiB
C
998 lines
24 KiB
C
/*
|
|
* IBM eServer i/pSeries Virtual SCSI Target Driver
|
|
* Copyright (C) 2003-2005 Dave Boutcher (boutcher@us.ibm.com) IBM Corp.
|
|
* Santiago Leon (santil@us.ibm.com) IBM Corp.
|
|
* Linda Xie (lxie@us.ibm.com) IBM Corp.
|
|
*
|
|
* Copyright (C) 2005-2006 FUJITA Tomonori <tomof@acm.org>
|
|
*
|
|
* This program is free software; you can redistribute it and/or modify
|
|
* it under the terms of the GNU General Public License as published by
|
|
* the Free Software Foundation; either version 2 of the License, or
|
|
* (at your option) any later version.
|
|
*
|
|
* This program is distributed in the hope that it will be useful,
|
|
* but WITHOUT ANY WARRANTY; without even the implied warranty of
|
|
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
|
|
* GNU General Public License for more details.
|
|
*
|
|
* You should have received a copy of the GNU General Public License
|
|
* along with this program; if not, write to the Free Software
|
|
* Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307
|
|
* USA
|
|
*/
|
|
#include <linux/interrupt.h>
|
|
#include <linux/module.h>
|
|
#include <scsi/scsi.h>
|
|
#include <scsi/scsi_host.h>
|
|
#include <scsi/scsi_transport_srp.h>
|
|
#include <scsi/scsi_tgt.h>
|
|
#include <scsi/libsrp.h>
|
|
#include <asm/hvcall.h>
|
|
#include <asm/iommu.h>
|
|
#include <asm/prom.h>
|
|
#include <asm/vio.h>
|
|
|
|
#include "ibmvscsi.h"
|
|
|
|
#define INITIAL_SRP_LIMIT 16
|
|
#define DEFAULT_MAX_SECTORS 256
|
|
|
|
#define TGT_NAME "ibmvstgt"
|
|
|
|
/*
|
|
* Hypervisor calls.
|
|
*/
|
|
#define h_copy_rdma(l, sa, sb, da, db) \
|
|
plpar_hcall_norets(H_COPY_RDMA, l, sa, sb, da, db)
|
|
#define h_send_crq(ua, l, h) \
|
|
plpar_hcall_norets(H_SEND_CRQ, ua, l, h)
|
|
#define h_reg_crq(ua, tok, sz)\
|
|
plpar_hcall_norets(H_REG_CRQ, ua, tok, sz);
|
|
#define h_free_crq(ua) \
|
|
plpar_hcall_norets(H_FREE_CRQ, ua);
|
|
|
|
/* tmp - will replace with SCSI logging stuff */
|
|
#define eprintk(fmt, args...) \
|
|
do { \
|
|
printk("%s(%d) " fmt, __FUNCTION__, __LINE__, ##args); \
|
|
} while (0)
|
|
/* #define dprintk eprintk */
|
|
#define dprintk(fmt, args...)
|
|
|
|
struct vio_port {
|
|
struct vio_dev *dma_dev;
|
|
|
|
struct crq_queue crq_queue;
|
|
struct work_struct crq_work;
|
|
|
|
unsigned long liobn;
|
|
unsigned long riobn;
|
|
struct srp_target *target;
|
|
|
|
struct srp_rport *rport;
|
|
};
|
|
|
|
static struct workqueue_struct *vtgtd;
|
|
static struct scsi_transport_template *ibmvstgt_transport_template;
|
|
|
|
/*
|
|
* These are fixed for the system and come from the Open Firmware device tree.
|
|
* We just store them here to save getting them every time.
|
|
*/
|
|
static char system_id[64] = "";
|
|
static char partition_name[97] = "UNKNOWN";
|
|
static unsigned int partition_number = -1;
|
|
|
|
static struct vio_port *target_to_port(struct srp_target *target)
|
|
{
|
|
return (struct vio_port *) target->ldata;
|
|
}
|
|
|
|
static inline union viosrp_iu *vio_iu(struct iu_entry *iue)
|
|
{
|
|
return (union viosrp_iu *) (iue->sbuf->buf);
|
|
}
|
|
|
|
static int send_iu(struct iu_entry *iue, uint64_t length, uint8_t format)
|
|
{
|
|
struct srp_target *target = iue->target;
|
|
struct vio_port *vport = target_to_port(target);
|
|
long rc, rc1;
|
|
union {
|
|
struct viosrp_crq cooked;
|
|
uint64_t raw[2];
|
|
} crq;
|
|
|
|
/* First copy the SRP */
|
|
rc = h_copy_rdma(length, vport->liobn, iue->sbuf->dma,
|
|
vport->riobn, iue->remote_token);
|
|
|
|
if (rc)
|
|
eprintk("Error %ld transferring data\n", rc);
|
|
|
|
crq.cooked.valid = 0x80;
|
|
crq.cooked.format = format;
|
|
crq.cooked.reserved = 0x00;
|
|
crq.cooked.timeout = 0x00;
|
|
crq.cooked.IU_length = length;
|
|
crq.cooked.IU_data_ptr = vio_iu(iue)->srp.rsp.tag;
|
|
|
|
if (rc == 0)
|
|
crq.cooked.status = 0x99; /* Just needs to be non-zero */
|
|
else
|
|
crq.cooked.status = 0x00;
|
|
|
|
rc1 = h_send_crq(vport->dma_dev->unit_address, crq.raw[0], crq.raw[1]);
|
|
|
|
if (rc1) {
|
|
eprintk("%ld sending response\n", rc1);
|
|
return rc1;
|
|
}
|
|
|
|
return rc;
|
|
}
|
|
|
|
#define SRP_RSP_SENSE_DATA_LEN 18
|
|
|
|
static int send_rsp(struct iu_entry *iue, struct scsi_cmnd *sc,
|
|
unsigned char status, unsigned char asc)
|
|
{
|
|
union viosrp_iu *iu = vio_iu(iue);
|
|
uint64_t tag = iu->srp.rsp.tag;
|
|
|
|
/* If the linked bit is on and status is good */
|
|
if (test_bit(V_LINKED, &iue->flags) && (status == NO_SENSE))
|
|
status = 0x10;
|
|
|
|
memset(iu, 0, sizeof(struct srp_rsp));
|
|
iu->srp.rsp.opcode = SRP_RSP;
|
|
iu->srp.rsp.req_lim_delta = 1;
|
|
iu->srp.rsp.tag = tag;
|
|
|
|
if (test_bit(V_DIOVER, &iue->flags))
|
|
iu->srp.rsp.flags |= SRP_RSP_FLAG_DIOVER;
|
|
|
|
iu->srp.rsp.data_in_res_cnt = 0;
|
|
iu->srp.rsp.data_out_res_cnt = 0;
|
|
|
|
iu->srp.rsp.flags &= ~SRP_RSP_FLAG_RSPVALID;
|
|
|
|
iu->srp.rsp.resp_data_len = 0;
|
|
iu->srp.rsp.status = status;
|
|
if (status) {
|
|
uint8_t *sense = iu->srp.rsp.data;
|
|
|
|
if (sc) {
|
|
iu->srp.rsp.flags |= SRP_RSP_FLAG_SNSVALID;
|
|
iu->srp.rsp.sense_data_len = SCSI_SENSE_BUFFERSIZE;
|
|
memcpy(sense, sc->sense_buffer, SCSI_SENSE_BUFFERSIZE);
|
|
} else {
|
|
iu->srp.rsp.status = SAM_STAT_CHECK_CONDITION;
|
|
iu->srp.rsp.flags |= SRP_RSP_FLAG_SNSVALID;
|
|
iu->srp.rsp.sense_data_len = SRP_RSP_SENSE_DATA_LEN;
|
|
|
|
/* Valid bit and 'current errors' */
|
|
sense[0] = (0x1 << 7 | 0x70);
|
|
/* Sense key */
|
|
sense[2] = status;
|
|
/* Additional sense length */
|
|
sense[7] = 0xa; /* 10 bytes */
|
|
/* Additional sense code */
|
|
sense[12] = asc;
|
|
}
|
|
}
|
|
|
|
send_iu(iue, sizeof(iu->srp.rsp) + SRP_RSP_SENSE_DATA_LEN,
|
|
VIOSRP_SRP_FORMAT);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void handle_cmd_queue(struct srp_target *target)
|
|
{
|
|
struct Scsi_Host *shost = target->shost;
|
|
struct srp_rport *rport = target_to_port(target)->rport;
|
|
struct iu_entry *iue;
|
|
struct srp_cmd *cmd;
|
|
unsigned long flags;
|
|
int err;
|
|
|
|
retry:
|
|
spin_lock_irqsave(&target->lock, flags);
|
|
|
|
list_for_each_entry(iue, &target->cmd_queue, ilist) {
|
|
if (!test_and_set_bit(V_FLYING, &iue->flags)) {
|
|
spin_unlock_irqrestore(&target->lock, flags);
|
|
cmd = iue->sbuf->buf;
|
|
err = srp_cmd_queue(shost, cmd, iue,
|
|
(unsigned long)rport, 0);
|
|
if (err) {
|
|
eprintk("cannot queue cmd %p %d\n", cmd, err);
|
|
srp_iu_put(iue);
|
|
}
|
|
goto retry;
|
|
}
|
|
}
|
|
|
|
spin_unlock_irqrestore(&target->lock, flags);
|
|
}
|
|
|
|
static int ibmvstgt_rdma(struct scsi_cmnd *sc, struct scatterlist *sg, int nsg,
|
|
struct srp_direct_buf *md, int nmd,
|
|
enum dma_data_direction dir, unsigned int rest)
|
|
{
|
|
struct iu_entry *iue = (struct iu_entry *) sc->SCp.ptr;
|
|
struct srp_target *target = iue->target;
|
|
struct vio_port *vport = target_to_port(target);
|
|
dma_addr_t token;
|
|
long err;
|
|
unsigned int done = 0;
|
|
int i, sidx, soff;
|
|
|
|
sidx = soff = 0;
|
|
token = sg_dma_address(sg + sidx);
|
|
|
|
for (i = 0; i < nmd && rest; i++) {
|
|
unsigned int mdone, mlen;
|
|
|
|
mlen = min(rest, md[i].len);
|
|
for (mdone = 0; mlen;) {
|
|
int slen = min(sg_dma_len(sg + sidx) - soff, mlen);
|
|
|
|
if (dir == DMA_TO_DEVICE)
|
|
err = h_copy_rdma(slen,
|
|
vport->riobn,
|
|
md[i].va + mdone,
|
|
vport->liobn,
|
|
token + soff);
|
|
else
|
|
err = h_copy_rdma(slen,
|
|
vport->liobn,
|
|
token + soff,
|
|
vport->riobn,
|
|
md[i].va + mdone);
|
|
|
|
if (err != H_SUCCESS) {
|
|
eprintk("rdma error %d %d %ld\n", dir, slen, err);
|
|
return -EIO;
|
|
}
|
|
|
|
mlen -= slen;
|
|
mdone += slen;
|
|
soff += slen;
|
|
done += slen;
|
|
|
|
if (soff == sg_dma_len(sg + sidx)) {
|
|
sidx++;
|
|
soff = 0;
|
|
token = sg_dma_address(sg + sidx);
|
|
|
|
if (sidx > nsg) {
|
|
eprintk("out of sg %p %d %d\n",
|
|
iue, sidx, nsg);
|
|
return -EIO;
|
|
}
|
|
}
|
|
};
|
|
|
|
rest -= mlen;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static int ibmvstgt_cmd_done(struct scsi_cmnd *sc,
|
|
void (*done)(struct scsi_cmnd *))
|
|
{
|
|
unsigned long flags;
|
|
struct iu_entry *iue = (struct iu_entry *) sc->SCp.ptr;
|
|
struct srp_target *target = iue->target;
|
|
int err = 0;
|
|
|
|
dprintk("%p %p %x %u\n", iue, target, vio_iu(iue)->srp.cmd.cdb[0],
|
|
cmd->usg_sg);
|
|
|
|
if (sc->use_sg)
|
|
err = srp_transfer_data(sc, &vio_iu(iue)->srp.cmd, ibmvstgt_rdma, 1, 1);
|
|
|
|
spin_lock_irqsave(&target->lock, flags);
|
|
list_del(&iue->ilist);
|
|
spin_unlock_irqrestore(&target->lock, flags);
|
|
|
|
if (err|| sc->result != SAM_STAT_GOOD) {
|
|
eprintk("operation failed %p %d %x\n",
|
|
iue, sc->result, vio_iu(iue)->srp.cmd.cdb[0]);
|
|
send_rsp(iue, sc, HARDWARE_ERROR, 0x00);
|
|
} else
|
|
send_rsp(iue, sc, NO_SENSE, 0x00);
|
|
|
|
done(sc);
|
|
srp_iu_put(iue);
|
|
return 0;
|
|
}
|
|
|
|
int send_adapter_info(struct iu_entry *iue,
|
|
dma_addr_t remote_buffer, uint16_t length)
|
|
{
|
|
struct srp_target *target = iue->target;
|
|
struct vio_port *vport = target_to_port(target);
|
|
struct Scsi_Host *shost = target->shost;
|
|
dma_addr_t data_token;
|
|
struct mad_adapter_info_data *info;
|
|
int err;
|
|
|
|
info = dma_alloc_coherent(target->dev, sizeof(*info), &data_token,
|
|
GFP_KERNEL);
|
|
if (!info) {
|
|
eprintk("bad dma_alloc_coherent %p\n", target);
|
|
return 1;
|
|
}
|
|
|
|
/* Get remote info */
|
|
err = h_copy_rdma(sizeof(*info), vport->riobn, remote_buffer,
|
|
vport->liobn, data_token);
|
|
if (err == H_SUCCESS) {
|
|
dprintk("Client connect: %s (%d)\n",
|
|
info->partition_name, info->partition_number);
|
|
}
|
|
|
|
memset(info, 0, sizeof(*info));
|
|
|
|
strcpy(info->srp_version, "16.a");
|
|
strncpy(info->partition_name, partition_name,
|
|
sizeof(info->partition_name));
|
|
info->partition_number = partition_number;
|
|
info->mad_version = 1;
|
|
info->os_type = 2;
|
|
info->port_max_txu[0] = shost->hostt->max_sectors << 9;
|
|
|
|
/* Send our info to remote */
|
|
err = h_copy_rdma(sizeof(*info), vport->liobn, data_token,
|
|
vport->riobn, remote_buffer);
|
|
|
|
dma_free_coherent(target->dev, sizeof(*info), info, data_token);
|
|
|
|
if (err != H_SUCCESS) {
|
|
eprintk("Error sending adapter info %d\n", err);
|
|
return 1;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static void process_login(struct iu_entry *iue)
|
|
{
|
|
union viosrp_iu *iu = vio_iu(iue);
|
|
struct srp_login_rsp *rsp = &iu->srp.login_rsp;
|
|
uint64_t tag = iu->srp.rsp.tag;
|
|
struct Scsi_Host *shost = iue->target->shost;
|
|
struct srp_target *target = host_to_srp_target(shost);
|
|
struct vio_port *vport = target_to_port(target);
|
|
struct srp_rport_identifiers ids;
|
|
|
|
memset(&ids, 0, sizeof(ids));
|
|
sprintf(ids.port_id, "%x", vport->dma_dev->unit_address);
|
|
ids.roles = SRP_RPORT_ROLE_INITIATOR;
|
|
if (!vport->rport)
|
|
vport->rport = srp_rport_add(shost, &ids);
|
|
|
|
/* TODO handle case that requested size is wrong and
|
|
* buffer format is wrong
|
|
*/
|
|
memset(iu, 0, sizeof(struct srp_login_rsp));
|
|
rsp->opcode = SRP_LOGIN_RSP;
|
|
rsp->req_lim_delta = INITIAL_SRP_LIMIT;
|
|
rsp->tag = tag;
|
|
rsp->max_it_iu_len = sizeof(union srp_iu);
|
|
rsp->max_ti_iu_len = sizeof(union srp_iu);
|
|
/* direct and indirect */
|
|
rsp->buf_fmt = SRP_BUF_FORMAT_DIRECT | SRP_BUF_FORMAT_INDIRECT;
|
|
|
|
send_iu(iue, sizeof(*rsp), VIOSRP_SRP_FORMAT);
|
|
}
|
|
|
|
static inline void queue_cmd(struct iu_entry *iue)
|
|
{
|
|
struct srp_target *target = iue->target;
|
|
unsigned long flags;
|
|
|
|
spin_lock_irqsave(&target->lock, flags);
|
|
list_add_tail(&iue->ilist, &target->cmd_queue);
|
|
spin_unlock_irqrestore(&target->lock, flags);
|
|
}
|
|
|
|
static int process_tsk_mgmt(struct iu_entry *iue)
|
|
{
|
|
union viosrp_iu *iu = vio_iu(iue);
|
|
int fn;
|
|
|
|
dprintk("%p %u\n", iue, iu->srp.tsk_mgmt.tsk_mgmt_func);
|
|
|
|
switch (iu->srp.tsk_mgmt.tsk_mgmt_func) {
|
|
case SRP_TSK_ABORT_TASK:
|
|
fn = ABORT_TASK;
|
|
break;
|
|
case SRP_TSK_ABORT_TASK_SET:
|
|
fn = ABORT_TASK_SET;
|
|
break;
|
|
case SRP_TSK_CLEAR_TASK_SET:
|
|
fn = CLEAR_TASK_SET;
|
|
break;
|
|
case SRP_TSK_LUN_RESET:
|
|
fn = LOGICAL_UNIT_RESET;
|
|
break;
|
|
case SRP_TSK_CLEAR_ACA:
|
|
fn = CLEAR_ACA;
|
|
break;
|
|
default:
|
|
fn = 0;
|
|
}
|
|
if (fn)
|
|
scsi_tgt_tsk_mgmt_request(iue->target->shost,
|
|
(unsigned long)iue->target->shost,
|
|
fn,
|
|
iu->srp.tsk_mgmt.task_tag,
|
|
(struct scsi_lun *) &iu->srp.tsk_mgmt.lun,
|
|
iue);
|
|
else
|
|
send_rsp(iue, NULL, ILLEGAL_REQUEST, 0x20);
|
|
|
|
return !fn;
|
|
}
|
|
|
|
static int process_mad_iu(struct iu_entry *iue)
|
|
{
|
|
union viosrp_iu *iu = vio_iu(iue);
|
|
struct viosrp_adapter_info *info;
|
|
struct viosrp_host_config *conf;
|
|
|
|
switch (iu->mad.empty_iu.common.type) {
|
|
case VIOSRP_EMPTY_IU_TYPE:
|
|
eprintk("%s\n", "Unsupported EMPTY MAD IU");
|
|
break;
|
|
case VIOSRP_ERROR_LOG_TYPE:
|
|
eprintk("%s\n", "Unsupported ERROR LOG MAD IU");
|
|
iu->mad.error_log.common.status = 1;
|
|
send_iu(iue, sizeof(iu->mad.error_log), VIOSRP_MAD_FORMAT);
|
|
break;
|
|
case VIOSRP_ADAPTER_INFO_TYPE:
|
|
info = &iu->mad.adapter_info;
|
|
info->common.status = send_adapter_info(iue, info->buffer,
|
|
info->common.length);
|
|
send_iu(iue, sizeof(*info), VIOSRP_MAD_FORMAT);
|
|
break;
|
|
case VIOSRP_HOST_CONFIG_TYPE:
|
|
conf = &iu->mad.host_config;
|
|
conf->common.status = 1;
|
|
send_iu(iue, sizeof(*conf), VIOSRP_MAD_FORMAT);
|
|
break;
|
|
default:
|
|
eprintk("Unknown type %u\n", iu->srp.rsp.opcode);
|
|
}
|
|
|
|
return 1;
|
|
}
|
|
|
|
static int process_srp_iu(struct iu_entry *iue)
|
|
{
|
|
union viosrp_iu *iu = vio_iu(iue);
|
|
int done = 1;
|
|
u8 opcode = iu->srp.rsp.opcode;
|
|
|
|
switch (opcode) {
|
|
case SRP_LOGIN_REQ:
|
|
process_login(iue);
|
|
break;
|
|
case SRP_TSK_MGMT:
|
|
done = process_tsk_mgmt(iue);
|
|
break;
|
|
case SRP_CMD:
|
|
queue_cmd(iue);
|
|
done = 0;
|
|
break;
|
|
case SRP_LOGIN_RSP:
|
|
case SRP_I_LOGOUT:
|
|
case SRP_T_LOGOUT:
|
|
case SRP_RSP:
|
|
case SRP_CRED_REQ:
|
|
case SRP_CRED_RSP:
|
|
case SRP_AER_REQ:
|
|
case SRP_AER_RSP:
|
|
eprintk("Unsupported type %u\n", opcode);
|
|
break;
|
|
default:
|
|
eprintk("Unknown type %u\n", opcode);
|
|
}
|
|
|
|
return done;
|
|
}
|
|
|
|
static void process_iu(struct viosrp_crq *crq, struct srp_target *target)
|
|
{
|
|
struct vio_port *vport = target_to_port(target);
|
|
struct iu_entry *iue;
|
|
long err;
|
|
int done = 1;
|
|
|
|
iue = srp_iu_get(target);
|
|
if (!iue) {
|
|
eprintk("Error getting IU from pool, %p\n", target);
|
|
return;
|
|
}
|
|
|
|
iue->remote_token = crq->IU_data_ptr;
|
|
|
|
err = h_copy_rdma(crq->IU_length, vport->riobn,
|
|
iue->remote_token, vport->liobn, iue->sbuf->dma);
|
|
|
|
if (err != H_SUCCESS) {
|
|
eprintk("%ld transferring data error %p\n", err, iue);
|
|
goto out;
|
|
}
|
|
|
|
if (crq->format == VIOSRP_MAD_FORMAT)
|
|
done = process_mad_iu(iue);
|
|
else
|
|
done = process_srp_iu(iue);
|
|
out:
|
|
if (done)
|
|
srp_iu_put(iue);
|
|
}
|
|
|
|
static irqreturn_t ibmvstgt_interrupt(int irq, void *data)
|
|
{
|
|
struct srp_target *target = (struct srp_target *) data;
|
|
struct vio_port *vport = target_to_port(target);
|
|
|
|
vio_disable_interrupts(vport->dma_dev);
|
|
queue_work(vtgtd, &vport->crq_work);
|
|
|
|
return IRQ_HANDLED;
|
|
}
|
|
|
|
static int crq_queue_create(struct crq_queue *queue, struct srp_target *target)
|
|
{
|
|
int err;
|
|
struct vio_port *vport = target_to_port(target);
|
|
|
|
queue->msgs = (struct viosrp_crq *) get_zeroed_page(GFP_KERNEL);
|
|
if (!queue->msgs)
|
|
goto malloc_failed;
|
|
queue->size = PAGE_SIZE / sizeof(*queue->msgs);
|
|
|
|
queue->msg_token = dma_map_single(target->dev, queue->msgs,
|
|
queue->size * sizeof(*queue->msgs),
|
|
DMA_BIDIRECTIONAL);
|
|
|
|
if (dma_mapping_error(queue->msg_token))
|
|
goto map_failed;
|
|
|
|
err = h_reg_crq(vport->dma_dev->unit_address, queue->msg_token,
|
|
PAGE_SIZE);
|
|
|
|
/* If the adapter was left active for some reason (like kexec)
|
|
* try freeing and re-registering
|
|
*/
|
|
if (err == H_RESOURCE) {
|
|
do {
|
|
err = h_free_crq(vport->dma_dev->unit_address);
|
|
} while (err == H_BUSY || H_IS_LONG_BUSY(err));
|
|
|
|
err = h_reg_crq(vport->dma_dev->unit_address, queue->msg_token,
|
|
PAGE_SIZE);
|
|
}
|
|
|
|
if (err != H_SUCCESS && err != 2) {
|
|
eprintk("Error 0x%x opening virtual adapter\n", err);
|
|
goto reg_crq_failed;
|
|
}
|
|
|
|
err = request_irq(vport->dma_dev->irq, &ibmvstgt_interrupt,
|
|
IRQF_DISABLED, "ibmvstgt", target);
|
|
if (err)
|
|
goto req_irq_failed;
|
|
|
|
vio_enable_interrupts(vport->dma_dev);
|
|
|
|
h_send_crq(vport->dma_dev->unit_address, 0xC001000000000000, 0);
|
|
|
|
queue->cur = 0;
|
|
spin_lock_init(&queue->lock);
|
|
|
|
return 0;
|
|
|
|
req_irq_failed:
|
|
do {
|
|
err = h_free_crq(vport->dma_dev->unit_address);
|
|
} while (err == H_BUSY || H_IS_LONG_BUSY(err));
|
|
|
|
reg_crq_failed:
|
|
dma_unmap_single(target->dev, queue->msg_token,
|
|
queue->size * sizeof(*queue->msgs), DMA_BIDIRECTIONAL);
|
|
map_failed:
|
|
free_page((unsigned long) queue->msgs);
|
|
|
|
malloc_failed:
|
|
return -ENOMEM;
|
|
}
|
|
|
|
static void crq_queue_destroy(struct srp_target *target)
|
|
{
|
|
struct vio_port *vport = target_to_port(target);
|
|
struct crq_queue *queue = &vport->crq_queue;
|
|
int err;
|
|
|
|
free_irq(vport->dma_dev->irq, target);
|
|
do {
|
|
err = h_free_crq(vport->dma_dev->unit_address);
|
|
} while (err == H_BUSY || H_IS_LONG_BUSY(err));
|
|
|
|
dma_unmap_single(target->dev, queue->msg_token,
|
|
queue->size * sizeof(*queue->msgs), DMA_BIDIRECTIONAL);
|
|
|
|
free_page((unsigned long) queue->msgs);
|
|
}
|
|
|
|
static void process_crq(struct viosrp_crq *crq, struct srp_target *target)
|
|
{
|
|
struct vio_port *vport = target_to_port(target);
|
|
dprintk("%x %x\n", crq->valid, crq->format);
|
|
|
|
switch (crq->valid) {
|
|
case 0xC0:
|
|
/* initialization */
|
|
switch (crq->format) {
|
|
case 0x01:
|
|
h_send_crq(vport->dma_dev->unit_address,
|
|
0xC002000000000000, 0);
|
|
break;
|
|
case 0x02:
|
|
break;
|
|
default:
|
|
eprintk("Unknown format %u\n", crq->format);
|
|
}
|
|
break;
|
|
case 0xFF:
|
|
/* transport event */
|
|
break;
|
|
case 0x80:
|
|
/* real payload */
|
|
switch (crq->format) {
|
|
case VIOSRP_SRP_FORMAT:
|
|
case VIOSRP_MAD_FORMAT:
|
|
process_iu(crq, target);
|
|
break;
|
|
case VIOSRP_OS400_FORMAT:
|
|
case VIOSRP_AIX_FORMAT:
|
|
case VIOSRP_LINUX_FORMAT:
|
|
case VIOSRP_INLINE_FORMAT:
|
|
eprintk("Unsupported format %u\n", crq->format);
|
|
break;
|
|
default:
|
|
eprintk("Unknown format %u\n", crq->format);
|
|
}
|
|
break;
|
|
default:
|
|
eprintk("unknown message type 0x%02x!?\n", crq->valid);
|
|
}
|
|
}
|
|
|
|
static inline struct viosrp_crq *next_crq(struct crq_queue *queue)
|
|
{
|
|
struct viosrp_crq *crq;
|
|
unsigned long flags;
|
|
|
|
spin_lock_irqsave(&queue->lock, flags);
|
|
crq = &queue->msgs[queue->cur];
|
|
if (crq->valid & 0x80) {
|
|
if (++queue->cur == queue->size)
|
|
queue->cur = 0;
|
|
} else
|
|
crq = NULL;
|
|
spin_unlock_irqrestore(&queue->lock, flags);
|
|
|
|
return crq;
|
|
}
|
|
|
|
static void handle_crq(struct work_struct *work)
|
|
{
|
|
struct vio_port *vport = container_of(work, struct vio_port, crq_work);
|
|
struct srp_target *target = vport->target;
|
|
struct viosrp_crq *crq;
|
|
int done = 0;
|
|
|
|
while (!done) {
|
|
while ((crq = next_crq(&vport->crq_queue)) != NULL) {
|
|
process_crq(crq, target);
|
|
crq->valid = 0x00;
|
|
}
|
|
|
|
vio_enable_interrupts(vport->dma_dev);
|
|
|
|
crq = next_crq(&vport->crq_queue);
|
|
if (crq) {
|
|
vio_disable_interrupts(vport->dma_dev);
|
|
process_crq(crq, target);
|
|
crq->valid = 0x00;
|
|
} else
|
|
done = 1;
|
|
}
|
|
|
|
handle_cmd_queue(target);
|
|
}
|
|
|
|
|
|
static int ibmvstgt_eh_abort_handler(struct scsi_cmnd *sc)
|
|
{
|
|
unsigned long flags;
|
|
struct iu_entry *iue = (struct iu_entry *) sc->SCp.ptr;
|
|
struct srp_target *target = iue->target;
|
|
|
|
dprintk("%p %p %x\n", iue, target, vio_iu(iue)->srp.cmd.cdb[0]);
|
|
|
|
spin_lock_irqsave(&target->lock, flags);
|
|
list_del(&iue->ilist);
|
|
spin_unlock_irqrestore(&target->lock, flags);
|
|
|
|
srp_iu_put(iue);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int ibmvstgt_tsk_mgmt_response(struct Scsi_Host *shost,
|
|
u64 itn_id, u64 mid, int result)
|
|
{
|
|
struct iu_entry *iue = (struct iu_entry *) ((void *) mid);
|
|
union viosrp_iu *iu = vio_iu(iue);
|
|
unsigned char status, asc;
|
|
|
|
eprintk("%p %d\n", iue, result);
|
|
status = NO_SENSE;
|
|
asc = 0;
|
|
|
|
switch (iu->srp.tsk_mgmt.tsk_mgmt_func) {
|
|
case SRP_TSK_ABORT_TASK:
|
|
asc = 0x14;
|
|
if (result)
|
|
status = ABORTED_COMMAND;
|
|
break;
|
|
default:
|
|
break;
|
|
}
|
|
|
|
send_rsp(iue, NULL, status, asc);
|
|
srp_iu_put(iue);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int ibmvstgt_it_nexus_response(struct Scsi_Host *shost, u64 itn_id,
|
|
int result)
|
|
{
|
|
struct srp_target *target = host_to_srp_target(shost);
|
|
struct vio_port *vport = target_to_port(target);
|
|
|
|
if (result) {
|
|
eprintk("%p %d\n", shost, result);
|
|
srp_rport_del(vport->rport);
|
|
vport->rport = NULL;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static ssize_t system_id_show(struct class_device *cdev, char *buf)
|
|
{
|
|
return snprintf(buf, PAGE_SIZE, "%s\n", system_id);
|
|
}
|
|
|
|
static ssize_t partition_number_show(struct class_device *cdev, char *buf)
|
|
{
|
|
return snprintf(buf, PAGE_SIZE, "%x\n", partition_number);
|
|
}
|
|
|
|
static ssize_t unit_address_show(struct class_device *cdev, char *buf)
|
|
{
|
|
struct Scsi_Host *shost = class_to_shost(cdev);
|
|
struct srp_target *target = host_to_srp_target(shost);
|
|
struct vio_port *vport = target_to_port(target);
|
|
return snprintf(buf, PAGE_SIZE, "%x\n", vport->dma_dev->unit_address);
|
|
}
|
|
|
|
static CLASS_DEVICE_ATTR(system_id, S_IRUGO, system_id_show, NULL);
|
|
static CLASS_DEVICE_ATTR(partition_number, S_IRUGO, partition_number_show, NULL);
|
|
static CLASS_DEVICE_ATTR(unit_address, S_IRUGO, unit_address_show, NULL);
|
|
|
|
static struct class_device_attribute *ibmvstgt_attrs[] = {
|
|
&class_device_attr_system_id,
|
|
&class_device_attr_partition_number,
|
|
&class_device_attr_unit_address,
|
|
NULL,
|
|
};
|
|
|
|
static struct scsi_host_template ibmvstgt_sht = {
|
|
.name = TGT_NAME,
|
|
.module = THIS_MODULE,
|
|
.can_queue = INITIAL_SRP_LIMIT,
|
|
.sg_tablesize = SG_ALL,
|
|
.use_clustering = DISABLE_CLUSTERING,
|
|
.max_sectors = DEFAULT_MAX_SECTORS,
|
|
.transfer_response = ibmvstgt_cmd_done,
|
|
.eh_abort_handler = ibmvstgt_eh_abort_handler,
|
|
.shost_attrs = ibmvstgt_attrs,
|
|
.proc_name = TGT_NAME,
|
|
.supported_mode = MODE_TARGET,
|
|
};
|
|
|
|
static int ibmvstgt_probe(struct vio_dev *dev, const struct vio_device_id *id)
|
|
{
|
|
struct Scsi_Host *shost;
|
|
struct srp_target *target;
|
|
struct vio_port *vport;
|
|
unsigned int *dma, dma_size;
|
|
int err = -ENOMEM;
|
|
|
|
vport = kzalloc(sizeof(struct vio_port), GFP_KERNEL);
|
|
if (!vport)
|
|
return err;
|
|
shost = scsi_host_alloc(&ibmvstgt_sht, sizeof(struct srp_target));
|
|
if (!shost)
|
|
goto free_vport;
|
|
shost->transportt = ibmvstgt_transport_template;
|
|
err = scsi_tgt_alloc_queue(shost);
|
|
if (err)
|
|
goto put_host;
|
|
|
|
target = host_to_srp_target(shost);
|
|
target->shost = shost;
|
|
vport->dma_dev = dev;
|
|
target->ldata = vport;
|
|
vport->target = target;
|
|
err = srp_target_alloc(target, &dev->dev, INITIAL_SRP_LIMIT,
|
|
SRP_MAX_IU_LEN);
|
|
if (err)
|
|
goto put_host;
|
|
|
|
dma = (unsigned int *) vio_get_attribute(dev, "ibm,my-dma-window",
|
|
&dma_size);
|
|
if (!dma || dma_size != 40) {
|
|
eprintk("Couldn't get window property %d\n", dma_size);
|
|
err = -EIO;
|
|
goto free_srp_target;
|
|
}
|
|
vport->liobn = dma[0];
|
|
vport->riobn = dma[5];
|
|
|
|
INIT_WORK(&vport->crq_work, handle_crq);
|
|
|
|
err = crq_queue_create(&vport->crq_queue, target);
|
|
if (err)
|
|
goto free_srp_target;
|
|
|
|
err = scsi_add_host(shost, target->dev);
|
|
if (err)
|
|
goto destroy_queue;
|
|
|
|
return 0;
|
|
destroy_queue:
|
|
crq_queue_destroy(target);
|
|
free_srp_target:
|
|
srp_target_free(target);
|
|
put_host:
|
|
scsi_host_put(shost);
|
|
free_vport:
|
|
kfree(vport);
|
|
return err;
|
|
}
|
|
|
|
static int ibmvstgt_remove(struct vio_dev *dev)
|
|
{
|
|
struct srp_target *target = (struct srp_target *) dev->dev.driver_data;
|
|
struct Scsi_Host *shost = target->shost;
|
|
struct vio_port *vport = target->ldata;
|
|
|
|
crq_queue_destroy(target);
|
|
srp_remove_host(shost);
|
|
scsi_remove_host(shost);
|
|
scsi_tgt_free_queue(shost);
|
|
srp_target_free(target);
|
|
kfree(vport);
|
|
scsi_host_put(shost);
|
|
return 0;
|
|
}
|
|
|
|
static struct vio_device_id ibmvstgt_device_table[] __devinitdata = {
|
|
{"v-scsi-host", "IBM,v-scsi-host"},
|
|
{"",""}
|
|
};
|
|
|
|
MODULE_DEVICE_TABLE(vio, ibmvstgt_device_table);
|
|
|
|
static struct vio_driver ibmvstgt_driver = {
|
|
.id_table = ibmvstgt_device_table,
|
|
.probe = ibmvstgt_probe,
|
|
.remove = ibmvstgt_remove,
|
|
.driver = {
|
|
.name = "ibmvscsis",
|
|
.owner = THIS_MODULE,
|
|
}
|
|
};
|
|
|
|
static int get_system_info(void)
|
|
{
|
|
struct device_node *rootdn;
|
|
const char *id, *model, *name;
|
|
const unsigned int *num;
|
|
|
|
rootdn = of_find_node_by_path("/");
|
|
if (!rootdn)
|
|
return -ENOENT;
|
|
|
|
model = of_get_property(rootdn, "model", NULL);
|
|
id = of_get_property(rootdn, "system-id", NULL);
|
|
if (model && id)
|
|
snprintf(system_id, sizeof(system_id), "%s-%s", model, id);
|
|
|
|
name = of_get_property(rootdn, "ibm,partition-name", NULL);
|
|
if (name)
|
|
strncpy(partition_name, name, sizeof(partition_name));
|
|
|
|
num = of_get_property(rootdn, "ibm,partition-no", NULL);
|
|
if (num)
|
|
partition_number = *num;
|
|
|
|
of_node_put(rootdn);
|
|
return 0;
|
|
}
|
|
|
|
static struct srp_function_template ibmvstgt_transport_functions = {
|
|
.tsk_mgmt_response = ibmvstgt_tsk_mgmt_response,
|
|
.it_nexus_response = ibmvstgt_it_nexus_response,
|
|
};
|
|
|
|
static int ibmvstgt_init(void)
|
|
{
|
|
int err = -ENOMEM;
|
|
|
|
printk("IBM eServer i/pSeries Virtual SCSI Target Driver\n");
|
|
|
|
ibmvstgt_transport_template =
|
|
srp_attach_transport(&ibmvstgt_transport_functions);
|
|
if (!ibmvstgt_transport_template)
|
|
return err;
|
|
|
|
vtgtd = create_workqueue("ibmvtgtd");
|
|
if (!vtgtd)
|
|
goto release_transport;
|
|
|
|
err = get_system_info();
|
|
if (err)
|
|
goto destroy_wq;
|
|
|
|
err = vio_register_driver(&ibmvstgt_driver);
|
|
if (err)
|
|
goto destroy_wq;
|
|
|
|
return 0;
|
|
destroy_wq:
|
|
destroy_workqueue(vtgtd);
|
|
release_transport:
|
|
srp_release_transport(ibmvstgt_transport_template);
|
|
return err;
|
|
}
|
|
|
|
static void ibmvstgt_exit(void)
|
|
{
|
|
printk("Unregister IBM virtual SCSI driver\n");
|
|
|
|
destroy_workqueue(vtgtd);
|
|
vio_unregister_driver(&ibmvstgt_driver);
|
|
srp_release_transport(ibmvstgt_transport_template);
|
|
}
|
|
|
|
MODULE_DESCRIPTION("IBM Virtual SCSI Target");
|
|
MODULE_AUTHOR("Santiago Leon");
|
|
MODULE_LICENSE("GPL");
|
|
|
|
module_init(ibmvstgt_init);
|
|
module_exit(ibmvstgt_exit);
|