mirror of
https://github.com/AuxXxilium/linux_dsm_epyc7002.git
synced 2024-12-02 05:36:43 +07:00
ddccf307a3
I increased the delay step by step until loading of mvsas reliably detected the drive 200 times in sequence. A much better approach would be to monitor the hardware for some flag which indicates that port detection has finished, but I do not have any hardware documentation. Signed-off-by: Reinhard Nissl <rnissl@gmx.de> Cc: Ke Wei <kewei@marvell.com> Signed-off-by: James Bottomley <James.Bottomley@HansenPartnership.com>
3223 lines
86 KiB
C
3223 lines
86 KiB
C
/*
|
|
mvsas.c - Marvell 88SE6440 SAS/SATA support
|
|
|
|
Copyright 2007 Red Hat, Inc.
|
|
Copyright 2008 Marvell. <kewei@marvell.com>
|
|
|
|
This program is free software; you can redistribute it and/or
|
|
modify it under the terms of the GNU General Public License as
|
|
published by the Free Software Foundation; either version 2,
|
|
or (at your option) any later version.
|
|
|
|
This program is distributed in the hope that it will be useful,
|
|
but WITHOUT ANY WARRANTY; without even the implied warranty
|
|
of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.
|
|
See the GNU General Public License for more details.
|
|
|
|
You should have received a copy of the GNU General Public
|
|
License along with this program; see the file COPYING. If not,
|
|
write to the Free Software Foundation, 675 Mass Ave, Cambridge,
|
|
MA 02139, USA.
|
|
|
|
---------------------------------------------------------------
|
|
|
|
Random notes:
|
|
* hardware supports controlling the endian-ness of data
|
|
structures. this permits elimination of all the le32_to_cpu()
|
|
and cpu_to_le32() conversions.
|
|
|
|
*/
|
|
|
|
#include <linux/kernel.h>
|
|
#include <linux/module.h>
|
|
#include <linux/pci.h>
|
|
#include <linux/interrupt.h>
|
|
#include <linux/spinlock.h>
|
|
#include <linux/delay.h>
|
|
#include <linux/dma-mapping.h>
|
|
#include <linux/ctype.h>
|
|
#include <scsi/libsas.h>
|
|
#include <scsi/scsi_tcq.h>
|
|
#include <scsi/sas_ata.h>
|
|
#include <asm/io.h>
|
|
|
|
#define DRV_NAME "mvsas"
|
|
#define DRV_VERSION "0.5.2"
|
|
#define _MV_DUMP 0
|
|
#define MVS_DISABLE_NVRAM
|
|
#define MVS_DISABLE_MSI
|
|
|
|
#define mr32(reg) readl(regs + MVS_##reg)
|
|
#define mw32(reg,val) writel((val), regs + MVS_##reg)
|
|
#define mw32_f(reg,val) do { \
|
|
writel((val), regs + MVS_##reg); \
|
|
readl(regs + MVS_##reg); \
|
|
} while (0)
|
|
|
|
#define MVS_ID_NOT_MAPPED 0x7f
|
|
#define MVS_CHIP_SLOT_SZ (1U << mvi->chip->slot_width)
|
|
|
|
/* offset for D2H FIS in the Received FIS List Structure */
|
|
#define SATA_RECEIVED_D2H_FIS(reg_set) \
|
|
((void *) mvi->rx_fis + 0x400 + 0x100 * reg_set + 0x40)
|
|
#define SATA_RECEIVED_PIO_FIS(reg_set) \
|
|
((void *) mvi->rx_fis + 0x400 + 0x100 * reg_set + 0x20)
|
|
#define UNASSOC_D2H_FIS(id) \
|
|
((void *) mvi->rx_fis + 0x100 * id)
|
|
|
|
#define for_each_phy(__lseq_mask, __mc, __lseq, __rest) \
|
|
for ((__mc) = (__lseq_mask), (__lseq) = 0; \
|
|
(__mc) != 0 && __rest; \
|
|
(++__lseq), (__mc) >>= 1)
|
|
|
|
/* driver compile-time configuration */
|
|
enum driver_configuration {
|
|
MVS_TX_RING_SZ = 1024, /* TX ring size (12-bit) */
|
|
MVS_RX_RING_SZ = 1024, /* RX ring size (12-bit) */
|
|
/* software requires power-of-2
|
|
ring size */
|
|
|
|
MVS_SLOTS = 512, /* command slots */
|
|
MVS_SLOT_BUF_SZ = 8192, /* cmd tbl + IU + status + PRD */
|
|
MVS_SSP_CMD_SZ = 64, /* SSP command table buffer size */
|
|
MVS_ATA_CMD_SZ = 96, /* SATA command table buffer size */
|
|
MVS_OAF_SZ = 64, /* Open address frame buffer size */
|
|
|
|
MVS_RX_FIS_COUNT = 17, /* Optional rx'd FISs (max 17) */
|
|
|
|
MVS_QUEUE_SIZE = 30, /* Support Queue depth */
|
|
MVS_CAN_QUEUE = MVS_SLOTS - 1, /* SCSI Queue depth */
|
|
};
|
|
|
|
/* unchangeable hardware details */
|
|
enum hardware_details {
|
|
MVS_MAX_PHYS = 8, /* max. possible phys */
|
|
MVS_MAX_PORTS = 8, /* max. possible ports */
|
|
MVS_RX_FISL_SZ = 0x400 + (MVS_RX_FIS_COUNT * 0x100),
|
|
};
|
|
|
|
/* peripheral registers (BAR2) */
|
|
enum peripheral_registers {
|
|
SPI_CTL = 0x10, /* EEPROM control */
|
|
SPI_CMD = 0x14, /* EEPROM command */
|
|
SPI_DATA = 0x18, /* EEPROM data */
|
|
};
|
|
|
|
enum peripheral_register_bits {
|
|
TWSI_RDY = (1U << 7), /* EEPROM interface ready */
|
|
TWSI_RD = (1U << 4), /* EEPROM read access */
|
|
|
|
SPI_ADDR_MASK = 0x3ffff, /* bits 17:0 */
|
|
};
|
|
|
|
/* enhanced mode registers (BAR4) */
|
|
enum hw_registers {
|
|
MVS_GBL_CTL = 0x04, /* global control */
|
|
MVS_GBL_INT_STAT = 0x08, /* global irq status */
|
|
MVS_GBL_PI = 0x0C, /* ports implemented bitmask */
|
|
MVS_GBL_PORT_TYPE = 0xa0, /* port type */
|
|
|
|
MVS_CTL = 0x100, /* SAS/SATA port configuration */
|
|
MVS_PCS = 0x104, /* SAS/SATA port control/status */
|
|
MVS_CMD_LIST_LO = 0x108, /* cmd list addr */
|
|
MVS_CMD_LIST_HI = 0x10C,
|
|
MVS_RX_FIS_LO = 0x110, /* RX FIS list addr */
|
|
MVS_RX_FIS_HI = 0x114,
|
|
|
|
MVS_TX_CFG = 0x120, /* TX configuration */
|
|
MVS_TX_LO = 0x124, /* TX (delivery) ring addr */
|
|
MVS_TX_HI = 0x128,
|
|
|
|
MVS_TX_PROD_IDX = 0x12C, /* TX producer pointer */
|
|
MVS_TX_CONS_IDX = 0x130, /* TX consumer pointer (RO) */
|
|
MVS_RX_CFG = 0x134, /* RX configuration */
|
|
MVS_RX_LO = 0x138, /* RX (completion) ring addr */
|
|
MVS_RX_HI = 0x13C,
|
|
MVS_RX_CONS_IDX = 0x140, /* RX consumer pointer (RO) */
|
|
|
|
MVS_INT_COAL = 0x148, /* Int coalescing config */
|
|
MVS_INT_COAL_TMOUT = 0x14C, /* Int coalescing timeout */
|
|
MVS_INT_STAT = 0x150, /* Central int status */
|
|
MVS_INT_MASK = 0x154, /* Central int enable */
|
|
MVS_INT_STAT_SRS = 0x158, /* SATA register set status */
|
|
MVS_INT_MASK_SRS = 0x15C,
|
|
|
|
/* ports 1-3 follow after this */
|
|
MVS_P0_INT_STAT = 0x160, /* port0 interrupt status */
|
|
MVS_P0_INT_MASK = 0x164, /* port0 interrupt mask */
|
|
MVS_P4_INT_STAT = 0x200, /* Port 4 interrupt status */
|
|
MVS_P4_INT_MASK = 0x204, /* Port 4 interrupt enable mask */
|
|
|
|
/* ports 1-3 follow after this */
|
|
MVS_P0_SER_CTLSTAT = 0x180, /* port0 serial control/status */
|
|
MVS_P4_SER_CTLSTAT = 0x220, /* port4 serial control/status */
|
|
|
|
MVS_CMD_ADDR = 0x1B8, /* Command register port (addr) */
|
|
MVS_CMD_DATA = 0x1BC, /* Command register port (data) */
|
|
|
|
/* ports 1-3 follow after this */
|
|
MVS_P0_CFG_ADDR = 0x1C0, /* port0 phy register address */
|
|
MVS_P0_CFG_DATA = 0x1C4, /* port0 phy register data */
|
|
MVS_P4_CFG_ADDR = 0x230, /* Port 4 config address */
|
|
MVS_P4_CFG_DATA = 0x234, /* Port 4 config data */
|
|
|
|
/* ports 1-3 follow after this */
|
|
MVS_P0_VSR_ADDR = 0x1E0, /* port0 VSR address */
|
|
MVS_P0_VSR_DATA = 0x1E4, /* port0 VSR data */
|
|
MVS_P4_VSR_ADDR = 0x250, /* port 4 VSR addr */
|
|
MVS_P4_VSR_DATA = 0x254, /* port 4 VSR data */
|
|
};
|
|
|
|
enum hw_register_bits {
|
|
/* MVS_GBL_CTL */
|
|
INT_EN = (1U << 1), /* Global int enable */
|
|
HBA_RST = (1U << 0), /* HBA reset */
|
|
|
|
/* MVS_GBL_INT_STAT */
|
|
INT_XOR = (1U << 4), /* XOR engine event */
|
|
INT_SAS_SATA = (1U << 0), /* SAS/SATA event */
|
|
|
|
/* MVS_GBL_PORT_TYPE */ /* shl for ports 1-3 */
|
|
SATA_TARGET = (1U << 16), /* port0 SATA target enable */
|
|
MODE_AUTO_DET_PORT7 = (1U << 15), /* port0 SAS/SATA autodetect */
|
|
MODE_AUTO_DET_PORT6 = (1U << 14),
|
|
MODE_AUTO_DET_PORT5 = (1U << 13),
|
|
MODE_AUTO_DET_PORT4 = (1U << 12),
|
|
MODE_AUTO_DET_PORT3 = (1U << 11),
|
|
MODE_AUTO_DET_PORT2 = (1U << 10),
|
|
MODE_AUTO_DET_PORT1 = (1U << 9),
|
|
MODE_AUTO_DET_PORT0 = (1U << 8),
|
|
MODE_AUTO_DET_EN = MODE_AUTO_DET_PORT0 | MODE_AUTO_DET_PORT1 |
|
|
MODE_AUTO_DET_PORT2 | MODE_AUTO_DET_PORT3 |
|
|
MODE_AUTO_DET_PORT4 | MODE_AUTO_DET_PORT5 |
|
|
MODE_AUTO_DET_PORT6 | MODE_AUTO_DET_PORT7,
|
|
MODE_SAS_PORT7_MASK = (1U << 7), /* port0 SAS(1), SATA(0) mode */
|
|
MODE_SAS_PORT6_MASK = (1U << 6),
|
|
MODE_SAS_PORT5_MASK = (1U << 5),
|
|
MODE_SAS_PORT4_MASK = (1U << 4),
|
|
MODE_SAS_PORT3_MASK = (1U << 3),
|
|
MODE_SAS_PORT2_MASK = (1U << 2),
|
|
MODE_SAS_PORT1_MASK = (1U << 1),
|
|
MODE_SAS_PORT0_MASK = (1U << 0),
|
|
MODE_SAS_SATA = MODE_SAS_PORT0_MASK | MODE_SAS_PORT1_MASK |
|
|
MODE_SAS_PORT2_MASK | MODE_SAS_PORT3_MASK |
|
|
MODE_SAS_PORT4_MASK | MODE_SAS_PORT5_MASK |
|
|
MODE_SAS_PORT6_MASK | MODE_SAS_PORT7_MASK,
|
|
|
|
/* SAS_MODE value may be
|
|
* dictated (in hw) by values
|
|
* of SATA_TARGET & AUTO_DET
|
|
*/
|
|
|
|
/* MVS_TX_CFG */
|
|
TX_EN = (1U << 16), /* Enable TX */
|
|
TX_RING_SZ_MASK = 0xfff, /* TX ring size, bits 11:0 */
|
|
|
|
/* MVS_RX_CFG */
|
|
RX_EN = (1U << 16), /* Enable RX */
|
|
RX_RING_SZ_MASK = 0xfff, /* RX ring size, bits 11:0 */
|
|
|
|
/* MVS_INT_COAL */
|
|
COAL_EN = (1U << 16), /* Enable int coalescing */
|
|
|
|
/* MVS_INT_STAT, MVS_INT_MASK */
|
|
CINT_I2C = (1U << 31), /* I2C event */
|
|
CINT_SW0 = (1U << 30), /* software event 0 */
|
|
CINT_SW1 = (1U << 29), /* software event 1 */
|
|
CINT_PRD_BC = (1U << 28), /* PRD BC err for read cmd */
|
|
CINT_DMA_PCIE = (1U << 27), /* DMA to PCIE timeout */
|
|
CINT_MEM = (1U << 26), /* int mem parity err */
|
|
CINT_I2C_SLAVE = (1U << 25), /* slave I2C event */
|
|
CINT_SRS = (1U << 3), /* SRS event */
|
|
CINT_CI_STOP = (1U << 1), /* cmd issue stopped */
|
|
CINT_DONE = (1U << 0), /* cmd completion */
|
|
|
|
/* shl for ports 1-3 */
|
|
CINT_PORT_STOPPED = (1U << 16), /* port0 stopped */
|
|
CINT_PORT = (1U << 8), /* port0 event */
|
|
CINT_PORT_MASK_OFFSET = 8,
|
|
CINT_PORT_MASK = (0xFF << CINT_PORT_MASK_OFFSET),
|
|
|
|
/* TX (delivery) ring bits */
|
|
TXQ_CMD_SHIFT = 29,
|
|
TXQ_CMD_SSP = 1, /* SSP protocol */
|
|
TXQ_CMD_SMP = 2, /* SMP protocol */
|
|
TXQ_CMD_STP = 3, /* STP/SATA protocol */
|
|
TXQ_CMD_SSP_FREE_LIST = 4, /* add to SSP targ free list */
|
|
TXQ_CMD_SLOT_RESET = 7, /* reset command slot */
|
|
TXQ_MODE_I = (1U << 28), /* mode: 0=target,1=initiator */
|
|
TXQ_PRIO_HI = (1U << 27), /* priority: 0=normal, 1=high */
|
|
TXQ_SRS_SHIFT = 20, /* SATA register set */
|
|
TXQ_SRS_MASK = 0x7f,
|
|
TXQ_PHY_SHIFT = 12, /* PHY bitmap */
|
|
TXQ_PHY_MASK = 0xff,
|
|
TXQ_SLOT_MASK = 0xfff, /* slot number */
|
|
|
|
/* RX (completion) ring bits */
|
|
RXQ_GOOD = (1U << 23), /* Response good */
|
|
RXQ_SLOT_RESET = (1U << 21), /* Slot reset complete */
|
|
RXQ_CMD_RX = (1U << 20), /* target cmd received */
|
|
RXQ_ATTN = (1U << 19), /* attention */
|
|
RXQ_RSP = (1U << 18), /* response frame xfer'd */
|
|
RXQ_ERR = (1U << 17), /* err info rec xfer'd */
|
|
RXQ_DONE = (1U << 16), /* cmd complete */
|
|
RXQ_SLOT_MASK = 0xfff, /* slot number */
|
|
|
|
/* mvs_cmd_hdr bits */
|
|
MCH_PRD_LEN_SHIFT = 16, /* 16-bit PRD table len */
|
|
MCH_SSP_FR_TYPE_SHIFT = 13, /* SSP frame type */
|
|
|
|
/* SSP initiator only */
|
|
MCH_SSP_FR_CMD = 0x0, /* COMMAND frame */
|
|
|
|
/* SSP initiator or target */
|
|
MCH_SSP_FR_TASK = 0x1, /* TASK frame */
|
|
|
|
/* SSP target only */
|
|
MCH_SSP_FR_XFER_RDY = 0x4, /* XFER_RDY frame */
|
|
MCH_SSP_FR_RESP = 0x5, /* RESPONSE frame */
|
|
MCH_SSP_FR_READ = 0x6, /* Read DATA frame(s) */
|
|
MCH_SSP_FR_READ_RESP = 0x7, /* ditto, plus RESPONSE */
|
|
|
|
MCH_PASSTHRU = (1U << 12), /* pass-through (SSP) */
|
|
MCH_FBURST = (1U << 11), /* first burst (SSP) */
|
|
MCH_CHK_LEN = (1U << 10), /* chk xfer len (SSP) */
|
|
MCH_RETRY = (1U << 9), /* tport layer retry (SSP) */
|
|
MCH_PROTECTION = (1U << 8), /* protection info rec (SSP) */
|
|
MCH_RESET = (1U << 7), /* Reset (STP/SATA) */
|
|
MCH_FPDMA = (1U << 6), /* First party DMA (STP/SATA) */
|
|
MCH_ATAPI = (1U << 5), /* ATAPI (STP/SATA) */
|
|
MCH_BIST = (1U << 4), /* BIST activate (STP/SATA) */
|
|
MCH_PMP_MASK = 0xf, /* PMP from cmd FIS (STP/SATA)*/
|
|
|
|
CCTL_RST = (1U << 5), /* port logic reset */
|
|
|
|
/* 0(LSB first), 1(MSB first) */
|
|
CCTL_ENDIAN_DATA = (1U << 3), /* PRD data */
|
|
CCTL_ENDIAN_RSP = (1U << 2), /* response frame */
|
|
CCTL_ENDIAN_OPEN = (1U << 1), /* open address frame */
|
|
CCTL_ENDIAN_CMD = (1U << 0), /* command table */
|
|
|
|
/* MVS_Px_SER_CTLSTAT (per-phy control) */
|
|
PHY_SSP_RST = (1U << 3), /* reset SSP link layer */
|
|
PHY_BCAST_CHG = (1U << 2), /* broadcast(change) notif */
|
|
PHY_RST_HARD = (1U << 1), /* hard reset + phy reset */
|
|
PHY_RST = (1U << 0), /* phy reset */
|
|
PHY_MIN_SPP_PHYS_LINK_RATE_MASK = (0xF << 8),
|
|
PHY_MAX_SPP_PHYS_LINK_RATE_MASK = (0xF << 12),
|
|
PHY_NEG_SPP_PHYS_LINK_RATE_MASK_OFFSET = (16),
|
|
PHY_NEG_SPP_PHYS_LINK_RATE_MASK =
|
|
(0xF << PHY_NEG_SPP_PHYS_LINK_RATE_MASK_OFFSET),
|
|
PHY_READY_MASK = (1U << 20),
|
|
|
|
/* MVS_Px_INT_STAT, MVS_Px_INT_MASK (per-phy events) */
|
|
PHYEV_DEC_ERR = (1U << 24), /* Phy Decoding Error */
|
|
PHYEV_UNASSOC_FIS = (1U << 19), /* unassociated FIS rx'd */
|
|
PHYEV_AN = (1U << 18), /* SATA async notification */
|
|
PHYEV_BIST_ACT = (1U << 17), /* BIST activate FIS */
|
|
PHYEV_SIG_FIS = (1U << 16), /* signature FIS */
|
|
PHYEV_POOF = (1U << 12), /* phy ready from 1 -> 0 */
|
|
PHYEV_IU_BIG = (1U << 11), /* IU too long err */
|
|
PHYEV_IU_SMALL = (1U << 10), /* IU too short err */
|
|
PHYEV_UNK_TAG = (1U << 9), /* unknown tag */
|
|
PHYEV_BROAD_CH = (1U << 8), /* broadcast(CHANGE) */
|
|
PHYEV_COMWAKE = (1U << 7), /* COMWAKE rx'd */
|
|
PHYEV_PORT_SEL = (1U << 6), /* port selector present */
|
|
PHYEV_HARD_RST = (1U << 5), /* hard reset rx'd */
|
|
PHYEV_ID_TMOUT = (1U << 4), /* identify timeout */
|
|
PHYEV_ID_FAIL = (1U << 3), /* identify failed */
|
|
PHYEV_ID_DONE = (1U << 2), /* identify done */
|
|
PHYEV_HARD_RST_DONE = (1U << 1), /* hard reset done */
|
|
PHYEV_RDY_CH = (1U << 0), /* phy ready changed state */
|
|
|
|
/* MVS_PCS */
|
|
PCS_EN_SATA_REG_SHIFT = (16), /* Enable SATA Register Set */
|
|
PCS_EN_PORT_XMT_SHIFT = (12), /* Enable Port Transmit */
|
|
PCS_EN_PORT_XMT_SHIFT2 = (8), /* For 6480 */
|
|
PCS_SATA_RETRY = (1U << 8), /* retry ctl FIS on R_ERR */
|
|
PCS_RSP_RX_EN = (1U << 7), /* raw response rx */
|
|
PCS_SELF_CLEAR = (1U << 5), /* self-clearing int mode */
|
|
PCS_FIS_RX_EN = (1U << 4), /* FIS rx enable */
|
|
PCS_CMD_STOP_ERR = (1U << 3), /* cmd stop-on-err enable */
|
|
PCS_CMD_RST = (1U << 1), /* reset cmd issue */
|
|
PCS_CMD_EN = (1U << 0), /* enable cmd issue */
|
|
|
|
/* Port n Attached Device Info */
|
|
PORT_DEV_SSP_TRGT = (1U << 19),
|
|
PORT_DEV_SMP_TRGT = (1U << 18),
|
|
PORT_DEV_STP_TRGT = (1U << 17),
|
|
PORT_DEV_SSP_INIT = (1U << 11),
|
|
PORT_DEV_SMP_INIT = (1U << 10),
|
|
PORT_DEV_STP_INIT = (1U << 9),
|
|
PORT_PHY_ID_MASK = (0xFFU << 24),
|
|
PORT_DEV_TRGT_MASK = (0x7U << 17),
|
|
PORT_DEV_INIT_MASK = (0x7U << 9),
|
|
PORT_DEV_TYPE_MASK = (0x7U << 0),
|
|
|
|
/* Port n PHY Status */
|
|
PHY_RDY = (1U << 2),
|
|
PHY_DW_SYNC = (1U << 1),
|
|
PHY_OOB_DTCTD = (1U << 0),
|
|
|
|
/* VSR */
|
|
/* PHYMODE 6 (CDB) */
|
|
PHY_MODE6_LATECLK = (1U << 29), /* Lock Clock */
|
|
PHY_MODE6_DTL_SPEED = (1U << 27), /* Digital Loop Speed */
|
|
PHY_MODE6_FC_ORDER = (1U << 26), /* Fibre Channel Mode Order*/
|
|
PHY_MODE6_MUCNT_EN = (1U << 24), /* u Count Enable */
|
|
PHY_MODE6_SEL_MUCNT_LEN = (1U << 22), /* Training Length Select */
|
|
PHY_MODE6_SELMUPI = (1U << 20), /* Phase Multi Select (init) */
|
|
PHY_MODE6_SELMUPF = (1U << 18), /* Phase Multi Select (final) */
|
|
PHY_MODE6_SELMUFF = (1U << 16), /* Freq Loop Multi Sel(final) */
|
|
PHY_MODE6_SELMUFI = (1U << 14), /* Freq Loop Multi Sel(init) */
|
|
PHY_MODE6_FREEZE_LOOP = (1U << 12), /* Freeze Rx CDR Loop */
|
|
PHY_MODE6_INT_RXFOFFS = (1U << 3), /* Rx CDR Freq Loop Enable */
|
|
PHY_MODE6_FRC_RXFOFFS = (1U << 2), /* Initial Rx CDR Offset */
|
|
PHY_MODE6_STAU_0D8 = (1U << 1), /* Rx CDR Freq Loop Saturate */
|
|
PHY_MODE6_RXSAT_DIS = (1U << 0), /* Saturate Ctl */
|
|
};
|
|
|
|
enum mvs_info_flags {
|
|
MVF_MSI = (1U << 0), /* MSI is enabled */
|
|
MVF_PHY_PWR_FIX = (1U << 1), /* bug workaround */
|
|
};
|
|
|
|
enum sas_cmd_port_registers {
|
|
CMD_CMRST_OOB_DET = 0x100, /* COMRESET OOB detect register */
|
|
CMD_CMWK_OOB_DET = 0x104, /* COMWAKE OOB detect register */
|
|
CMD_CMSAS_OOB_DET = 0x108, /* COMSAS OOB detect register */
|
|
CMD_BRST_OOB_DET = 0x10c, /* burst OOB detect register */
|
|
CMD_OOB_SPACE = 0x110, /* OOB space control register */
|
|
CMD_OOB_BURST = 0x114, /* OOB burst control register */
|
|
CMD_PHY_TIMER = 0x118, /* PHY timer control register */
|
|
CMD_PHY_CONFIG0 = 0x11c, /* PHY config register 0 */
|
|
CMD_PHY_CONFIG1 = 0x120, /* PHY config register 1 */
|
|
CMD_SAS_CTL0 = 0x124, /* SAS control register 0 */
|
|
CMD_SAS_CTL1 = 0x128, /* SAS control register 1 */
|
|
CMD_SAS_CTL2 = 0x12c, /* SAS control register 2 */
|
|
CMD_SAS_CTL3 = 0x130, /* SAS control register 3 */
|
|
CMD_ID_TEST = 0x134, /* ID test register */
|
|
CMD_PL_TIMER = 0x138, /* PL timer register */
|
|
CMD_WD_TIMER = 0x13c, /* WD timer register */
|
|
CMD_PORT_SEL_COUNT = 0x140, /* port selector count register */
|
|
CMD_APP_MEM_CTL = 0x144, /* Application Memory Control */
|
|
CMD_XOR_MEM_CTL = 0x148, /* XOR Block Memory Control */
|
|
CMD_DMA_MEM_CTL = 0x14c, /* DMA Block Memory Control */
|
|
CMD_PORT_MEM_CTL0 = 0x150, /* Port Memory Control 0 */
|
|
CMD_PORT_MEM_CTL1 = 0x154, /* Port Memory Control 1 */
|
|
CMD_SATA_PORT_MEM_CTL0 = 0x158, /* SATA Port Memory Control 0 */
|
|
CMD_SATA_PORT_MEM_CTL1 = 0x15c, /* SATA Port Memory Control 1 */
|
|
CMD_XOR_MEM_BIST_CTL = 0x160, /* XOR Memory BIST Control */
|
|
CMD_XOR_MEM_BIST_STAT = 0x164, /* XOR Memroy BIST Status */
|
|
CMD_DMA_MEM_BIST_CTL = 0x168, /* DMA Memory BIST Control */
|
|
CMD_DMA_MEM_BIST_STAT = 0x16c, /* DMA Memory BIST Status */
|
|
CMD_PORT_MEM_BIST_CTL = 0x170, /* Port Memory BIST Control */
|
|
CMD_PORT_MEM_BIST_STAT0 = 0x174, /* Port Memory BIST Status 0 */
|
|
CMD_PORT_MEM_BIST_STAT1 = 0x178, /* Port Memory BIST Status 1 */
|
|
CMD_STP_MEM_BIST_CTL = 0x17c, /* STP Memory BIST Control */
|
|
CMD_STP_MEM_BIST_STAT0 = 0x180, /* STP Memory BIST Status 0 */
|
|
CMD_STP_MEM_BIST_STAT1 = 0x184, /* STP Memory BIST Status 1 */
|
|
CMD_RESET_COUNT = 0x188, /* Reset Count */
|
|
CMD_MONTR_DATA_SEL = 0x18C, /* Monitor Data/Select */
|
|
CMD_PLL_PHY_CONFIG = 0x190, /* PLL/PHY Configuration */
|
|
CMD_PHY_CTL = 0x194, /* PHY Control and Status */
|
|
CMD_PHY_TEST_COUNT0 = 0x198, /* Phy Test Count 0 */
|
|
CMD_PHY_TEST_COUNT1 = 0x19C, /* Phy Test Count 1 */
|
|
CMD_PHY_TEST_COUNT2 = 0x1A0, /* Phy Test Count 2 */
|
|
CMD_APP_ERR_CONFIG = 0x1A4, /* Application Error Configuration */
|
|
CMD_PND_FIFO_CTL0 = 0x1A8, /* Pending FIFO Control 0 */
|
|
CMD_HOST_CTL = 0x1AC, /* Host Control Status */
|
|
CMD_HOST_WR_DATA = 0x1B0, /* Host Write Data */
|
|
CMD_HOST_RD_DATA = 0x1B4, /* Host Read Data */
|
|
CMD_PHY_MODE_21 = 0x1B8, /* Phy Mode 21 */
|
|
CMD_SL_MODE0 = 0x1BC, /* SL Mode 0 */
|
|
CMD_SL_MODE1 = 0x1C0, /* SL Mode 1 */
|
|
CMD_PND_FIFO_CTL1 = 0x1C4, /* Pending FIFO Control 1 */
|
|
};
|
|
|
|
/* SAS/SATA configuration port registers, aka phy registers */
|
|
enum sas_sata_config_port_regs {
|
|
PHYR_IDENTIFY = 0x00, /* info for IDENTIFY frame */
|
|
PHYR_ADDR_LO = 0x04, /* my SAS address (low) */
|
|
PHYR_ADDR_HI = 0x08, /* my SAS address (high) */
|
|
PHYR_ATT_DEV_INFO = 0x0C, /* attached device info */
|
|
PHYR_ATT_ADDR_LO = 0x10, /* attached dev SAS addr (low) */
|
|
PHYR_ATT_ADDR_HI = 0x14, /* attached dev SAS addr (high) */
|
|
PHYR_SATA_CTL = 0x18, /* SATA control */
|
|
PHYR_PHY_STAT = 0x1C, /* PHY status */
|
|
PHYR_SATA_SIG0 = 0x20, /*port SATA signature FIS(Byte 0-3) */
|
|
PHYR_SATA_SIG1 = 0x24, /*port SATA signature FIS(Byte 4-7) */
|
|
PHYR_SATA_SIG2 = 0x28, /*port SATA signature FIS(Byte 8-11) */
|
|
PHYR_SATA_SIG3 = 0x2c, /*port SATA signature FIS(Byte 12-15) */
|
|
PHYR_R_ERR_COUNT = 0x30, /* port R_ERR count register */
|
|
PHYR_CRC_ERR_COUNT = 0x34, /* port CRC error count register */
|
|
PHYR_WIDE_PORT = 0x38, /* wide port participating */
|
|
PHYR_CURRENT0 = 0x80, /* current connection info 0 */
|
|
PHYR_CURRENT1 = 0x84, /* current connection info 1 */
|
|
PHYR_CURRENT2 = 0x88, /* current connection info 2 */
|
|
};
|
|
|
|
/* SAS/SATA Vendor Specific Port Registers */
|
|
enum sas_sata_vsp_regs {
|
|
VSR_PHY_STAT = 0x00, /* Phy Status */
|
|
VSR_PHY_MODE1 = 0x01, /* phy tx */
|
|
VSR_PHY_MODE2 = 0x02, /* tx scc */
|
|
VSR_PHY_MODE3 = 0x03, /* pll */
|
|
VSR_PHY_MODE4 = 0x04, /* VCO */
|
|
VSR_PHY_MODE5 = 0x05, /* Rx */
|
|
VSR_PHY_MODE6 = 0x06, /* CDR */
|
|
VSR_PHY_MODE7 = 0x07, /* Impedance */
|
|
VSR_PHY_MODE8 = 0x08, /* Voltage */
|
|
VSR_PHY_MODE9 = 0x09, /* Test */
|
|
VSR_PHY_MODE10 = 0x0A, /* Power */
|
|
VSR_PHY_MODE11 = 0x0B, /* Phy Mode */
|
|
VSR_PHY_VS0 = 0x0C, /* Vednor Specific 0 */
|
|
VSR_PHY_VS1 = 0x0D, /* Vednor Specific 1 */
|
|
};
|
|
|
|
enum pci_cfg_registers {
|
|
PCR_PHY_CTL = 0x40,
|
|
PCR_PHY_CTL2 = 0x90,
|
|
PCR_DEV_CTRL = 0xE8,
|
|
};
|
|
|
|
enum pci_cfg_register_bits {
|
|
PCTL_PWR_ON = (0xFU << 24),
|
|
PCTL_OFF = (0xFU << 12),
|
|
PRD_REQ_SIZE = (0x4000),
|
|
PRD_REQ_MASK = (0x00007000),
|
|
};
|
|
|
|
enum nvram_layout_offsets {
|
|
NVR_SIG = 0x00, /* 0xAA, 0x55 */
|
|
NVR_SAS_ADDR = 0x02, /* 8-byte SAS address */
|
|
};
|
|
|
|
enum chip_flavors {
|
|
chip_6320,
|
|
chip_6440,
|
|
chip_6480,
|
|
};
|
|
|
|
enum port_type {
|
|
PORT_TYPE_SAS = (1L << 1),
|
|
PORT_TYPE_SATA = (1L << 0),
|
|
};
|
|
|
|
/* Command Table Format */
|
|
enum ct_format {
|
|
/* SSP */
|
|
SSP_F_H = 0x00,
|
|
SSP_F_IU = 0x18,
|
|
SSP_F_MAX = 0x4D,
|
|
/* STP */
|
|
STP_CMD_FIS = 0x00,
|
|
STP_ATAPI_CMD = 0x40,
|
|
STP_F_MAX = 0x10,
|
|
/* SMP */
|
|
SMP_F_T = 0x00,
|
|
SMP_F_DEP = 0x01,
|
|
SMP_F_MAX = 0x101,
|
|
};
|
|
|
|
enum status_buffer {
|
|
SB_EIR_OFF = 0x00, /* Error Information Record */
|
|
SB_RFB_OFF = 0x08, /* Response Frame Buffer */
|
|
SB_RFB_MAX = 0x400, /* RFB size*/
|
|
};
|
|
|
|
enum error_info_rec {
|
|
CMD_ISS_STPD = (1U << 31), /* Cmd Issue Stopped */
|
|
CMD_PI_ERR = (1U << 30), /* Protection info error. see flags2 */
|
|
RSP_OVER = (1U << 29), /* rsp buffer overflow */
|
|
RETRY_LIM = (1U << 28), /* FIS/frame retry limit exceeded */
|
|
UNK_FIS = (1U << 27), /* unknown FIS */
|
|
DMA_TERM = (1U << 26), /* DMA terminate primitive rx'd */
|
|
SYNC_ERR = (1U << 25), /* SYNC rx'd during frame xmit */
|
|
TFILE_ERR = (1U << 24), /* SATA taskfile Error bit set */
|
|
R_ERR = (1U << 23), /* SATA returned R_ERR prim */
|
|
RD_OFS = (1U << 20), /* Read DATA frame invalid offset */
|
|
XFER_RDY_OFS = (1U << 19), /* XFER_RDY offset error */
|
|
UNEXP_XFER_RDY = (1U << 18), /* unexpected XFER_RDY error */
|
|
DATA_OVER_UNDER = (1U << 16), /* data overflow/underflow */
|
|
INTERLOCK = (1U << 15), /* interlock error */
|
|
NAK = (1U << 14), /* NAK rx'd */
|
|
ACK_NAK_TO = (1U << 13), /* ACK/NAK timeout */
|
|
CXN_CLOSED = (1U << 12), /* cxn closed w/out ack/nak */
|
|
OPEN_TO = (1U << 11), /* I_T nexus lost, open cxn timeout */
|
|
PATH_BLOCKED = (1U << 10), /* I_T nexus lost, pathway blocked */
|
|
NO_DEST = (1U << 9), /* I_T nexus lost, no destination */
|
|
STP_RES_BSY = (1U << 8), /* STP resources busy */
|
|
BREAK = (1U << 7), /* break received */
|
|
BAD_DEST = (1U << 6), /* bad destination */
|
|
BAD_PROTO = (1U << 5), /* protocol not supported */
|
|
BAD_RATE = (1U << 4), /* cxn rate not supported */
|
|
WRONG_DEST = (1U << 3), /* wrong destination error */
|
|
CREDIT_TO = (1U << 2), /* credit timeout */
|
|
WDOG_TO = (1U << 1), /* watchdog timeout */
|
|
BUF_PAR = (1U << 0), /* buffer parity error */
|
|
};
|
|
|
|
enum error_info_rec_2 {
|
|
SLOT_BSY_ERR = (1U << 31), /* Slot Busy Error */
|
|
GRD_CHK_ERR = (1U << 14), /* Guard Check Error */
|
|
APP_CHK_ERR = (1U << 13), /* Application Check error */
|
|
REF_CHK_ERR = (1U << 12), /* Reference Check Error */
|
|
USR_BLK_NM = (1U << 0), /* User Block Number */
|
|
};
|
|
|
|
struct mvs_chip_info {
|
|
u32 n_phy;
|
|
u32 srs_sz;
|
|
u32 slot_width;
|
|
};
|
|
|
|
struct mvs_err_info {
|
|
__le32 flags;
|
|
__le32 flags2;
|
|
};
|
|
|
|
struct mvs_prd {
|
|
__le64 addr; /* 64-bit buffer address */
|
|
__le32 reserved;
|
|
__le32 len; /* 16-bit length */
|
|
};
|
|
|
|
struct mvs_cmd_hdr {
|
|
__le32 flags; /* PRD tbl len; SAS, SATA ctl */
|
|
__le32 lens; /* cmd, max resp frame len */
|
|
__le32 tags; /* targ port xfer tag; tag */
|
|
__le32 data_len; /* data xfer len */
|
|
__le64 cmd_tbl; /* command table address */
|
|
__le64 open_frame; /* open addr frame address */
|
|
__le64 status_buf; /* status buffer address */
|
|
__le64 prd_tbl; /* PRD tbl address */
|
|
__le32 reserved[4];
|
|
};
|
|
|
|
struct mvs_port {
|
|
struct asd_sas_port sas_port;
|
|
u8 port_attached;
|
|
u8 taskfileset;
|
|
u8 wide_port_phymap;
|
|
struct list_head list;
|
|
};
|
|
|
|
struct mvs_phy {
|
|
struct mvs_port *port;
|
|
struct asd_sas_phy sas_phy;
|
|
struct sas_identify identify;
|
|
struct scsi_device *sdev;
|
|
u64 dev_sas_addr;
|
|
u64 att_dev_sas_addr;
|
|
u32 att_dev_info;
|
|
u32 dev_info;
|
|
u32 phy_type;
|
|
u32 phy_status;
|
|
u32 irq_status;
|
|
u32 frame_rcvd_size;
|
|
u8 frame_rcvd[32];
|
|
u8 phy_attached;
|
|
enum sas_linkrate minimum_linkrate;
|
|
enum sas_linkrate maximum_linkrate;
|
|
};
|
|
|
|
struct mvs_slot_info {
|
|
struct list_head list;
|
|
struct sas_task *task;
|
|
u32 n_elem;
|
|
u32 tx;
|
|
|
|
/* DMA buffer for storing cmd tbl, open addr frame, status buffer,
|
|
* and PRD table
|
|
*/
|
|
void *buf;
|
|
dma_addr_t buf_dma;
|
|
#if _MV_DUMP
|
|
u32 cmd_size;
|
|
#endif
|
|
|
|
void *response;
|
|
struct mvs_port *port;
|
|
};
|
|
|
|
struct mvs_info {
|
|
unsigned long flags;
|
|
|
|
spinlock_t lock; /* host-wide lock */
|
|
struct pci_dev *pdev; /* our device */
|
|
void __iomem *regs; /* enhanced mode registers */
|
|
void __iomem *peri_regs; /* peripheral registers */
|
|
|
|
u8 sas_addr[SAS_ADDR_SIZE];
|
|
struct sas_ha_struct sas; /* SCSI/SAS glue */
|
|
struct Scsi_Host *shost;
|
|
|
|
__le32 *tx; /* TX (delivery) DMA ring */
|
|
dma_addr_t tx_dma;
|
|
u32 tx_prod; /* cached next-producer idx */
|
|
|
|
__le32 *rx; /* RX (completion) DMA ring */
|
|
dma_addr_t rx_dma;
|
|
u32 rx_cons; /* RX consumer idx */
|
|
|
|
__le32 *rx_fis; /* RX'd FIS area */
|
|
dma_addr_t rx_fis_dma;
|
|
|
|
struct mvs_cmd_hdr *slot; /* DMA command header slots */
|
|
dma_addr_t slot_dma;
|
|
|
|
const struct mvs_chip_info *chip;
|
|
|
|
u8 tags[MVS_SLOTS];
|
|
struct mvs_slot_info slot_info[MVS_SLOTS];
|
|
/* further per-slot information */
|
|
struct mvs_phy phy[MVS_MAX_PHYS];
|
|
struct mvs_port port[MVS_MAX_PHYS];
|
|
#ifdef MVS_USE_TASKLET
|
|
struct tasklet_struct tasklet;
|
|
#endif
|
|
};
|
|
|
|
static int mvs_phy_control(struct asd_sas_phy *sas_phy, enum phy_func func,
|
|
void *funcdata);
|
|
static u32 mvs_read_phy_ctl(struct mvs_info *mvi, u32 port);
|
|
static void mvs_write_phy_ctl(struct mvs_info *mvi, u32 port, u32 val);
|
|
static u32 mvs_read_port_irq_stat(struct mvs_info *mvi, u32 port);
|
|
static void mvs_write_port_irq_stat(struct mvs_info *mvi, u32 port, u32 val);
|
|
static void mvs_write_port_irq_mask(struct mvs_info *mvi, u32 port, u32 val);
|
|
static u32 mvs_read_port_irq_mask(struct mvs_info *mvi, u32 port);
|
|
|
|
static u32 mvs_is_phy_ready(struct mvs_info *mvi, int i);
|
|
static void mvs_detect_porttype(struct mvs_info *mvi, int i);
|
|
static void mvs_update_phyinfo(struct mvs_info *mvi, int i, int get_st);
|
|
static void mvs_release_task(struct mvs_info *mvi, int phy_no);
|
|
|
|
static int mvs_scan_finished(struct Scsi_Host *, unsigned long);
|
|
static void mvs_scan_start(struct Scsi_Host *);
|
|
static int mvs_slave_configure(struct scsi_device *sdev);
|
|
|
|
static struct scsi_transport_template *mvs_stt;
|
|
|
|
static const struct mvs_chip_info mvs_chips[] = {
|
|
[chip_6320] = { 2, 16, 9 },
|
|
[chip_6440] = { 4, 16, 9 },
|
|
[chip_6480] = { 8, 32, 10 },
|
|
};
|
|
|
|
static struct scsi_host_template mvs_sht = {
|
|
.module = THIS_MODULE,
|
|
.name = DRV_NAME,
|
|
.queuecommand = sas_queuecommand,
|
|
.target_alloc = sas_target_alloc,
|
|
.slave_configure = mvs_slave_configure,
|
|
.slave_destroy = sas_slave_destroy,
|
|
.scan_finished = mvs_scan_finished,
|
|
.scan_start = mvs_scan_start,
|
|
.change_queue_depth = sas_change_queue_depth,
|
|
.change_queue_type = sas_change_queue_type,
|
|
.bios_param = sas_bios_param,
|
|
.can_queue = 1,
|
|
.cmd_per_lun = 1,
|
|
.this_id = -1,
|
|
.sg_tablesize = SG_ALL,
|
|
.max_sectors = SCSI_DEFAULT_MAX_SECTORS,
|
|
.use_clustering = ENABLE_CLUSTERING,
|
|
.eh_device_reset_handler = sas_eh_device_reset_handler,
|
|
.eh_bus_reset_handler = sas_eh_bus_reset_handler,
|
|
.slave_alloc = sas_slave_alloc,
|
|
.target_destroy = sas_target_destroy,
|
|
.ioctl = sas_ioctl,
|
|
};
|
|
|
|
static void mvs_hexdump(u32 size, u8 *data, u32 baseaddr)
|
|
{
|
|
u32 i;
|
|
u32 run;
|
|
u32 offset;
|
|
|
|
offset = 0;
|
|
while (size) {
|
|
printk("%08X : ", baseaddr + offset);
|
|
if (size >= 16)
|
|
run = 16;
|
|
else
|
|
run = size;
|
|
size -= run;
|
|
for (i = 0; i < 16; i++) {
|
|
if (i < run)
|
|
printk("%02X ", (u32)data[i]);
|
|
else
|
|
printk(" ");
|
|
}
|
|
printk(": ");
|
|
for (i = 0; i < run; i++)
|
|
printk("%c", isalnum(data[i]) ? data[i] : '.');
|
|
printk("\n");
|
|
data = &data[16];
|
|
offset += run;
|
|
}
|
|
printk("\n");
|
|
}
|
|
|
|
#if _MV_DUMP
|
|
static void mvs_hba_sb_dump(struct mvs_info *mvi, u32 tag,
|
|
enum sas_protocol proto)
|
|
{
|
|
u32 offset;
|
|
struct pci_dev *pdev = mvi->pdev;
|
|
struct mvs_slot_info *slot = &mvi->slot_info[tag];
|
|
|
|
offset = slot->cmd_size + MVS_OAF_SZ +
|
|
sizeof(struct mvs_prd) * slot->n_elem;
|
|
dev_printk(KERN_DEBUG, &pdev->dev, "+---->Status buffer[%d] :\n",
|
|
tag);
|
|
mvs_hexdump(32, (u8 *) slot->response,
|
|
(u32) slot->buf_dma + offset);
|
|
}
|
|
#endif
|
|
|
|
static void mvs_hba_memory_dump(struct mvs_info *mvi, u32 tag,
|
|
enum sas_protocol proto)
|
|
{
|
|
#if _MV_DUMP
|
|
u32 sz, w_ptr;
|
|
u64 addr;
|
|
void __iomem *regs = mvi->regs;
|
|
struct pci_dev *pdev = mvi->pdev;
|
|
struct mvs_slot_info *slot = &mvi->slot_info[tag];
|
|
|
|
/*Delivery Queue */
|
|
sz = mr32(TX_CFG) & TX_RING_SZ_MASK;
|
|
w_ptr = slot->tx;
|
|
addr = mr32(TX_HI) << 16 << 16 | mr32(TX_LO);
|
|
dev_printk(KERN_DEBUG, &pdev->dev,
|
|
"Delivery Queue Size=%04d , WRT_PTR=%04X\n", sz, w_ptr);
|
|
dev_printk(KERN_DEBUG, &pdev->dev,
|
|
"Delivery Queue Base Address=0x%llX (PA)"
|
|
"(tx_dma=0x%llX), Entry=%04d\n",
|
|
addr, mvi->tx_dma, w_ptr);
|
|
mvs_hexdump(sizeof(u32), (u8 *)(&mvi->tx[mvi->tx_prod]),
|
|
(u32) mvi->tx_dma + sizeof(u32) * w_ptr);
|
|
/*Command List */
|
|
addr = mvi->slot_dma;
|
|
dev_printk(KERN_DEBUG, &pdev->dev,
|
|
"Command List Base Address=0x%llX (PA)"
|
|
"(slot_dma=0x%llX), Header=%03d\n",
|
|
addr, slot->buf_dma, tag);
|
|
dev_printk(KERN_DEBUG, &pdev->dev, "Command Header[%03d]:\n", tag);
|
|
/*mvs_cmd_hdr */
|
|
mvs_hexdump(sizeof(struct mvs_cmd_hdr), (u8 *)(&mvi->slot[tag]),
|
|
(u32) mvi->slot_dma + tag * sizeof(struct mvs_cmd_hdr));
|
|
/*1.command table area */
|
|
dev_printk(KERN_DEBUG, &pdev->dev, "+---->Command Table :\n");
|
|
mvs_hexdump(slot->cmd_size, (u8 *) slot->buf, (u32) slot->buf_dma);
|
|
/*2.open address frame area */
|
|
dev_printk(KERN_DEBUG, &pdev->dev, "+---->Open Address Frame :\n");
|
|
mvs_hexdump(MVS_OAF_SZ, (u8 *) slot->buf + slot->cmd_size,
|
|
(u32) slot->buf_dma + slot->cmd_size);
|
|
/*3.status buffer */
|
|
mvs_hba_sb_dump(mvi, tag, proto);
|
|
/*4.PRD table */
|
|
dev_printk(KERN_DEBUG, &pdev->dev, "+---->PRD table :\n");
|
|
mvs_hexdump(sizeof(struct mvs_prd) * slot->n_elem,
|
|
(u8 *) slot->buf + slot->cmd_size + MVS_OAF_SZ,
|
|
(u32) slot->buf_dma + slot->cmd_size + MVS_OAF_SZ);
|
|
#endif
|
|
}
|
|
|
|
static void mvs_hba_cq_dump(struct mvs_info *mvi)
|
|
{
|
|
#if (_MV_DUMP > 2)
|
|
u64 addr;
|
|
void __iomem *regs = mvi->regs;
|
|
struct pci_dev *pdev = mvi->pdev;
|
|
u32 entry = mvi->rx_cons + 1;
|
|
u32 rx_desc = le32_to_cpu(mvi->rx[entry]);
|
|
|
|
/*Completion Queue */
|
|
addr = mr32(RX_HI) << 16 << 16 | mr32(RX_LO);
|
|
dev_printk(KERN_DEBUG, &pdev->dev, "Completion Task = 0x%p\n",
|
|
mvi->slot_info[rx_desc & RXQ_SLOT_MASK].task);
|
|
dev_printk(KERN_DEBUG, &pdev->dev,
|
|
"Completion List Base Address=0x%llX (PA), "
|
|
"CQ_Entry=%04d, CQ_WP=0x%08X\n",
|
|
addr, entry - 1, mvi->rx[0]);
|
|
mvs_hexdump(sizeof(u32), (u8 *)(&rx_desc),
|
|
mvi->rx_dma + sizeof(u32) * entry);
|
|
#endif
|
|
}
|
|
|
|
static void mvs_hba_interrupt_enable(struct mvs_info *mvi)
|
|
{
|
|
void __iomem *regs = mvi->regs;
|
|
u32 tmp;
|
|
|
|
tmp = mr32(GBL_CTL);
|
|
|
|
mw32(GBL_CTL, tmp | INT_EN);
|
|
}
|
|
|
|
static void mvs_hba_interrupt_disable(struct mvs_info *mvi)
|
|
{
|
|
void __iomem *regs = mvi->regs;
|
|
u32 tmp;
|
|
|
|
tmp = mr32(GBL_CTL);
|
|
|
|
mw32(GBL_CTL, tmp & ~INT_EN);
|
|
}
|
|
|
|
static int mvs_int_rx(struct mvs_info *mvi, bool self_clear);
|
|
|
|
/* move to PCI layer or libata core? */
|
|
static int pci_go_64(struct pci_dev *pdev)
|
|
{
|
|
int rc;
|
|
|
|
if (!pci_set_dma_mask(pdev, DMA_64BIT_MASK)) {
|
|
rc = pci_set_consistent_dma_mask(pdev, DMA_64BIT_MASK);
|
|
if (rc) {
|
|
rc = pci_set_consistent_dma_mask(pdev, DMA_32BIT_MASK);
|
|
if (rc) {
|
|
dev_printk(KERN_ERR, &pdev->dev,
|
|
"64-bit DMA enable failed\n");
|
|
return rc;
|
|
}
|
|
}
|
|
} else {
|
|
rc = pci_set_dma_mask(pdev, DMA_32BIT_MASK);
|
|
if (rc) {
|
|
dev_printk(KERN_ERR, &pdev->dev,
|
|
"32-bit DMA enable failed\n");
|
|
return rc;
|
|
}
|
|
rc = pci_set_consistent_dma_mask(pdev, DMA_32BIT_MASK);
|
|
if (rc) {
|
|
dev_printk(KERN_ERR, &pdev->dev,
|
|
"32-bit consistent DMA enable failed\n");
|
|
return rc;
|
|
}
|
|
}
|
|
|
|
return rc;
|
|
}
|
|
|
|
static int mvs_find_tag(struct mvs_info *mvi, struct sas_task *task, u32 *tag)
|
|
{
|
|
if (task->lldd_task) {
|
|
struct mvs_slot_info *slot;
|
|
slot = (struct mvs_slot_info *) task->lldd_task;
|
|
*tag = slot - mvi->slot_info;
|
|
return 1;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static void mvs_tag_clear(struct mvs_info *mvi, u32 tag)
|
|
{
|
|
void *bitmap = (void *) &mvi->tags;
|
|
clear_bit(tag, bitmap);
|
|
}
|
|
|
|
static void mvs_tag_free(struct mvs_info *mvi, u32 tag)
|
|
{
|
|
mvs_tag_clear(mvi, tag);
|
|
}
|
|
|
|
static void mvs_tag_set(struct mvs_info *mvi, unsigned int tag)
|
|
{
|
|
void *bitmap = (void *) &mvi->tags;
|
|
set_bit(tag, bitmap);
|
|
}
|
|
|
|
static int mvs_tag_alloc(struct mvs_info *mvi, u32 *tag_out)
|
|
{
|
|
unsigned int index, tag;
|
|
void *bitmap = (void *) &mvi->tags;
|
|
|
|
index = find_first_zero_bit(bitmap, MVS_SLOTS);
|
|
tag = index;
|
|
if (tag >= MVS_SLOTS)
|
|
return -SAS_QUEUE_FULL;
|
|
mvs_tag_set(mvi, tag);
|
|
*tag_out = tag;
|
|
return 0;
|
|
}
|
|
|
|
static void mvs_tag_init(struct mvs_info *mvi)
|
|
{
|
|
int i;
|
|
for (i = 0; i < MVS_SLOTS; ++i)
|
|
mvs_tag_clear(mvi, i);
|
|
}
|
|
|
|
#ifndef MVS_DISABLE_NVRAM
|
|
static int mvs_eep_read(void __iomem *regs, u32 addr, u32 *data)
|
|
{
|
|
int timeout = 1000;
|
|
|
|
if (addr & ~SPI_ADDR_MASK)
|
|
return -EINVAL;
|
|
|
|
writel(addr, regs + SPI_CMD);
|
|
writel(TWSI_RD, regs + SPI_CTL);
|
|
|
|
while (timeout-- > 0) {
|
|
if (readl(regs + SPI_CTL) & TWSI_RDY) {
|
|
*data = readl(regs + SPI_DATA);
|
|
return 0;
|
|
}
|
|
|
|
udelay(10);
|
|
}
|
|
|
|
return -EBUSY;
|
|
}
|
|
|
|
static int mvs_eep_read_buf(void __iomem *regs, u32 addr,
|
|
void *buf, u32 buflen)
|
|
{
|
|
u32 addr_end, tmp_addr, i, j;
|
|
u32 tmp = 0;
|
|
int rc;
|
|
u8 *tmp8, *buf8 = buf;
|
|
|
|
addr_end = addr + buflen;
|
|
tmp_addr = ALIGN(addr, 4);
|
|
if (addr > 0xff)
|
|
return -EINVAL;
|
|
|
|
j = addr & 0x3;
|
|
if (j) {
|
|
rc = mvs_eep_read(regs, tmp_addr, &tmp);
|
|
if (rc)
|
|
return rc;
|
|
|
|
tmp8 = (u8 *)&tmp;
|
|
for (i = j; i < 4; i++)
|
|
*buf8++ = tmp8[i];
|
|
|
|
tmp_addr += 4;
|
|
}
|
|
|
|
for (j = ALIGN(addr_end, 4); tmp_addr < j; tmp_addr += 4) {
|
|
rc = mvs_eep_read(regs, tmp_addr, &tmp);
|
|
if (rc)
|
|
return rc;
|
|
|
|
memcpy(buf8, &tmp, 4);
|
|
buf8 += 4;
|
|
}
|
|
|
|
if (tmp_addr < addr_end) {
|
|
rc = mvs_eep_read(regs, tmp_addr, &tmp);
|
|
if (rc)
|
|
return rc;
|
|
|
|
tmp8 = (u8 *)&tmp;
|
|
j = addr_end - tmp_addr;
|
|
for (i = 0; i < j; i++)
|
|
*buf8++ = tmp8[i];
|
|
|
|
tmp_addr += 4;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
#endif
|
|
|
|
static int mvs_nvram_read(struct mvs_info *mvi, u32 addr,
|
|
void *buf, u32 buflen)
|
|
{
|
|
#ifndef MVS_DISABLE_NVRAM
|
|
void __iomem *regs = mvi->regs;
|
|
int rc, i;
|
|
u32 sum;
|
|
u8 hdr[2], *tmp;
|
|
const char *msg;
|
|
|
|
rc = mvs_eep_read_buf(regs, addr, &hdr, 2);
|
|
if (rc) {
|
|
msg = "nvram hdr read failed";
|
|
goto err_out;
|
|
}
|
|
rc = mvs_eep_read_buf(regs, addr + 2, buf, buflen);
|
|
if (rc) {
|
|
msg = "nvram read failed";
|
|
goto err_out;
|
|
}
|
|
|
|
if (hdr[0] != 0x5A) {
|
|
/* entry id */
|
|
msg = "invalid nvram entry id";
|
|
rc = -ENOENT;
|
|
goto err_out;
|
|
}
|
|
|
|
tmp = buf;
|
|
sum = ((u32)hdr[0]) + ((u32)hdr[1]);
|
|
for (i = 0; i < buflen; i++)
|
|
sum += ((u32)tmp[i]);
|
|
|
|
if (sum) {
|
|
msg = "nvram checksum failure";
|
|
rc = -EILSEQ;
|
|
goto err_out;
|
|
}
|
|
|
|
return 0;
|
|
|
|
err_out:
|
|
dev_printk(KERN_ERR, &mvi->pdev->dev, "%s", msg);
|
|
return rc;
|
|
#else
|
|
/* FIXME , For SAS target mode */
|
|
memcpy(buf, "\x50\x05\x04\x30\x11\xab\x00\x00", 8);
|
|
return 0;
|
|
#endif
|
|
}
|
|
|
|
static void mvs_bytes_dmaed(struct mvs_info *mvi, int i)
|
|
{
|
|
struct mvs_phy *phy = &mvi->phy[i];
|
|
struct asd_sas_phy *sas_phy = mvi->sas.sas_phy[i];
|
|
|
|
if (!phy->phy_attached)
|
|
return;
|
|
|
|
if (sas_phy->phy) {
|
|
struct sas_phy *sphy = sas_phy->phy;
|
|
|
|
sphy->negotiated_linkrate = sas_phy->linkrate;
|
|
sphy->minimum_linkrate = phy->minimum_linkrate;
|
|
sphy->minimum_linkrate_hw = SAS_LINK_RATE_1_5_GBPS;
|
|
sphy->maximum_linkrate = phy->maximum_linkrate;
|
|
sphy->maximum_linkrate_hw = SAS_LINK_RATE_3_0_GBPS;
|
|
}
|
|
|
|
if (phy->phy_type & PORT_TYPE_SAS) {
|
|
struct sas_identify_frame *id;
|
|
|
|
id = (struct sas_identify_frame *)phy->frame_rcvd;
|
|
id->dev_type = phy->identify.device_type;
|
|
id->initiator_bits = SAS_PROTOCOL_ALL;
|
|
id->target_bits = phy->identify.target_port_protocols;
|
|
} else if (phy->phy_type & PORT_TYPE_SATA) {
|
|
/* TODO */
|
|
}
|
|
mvi->sas.sas_phy[i]->frame_rcvd_size = phy->frame_rcvd_size;
|
|
mvi->sas.notify_port_event(mvi->sas.sas_phy[i],
|
|
PORTE_BYTES_DMAED);
|
|
}
|
|
|
|
static int mvs_scan_finished(struct Scsi_Host *shost, unsigned long time)
|
|
{
|
|
/* give the phy enabling interrupt event time to come in (1s
|
|
* is empirically about all it takes) */
|
|
if (time < HZ)
|
|
return 0;
|
|
/* Wait for discovery to finish */
|
|
scsi_flush_work(shost);
|
|
return 1;
|
|
}
|
|
|
|
static void mvs_scan_start(struct Scsi_Host *shost)
|
|
{
|
|
int i;
|
|
struct mvs_info *mvi = SHOST_TO_SAS_HA(shost)->lldd_ha;
|
|
|
|
for (i = 0; i < mvi->chip->n_phy; ++i) {
|
|
mvs_bytes_dmaed(mvi, i);
|
|
}
|
|
}
|
|
|
|
static int mvs_slave_configure(struct scsi_device *sdev)
|
|
{
|
|
struct domain_device *dev = sdev_to_domain_dev(sdev);
|
|
int ret = sas_slave_configure(sdev);
|
|
|
|
if (ret)
|
|
return ret;
|
|
|
|
if (dev_is_sata(dev)) {
|
|
/* struct ata_port *ap = dev->sata_dev.ap; */
|
|
/* struct ata_device *adev = ap->link.device; */
|
|
|
|
/* clamp at no NCQ for the time being */
|
|
/* adev->flags |= ATA_DFLAG_NCQ_OFF; */
|
|
scsi_adjust_queue_depth(sdev, MSG_SIMPLE_TAG, 1);
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static void mvs_int_port(struct mvs_info *mvi, int phy_no, u32 events)
|
|
{
|
|
struct pci_dev *pdev = mvi->pdev;
|
|
struct sas_ha_struct *sas_ha = &mvi->sas;
|
|
struct mvs_phy *phy = &mvi->phy[phy_no];
|
|
struct asd_sas_phy *sas_phy = &phy->sas_phy;
|
|
|
|
phy->irq_status = mvs_read_port_irq_stat(mvi, phy_no);
|
|
/*
|
|
* events is port event now ,
|
|
* we need check the interrupt status which belongs to per port.
|
|
*/
|
|
dev_printk(KERN_DEBUG, &pdev->dev,
|
|
"Port %d Event = %X\n",
|
|
phy_no, phy->irq_status);
|
|
|
|
if (phy->irq_status & (PHYEV_POOF | PHYEV_DEC_ERR)) {
|
|
mvs_release_task(mvi, phy_no);
|
|
if (!mvs_is_phy_ready(mvi, phy_no)) {
|
|
sas_phy_disconnected(sas_phy);
|
|
sas_ha->notify_phy_event(sas_phy, PHYE_LOSS_OF_SIGNAL);
|
|
dev_printk(KERN_INFO, &pdev->dev,
|
|
"Port %d Unplug Notice\n", phy_no);
|
|
|
|
} else
|
|
mvs_phy_control(sas_phy, PHY_FUNC_LINK_RESET, NULL);
|
|
}
|
|
if (!(phy->irq_status & PHYEV_DEC_ERR)) {
|
|
if (phy->irq_status & PHYEV_COMWAKE) {
|
|
u32 tmp = mvs_read_port_irq_mask(mvi, phy_no);
|
|
mvs_write_port_irq_mask(mvi, phy_no,
|
|
tmp | PHYEV_SIG_FIS);
|
|
}
|
|
if (phy->irq_status & (PHYEV_SIG_FIS | PHYEV_ID_DONE)) {
|
|
phy->phy_status = mvs_is_phy_ready(mvi, phy_no);
|
|
if (phy->phy_status) {
|
|
mvs_detect_porttype(mvi, phy_no);
|
|
|
|
if (phy->phy_type & PORT_TYPE_SATA) {
|
|
u32 tmp = mvs_read_port_irq_mask(mvi,
|
|
phy_no);
|
|
tmp &= ~PHYEV_SIG_FIS;
|
|
mvs_write_port_irq_mask(mvi,
|
|
phy_no, tmp);
|
|
}
|
|
|
|
mvs_update_phyinfo(mvi, phy_no, 0);
|
|
sas_ha->notify_phy_event(sas_phy,
|
|
PHYE_OOB_DONE);
|
|
mvs_bytes_dmaed(mvi, phy_no);
|
|
} else {
|
|
dev_printk(KERN_DEBUG, &pdev->dev,
|
|
"plugin interrupt but phy is gone\n");
|
|
mvs_phy_control(sas_phy, PHY_FUNC_LINK_RESET,
|
|
NULL);
|
|
}
|
|
} else if (phy->irq_status & PHYEV_BROAD_CH) {
|
|
mvs_release_task(mvi, phy_no);
|
|
sas_ha->notify_port_event(sas_phy,
|
|
PORTE_BROADCAST_RCVD);
|
|
}
|
|
}
|
|
mvs_write_port_irq_stat(mvi, phy_no, phy->irq_status);
|
|
}
|
|
|
|
static void mvs_int_sata(struct mvs_info *mvi)
|
|
{
|
|
u32 tmp;
|
|
void __iomem *regs = mvi->regs;
|
|
tmp = mr32(INT_STAT_SRS);
|
|
mw32(INT_STAT_SRS, tmp & 0xFFFF);
|
|
}
|
|
|
|
static void mvs_slot_reset(struct mvs_info *mvi, struct sas_task *task,
|
|
u32 slot_idx)
|
|
{
|
|
void __iomem *regs = mvi->regs;
|
|
struct domain_device *dev = task->dev;
|
|
struct asd_sas_port *sas_port = dev->port;
|
|
struct mvs_port *port = mvi->slot_info[slot_idx].port;
|
|
u32 reg_set, phy_mask;
|
|
|
|
if (!sas_protocol_ata(task->task_proto)) {
|
|
reg_set = 0;
|
|
phy_mask = (port->wide_port_phymap) ? port->wide_port_phymap :
|
|
sas_port->phy_mask;
|
|
} else {
|
|
reg_set = port->taskfileset;
|
|
phy_mask = sas_port->phy_mask;
|
|
}
|
|
mvi->tx[mvi->tx_prod] = cpu_to_le32(TXQ_MODE_I | slot_idx |
|
|
(TXQ_CMD_SLOT_RESET << TXQ_CMD_SHIFT) |
|
|
(phy_mask << TXQ_PHY_SHIFT) |
|
|
(reg_set << TXQ_SRS_SHIFT));
|
|
|
|
mw32(TX_PROD_IDX, mvi->tx_prod);
|
|
mvi->tx_prod = (mvi->tx_prod + 1) & (MVS_CHIP_SLOT_SZ - 1);
|
|
}
|
|
|
|
static int mvs_sata_done(struct mvs_info *mvi, struct sas_task *task,
|
|
u32 slot_idx, int err)
|
|
{
|
|
struct mvs_port *port = mvi->slot_info[slot_idx].port;
|
|
struct task_status_struct *tstat = &task->task_status;
|
|
struct ata_task_resp *resp = (struct ata_task_resp *)tstat->buf;
|
|
int stat = SAM_GOOD;
|
|
|
|
resp->frame_len = sizeof(struct dev_to_host_fis);
|
|
memcpy(&resp->ending_fis[0],
|
|
SATA_RECEIVED_D2H_FIS(port->taskfileset),
|
|
sizeof(struct dev_to_host_fis));
|
|
tstat->buf_valid_size = sizeof(*resp);
|
|
if (unlikely(err))
|
|
stat = SAS_PROTO_RESPONSE;
|
|
return stat;
|
|
}
|
|
|
|
static void mvs_slot_free(struct mvs_info *mvi, u32 rx_desc)
|
|
{
|
|
u32 slot_idx = rx_desc & RXQ_SLOT_MASK;
|
|
mvs_tag_clear(mvi, slot_idx);
|
|
}
|
|
|
|
static void mvs_slot_task_free(struct mvs_info *mvi, struct sas_task *task,
|
|
struct mvs_slot_info *slot, u32 slot_idx)
|
|
{
|
|
if (!sas_protocol_ata(task->task_proto))
|
|
if (slot->n_elem)
|
|
pci_unmap_sg(mvi->pdev, task->scatter,
|
|
slot->n_elem, task->data_dir);
|
|
|
|
switch (task->task_proto) {
|
|
case SAS_PROTOCOL_SMP:
|
|
pci_unmap_sg(mvi->pdev, &task->smp_task.smp_resp, 1,
|
|
PCI_DMA_FROMDEVICE);
|
|
pci_unmap_sg(mvi->pdev, &task->smp_task.smp_req, 1,
|
|
PCI_DMA_TODEVICE);
|
|
break;
|
|
|
|
case SAS_PROTOCOL_SATA:
|
|
case SAS_PROTOCOL_STP:
|
|
case SAS_PROTOCOL_SSP:
|
|
default:
|
|
/* do nothing */
|
|
break;
|
|
}
|
|
list_del(&slot->list);
|
|
task->lldd_task = NULL;
|
|
slot->task = NULL;
|
|
slot->port = NULL;
|
|
}
|
|
|
|
static int mvs_slot_err(struct mvs_info *mvi, struct sas_task *task,
|
|
u32 slot_idx)
|
|
{
|
|
struct mvs_slot_info *slot = &mvi->slot_info[slot_idx];
|
|
u32 err_dw0 = le32_to_cpu(*(u32 *) (slot->response));
|
|
u32 err_dw1 = le32_to_cpu(*(u32 *) (slot->response + 4));
|
|
int stat = SAM_CHECK_COND;
|
|
|
|
if (err_dw1 & SLOT_BSY_ERR) {
|
|
stat = SAS_QUEUE_FULL;
|
|
mvs_slot_reset(mvi, task, slot_idx);
|
|
}
|
|
switch (task->task_proto) {
|
|
case SAS_PROTOCOL_SSP:
|
|
break;
|
|
case SAS_PROTOCOL_SMP:
|
|
break;
|
|
case SAS_PROTOCOL_SATA:
|
|
case SAS_PROTOCOL_STP:
|
|
case SAS_PROTOCOL_SATA | SAS_PROTOCOL_STP:
|
|
if (err_dw0 & TFILE_ERR)
|
|
stat = mvs_sata_done(mvi, task, slot_idx, 1);
|
|
break;
|
|
default:
|
|
break;
|
|
}
|
|
|
|
mvs_hexdump(16, (u8 *) slot->response, 0);
|
|
return stat;
|
|
}
|
|
|
|
static int mvs_slot_complete(struct mvs_info *mvi, u32 rx_desc, u32 flags)
|
|
{
|
|
u32 slot_idx = rx_desc & RXQ_SLOT_MASK;
|
|
struct mvs_slot_info *slot = &mvi->slot_info[slot_idx];
|
|
struct sas_task *task = slot->task;
|
|
struct task_status_struct *tstat;
|
|
struct mvs_port *port;
|
|
bool aborted;
|
|
void *to;
|
|
|
|
if (unlikely(!task || !task->lldd_task))
|
|
return -1;
|
|
|
|
mvs_hba_cq_dump(mvi);
|
|
|
|
spin_lock(&task->task_state_lock);
|
|
aborted = task->task_state_flags & SAS_TASK_STATE_ABORTED;
|
|
if (!aborted) {
|
|
task->task_state_flags &=
|
|
~(SAS_TASK_STATE_PENDING | SAS_TASK_AT_INITIATOR);
|
|
task->task_state_flags |= SAS_TASK_STATE_DONE;
|
|
}
|
|
spin_unlock(&task->task_state_lock);
|
|
|
|
if (aborted) {
|
|
mvs_slot_task_free(mvi, task, slot, slot_idx);
|
|
mvs_slot_free(mvi, rx_desc);
|
|
return -1;
|
|
}
|
|
|
|
port = slot->port;
|
|
tstat = &task->task_status;
|
|
memset(tstat, 0, sizeof(*tstat));
|
|
tstat->resp = SAS_TASK_COMPLETE;
|
|
|
|
if (unlikely(!port->port_attached || flags)) {
|
|
mvs_slot_err(mvi, task, slot_idx);
|
|
if (!sas_protocol_ata(task->task_proto))
|
|
tstat->stat = SAS_PHY_DOWN;
|
|
goto out;
|
|
}
|
|
|
|
/* error info record present */
|
|
if (unlikely((rx_desc & RXQ_ERR) && (*(u64 *) slot->response))) {
|
|
tstat->stat = mvs_slot_err(mvi, task, slot_idx);
|
|
goto out;
|
|
}
|
|
|
|
switch (task->task_proto) {
|
|
case SAS_PROTOCOL_SSP:
|
|
/* hw says status == 0, datapres == 0 */
|
|
if (rx_desc & RXQ_GOOD) {
|
|
tstat->stat = SAM_GOOD;
|
|
tstat->resp = SAS_TASK_COMPLETE;
|
|
}
|
|
/* response frame present */
|
|
else if (rx_desc & RXQ_RSP) {
|
|
struct ssp_response_iu *iu =
|
|
slot->response + sizeof(struct mvs_err_info);
|
|
sas_ssp_task_response(&mvi->pdev->dev, task, iu);
|
|
}
|
|
|
|
/* should never happen? */
|
|
else
|
|
tstat->stat = SAM_CHECK_COND;
|
|
break;
|
|
|
|
case SAS_PROTOCOL_SMP: {
|
|
struct scatterlist *sg_resp = &task->smp_task.smp_resp;
|
|
tstat->stat = SAM_GOOD;
|
|
to = kmap_atomic(sg_page(sg_resp), KM_IRQ0);
|
|
memcpy(to + sg_resp->offset,
|
|
slot->response + sizeof(struct mvs_err_info),
|
|
sg_dma_len(sg_resp));
|
|
kunmap_atomic(to, KM_IRQ0);
|
|
break;
|
|
}
|
|
|
|
case SAS_PROTOCOL_SATA:
|
|
case SAS_PROTOCOL_STP:
|
|
case SAS_PROTOCOL_SATA | SAS_PROTOCOL_STP: {
|
|
tstat->stat = mvs_sata_done(mvi, task, slot_idx, 0);
|
|
break;
|
|
}
|
|
|
|
default:
|
|
tstat->stat = SAM_CHECK_COND;
|
|
break;
|
|
}
|
|
|
|
out:
|
|
mvs_slot_task_free(mvi, task, slot, slot_idx);
|
|
if (unlikely(tstat->stat != SAS_QUEUE_FULL))
|
|
mvs_slot_free(mvi, rx_desc);
|
|
|
|
spin_unlock(&mvi->lock);
|
|
task->task_done(task);
|
|
spin_lock(&mvi->lock);
|
|
return tstat->stat;
|
|
}
|
|
|
|
static void mvs_release_task(struct mvs_info *mvi, int phy_no)
|
|
{
|
|
struct list_head *pos, *n;
|
|
struct mvs_slot_info *slot;
|
|
struct mvs_phy *phy = &mvi->phy[phy_no];
|
|
struct mvs_port *port = phy->port;
|
|
u32 rx_desc;
|
|
|
|
if (!port)
|
|
return;
|
|
|
|
list_for_each_safe(pos, n, &port->list) {
|
|
slot = container_of(pos, struct mvs_slot_info, list);
|
|
rx_desc = (u32) (slot - mvi->slot_info);
|
|
mvs_slot_complete(mvi, rx_desc, 1);
|
|
}
|
|
}
|
|
|
|
static void mvs_int_full(struct mvs_info *mvi)
|
|
{
|
|
void __iomem *regs = mvi->regs;
|
|
u32 tmp, stat;
|
|
int i;
|
|
|
|
stat = mr32(INT_STAT);
|
|
|
|
mvs_int_rx(mvi, false);
|
|
|
|
for (i = 0; i < MVS_MAX_PORTS; i++) {
|
|
tmp = (stat >> i) & (CINT_PORT | CINT_PORT_STOPPED);
|
|
if (tmp)
|
|
mvs_int_port(mvi, i, tmp);
|
|
}
|
|
|
|
if (stat & CINT_SRS)
|
|
mvs_int_sata(mvi);
|
|
|
|
mw32(INT_STAT, stat);
|
|
}
|
|
|
|
static int mvs_int_rx(struct mvs_info *mvi, bool self_clear)
|
|
{
|
|
void __iomem *regs = mvi->regs;
|
|
u32 rx_prod_idx, rx_desc;
|
|
bool attn = false;
|
|
struct pci_dev *pdev = mvi->pdev;
|
|
|
|
/* the first dword in the RX ring is special: it contains
|
|
* a mirror of the hardware's RX producer index, so that
|
|
* we don't have to stall the CPU reading that register.
|
|
* The actual RX ring is offset by one dword, due to this.
|
|
*/
|
|
rx_prod_idx = mvi->rx_cons;
|
|
mvi->rx_cons = le32_to_cpu(mvi->rx[0]);
|
|
if (mvi->rx_cons == 0xfff) /* h/w hasn't touched RX ring yet */
|
|
return 0;
|
|
|
|
/* The CMPL_Q may come late, read from register and try again
|
|
* note: if coalescing is enabled,
|
|
* it will need to read from register every time for sure
|
|
*/
|
|
if (mvi->rx_cons == rx_prod_idx)
|
|
mvi->rx_cons = mr32(RX_CONS_IDX) & RX_RING_SZ_MASK;
|
|
|
|
if (mvi->rx_cons == rx_prod_idx)
|
|
return 0;
|
|
|
|
while (mvi->rx_cons != rx_prod_idx) {
|
|
|
|
/* increment our internal RX consumer pointer */
|
|
rx_prod_idx = (rx_prod_idx + 1) & (MVS_RX_RING_SZ - 1);
|
|
|
|
rx_desc = le32_to_cpu(mvi->rx[rx_prod_idx + 1]);
|
|
|
|
if (likely(rx_desc & RXQ_DONE))
|
|
mvs_slot_complete(mvi, rx_desc, 0);
|
|
if (rx_desc & RXQ_ATTN) {
|
|
attn = true;
|
|
dev_printk(KERN_DEBUG, &pdev->dev, "ATTN %X\n",
|
|
rx_desc);
|
|
} else if (rx_desc & RXQ_ERR) {
|
|
if (!(rx_desc & RXQ_DONE))
|
|
mvs_slot_complete(mvi, rx_desc, 0);
|
|
dev_printk(KERN_DEBUG, &pdev->dev, "RXQ_ERR %X\n",
|
|
rx_desc);
|
|
} else if (rx_desc & RXQ_SLOT_RESET) {
|
|
dev_printk(KERN_DEBUG, &pdev->dev, "Slot reset[%X]\n",
|
|
rx_desc);
|
|
mvs_slot_free(mvi, rx_desc);
|
|
}
|
|
}
|
|
|
|
if (attn && self_clear)
|
|
mvs_int_full(mvi);
|
|
|
|
return 0;
|
|
}
|
|
|
|
#ifdef MVS_USE_TASKLET
|
|
static void mvs_tasklet(unsigned long data)
|
|
{
|
|
struct mvs_info *mvi = (struct mvs_info *) data;
|
|
unsigned long flags;
|
|
|
|
spin_lock_irqsave(&mvi->lock, flags);
|
|
|
|
#ifdef MVS_DISABLE_MSI
|
|
mvs_int_full(mvi);
|
|
#else
|
|
mvs_int_rx(mvi, true);
|
|
#endif
|
|
spin_unlock_irqrestore(&mvi->lock, flags);
|
|
}
|
|
#endif
|
|
|
|
static irqreturn_t mvs_interrupt(int irq, void *opaque)
|
|
{
|
|
struct mvs_info *mvi = opaque;
|
|
void __iomem *regs = mvi->regs;
|
|
u32 stat;
|
|
|
|
stat = mr32(GBL_INT_STAT);
|
|
|
|
if (stat == 0 || stat == 0xffffffff)
|
|
return IRQ_NONE;
|
|
|
|
/* clear CMD_CMPLT ASAP */
|
|
mw32_f(INT_STAT, CINT_DONE);
|
|
|
|
#ifndef MVS_USE_TASKLET
|
|
spin_lock(&mvi->lock);
|
|
|
|
mvs_int_full(mvi);
|
|
|
|
spin_unlock(&mvi->lock);
|
|
#else
|
|
tasklet_schedule(&mvi->tasklet);
|
|
#endif
|
|
return IRQ_HANDLED;
|
|
}
|
|
|
|
#ifndef MVS_DISABLE_MSI
|
|
static irqreturn_t mvs_msi_interrupt(int irq, void *opaque)
|
|
{
|
|
struct mvs_info *mvi = opaque;
|
|
|
|
#ifndef MVS_USE_TASKLET
|
|
spin_lock(&mvi->lock);
|
|
|
|
mvs_int_rx(mvi, true);
|
|
|
|
spin_unlock(&mvi->lock);
|
|
#else
|
|
tasklet_schedule(&mvi->tasklet);
|
|
#endif
|
|
return IRQ_HANDLED;
|
|
}
|
|
#endif
|
|
|
|
struct mvs_task_exec_info {
|
|
struct sas_task *task;
|
|
struct mvs_cmd_hdr *hdr;
|
|
struct mvs_port *port;
|
|
u32 tag;
|
|
int n_elem;
|
|
};
|
|
|
|
static int mvs_task_prep_smp(struct mvs_info *mvi,
|
|
struct mvs_task_exec_info *tei)
|
|
{
|
|
int elem, rc, i;
|
|
struct sas_task *task = tei->task;
|
|
struct mvs_cmd_hdr *hdr = tei->hdr;
|
|
struct scatterlist *sg_req, *sg_resp;
|
|
u32 req_len, resp_len, tag = tei->tag;
|
|
void *buf_tmp;
|
|
u8 *buf_oaf;
|
|
dma_addr_t buf_tmp_dma;
|
|
struct mvs_prd *buf_prd;
|
|
struct scatterlist *sg;
|
|
struct mvs_slot_info *slot = &mvi->slot_info[tag];
|
|
struct asd_sas_port *sas_port = task->dev->port;
|
|
u32 flags = (tei->n_elem << MCH_PRD_LEN_SHIFT);
|
|
#if _MV_DUMP
|
|
u8 *buf_cmd;
|
|
void *from;
|
|
#endif
|
|
/*
|
|
* DMA-map SMP request, response buffers
|
|
*/
|
|
sg_req = &task->smp_task.smp_req;
|
|
elem = pci_map_sg(mvi->pdev, sg_req, 1, PCI_DMA_TODEVICE);
|
|
if (!elem)
|
|
return -ENOMEM;
|
|
req_len = sg_dma_len(sg_req);
|
|
|
|
sg_resp = &task->smp_task.smp_resp;
|
|
elem = pci_map_sg(mvi->pdev, sg_resp, 1, PCI_DMA_FROMDEVICE);
|
|
if (!elem) {
|
|
rc = -ENOMEM;
|
|
goto err_out;
|
|
}
|
|
resp_len = sg_dma_len(sg_resp);
|
|
|
|
/* must be in dwords */
|
|
if ((req_len & 0x3) || (resp_len & 0x3)) {
|
|
rc = -EINVAL;
|
|
goto err_out_2;
|
|
}
|
|
|
|
/*
|
|
* arrange MVS_SLOT_BUF_SZ-sized DMA buffer according to our needs
|
|
*/
|
|
|
|
/* region 1: command table area (MVS_SSP_CMD_SZ bytes) ************** */
|
|
buf_tmp = slot->buf;
|
|
buf_tmp_dma = slot->buf_dma;
|
|
|
|
#if _MV_DUMP
|
|
buf_cmd = buf_tmp;
|
|
hdr->cmd_tbl = cpu_to_le64(buf_tmp_dma);
|
|
buf_tmp += req_len;
|
|
buf_tmp_dma += req_len;
|
|
slot->cmd_size = req_len;
|
|
#else
|
|
hdr->cmd_tbl = cpu_to_le64(sg_dma_address(sg_req));
|
|
#endif
|
|
|
|
/* region 2: open address frame area (MVS_OAF_SZ bytes) ********* */
|
|
buf_oaf = buf_tmp;
|
|
hdr->open_frame = cpu_to_le64(buf_tmp_dma);
|
|
|
|
buf_tmp += MVS_OAF_SZ;
|
|
buf_tmp_dma += MVS_OAF_SZ;
|
|
|
|
/* region 3: PRD table ********************************************* */
|
|
buf_prd = buf_tmp;
|
|
if (tei->n_elem)
|
|
hdr->prd_tbl = cpu_to_le64(buf_tmp_dma);
|
|
else
|
|
hdr->prd_tbl = 0;
|
|
|
|
i = sizeof(struct mvs_prd) * tei->n_elem;
|
|
buf_tmp += i;
|
|
buf_tmp_dma += i;
|
|
|
|
/* region 4: status buffer (larger the PRD, smaller this buf) ****** */
|
|
slot->response = buf_tmp;
|
|
hdr->status_buf = cpu_to_le64(buf_tmp_dma);
|
|
|
|
/*
|
|
* Fill in TX ring and command slot header
|
|
*/
|
|
slot->tx = mvi->tx_prod;
|
|
mvi->tx[mvi->tx_prod] = cpu_to_le32((TXQ_CMD_SMP << TXQ_CMD_SHIFT) |
|
|
TXQ_MODE_I | tag |
|
|
(sas_port->phy_mask << TXQ_PHY_SHIFT));
|
|
|
|
hdr->flags |= flags;
|
|
hdr->lens = cpu_to_le32(((resp_len / 4) << 16) | ((req_len - 4) / 4));
|
|
hdr->tags = cpu_to_le32(tag);
|
|
hdr->data_len = 0;
|
|
|
|
/* generate open address frame hdr (first 12 bytes) */
|
|
buf_oaf[0] = (1 << 7) | (0 << 4) | 0x01; /* initiator, SMP, ftype 1h */
|
|
buf_oaf[1] = task->dev->linkrate & 0xf;
|
|
*(u16 *)(buf_oaf + 2) = 0xFFFF; /* SAS SPEC */
|
|
memcpy(buf_oaf + 4, task->dev->sas_addr, SAS_ADDR_SIZE);
|
|
|
|
/* fill in PRD (scatter/gather) table, if any */
|
|
for_each_sg(task->scatter, sg, tei->n_elem, i) {
|
|
buf_prd->addr = cpu_to_le64(sg_dma_address(sg));
|
|
buf_prd->len = cpu_to_le32(sg_dma_len(sg));
|
|
buf_prd++;
|
|
}
|
|
|
|
#if _MV_DUMP
|
|
/* copy cmd table */
|
|
from = kmap_atomic(sg_page(sg_req), KM_IRQ0);
|
|
memcpy(buf_cmd, from + sg_req->offset, req_len);
|
|
kunmap_atomic(from, KM_IRQ0);
|
|
#endif
|
|
return 0;
|
|
|
|
err_out_2:
|
|
pci_unmap_sg(mvi->pdev, &tei->task->smp_task.smp_resp, 1,
|
|
PCI_DMA_FROMDEVICE);
|
|
err_out:
|
|
pci_unmap_sg(mvi->pdev, &tei->task->smp_task.smp_req, 1,
|
|
PCI_DMA_TODEVICE);
|
|
return rc;
|
|
}
|
|
|
|
static void mvs_free_reg_set(struct mvs_info *mvi, struct mvs_port *port)
|
|
{
|
|
void __iomem *regs = mvi->regs;
|
|
u32 tmp, offs;
|
|
u8 *tfs = &port->taskfileset;
|
|
|
|
if (*tfs == MVS_ID_NOT_MAPPED)
|
|
return;
|
|
|
|
offs = 1U << ((*tfs & 0x0f) + PCS_EN_SATA_REG_SHIFT);
|
|
if (*tfs < 16) {
|
|
tmp = mr32(PCS);
|
|
mw32(PCS, tmp & ~offs);
|
|
} else {
|
|
tmp = mr32(CTL);
|
|
mw32(CTL, tmp & ~offs);
|
|
}
|
|
|
|
tmp = mr32(INT_STAT_SRS) & (1U << *tfs);
|
|
if (tmp)
|
|
mw32(INT_STAT_SRS, tmp);
|
|
|
|
*tfs = MVS_ID_NOT_MAPPED;
|
|
}
|
|
|
|
static u8 mvs_assign_reg_set(struct mvs_info *mvi, struct mvs_port *port)
|
|
{
|
|
int i;
|
|
u32 tmp, offs;
|
|
void __iomem *regs = mvi->regs;
|
|
|
|
if (port->taskfileset != MVS_ID_NOT_MAPPED)
|
|
return 0;
|
|
|
|
tmp = mr32(PCS);
|
|
|
|
for (i = 0; i < mvi->chip->srs_sz; i++) {
|
|
if (i == 16)
|
|
tmp = mr32(CTL);
|
|
offs = 1U << ((i & 0x0f) + PCS_EN_SATA_REG_SHIFT);
|
|
if (!(tmp & offs)) {
|
|
port->taskfileset = i;
|
|
|
|
if (i < 16)
|
|
mw32(PCS, tmp | offs);
|
|
else
|
|
mw32(CTL, tmp | offs);
|
|
tmp = mr32(INT_STAT_SRS) & (1U << i);
|
|
if (tmp)
|
|
mw32(INT_STAT_SRS, tmp);
|
|
return 0;
|
|
}
|
|
}
|
|
return MVS_ID_NOT_MAPPED;
|
|
}
|
|
|
|
static u32 mvs_get_ncq_tag(struct sas_task *task, u32 *tag)
|
|
{
|
|
struct ata_queued_cmd *qc = task->uldd_task;
|
|
|
|
if (qc) {
|
|
if (qc->tf.command == ATA_CMD_FPDMA_WRITE ||
|
|
qc->tf.command == ATA_CMD_FPDMA_READ) {
|
|
*tag = qc->tag;
|
|
return 1;
|
|
}
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int mvs_task_prep_ata(struct mvs_info *mvi,
|
|
struct mvs_task_exec_info *tei)
|
|
{
|
|
struct sas_task *task = tei->task;
|
|
struct domain_device *dev = task->dev;
|
|
struct mvs_cmd_hdr *hdr = tei->hdr;
|
|
struct asd_sas_port *sas_port = dev->port;
|
|
struct mvs_slot_info *slot;
|
|
struct scatterlist *sg;
|
|
struct mvs_prd *buf_prd;
|
|
struct mvs_port *port = tei->port;
|
|
u32 tag = tei->tag;
|
|
u32 flags = (tei->n_elem << MCH_PRD_LEN_SHIFT);
|
|
void *buf_tmp;
|
|
u8 *buf_cmd, *buf_oaf;
|
|
dma_addr_t buf_tmp_dma;
|
|
u32 i, req_len, resp_len;
|
|
const u32 max_resp_len = SB_RFB_MAX;
|
|
|
|
if (mvs_assign_reg_set(mvi, port) == MVS_ID_NOT_MAPPED)
|
|
return -EBUSY;
|
|
|
|
slot = &mvi->slot_info[tag];
|
|
slot->tx = mvi->tx_prod;
|
|
mvi->tx[mvi->tx_prod] = cpu_to_le32(TXQ_MODE_I | tag |
|
|
(TXQ_CMD_STP << TXQ_CMD_SHIFT) |
|
|
(sas_port->phy_mask << TXQ_PHY_SHIFT) |
|
|
(port->taskfileset << TXQ_SRS_SHIFT));
|
|
|
|
if (task->ata_task.use_ncq)
|
|
flags |= MCH_FPDMA;
|
|
if (dev->sata_dev.command_set == ATAPI_COMMAND_SET) {
|
|
if (task->ata_task.fis.command != ATA_CMD_ID_ATAPI)
|
|
flags |= MCH_ATAPI;
|
|
}
|
|
|
|
/* FIXME: fill in port multiplier number */
|
|
|
|
hdr->flags = cpu_to_le32(flags);
|
|
|
|
/* FIXME: the low order order 5 bits for the TAG if enable NCQ */
|
|
if (task->ata_task.use_ncq && mvs_get_ncq_tag(task, &hdr->tags))
|
|
task->ata_task.fis.sector_count |= hdr->tags << 3;
|
|
else
|
|
hdr->tags = cpu_to_le32(tag);
|
|
hdr->data_len = cpu_to_le32(task->total_xfer_len);
|
|
|
|
/*
|
|
* arrange MVS_SLOT_BUF_SZ-sized DMA buffer according to our needs
|
|
*/
|
|
|
|
/* region 1: command table area (MVS_ATA_CMD_SZ bytes) ************** */
|
|
buf_cmd = buf_tmp = slot->buf;
|
|
buf_tmp_dma = slot->buf_dma;
|
|
|
|
hdr->cmd_tbl = cpu_to_le64(buf_tmp_dma);
|
|
|
|
buf_tmp += MVS_ATA_CMD_SZ;
|
|
buf_tmp_dma += MVS_ATA_CMD_SZ;
|
|
#if _MV_DUMP
|
|
slot->cmd_size = MVS_ATA_CMD_SZ;
|
|
#endif
|
|
|
|
/* region 2: open address frame area (MVS_OAF_SZ bytes) ********* */
|
|
/* used for STP. unused for SATA? */
|
|
buf_oaf = buf_tmp;
|
|
hdr->open_frame = cpu_to_le64(buf_tmp_dma);
|
|
|
|
buf_tmp += MVS_OAF_SZ;
|
|
buf_tmp_dma += MVS_OAF_SZ;
|
|
|
|
/* region 3: PRD table ********************************************* */
|
|
buf_prd = buf_tmp;
|
|
if (tei->n_elem)
|
|
hdr->prd_tbl = cpu_to_le64(buf_tmp_dma);
|
|
else
|
|
hdr->prd_tbl = 0;
|
|
|
|
i = sizeof(struct mvs_prd) * tei->n_elem;
|
|
buf_tmp += i;
|
|
buf_tmp_dma += i;
|
|
|
|
/* region 4: status buffer (larger the PRD, smaller this buf) ****** */
|
|
/* FIXME: probably unused, for SATA. kept here just in case
|
|
* we get a STP/SATA error information record
|
|
*/
|
|
slot->response = buf_tmp;
|
|
hdr->status_buf = cpu_to_le64(buf_tmp_dma);
|
|
|
|
req_len = sizeof(struct host_to_dev_fis);
|
|
resp_len = MVS_SLOT_BUF_SZ - MVS_ATA_CMD_SZ -
|
|
sizeof(struct mvs_err_info) - i;
|
|
|
|
/* request, response lengths */
|
|
resp_len = min(resp_len, max_resp_len);
|
|
hdr->lens = cpu_to_le32(((resp_len / 4) << 16) | (req_len / 4));
|
|
|
|
task->ata_task.fis.flags |= 0x80; /* C=1: update ATA cmd reg */
|
|
/* fill in command FIS and ATAPI CDB */
|
|
memcpy(buf_cmd, &task->ata_task.fis, sizeof(struct host_to_dev_fis));
|
|
if (dev->sata_dev.command_set == ATAPI_COMMAND_SET)
|
|
memcpy(buf_cmd + STP_ATAPI_CMD,
|
|
task->ata_task.atapi_packet, 16);
|
|
|
|
/* generate open address frame hdr (first 12 bytes) */
|
|
buf_oaf[0] = (1 << 7) | (2 << 4) | 0x1; /* initiator, STP, ftype 1h */
|
|
buf_oaf[1] = task->dev->linkrate & 0xf;
|
|
*(u16 *)(buf_oaf + 2) = cpu_to_be16(tag);
|
|
memcpy(buf_oaf + 4, task->dev->sas_addr, SAS_ADDR_SIZE);
|
|
|
|
/* fill in PRD (scatter/gather) table, if any */
|
|
for_each_sg(task->scatter, sg, tei->n_elem, i) {
|
|
buf_prd->addr = cpu_to_le64(sg_dma_address(sg));
|
|
buf_prd->len = cpu_to_le32(sg_dma_len(sg));
|
|
buf_prd++;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int mvs_task_prep_ssp(struct mvs_info *mvi,
|
|
struct mvs_task_exec_info *tei)
|
|
{
|
|
struct sas_task *task = tei->task;
|
|
struct mvs_cmd_hdr *hdr = tei->hdr;
|
|
struct mvs_port *port = tei->port;
|
|
struct mvs_slot_info *slot;
|
|
struct scatterlist *sg;
|
|
struct mvs_prd *buf_prd;
|
|
struct ssp_frame_hdr *ssp_hdr;
|
|
void *buf_tmp;
|
|
u8 *buf_cmd, *buf_oaf, fburst = 0;
|
|
dma_addr_t buf_tmp_dma;
|
|
u32 flags;
|
|
u32 resp_len, req_len, i, tag = tei->tag;
|
|
const u32 max_resp_len = SB_RFB_MAX;
|
|
u8 phy_mask;
|
|
|
|
slot = &mvi->slot_info[tag];
|
|
|
|
phy_mask = (port->wide_port_phymap) ? port->wide_port_phymap :
|
|
task->dev->port->phy_mask;
|
|
slot->tx = mvi->tx_prod;
|
|
mvi->tx[mvi->tx_prod] = cpu_to_le32(TXQ_MODE_I | tag |
|
|
(TXQ_CMD_SSP << TXQ_CMD_SHIFT) |
|
|
(phy_mask << TXQ_PHY_SHIFT));
|
|
|
|
flags = MCH_RETRY;
|
|
if (task->ssp_task.enable_first_burst) {
|
|
flags |= MCH_FBURST;
|
|
fburst = (1 << 7);
|
|
}
|
|
hdr->flags = cpu_to_le32(flags |
|
|
(tei->n_elem << MCH_PRD_LEN_SHIFT) |
|
|
(MCH_SSP_FR_CMD << MCH_SSP_FR_TYPE_SHIFT));
|
|
|
|
hdr->tags = cpu_to_le32(tag);
|
|
hdr->data_len = cpu_to_le32(task->total_xfer_len);
|
|
|
|
/*
|
|
* arrange MVS_SLOT_BUF_SZ-sized DMA buffer according to our needs
|
|
*/
|
|
|
|
/* region 1: command table area (MVS_SSP_CMD_SZ bytes) ************** */
|
|
buf_cmd = buf_tmp = slot->buf;
|
|
buf_tmp_dma = slot->buf_dma;
|
|
|
|
hdr->cmd_tbl = cpu_to_le64(buf_tmp_dma);
|
|
|
|
buf_tmp += MVS_SSP_CMD_SZ;
|
|
buf_tmp_dma += MVS_SSP_CMD_SZ;
|
|
#if _MV_DUMP
|
|
slot->cmd_size = MVS_SSP_CMD_SZ;
|
|
#endif
|
|
|
|
/* region 2: open address frame area (MVS_OAF_SZ bytes) ********* */
|
|
buf_oaf = buf_tmp;
|
|
hdr->open_frame = cpu_to_le64(buf_tmp_dma);
|
|
|
|
buf_tmp += MVS_OAF_SZ;
|
|
buf_tmp_dma += MVS_OAF_SZ;
|
|
|
|
/* region 3: PRD table ********************************************* */
|
|
buf_prd = buf_tmp;
|
|
if (tei->n_elem)
|
|
hdr->prd_tbl = cpu_to_le64(buf_tmp_dma);
|
|
else
|
|
hdr->prd_tbl = 0;
|
|
|
|
i = sizeof(struct mvs_prd) * tei->n_elem;
|
|
buf_tmp += i;
|
|
buf_tmp_dma += i;
|
|
|
|
/* region 4: status buffer (larger the PRD, smaller this buf) ****** */
|
|
slot->response = buf_tmp;
|
|
hdr->status_buf = cpu_to_le64(buf_tmp_dma);
|
|
|
|
resp_len = MVS_SLOT_BUF_SZ - MVS_SSP_CMD_SZ - MVS_OAF_SZ -
|
|
sizeof(struct mvs_err_info) - i;
|
|
resp_len = min(resp_len, max_resp_len);
|
|
|
|
req_len = sizeof(struct ssp_frame_hdr) + 28;
|
|
|
|
/* request, response lengths */
|
|
hdr->lens = cpu_to_le32(((resp_len / 4) << 16) | (req_len / 4));
|
|
|
|
/* generate open address frame hdr (first 12 bytes) */
|
|
buf_oaf[0] = (1 << 7) | (1 << 4) | 0x1; /* initiator, SSP, ftype 1h */
|
|
buf_oaf[1] = task->dev->linkrate & 0xf;
|
|
*(u16 *)(buf_oaf + 2) = cpu_to_be16(tag);
|
|
memcpy(buf_oaf + 4, task->dev->sas_addr, SAS_ADDR_SIZE);
|
|
|
|
/* fill in SSP frame header (Command Table.SSP frame header) */
|
|
ssp_hdr = (struct ssp_frame_hdr *)buf_cmd;
|
|
ssp_hdr->frame_type = SSP_COMMAND;
|
|
memcpy(ssp_hdr->hashed_dest_addr, task->dev->hashed_sas_addr,
|
|
HASHED_SAS_ADDR_SIZE);
|
|
memcpy(ssp_hdr->hashed_src_addr,
|
|
task->dev->port->ha->hashed_sas_addr, HASHED_SAS_ADDR_SIZE);
|
|
ssp_hdr->tag = cpu_to_be16(tag);
|
|
|
|
/* fill in command frame IU */
|
|
buf_cmd += sizeof(*ssp_hdr);
|
|
memcpy(buf_cmd, &task->ssp_task.LUN, 8);
|
|
buf_cmd[9] = fburst | task->ssp_task.task_attr |
|
|
(task->ssp_task.task_prio << 3);
|
|
memcpy(buf_cmd + 12, &task->ssp_task.cdb, 16);
|
|
|
|
/* fill in PRD (scatter/gather) table, if any */
|
|
for_each_sg(task->scatter, sg, tei->n_elem, i) {
|
|
buf_prd->addr = cpu_to_le64(sg_dma_address(sg));
|
|
buf_prd->len = cpu_to_le32(sg_dma_len(sg));
|
|
buf_prd++;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int mvs_task_exec(struct sas_task *task, const int num, gfp_t gfp_flags)
|
|
{
|
|
struct domain_device *dev = task->dev;
|
|
struct mvs_info *mvi = dev->port->ha->lldd_ha;
|
|
struct pci_dev *pdev = mvi->pdev;
|
|
void __iomem *regs = mvi->regs;
|
|
struct mvs_task_exec_info tei;
|
|
struct sas_task *t = task;
|
|
struct mvs_slot_info *slot;
|
|
u32 tag = 0xdeadbeef, rc, n_elem = 0;
|
|
unsigned long flags;
|
|
u32 n = num, pass = 0;
|
|
|
|
spin_lock_irqsave(&mvi->lock, flags);
|
|
do {
|
|
dev = t->dev;
|
|
tei.port = &mvi->port[dev->port->id];
|
|
|
|
if (!tei.port->port_attached) {
|
|
if (sas_protocol_ata(t->task_proto)) {
|
|
rc = SAS_PHY_DOWN;
|
|
goto out_done;
|
|
} else {
|
|
struct task_status_struct *ts = &t->task_status;
|
|
ts->resp = SAS_TASK_UNDELIVERED;
|
|
ts->stat = SAS_PHY_DOWN;
|
|
t->task_done(t);
|
|
if (n > 1)
|
|
t = list_entry(t->list.next,
|
|
struct sas_task, list);
|
|
continue;
|
|
}
|
|
}
|
|
|
|
if (!sas_protocol_ata(t->task_proto)) {
|
|
if (t->num_scatter) {
|
|
n_elem = pci_map_sg(mvi->pdev, t->scatter,
|
|
t->num_scatter,
|
|
t->data_dir);
|
|
if (!n_elem) {
|
|
rc = -ENOMEM;
|
|
goto err_out;
|
|
}
|
|
}
|
|
} else {
|
|
n_elem = t->num_scatter;
|
|
}
|
|
|
|
rc = mvs_tag_alloc(mvi, &tag);
|
|
if (rc)
|
|
goto err_out;
|
|
|
|
slot = &mvi->slot_info[tag];
|
|
t->lldd_task = NULL;
|
|
slot->n_elem = n_elem;
|
|
memset(slot->buf, 0, MVS_SLOT_BUF_SZ);
|
|
tei.task = t;
|
|
tei.hdr = &mvi->slot[tag];
|
|
tei.tag = tag;
|
|
tei.n_elem = n_elem;
|
|
|
|
switch (t->task_proto) {
|
|
case SAS_PROTOCOL_SMP:
|
|
rc = mvs_task_prep_smp(mvi, &tei);
|
|
break;
|
|
case SAS_PROTOCOL_SSP:
|
|
rc = mvs_task_prep_ssp(mvi, &tei);
|
|
break;
|
|
case SAS_PROTOCOL_SATA:
|
|
case SAS_PROTOCOL_STP:
|
|
case SAS_PROTOCOL_SATA | SAS_PROTOCOL_STP:
|
|
rc = mvs_task_prep_ata(mvi, &tei);
|
|
break;
|
|
default:
|
|
dev_printk(KERN_ERR, &pdev->dev,
|
|
"unknown sas_task proto: 0x%x\n",
|
|
t->task_proto);
|
|
rc = -EINVAL;
|
|
break;
|
|
}
|
|
|
|
if (rc)
|
|
goto err_out_tag;
|
|
|
|
slot->task = t;
|
|
slot->port = tei.port;
|
|
t->lldd_task = (void *) slot;
|
|
list_add_tail(&slot->list, &slot->port->list);
|
|
/* TODO: select normal or high priority */
|
|
|
|
spin_lock(&t->task_state_lock);
|
|
t->task_state_flags |= SAS_TASK_AT_INITIATOR;
|
|
spin_unlock(&t->task_state_lock);
|
|
|
|
mvs_hba_memory_dump(mvi, tag, t->task_proto);
|
|
|
|
++pass;
|
|
mvi->tx_prod = (mvi->tx_prod + 1) & (MVS_CHIP_SLOT_SZ - 1);
|
|
if (n > 1)
|
|
t = list_entry(t->list.next, struct sas_task, list);
|
|
} while (--n);
|
|
|
|
rc = 0;
|
|
goto out_done;
|
|
|
|
err_out_tag:
|
|
mvs_tag_free(mvi, tag);
|
|
err_out:
|
|
dev_printk(KERN_ERR, &pdev->dev, "mvsas exec failed[%d]!\n", rc);
|
|
if (!sas_protocol_ata(t->task_proto))
|
|
if (n_elem)
|
|
pci_unmap_sg(mvi->pdev, t->scatter, n_elem,
|
|
t->data_dir);
|
|
out_done:
|
|
if (pass)
|
|
mw32(TX_PROD_IDX, (mvi->tx_prod - 1) & (MVS_CHIP_SLOT_SZ - 1));
|
|
spin_unlock_irqrestore(&mvi->lock, flags);
|
|
return rc;
|
|
}
|
|
|
|
static int mvs_task_abort(struct sas_task *task)
|
|
{
|
|
int rc;
|
|
unsigned long flags;
|
|
struct mvs_info *mvi = task->dev->port->ha->lldd_ha;
|
|
struct pci_dev *pdev = mvi->pdev;
|
|
int tag;
|
|
|
|
spin_lock_irqsave(&task->task_state_lock, flags);
|
|
if (task->task_state_flags & SAS_TASK_STATE_DONE) {
|
|
rc = TMF_RESP_FUNC_COMPLETE;
|
|
spin_unlock_irqrestore(&task->task_state_lock, flags);
|
|
goto out_done;
|
|
}
|
|
spin_unlock_irqrestore(&task->task_state_lock, flags);
|
|
|
|
switch (task->task_proto) {
|
|
case SAS_PROTOCOL_SMP:
|
|
dev_printk(KERN_DEBUG, &pdev->dev, "SMP Abort! \n");
|
|
break;
|
|
case SAS_PROTOCOL_SSP:
|
|
dev_printk(KERN_DEBUG, &pdev->dev, "SSP Abort! \n");
|
|
break;
|
|
case SAS_PROTOCOL_SATA:
|
|
case SAS_PROTOCOL_STP:
|
|
case SAS_PROTOCOL_SATA | SAS_PROTOCOL_STP:{
|
|
dev_printk(KERN_DEBUG, &pdev->dev, "STP Abort! \n");
|
|
#if _MV_DUMP
|
|
dev_printk(KERN_DEBUG, &pdev->dev, "Dump D2H FIS: \n");
|
|
mvs_hexdump(sizeof(struct host_to_dev_fis),
|
|
(void *)&task->ata_task.fis, 0);
|
|
dev_printk(KERN_DEBUG, &pdev->dev, "Dump ATAPI Cmd : \n");
|
|
mvs_hexdump(16, task->ata_task.atapi_packet, 0);
|
|
#endif
|
|
spin_lock_irqsave(&task->task_state_lock, flags);
|
|
if (task->task_state_flags & SAS_TASK_NEED_DEV_RESET) {
|
|
/* TODO */
|
|
;
|
|
}
|
|
spin_unlock_irqrestore(&task->task_state_lock, flags);
|
|
break;
|
|
}
|
|
default:
|
|
break;
|
|
}
|
|
|
|
if (mvs_find_tag(mvi, task, &tag)) {
|
|
spin_lock_irqsave(&mvi->lock, flags);
|
|
mvs_slot_task_free(mvi, task, &mvi->slot_info[tag], tag);
|
|
spin_unlock_irqrestore(&mvi->lock, flags);
|
|
}
|
|
if (!mvs_task_exec(task, 1, GFP_ATOMIC))
|
|
rc = TMF_RESP_FUNC_COMPLETE;
|
|
else
|
|
rc = TMF_RESP_FUNC_FAILED;
|
|
out_done:
|
|
return rc;
|
|
}
|
|
|
|
static void mvs_free(struct mvs_info *mvi)
|
|
{
|
|
int i;
|
|
|
|
if (!mvi)
|
|
return;
|
|
|
|
for (i = 0; i < MVS_SLOTS; i++) {
|
|
struct mvs_slot_info *slot = &mvi->slot_info[i];
|
|
|
|
if (slot->buf)
|
|
dma_free_coherent(&mvi->pdev->dev, MVS_SLOT_BUF_SZ,
|
|
slot->buf, slot->buf_dma);
|
|
}
|
|
|
|
if (mvi->tx)
|
|
dma_free_coherent(&mvi->pdev->dev,
|
|
sizeof(*mvi->tx) * MVS_CHIP_SLOT_SZ,
|
|
mvi->tx, mvi->tx_dma);
|
|
if (mvi->rx_fis)
|
|
dma_free_coherent(&mvi->pdev->dev, MVS_RX_FISL_SZ,
|
|
mvi->rx_fis, mvi->rx_fis_dma);
|
|
if (mvi->rx)
|
|
dma_free_coherent(&mvi->pdev->dev,
|
|
sizeof(*mvi->rx) * (MVS_RX_RING_SZ + 1),
|
|
mvi->rx, mvi->rx_dma);
|
|
if (mvi->slot)
|
|
dma_free_coherent(&mvi->pdev->dev,
|
|
sizeof(*mvi->slot) * MVS_SLOTS,
|
|
mvi->slot, mvi->slot_dma);
|
|
#ifdef MVS_ENABLE_PERI
|
|
if (mvi->peri_regs)
|
|
iounmap(mvi->peri_regs);
|
|
#endif
|
|
if (mvi->regs)
|
|
iounmap(mvi->regs);
|
|
if (mvi->shost)
|
|
scsi_host_put(mvi->shost);
|
|
kfree(mvi->sas.sas_port);
|
|
kfree(mvi->sas.sas_phy);
|
|
kfree(mvi);
|
|
}
|
|
|
|
/* FIXME: locking? */
|
|
static int mvs_phy_control(struct asd_sas_phy *sas_phy, enum phy_func func,
|
|
void *funcdata)
|
|
{
|
|
struct mvs_info *mvi = sas_phy->ha->lldd_ha;
|
|
int rc = 0, phy_id = sas_phy->id;
|
|
u32 tmp;
|
|
|
|
tmp = mvs_read_phy_ctl(mvi, phy_id);
|
|
|
|
switch (func) {
|
|
case PHY_FUNC_SET_LINK_RATE:{
|
|
struct sas_phy_linkrates *rates = funcdata;
|
|
u32 lrmin = 0, lrmax = 0;
|
|
|
|
lrmin = (rates->minimum_linkrate << 8);
|
|
lrmax = (rates->maximum_linkrate << 12);
|
|
|
|
if (lrmin) {
|
|
tmp &= ~(0xf << 8);
|
|
tmp |= lrmin;
|
|
}
|
|
if (lrmax) {
|
|
tmp &= ~(0xf << 12);
|
|
tmp |= lrmax;
|
|
}
|
|
mvs_write_phy_ctl(mvi, phy_id, tmp);
|
|
break;
|
|
}
|
|
|
|
case PHY_FUNC_HARD_RESET:
|
|
if (tmp & PHY_RST_HARD)
|
|
break;
|
|
mvs_write_phy_ctl(mvi, phy_id, tmp | PHY_RST_HARD);
|
|
break;
|
|
|
|
case PHY_FUNC_LINK_RESET:
|
|
mvs_write_phy_ctl(mvi, phy_id, tmp | PHY_RST);
|
|
break;
|
|
|
|
case PHY_FUNC_DISABLE:
|
|
case PHY_FUNC_RELEASE_SPINUP_HOLD:
|
|
default:
|
|
rc = -EOPNOTSUPP;
|
|
}
|
|
|
|
return rc;
|
|
}
|
|
|
|
static void __devinit mvs_phy_init(struct mvs_info *mvi, int phy_id)
|
|
{
|
|
struct mvs_phy *phy = &mvi->phy[phy_id];
|
|
struct asd_sas_phy *sas_phy = &phy->sas_phy;
|
|
|
|
sas_phy->enabled = (phy_id < mvi->chip->n_phy) ? 1 : 0;
|
|
sas_phy->class = SAS;
|
|
sas_phy->iproto = SAS_PROTOCOL_ALL;
|
|
sas_phy->tproto = 0;
|
|
sas_phy->type = PHY_TYPE_PHYSICAL;
|
|
sas_phy->role = PHY_ROLE_INITIATOR;
|
|
sas_phy->oob_mode = OOB_NOT_CONNECTED;
|
|
sas_phy->linkrate = SAS_LINK_RATE_UNKNOWN;
|
|
|
|
sas_phy->id = phy_id;
|
|
sas_phy->sas_addr = &mvi->sas_addr[0];
|
|
sas_phy->frame_rcvd = &phy->frame_rcvd[0];
|
|
sas_phy->ha = &mvi->sas;
|
|
sas_phy->lldd_phy = phy;
|
|
}
|
|
|
|
static struct mvs_info *__devinit mvs_alloc(struct pci_dev *pdev,
|
|
const struct pci_device_id *ent)
|
|
{
|
|
struct mvs_info *mvi;
|
|
unsigned long res_start, res_len, res_flag;
|
|
struct asd_sas_phy **arr_phy;
|
|
struct asd_sas_port **arr_port;
|
|
const struct mvs_chip_info *chip = &mvs_chips[ent->driver_data];
|
|
int i;
|
|
|
|
/*
|
|
* alloc and init our per-HBA mvs_info struct
|
|
*/
|
|
|
|
mvi = kzalloc(sizeof(*mvi), GFP_KERNEL);
|
|
if (!mvi)
|
|
return NULL;
|
|
|
|
spin_lock_init(&mvi->lock);
|
|
#ifdef MVS_USE_TASKLET
|
|
tasklet_init(&mvi->tasklet, mvs_tasklet, (unsigned long)mvi);
|
|
#endif
|
|
mvi->pdev = pdev;
|
|
mvi->chip = chip;
|
|
|
|
if (pdev->device == 0x6440 && pdev->revision == 0)
|
|
mvi->flags |= MVF_PHY_PWR_FIX;
|
|
|
|
/*
|
|
* alloc and init SCSI, SAS glue
|
|
*/
|
|
|
|
mvi->shost = scsi_host_alloc(&mvs_sht, sizeof(void *));
|
|
if (!mvi->shost)
|
|
goto err_out;
|
|
|
|
arr_phy = kcalloc(MVS_MAX_PHYS, sizeof(void *), GFP_KERNEL);
|
|
arr_port = kcalloc(MVS_MAX_PHYS, sizeof(void *), GFP_KERNEL);
|
|
if (!arr_phy || !arr_port)
|
|
goto err_out;
|
|
|
|
for (i = 0; i < MVS_MAX_PHYS; i++) {
|
|
mvs_phy_init(mvi, i);
|
|
arr_phy[i] = &mvi->phy[i].sas_phy;
|
|
arr_port[i] = &mvi->port[i].sas_port;
|
|
mvi->port[i].taskfileset = MVS_ID_NOT_MAPPED;
|
|
mvi->port[i].wide_port_phymap = 0;
|
|
mvi->port[i].port_attached = 0;
|
|
INIT_LIST_HEAD(&mvi->port[i].list);
|
|
}
|
|
|
|
SHOST_TO_SAS_HA(mvi->shost) = &mvi->sas;
|
|
mvi->shost->transportt = mvs_stt;
|
|
mvi->shost->max_id = 21;
|
|
mvi->shost->max_lun = ~0;
|
|
mvi->shost->max_channel = 0;
|
|
mvi->shost->max_cmd_len = 16;
|
|
|
|
mvi->sas.sas_ha_name = DRV_NAME;
|
|
mvi->sas.dev = &pdev->dev;
|
|
mvi->sas.lldd_module = THIS_MODULE;
|
|
mvi->sas.sas_addr = &mvi->sas_addr[0];
|
|
mvi->sas.sas_phy = arr_phy;
|
|
mvi->sas.sas_port = arr_port;
|
|
mvi->sas.num_phys = chip->n_phy;
|
|
mvi->sas.lldd_max_execute_num = 1;
|
|
mvi->sas.lldd_queue_size = MVS_QUEUE_SIZE;
|
|
mvi->shost->can_queue = MVS_CAN_QUEUE;
|
|
mvi->shost->cmd_per_lun = MVS_SLOTS / mvi->sas.num_phys;
|
|
mvi->sas.lldd_ha = mvi;
|
|
mvi->sas.core.shost = mvi->shost;
|
|
|
|
mvs_tag_init(mvi);
|
|
|
|
/*
|
|
* ioremap main and peripheral registers
|
|
*/
|
|
|
|
#ifdef MVS_ENABLE_PERI
|
|
res_start = pci_resource_start(pdev, 2);
|
|
res_len = pci_resource_len(pdev, 2);
|
|
if (!res_start || !res_len)
|
|
goto err_out;
|
|
|
|
mvi->peri_regs = ioremap_nocache(res_start, res_len);
|
|
if (!mvi->peri_regs)
|
|
goto err_out;
|
|
#endif
|
|
|
|
res_start = pci_resource_start(pdev, 4);
|
|
res_len = pci_resource_len(pdev, 4);
|
|
if (!res_start || !res_len)
|
|
goto err_out;
|
|
|
|
res_flag = pci_resource_flags(pdev, 4);
|
|
if (res_flag & IORESOURCE_CACHEABLE)
|
|
mvi->regs = ioremap(res_start, res_len);
|
|
else
|
|
mvi->regs = ioremap_nocache(res_start, res_len);
|
|
|
|
if (!mvi->regs)
|
|
goto err_out;
|
|
|
|
/*
|
|
* alloc and init our DMA areas
|
|
*/
|
|
|
|
mvi->tx = dma_alloc_coherent(&pdev->dev,
|
|
sizeof(*mvi->tx) * MVS_CHIP_SLOT_SZ,
|
|
&mvi->tx_dma, GFP_KERNEL);
|
|
if (!mvi->tx)
|
|
goto err_out;
|
|
memset(mvi->tx, 0, sizeof(*mvi->tx) * MVS_CHIP_SLOT_SZ);
|
|
|
|
mvi->rx_fis = dma_alloc_coherent(&pdev->dev, MVS_RX_FISL_SZ,
|
|
&mvi->rx_fis_dma, GFP_KERNEL);
|
|
if (!mvi->rx_fis)
|
|
goto err_out;
|
|
memset(mvi->rx_fis, 0, MVS_RX_FISL_SZ);
|
|
|
|
mvi->rx = dma_alloc_coherent(&pdev->dev,
|
|
sizeof(*mvi->rx) * (MVS_RX_RING_SZ + 1),
|
|
&mvi->rx_dma, GFP_KERNEL);
|
|
if (!mvi->rx)
|
|
goto err_out;
|
|
memset(mvi->rx, 0, sizeof(*mvi->rx) * (MVS_RX_RING_SZ + 1));
|
|
|
|
mvi->rx[0] = cpu_to_le32(0xfff);
|
|
mvi->rx_cons = 0xfff;
|
|
|
|
mvi->slot = dma_alloc_coherent(&pdev->dev,
|
|
sizeof(*mvi->slot) * MVS_SLOTS,
|
|
&mvi->slot_dma, GFP_KERNEL);
|
|
if (!mvi->slot)
|
|
goto err_out;
|
|
memset(mvi->slot, 0, sizeof(*mvi->slot) * MVS_SLOTS);
|
|
|
|
for (i = 0; i < MVS_SLOTS; i++) {
|
|
struct mvs_slot_info *slot = &mvi->slot_info[i];
|
|
|
|
slot->buf = dma_alloc_coherent(&pdev->dev, MVS_SLOT_BUF_SZ,
|
|
&slot->buf_dma, GFP_KERNEL);
|
|
if (!slot->buf)
|
|
goto err_out;
|
|
memset(slot->buf, 0, MVS_SLOT_BUF_SZ);
|
|
}
|
|
|
|
/* finally, read NVRAM to get our SAS address */
|
|
if (mvs_nvram_read(mvi, NVR_SAS_ADDR, &mvi->sas_addr, 8))
|
|
goto err_out;
|
|
return mvi;
|
|
|
|
err_out:
|
|
mvs_free(mvi);
|
|
return NULL;
|
|
}
|
|
|
|
static u32 mvs_cr32(void __iomem *regs, u32 addr)
|
|
{
|
|
mw32(CMD_ADDR, addr);
|
|
return mr32(CMD_DATA);
|
|
}
|
|
|
|
static void mvs_cw32(void __iomem *regs, u32 addr, u32 val)
|
|
{
|
|
mw32(CMD_ADDR, addr);
|
|
mw32(CMD_DATA, val);
|
|
}
|
|
|
|
static u32 mvs_read_phy_ctl(struct mvs_info *mvi, u32 port)
|
|
{
|
|
void __iomem *regs = mvi->regs;
|
|
return (port < 4)?mr32(P0_SER_CTLSTAT + port * 4):
|
|
mr32(P4_SER_CTLSTAT + (port - 4) * 4);
|
|
}
|
|
|
|
static void mvs_write_phy_ctl(struct mvs_info *mvi, u32 port, u32 val)
|
|
{
|
|
void __iomem *regs = mvi->regs;
|
|
if (port < 4)
|
|
mw32(P0_SER_CTLSTAT + port * 4, val);
|
|
else
|
|
mw32(P4_SER_CTLSTAT + (port - 4) * 4, val);
|
|
}
|
|
|
|
static u32 mvs_read_port(struct mvs_info *mvi, u32 off, u32 off2, u32 port)
|
|
{
|
|
void __iomem *regs = mvi->regs + off;
|
|
void __iomem *regs2 = mvi->regs + off2;
|
|
return (port < 4)?readl(regs + port * 8):
|
|
readl(regs2 + (port - 4) * 8);
|
|
}
|
|
|
|
static void mvs_write_port(struct mvs_info *mvi, u32 off, u32 off2,
|
|
u32 port, u32 val)
|
|
{
|
|
void __iomem *regs = mvi->regs + off;
|
|
void __iomem *regs2 = mvi->regs + off2;
|
|
if (port < 4)
|
|
writel(val, regs + port * 8);
|
|
else
|
|
writel(val, regs2 + (port - 4) * 8);
|
|
}
|
|
|
|
static u32 mvs_read_port_cfg_data(struct mvs_info *mvi, u32 port)
|
|
{
|
|
return mvs_read_port(mvi, MVS_P0_CFG_DATA, MVS_P4_CFG_DATA, port);
|
|
}
|
|
|
|
static void mvs_write_port_cfg_data(struct mvs_info *mvi, u32 port, u32 val)
|
|
{
|
|
mvs_write_port(mvi, MVS_P0_CFG_DATA, MVS_P4_CFG_DATA, port, val);
|
|
}
|
|
|
|
static void mvs_write_port_cfg_addr(struct mvs_info *mvi, u32 port, u32 addr)
|
|
{
|
|
mvs_write_port(mvi, MVS_P0_CFG_ADDR, MVS_P4_CFG_ADDR, port, addr);
|
|
}
|
|
|
|
static u32 mvs_read_port_vsr_data(struct mvs_info *mvi, u32 port)
|
|
{
|
|
return mvs_read_port(mvi, MVS_P0_VSR_DATA, MVS_P4_VSR_DATA, port);
|
|
}
|
|
|
|
static void mvs_write_port_vsr_data(struct mvs_info *mvi, u32 port, u32 val)
|
|
{
|
|
mvs_write_port(mvi, MVS_P0_VSR_DATA, MVS_P4_VSR_DATA, port, val);
|
|
}
|
|
|
|
static void mvs_write_port_vsr_addr(struct mvs_info *mvi, u32 port, u32 addr)
|
|
{
|
|
mvs_write_port(mvi, MVS_P0_VSR_ADDR, MVS_P4_VSR_ADDR, port, addr);
|
|
}
|
|
|
|
static u32 mvs_read_port_irq_stat(struct mvs_info *mvi, u32 port)
|
|
{
|
|
return mvs_read_port(mvi, MVS_P0_INT_STAT, MVS_P4_INT_STAT, port);
|
|
}
|
|
|
|
static void mvs_write_port_irq_stat(struct mvs_info *mvi, u32 port, u32 val)
|
|
{
|
|
mvs_write_port(mvi, MVS_P0_INT_STAT, MVS_P4_INT_STAT, port, val);
|
|
}
|
|
|
|
static u32 mvs_read_port_irq_mask(struct mvs_info *mvi, u32 port)
|
|
{
|
|
return mvs_read_port(mvi, MVS_P0_INT_MASK, MVS_P4_INT_MASK, port);
|
|
}
|
|
|
|
static void mvs_write_port_irq_mask(struct mvs_info *mvi, u32 port, u32 val)
|
|
{
|
|
mvs_write_port(mvi, MVS_P0_INT_MASK, MVS_P4_INT_MASK, port, val);
|
|
}
|
|
|
|
static void __devinit mvs_phy_hacks(struct mvs_info *mvi)
|
|
{
|
|
void __iomem *regs = mvi->regs;
|
|
u32 tmp;
|
|
|
|
/* workaround for SATA R-ERR, to ignore phy glitch */
|
|
tmp = mvs_cr32(regs, CMD_PHY_TIMER);
|
|
tmp &= ~(1 << 9);
|
|
tmp |= (1 << 10);
|
|
mvs_cw32(regs, CMD_PHY_TIMER, tmp);
|
|
|
|
/* enable retry 127 times */
|
|
mvs_cw32(regs, CMD_SAS_CTL1, 0x7f7f);
|
|
|
|
/* extend open frame timeout to max */
|
|
tmp = mvs_cr32(regs, CMD_SAS_CTL0);
|
|
tmp &= ~0xffff;
|
|
tmp |= 0x3fff;
|
|
mvs_cw32(regs, CMD_SAS_CTL0, tmp);
|
|
|
|
/* workaround for WDTIMEOUT , set to 550 ms */
|
|
mvs_cw32(regs, CMD_WD_TIMER, 0x86470);
|
|
|
|
/* not to halt for different port op during wideport link change */
|
|
mvs_cw32(regs, CMD_APP_ERR_CONFIG, 0xffefbf7d);
|
|
|
|
/* workaround for Seagate disk not-found OOB sequence, recv
|
|
* COMINIT before sending out COMWAKE */
|
|
tmp = mvs_cr32(regs, CMD_PHY_MODE_21);
|
|
tmp &= 0x0000ffff;
|
|
tmp |= 0x00fa0000;
|
|
mvs_cw32(regs, CMD_PHY_MODE_21, tmp);
|
|
|
|
tmp = mvs_cr32(regs, CMD_PHY_TIMER);
|
|
tmp &= 0x1fffffff;
|
|
tmp |= (2U << 29); /* 8 ms retry */
|
|
mvs_cw32(regs, CMD_PHY_TIMER, tmp);
|
|
|
|
/* TEST - for phy decoding error, adjust voltage levels */
|
|
mw32(P0_VSR_ADDR + 0, 0x8);
|
|
mw32(P0_VSR_DATA + 0, 0x2F0);
|
|
|
|
mw32(P0_VSR_ADDR + 8, 0x8);
|
|
mw32(P0_VSR_DATA + 8, 0x2F0);
|
|
|
|
mw32(P0_VSR_ADDR + 16, 0x8);
|
|
mw32(P0_VSR_DATA + 16, 0x2F0);
|
|
|
|
mw32(P0_VSR_ADDR + 24, 0x8);
|
|
mw32(P0_VSR_DATA + 24, 0x2F0);
|
|
|
|
}
|
|
|
|
static void mvs_enable_xmt(struct mvs_info *mvi, int PhyId)
|
|
{
|
|
void __iomem *regs = mvi->regs;
|
|
u32 tmp;
|
|
|
|
tmp = mr32(PCS);
|
|
if (mvi->chip->n_phy <= 4)
|
|
tmp |= 1 << (PhyId + PCS_EN_PORT_XMT_SHIFT);
|
|
else
|
|
tmp |= 1 << (PhyId + PCS_EN_PORT_XMT_SHIFT2);
|
|
mw32(PCS, tmp);
|
|
}
|
|
|
|
static void mvs_detect_porttype(struct mvs_info *mvi, int i)
|
|
{
|
|
void __iomem *regs = mvi->regs;
|
|
u32 reg;
|
|
struct mvs_phy *phy = &mvi->phy[i];
|
|
|
|
/* TODO check & save device type */
|
|
reg = mr32(GBL_PORT_TYPE);
|
|
|
|
if (reg & MODE_SAS_SATA & (1 << i))
|
|
phy->phy_type |= PORT_TYPE_SAS;
|
|
else
|
|
phy->phy_type |= PORT_TYPE_SATA;
|
|
}
|
|
|
|
static void *mvs_get_d2h_reg(struct mvs_info *mvi, int i, void *buf)
|
|
{
|
|
u32 *s = (u32 *) buf;
|
|
|
|
if (!s)
|
|
return NULL;
|
|
|
|
mvs_write_port_cfg_addr(mvi, i, PHYR_SATA_SIG3);
|
|
s[3] = mvs_read_port_cfg_data(mvi, i);
|
|
|
|
mvs_write_port_cfg_addr(mvi, i, PHYR_SATA_SIG2);
|
|
s[2] = mvs_read_port_cfg_data(mvi, i);
|
|
|
|
mvs_write_port_cfg_addr(mvi, i, PHYR_SATA_SIG1);
|
|
s[1] = mvs_read_port_cfg_data(mvi, i);
|
|
|
|
mvs_write_port_cfg_addr(mvi, i, PHYR_SATA_SIG0);
|
|
s[0] = mvs_read_port_cfg_data(mvi, i);
|
|
|
|
return (void *)s;
|
|
}
|
|
|
|
static u32 mvs_is_sig_fis_received(u32 irq_status)
|
|
{
|
|
return irq_status & PHYEV_SIG_FIS;
|
|
}
|
|
|
|
static void mvs_update_wideport(struct mvs_info *mvi, int i)
|
|
{
|
|
struct mvs_phy *phy = &mvi->phy[i];
|
|
struct mvs_port *port = phy->port;
|
|
int j, no;
|
|
|
|
for_each_phy(port->wide_port_phymap, no, j, mvi->chip->n_phy)
|
|
if (no & 1) {
|
|
mvs_write_port_cfg_addr(mvi, no, PHYR_WIDE_PORT);
|
|
mvs_write_port_cfg_data(mvi, no,
|
|
port->wide_port_phymap);
|
|
} else {
|
|
mvs_write_port_cfg_addr(mvi, no, PHYR_WIDE_PORT);
|
|
mvs_write_port_cfg_data(mvi, no, 0);
|
|
}
|
|
}
|
|
|
|
static u32 mvs_is_phy_ready(struct mvs_info *mvi, int i)
|
|
{
|
|
u32 tmp;
|
|
struct mvs_phy *phy = &mvi->phy[i];
|
|
struct mvs_port *port = phy->port;;
|
|
|
|
tmp = mvs_read_phy_ctl(mvi, i);
|
|
|
|
if ((tmp & PHY_READY_MASK) && !(phy->irq_status & PHYEV_POOF)) {
|
|
if (!port)
|
|
phy->phy_attached = 1;
|
|
return tmp;
|
|
}
|
|
|
|
if (port) {
|
|
if (phy->phy_type & PORT_TYPE_SAS) {
|
|
port->wide_port_phymap &= ~(1U << i);
|
|
if (!port->wide_port_phymap)
|
|
port->port_attached = 0;
|
|
mvs_update_wideport(mvi, i);
|
|
} else if (phy->phy_type & PORT_TYPE_SATA)
|
|
port->port_attached = 0;
|
|
mvs_free_reg_set(mvi, phy->port);
|
|
phy->port = NULL;
|
|
phy->phy_attached = 0;
|
|
phy->phy_type &= ~(PORT_TYPE_SAS | PORT_TYPE_SATA);
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
static void mvs_update_phyinfo(struct mvs_info *mvi, int i,
|
|
int get_st)
|
|
{
|
|
struct mvs_phy *phy = &mvi->phy[i];
|
|
struct pci_dev *pdev = mvi->pdev;
|
|
u32 tmp;
|
|
u64 tmp64;
|
|
|
|
mvs_write_port_cfg_addr(mvi, i, PHYR_IDENTIFY);
|
|
phy->dev_info = mvs_read_port_cfg_data(mvi, i);
|
|
|
|
mvs_write_port_cfg_addr(mvi, i, PHYR_ADDR_HI);
|
|
phy->dev_sas_addr = (u64) mvs_read_port_cfg_data(mvi, i) << 32;
|
|
|
|
mvs_write_port_cfg_addr(mvi, i, PHYR_ADDR_LO);
|
|
phy->dev_sas_addr |= mvs_read_port_cfg_data(mvi, i);
|
|
|
|
if (get_st) {
|
|
phy->irq_status = mvs_read_port_irq_stat(mvi, i);
|
|
phy->phy_status = mvs_is_phy_ready(mvi, i);
|
|
}
|
|
|
|
if (phy->phy_status) {
|
|
u32 phy_st;
|
|
struct asd_sas_phy *sas_phy = mvi->sas.sas_phy[i];
|
|
|
|
mvs_write_port_cfg_addr(mvi, i, PHYR_PHY_STAT);
|
|
phy_st = mvs_read_port_cfg_data(mvi, i);
|
|
|
|
sas_phy->linkrate =
|
|
(phy->phy_status & PHY_NEG_SPP_PHYS_LINK_RATE_MASK) >>
|
|
PHY_NEG_SPP_PHYS_LINK_RATE_MASK_OFFSET;
|
|
phy->minimum_linkrate =
|
|
(phy->phy_status &
|
|
PHY_MIN_SPP_PHYS_LINK_RATE_MASK) >> 8;
|
|
phy->maximum_linkrate =
|
|
(phy->phy_status &
|
|
PHY_MAX_SPP_PHYS_LINK_RATE_MASK) >> 12;
|
|
|
|
if (phy->phy_type & PORT_TYPE_SAS) {
|
|
/* Updated attached_sas_addr */
|
|
mvs_write_port_cfg_addr(mvi, i, PHYR_ATT_ADDR_HI);
|
|
phy->att_dev_sas_addr =
|
|
(u64) mvs_read_port_cfg_data(mvi, i) << 32;
|
|
mvs_write_port_cfg_addr(mvi, i, PHYR_ATT_ADDR_LO);
|
|
phy->att_dev_sas_addr |= mvs_read_port_cfg_data(mvi, i);
|
|
mvs_write_port_cfg_addr(mvi, i, PHYR_ATT_DEV_INFO);
|
|
phy->att_dev_info = mvs_read_port_cfg_data(mvi, i);
|
|
phy->identify.device_type =
|
|
phy->att_dev_info & PORT_DEV_TYPE_MASK;
|
|
|
|
if (phy->identify.device_type == SAS_END_DEV)
|
|
phy->identify.target_port_protocols =
|
|
SAS_PROTOCOL_SSP;
|
|
else if (phy->identify.device_type != NO_DEVICE)
|
|
phy->identify.target_port_protocols =
|
|
SAS_PROTOCOL_SMP;
|
|
if (phy_st & PHY_OOB_DTCTD)
|
|
sas_phy->oob_mode = SAS_OOB_MODE;
|
|
phy->frame_rcvd_size =
|
|
sizeof(struct sas_identify_frame);
|
|
} else if (phy->phy_type & PORT_TYPE_SATA) {
|
|
phy->identify.target_port_protocols = SAS_PROTOCOL_STP;
|
|
if (mvs_is_sig_fis_received(phy->irq_status)) {
|
|
phy->att_dev_sas_addr = i; /* temp */
|
|
if (phy_st & PHY_OOB_DTCTD)
|
|
sas_phy->oob_mode = SATA_OOB_MODE;
|
|
phy->frame_rcvd_size =
|
|
sizeof(struct dev_to_host_fis);
|
|
mvs_get_d2h_reg(mvi, i,
|
|
(void *)sas_phy->frame_rcvd);
|
|
} else {
|
|
dev_printk(KERN_DEBUG, &pdev->dev,
|
|
"No sig fis\n");
|
|
phy->phy_type &= ~(PORT_TYPE_SATA);
|
|
goto out_done;
|
|
}
|
|
}
|
|
tmp64 = cpu_to_be64(phy->att_dev_sas_addr);
|
|
memcpy(sas_phy->attached_sas_addr, &tmp64, SAS_ADDR_SIZE);
|
|
|
|
dev_printk(KERN_DEBUG, &pdev->dev,
|
|
"phy[%d] Get Attached Address 0x%llX ,"
|
|
" SAS Address 0x%llX\n",
|
|
i,
|
|
(unsigned long long)phy->att_dev_sas_addr,
|
|
(unsigned long long)phy->dev_sas_addr);
|
|
dev_printk(KERN_DEBUG, &pdev->dev,
|
|
"Rate = %x , type = %d\n",
|
|
sas_phy->linkrate, phy->phy_type);
|
|
|
|
/* workaround for HW phy decoding error on 1.5g disk drive */
|
|
mvs_write_port_vsr_addr(mvi, i, VSR_PHY_MODE6);
|
|
tmp = mvs_read_port_vsr_data(mvi, i);
|
|
if (((phy->phy_status & PHY_NEG_SPP_PHYS_LINK_RATE_MASK) >>
|
|
PHY_NEG_SPP_PHYS_LINK_RATE_MASK_OFFSET) ==
|
|
SAS_LINK_RATE_1_5_GBPS)
|
|
tmp &= ~PHY_MODE6_LATECLK;
|
|
else
|
|
tmp |= PHY_MODE6_LATECLK;
|
|
mvs_write_port_vsr_data(mvi, i, tmp);
|
|
|
|
}
|
|
out_done:
|
|
if (get_st)
|
|
mvs_write_port_irq_stat(mvi, i, phy->irq_status);
|
|
}
|
|
|
|
static void mvs_port_formed(struct asd_sas_phy *sas_phy)
|
|
{
|
|
struct sas_ha_struct *sas_ha = sas_phy->ha;
|
|
struct mvs_info *mvi = sas_ha->lldd_ha;
|
|
struct asd_sas_port *sas_port = sas_phy->port;
|
|
struct mvs_phy *phy = sas_phy->lldd_phy;
|
|
struct mvs_port *port = &mvi->port[sas_port->id];
|
|
unsigned long flags;
|
|
|
|
spin_lock_irqsave(&mvi->lock, flags);
|
|
port->port_attached = 1;
|
|
phy->port = port;
|
|
port->taskfileset = MVS_ID_NOT_MAPPED;
|
|
if (phy->phy_type & PORT_TYPE_SAS) {
|
|
port->wide_port_phymap = sas_port->phy_mask;
|
|
mvs_update_wideport(mvi, sas_phy->id);
|
|
}
|
|
spin_unlock_irqrestore(&mvi->lock, flags);
|
|
}
|
|
|
|
static int mvs_I_T_nexus_reset(struct domain_device *dev)
|
|
{
|
|
return TMF_RESP_FUNC_FAILED;
|
|
}
|
|
|
|
static int __devinit mvs_hw_init(struct mvs_info *mvi)
|
|
{
|
|
void __iomem *regs = mvi->regs;
|
|
int i;
|
|
u32 tmp, cctl;
|
|
|
|
/* make sure interrupts are masked immediately (paranoia) */
|
|
mw32(GBL_CTL, 0);
|
|
tmp = mr32(GBL_CTL);
|
|
|
|
/* Reset Controller */
|
|
if (!(tmp & HBA_RST)) {
|
|
if (mvi->flags & MVF_PHY_PWR_FIX) {
|
|
pci_read_config_dword(mvi->pdev, PCR_PHY_CTL, &tmp);
|
|
tmp &= ~PCTL_PWR_ON;
|
|
tmp |= PCTL_OFF;
|
|
pci_write_config_dword(mvi->pdev, PCR_PHY_CTL, tmp);
|
|
|
|
pci_read_config_dword(mvi->pdev, PCR_PHY_CTL2, &tmp);
|
|
tmp &= ~PCTL_PWR_ON;
|
|
tmp |= PCTL_OFF;
|
|
pci_write_config_dword(mvi->pdev, PCR_PHY_CTL2, tmp);
|
|
}
|
|
|
|
/* global reset, incl. COMRESET/H_RESET_N (self-clearing) */
|
|
mw32_f(GBL_CTL, HBA_RST);
|
|
}
|
|
|
|
/* wait for reset to finish; timeout is just a guess */
|
|
i = 1000;
|
|
while (i-- > 0) {
|
|
msleep(10);
|
|
|
|
if (!(mr32(GBL_CTL) & HBA_RST))
|
|
break;
|
|
}
|
|
if (mr32(GBL_CTL) & HBA_RST) {
|
|
dev_printk(KERN_ERR, &mvi->pdev->dev, "HBA reset failed\n");
|
|
return -EBUSY;
|
|
}
|
|
|
|
/* Init Chip */
|
|
/* make sure RST is set; HBA_RST /should/ have done that for us */
|
|
cctl = mr32(CTL);
|
|
if (cctl & CCTL_RST)
|
|
cctl &= ~CCTL_RST;
|
|
else
|
|
mw32_f(CTL, cctl | CCTL_RST);
|
|
|
|
/* write to device control _AND_ device status register? - A.C. */
|
|
pci_read_config_dword(mvi->pdev, PCR_DEV_CTRL, &tmp);
|
|
tmp &= ~PRD_REQ_MASK;
|
|
tmp |= PRD_REQ_SIZE;
|
|
pci_write_config_dword(mvi->pdev, PCR_DEV_CTRL, tmp);
|
|
|
|
pci_read_config_dword(mvi->pdev, PCR_PHY_CTL, &tmp);
|
|
tmp |= PCTL_PWR_ON;
|
|
tmp &= ~PCTL_OFF;
|
|
pci_write_config_dword(mvi->pdev, PCR_PHY_CTL, tmp);
|
|
|
|
pci_read_config_dword(mvi->pdev, PCR_PHY_CTL2, &tmp);
|
|
tmp |= PCTL_PWR_ON;
|
|
tmp &= ~PCTL_OFF;
|
|
pci_write_config_dword(mvi->pdev, PCR_PHY_CTL2, tmp);
|
|
|
|
mw32_f(CTL, cctl);
|
|
|
|
/* reset control */
|
|
mw32(PCS, 0); /*MVS_PCS */
|
|
|
|
mvs_phy_hacks(mvi);
|
|
|
|
mw32(CMD_LIST_LO, mvi->slot_dma);
|
|
mw32(CMD_LIST_HI, (mvi->slot_dma >> 16) >> 16);
|
|
|
|
mw32(RX_FIS_LO, mvi->rx_fis_dma);
|
|
mw32(RX_FIS_HI, (mvi->rx_fis_dma >> 16) >> 16);
|
|
|
|
mw32(TX_CFG, MVS_CHIP_SLOT_SZ);
|
|
mw32(TX_LO, mvi->tx_dma);
|
|
mw32(TX_HI, (mvi->tx_dma >> 16) >> 16);
|
|
|
|
mw32(RX_CFG, MVS_RX_RING_SZ);
|
|
mw32(RX_LO, mvi->rx_dma);
|
|
mw32(RX_HI, (mvi->rx_dma >> 16) >> 16);
|
|
|
|
/* enable auto port detection */
|
|
mw32(GBL_PORT_TYPE, MODE_AUTO_DET_EN);
|
|
msleep(1100);
|
|
/* init and reset phys */
|
|
for (i = 0; i < mvi->chip->n_phy; i++) {
|
|
u32 lo = be32_to_cpu(*(u32 *)&mvi->sas_addr[4]);
|
|
u32 hi = be32_to_cpu(*(u32 *)&mvi->sas_addr[0]);
|
|
|
|
mvs_detect_porttype(mvi, i);
|
|
|
|
/* set phy local SAS address */
|
|
mvs_write_port_cfg_addr(mvi, i, PHYR_ADDR_LO);
|
|
mvs_write_port_cfg_data(mvi, i, lo);
|
|
mvs_write_port_cfg_addr(mvi, i, PHYR_ADDR_HI);
|
|
mvs_write_port_cfg_data(mvi, i, hi);
|
|
|
|
/* reset phy */
|
|
tmp = mvs_read_phy_ctl(mvi, i);
|
|
tmp |= PHY_RST;
|
|
mvs_write_phy_ctl(mvi, i, tmp);
|
|
}
|
|
|
|
msleep(100);
|
|
|
|
for (i = 0; i < mvi->chip->n_phy; i++) {
|
|
/* clear phy int status */
|
|
tmp = mvs_read_port_irq_stat(mvi, i);
|
|
tmp &= ~PHYEV_SIG_FIS;
|
|
mvs_write_port_irq_stat(mvi, i, tmp);
|
|
|
|
/* set phy int mask */
|
|
tmp = PHYEV_RDY_CH | PHYEV_BROAD_CH | PHYEV_UNASSOC_FIS |
|
|
PHYEV_ID_DONE | PHYEV_DEC_ERR;
|
|
mvs_write_port_irq_mask(mvi, i, tmp);
|
|
|
|
msleep(100);
|
|
mvs_update_phyinfo(mvi, i, 1);
|
|
mvs_enable_xmt(mvi, i);
|
|
}
|
|
|
|
/* FIXME: update wide port bitmaps */
|
|
|
|
/* little endian for open address and command table, etc. */
|
|
/* A.C.
|
|
* it seems that ( from the spec ) turning on big-endian won't
|
|
* do us any good on big-endian machines, need further confirmation
|
|
*/
|
|
cctl = mr32(CTL);
|
|
cctl |= CCTL_ENDIAN_CMD;
|
|
cctl |= CCTL_ENDIAN_DATA;
|
|
cctl &= ~CCTL_ENDIAN_OPEN;
|
|
cctl |= CCTL_ENDIAN_RSP;
|
|
mw32_f(CTL, cctl);
|
|
|
|
/* reset CMD queue */
|
|
tmp = mr32(PCS);
|
|
tmp |= PCS_CMD_RST;
|
|
mw32(PCS, tmp);
|
|
/* interrupt coalescing may cause missing HW interrput in some case,
|
|
* and the max count is 0x1ff, while our max slot is 0x200,
|
|
* it will make count 0.
|
|
*/
|
|
tmp = 0;
|
|
mw32(INT_COAL, tmp);
|
|
|
|
tmp = 0x100;
|
|
mw32(INT_COAL_TMOUT, tmp);
|
|
|
|
/* ladies and gentlemen, start your engines */
|
|
mw32(TX_CFG, 0);
|
|
mw32(TX_CFG, MVS_CHIP_SLOT_SZ | TX_EN);
|
|
mw32(RX_CFG, MVS_RX_RING_SZ | RX_EN);
|
|
/* enable CMD/CMPL_Q/RESP mode */
|
|
mw32(PCS, PCS_SATA_RETRY | PCS_FIS_RX_EN | PCS_CMD_EN);
|
|
|
|
/* enable completion queue interrupt */
|
|
tmp = (CINT_PORT_MASK | CINT_DONE | CINT_MEM | CINT_SRS);
|
|
mw32(INT_MASK, tmp);
|
|
|
|
/* Enable SRS interrupt */
|
|
mw32(INT_MASK_SRS, 0xFF);
|
|
return 0;
|
|
}
|
|
|
|
static void __devinit mvs_print_info(struct mvs_info *mvi)
|
|
{
|
|
struct pci_dev *pdev = mvi->pdev;
|
|
static int printed_version;
|
|
|
|
if (!printed_version++)
|
|
dev_printk(KERN_INFO, &pdev->dev, "version " DRV_VERSION "\n");
|
|
|
|
dev_printk(KERN_INFO, &pdev->dev, "%u phys, addr %llx\n",
|
|
mvi->chip->n_phy, SAS_ADDR(mvi->sas_addr));
|
|
}
|
|
|
|
static int __devinit mvs_pci_init(struct pci_dev *pdev,
|
|
const struct pci_device_id *ent)
|
|
{
|
|
int rc;
|
|
struct mvs_info *mvi;
|
|
irq_handler_t irq_handler = mvs_interrupt;
|
|
|
|
rc = pci_enable_device(pdev);
|
|
if (rc)
|
|
return rc;
|
|
|
|
pci_set_master(pdev);
|
|
|
|
rc = pci_request_regions(pdev, DRV_NAME);
|
|
if (rc)
|
|
goto err_out_disable;
|
|
|
|
rc = pci_go_64(pdev);
|
|
if (rc)
|
|
goto err_out_regions;
|
|
|
|
mvi = mvs_alloc(pdev, ent);
|
|
if (!mvi) {
|
|
rc = -ENOMEM;
|
|
goto err_out_regions;
|
|
}
|
|
|
|
rc = mvs_hw_init(mvi);
|
|
if (rc)
|
|
goto err_out_mvi;
|
|
|
|
#ifndef MVS_DISABLE_MSI
|
|
if (!pci_enable_msi(pdev)) {
|
|
u32 tmp;
|
|
void __iomem *regs = mvi->regs;
|
|
mvi->flags |= MVF_MSI;
|
|
irq_handler = mvs_msi_interrupt;
|
|
tmp = mr32(PCS);
|
|
mw32(PCS, tmp | PCS_SELF_CLEAR);
|
|
}
|
|
#endif
|
|
|
|
rc = request_irq(pdev->irq, irq_handler, IRQF_SHARED, DRV_NAME, mvi);
|
|
if (rc)
|
|
goto err_out_msi;
|
|
|
|
rc = scsi_add_host(mvi->shost, &pdev->dev);
|
|
if (rc)
|
|
goto err_out_irq;
|
|
|
|
rc = sas_register_ha(&mvi->sas);
|
|
if (rc)
|
|
goto err_out_shost;
|
|
|
|
pci_set_drvdata(pdev, mvi);
|
|
|
|
mvs_print_info(mvi);
|
|
|
|
mvs_hba_interrupt_enable(mvi);
|
|
|
|
scsi_scan_host(mvi->shost);
|
|
|
|
return 0;
|
|
|
|
err_out_shost:
|
|
scsi_remove_host(mvi->shost);
|
|
err_out_irq:
|
|
free_irq(pdev->irq, mvi);
|
|
err_out_msi:
|
|
if (mvi->flags |= MVF_MSI)
|
|
pci_disable_msi(pdev);
|
|
err_out_mvi:
|
|
mvs_free(mvi);
|
|
err_out_regions:
|
|
pci_release_regions(pdev);
|
|
err_out_disable:
|
|
pci_disable_device(pdev);
|
|
return rc;
|
|
}
|
|
|
|
static void __devexit mvs_pci_remove(struct pci_dev *pdev)
|
|
{
|
|
struct mvs_info *mvi = pci_get_drvdata(pdev);
|
|
|
|
pci_set_drvdata(pdev, NULL);
|
|
|
|
if (mvi) {
|
|
sas_unregister_ha(&mvi->sas);
|
|
mvs_hba_interrupt_disable(mvi);
|
|
sas_remove_host(mvi->shost);
|
|
scsi_remove_host(mvi->shost);
|
|
|
|
free_irq(pdev->irq, mvi);
|
|
if (mvi->flags & MVF_MSI)
|
|
pci_disable_msi(pdev);
|
|
mvs_free(mvi);
|
|
pci_release_regions(pdev);
|
|
}
|
|
pci_disable_device(pdev);
|
|
}
|
|
|
|
static struct sas_domain_function_template mvs_transport_ops = {
|
|
.lldd_execute_task = mvs_task_exec,
|
|
.lldd_control_phy = mvs_phy_control,
|
|
.lldd_abort_task = mvs_task_abort,
|
|
.lldd_port_formed = mvs_port_formed,
|
|
.lldd_I_T_nexus_reset = mvs_I_T_nexus_reset,
|
|
};
|
|
|
|
static struct pci_device_id __devinitdata mvs_pci_table[] = {
|
|
{ PCI_VDEVICE(MARVELL, 0x6320), chip_6320 },
|
|
{ PCI_VDEVICE(MARVELL, 0x6340), chip_6440 },
|
|
{
|
|
.vendor = PCI_VENDOR_ID_MARVELL,
|
|
.device = 0x6440,
|
|
.subvendor = PCI_ANY_ID,
|
|
.subdevice = 0x6480,
|
|
.class = 0,
|
|
.class_mask = 0,
|
|
.driver_data = chip_6480,
|
|
},
|
|
{ PCI_VDEVICE(MARVELL, 0x6440), chip_6440 },
|
|
{ PCI_VDEVICE(MARVELL, 0x6480), chip_6480 },
|
|
|
|
{ } /* terminate list */
|
|
};
|
|
|
|
static struct pci_driver mvs_pci_driver = {
|
|
.name = DRV_NAME,
|
|
.id_table = mvs_pci_table,
|
|
.probe = mvs_pci_init,
|
|
.remove = __devexit_p(mvs_pci_remove),
|
|
};
|
|
|
|
static int __init mvs_init(void)
|
|
{
|
|
int rc;
|
|
|
|
mvs_stt = sas_domain_attach_transport(&mvs_transport_ops);
|
|
if (!mvs_stt)
|
|
return -ENOMEM;
|
|
|
|
rc = pci_register_driver(&mvs_pci_driver);
|
|
if (rc)
|
|
goto err_out;
|
|
|
|
return 0;
|
|
|
|
err_out:
|
|
sas_release_transport(mvs_stt);
|
|
return rc;
|
|
}
|
|
|
|
static void __exit mvs_exit(void)
|
|
{
|
|
pci_unregister_driver(&mvs_pci_driver);
|
|
sas_release_transport(mvs_stt);
|
|
}
|
|
|
|
module_init(mvs_init);
|
|
module_exit(mvs_exit);
|
|
|
|
MODULE_AUTHOR("Jeff Garzik <jgarzik@pobox.com>");
|
|
MODULE_DESCRIPTION("Marvell 88SE6440 SAS/SATA controller driver");
|
|
MODULE_VERSION(DRV_VERSION);
|
|
MODULE_LICENSE("GPL");
|
|
MODULE_DEVICE_TABLE(pci, mvs_pci_table);
|