mirror of
https://github.com/AuxXxilium/linux_dsm_epyc7002.git
synced 2024-11-25 01:00:58 +07:00
96f339c6b9
CONFIG_SGI_MBCS is enabled in generic kernels, but the driver may oops some other platforms. Check whether we are running on sn2 and bail out if we are not before doing anything dangerous. Acked-by: Bruce Losure <blosure@americas.sgi.com> Signed-off-by: Greg Edwards <edwardsg@sgi.com> Signed-off-by: Tony Luck <tony.luck@intel.com>
853 lines
20 KiB
C
853 lines
20 KiB
C
/*
|
|
* This file is subject to the terms and conditions of the GNU General Public
|
|
* License. See the file "COPYING" in the main directory of this archive
|
|
* for more details.
|
|
*
|
|
* Copyright (c) 2005 Silicon Graphics, Inc. All rights reserved.
|
|
*/
|
|
|
|
/*
|
|
* MOATB Core Services driver.
|
|
*/
|
|
|
|
#include <linux/config.h>
|
|
#include <linux/interrupt.h>
|
|
#include <linux/module.h>
|
|
#include <linux/moduleparam.h>
|
|
#include <linux/types.h>
|
|
#include <linux/ioport.h>
|
|
#include <linux/notifier.h>
|
|
#include <linux/reboot.h>
|
|
#include <linux/init.h>
|
|
#include <linux/fs.h>
|
|
#include <linux/delay.h>
|
|
#include <linux/device.h>
|
|
#include <linux/mm.h>
|
|
#include <linux/uio.h>
|
|
#include <asm/io.h>
|
|
#include <asm/uaccess.h>
|
|
#include <asm/system.h>
|
|
#include <asm/pgtable.h>
|
|
#include <asm/sn/addrs.h>
|
|
#include <asm/sn/intr.h>
|
|
#include <asm/sn/tiocx.h>
|
|
#include "mbcs.h"
|
|
|
|
#define MBCS_DEBUG 0
|
|
#if MBCS_DEBUG
|
|
#define DBG(fmt...) printk(KERN_ALERT fmt)
|
|
#else
|
|
#define DBG(fmt...)
|
|
#endif
|
|
int mbcs_major;
|
|
|
|
LIST_HEAD(soft_list);
|
|
|
|
/*
|
|
* file operations
|
|
*/
|
|
struct file_operations mbcs_ops = {
|
|
.open = mbcs_open,
|
|
.llseek = mbcs_sram_llseek,
|
|
.read = mbcs_sram_read,
|
|
.write = mbcs_sram_write,
|
|
.mmap = mbcs_gscr_mmap,
|
|
};
|
|
|
|
struct mbcs_callback_arg {
|
|
int minor;
|
|
struct cx_dev *cx_dev;
|
|
};
|
|
|
|
static inline void mbcs_getdma_init(struct getdma *gdma)
|
|
{
|
|
memset(gdma, 0, sizeof(struct getdma));
|
|
gdma->DoneIntEnable = 1;
|
|
}
|
|
|
|
static inline void mbcs_putdma_init(struct putdma *pdma)
|
|
{
|
|
memset(pdma, 0, sizeof(struct putdma));
|
|
pdma->DoneIntEnable = 1;
|
|
}
|
|
|
|
static inline void mbcs_algo_init(struct algoblock *algo_soft)
|
|
{
|
|
memset(algo_soft, 0, sizeof(struct algoblock));
|
|
}
|
|
|
|
static inline void mbcs_getdma_set(void *mmr,
|
|
uint64_t hostAddr,
|
|
uint64_t localAddr,
|
|
uint64_t localRamSel,
|
|
uint64_t numPkts,
|
|
uint64_t amoEnable,
|
|
uint64_t intrEnable,
|
|
uint64_t peerIO,
|
|
uint64_t amoHostDest,
|
|
uint64_t amoModType, uint64_t intrHostDest,
|
|
uint64_t intrVector)
|
|
{
|
|
union dma_control rdma_control;
|
|
union dma_amo_dest amo_dest;
|
|
union intr_dest intr_dest;
|
|
union dma_localaddr local_addr;
|
|
union dma_hostaddr host_addr;
|
|
|
|
rdma_control.dma_control_reg = 0;
|
|
amo_dest.dma_amo_dest_reg = 0;
|
|
intr_dest.intr_dest_reg = 0;
|
|
local_addr.dma_localaddr_reg = 0;
|
|
host_addr.dma_hostaddr_reg = 0;
|
|
|
|
host_addr.dma_sys_addr = hostAddr;
|
|
MBCS_MMR_SET(mmr, MBCS_RD_DMA_SYS_ADDR, host_addr.dma_hostaddr_reg);
|
|
|
|
local_addr.dma_ram_addr = localAddr;
|
|
local_addr.dma_ram_sel = localRamSel;
|
|
MBCS_MMR_SET(mmr, MBCS_RD_DMA_LOC_ADDR, local_addr.dma_localaddr_reg);
|
|
|
|
rdma_control.dma_op_length = numPkts;
|
|
rdma_control.done_amo_en = amoEnable;
|
|
rdma_control.done_int_en = intrEnable;
|
|
rdma_control.pio_mem_n = peerIO;
|
|
MBCS_MMR_SET(mmr, MBCS_RD_DMA_CTRL, rdma_control.dma_control_reg);
|
|
|
|
amo_dest.dma_amo_sys_addr = amoHostDest;
|
|
amo_dest.dma_amo_mod_type = amoModType;
|
|
MBCS_MMR_SET(mmr, MBCS_RD_DMA_AMO_DEST, amo_dest.dma_amo_dest_reg);
|
|
|
|
intr_dest.address = intrHostDest;
|
|
intr_dest.int_vector = intrVector;
|
|
MBCS_MMR_SET(mmr, MBCS_RD_DMA_INT_DEST, intr_dest.intr_dest_reg);
|
|
|
|
}
|
|
|
|
static inline void mbcs_putdma_set(void *mmr,
|
|
uint64_t hostAddr,
|
|
uint64_t localAddr,
|
|
uint64_t localRamSel,
|
|
uint64_t numPkts,
|
|
uint64_t amoEnable,
|
|
uint64_t intrEnable,
|
|
uint64_t peerIO,
|
|
uint64_t amoHostDest,
|
|
uint64_t amoModType,
|
|
uint64_t intrHostDest, uint64_t intrVector)
|
|
{
|
|
union dma_control wdma_control;
|
|
union dma_amo_dest amo_dest;
|
|
union intr_dest intr_dest;
|
|
union dma_localaddr local_addr;
|
|
union dma_hostaddr host_addr;
|
|
|
|
wdma_control.dma_control_reg = 0;
|
|
amo_dest.dma_amo_dest_reg = 0;
|
|
intr_dest.intr_dest_reg = 0;
|
|
local_addr.dma_localaddr_reg = 0;
|
|
host_addr.dma_hostaddr_reg = 0;
|
|
|
|
host_addr.dma_sys_addr = hostAddr;
|
|
MBCS_MMR_SET(mmr, MBCS_WR_DMA_SYS_ADDR, host_addr.dma_hostaddr_reg);
|
|
|
|
local_addr.dma_ram_addr = localAddr;
|
|
local_addr.dma_ram_sel = localRamSel;
|
|
MBCS_MMR_SET(mmr, MBCS_WR_DMA_LOC_ADDR, local_addr.dma_localaddr_reg);
|
|
|
|
wdma_control.dma_op_length = numPkts;
|
|
wdma_control.done_amo_en = amoEnable;
|
|
wdma_control.done_int_en = intrEnable;
|
|
wdma_control.pio_mem_n = peerIO;
|
|
MBCS_MMR_SET(mmr, MBCS_WR_DMA_CTRL, wdma_control.dma_control_reg);
|
|
|
|
amo_dest.dma_amo_sys_addr = amoHostDest;
|
|
amo_dest.dma_amo_mod_type = amoModType;
|
|
MBCS_MMR_SET(mmr, MBCS_WR_DMA_AMO_DEST, amo_dest.dma_amo_dest_reg);
|
|
|
|
intr_dest.address = intrHostDest;
|
|
intr_dest.int_vector = intrVector;
|
|
MBCS_MMR_SET(mmr, MBCS_WR_DMA_INT_DEST, intr_dest.intr_dest_reg);
|
|
|
|
}
|
|
|
|
static inline void mbcs_algo_set(void *mmr,
|
|
uint64_t amoHostDest,
|
|
uint64_t amoModType,
|
|
uint64_t intrHostDest,
|
|
uint64_t intrVector, uint64_t algoStepCount)
|
|
{
|
|
union dma_amo_dest amo_dest;
|
|
union intr_dest intr_dest;
|
|
union algo_step step;
|
|
|
|
step.algo_step_reg = 0;
|
|
intr_dest.intr_dest_reg = 0;
|
|
amo_dest.dma_amo_dest_reg = 0;
|
|
|
|
amo_dest.dma_amo_sys_addr = amoHostDest;
|
|
amo_dest.dma_amo_mod_type = amoModType;
|
|
MBCS_MMR_SET(mmr, MBCS_ALG_AMO_DEST, amo_dest.dma_amo_dest_reg);
|
|
|
|
intr_dest.address = intrHostDest;
|
|
intr_dest.int_vector = intrVector;
|
|
MBCS_MMR_SET(mmr, MBCS_ALG_INT_DEST, intr_dest.intr_dest_reg);
|
|
|
|
step.alg_step_cnt = algoStepCount;
|
|
MBCS_MMR_SET(mmr, MBCS_ALG_STEP, step.algo_step_reg);
|
|
}
|
|
|
|
static inline int mbcs_getdma_start(struct mbcs_soft *soft)
|
|
{
|
|
void *mmr_base;
|
|
struct getdma *gdma;
|
|
uint64_t numPkts;
|
|
union cm_control cm_control;
|
|
|
|
mmr_base = soft->mmr_base;
|
|
gdma = &soft->getdma;
|
|
|
|
/* check that host address got setup */
|
|
if (!gdma->hostAddr)
|
|
return -1;
|
|
|
|
numPkts =
|
|
(gdma->bytes + (MBCS_CACHELINE_SIZE - 1)) / MBCS_CACHELINE_SIZE;
|
|
|
|
/* program engine */
|
|
mbcs_getdma_set(mmr_base, tiocx_dma_addr(gdma->hostAddr),
|
|
gdma->localAddr,
|
|
(gdma->localAddr < MB2) ? 0 :
|
|
(gdma->localAddr < MB4) ? 1 :
|
|
(gdma->localAddr < MB6) ? 2 : 3,
|
|
numPkts,
|
|
gdma->DoneAmoEnable,
|
|
gdma->DoneIntEnable,
|
|
gdma->peerIO,
|
|
gdma->amoHostDest,
|
|
gdma->amoModType,
|
|
gdma->intrHostDest, gdma->intrVector);
|
|
|
|
/* start engine */
|
|
cm_control.cm_control_reg = MBCS_MMR_GET(mmr_base, MBCS_CM_CONTROL);
|
|
cm_control.rd_dma_go = 1;
|
|
MBCS_MMR_SET(mmr_base, MBCS_CM_CONTROL, cm_control.cm_control_reg);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
static inline int mbcs_putdma_start(struct mbcs_soft *soft)
|
|
{
|
|
void *mmr_base;
|
|
struct putdma *pdma;
|
|
uint64_t numPkts;
|
|
union cm_control cm_control;
|
|
|
|
mmr_base = soft->mmr_base;
|
|
pdma = &soft->putdma;
|
|
|
|
/* check that host address got setup */
|
|
if (!pdma->hostAddr)
|
|
return -1;
|
|
|
|
numPkts =
|
|
(pdma->bytes + (MBCS_CACHELINE_SIZE - 1)) / MBCS_CACHELINE_SIZE;
|
|
|
|
/* program engine */
|
|
mbcs_putdma_set(mmr_base, tiocx_dma_addr(pdma->hostAddr),
|
|
pdma->localAddr,
|
|
(pdma->localAddr < MB2) ? 0 :
|
|
(pdma->localAddr < MB4) ? 1 :
|
|
(pdma->localAddr < MB6) ? 2 : 3,
|
|
numPkts,
|
|
pdma->DoneAmoEnable,
|
|
pdma->DoneIntEnable,
|
|
pdma->peerIO,
|
|
pdma->amoHostDest,
|
|
pdma->amoModType,
|
|
pdma->intrHostDest, pdma->intrVector);
|
|
|
|
/* start engine */
|
|
cm_control.cm_control_reg = MBCS_MMR_GET(mmr_base, MBCS_CM_CONTROL);
|
|
cm_control.wr_dma_go = 1;
|
|
MBCS_MMR_SET(mmr_base, MBCS_CM_CONTROL, cm_control.cm_control_reg);
|
|
|
|
return 0;
|
|
|
|
}
|
|
|
|
static inline int mbcs_algo_start(struct mbcs_soft *soft)
|
|
{
|
|
struct algoblock *algo_soft = &soft->algo;
|
|
void *mmr_base = soft->mmr_base;
|
|
union cm_control cm_control;
|
|
|
|
if (down_interruptible(&soft->algolock))
|
|
return -ERESTARTSYS;
|
|
|
|
atomic_set(&soft->algo_done, 0);
|
|
|
|
mbcs_algo_set(mmr_base,
|
|
algo_soft->amoHostDest,
|
|
algo_soft->amoModType,
|
|
algo_soft->intrHostDest,
|
|
algo_soft->intrVector, algo_soft->algoStepCount);
|
|
|
|
/* start algorithm */
|
|
cm_control.cm_control_reg = MBCS_MMR_GET(mmr_base, MBCS_CM_CONTROL);
|
|
cm_control.alg_done_int_en = 1;
|
|
cm_control.alg_go = 1;
|
|
MBCS_MMR_SET(mmr_base, MBCS_CM_CONTROL, cm_control.cm_control_reg);
|
|
|
|
up(&soft->algolock);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static inline ssize_t
|
|
do_mbcs_sram_dmawrite(struct mbcs_soft *soft, uint64_t hostAddr,
|
|
size_t len, loff_t * off)
|
|
{
|
|
int rv = 0;
|
|
|
|
if (down_interruptible(&soft->dmawritelock))
|
|
return -ERESTARTSYS;
|
|
|
|
atomic_set(&soft->dmawrite_done, 0);
|
|
|
|
soft->putdma.hostAddr = hostAddr;
|
|
soft->putdma.localAddr = *off;
|
|
soft->putdma.bytes = len;
|
|
|
|
if (mbcs_putdma_start(soft) < 0) {
|
|
DBG(KERN_ALERT "do_mbcs_sram_dmawrite: "
|
|
"mbcs_putdma_start failed\n");
|
|
rv = -EAGAIN;
|
|
goto dmawrite_exit;
|
|
}
|
|
|
|
if (wait_event_interruptible(soft->dmawrite_queue,
|
|
atomic_read(&soft->dmawrite_done))) {
|
|
rv = -ERESTARTSYS;
|
|
goto dmawrite_exit;
|
|
}
|
|
|
|
rv = len;
|
|
*off += len;
|
|
|
|
dmawrite_exit:
|
|
up(&soft->dmawritelock);
|
|
|
|
return rv;
|
|
}
|
|
|
|
static inline ssize_t
|
|
do_mbcs_sram_dmaread(struct mbcs_soft *soft, uint64_t hostAddr,
|
|
size_t len, loff_t * off)
|
|
{
|
|
int rv = 0;
|
|
|
|
if (down_interruptible(&soft->dmareadlock))
|
|
return -ERESTARTSYS;
|
|
|
|
atomic_set(&soft->dmawrite_done, 0);
|
|
|
|
soft->getdma.hostAddr = hostAddr;
|
|
soft->getdma.localAddr = *off;
|
|
soft->getdma.bytes = len;
|
|
|
|
if (mbcs_getdma_start(soft) < 0) {
|
|
DBG(KERN_ALERT "mbcs_strategy: mbcs_getdma_start failed\n");
|
|
rv = -EAGAIN;
|
|
goto dmaread_exit;
|
|
}
|
|
|
|
if (wait_event_interruptible(soft->dmaread_queue,
|
|
atomic_read(&soft->dmaread_done))) {
|
|
rv = -ERESTARTSYS;
|
|
goto dmaread_exit;
|
|
}
|
|
|
|
rv = len;
|
|
*off += len;
|
|
|
|
dmaread_exit:
|
|
up(&soft->dmareadlock);
|
|
|
|
return rv;
|
|
}
|
|
|
|
int mbcs_open(struct inode *ip, struct file *fp)
|
|
{
|
|
struct mbcs_soft *soft;
|
|
int minor;
|
|
|
|
minor = iminor(ip);
|
|
|
|
list_for_each_entry(soft, &soft_list, list) {
|
|
if (soft->nasid == minor) {
|
|
fp->private_data = soft->cxdev;
|
|
return 0;
|
|
}
|
|
}
|
|
|
|
return -ENODEV;
|
|
}
|
|
|
|
ssize_t mbcs_sram_read(struct file * fp, char __user *buf, size_t len, loff_t * off)
|
|
{
|
|
struct cx_dev *cx_dev = fp->private_data;
|
|
struct mbcs_soft *soft = cx_dev->soft;
|
|
uint64_t hostAddr;
|
|
int rv = 0;
|
|
|
|
hostAddr = __get_dma_pages(GFP_KERNEL, get_order(len));
|
|
if (hostAddr == 0)
|
|
return -ENOMEM;
|
|
|
|
rv = do_mbcs_sram_dmawrite(soft, hostAddr, len, off);
|
|
if (rv < 0)
|
|
goto exit;
|
|
|
|
if (copy_to_user(buf, (void *)hostAddr, len))
|
|
rv = -EFAULT;
|
|
|
|
exit:
|
|
free_pages(hostAddr, get_order(len));
|
|
|
|
return rv;
|
|
}
|
|
|
|
ssize_t
|
|
mbcs_sram_write(struct file * fp, const char __user *buf, size_t len, loff_t * off)
|
|
{
|
|
struct cx_dev *cx_dev = fp->private_data;
|
|
struct mbcs_soft *soft = cx_dev->soft;
|
|
uint64_t hostAddr;
|
|
int rv = 0;
|
|
|
|
hostAddr = __get_dma_pages(GFP_KERNEL, get_order(len));
|
|
if (hostAddr == 0)
|
|
return -ENOMEM;
|
|
|
|
if (copy_from_user((void *)hostAddr, buf, len)) {
|
|
rv = -EFAULT;
|
|
goto exit;
|
|
}
|
|
|
|
rv = do_mbcs_sram_dmaread(soft, hostAddr, len, off);
|
|
|
|
exit:
|
|
free_pages(hostAddr, get_order(len));
|
|
|
|
return rv;
|
|
}
|
|
|
|
loff_t mbcs_sram_llseek(struct file * filp, loff_t off, int whence)
|
|
{
|
|
loff_t newpos;
|
|
|
|
switch (whence) {
|
|
case 0: /* SEEK_SET */
|
|
newpos = off;
|
|
break;
|
|
|
|
case 1: /* SEEK_CUR */
|
|
newpos = filp->f_pos + off;
|
|
break;
|
|
|
|
case 2: /* SEEK_END */
|
|
newpos = MBCS_SRAM_SIZE + off;
|
|
break;
|
|
|
|
default: /* can't happen */
|
|
return -EINVAL;
|
|
}
|
|
|
|
if (newpos < 0)
|
|
return -EINVAL;
|
|
|
|
filp->f_pos = newpos;
|
|
|
|
return newpos;
|
|
}
|
|
|
|
static uint64_t mbcs_pioaddr(struct mbcs_soft *soft, uint64_t offset)
|
|
{
|
|
uint64_t mmr_base;
|
|
|
|
mmr_base = (uint64_t) (soft->mmr_base + offset);
|
|
|
|
return mmr_base;
|
|
}
|
|
|
|
static void mbcs_debug_pioaddr_set(struct mbcs_soft *soft)
|
|
{
|
|
soft->debug_addr = mbcs_pioaddr(soft, MBCS_DEBUG_START);
|
|
}
|
|
|
|
static void mbcs_gscr_pioaddr_set(struct mbcs_soft *soft)
|
|
{
|
|
soft->gscr_addr = mbcs_pioaddr(soft, MBCS_GSCR_START);
|
|
}
|
|
|
|
int mbcs_gscr_mmap(struct file *fp, struct vm_area_struct *vma)
|
|
{
|
|
struct cx_dev *cx_dev = fp->private_data;
|
|
struct mbcs_soft *soft = cx_dev->soft;
|
|
|
|
if (vma->vm_pgoff != 0)
|
|
return -EINVAL;
|
|
|
|
vma->vm_page_prot = pgprot_noncached(vma->vm_page_prot);
|
|
|
|
/* Remap-pfn-range will mark the range VM_IO and VM_RESERVED */
|
|
if (remap_pfn_range(vma,
|
|
vma->vm_start,
|
|
__pa(soft->gscr_addr) >> PAGE_SHIFT,
|
|
PAGE_SIZE,
|
|
vma->vm_page_prot))
|
|
return -EAGAIN;
|
|
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* mbcs_completion_intr_handler - Primary completion handler.
|
|
* @irq: irq
|
|
* @arg: soft struct for device
|
|
* @ep: regs
|
|
*
|
|
*/
|
|
static irqreturn_t
|
|
mbcs_completion_intr_handler(int irq, void *arg, struct pt_regs *ep)
|
|
{
|
|
struct mbcs_soft *soft = (struct mbcs_soft *)arg;
|
|
void *mmr_base;
|
|
union cm_status cm_status;
|
|
union cm_control cm_control;
|
|
|
|
mmr_base = soft->mmr_base;
|
|
cm_status.cm_status_reg = MBCS_MMR_GET(mmr_base, MBCS_CM_STATUS);
|
|
|
|
if (cm_status.rd_dma_done) {
|
|
/* stop dma-read engine, clear status */
|
|
cm_control.cm_control_reg =
|
|
MBCS_MMR_GET(mmr_base, MBCS_CM_CONTROL);
|
|
cm_control.rd_dma_clr = 1;
|
|
MBCS_MMR_SET(mmr_base, MBCS_CM_CONTROL,
|
|
cm_control.cm_control_reg);
|
|
atomic_set(&soft->dmaread_done, 1);
|
|
wake_up(&soft->dmaread_queue);
|
|
}
|
|
if (cm_status.wr_dma_done) {
|
|
/* stop dma-write engine, clear status */
|
|
cm_control.cm_control_reg =
|
|
MBCS_MMR_GET(mmr_base, MBCS_CM_CONTROL);
|
|
cm_control.wr_dma_clr = 1;
|
|
MBCS_MMR_SET(mmr_base, MBCS_CM_CONTROL,
|
|
cm_control.cm_control_reg);
|
|
atomic_set(&soft->dmawrite_done, 1);
|
|
wake_up(&soft->dmawrite_queue);
|
|
}
|
|
if (cm_status.alg_done) {
|
|
/* clear status */
|
|
cm_control.cm_control_reg =
|
|
MBCS_MMR_GET(mmr_base, MBCS_CM_CONTROL);
|
|
cm_control.alg_done_clr = 1;
|
|
MBCS_MMR_SET(mmr_base, MBCS_CM_CONTROL,
|
|
cm_control.cm_control_reg);
|
|
atomic_set(&soft->algo_done, 1);
|
|
wake_up(&soft->algo_queue);
|
|
}
|
|
|
|
return IRQ_HANDLED;
|
|
}
|
|
|
|
/**
|
|
* mbcs_intr_alloc - Allocate interrupts.
|
|
* @dev: device pointer
|
|
*
|
|
*/
|
|
static int mbcs_intr_alloc(struct cx_dev *dev)
|
|
{
|
|
struct sn_irq_info *sn_irq;
|
|
struct mbcs_soft *soft;
|
|
struct getdma *getdma;
|
|
struct putdma *putdma;
|
|
struct algoblock *algo;
|
|
|
|
soft = dev->soft;
|
|
getdma = &soft->getdma;
|
|
putdma = &soft->putdma;
|
|
algo = &soft->algo;
|
|
|
|
soft->get_sn_irq = NULL;
|
|
soft->put_sn_irq = NULL;
|
|
soft->algo_sn_irq = NULL;
|
|
|
|
sn_irq = tiocx_irq_alloc(dev->cx_id.nasid, TIOCX_CORELET, -1, -1, -1);
|
|
if (sn_irq == NULL)
|
|
return -EAGAIN;
|
|
soft->get_sn_irq = sn_irq;
|
|
getdma->intrHostDest = sn_irq->irq_xtalkaddr;
|
|
getdma->intrVector = sn_irq->irq_irq;
|
|
if (request_irq(sn_irq->irq_irq,
|
|
(void *)mbcs_completion_intr_handler, SA_SHIRQ,
|
|
"MBCS get intr", (void *)soft)) {
|
|
tiocx_irq_free(soft->get_sn_irq);
|
|
return -EAGAIN;
|
|
}
|
|
|
|
sn_irq = tiocx_irq_alloc(dev->cx_id.nasid, TIOCX_CORELET, -1, -1, -1);
|
|
if (sn_irq == NULL) {
|
|
free_irq(soft->get_sn_irq->irq_irq, soft);
|
|
tiocx_irq_free(soft->get_sn_irq);
|
|
return -EAGAIN;
|
|
}
|
|
soft->put_sn_irq = sn_irq;
|
|
putdma->intrHostDest = sn_irq->irq_xtalkaddr;
|
|
putdma->intrVector = sn_irq->irq_irq;
|
|
if (request_irq(sn_irq->irq_irq,
|
|
(void *)mbcs_completion_intr_handler, SA_SHIRQ,
|
|
"MBCS put intr", (void *)soft)) {
|
|
tiocx_irq_free(soft->put_sn_irq);
|
|
free_irq(soft->get_sn_irq->irq_irq, soft);
|
|
tiocx_irq_free(soft->get_sn_irq);
|
|
return -EAGAIN;
|
|
}
|
|
|
|
sn_irq = tiocx_irq_alloc(dev->cx_id.nasid, TIOCX_CORELET, -1, -1, -1);
|
|
if (sn_irq == NULL) {
|
|
free_irq(soft->put_sn_irq->irq_irq, soft);
|
|
tiocx_irq_free(soft->put_sn_irq);
|
|
free_irq(soft->get_sn_irq->irq_irq, soft);
|
|
tiocx_irq_free(soft->get_sn_irq);
|
|
return -EAGAIN;
|
|
}
|
|
soft->algo_sn_irq = sn_irq;
|
|
algo->intrHostDest = sn_irq->irq_xtalkaddr;
|
|
algo->intrVector = sn_irq->irq_irq;
|
|
if (request_irq(sn_irq->irq_irq,
|
|
(void *)mbcs_completion_intr_handler, SA_SHIRQ,
|
|
"MBCS algo intr", (void *)soft)) {
|
|
tiocx_irq_free(soft->algo_sn_irq);
|
|
free_irq(soft->put_sn_irq->irq_irq, soft);
|
|
tiocx_irq_free(soft->put_sn_irq);
|
|
free_irq(soft->get_sn_irq->irq_irq, soft);
|
|
tiocx_irq_free(soft->get_sn_irq);
|
|
return -EAGAIN;
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* mbcs_intr_dealloc - Remove interrupts.
|
|
* @dev: device pointer
|
|
*
|
|
*/
|
|
static void mbcs_intr_dealloc(struct cx_dev *dev)
|
|
{
|
|
struct mbcs_soft *soft;
|
|
|
|
soft = dev->soft;
|
|
|
|
free_irq(soft->get_sn_irq->irq_irq, soft);
|
|
tiocx_irq_free(soft->get_sn_irq);
|
|
free_irq(soft->put_sn_irq->irq_irq, soft);
|
|
tiocx_irq_free(soft->put_sn_irq);
|
|
free_irq(soft->algo_sn_irq->irq_irq, soft);
|
|
tiocx_irq_free(soft->algo_sn_irq);
|
|
}
|
|
|
|
static inline int mbcs_hw_init(struct mbcs_soft *soft)
|
|
{
|
|
void *mmr_base = soft->mmr_base;
|
|
union cm_control cm_control;
|
|
union cm_req_timeout cm_req_timeout;
|
|
uint64_t err_stat;
|
|
|
|
cm_req_timeout.cm_req_timeout_reg =
|
|
MBCS_MMR_GET(mmr_base, MBCS_CM_REQ_TOUT);
|
|
|
|
cm_req_timeout.time_out = MBCS_CM_CONTROL_REQ_TOUT_MASK;
|
|
MBCS_MMR_SET(mmr_base, MBCS_CM_REQ_TOUT,
|
|
cm_req_timeout.cm_req_timeout_reg);
|
|
|
|
mbcs_gscr_pioaddr_set(soft);
|
|
mbcs_debug_pioaddr_set(soft);
|
|
|
|
/* clear errors */
|
|
err_stat = MBCS_MMR_GET(mmr_base, MBCS_CM_ERR_STAT);
|
|
MBCS_MMR_SET(mmr_base, MBCS_CM_CLR_ERR_STAT, err_stat);
|
|
MBCS_MMR_ZERO(mmr_base, MBCS_CM_ERROR_DETAIL1);
|
|
|
|
/* enable interrupts */
|
|
/* turn off 2^23 (INT_EN_PIO_REQ_ADDR_INV) */
|
|
MBCS_MMR_SET(mmr_base, MBCS_CM_ERR_INT_EN, 0x3ffffff7e00ffUL);
|
|
|
|
/* arm status regs and clear engines */
|
|
cm_control.cm_control_reg = MBCS_MMR_GET(mmr_base, MBCS_CM_CONTROL);
|
|
cm_control.rearm_stat_regs = 1;
|
|
cm_control.alg_clr = 1;
|
|
cm_control.wr_dma_clr = 1;
|
|
cm_control.rd_dma_clr = 1;
|
|
|
|
MBCS_MMR_SET(mmr_base, MBCS_CM_CONTROL, cm_control.cm_control_reg);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static ssize_t show_algo(struct device *dev, struct device_attribute *attr, char *buf)
|
|
{
|
|
struct cx_dev *cx_dev = to_cx_dev(dev);
|
|
struct mbcs_soft *soft = cx_dev->soft;
|
|
uint64_t debug0;
|
|
|
|
/*
|
|
* By convention, the first debug register contains the
|
|
* algorithm number and revision.
|
|
*/
|
|
debug0 = *(uint64_t *) soft->debug_addr;
|
|
|
|
return sprintf(buf, "0x%lx 0x%lx\n",
|
|
(debug0 >> 32), (debug0 & 0xffffffff));
|
|
}
|
|
|
|
static ssize_t store_algo(struct device *dev, struct device_attribute *attr, const char *buf, size_t count)
|
|
{
|
|
int n;
|
|
struct cx_dev *cx_dev = to_cx_dev(dev);
|
|
struct mbcs_soft *soft = cx_dev->soft;
|
|
|
|
if (count <= 0)
|
|
return 0;
|
|
|
|
n = simple_strtoul(buf, NULL, 0);
|
|
|
|
if (n == 1) {
|
|
mbcs_algo_start(soft);
|
|
if (wait_event_interruptible(soft->algo_queue,
|
|
atomic_read(&soft->algo_done)))
|
|
return -ERESTARTSYS;
|
|
}
|
|
|
|
return count;
|
|
}
|
|
|
|
DEVICE_ATTR(algo, 0644, show_algo, store_algo);
|
|
|
|
/**
|
|
* mbcs_probe - Initialize for device
|
|
* @dev: device pointer
|
|
* @device_id: id table pointer
|
|
*
|
|
*/
|
|
static int mbcs_probe(struct cx_dev *dev, const struct cx_device_id *id)
|
|
{
|
|
struct mbcs_soft *soft;
|
|
|
|
dev->soft = NULL;
|
|
|
|
soft = kzalloc(sizeof(struct mbcs_soft), GFP_KERNEL);
|
|
if (soft == NULL)
|
|
return -ENOMEM;
|
|
|
|
soft->nasid = dev->cx_id.nasid;
|
|
list_add(&soft->list, &soft_list);
|
|
soft->mmr_base = (void *)tiocx_swin_base(dev->cx_id.nasid);
|
|
dev->soft = soft;
|
|
soft->cxdev = dev;
|
|
|
|
init_waitqueue_head(&soft->dmawrite_queue);
|
|
init_waitqueue_head(&soft->dmaread_queue);
|
|
init_waitqueue_head(&soft->algo_queue);
|
|
|
|
init_MUTEX(&soft->dmawritelock);
|
|
init_MUTEX(&soft->dmareadlock);
|
|
init_MUTEX(&soft->algolock);
|
|
|
|
mbcs_getdma_init(&soft->getdma);
|
|
mbcs_putdma_init(&soft->putdma);
|
|
mbcs_algo_init(&soft->algo);
|
|
|
|
mbcs_hw_init(soft);
|
|
|
|
/* Allocate interrupts */
|
|
mbcs_intr_alloc(dev);
|
|
|
|
device_create_file(&dev->dev, &dev_attr_algo);
|
|
|
|
return 0;
|
|
}
|
|
|
|
static int mbcs_remove(struct cx_dev *dev)
|
|
{
|
|
if (dev->soft) {
|
|
mbcs_intr_dealloc(dev);
|
|
kfree(dev->soft);
|
|
}
|
|
|
|
device_remove_file(&dev->dev, &dev_attr_algo);
|
|
|
|
return 0;
|
|
}
|
|
|
|
const struct cx_device_id __devinitdata mbcs_id_table[] = {
|
|
{
|
|
.part_num = MBCS_PART_NUM,
|
|
.mfg_num = MBCS_MFG_NUM,
|
|
},
|
|
{
|
|
.part_num = MBCS_PART_NUM_ALG0,
|
|
.mfg_num = MBCS_MFG_NUM,
|
|
},
|
|
{0, 0}
|
|
};
|
|
|
|
MODULE_DEVICE_TABLE(cx, mbcs_id_table);
|
|
|
|
struct cx_drv mbcs_driver = {
|
|
.name = DEVICE_NAME,
|
|
.id_table = mbcs_id_table,
|
|
.probe = mbcs_probe,
|
|
.remove = mbcs_remove,
|
|
};
|
|
|
|
static void __exit mbcs_exit(void)
|
|
{
|
|
int rv;
|
|
|
|
rv = unregister_chrdev(mbcs_major, DEVICE_NAME);
|
|
if (rv < 0)
|
|
DBG(KERN_ALERT "Error in unregister_chrdev: %d\n", rv);
|
|
|
|
cx_driver_unregister(&mbcs_driver);
|
|
}
|
|
|
|
static int __init mbcs_init(void)
|
|
{
|
|
int rv;
|
|
|
|
if (!ia64_platform_is("sn2"))
|
|
return -ENODEV;
|
|
|
|
// Put driver into chrdevs[]. Get major number.
|
|
rv = register_chrdev(mbcs_major, DEVICE_NAME, &mbcs_ops);
|
|
if (rv < 0) {
|
|
DBG(KERN_ALERT "mbcs_init: can't get major number. %d\n", rv);
|
|
return rv;
|
|
}
|
|
mbcs_major = rv;
|
|
|
|
return cx_driver_register(&mbcs_driver);
|
|
}
|
|
|
|
module_init(mbcs_init);
|
|
module_exit(mbcs_exit);
|
|
|
|
MODULE_AUTHOR("Bruce Losure <blosure@sgi.com>");
|
|
MODULE_DESCRIPTION("Driver for MOATB Core Services");
|
|
MODULE_LICENSE("GPL");
|