mirror of
https://github.com/AuxXxilium/linux_dsm_epyc7002.git
synced 2024-12-27 07:05:08 +07:00
c30362cc32
Implement a circular buffer on allocated system memory. Read and write indices are stored on the control block which is also shared between the device and the host. Two new functions are exported from the DMA module: mei_dma_ring_write, and mei_dma_ring_empty_slots. The former simply copy a packet on the TX DMA circular buffer and later, returns the number of empty slots on the TX DMA circular buffer. Signed-off-by: Tomas Winkler <tomas.winkler@intel.com> Signed-off-by: Alexander Usyskin <alexander.usyskin@intel.com> Signed-off-by: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
270 lines
5.8 KiB
C
270 lines
5.8 KiB
C
// SPDX-License-Identifier: GPL-2.0
|
|
/*
|
|
* Copyright(c) 2016 - 2018 Intel Corporation. All rights reserved.
|
|
*/
|
|
#include <linux/dma-mapping.h>
|
|
#include <linux/mei.h>
|
|
|
|
#include "mei_dev.h"
|
|
|
|
/**
|
|
* mei_dmam_dscr_alloc() - allocate a managed coherent buffer
|
|
* for the dma descriptor
|
|
* @dev: mei_device
|
|
* @dscr: dma descriptor
|
|
*
|
|
* Return:
|
|
* * 0 - on success or zero allocation request
|
|
* * -EINVAL - if size is not power of 2
|
|
* * -ENOMEM - of allocation has failed
|
|
*/
|
|
static int mei_dmam_dscr_alloc(struct mei_device *dev,
|
|
struct mei_dma_dscr *dscr)
|
|
{
|
|
if (!dscr->size)
|
|
return 0;
|
|
|
|
if (WARN_ON(!is_power_of_2(dscr->size)))
|
|
return -EINVAL;
|
|
|
|
if (dscr->vaddr)
|
|
return 0;
|
|
|
|
dscr->vaddr = dmam_alloc_coherent(dev->dev, dscr->size, &dscr->daddr,
|
|
GFP_KERNEL);
|
|
if (!dscr->vaddr)
|
|
return -ENOMEM;
|
|
|
|
return 0;
|
|
}
|
|
|
|
/**
|
|
* mei_dmam_dscr_free() - free a managed coherent buffer
|
|
* from the dma descriptor
|
|
* @dev: mei_device
|
|
* @dscr: dma descriptor
|
|
*/
|
|
static void mei_dmam_dscr_free(struct mei_device *dev,
|
|
struct mei_dma_dscr *dscr)
|
|
{
|
|
if (!dscr->vaddr)
|
|
return;
|
|
|
|
dmam_free_coherent(dev->dev, dscr->size, dscr->vaddr, dscr->daddr);
|
|
dscr->vaddr = NULL;
|
|
}
|
|
|
|
/**
|
|
* mei_dmam_ring_free() - free dma ring buffers
|
|
* @dev: mei device
|
|
*/
|
|
void mei_dmam_ring_free(struct mei_device *dev)
|
|
{
|
|
int i;
|
|
|
|
for (i = 0; i < DMA_DSCR_NUM; i++)
|
|
mei_dmam_dscr_free(dev, &dev->dr_dscr[i]);
|
|
}
|
|
|
|
/**
|
|
* mei_dmam_ring_alloc() - allocate dma ring buffers
|
|
* @dev: mei device
|
|
*
|
|
* Return: -ENOMEM on allocation failure 0 otherwise
|
|
*/
|
|
int mei_dmam_ring_alloc(struct mei_device *dev)
|
|
{
|
|
int i;
|
|
|
|
for (i = 0; i < DMA_DSCR_NUM; i++)
|
|
if (mei_dmam_dscr_alloc(dev, &dev->dr_dscr[i]))
|
|
goto err;
|
|
|
|
return 0;
|
|
|
|
err:
|
|
mei_dmam_ring_free(dev);
|
|
return -ENOMEM;
|
|
}
|
|
|
|
/**
|
|
* mei_dma_ring_is_allocated() - check if dma ring is allocated
|
|
* @dev: mei device
|
|
*
|
|
* Return: true if dma ring is allocated
|
|
*/
|
|
bool mei_dma_ring_is_allocated(struct mei_device *dev)
|
|
{
|
|
return !!dev->dr_dscr[DMA_DSCR_HOST].vaddr;
|
|
}
|
|
|
|
static inline
|
|
struct hbm_dma_ring_ctrl *mei_dma_ring_ctrl(struct mei_device *dev)
|
|
{
|
|
return (struct hbm_dma_ring_ctrl *)dev->dr_dscr[DMA_DSCR_CTRL].vaddr;
|
|
}
|
|
|
|
/**
|
|
* mei_dma_ring_reset() - reset the dma control block
|
|
* @dev: mei device
|
|
*/
|
|
void mei_dma_ring_reset(struct mei_device *dev)
|
|
{
|
|
struct hbm_dma_ring_ctrl *ctrl = mei_dma_ring_ctrl(dev);
|
|
|
|
if (!ctrl)
|
|
return;
|
|
|
|
memset(ctrl, 0, sizeof(*ctrl));
|
|
}
|
|
|
|
/**
|
|
* mei_dma_copy_from() - copy from dma ring into buffer
|
|
* @dev: mei device
|
|
* @buf: data buffer
|
|
* @offset: offset in slots.
|
|
* @n: number of slots to copy.
|
|
*/
|
|
static size_t mei_dma_copy_from(struct mei_device *dev, unsigned char *buf,
|
|
u32 offset, u32 n)
|
|
{
|
|
unsigned char *dbuf = dev->dr_dscr[DMA_DSCR_DEVICE].vaddr;
|
|
|
|
size_t b_offset = offset << 2;
|
|
size_t b_n = n << 2;
|
|
|
|
memcpy(buf, dbuf + b_offset, b_n);
|
|
|
|
return b_n;
|
|
}
|
|
|
|
/**
|
|
* mei_dma_copy_to() - copy to a buffer to the dma ring
|
|
* @dev: mei device
|
|
* @buf: data buffer
|
|
* @offset: offset in slots.
|
|
* @n: number of slots to copy.
|
|
*/
|
|
static size_t mei_dma_copy_to(struct mei_device *dev, unsigned char *buf,
|
|
u32 offset, u32 n)
|
|
{
|
|
unsigned char *hbuf = dev->dr_dscr[DMA_DSCR_HOST].vaddr;
|
|
|
|
size_t b_offset = offset << 2;
|
|
size_t b_n = n << 2;
|
|
|
|
memcpy(hbuf + b_offset, buf, b_n);
|
|
|
|
return b_n;
|
|
}
|
|
|
|
/**
|
|
* mei_dma_ring_read() - read data from the ring
|
|
* @dev: mei device
|
|
* @buf: buffer to read into: may be NULL in case of droping the data.
|
|
* @len: length to read.
|
|
*/
|
|
void mei_dma_ring_read(struct mei_device *dev, unsigned char *buf, u32 len)
|
|
{
|
|
struct hbm_dma_ring_ctrl *ctrl = mei_dma_ring_ctrl(dev);
|
|
u32 dbuf_depth;
|
|
u32 rd_idx, rem, slots;
|
|
|
|
if (WARN_ON(!ctrl))
|
|
return;
|
|
|
|
dev_dbg(dev->dev, "reading from dma %u bytes\n", len);
|
|
|
|
if (!len)
|
|
return;
|
|
|
|
dbuf_depth = dev->dr_dscr[DMA_DSCR_DEVICE].size >> 2;
|
|
rd_idx = READ_ONCE(ctrl->dbuf_rd_idx) & (dbuf_depth - 1);
|
|
slots = mei_data2slots(len);
|
|
|
|
/* if buf is NULL we drop the packet by advancing the pointer.*/
|
|
if (!buf)
|
|
goto out;
|
|
|
|
if (rd_idx + slots > dbuf_depth) {
|
|
buf += mei_dma_copy_from(dev, buf, rd_idx, dbuf_depth - rd_idx);
|
|
rem = slots - (dbuf_depth - rd_idx);
|
|
rd_idx = 0;
|
|
} else {
|
|
rem = slots;
|
|
}
|
|
|
|
mei_dma_copy_from(dev, buf, rd_idx, rem);
|
|
out:
|
|
WRITE_ONCE(ctrl->dbuf_rd_idx, ctrl->dbuf_rd_idx + slots);
|
|
}
|
|
|
|
static inline u32 mei_dma_ring_hbuf_depth(struct mei_device *dev)
|
|
{
|
|
return dev->dr_dscr[DMA_DSCR_HOST].size >> 2;
|
|
}
|
|
|
|
/**
|
|
* mei_dma_ring_empty_slots() - calaculate number of empty slots in dma ring
|
|
* @dev: mei_device
|
|
*
|
|
* Return: number of empty slots
|
|
*/
|
|
u32 mei_dma_ring_empty_slots(struct mei_device *dev)
|
|
{
|
|
struct hbm_dma_ring_ctrl *ctrl = mei_dma_ring_ctrl(dev);
|
|
u32 wr_idx, rd_idx, hbuf_depth, empty;
|
|
|
|
if (!mei_dma_ring_is_allocated(dev))
|
|
return 0;
|
|
|
|
if (WARN_ON(!ctrl))
|
|
return 0;
|
|
|
|
/* easier to work in slots */
|
|
hbuf_depth = mei_dma_ring_hbuf_depth(dev);
|
|
rd_idx = READ_ONCE(ctrl->hbuf_rd_idx);
|
|
wr_idx = READ_ONCE(ctrl->hbuf_wr_idx);
|
|
|
|
if (rd_idx > wr_idx)
|
|
empty = rd_idx - wr_idx;
|
|
else
|
|
empty = hbuf_depth - (wr_idx - rd_idx);
|
|
|
|
return empty;
|
|
}
|
|
|
|
/**
|
|
* mei_dma_ring_write - write data to dma ring host buffer
|
|
*
|
|
* @dev: mei_device
|
|
* @buf: data will be written
|
|
* @len: data length
|
|
*/
|
|
void mei_dma_ring_write(struct mei_device *dev, unsigned char *buf, u32 len)
|
|
{
|
|
struct hbm_dma_ring_ctrl *ctrl = mei_dma_ring_ctrl(dev);
|
|
u32 hbuf_depth;
|
|
u32 wr_idx, rem, slots;
|
|
|
|
if (WARN_ON(!ctrl))
|
|
return;
|
|
|
|
dev_dbg(dev->dev, "writing to dma %u bytes\n", len);
|
|
hbuf_depth = mei_dma_ring_hbuf_depth(dev);
|
|
wr_idx = READ_ONCE(ctrl->hbuf_wr_idx) & (hbuf_depth - 1);
|
|
slots = mei_data2slots(len);
|
|
|
|
if (wr_idx + slots > hbuf_depth) {
|
|
buf += mei_dma_copy_to(dev, buf, wr_idx, hbuf_depth - wr_idx);
|
|
rem = slots - (hbuf_depth - wr_idx);
|
|
wr_idx = 0;
|
|
} else {
|
|
rem = slots;
|
|
}
|
|
|
|
mei_dma_copy_to(dev, buf, wr_idx, rem);
|
|
|
|
WRITE_ONCE(ctrl->hbuf_wr_idx, ctrl->hbuf_wr_idx + slots);
|
|
}
|