linux_dsm_epyc7002/drivers/soc/ti/knav_qmss_queue.c
Sandeep Nair 41f93af900 soc: ti: add Keystone Navigator QMSS driver
The QMSS (Queue Manager Sub System) found on Keystone SOCs is one of
the main hardware sub system which forms the backbone of the Keystone
Multi-core Navigator. QMSS consist of queue managers, packed-data structure
processors(PDSP), linking RAM, descriptor pools and infrastructure
Packet DMA.

The Queue Manager is a hardware module that is responsible for accelerating
management of the packet queues. Packets are queued/de-queued by writing or
reading descriptor address to a particular memory mapped location. The PDSPs
perform QMSS related functions like accumulation, QoS, or event management.
Linking RAM registers are used to link the descriptors which are stored in
descriptor RAM. Descriptor RAM is configurable as internal or external memory.

The QMSS driver manages the PDSP setups, linking RAM regions,
queue pool management (allocation, push, pop and notify) and descriptor
pool management. The specifics on the device tree bindings for
QMSS can be found in:
	Documentation/devicetree/bindings/soc/keystone-navigator-qmss.txt

Cc: Greg Kroah-Hartman <gregkh@linuxfoundation.org>
Cc: Kumar Gala <galak@codeaurora.org>
Cc: Olof Johansson <olof@lixom.net>
Cc: Arnd Bergmann <arnd@arndb.de>
Cc: Grant Likely <grant.likely@linaro.org>
Cc: Rob Herring <robh+dt@kernel.org>
Cc: Mark Rutland <mark.rutland@arm.com>
Signed-off-by: Sandeep Nair <sandeep_n@ti.com>
Signed-off-by: Santosh Shilimkar <santosh.shilimkar@ti.com>
2014-09-24 09:49:14 -04:00

1817 lines
44 KiB
C

/*
* Keystone Queue Manager subsystem driver
*
* Copyright (C) 2014 Texas Instruments Incorporated - http://www.ti.com
* Authors: Sandeep Nair <sandeep_n@ti.com>
* Cyril Chemparathy <cyril@ti.com>
* Santosh Shilimkar <santosh.shilimkar@ti.com>
*
* This program is free software; you can redistribute it and/or
* modify it under the terms of the GNU General Public License
* version 2 as published by the Free Software Foundation.
*
* This program is distributed in the hope that it will be useful, but
* WITHOUT ANY WARRANTY; without even the implied warranty of
* MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
* General Public License for more details.
*/
#include <linux/kernel.h>
#include <linux/module.h>
#include <linux/device.h>
#include <linux/clk.h>
#include <linux/io.h>
#include <linux/interrupt.h>
#include <linux/bitops.h>
#include <linux/slab.h>
#include <linux/spinlock.h>
#include <linux/platform_device.h>
#include <linux/dma-mapping.h>
#include <linux/of.h>
#include <linux/of_irq.h>
#include <linux/of_device.h>
#include <linux/of_address.h>
#include <linux/pm_runtime.h>
#include <linux/firmware.h>
#include <linux/debugfs.h>
#include <linux/seq_file.h>
#include <linux/string.h>
#include <linux/soc/ti/knav_qmss.h>
#include "knav_qmss.h"
static struct knav_device *kdev;
static DEFINE_MUTEX(knav_dev_lock);
/* Queue manager register indices in DTS */
#define KNAV_QUEUE_PEEK_REG_INDEX 0
#define KNAV_QUEUE_STATUS_REG_INDEX 1
#define KNAV_QUEUE_CONFIG_REG_INDEX 2
#define KNAV_QUEUE_REGION_REG_INDEX 3
#define KNAV_QUEUE_PUSH_REG_INDEX 4
#define KNAV_QUEUE_POP_REG_INDEX 5
/* PDSP register indices in DTS */
#define KNAV_QUEUE_PDSP_IRAM_REG_INDEX 0
#define KNAV_QUEUE_PDSP_REGS_REG_INDEX 1
#define KNAV_QUEUE_PDSP_INTD_REG_INDEX 2
#define KNAV_QUEUE_PDSP_CMD_REG_INDEX 3
#define knav_queue_idx_to_inst(kdev, idx) \
(kdev->instances + (idx << kdev->inst_shift))
#define for_each_handle_rcu(qh, inst) \
list_for_each_entry_rcu(qh, &inst->handles, list)
#define for_each_instance(idx, inst, kdev) \
for (idx = 0, inst = kdev->instances; \
idx < (kdev)->num_queues_in_use; \
idx++, inst = knav_queue_idx_to_inst(kdev, idx))
/**
* knav_queue_notify: qmss queue notfier call
*
* @inst: qmss queue instance like accumulator
*/
void knav_queue_notify(struct knav_queue_inst *inst)
{
struct knav_queue *qh;
if (!inst)
return;
rcu_read_lock();
for_each_handle_rcu(qh, inst) {
if (atomic_read(&qh->notifier_enabled) <= 0)
continue;
if (WARN_ON(!qh->notifier_fn))
continue;
atomic_inc(&qh->stats.notifies);
qh->notifier_fn(qh->notifier_fn_arg);
}
rcu_read_unlock();
}
EXPORT_SYMBOL_GPL(knav_queue_notify);
static irqreturn_t knav_queue_int_handler(int irq, void *_instdata)
{
struct knav_queue_inst *inst = _instdata;
knav_queue_notify(inst);
return IRQ_HANDLED;
}
static int knav_queue_setup_irq(struct knav_range_info *range,
struct knav_queue_inst *inst)
{
unsigned queue = inst->id - range->queue_base;
unsigned long cpu_map;
int ret = 0, irq;
if (range->flags & RANGE_HAS_IRQ) {
irq = range->irqs[queue].irq;
cpu_map = range->irqs[queue].cpu_map;
ret = request_irq(irq, knav_queue_int_handler, 0,
inst->irq_name, inst);
if (ret)
return ret;
disable_irq(irq);
if (cpu_map) {
ret = irq_set_affinity_hint(irq, to_cpumask(&cpu_map));
if (ret) {
dev_warn(range->kdev->dev,
"Failed to set IRQ affinity\n");
return ret;
}
}
}
return ret;
}
static void knav_queue_free_irq(struct knav_queue_inst *inst)
{
struct knav_range_info *range = inst->range;
unsigned queue = inst->id - inst->range->queue_base;
int irq;
if (range->flags & RANGE_HAS_IRQ) {
irq = range->irqs[queue].irq;
irq_set_affinity_hint(irq, NULL);
free_irq(irq, inst);
}
}
static inline bool knav_queue_is_busy(struct knav_queue_inst *inst)
{
return !list_empty(&inst->handles);
}
static inline bool knav_queue_is_reserved(struct knav_queue_inst *inst)
{
return inst->range->flags & RANGE_RESERVED;
}
static inline bool knav_queue_is_shared(struct knav_queue_inst *inst)
{
struct knav_queue *tmp;
rcu_read_lock();
for_each_handle_rcu(tmp, inst) {
if (tmp->flags & KNAV_QUEUE_SHARED) {
rcu_read_unlock();
return true;
}
}
rcu_read_unlock();
return false;
}
static inline bool knav_queue_match_type(struct knav_queue_inst *inst,
unsigned type)
{
if ((type == KNAV_QUEUE_QPEND) &&
(inst->range->flags & RANGE_HAS_IRQ)) {
return true;
} else if ((type == KNAV_QUEUE_ACC) &&
(inst->range->flags & RANGE_HAS_ACCUMULATOR)) {
return true;
} else if ((type == KNAV_QUEUE_GP) &&
!(inst->range->flags &
(RANGE_HAS_ACCUMULATOR | RANGE_HAS_IRQ))) {
return true;
}
return false;
}
static inline struct knav_queue_inst *
knav_queue_match_id_to_inst(struct knav_device *kdev, unsigned id)
{
struct knav_queue_inst *inst;
int idx;
for_each_instance(idx, inst, kdev) {
if (inst->id == id)
return inst;
}
return NULL;
}
static inline struct knav_queue_inst *knav_queue_find_by_id(int id)
{
if (kdev->base_id <= id &&
kdev->base_id + kdev->num_queues > id) {
id -= kdev->base_id;
return knav_queue_match_id_to_inst(kdev, id);
}
return NULL;
}
static struct knav_queue *__knav_queue_open(struct knav_queue_inst *inst,
const char *name, unsigned flags)
{
struct knav_queue *qh;
unsigned id;
int ret = 0;
qh = devm_kzalloc(inst->kdev->dev, sizeof(*qh), GFP_KERNEL);
if (!qh)
return ERR_PTR(-ENOMEM);
qh->flags = flags;
qh->inst = inst;
id = inst->id - inst->qmgr->start_queue;
qh->reg_push = &inst->qmgr->reg_push[id];
qh->reg_pop = &inst->qmgr->reg_pop[id];
qh->reg_peek = &inst->qmgr->reg_peek[id];
/* first opener? */
if (!knav_queue_is_busy(inst)) {
struct knav_range_info *range = inst->range;
inst->name = kstrndup(name, KNAV_NAME_SIZE, GFP_KERNEL);
if (range->ops && range->ops->open_queue)
ret = range->ops->open_queue(range, inst, flags);
if (ret) {
devm_kfree(inst->kdev->dev, qh);
return ERR_PTR(ret);
}
}
list_add_tail_rcu(&qh->list, &inst->handles);
return qh;
}
static struct knav_queue *
knav_queue_open_by_id(const char *name, unsigned id, unsigned flags)
{
struct knav_queue_inst *inst;
struct knav_queue *qh;
mutex_lock(&knav_dev_lock);
qh = ERR_PTR(-ENODEV);
inst = knav_queue_find_by_id(id);
if (!inst)
goto unlock_ret;
qh = ERR_PTR(-EEXIST);
if (!(flags & KNAV_QUEUE_SHARED) && knav_queue_is_busy(inst))
goto unlock_ret;
qh = ERR_PTR(-EBUSY);
if ((flags & KNAV_QUEUE_SHARED) &&
(knav_queue_is_busy(inst) && !knav_queue_is_shared(inst)))
goto unlock_ret;
qh = __knav_queue_open(inst, name, flags);
unlock_ret:
mutex_unlock(&knav_dev_lock);
return qh;
}
static struct knav_queue *knav_queue_open_by_type(const char *name,
unsigned type, unsigned flags)
{
struct knav_queue_inst *inst;
struct knav_queue *qh = ERR_PTR(-EINVAL);
int idx;
mutex_lock(&knav_dev_lock);
for_each_instance(idx, inst, kdev) {
if (knav_queue_is_reserved(inst))
continue;
if (!knav_queue_match_type(inst, type))
continue;
if (knav_queue_is_busy(inst))
continue;
qh = __knav_queue_open(inst, name, flags);
goto unlock_ret;
}
unlock_ret:
mutex_unlock(&knav_dev_lock);
return qh;
}
static void knav_queue_set_notify(struct knav_queue_inst *inst, bool enabled)
{
struct knav_range_info *range = inst->range;
if (range->ops && range->ops->set_notify)
range->ops->set_notify(range, inst, enabled);
}
static int knav_queue_enable_notifier(struct knav_queue *qh)
{
struct knav_queue_inst *inst = qh->inst;
bool first;
if (WARN_ON(!qh->notifier_fn))
return -EINVAL;
/* Adjust the per handle notifier count */
first = (atomic_inc_return(&qh->notifier_enabled) == 1);
if (!first)
return 0; /* nothing to do */
/* Now adjust the per instance notifier count */
first = (atomic_inc_return(&inst->num_notifiers) == 1);
if (first)
knav_queue_set_notify(inst, true);
return 0;
}
static int knav_queue_disable_notifier(struct knav_queue *qh)
{
struct knav_queue_inst *inst = qh->inst;
bool last;
last = (atomic_dec_return(&qh->notifier_enabled) == 0);
if (!last)
return 0; /* nothing to do */
last = (atomic_dec_return(&inst->num_notifiers) == 0);
if (last)
knav_queue_set_notify(inst, false);
return 0;
}
static int knav_queue_set_notifier(struct knav_queue *qh,
struct knav_queue_notify_config *cfg)
{
knav_queue_notify_fn old_fn = qh->notifier_fn;
if (!cfg)
return -EINVAL;
if (!(qh->inst->range->flags & (RANGE_HAS_ACCUMULATOR | RANGE_HAS_IRQ)))
return -ENOTSUPP;
if (!cfg->fn && old_fn)
knav_queue_disable_notifier(qh);
qh->notifier_fn = cfg->fn;
qh->notifier_fn_arg = cfg->fn_arg;
if (cfg->fn && !old_fn)
knav_queue_enable_notifier(qh);
return 0;
}
static int knav_gp_set_notify(struct knav_range_info *range,
struct knav_queue_inst *inst,
bool enabled)
{
unsigned queue;
if (range->flags & RANGE_HAS_IRQ) {
queue = inst->id - range->queue_base;
if (enabled)
enable_irq(range->irqs[queue].irq);
else
disable_irq_nosync(range->irqs[queue].irq);
}
return 0;
}
static int knav_gp_open_queue(struct knav_range_info *range,
struct knav_queue_inst *inst, unsigned flags)
{
return knav_queue_setup_irq(range, inst);
}
static int knav_gp_close_queue(struct knav_range_info *range,
struct knav_queue_inst *inst)
{
knav_queue_free_irq(inst);
return 0;
}
struct knav_range_ops knav_gp_range_ops = {
.set_notify = knav_gp_set_notify,
.open_queue = knav_gp_open_queue,
.close_queue = knav_gp_close_queue,
};
static int knav_queue_get_count(void *qhandle)
{
struct knav_queue *qh = qhandle;
struct knav_queue_inst *inst = qh->inst;
return readl_relaxed(&qh->reg_peek[0].entry_count) +
atomic_read(&inst->desc_count);
}
static void knav_queue_debug_show_instance(struct seq_file *s,
struct knav_queue_inst *inst)
{
struct knav_device *kdev = inst->kdev;
struct knav_queue *qh;
if (!knav_queue_is_busy(inst))
return;
seq_printf(s, "\tqueue id %d (%s)\n",
kdev->base_id + inst->id, inst->name);
for_each_handle_rcu(qh, inst) {
seq_printf(s, "\t\thandle %p: ", qh);
seq_printf(s, "pushes %8d, ",
atomic_read(&qh->stats.pushes));
seq_printf(s, "pops %8d, ",
atomic_read(&qh->stats.pops));
seq_printf(s, "count %8d, ",
knav_queue_get_count(qh));
seq_printf(s, "notifies %8d, ",
atomic_read(&qh->stats.notifies));
seq_printf(s, "push errors %8d, ",
atomic_read(&qh->stats.push_errors));
seq_printf(s, "pop errors %8d\n",
atomic_read(&qh->stats.pop_errors));
}
}
static int knav_queue_debug_show(struct seq_file *s, void *v)
{
struct knav_queue_inst *inst;
int idx;
mutex_lock(&knav_dev_lock);
seq_printf(s, "%s: %u-%u\n",
dev_name(kdev->dev), kdev->base_id,
kdev->base_id + kdev->num_queues - 1);
for_each_instance(idx, inst, kdev)
knav_queue_debug_show_instance(s, inst);
mutex_unlock(&knav_dev_lock);
return 0;
}
static int knav_queue_debug_open(struct inode *inode, struct file *file)
{
return single_open(file, knav_queue_debug_show, NULL);
}
static const struct file_operations knav_queue_debug_ops = {
.open = knav_queue_debug_open,
.read = seq_read,
.llseek = seq_lseek,
.release = single_release,
};
static inline int knav_queue_pdsp_wait(u32 * __iomem addr, unsigned timeout,
u32 flags)
{
unsigned long end;
u32 val = 0;
end = jiffies + msecs_to_jiffies(timeout);
while (time_after(end, jiffies)) {
val = readl_relaxed(addr);
if (flags)
val &= flags;
if (!val)
break;
cpu_relax();
}
return val ? -ETIMEDOUT : 0;
}
static int knav_queue_flush(struct knav_queue *qh)
{
struct knav_queue_inst *inst = qh->inst;
unsigned id = inst->id - inst->qmgr->start_queue;
atomic_set(&inst->desc_count, 0);
writel_relaxed(0, &inst->qmgr->reg_push[id].ptr_size_thresh);
return 0;
}
/**
* knav_queue_open() - open a hardware queue
* @name - name to give the queue handle
* @id - desired queue number if any or specifes the type
* of queue
* @flags - the following flags are applicable to queues:
* KNAV_QUEUE_SHARED - allow the queue to be shared. Queues are
* exclusive by default.
* Subsequent attempts to open a shared queue should
* also have this flag.
*
* Returns a handle to the open hardware queue if successful. Use IS_ERR()
* to check the returned value for error codes.
*/
void *knav_queue_open(const char *name, unsigned id,
unsigned flags)
{
struct knav_queue *qh = ERR_PTR(-EINVAL);
switch (id) {
case KNAV_QUEUE_QPEND:
case KNAV_QUEUE_ACC:
case KNAV_QUEUE_GP:
qh = knav_queue_open_by_type(name, id, flags);
break;
default:
qh = knav_queue_open_by_id(name, id, flags);
break;
}
return qh;
}
EXPORT_SYMBOL_GPL(knav_queue_open);
/**
* knav_queue_close() - close a hardware queue handle
* @qh - handle to close
*/
void knav_queue_close(void *qhandle)
{
struct knav_queue *qh = qhandle;
struct knav_queue_inst *inst = qh->inst;
while (atomic_read(&qh->notifier_enabled) > 0)
knav_queue_disable_notifier(qh);
mutex_lock(&knav_dev_lock);
list_del_rcu(&qh->list);
mutex_unlock(&knav_dev_lock);
synchronize_rcu();
if (!knav_queue_is_busy(inst)) {
struct knav_range_info *range = inst->range;
if (range->ops && range->ops->close_queue)
range->ops->close_queue(range, inst);
}
devm_kfree(inst->kdev->dev, qh);
}
EXPORT_SYMBOL_GPL(knav_queue_close);
/**
* knav_queue_device_control() - Perform control operations on a queue
* @qh - queue handle
* @cmd - control commands
* @arg - command argument
*
* Returns 0 on success, errno otherwise.
*/
int knav_queue_device_control(void *qhandle, enum knav_queue_ctrl_cmd cmd,
unsigned long arg)
{
struct knav_queue *qh = qhandle;
struct knav_queue_notify_config *cfg;
int ret;
switch ((int)cmd) {
case KNAV_QUEUE_GET_ID:
ret = qh->inst->kdev->base_id + qh->inst->id;
break;
case KNAV_QUEUE_FLUSH:
ret = knav_queue_flush(qh);
break;
case KNAV_QUEUE_SET_NOTIFIER:
cfg = (void *)arg;
ret = knav_queue_set_notifier(qh, cfg);
break;
case KNAV_QUEUE_ENABLE_NOTIFY:
ret = knav_queue_enable_notifier(qh);
break;
case KNAV_QUEUE_DISABLE_NOTIFY:
ret = knav_queue_disable_notifier(qh);
break;
case KNAV_QUEUE_GET_COUNT:
ret = knav_queue_get_count(qh);
break;
default:
ret = -ENOTSUPP;
break;
}
return ret;
}
EXPORT_SYMBOL_GPL(knav_queue_device_control);
/**
* knav_queue_push() - push data (or descriptor) to the tail of a queue
* @qh - hardware queue handle
* @data - data to push
* @size - size of data to push
* @flags - can be used to pass additional information
*
* Returns 0 on success, errno otherwise.
*/
int knav_queue_push(void *qhandle, dma_addr_t dma,
unsigned size, unsigned flags)
{
struct knav_queue *qh = qhandle;
u32 val;
val = (u32)dma | ((size / 16) - 1);
writel_relaxed(val, &qh->reg_push[0].ptr_size_thresh);
atomic_inc(&qh->stats.pushes);
return 0;
}
/**
* knav_queue_pop() - pop data (or descriptor) from the head of a queue
* @qh - hardware queue handle
* @size - (optional) size of the data pop'ed.
*
* Returns a DMA address on success, 0 on failure.
*/
dma_addr_t knav_queue_pop(void *qhandle, unsigned *size)
{
struct knav_queue *qh = qhandle;
struct knav_queue_inst *inst = qh->inst;
dma_addr_t dma;
u32 val, idx;
/* are we accumulated? */
if (inst->descs) {
if (unlikely(atomic_dec_return(&inst->desc_count) < 0)) {
atomic_inc(&inst->desc_count);
return 0;
}
idx = atomic_inc_return(&inst->desc_head);
idx &= ACC_DESCS_MASK;
val = inst->descs[idx];
} else {
val = readl_relaxed(&qh->reg_pop[0].ptr_size_thresh);
if (unlikely(!val))
return 0;
}
dma = val & DESC_PTR_MASK;
if (size)
*size = ((val & DESC_SIZE_MASK) + 1) * 16;
atomic_inc(&qh->stats.pops);
return dma;
}
/* carve out descriptors and push into queue */
static void kdesc_fill_pool(struct knav_pool *pool)
{
struct knav_region *region;
int i;
region = pool->region;
pool->desc_size = region->desc_size;
for (i = 0; i < pool->num_desc; i++) {
int index = pool->region_offset + i;
dma_addr_t dma_addr;
unsigned dma_size;
dma_addr = region->dma_start + (region->desc_size * index);
dma_size = ALIGN(pool->desc_size, SMP_CACHE_BYTES);
dma_sync_single_for_device(pool->dev, dma_addr, dma_size,
DMA_TO_DEVICE);
knav_queue_push(pool->queue, dma_addr, dma_size, 0);
}
}
/* pop out descriptors and close the queue */
static void kdesc_empty_pool(struct knav_pool *pool)
{
dma_addr_t dma;
unsigned size;
void *desc;
int i;
if (!pool->queue)
return;
for (i = 0;; i++) {
dma = knav_queue_pop(pool->queue, &size);
if (!dma)
break;
desc = knav_pool_desc_dma_to_virt(pool, dma);
if (!desc) {
dev_dbg(pool->kdev->dev,
"couldn't unmap desc, continuing\n");
continue;
}
}
WARN_ON(i != pool->num_desc);
knav_queue_close(pool->queue);
}
/* Get the DMA address of a descriptor */
dma_addr_t knav_pool_desc_virt_to_dma(void *ph, void *virt)
{
struct knav_pool *pool = ph;
return pool->region->dma_start + (virt - pool->region->virt_start);
}
void *knav_pool_desc_dma_to_virt(void *ph, dma_addr_t dma)
{
struct knav_pool *pool = ph;
return pool->region->virt_start + (dma - pool->region->dma_start);
}
/**
* knav_pool_create() - Create a pool of descriptors
* @name - name to give the pool handle
* @num_desc - numbers of descriptors in the pool
* @region_id - QMSS region id from which the descriptors are to be
* allocated.
*
* Returns a pool handle on success.
* Use IS_ERR_OR_NULL() to identify error values on return.
*/
void *knav_pool_create(const char *name,
int num_desc, int region_id)
{
struct knav_region *reg_itr, *region = NULL;
struct knav_pool *pool, *pi;
struct list_head *node;
unsigned last_offset;
bool slot_found;
int ret;
if (!kdev->dev)
return ERR_PTR(-ENODEV);
pool = devm_kzalloc(kdev->dev, sizeof(*pool), GFP_KERNEL);
if (!pool) {
dev_err(kdev->dev, "out of memory allocating pool\n");
return ERR_PTR(-ENOMEM);
}
for_each_region(kdev, reg_itr) {
if (reg_itr->id != region_id)
continue;
region = reg_itr;
break;
}
if (!region) {
dev_err(kdev->dev, "region-id(%d) not found\n", region_id);
ret = -EINVAL;
goto err;
}
pool->queue = knav_queue_open(name, KNAV_QUEUE_GP, 0);
if (IS_ERR_OR_NULL(pool->queue)) {
dev_err(kdev->dev,
"failed to open queue for pool(%s), error %ld\n",
name, PTR_ERR(pool->queue));
ret = PTR_ERR(pool->queue);
goto err;
}
pool->name = kstrndup(name, KNAV_NAME_SIZE, GFP_KERNEL);
pool->kdev = kdev;
pool->dev = kdev->dev;
mutex_lock(&knav_dev_lock);
if (num_desc > (region->num_desc - region->used_desc)) {
dev_err(kdev->dev, "out of descs in region(%d) for pool(%s)\n",
region_id, name);
ret = -ENOMEM;
goto err;
}
/* Region maintains a sorted (by region offset) list of pools
* use the first free slot which is large enough to accomodate
* the request
*/
last_offset = 0;
slot_found = false;
node = &region->pools;
list_for_each_entry(pi, &region->pools, region_inst) {
if ((pi->region_offset - last_offset) >= num_desc) {
slot_found = true;
break;
}
last_offset = pi->region_offset + pi->num_desc;
}
node = &pi->region_inst;
if (slot_found) {
pool->region = region;
pool->num_desc = num_desc;
pool->region_offset = last_offset;
region->used_desc += num_desc;
list_add_tail(&pool->list, &kdev->pools);
list_add_tail(&pool->region_inst, node);
} else {
dev_err(kdev->dev, "pool(%s) create failed: fragmented desc pool in region(%d)\n",
name, region_id);
ret = -ENOMEM;
goto err;
}
mutex_unlock(&knav_dev_lock);
kdesc_fill_pool(pool);
return pool;
err:
mutex_unlock(&knav_dev_lock);
kfree(pool->name);
devm_kfree(kdev->dev, pool);
return ERR_PTR(ret);
}
EXPORT_SYMBOL_GPL(knav_pool_create);
/**
* knav_pool_destroy() - Free a pool of descriptors
* @pool - pool handle
*/
void knav_pool_destroy(void *ph)
{
struct knav_pool *pool = ph;
if (!pool)
return;
if (!pool->region)
return;
kdesc_empty_pool(pool);
mutex_lock(&knav_dev_lock);
pool->region->used_desc -= pool->num_desc;
list_del(&pool->region_inst);
list_del(&pool->list);
mutex_unlock(&knav_dev_lock);
kfree(pool->name);
devm_kfree(kdev->dev, pool);
}
EXPORT_SYMBOL_GPL(knav_pool_destroy);
/**
* knav_pool_desc_get() - Get a descriptor from the pool
* @pool - pool handle
*
* Returns descriptor from the pool.
*/
void *knav_pool_desc_get(void *ph)
{
struct knav_pool *pool = ph;
dma_addr_t dma;
unsigned size;
void *data;
dma = knav_queue_pop(pool->queue, &size);
if (unlikely(!dma))
return ERR_PTR(-ENOMEM);
data = knav_pool_desc_dma_to_virt(pool, dma);
return data;
}
/**
* knav_pool_desc_put() - return a descriptor to the pool
* @pool - pool handle
*/
void knav_pool_desc_put(void *ph, void *desc)
{
struct knav_pool *pool = ph;
dma_addr_t dma;
dma = knav_pool_desc_virt_to_dma(pool, desc);
knav_queue_push(pool->queue, dma, pool->region->desc_size, 0);
}
/**
* knav_pool_desc_map() - Map descriptor for DMA transfer
* @pool - pool handle
* @desc - address of descriptor to map
* @size - size of descriptor to map
* @dma - DMA address return pointer
* @dma_sz - adjusted return pointer
*
* Returns 0 on success, errno otherwise.
*/
int knav_pool_desc_map(void *ph, void *desc, unsigned size,
dma_addr_t *dma, unsigned *dma_sz)
{
struct knav_pool *pool = ph;
*dma = knav_pool_desc_virt_to_dma(pool, desc);
size = min(size, pool->region->desc_size);
size = ALIGN(size, SMP_CACHE_BYTES);
*dma_sz = size;
dma_sync_single_for_device(pool->dev, *dma, size, DMA_TO_DEVICE);
/* Ensure the descriptor reaches to the memory */
__iowmb();
return 0;
}
/**
* knav_pool_desc_unmap() - Unmap descriptor after DMA transfer
* @pool - pool handle
* @dma - DMA address of descriptor to unmap
* @dma_sz - size of descriptor to unmap
*
* Returns descriptor address on success, Use IS_ERR_OR_NULL() to identify
* error values on return.
*/
void *knav_pool_desc_unmap(void *ph, dma_addr_t dma, unsigned dma_sz)
{
struct knav_pool *pool = ph;
unsigned desc_sz;
void *desc;
desc_sz = min(dma_sz, pool->region->desc_size);
desc = knav_pool_desc_dma_to_virt(pool, dma);
dma_sync_single_for_cpu(pool->dev, dma, desc_sz, DMA_FROM_DEVICE);
prefetch(desc);
return desc;
}
/**
* knav_pool_count() - Get the number of descriptors in pool.
* @pool - pool handle
* Returns number of elements in the pool.
*/
int knav_pool_count(void *ph)
{
struct knav_pool *pool = ph;
return knav_queue_get_count(pool->queue);
}
static void knav_queue_setup_region(struct knav_device *kdev,
struct knav_region *region)
{
unsigned hw_num_desc, hw_desc_size, size;
struct knav_reg_region __iomem *regs;
struct knav_qmgr_info *qmgr;
struct knav_pool *pool;
int id = region->id;
struct page *page;
/* unused region? */
if (!region->num_desc) {
dev_warn(kdev->dev, "unused region %s\n", region->name);
return;
}
/* get hardware descriptor value */
hw_num_desc = ilog2(region->num_desc - 1) + 1;
/* did we force fit ourselves into nothingness? */
if (region->num_desc < 32) {
region->num_desc = 0;
dev_warn(kdev->dev, "too few descriptors in region %s\n",
region->name);
return;
}
size = region->num_desc * region->desc_size;
region->virt_start = alloc_pages_exact(size, GFP_KERNEL | GFP_DMA |
GFP_DMA32);
if (!region->virt_start) {
region->num_desc = 0;
dev_err(kdev->dev, "memory alloc failed for region %s\n",
region->name);
return;
}
region->virt_end = region->virt_start + size;
page = virt_to_page(region->virt_start);
region->dma_start = dma_map_page(kdev->dev, page, 0, size,
DMA_BIDIRECTIONAL);
if (dma_mapping_error(kdev->dev, region->dma_start)) {
dev_err(kdev->dev, "dma map failed for region %s\n",
region->name);
goto fail;
}
region->dma_end = region->dma_start + size;
pool = devm_kzalloc(kdev->dev, sizeof(*pool), GFP_KERNEL);
if (!pool) {
dev_err(kdev->dev, "out of memory allocating dummy pool\n");
goto fail;
}
pool->num_desc = 0;
pool->region_offset = region->num_desc;
list_add(&pool->region_inst, &region->pools);
dev_dbg(kdev->dev,
"region %s (%d): size:%d, link:%d@%d, phys:%08x-%08x, virt:%p-%p\n",
region->name, id, region->desc_size, region->num_desc,
region->link_index, region->dma_start, region->dma_end,
region->virt_start, region->virt_end);
hw_desc_size = (region->desc_size / 16) - 1;
hw_num_desc -= 5;
for_each_qmgr(kdev, qmgr) {
regs = qmgr->reg_region + id;
writel_relaxed(region->dma_start, &regs->base);
writel_relaxed(region->link_index, &regs->start_index);
writel_relaxed(hw_desc_size << 16 | hw_num_desc,
&regs->size_count);
}
return;
fail:
if (region->dma_start)
dma_unmap_page(kdev->dev, region->dma_start, size,
DMA_BIDIRECTIONAL);
if (region->virt_start)
free_pages_exact(region->virt_start, size);
region->num_desc = 0;
return;
}
static const char *knav_queue_find_name(struct device_node *node)
{
const char *name;
if (of_property_read_string(node, "label", &name) < 0)
name = node->name;
if (!name)
name = "unknown";
return name;
}
static int knav_queue_setup_regions(struct knav_device *kdev,
struct device_node *regions)
{
struct device *dev = kdev->dev;
struct knav_region *region;
struct device_node *child;
u32 temp[2];
int ret;
for_each_child_of_node(regions, child) {
region = devm_kzalloc(dev, sizeof(*region), GFP_KERNEL);
if (!region) {
dev_err(dev, "out of memory allocating region\n");
return -ENOMEM;
}
region->name = knav_queue_find_name(child);
of_property_read_u32(child, "id", &region->id);
ret = of_property_read_u32_array(child, "region-spec", temp, 2);
if (!ret) {
region->num_desc = temp[0];
region->desc_size = temp[1];
} else {
dev_err(dev, "invalid region info %s\n", region->name);
devm_kfree(dev, region);
continue;
}
if (!of_get_property(child, "link-index", NULL)) {
dev_err(dev, "No link info for %s\n", region->name);
devm_kfree(dev, region);
continue;
}
ret = of_property_read_u32(child, "link-index",
&region->link_index);
if (ret) {
dev_err(dev, "link index not found for %s\n",
region->name);
devm_kfree(dev, region);
continue;
}
INIT_LIST_HEAD(&region->pools);
list_add_tail(&region->list, &kdev->regions);
}
if (list_empty(&kdev->regions)) {
dev_err(dev, "no valid region information found\n");
return -ENODEV;
}
/* Next, we run through the regions and set things up */
for_each_region(kdev, region)
knav_queue_setup_region(kdev, region);
return 0;
}
static int knav_get_link_ram(struct knav_device *kdev,
const char *name,
struct knav_link_ram_block *block)
{
struct platform_device *pdev = to_platform_device(kdev->dev);
struct device_node *node = pdev->dev.of_node;
u32 temp[2];
/*
* Note: link ram resources are specified in "entry" sized units. In
* reality, although entries are ~40bits in hardware, we treat them as
* 64-bit entities here.
*
* For example, to specify the internal link ram for Keystone-I class
* devices, we would set the linkram0 resource to 0x80000-0x83fff.
*
* This gets a bit weird when other link rams are used. For example,
* if the range specified is 0x0c000000-0x0c003fff (i.e., 16K entries
* in MSMC SRAM), the actual memory used is 0x0c000000-0x0c020000,
* which accounts for 64-bits per entry, for 16K entries.
*/
if (!of_property_read_u32_array(node, name , temp, 2)) {
if (temp[0]) {
/*
* queue_base specified => using internal or onchip
* link ram WARNING - we do not "reserve" this block
*/
block->phys = (dma_addr_t)temp[0];
block->virt = NULL;
block->size = temp[1];
} else {
block->size = temp[1];
/* queue_base not specific => allocate requested size */
block->virt = dmam_alloc_coherent(kdev->dev,
8 * block->size, &block->phys,
GFP_KERNEL);
if (!block->virt) {
dev_err(kdev->dev, "failed to alloc linkram\n");
return -ENOMEM;
}
}
} else {
return -ENODEV;
}
return 0;
}
static int knav_queue_setup_link_ram(struct knav_device *kdev)
{
struct knav_link_ram_block *block;
struct knav_qmgr_info *qmgr;
for_each_qmgr(kdev, qmgr) {
block = &kdev->link_rams[0];
dev_dbg(kdev->dev, "linkram0: phys:%x, virt:%p, size:%x\n",
block->phys, block->virt, block->size);
writel_relaxed(block->phys, &qmgr->reg_config->link_ram_base0);
writel_relaxed(block->size, &qmgr->reg_config->link_ram_size0);
block++;
if (!block->size)
return 0;
dev_dbg(kdev->dev, "linkram1: phys:%x, virt:%p, size:%x\n",
block->phys, block->virt, block->size);
writel_relaxed(block->phys, &qmgr->reg_config->link_ram_base1);
}
return 0;
}
static int knav_setup_queue_range(struct knav_device *kdev,
struct device_node *node)
{
struct device *dev = kdev->dev;
struct knav_range_info *range;
struct knav_qmgr_info *qmgr;
u32 temp[2], start, end, id, index;
int ret, i;
range = devm_kzalloc(dev, sizeof(*range), GFP_KERNEL);
if (!range) {
dev_err(dev, "out of memory allocating range\n");
return -ENOMEM;
}
range->kdev = kdev;
range->name = knav_queue_find_name(node);
ret = of_property_read_u32_array(node, "qrange", temp, 2);
if (!ret) {
range->queue_base = temp[0] - kdev->base_id;
range->num_queues = temp[1];
} else {
dev_err(dev, "invalid queue range %s\n", range->name);
devm_kfree(dev, range);
return -EINVAL;
}
for (i = 0; i < RANGE_MAX_IRQS; i++) {
struct of_phandle_args oirq;
if (of_irq_parse_one(node, i, &oirq))
break;
range->irqs[i].irq = irq_create_of_mapping(&oirq);
if (range->irqs[i].irq == IRQ_NONE)
break;
range->num_irqs++;
if (oirq.args_count == 3)
range->irqs[i].cpu_map =
(oirq.args[2] & 0x0000ff00) >> 8;
}
range->num_irqs = min(range->num_irqs, range->num_queues);
if (range->num_irqs)
range->flags |= RANGE_HAS_IRQ;
if (of_get_property(node, "qalloc-by-id", NULL))
range->flags |= RANGE_RESERVED;
if (of_get_property(node, "accumulator", NULL)) {
ret = knav_init_acc_range(kdev, node, range);
if (ret < 0) {
devm_kfree(dev, range);
return ret;
}
} else {
range->ops = &knav_gp_range_ops;
}
/* set threshold to 1, and flush out the queues */
for_each_qmgr(kdev, qmgr) {
start = max(qmgr->start_queue, range->queue_base);
end = min(qmgr->start_queue + qmgr->num_queues,
range->queue_base + range->num_queues);
for (id = start; id < end; id++) {
index = id - qmgr->start_queue;
writel_relaxed(THRESH_GTE | 1,
&qmgr->reg_peek[index].ptr_size_thresh);
writel_relaxed(0,
&qmgr->reg_push[index].ptr_size_thresh);
}
}
list_add_tail(&range->list, &kdev->queue_ranges);
dev_dbg(dev, "added range %s: %d-%d, %d irqs%s%s%s\n",
range->name, range->queue_base,
range->queue_base + range->num_queues - 1,
range->num_irqs,
(range->flags & RANGE_HAS_IRQ) ? ", has irq" : "",
(range->flags & RANGE_RESERVED) ? ", reserved" : "",
(range->flags & RANGE_HAS_ACCUMULATOR) ? ", acc" : "");
kdev->num_queues_in_use += range->num_queues;
return 0;
}
static int knav_setup_queue_pools(struct knav_device *kdev,
struct device_node *queue_pools)
{
struct device_node *type, *range;
int ret;
for_each_child_of_node(queue_pools, type) {
for_each_child_of_node(type, range) {
ret = knav_setup_queue_range(kdev, range);
/* return value ignored, we init the rest... */
}
}
/* ... and barf if they all failed! */
if (list_empty(&kdev->queue_ranges)) {
dev_err(kdev->dev, "no valid queue range found\n");
return -ENODEV;
}
return 0;
}
static void knav_free_queue_range(struct knav_device *kdev,
struct knav_range_info *range)
{
if (range->ops && range->ops->free_range)
range->ops->free_range(range);
list_del(&range->list);
devm_kfree(kdev->dev, range);
}
static void knav_free_queue_ranges(struct knav_device *kdev)
{
struct knav_range_info *range;
for (;;) {
range = first_queue_range(kdev);
if (!range)
break;
knav_free_queue_range(kdev, range);
}
}
static void knav_queue_free_regions(struct knav_device *kdev)
{
struct knav_region *region;
struct knav_pool *pool;
unsigned size;
for (;;) {
region = first_region(kdev);
if (!region)
break;
list_for_each_entry(pool, &region->pools, region_inst)
knav_pool_destroy(pool);
size = region->virt_end - region->virt_start;
if (size)
free_pages_exact(region->virt_start, size);
list_del(&region->list);
devm_kfree(kdev->dev, region);
}
}
static void __iomem *knav_queue_map_reg(struct knav_device *kdev,
struct device_node *node, int index)
{
struct resource res;
void __iomem *regs;
int ret;
ret = of_address_to_resource(node, index, &res);
if (ret) {
dev_err(kdev->dev, "Can't translate of node(%s) address for index(%d)\n",
node->name, index);
return ERR_PTR(ret);
}
regs = devm_ioremap_resource(kdev->dev, &res);
if (IS_ERR(regs))
dev_err(kdev->dev, "Failed to map register base for index(%d) node(%s)\n",
index, node->name);
return regs;
}
static int knav_queue_init_qmgrs(struct knav_device *kdev,
struct device_node *qmgrs)
{
struct device *dev = kdev->dev;
struct knav_qmgr_info *qmgr;
struct device_node *child;
u32 temp[2];
int ret;
for_each_child_of_node(qmgrs, child) {
qmgr = devm_kzalloc(dev, sizeof(*qmgr), GFP_KERNEL);
if (!qmgr) {
dev_err(dev, "out of memory allocating qmgr\n");
return -ENOMEM;
}
ret = of_property_read_u32_array(child, "managed-queues",
temp, 2);
if (!ret) {
qmgr->start_queue = temp[0];
qmgr->num_queues = temp[1];
} else {
dev_err(dev, "invalid qmgr queue range\n");
devm_kfree(dev, qmgr);
continue;
}
dev_info(dev, "qmgr start queue %d, number of queues %d\n",
qmgr->start_queue, qmgr->num_queues);
qmgr->reg_peek =
knav_queue_map_reg(kdev, child,
KNAV_QUEUE_PEEK_REG_INDEX);
qmgr->reg_status =
knav_queue_map_reg(kdev, child,
KNAV_QUEUE_STATUS_REG_INDEX);
qmgr->reg_config =
knav_queue_map_reg(kdev, child,
KNAV_QUEUE_CONFIG_REG_INDEX);
qmgr->reg_region =
knav_queue_map_reg(kdev, child,
KNAV_QUEUE_REGION_REG_INDEX);
qmgr->reg_push =
knav_queue_map_reg(kdev, child,
KNAV_QUEUE_PUSH_REG_INDEX);
qmgr->reg_pop =
knav_queue_map_reg(kdev, child,
KNAV_QUEUE_POP_REG_INDEX);
if (IS_ERR(qmgr->reg_peek) || IS_ERR(qmgr->reg_status) ||
IS_ERR(qmgr->reg_config) || IS_ERR(qmgr->reg_region) ||
IS_ERR(qmgr->reg_push) || IS_ERR(qmgr->reg_pop)) {
dev_err(dev, "failed to map qmgr regs\n");
if (!IS_ERR(qmgr->reg_peek))
devm_iounmap(dev, qmgr->reg_peek);
if (!IS_ERR(qmgr->reg_status))
devm_iounmap(dev, qmgr->reg_status);
if (!IS_ERR(qmgr->reg_config))
devm_iounmap(dev, qmgr->reg_config);
if (!IS_ERR(qmgr->reg_region))
devm_iounmap(dev, qmgr->reg_region);
if (!IS_ERR(qmgr->reg_push))
devm_iounmap(dev, qmgr->reg_push);
if (!IS_ERR(qmgr->reg_pop))
devm_iounmap(dev, qmgr->reg_pop);
devm_kfree(dev, qmgr);
continue;
}
list_add_tail(&qmgr->list, &kdev->qmgrs);
dev_info(dev, "added qmgr start queue %d, num of queues %d, reg_peek %p, reg_status %p, reg_config %p, reg_region %p, reg_push %p, reg_pop %p\n",
qmgr->start_queue, qmgr->num_queues,
qmgr->reg_peek, qmgr->reg_status,
qmgr->reg_config, qmgr->reg_region,
qmgr->reg_push, qmgr->reg_pop);
}
return 0;
}
static int knav_queue_init_pdsps(struct knav_device *kdev,
struct device_node *pdsps)
{
struct device *dev = kdev->dev;
struct knav_pdsp_info *pdsp;
struct device_node *child;
int ret;
for_each_child_of_node(pdsps, child) {
pdsp = devm_kzalloc(dev, sizeof(*pdsp), GFP_KERNEL);
if (!pdsp) {
dev_err(dev, "out of memory allocating pdsp\n");
return -ENOMEM;
}
pdsp->name = knav_queue_find_name(child);
ret = of_property_read_string(child, "firmware",
&pdsp->firmware);
if (ret < 0 || !pdsp->firmware) {
dev_err(dev, "unknown firmware for pdsp %s\n",
pdsp->name);
devm_kfree(dev, pdsp);
continue;
}
dev_dbg(dev, "pdsp name %s fw name :%s\n", pdsp->name,
pdsp->firmware);
pdsp->iram =
knav_queue_map_reg(kdev, child,
KNAV_QUEUE_PDSP_IRAM_REG_INDEX);
pdsp->regs =
knav_queue_map_reg(kdev, child,
KNAV_QUEUE_PDSP_REGS_REG_INDEX);
pdsp->intd =
knav_queue_map_reg(kdev, child,
KNAV_QUEUE_PDSP_INTD_REG_INDEX);
pdsp->command =
knav_queue_map_reg(kdev, child,
KNAV_QUEUE_PDSP_CMD_REG_INDEX);
if (IS_ERR(pdsp->command) || IS_ERR(pdsp->iram) ||
IS_ERR(pdsp->regs) || IS_ERR(pdsp->intd)) {
dev_err(dev, "failed to map pdsp %s regs\n",
pdsp->name);
if (!IS_ERR(pdsp->command))
devm_iounmap(dev, pdsp->command);
if (!IS_ERR(pdsp->iram))
devm_iounmap(dev, pdsp->iram);
if (!IS_ERR(pdsp->regs))
devm_iounmap(dev, pdsp->regs);
if (!IS_ERR(pdsp->intd))
devm_iounmap(dev, pdsp->intd);
devm_kfree(dev, pdsp);
continue;
}
of_property_read_u32(child, "id", &pdsp->id);
list_add_tail(&pdsp->list, &kdev->pdsps);
dev_dbg(dev, "added pdsp %s: command %p, iram %p, regs %p, intd %p, firmware %s\n",
pdsp->name, pdsp->command, pdsp->iram, pdsp->regs,
pdsp->intd, pdsp->firmware);
}
return 0;
}
static int knav_queue_stop_pdsp(struct knav_device *kdev,
struct knav_pdsp_info *pdsp)
{
u32 val, timeout = 1000;
int ret;
val = readl_relaxed(&pdsp->regs->control) & ~PDSP_CTRL_ENABLE;
writel_relaxed(val, &pdsp->regs->control);
ret = knav_queue_pdsp_wait(&pdsp->regs->control, timeout,
PDSP_CTRL_RUNNING);
if (ret < 0) {
dev_err(kdev->dev, "timed out on pdsp %s stop\n", pdsp->name);
return ret;
}
return 0;
}
static int knav_queue_load_pdsp(struct knav_device *kdev,
struct knav_pdsp_info *pdsp)
{
int i, ret, fwlen;
const struct firmware *fw;
u32 *fwdata;
ret = request_firmware(&fw, pdsp->firmware, kdev->dev);
if (ret) {
dev_err(kdev->dev, "failed to get firmware %s for pdsp %s\n",
pdsp->firmware, pdsp->name);
return ret;
}
writel_relaxed(pdsp->id + 1, pdsp->command + 0x18);
/* download the firmware */
fwdata = (u32 *)fw->data;
fwlen = (fw->size + sizeof(u32) - 1) / sizeof(u32);
for (i = 0; i < fwlen; i++)
writel_relaxed(be32_to_cpu(fwdata[i]), pdsp->iram + i);
release_firmware(fw);
return 0;
}
static int knav_queue_start_pdsp(struct knav_device *kdev,
struct knav_pdsp_info *pdsp)
{
u32 val, timeout = 1000;
int ret;
/* write a command for sync */
writel_relaxed(0xffffffff, pdsp->command);
while (readl_relaxed(pdsp->command) != 0xffffffff)
cpu_relax();
/* soft reset the PDSP */
val = readl_relaxed(&pdsp->regs->control);
val &= ~(PDSP_CTRL_PC_MASK | PDSP_CTRL_SOFT_RESET);
writel_relaxed(val, &pdsp->regs->control);
/* enable pdsp */
val = readl_relaxed(&pdsp->regs->control) | PDSP_CTRL_ENABLE;
writel_relaxed(val, &pdsp->regs->control);
/* wait for command register to clear */
ret = knav_queue_pdsp_wait(pdsp->command, timeout, 0);
if (ret < 0) {
dev_err(kdev->dev,
"timed out on pdsp %s command register wait\n",
pdsp->name);
return ret;
}
return 0;
}
static void knav_queue_stop_pdsps(struct knav_device *kdev)
{
struct knav_pdsp_info *pdsp;
/* disable all pdsps */
for_each_pdsp(kdev, pdsp)
knav_queue_stop_pdsp(kdev, pdsp);
}
static int knav_queue_start_pdsps(struct knav_device *kdev)
{
struct knav_pdsp_info *pdsp;
int ret;
knav_queue_stop_pdsps(kdev);
/* now load them all */
for_each_pdsp(kdev, pdsp) {
ret = knav_queue_load_pdsp(kdev, pdsp);
if (ret < 0)
return ret;
}
for_each_pdsp(kdev, pdsp) {
ret = knav_queue_start_pdsp(kdev, pdsp);
WARN_ON(ret);
}
return 0;
}
static inline struct knav_qmgr_info *knav_find_qmgr(unsigned id)
{
struct knav_qmgr_info *qmgr;
for_each_qmgr(kdev, qmgr) {
if ((id >= qmgr->start_queue) &&
(id < qmgr->start_queue + qmgr->num_queues))
return qmgr;
}
return NULL;
}
static int knav_queue_init_queue(struct knav_device *kdev,
struct knav_range_info *range,
struct knav_queue_inst *inst,
unsigned id)
{
char irq_name[KNAV_NAME_SIZE];
inst->qmgr = knav_find_qmgr(id);
if (!inst->qmgr)
return -1;
INIT_LIST_HEAD(&inst->handles);
inst->kdev = kdev;
inst->range = range;
inst->irq_num = -1;
inst->id = id;
scnprintf(irq_name, sizeof(irq_name), "hwqueue-%d", id);
inst->irq_name = kstrndup(irq_name, sizeof(irq_name), GFP_KERNEL);
if (range->ops && range->ops->init_queue)
return range->ops->init_queue(range, inst);
else
return 0;
}
static int knav_queue_init_queues(struct knav_device *kdev)
{
struct knav_range_info *range;
int size, id, base_idx;
int idx = 0, ret = 0;
/* how much do we need for instance data? */
size = sizeof(struct knav_queue_inst);
/* round this up to a power of 2, keep the index to instance
* arithmetic fast.
* */
kdev->inst_shift = order_base_2(size);
size = (1 << kdev->inst_shift) * kdev->num_queues_in_use;
kdev->instances = devm_kzalloc(kdev->dev, size, GFP_KERNEL);
if (!kdev->instances)
return -1;
for_each_queue_range(kdev, range) {
if (range->ops && range->ops->init_range)
range->ops->init_range(range);
base_idx = idx;
for (id = range->queue_base;
id < range->queue_base + range->num_queues; id++, idx++) {
ret = knav_queue_init_queue(kdev, range,
knav_queue_idx_to_inst(kdev, idx), id);
if (ret < 0)
return ret;
}
range->queue_base_inst =
knav_queue_idx_to_inst(kdev, base_idx);
}
return 0;
}
static int knav_queue_probe(struct platform_device *pdev)
{
struct device_node *node = pdev->dev.of_node;
struct device_node *qmgrs, *queue_pools, *regions, *pdsps;
struct device *dev = &pdev->dev;
u32 temp[2];
int ret;
if (!node) {
dev_err(dev, "device tree info unavailable\n");
return -ENODEV;
}
kdev = devm_kzalloc(dev, sizeof(struct knav_device), GFP_KERNEL);
if (!kdev) {
dev_err(dev, "memory allocation failed\n");
return -ENOMEM;
}
platform_set_drvdata(pdev, kdev);
kdev->dev = dev;
INIT_LIST_HEAD(&kdev->queue_ranges);
INIT_LIST_HEAD(&kdev->qmgrs);
INIT_LIST_HEAD(&kdev->pools);
INIT_LIST_HEAD(&kdev->regions);
INIT_LIST_HEAD(&kdev->pdsps);
pm_runtime_enable(&pdev->dev);
ret = pm_runtime_get_sync(&pdev->dev);
if (ret < 0) {
dev_err(dev, "Failed to enable QMSS\n");
return ret;
}
if (of_property_read_u32_array(node, "queue-range", temp, 2)) {
dev_err(dev, "queue-range not specified\n");
ret = -ENODEV;
goto err;
}
kdev->base_id = temp[0];
kdev->num_queues = temp[1];
/* Initialize queue managers using device tree configuration */
qmgrs = of_get_child_by_name(node, "qmgrs");
if (!qmgrs) {
dev_err(dev, "queue manager info not specified\n");
ret = -ENODEV;
goto err;
}
ret = knav_queue_init_qmgrs(kdev, qmgrs);
of_node_put(qmgrs);
if (ret)
goto err;
/* get pdsp configuration values from device tree */
pdsps = of_get_child_by_name(node, "pdsps");
if (pdsps) {
ret = knav_queue_init_pdsps(kdev, pdsps);
if (ret)
goto err;
ret = knav_queue_start_pdsps(kdev);
if (ret)
goto err;
}
of_node_put(pdsps);
/* get usable queue range values from device tree */
queue_pools = of_get_child_by_name(node, "queue-pools");
if (!queue_pools) {
dev_err(dev, "queue-pools not specified\n");
ret = -ENODEV;
goto err;
}
ret = knav_setup_queue_pools(kdev, queue_pools);
of_node_put(queue_pools);
if (ret)
goto err;
ret = knav_get_link_ram(kdev, "linkram0", &kdev->link_rams[0]);
if (ret) {
dev_err(kdev->dev, "could not setup linking ram\n");
goto err;
}
ret = knav_get_link_ram(kdev, "linkram1", &kdev->link_rams[1]);
if (ret) {
/*
* nothing really, we have one linking ram already, so we just
* live within our means
*/
}
ret = knav_queue_setup_link_ram(kdev);
if (ret)
goto err;
regions = of_get_child_by_name(node, "descriptor-regions");
if (!regions) {
dev_err(dev, "descriptor-regions not specified\n");
goto err;
}
ret = knav_queue_setup_regions(kdev, regions);
of_node_put(regions);
if (ret)
goto err;
ret = knav_queue_init_queues(kdev);
if (ret < 0) {
dev_err(dev, "hwqueue initialization failed\n");
goto err;
}
debugfs_create_file("qmss", S_IFREG | S_IRUGO, NULL, NULL,
&knav_queue_debug_ops);
return 0;
err:
knav_queue_stop_pdsps(kdev);
knav_queue_free_regions(kdev);
knav_free_queue_ranges(kdev);
pm_runtime_put_sync(&pdev->dev);
pm_runtime_disable(&pdev->dev);
return ret;
}
static int knav_queue_remove(struct platform_device *pdev)
{
/* TODO: Free resources */
pm_runtime_put_sync(&pdev->dev);
pm_runtime_disable(&pdev->dev);
return 0;
}
/* Match table for of_platform binding */
static struct of_device_id keystone_qmss_of_match[] = {
{ .compatible = "ti,keystone-navigator-qmss", },
{},
};
MODULE_DEVICE_TABLE(of, keystone_qmss_of_match);
static struct platform_driver keystone_qmss_driver = {
.probe = knav_queue_probe,
.remove = knav_queue_remove,
.driver = {
.name = "keystone-navigator-qmss",
.owner = THIS_MODULE,
.of_match_table = keystone_qmss_of_match,
},
};
module_platform_driver(keystone_qmss_driver);
MODULE_LICENSE("GPL v2");
MODULE_DESCRIPTION("TI QMSS driver for Keystone SOCs");
MODULE_AUTHOR("Sandeep Nair <sandeep_n@ti.com>");
MODULE_AUTHOR("Santosh Shilimkar <santosh.shilimkar@ti.com>");