2019-10-03 02:42:41 +07:00
|
|
|
// SPDX-License-Identifier: GPL-2.0
|
|
|
|
/*
|
|
|
|
* I/O Address Space ID allocator. There is one global IOASID space, split into
|
|
|
|
* subsets. Users create a subset with DECLARE_IOASID_SET, then allocate and
|
|
|
|
* free IOASIDs with ioasid_alloc and ioasid_free.
|
|
|
|
*/
|
|
|
|
#include <linux/ioasid.h>
|
|
|
|
#include <linux/module.h>
|
|
|
|
#include <linux/slab.h>
|
|
|
|
#include <linux/spinlock.h>
|
|
|
|
#include <linux/xarray.h>
|
|
|
|
|
|
|
|
struct ioasid_data {
|
|
|
|
ioasid_t id;
|
|
|
|
struct ioasid_set *set;
|
|
|
|
void *private;
|
|
|
|
struct rcu_head rcu;
|
|
|
|
};
|
|
|
|
|
2019-10-03 02:42:42 +07:00
|
|
|
/*
|
|
|
|
* struct ioasid_allocator_data - Internal data structure to hold information
|
|
|
|
* about an allocator. There are two types of allocators:
|
|
|
|
*
|
|
|
|
* - Default allocator always has its own XArray to track the IOASIDs allocated.
|
|
|
|
* - Custom allocators may share allocation helpers with different private data.
|
|
|
|
* Custom allocators that share the same helper functions also share the same
|
|
|
|
* XArray.
|
|
|
|
* Rules:
|
|
|
|
* 1. Default allocator is always available, not dynamically registered. This is
|
|
|
|
* to prevent race conditions with early boot code that want to register
|
|
|
|
* custom allocators or allocate IOASIDs.
|
|
|
|
* 2. Custom allocators take precedence over the default allocator.
|
|
|
|
* 3. When all custom allocators sharing the same helper functions are
|
|
|
|
* unregistered (e.g. due to hotplug), all outstanding IOASIDs must be
|
|
|
|
* freed. Otherwise, outstanding IOASIDs will be lost and orphaned.
|
|
|
|
* 4. When switching between custom allocators sharing the same helper
|
|
|
|
* functions, outstanding IOASIDs are preserved.
|
|
|
|
* 5. When switching between custom allocator and default allocator, all IOASIDs
|
|
|
|
* must be freed to ensure unadulterated space for the new allocator.
|
|
|
|
*
|
|
|
|
* @ops: allocator helper functions and its data
|
|
|
|
* @list: registered custom allocators
|
|
|
|
* @slist: allocators share the same ops but different data
|
|
|
|
* @flags: attributes of the allocator
|
|
|
|
* @xa: xarray holds the IOASID space
|
|
|
|
* @rcu: used for kfree_rcu when unregistering allocator
|
|
|
|
*/
|
|
|
|
struct ioasid_allocator_data {
|
|
|
|
struct ioasid_allocator_ops *ops;
|
|
|
|
struct list_head list;
|
|
|
|
struct list_head slist;
|
|
|
|
#define IOASID_ALLOCATOR_CUSTOM BIT(0) /* Needs framework to track results */
|
|
|
|
unsigned long flags;
|
|
|
|
struct xarray xa;
|
|
|
|
struct rcu_head rcu;
|
|
|
|
};
|
|
|
|
|
|
|
|
static DEFINE_SPINLOCK(ioasid_allocator_lock);
|
|
|
|
static LIST_HEAD(allocators_list);
|
|
|
|
|
|
|
|
static ioasid_t default_alloc(ioasid_t min, ioasid_t max, void *opaque);
|
|
|
|
static void default_free(ioasid_t ioasid, void *opaque);
|
|
|
|
|
|
|
|
static struct ioasid_allocator_ops default_ops = {
|
|
|
|
.alloc = default_alloc,
|
|
|
|
.free = default_free,
|
|
|
|
};
|
|
|
|
|
|
|
|
static struct ioasid_allocator_data default_allocator = {
|
|
|
|
.ops = &default_ops,
|
|
|
|
.flags = 0,
|
|
|
|
.xa = XARRAY_INIT(ioasid_xa, XA_FLAGS_ALLOC),
|
|
|
|
};
|
|
|
|
|
|
|
|
static struct ioasid_allocator_data *active_allocator = &default_allocator;
|
|
|
|
|
|
|
|
static ioasid_t default_alloc(ioasid_t min, ioasid_t max, void *opaque)
|
|
|
|
{
|
|
|
|
ioasid_t id;
|
|
|
|
|
|
|
|
if (xa_alloc(&default_allocator.xa, &id, opaque, XA_LIMIT(min, max), GFP_ATOMIC)) {
|
|
|
|
pr_err("Failed to alloc ioasid from %d to %d\n", min, max);
|
|
|
|
return INVALID_IOASID;
|
|
|
|
}
|
|
|
|
|
|
|
|
return id;
|
|
|
|
}
|
|
|
|
|
|
|
|
static void default_free(ioasid_t ioasid, void *opaque)
|
|
|
|
{
|
|
|
|
struct ioasid_data *ioasid_data;
|
|
|
|
|
|
|
|
ioasid_data = xa_erase(&default_allocator.xa, ioasid);
|
|
|
|
kfree_rcu(ioasid_data, rcu);
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Allocate and initialize a new custom allocator with its helper functions */
|
|
|
|
static struct ioasid_allocator_data *ioasid_alloc_allocator(struct ioasid_allocator_ops *ops)
|
|
|
|
{
|
|
|
|
struct ioasid_allocator_data *ia_data;
|
|
|
|
|
|
|
|
ia_data = kzalloc(sizeof(*ia_data), GFP_ATOMIC);
|
|
|
|
if (!ia_data)
|
|
|
|
return NULL;
|
|
|
|
|
|
|
|
xa_init_flags(&ia_data->xa, XA_FLAGS_ALLOC);
|
|
|
|
INIT_LIST_HEAD(&ia_data->slist);
|
|
|
|
ia_data->flags |= IOASID_ALLOCATOR_CUSTOM;
|
|
|
|
ia_data->ops = ops;
|
|
|
|
|
|
|
|
/* For tracking custom allocators that share the same ops */
|
|
|
|
list_add_tail(&ops->list, &ia_data->slist);
|
|
|
|
|
|
|
|
return ia_data;
|
|
|
|
}
|
|
|
|
|
|
|
|
static bool use_same_ops(struct ioasid_allocator_ops *a, struct ioasid_allocator_ops *b)
|
|
|
|
{
|
|
|
|
return (a->free == b->free) && (a->alloc == b->alloc);
|
|
|
|
}
|
|
|
|
|
|
|
|
/**
|
|
|
|
* ioasid_register_allocator - register a custom allocator
|
|
|
|
* @ops: the custom allocator ops to be registered
|
|
|
|
*
|
|
|
|
* Custom allocators take precedence over the default xarray based allocator.
|
|
|
|
* Private data associated with the IOASID allocated by the custom allocators
|
|
|
|
* are managed by IOASID framework similar to data stored in xa by default
|
|
|
|
* allocator.
|
|
|
|
*
|
|
|
|
* There can be multiple allocators registered but only one is active. In case
|
|
|
|
* of runtime removal of a custom allocator, the next one is activated based
|
|
|
|
* on the registration ordering.
|
|
|
|
*
|
|
|
|
* Multiple allocators can share the same alloc() function, in this case the
|
|
|
|
* IOASID space is shared.
|
|
|
|
*/
|
|
|
|
int ioasid_register_allocator(struct ioasid_allocator_ops *ops)
|
|
|
|
{
|
|
|
|
struct ioasid_allocator_data *ia_data;
|
|
|
|
struct ioasid_allocator_data *pallocator;
|
|
|
|
int ret = 0;
|
|
|
|
|
|
|
|
spin_lock(&ioasid_allocator_lock);
|
|
|
|
|
|
|
|
ia_data = ioasid_alloc_allocator(ops);
|
|
|
|
if (!ia_data) {
|
|
|
|
ret = -ENOMEM;
|
|
|
|
goto out_unlock;
|
|
|
|
}
|
|
|
|
|
|
|
|
/*
|
|
|
|
* No particular preference, we activate the first one and keep
|
|
|
|
* the later registered allocators in a list in case the first one gets
|
|
|
|
* removed due to hotplug.
|
|
|
|
*/
|
|
|
|
if (list_empty(&allocators_list)) {
|
|
|
|
WARN_ON(active_allocator != &default_allocator);
|
|
|
|
/* Use this new allocator if default is not active */
|
|
|
|
if (xa_empty(&active_allocator->xa)) {
|
|
|
|
rcu_assign_pointer(active_allocator, ia_data);
|
|
|
|
list_add_tail(&ia_data->list, &allocators_list);
|
|
|
|
goto out_unlock;
|
|
|
|
}
|
|
|
|
pr_warn("Default allocator active with outstanding IOASID\n");
|
|
|
|
ret = -EAGAIN;
|
|
|
|
goto out_free;
|
|
|
|
}
|
|
|
|
|
|
|
|
/* Check if the allocator is already registered */
|
|
|
|
list_for_each_entry(pallocator, &allocators_list, list) {
|
|
|
|
if (pallocator->ops == ops) {
|
|
|
|
pr_err("IOASID allocator already registered\n");
|
|
|
|
ret = -EEXIST;
|
|
|
|
goto out_free;
|
|
|
|
} else if (use_same_ops(pallocator->ops, ops)) {
|
|
|
|
/*
|
|
|
|
* If the new allocator shares the same ops,
|
|
|
|
* then they will share the same IOASID space.
|
|
|
|
* We should put them under the same xarray.
|
|
|
|
*/
|
|
|
|
list_add_tail(&ops->list, &pallocator->slist);
|
|
|
|
goto out_free;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
list_add_tail(&ia_data->list, &allocators_list);
|
|
|
|
|
|
|
|
spin_unlock(&ioasid_allocator_lock);
|
|
|
|
return 0;
|
|
|
|
out_free:
|
|
|
|
kfree(ia_data);
|
|
|
|
out_unlock:
|
|
|
|
spin_unlock(&ioasid_allocator_lock);
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL_GPL(ioasid_register_allocator);
|
|
|
|
|
|
|
|
/**
|
|
|
|
* ioasid_unregister_allocator - Remove a custom IOASID allocator ops
|
|
|
|
* @ops: the custom allocator to be removed
|
|
|
|
*
|
|
|
|
* Remove an allocator from the list, activate the next allocator in
|
|
|
|
* the order it was registered. Or revert to default allocator if all
|
|
|
|
* custom allocators are unregistered without outstanding IOASIDs.
|
|
|
|
*/
|
|
|
|
void ioasid_unregister_allocator(struct ioasid_allocator_ops *ops)
|
|
|
|
{
|
|
|
|
struct ioasid_allocator_data *pallocator;
|
|
|
|
struct ioasid_allocator_ops *sops;
|
|
|
|
|
|
|
|
spin_lock(&ioasid_allocator_lock);
|
|
|
|
if (list_empty(&allocators_list)) {
|
|
|
|
pr_warn("No custom IOASID allocators active!\n");
|
|
|
|
goto exit_unlock;
|
|
|
|
}
|
|
|
|
|
|
|
|
list_for_each_entry(pallocator, &allocators_list, list) {
|
|
|
|
if (!use_same_ops(pallocator->ops, ops))
|
|
|
|
continue;
|
|
|
|
|
|
|
|
if (list_is_singular(&pallocator->slist)) {
|
|
|
|
/* No shared helper functions */
|
|
|
|
list_del(&pallocator->list);
|
|
|
|
/*
|
|
|
|
* All IOASIDs should have been freed before
|
|
|
|
* the last allocator that shares the same ops
|
|
|
|
* is unregistered.
|
|
|
|
*/
|
|
|
|
WARN_ON(!xa_empty(&pallocator->xa));
|
|
|
|
if (list_empty(&allocators_list)) {
|
|
|
|
pr_info("No custom IOASID allocators, switch to default.\n");
|
|
|
|
rcu_assign_pointer(active_allocator, &default_allocator);
|
|
|
|
} else if (pallocator == active_allocator) {
|
|
|
|
rcu_assign_pointer(active_allocator,
|
|
|
|
list_first_entry(&allocators_list,
|
|
|
|
struct ioasid_allocator_data, list));
|
|
|
|
pr_info("IOASID allocator changed");
|
|
|
|
}
|
|
|
|
kfree_rcu(pallocator, rcu);
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
/*
|
|
|
|
* Find the matching shared ops to delete,
|
|
|
|
* but keep outstanding IOASIDs
|
|
|
|
*/
|
|
|
|
list_for_each_entry(sops, &pallocator->slist, list) {
|
|
|
|
if (sops == ops) {
|
|
|
|
list_del(&ops->list);
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
}
|
|
|
|
break;
|
|
|
|
}
|
|
|
|
|
|
|
|
exit_unlock:
|
|
|
|
spin_unlock(&ioasid_allocator_lock);
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL_GPL(ioasid_unregister_allocator);
|
2019-10-03 02:42:41 +07:00
|
|
|
|
|
|
|
/**
|
|
|
|
* ioasid_set_data - Set private data for an allocated ioasid
|
|
|
|
* @ioasid: the ID to set data
|
|
|
|
* @data: the private data
|
|
|
|
*
|
|
|
|
* For IOASID that is already allocated, private data can be set
|
|
|
|
* via this API. Future lookup can be done via ioasid_find.
|
|
|
|
*/
|
|
|
|
int ioasid_set_data(ioasid_t ioasid, void *data)
|
|
|
|
{
|
|
|
|
struct ioasid_data *ioasid_data;
|
|
|
|
int ret = 0;
|
|
|
|
|
2019-10-03 02:42:42 +07:00
|
|
|
spin_lock(&ioasid_allocator_lock);
|
|
|
|
ioasid_data = xa_load(&active_allocator->xa, ioasid);
|
2019-10-03 02:42:41 +07:00
|
|
|
if (ioasid_data)
|
|
|
|
rcu_assign_pointer(ioasid_data->private, data);
|
|
|
|
else
|
|
|
|
ret = -ENOENT;
|
2019-10-03 02:42:42 +07:00
|
|
|
spin_unlock(&ioasid_allocator_lock);
|
2019-10-03 02:42:41 +07:00
|
|
|
|
|
|
|
/*
|
|
|
|
* Wait for readers to stop accessing the old private data, so the
|
|
|
|
* caller can free it.
|
|
|
|
*/
|
|
|
|
if (!ret)
|
|
|
|
synchronize_rcu();
|
|
|
|
|
|
|
|
return ret;
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL_GPL(ioasid_set_data);
|
|
|
|
|
|
|
|
/**
|
|
|
|
* ioasid_alloc - Allocate an IOASID
|
|
|
|
* @set: the IOASID set
|
|
|
|
* @min: the minimum ID (inclusive)
|
|
|
|
* @max: the maximum ID (inclusive)
|
|
|
|
* @private: data private to the caller
|
|
|
|
*
|
|
|
|
* Allocate an ID between @min and @max. The @private pointer is stored
|
|
|
|
* internally and can be retrieved with ioasid_find().
|
|
|
|
*
|
|
|
|
* Return: the allocated ID on success, or %INVALID_IOASID on failure.
|
|
|
|
*/
|
|
|
|
ioasid_t ioasid_alloc(struct ioasid_set *set, ioasid_t min, ioasid_t max,
|
|
|
|
void *private)
|
|
|
|
{
|
|
|
|
struct ioasid_data *data;
|
2019-10-03 02:42:42 +07:00
|
|
|
void *adata;
|
2019-10-03 02:42:41 +07:00
|
|
|
ioasid_t id;
|
|
|
|
|
|
|
|
data = kzalloc(sizeof(*data), GFP_ATOMIC);
|
|
|
|
if (!data)
|
|
|
|
return INVALID_IOASID;
|
|
|
|
|
|
|
|
data->set = set;
|
|
|
|
data->private = private;
|
|
|
|
|
2019-10-03 02:42:42 +07:00
|
|
|
/*
|
|
|
|
* Custom allocator needs allocator data to perform platform specific
|
|
|
|
* operations.
|
|
|
|
*/
|
|
|
|
spin_lock(&ioasid_allocator_lock);
|
|
|
|
adata = active_allocator->flags & IOASID_ALLOCATOR_CUSTOM ? active_allocator->ops->pdata : data;
|
|
|
|
id = active_allocator->ops->alloc(min, max, adata);
|
|
|
|
if (id == INVALID_IOASID) {
|
|
|
|
pr_err("Failed ASID allocation %lu\n", active_allocator->flags);
|
|
|
|
goto exit_free;
|
|
|
|
}
|
|
|
|
|
|
|
|
if ((active_allocator->flags & IOASID_ALLOCATOR_CUSTOM) &&
|
|
|
|
xa_alloc(&active_allocator->xa, &id, data, XA_LIMIT(id, id), GFP_ATOMIC)) {
|
|
|
|
/* Custom allocator needs framework to store and track allocation results */
|
|
|
|
pr_err("Failed to alloc ioasid from %d\n", id);
|
|
|
|
active_allocator->ops->free(id, active_allocator->ops->pdata);
|
2019-10-03 02:42:41 +07:00
|
|
|
goto exit_free;
|
|
|
|
}
|
|
|
|
data->id = id;
|
|
|
|
|
2019-10-03 02:42:42 +07:00
|
|
|
spin_unlock(&ioasid_allocator_lock);
|
2019-10-03 02:42:41 +07:00
|
|
|
return id;
|
|
|
|
exit_free:
|
2019-10-03 02:42:42 +07:00
|
|
|
spin_unlock(&ioasid_allocator_lock);
|
2019-10-03 02:42:41 +07:00
|
|
|
kfree(data);
|
|
|
|
return INVALID_IOASID;
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL_GPL(ioasid_alloc);
|
|
|
|
|
|
|
|
/**
|
|
|
|
* ioasid_free - Free an IOASID
|
|
|
|
* @ioasid: the ID to remove
|
|
|
|
*/
|
|
|
|
void ioasid_free(ioasid_t ioasid)
|
|
|
|
{
|
|
|
|
struct ioasid_data *ioasid_data;
|
|
|
|
|
2019-10-03 02:42:42 +07:00
|
|
|
spin_lock(&ioasid_allocator_lock);
|
|
|
|
ioasid_data = xa_load(&active_allocator->xa, ioasid);
|
|
|
|
if (!ioasid_data) {
|
|
|
|
pr_err("Trying to free unknown IOASID %u\n", ioasid);
|
|
|
|
goto exit_unlock;
|
|
|
|
}
|
2019-10-03 02:42:41 +07:00
|
|
|
|
2019-10-03 02:42:42 +07:00
|
|
|
active_allocator->ops->free(ioasid, active_allocator->ops->pdata);
|
|
|
|
/* Custom allocator needs additional steps to free the xa element */
|
|
|
|
if (active_allocator->flags & IOASID_ALLOCATOR_CUSTOM) {
|
|
|
|
ioasid_data = xa_erase(&active_allocator->xa, ioasid);
|
|
|
|
kfree_rcu(ioasid_data, rcu);
|
|
|
|
}
|
|
|
|
|
|
|
|
exit_unlock:
|
|
|
|
spin_unlock(&ioasid_allocator_lock);
|
2019-10-03 02:42:41 +07:00
|
|
|
}
|
|
|
|
EXPORT_SYMBOL_GPL(ioasid_free);
|
|
|
|
|
|
|
|
/**
|
|
|
|
* ioasid_find - Find IOASID data
|
|
|
|
* @set: the IOASID set
|
|
|
|
* @ioasid: the IOASID to find
|
|
|
|
* @getter: function to call on the found object
|
|
|
|
*
|
|
|
|
* The optional getter function allows to take a reference to the found object
|
|
|
|
* under the rcu lock. The function can also check if the object is still valid:
|
|
|
|
* if @getter returns false, then the object is invalid and NULL is returned.
|
|
|
|
*
|
|
|
|
* If the IOASID exists, return the private pointer passed to ioasid_alloc.
|
|
|
|
* Private data can be NULL if not set. Return an error if the IOASID is not
|
|
|
|
* found, or if @set is not NULL and the IOASID does not belong to the set.
|
|
|
|
*/
|
|
|
|
void *ioasid_find(struct ioasid_set *set, ioasid_t ioasid,
|
|
|
|
bool (*getter)(void *))
|
|
|
|
{
|
|
|
|
void *priv;
|
|
|
|
struct ioasid_data *ioasid_data;
|
2019-10-03 02:42:42 +07:00
|
|
|
struct ioasid_allocator_data *idata;
|
2019-10-03 02:42:41 +07:00
|
|
|
|
|
|
|
rcu_read_lock();
|
2019-10-03 02:42:42 +07:00
|
|
|
idata = rcu_dereference(active_allocator);
|
|
|
|
ioasid_data = xa_load(&idata->xa, ioasid);
|
2019-10-03 02:42:41 +07:00
|
|
|
if (!ioasid_data) {
|
|
|
|
priv = ERR_PTR(-ENOENT);
|
|
|
|
goto unlock;
|
|
|
|
}
|
|
|
|
if (set && ioasid_data->set != set) {
|
|
|
|
/* data found but does not belong to the set */
|
|
|
|
priv = ERR_PTR(-EACCES);
|
|
|
|
goto unlock;
|
|
|
|
}
|
|
|
|
/* Now IOASID and its set is verified, we can return the private data */
|
|
|
|
priv = rcu_dereference(ioasid_data->private);
|
|
|
|
if (getter && !getter(priv))
|
|
|
|
priv = NULL;
|
|
|
|
unlock:
|
|
|
|
rcu_read_unlock();
|
|
|
|
|
|
|
|
return priv;
|
|
|
|
}
|
|
|
|
EXPORT_SYMBOL_GPL(ioasid_find);
|
|
|
|
|
|
|
|
MODULE_AUTHOR("Jean-Philippe Brucker <jean-philippe.brucker@arm.com>");
|
|
|
|
MODULE_AUTHOR("Jacob Pan <jacob.jun.pan@linux.intel.com>");
|
|
|
|
MODULE_DESCRIPTION("IO Address Space ID (IOASID) allocator");
|
|
|
|
MODULE_LICENSE("GPL");
|