mirror of
https://github.com/AuxXxilium/linux_dsm_epyc7002.git
synced 2024-11-24 14:00:58 +07:00
** NOW WITH TESTING! **
Two fixes which got lost in my recent distraction. One is a weird cpumask function which needed to be rewritten, the other is a module bug which is cc:stable. Thanks, Rusty. -----BEGIN PGP SIGNATURE----- Version: GnuPG v1 iQIcBAABAgAGBQJVaBENAAoJENkgDmzRrbjxxL4QAJMFwo21VN8rwIsEJ2P/Yh4u YXxJtnbrSPZtyad8J4G6FGOOfM7ImkkADhGJE8MN05goIFmeORWduAiozBtZBfo3 OVpeo0HIGTEMXq/QCxSQsDhP9MSeWV592vjhlqQJ2KhU9Gpstc/Ub9ArVWuY3FD3 CFN6ciw+5DIhoc6jMI2P9XX7jpR4VOBu320j+3lQ1QZ1aEZIaPefWH+VYuIZXirq E6N4yKgTahKb1Clr0DS6EB2Z5g+upNzFf4WBHaChP5EklwatZkHAOvzfSLWcbShI ochGV5LBPcn7ruqOD5mR4LGkxfQSYPCKCKihmenD/EVoO/dshKOQREfsqRXNsh5X xk4yx/VCy68ubIjx7FIDL18qDvJrX82+Z2bYZbENvKrVinaQ7MWB+CokK0fNW0ai ZMP5s32vSUZMMIIE7+fS4n3BLUxOpLZC8S0wIac19jNKzCHVTuhnUolCHk11zQLk IIDHEJwzvWtPjKOyUyd7HG0bYeczwf8DZgHg+xom9BNbHbK3Jk5d1Sibjgf8eGg+ O36XR8FYYvqHwqqrPKSSaWoLj578/IWyHZg/V4tQ2HWi189BVHk6Iw2knftsvvPw pBu2AdbRSLLD+X/pwrdmm+xgytjUIr1X/Qnwj/eE5MvB/vaVVwV0OjapU/Z6S+dL JrZGvbWcviyjpvGD+vG1 =wuP+ -----END PGP SIGNATURE----- Merge tag 'fixes-for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/rusty/linux Pull fixes for cpumask and modules from Rusty Russell: "** NOW WITH TESTING! ** Two fixes which got lost in my recent distraction. One is a weird cpumask function which needed to be rewritten, the other is a module bug which is cc:stable" * tag 'fixes-for-linus' of git://git.kernel.org/pub/scm/linux/kernel/git/rusty/linux: cpumask_set_cpu_local_first => cpumask_local_spread, lament module: Call module notifier on failure after complete_formation()
This commit is contained in:
commit
6e49ba1bb1
@ -2358,11 +2358,11 @@ static int be_evt_queues_create(struct be_adapter *adapter)
|
||||
adapter->cfg_num_qs);
|
||||
|
||||
for_all_evt_queues(adapter, eqo, i) {
|
||||
int numa_node = dev_to_node(&adapter->pdev->dev);
|
||||
if (!zalloc_cpumask_var(&eqo->affinity_mask, GFP_KERNEL))
|
||||
return -ENOMEM;
|
||||
cpumask_set_cpu_local_first(i, dev_to_node(&adapter->pdev->dev),
|
||||
eqo->affinity_mask);
|
||||
|
||||
cpumask_set_cpu(cpumask_local_spread(i, numa_node),
|
||||
eqo->affinity_mask);
|
||||
netif_napi_add(adapter->netdev, &eqo->napi, be_poll,
|
||||
BE_NAPI_WEIGHT);
|
||||
napi_hash_add(&eqo->napi);
|
||||
|
@ -1501,17 +1501,13 @@ static int mlx4_en_init_affinity_hint(struct mlx4_en_priv *priv, int ring_idx)
|
||||
{
|
||||
struct mlx4_en_rx_ring *ring = priv->rx_ring[ring_idx];
|
||||
int numa_node = priv->mdev->dev->numa_node;
|
||||
int ret = 0;
|
||||
|
||||
if (!zalloc_cpumask_var(&ring->affinity_mask, GFP_KERNEL))
|
||||
return -ENOMEM;
|
||||
|
||||
ret = cpumask_set_cpu_local_first(ring_idx, numa_node,
|
||||
ring->affinity_mask);
|
||||
if (ret)
|
||||
free_cpumask_var(ring->affinity_mask);
|
||||
|
||||
return ret;
|
||||
cpumask_set_cpu(cpumask_local_spread(ring_idx, numa_node),
|
||||
ring->affinity_mask);
|
||||
return 0;
|
||||
}
|
||||
|
||||
static void mlx4_en_free_affinity_hint(struct mlx4_en_priv *priv, int ring_idx)
|
||||
|
@ -144,9 +144,9 @@ int mlx4_en_create_tx_ring(struct mlx4_en_priv *priv,
|
||||
ring->queue_index = queue_index;
|
||||
|
||||
if (queue_index < priv->num_tx_rings_p_up)
|
||||
cpumask_set_cpu_local_first(queue_index,
|
||||
priv->mdev->dev->numa_node,
|
||||
&ring->affinity_mask);
|
||||
cpumask_set_cpu(cpumask_local_spread(queue_index,
|
||||
priv->mdev->dev->numa_node),
|
||||
&ring->affinity_mask);
|
||||
|
||||
*pring = ring;
|
||||
return 0;
|
||||
|
@ -151,10 +151,8 @@ static inline unsigned int cpumask_any_but(const struct cpumask *mask,
|
||||
return 1;
|
||||
}
|
||||
|
||||
static inline int cpumask_set_cpu_local_first(int i, int numa_node, cpumask_t *dstp)
|
||||
static inline unsigned int cpumask_local_spread(unsigned int i, int node)
|
||||
{
|
||||
set_bit(0, cpumask_bits(dstp));
|
||||
|
||||
return 0;
|
||||
}
|
||||
|
||||
@ -208,7 +206,7 @@ static inline unsigned int cpumask_next_zero(int n, const struct cpumask *srcp)
|
||||
|
||||
int cpumask_next_and(int n, const struct cpumask *, const struct cpumask *);
|
||||
int cpumask_any_but(const struct cpumask *mask, unsigned int cpu);
|
||||
int cpumask_set_cpu_local_first(int i, int numa_node, cpumask_t *dstp);
|
||||
unsigned int cpumask_local_spread(unsigned int i, int node);
|
||||
|
||||
/**
|
||||
* for_each_cpu - iterate over every cpu in a mask
|
||||
|
@ -3370,6 +3370,9 @@ static int load_module(struct load_info *info, const char __user *uargs,
|
||||
module_bug_cleanup(mod);
|
||||
mutex_unlock(&module_mutex);
|
||||
|
||||
blocking_notifier_call_chain(&module_notify_list,
|
||||
MODULE_STATE_GOING, mod);
|
||||
|
||||
/* we can't deallocate the module until we clear memory protection */
|
||||
unset_module_init_ro_nx(mod);
|
||||
unset_module_core_ro_nx(mod);
|
||||
|
@ -139,64 +139,42 @@ void __init free_bootmem_cpumask_var(cpumask_var_t mask)
|
||||
#endif
|
||||
|
||||
/**
|
||||
* cpumask_set_cpu_local_first - set i'th cpu with local numa cpu's first
|
||||
*
|
||||
* cpumask_local_spread - select the i'th cpu with local numa cpu's first
|
||||
* @i: index number
|
||||
* @numa_node: local numa_node
|
||||
* @dstp: cpumask with the relevant cpu bit set according to the policy
|
||||
* @node: local numa_node
|
||||
*
|
||||
* This function sets the cpumask according to a numa aware policy.
|
||||
* cpumask could be used as an affinity hint for the IRQ related to a
|
||||
* queue. When the policy is to spread queues across cores - local cores
|
||||
* first.
|
||||
* This function selects an online CPU according to a numa aware policy;
|
||||
* local cpus are returned first, followed by non-local ones, then it
|
||||
* wraps around.
|
||||
*
|
||||
* Returns 0 on success, -ENOMEM for no memory, and -EAGAIN when failed to set
|
||||
* the cpu bit and need to re-call the function.
|
||||
* It's not very efficient, but useful for setup.
|
||||
*/
|
||||
int cpumask_set_cpu_local_first(int i, int numa_node, cpumask_t *dstp)
|
||||
unsigned int cpumask_local_spread(unsigned int i, int node)
|
||||
{
|
||||
cpumask_var_t mask;
|
||||
int cpu;
|
||||
int ret = 0;
|
||||
|
||||
if (!zalloc_cpumask_var(&mask, GFP_KERNEL))
|
||||
return -ENOMEM;
|
||||
|
||||
/* Wrap: we always want a cpu. */
|
||||
i %= num_online_cpus();
|
||||
|
||||
if (numa_node == -1 || !cpumask_of_node(numa_node)) {
|
||||
/* Use all online cpu's for non numa aware system */
|
||||
cpumask_copy(mask, cpu_online_mask);
|
||||
if (node == -1) {
|
||||
for_each_cpu(cpu, cpu_online_mask)
|
||||
if (i-- == 0)
|
||||
return cpu;
|
||||
} else {
|
||||
int n;
|
||||
/* NUMA first. */
|
||||
for_each_cpu_and(cpu, cpumask_of_node(node), cpu_online_mask)
|
||||
if (i-- == 0)
|
||||
return cpu;
|
||||
|
||||
cpumask_and(mask,
|
||||
cpumask_of_node(numa_node), cpu_online_mask);
|
||||
for_each_cpu(cpu, cpu_online_mask) {
|
||||
/* Skip NUMA nodes, done above. */
|
||||
if (cpumask_test_cpu(cpu, cpumask_of_node(node)))
|
||||
continue;
|
||||
|
||||
n = cpumask_weight(mask);
|
||||
if (i >= n) {
|
||||
i -= n;
|
||||
|
||||
/* If index > number of local cpu's, mask out local
|
||||
* cpu's
|
||||
*/
|
||||
cpumask_andnot(mask, cpu_online_mask, mask);
|
||||
if (i-- == 0)
|
||||
return cpu;
|
||||
}
|
||||
}
|
||||
|
||||
for_each_cpu(cpu, mask) {
|
||||
if (--i < 0)
|
||||
goto out;
|
||||
}
|
||||
|
||||
ret = -EAGAIN;
|
||||
|
||||
out:
|
||||
free_cpumask_var(mask);
|
||||
|
||||
if (!ret)
|
||||
cpumask_set_cpu(cpu, dstp);
|
||||
|
||||
return ret;
|
||||
BUG();
|
||||
}
|
||||
EXPORT_SYMBOL(cpumask_set_cpu_local_first);
|
||||
EXPORT_SYMBOL(cpumask_local_spread);
|
||||
|
Loading…
Reference in New Issue
Block a user