mirror of
https://mirrors.bfsu.edu.cn/git/linux.git
synced 2025-01-19 20:34:20 +08:00
dmaengine: centralize channel allocation, introduce dma_find_channel
Allowing multiple clients to each define their own channel allocation scheme quickly leads to a pathological situation. For memory-to-memory offload all clients can share a central allocator. This simply moves the existing async_tx allocator to dmaengine with minimal fixups: * async_tx.c:get_chan_ref_by_cap --> dmaengine.c:nth_chan * async_tx.c:async_tx_rebalance --> dmaengine.c:dma_channel_rebalance * split out common code from async_tx.c:__async_tx_find_channel --> dma_find_channel Reviewed-by: Andrew Morton <akpm@linux-foundation.org> Signed-off-by: Dan Williams <dan.j.williams@intel.com>
This commit is contained in:
parent
6f49a57aa5
commit
bec085134e
@ -37,26 +37,11 @@ static struct dma_client async_tx_dma = {
|
|||||||
/* .cap_mask == 0 defaults to all channels */
|
/* .cap_mask == 0 defaults to all channels */
|
||||||
};
|
};
|
||||||
|
|
||||||
/**
|
|
||||||
* dma_cap_mask_all - enable iteration over all operation types
|
|
||||||
*/
|
|
||||||
static dma_cap_mask_t dma_cap_mask_all;
|
|
||||||
|
|
||||||
/**
|
|
||||||
* chan_ref_percpu - tracks channel allocations per core/opertion
|
|
||||||
*/
|
|
||||||
struct chan_ref_percpu {
|
|
||||||
struct dma_chan_ref *ref;
|
|
||||||
};
|
|
||||||
|
|
||||||
static int channel_table_initialized;
|
|
||||||
static struct chan_ref_percpu *channel_table[DMA_TX_TYPE_END];
|
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* async_tx_lock - protect modification of async_tx_master_list and serialize
|
* async_tx_lock - protect modification of async_tx_master_list and serialize
|
||||||
* rebalance operations
|
* rebalance operations
|
||||||
*/
|
*/
|
||||||
static spinlock_t async_tx_lock;
|
static DEFINE_SPINLOCK(async_tx_lock);
|
||||||
|
|
||||||
static LIST_HEAD(async_tx_master_list);
|
static LIST_HEAD(async_tx_master_list);
|
||||||
|
|
||||||
@ -89,85 +74,6 @@ init_dma_chan_ref(struct dma_chan_ref *ref, struct dma_chan *chan)
|
|||||||
atomic_set(&ref->count, 0);
|
atomic_set(&ref->count, 0);
|
||||||
}
|
}
|
||||||
|
|
||||||
/**
|
|
||||||
* get_chan_ref_by_cap - returns the nth channel of the given capability
|
|
||||||
* defaults to returning the channel with the desired capability and the
|
|
||||||
* lowest reference count if the index can not be satisfied
|
|
||||||
* @cap: capability to match
|
|
||||||
* @index: nth channel desired, passing -1 has the effect of forcing the
|
|
||||||
* default return value
|
|
||||||
*/
|
|
||||||
static struct dma_chan_ref *
|
|
||||||
get_chan_ref_by_cap(enum dma_transaction_type cap, int index)
|
|
||||||
{
|
|
||||||
struct dma_chan_ref *ret_ref = NULL, *min_ref = NULL, *ref;
|
|
||||||
|
|
||||||
rcu_read_lock();
|
|
||||||
list_for_each_entry_rcu(ref, &async_tx_master_list, node)
|
|
||||||
if (dma_has_cap(cap, ref->chan->device->cap_mask)) {
|
|
||||||
if (!min_ref)
|
|
||||||
min_ref = ref;
|
|
||||||
else if (atomic_read(&ref->count) <
|
|
||||||
atomic_read(&min_ref->count))
|
|
||||||
min_ref = ref;
|
|
||||||
|
|
||||||
if (index-- == 0) {
|
|
||||||
ret_ref = ref;
|
|
||||||
break;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
rcu_read_unlock();
|
|
||||||
|
|
||||||
if (!ret_ref)
|
|
||||||
ret_ref = min_ref;
|
|
||||||
|
|
||||||
if (ret_ref)
|
|
||||||
atomic_inc(&ret_ref->count);
|
|
||||||
|
|
||||||
return ret_ref;
|
|
||||||
}
|
|
||||||
|
|
||||||
/**
|
|
||||||
* async_tx_rebalance - redistribute the available channels, optimize
|
|
||||||
* for cpu isolation in the SMP case, and opertaion isolation in the
|
|
||||||
* uniprocessor case
|
|
||||||
*/
|
|
||||||
static void async_tx_rebalance(void)
|
|
||||||
{
|
|
||||||
int cpu, cap, cpu_idx = 0;
|
|
||||||
unsigned long flags;
|
|
||||||
|
|
||||||
if (!channel_table_initialized)
|
|
||||||
return;
|
|
||||||
|
|
||||||
spin_lock_irqsave(&async_tx_lock, flags);
|
|
||||||
|
|
||||||
/* undo the last distribution */
|
|
||||||
for_each_dma_cap_mask(cap, dma_cap_mask_all)
|
|
||||||
for_each_possible_cpu(cpu) {
|
|
||||||
struct dma_chan_ref *ref =
|
|
||||||
per_cpu_ptr(channel_table[cap], cpu)->ref;
|
|
||||||
if (ref) {
|
|
||||||
atomic_set(&ref->count, 0);
|
|
||||||
per_cpu_ptr(channel_table[cap], cpu)->ref =
|
|
||||||
NULL;
|
|
||||||
}
|
|
||||||
}
|
|
||||||
|
|
||||||
for_each_dma_cap_mask(cap, dma_cap_mask_all)
|
|
||||||
for_each_online_cpu(cpu) {
|
|
||||||
struct dma_chan_ref *new;
|
|
||||||
if (NR_CPUS > 1)
|
|
||||||
new = get_chan_ref_by_cap(cap, cpu_idx++);
|
|
||||||
else
|
|
||||||
new = get_chan_ref_by_cap(cap, -1);
|
|
||||||
|
|
||||||
per_cpu_ptr(channel_table[cap], cpu)->ref = new;
|
|
||||||
}
|
|
||||||
|
|
||||||
spin_unlock_irqrestore(&async_tx_lock, flags);
|
|
||||||
}
|
|
||||||
|
|
||||||
static enum dma_state_client
|
static enum dma_state_client
|
||||||
dma_channel_add_remove(struct dma_client *client,
|
dma_channel_add_remove(struct dma_client *client,
|
||||||
struct dma_chan *chan, enum dma_state state)
|
struct dma_chan *chan, enum dma_state state)
|
||||||
@ -211,8 +117,6 @@ dma_channel_add_remove(struct dma_client *client,
|
|||||||
" (-ENOMEM)\n");
|
" (-ENOMEM)\n");
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
async_tx_rebalance();
|
|
||||||
break;
|
break;
|
||||||
case DMA_RESOURCE_REMOVED:
|
case DMA_RESOURCE_REMOVED:
|
||||||
found = 0;
|
found = 0;
|
||||||
@ -233,8 +137,6 @@ dma_channel_add_remove(struct dma_client *client,
|
|||||||
ack = DMA_ACK;
|
ack = DMA_ACK;
|
||||||
else
|
else
|
||||||
break;
|
break;
|
||||||
|
|
||||||
async_tx_rebalance();
|
|
||||||
break;
|
break;
|
||||||
case DMA_RESOURCE_SUSPEND:
|
case DMA_RESOURCE_SUSPEND:
|
||||||
case DMA_RESOURCE_RESUME:
|
case DMA_RESOURCE_RESUME:
|
||||||
@ -248,51 +150,18 @@ dma_channel_add_remove(struct dma_client *client,
|
|||||||
return ack;
|
return ack;
|
||||||
}
|
}
|
||||||
|
|
||||||
static int __init
|
static int __init async_tx_init(void)
|
||||||
async_tx_init(void)
|
|
||||||
{
|
{
|
||||||
enum dma_transaction_type cap;
|
|
||||||
|
|
||||||
spin_lock_init(&async_tx_lock);
|
|
||||||
bitmap_fill(dma_cap_mask_all.bits, DMA_TX_TYPE_END);
|
|
||||||
|
|
||||||
/* an interrupt will never be an explicit operation type.
|
|
||||||
* clearing this bit prevents allocation to a slot in 'channel_table'
|
|
||||||
*/
|
|
||||||
clear_bit(DMA_INTERRUPT, dma_cap_mask_all.bits);
|
|
||||||
|
|
||||||
for_each_dma_cap_mask(cap, dma_cap_mask_all) {
|
|
||||||
channel_table[cap] = alloc_percpu(struct chan_ref_percpu);
|
|
||||||
if (!channel_table[cap])
|
|
||||||
goto err;
|
|
||||||
}
|
|
||||||
|
|
||||||
channel_table_initialized = 1;
|
|
||||||
dma_async_client_register(&async_tx_dma);
|
dma_async_client_register(&async_tx_dma);
|
||||||
dma_async_client_chan_request(&async_tx_dma);
|
dma_async_client_chan_request(&async_tx_dma);
|
||||||
|
|
||||||
printk(KERN_INFO "async_tx: api initialized (async)\n");
|
printk(KERN_INFO "async_tx: api initialized (async)\n");
|
||||||
|
|
||||||
return 0;
|
return 0;
|
||||||
err:
|
|
||||||
printk(KERN_ERR "async_tx: initialization failure\n");
|
|
||||||
|
|
||||||
while (--cap >= 0)
|
|
||||||
free_percpu(channel_table[cap]);
|
|
||||||
|
|
||||||
return 1;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
static void __exit async_tx_exit(void)
|
static void __exit async_tx_exit(void)
|
||||||
{
|
{
|
||||||
enum dma_transaction_type cap;
|
|
||||||
|
|
||||||
channel_table_initialized = 0;
|
|
||||||
|
|
||||||
for_each_dma_cap_mask(cap, dma_cap_mask_all)
|
|
||||||
if (channel_table[cap])
|
|
||||||
free_percpu(channel_table[cap]);
|
|
||||||
|
|
||||||
dma_async_client_unregister(&async_tx_dma);
|
dma_async_client_unregister(&async_tx_dma);
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -308,16 +177,9 @@ __async_tx_find_channel(struct dma_async_tx_descriptor *depend_tx,
|
|||||||
{
|
{
|
||||||
/* see if we can keep the chain on one channel */
|
/* see if we can keep the chain on one channel */
|
||||||
if (depend_tx &&
|
if (depend_tx &&
|
||||||
dma_has_cap(tx_type, depend_tx->chan->device->cap_mask))
|
dma_has_cap(tx_type, depend_tx->chan->device->cap_mask))
|
||||||
return depend_tx->chan;
|
return depend_tx->chan;
|
||||||
else if (likely(channel_table_initialized)) {
|
return dma_find_channel(tx_type);
|
||||||
struct dma_chan_ref *ref;
|
|
||||||
int cpu = get_cpu();
|
|
||||||
ref = per_cpu_ptr(channel_table[tx_type], cpu)->ref;
|
|
||||||
put_cpu();
|
|
||||||
return ref ? ref->chan : NULL;
|
|
||||||
} else
|
|
||||||
return NULL;
|
|
||||||
}
|
}
|
||||||
EXPORT_SYMBOL_GPL(__async_tx_find_channel);
|
EXPORT_SYMBOL_GPL(__async_tx_find_channel);
|
||||||
#else
|
#else
|
||||||
|
@ -294,6 +294,164 @@ static void dma_chan_release(struct dma_chan *chan)
|
|||||||
call_rcu(&chan->rcu, dma_chan_free_rcu);
|
call_rcu(&chan->rcu, dma_chan_free_rcu);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* dma_cap_mask_all - enable iteration over all operation types
|
||||||
|
*/
|
||||||
|
static dma_cap_mask_t dma_cap_mask_all;
|
||||||
|
|
||||||
|
/**
|
||||||
|
* dma_chan_tbl_ent - tracks channel allocations per core/operation
|
||||||
|
* @chan - associated channel for this entry
|
||||||
|
*/
|
||||||
|
struct dma_chan_tbl_ent {
|
||||||
|
struct dma_chan *chan;
|
||||||
|
};
|
||||||
|
|
||||||
|
/**
|
||||||
|
* channel_table - percpu lookup table for memory-to-memory offload providers
|
||||||
|
*/
|
||||||
|
static struct dma_chan_tbl_ent *channel_table[DMA_TX_TYPE_END];
|
||||||
|
|
||||||
|
static int __init dma_channel_table_init(void)
|
||||||
|
{
|
||||||
|
enum dma_transaction_type cap;
|
||||||
|
int err = 0;
|
||||||
|
|
||||||
|
bitmap_fill(dma_cap_mask_all.bits, DMA_TX_TYPE_END);
|
||||||
|
|
||||||
|
/* 'interrupt' and 'slave' are channel capabilities, but are not
|
||||||
|
* associated with an operation so they do not need an entry in the
|
||||||
|
* channel_table
|
||||||
|
*/
|
||||||
|
clear_bit(DMA_INTERRUPT, dma_cap_mask_all.bits);
|
||||||
|
clear_bit(DMA_SLAVE, dma_cap_mask_all.bits);
|
||||||
|
|
||||||
|
for_each_dma_cap_mask(cap, dma_cap_mask_all) {
|
||||||
|
channel_table[cap] = alloc_percpu(struct dma_chan_tbl_ent);
|
||||||
|
if (!channel_table[cap]) {
|
||||||
|
err = -ENOMEM;
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
|
if (err) {
|
||||||
|
pr_err("dmaengine: initialization failure\n");
|
||||||
|
for_each_dma_cap_mask(cap, dma_cap_mask_all)
|
||||||
|
if (channel_table[cap])
|
||||||
|
free_percpu(channel_table[cap]);
|
||||||
|
}
|
||||||
|
|
||||||
|
return err;
|
||||||
|
}
|
||||||
|
subsys_initcall(dma_channel_table_init);
|
||||||
|
|
||||||
|
/**
|
||||||
|
* dma_find_channel - find a channel to carry out the operation
|
||||||
|
* @tx_type: transaction type
|
||||||
|
*/
|
||||||
|
struct dma_chan *dma_find_channel(enum dma_transaction_type tx_type)
|
||||||
|
{
|
||||||
|
struct dma_chan *chan;
|
||||||
|
int cpu;
|
||||||
|
|
||||||
|
WARN_ONCE(dmaengine_ref_count == 0,
|
||||||
|
"client called %s without a reference", __func__);
|
||||||
|
|
||||||
|
cpu = get_cpu();
|
||||||
|
chan = per_cpu_ptr(channel_table[tx_type], cpu)->chan;
|
||||||
|
put_cpu();
|
||||||
|
|
||||||
|
return chan;
|
||||||
|
}
|
||||||
|
EXPORT_SYMBOL(dma_find_channel);
|
||||||
|
|
||||||
|
/**
|
||||||
|
* nth_chan - returns the nth channel of the given capability
|
||||||
|
* @cap: capability to match
|
||||||
|
* @n: nth channel desired
|
||||||
|
*
|
||||||
|
* Defaults to returning the channel with the desired capability and the
|
||||||
|
* lowest reference count when 'n' cannot be satisfied. Must be called
|
||||||
|
* under dma_list_mutex.
|
||||||
|
*/
|
||||||
|
static struct dma_chan *nth_chan(enum dma_transaction_type cap, int n)
|
||||||
|
{
|
||||||
|
struct dma_device *device;
|
||||||
|
struct dma_chan *chan;
|
||||||
|
struct dma_chan *ret = NULL;
|
||||||
|
struct dma_chan *min = NULL;
|
||||||
|
|
||||||
|
list_for_each_entry(device, &dma_device_list, global_node) {
|
||||||
|
if (!dma_has_cap(cap, device->cap_mask))
|
||||||
|
continue;
|
||||||
|
list_for_each_entry(chan, &device->channels, device_node) {
|
||||||
|
if (!chan->client_count)
|
||||||
|
continue;
|
||||||
|
if (!min)
|
||||||
|
min = chan;
|
||||||
|
else if (chan->table_count < min->table_count)
|
||||||
|
min = chan;
|
||||||
|
|
||||||
|
if (n-- == 0) {
|
||||||
|
ret = chan;
|
||||||
|
break; /* done */
|
||||||
|
}
|
||||||
|
}
|
||||||
|
if (ret)
|
||||||
|
break; /* done */
|
||||||
|
}
|
||||||
|
|
||||||
|
if (!ret)
|
||||||
|
ret = min;
|
||||||
|
|
||||||
|
if (ret)
|
||||||
|
ret->table_count++;
|
||||||
|
|
||||||
|
return ret;
|
||||||
|
}
|
||||||
|
|
||||||
|
/**
|
||||||
|
* dma_channel_rebalance - redistribute the available channels
|
||||||
|
*
|
||||||
|
* Optimize for cpu isolation (each cpu gets a dedicated channel for an
|
||||||
|
* operation type) in the SMP case, and operation isolation (avoid
|
||||||
|
* multi-tasking channels) in the non-SMP case. Must be called under
|
||||||
|
* dma_list_mutex.
|
||||||
|
*/
|
||||||
|
static void dma_channel_rebalance(void)
|
||||||
|
{
|
||||||
|
struct dma_chan *chan;
|
||||||
|
struct dma_device *device;
|
||||||
|
int cpu;
|
||||||
|
int cap;
|
||||||
|
int n;
|
||||||
|
|
||||||
|
/* undo the last distribution */
|
||||||
|
for_each_dma_cap_mask(cap, dma_cap_mask_all)
|
||||||
|
for_each_possible_cpu(cpu)
|
||||||
|
per_cpu_ptr(channel_table[cap], cpu)->chan = NULL;
|
||||||
|
|
||||||
|
list_for_each_entry(device, &dma_device_list, global_node)
|
||||||
|
list_for_each_entry(chan, &device->channels, device_node)
|
||||||
|
chan->table_count = 0;
|
||||||
|
|
||||||
|
/* don't populate the channel_table if no clients are available */
|
||||||
|
if (!dmaengine_ref_count)
|
||||||
|
return;
|
||||||
|
|
||||||
|
/* redistribute available channels */
|
||||||
|
n = 0;
|
||||||
|
for_each_dma_cap_mask(cap, dma_cap_mask_all)
|
||||||
|
for_each_online_cpu(cpu) {
|
||||||
|
if (num_possible_cpus() > 1)
|
||||||
|
chan = nth_chan(cap, n++);
|
||||||
|
else
|
||||||
|
chan = nth_chan(cap, -1);
|
||||||
|
|
||||||
|
per_cpu_ptr(channel_table[cap], cpu)->chan = chan;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* dma_chans_notify_available - broadcast available channels to the clients
|
* dma_chans_notify_available - broadcast available channels to the clients
|
||||||
*/
|
*/
|
||||||
@ -339,7 +497,12 @@ void dma_async_client_register(struct dma_client *client)
|
|||||||
dev_name(&chan->dev), err);
|
dev_name(&chan->dev), err);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
/* if this is the first reference and there were channels
|
||||||
|
* waiting we need to rebalance to get those channels
|
||||||
|
* incorporated into the channel table
|
||||||
|
*/
|
||||||
|
if (dmaengine_ref_count == 1)
|
||||||
|
dma_channel_rebalance();
|
||||||
list_add_tail(&client->global_node, &dma_client_list);
|
list_add_tail(&client->global_node, &dma_client_list);
|
||||||
mutex_unlock(&dma_list_mutex);
|
mutex_unlock(&dma_list_mutex);
|
||||||
}
|
}
|
||||||
@ -473,6 +636,7 @@ int dma_async_device_register(struct dma_device *device)
|
|||||||
}
|
}
|
||||||
}
|
}
|
||||||
list_add_tail(&device->global_node, &dma_device_list);
|
list_add_tail(&device->global_node, &dma_device_list);
|
||||||
|
dma_channel_rebalance();
|
||||||
mutex_unlock(&dma_list_mutex);
|
mutex_unlock(&dma_list_mutex);
|
||||||
|
|
||||||
dma_clients_notify_available();
|
dma_clients_notify_available();
|
||||||
@ -514,6 +678,7 @@ void dma_async_device_unregister(struct dma_device *device)
|
|||||||
|
|
||||||
mutex_lock(&dma_list_mutex);
|
mutex_lock(&dma_list_mutex);
|
||||||
list_del(&device->global_node);
|
list_del(&device->global_node);
|
||||||
|
dma_channel_rebalance();
|
||||||
mutex_unlock(&dma_list_mutex);
|
mutex_unlock(&dma_list_mutex);
|
||||||
|
|
||||||
list_for_each_entry(chan, &device->channels, device_node) {
|
list_for_each_entry(chan, &device->channels, device_node) {
|
||||||
@ -768,3 +933,4 @@ static int __init dma_bus_init(void)
|
|||||||
}
|
}
|
||||||
subsys_initcall(dma_bus_init);
|
subsys_initcall(dma_bus_init);
|
||||||
|
|
||||||
|
|
||||||
|
@ -182,6 +182,7 @@ struct dma_chan_percpu {
|
|||||||
* @device_node: used to add this to the device chan list
|
* @device_node: used to add this to the device chan list
|
||||||
* @local: per-cpu pointer to a struct dma_chan_percpu
|
* @local: per-cpu pointer to a struct dma_chan_percpu
|
||||||
* @client-count: how many clients are using this channel
|
* @client-count: how many clients are using this channel
|
||||||
|
* @table_count: number of appearances in the mem-to-mem allocation table
|
||||||
*/
|
*/
|
||||||
struct dma_chan {
|
struct dma_chan {
|
||||||
struct dma_device *device;
|
struct dma_device *device;
|
||||||
@ -198,6 +199,7 @@ struct dma_chan {
|
|||||||
struct list_head device_node;
|
struct list_head device_node;
|
||||||
struct dma_chan_percpu *local;
|
struct dma_chan_percpu *local;
|
||||||
int client_count;
|
int client_count;
|
||||||
|
int table_count;
|
||||||
};
|
};
|
||||||
|
|
||||||
#define to_dma_chan(p) container_of(p, struct dma_chan, dev)
|
#define to_dma_chan(p) container_of(p, struct dma_chan, dev)
|
||||||
@ -468,6 +470,7 @@ static inline enum dma_status dma_wait_for_async_tx(struct dma_async_tx_descript
|
|||||||
int dma_async_device_register(struct dma_device *device);
|
int dma_async_device_register(struct dma_device *device);
|
||||||
void dma_async_device_unregister(struct dma_device *device);
|
void dma_async_device_unregister(struct dma_device *device);
|
||||||
void dma_run_dependencies(struct dma_async_tx_descriptor *tx);
|
void dma_run_dependencies(struct dma_async_tx_descriptor *tx);
|
||||||
|
struct dma_chan *dma_find_channel(enum dma_transaction_type tx_type);
|
||||||
|
|
||||||
/* --- Helper iov-locking functions --- */
|
/* --- Helper iov-locking functions --- */
|
||||||
|
|
||||||
|
Loading…
Reference in New Issue
Block a user