mirror of
https://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
synced 2025-01-10 07:10:27 +00:00
dmaengine: break out channel registration
In preparation for dynamic channel registration, the code segment that does the channel registration is broken out to its own function. Signed-off-by: Dave Jiang <dave.jiang@intel.com> Link: https://lore.kernel.org/r/157965022778.73301.8929944324898985438.stgit@djiang5-desk3.ch.intel.com Signed-off-by: Vinod Koul <vkoul@kernel.org>
This commit is contained in:
parent
232bb01bb8
commit
d2fb0a0438
@ -926,6 +926,79 @@ static int get_dma_id(struct dma_device *device)
|
|||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
static int __dma_async_device_channel_register(struct dma_device *device,
|
||||||
|
struct dma_chan *chan,
|
||||||
|
int chan_id)
|
||||||
|
{
|
||||||
|
int rc = 0;
|
||||||
|
int chancnt = device->chancnt;
|
||||||
|
atomic_t *idr_ref;
|
||||||
|
struct dma_chan *tchan;
|
||||||
|
|
||||||
|
tchan = list_first_entry_or_null(&device->channels,
|
||||||
|
struct dma_chan, device_node);
|
||||||
|
if (tchan->dev) {
|
||||||
|
idr_ref = tchan->dev->idr_ref;
|
||||||
|
} else {
|
||||||
|
idr_ref = kmalloc(sizeof(*idr_ref), GFP_KERNEL);
|
||||||
|
if (!idr_ref)
|
||||||
|
return -ENOMEM;
|
||||||
|
atomic_set(idr_ref, 0);
|
||||||
|
}
|
||||||
|
|
||||||
|
chan->local = alloc_percpu(typeof(*chan->local));
|
||||||
|
if (!chan->local)
|
||||||
|
goto err_out;
|
||||||
|
chan->dev = kzalloc(sizeof(*chan->dev), GFP_KERNEL);
|
||||||
|
if (!chan->dev) {
|
||||||
|
free_percpu(chan->local);
|
||||||
|
chan->local = NULL;
|
||||||
|
goto err_out;
|
||||||
|
}
|
||||||
|
|
||||||
|
/*
|
||||||
|
* When the chan_id is a negative value, we are dynamically adding
|
||||||
|
* the channel. Otherwise we are static enumerating.
|
||||||
|
*/
|
||||||
|
chan->chan_id = chan_id < 0 ? chancnt : chan_id;
|
||||||
|
chan->dev->device.class = &dma_devclass;
|
||||||
|
chan->dev->device.parent = device->dev;
|
||||||
|
chan->dev->chan = chan;
|
||||||
|
chan->dev->idr_ref = idr_ref;
|
||||||
|
chan->dev->dev_id = device->dev_id;
|
||||||
|
atomic_inc(idr_ref);
|
||||||
|
dev_set_name(&chan->dev->device, "dma%dchan%d",
|
||||||
|
device->dev_id, chan->chan_id);
|
||||||
|
|
||||||
|
rc = device_register(&chan->dev->device);
|
||||||
|
if (rc)
|
||||||
|
goto err_out;
|
||||||
|
chan->client_count = 0;
|
||||||
|
device->chancnt = chan->chan_id + 1;
|
||||||
|
|
||||||
|
return 0;
|
||||||
|
|
||||||
|
err_out:
|
||||||
|
free_percpu(chan->local);
|
||||||
|
kfree(chan->dev);
|
||||||
|
if (atomic_dec_return(idr_ref) == 0)
|
||||||
|
kfree(idr_ref);
|
||||||
|
return rc;
|
||||||
|
}
|
||||||
|
|
||||||
|
static void __dma_async_device_channel_unregister(struct dma_device *device,
|
||||||
|
struct dma_chan *chan)
|
||||||
|
{
|
||||||
|
WARN_ONCE(!device->device_release && chan->client_count,
|
||||||
|
"%s called while %d clients hold a reference\n",
|
||||||
|
__func__, chan->client_count);
|
||||||
|
mutex_lock(&dma_list_mutex);
|
||||||
|
chan->dev->chan = NULL;
|
||||||
|
mutex_unlock(&dma_list_mutex);
|
||||||
|
device_unregister(&chan->dev->device);
|
||||||
|
free_percpu(chan->local);
|
||||||
|
}
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* dma_async_device_register - registers DMA devices found
|
* dma_async_device_register - registers DMA devices found
|
||||||
* @device: &dma_device
|
* @device: &dma_device
|
||||||
@ -936,9 +1009,8 @@ static int get_dma_id(struct dma_device *device)
|
|||||||
*/
|
*/
|
||||||
int dma_async_device_register(struct dma_device *device)
|
int dma_async_device_register(struct dma_device *device)
|
||||||
{
|
{
|
||||||
int chancnt = 0, rc;
|
int rc, i = 0;
|
||||||
struct dma_chan* chan;
|
struct dma_chan* chan;
|
||||||
atomic_t *idr_ref;
|
|
||||||
|
|
||||||
if (!device)
|
if (!device)
|
||||||
return -ENODEV;
|
return -ENODEV;
|
||||||
@ -1038,59 +1110,23 @@ int dma_async_device_register(struct dma_device *device)
|
|||||||
if (device_has_all_tx_types(device))
|
if (device_has_all_tx_types(device))
|
||||||
dma_cap_set(DMA_ASYNC_TX, device->cap_mask);
|
dma_cap_set(DMA_ASYNC_TX, device->cap_mask);
|
||||||
|
|
||||||
idr_ref = kmalloc(sizeof(*idr_ref), GFP_KERNEL);
|
|
||||||
if (!idr_ref)
|
|
||||||
return -ENOMEM;
|
|
||||||
rc = get_dma_id(device);
|
rc = get_dma_id(device);
|
||||||
if (rc != 0) {
|
if (rc != 0)
|
||||||
kfree(idr_ref);
|
|
||||||
return rc;
|
return rc;
|
||||||
}
|
|
||||||
|
|
||||||
atomic_set(idr_ref, 0);
|
|
||||||
|
|
||||||
/* represent channels in sysfs. Probably want devs too */
|
/* represent channels in sysfs. Probably want devs too */
|
||||||
list_for_each_entry(chan, &device->channels, device_node) {
|
list_for_each_entry(chan, &device->channels, device_node) {
|
||||||
rc = -ENOMEM;
|
rc = __dma_async_device_channel_register(device, chan, i++);
|
||||||
chan->local = alloc_percpu(typeof(*chan->local));
|
if (rc < 0)
|
||||||
if (chan->local == NULL)
|
|
||||||
goto err_out;
|
goto err_out;
|
||||||
chan->dev = kzalloc(sizeof(*chan->dev), GFP_KERNEL);
|
|
||||||
if (chan->dev == NULL) {
|
|
||||||
free_percpu(chan->local);
|
|
||||||
chan->local = NULL;
|
|
||||||
goto err_out;
|
|
||||||
}
|
|
||||||
|
|
||||||
chan->chan_id = chancnt++;
|
|
||||||
chan->dev->device.class = &dma_devclass;
|
|
||||||
chan->dev->device.parent = device->dev;
|
|
||||||
chan->dev->chan = chan;
|
|
||||||
chan->dev->idr_ref = idr_ref;
|
|
||||||
chan->dev->dev_id = device->dev_id;
|
|
||||||
atomic_inc(idr_ref);
|
|
||||||
dev_set_name(&chan->dev->device, "dma%dchan%d",
|
|
||||||
device->dev_id, chan->chan_id);
|
|
||||||
|
|
||||||
rc = device_register(&chan->dev->device);
|
|
||||||
if (rc) {
|
|
||||||
free_percpu(chan->local);
|
|
||||||
chan->local = NULL;
|
|
||||||
kfree(chan->dev);
|
|
||||||
atomic_dec(idr_ref);
|
|
||||||
goto err_out;
|
|
||||||
}
|
|
||||||
chan->client_count = 0;
|
|
||||||
}
|
}
|
||||||
|
|
||||||
if (!chancnt) {
|
if (!device->chancnt) {
|
||||||
dev_err(device->dev, "%s: device has no channels!\n", __func__);
|
dev_err(device->dev, "%s: device has no channels!\n", __func__);
|
||||||
rc = -ENODEV;
|
rc = -ENODEV;
|
||||||
goto err_out;
|
goto err_out;
|
||||||
}
|
}
|
||||||
|
|
||||||
device->chancnt = chancnt;
|
|
||||||
|
|
||||||
mutex_lock(&dma_list_mutex);
|
mutex_lock(&dma_list_mutex);
|
||||||
/* take references on public channels */
|
/* take references on public channels */
|
||||||
if (dmaengine_ref_count && !dma_has_cap(DMA_PRIVATE, device->cap_mask))
|
if (dmaengine_ref_count && !dma_has_cap(DMA_PRIVATE, device->cap_mask))
|
||||||
@ -1118,9 +1154,8 @@ int dma_async_device_register(struct dma_device *device)
|
|||||||
|
|
||||||
err_out:
|
err_out:
|
||||||
/* if we never registered a channel just release the idr */
|
/* if we never registered a channel just release the idr */
|
||||||
if (atomic_read(idr_ref) == 0) {
|
if (!device->chancnt) {
|
||||||
ida_free(&dma_ida, device->dev_id);
|
ida_free(&dma_ida, device->dev_id);
|
||||||
kfree(idr_ref);
|
|
||||||
return rc;
|
return rc;
|
||||||
}
|
}
|
||||||
|
|
||||||
@ -1148,16 +1183,8 @@ void dma_async_device_unregister(struct dma_device *device)
|
|||||||
{
|
{
|
||||||
struct dma_chan *chan;
|
struct dma_chan *chan;
|
||||||
|
|
||||||
list_for_each_entry(chan, &device->channels, device_node) {
|
list_for_each_entry(chan, &device->channels, device_node)
|
||||||
WARN_ONCE(!device->device_release && chan->client_count,
|
__dma_async_device_channel_unregister(device, chan);
|
||||||
"%s called while %d clients hold a reference\n",
|
|
||||||
__func__, chan->client_count);
|
|
||||||
mutex_lock(&dma_list_mutex);
|
|
||||||
chan->dev->chan = NULL;
|
|
||||||
mutex_unlock(&dma_list_mutex);
|
|
||||||
device_unregister(&chan->dev->device);
|
|
||||||
free_percpu(chan->local);
|
|
||||||
}
|
|
||||||
|
|
||||||
mutex_lock(&dma_list_mutex);
|
mutex_lock(&dma_list_mutex);
|
||||||
/*
|
/*
|
||||||
|
Loading…
x
Reference in New Issue
Block a user