On Tue, Jun 15, 2021 at 01:21:17PM +0200, Christian König wrote:
Daniel pointed me towards this function and there are multiple obvious problems in the implementation.
First of all the retry loop is not working as intended. In general the retry makes only sense if you grab the reference first and then check the retry. Then we skipped checking the exclusive fence when shared fences were present. And last the whole implementation was unnecessary complex and rather hard to understand which could lead to probably unexpected behavior of the IOCTL.
Fix all this by reworking the implementation from scratch.
Can't we split this a bit?
The other thing I'm wondering, instead of open-coding this and breaking our heads trying to make sure we got it right. Can't we reuse dma_resv_get_fences? That's what a lot of drivers use already to get a consistent copy of the fence set without holding the lock.
I think then the actual semantics, i.e. do we need to include the exclusive fence or not, stick out more. -Daniel
Only mildly tested and needs a thoughtful review of the code.
Signed-off-by: Christian König christian.koenig@amd.com
drivers/dma-buf/dma-buf.c | 132 +++++++++++++++----------------------- include/linux/dma-buf.h | 2 +- 2 files changed, 54 insertions(+), 80 deletions(-)
diff --git a/drivers/dma-buf/dma-buf.c b/drivers/dma-buf/dma-buf.c index 511fe0d217a0..1bd00e18291f 100644 --- a/drivers/dma-buf/dma-buf.c +++ b/drivers/dma-buf/dma-buf.c @@ -72,7 +72,7 @@ static void dma_buf_release(struct dentry *dentry) * If you hit this BUG() it means someone dropped their ref to the * dma-buf while still having pending operation to the buffer. */
- BUG_ON(dmabuf->cb_shared.active || dmabuf->cb_excl.active);
BUG_ON(dmabuf->cb_in.active || dmabuf->cb_out.active);
dmabuf->ops->release(dmabuf);
@@ -206,12 +206,15 @@ static void dma_buf_poll_cb(struct dma_fence *fence, struct dma_fence_cb *cb)
static __poll_t dma_buf_poll(struct file *file, poll_table *poll) {
- struct dma_buf_poll_cb_t *dcb; struct dma_buf *dmabuf; struct dma_resv *resv; struct dma_resv_list *fobj; struct dma_fence *fence_excl;
- __poll_t events; unsigned shared_count, seq;
struct dma_fence *fence;
__poll_t events;
int r, i;
dmabuf = file->private_data; if (!dmabuf || !dmabuf->resv)
@@ -225,99 +228,70 @@ static __poll_t dma_buf_poll(struct file *file, poll_table *poll) if (!events) return 0;
- dcb = events & EPOLLOUT ? &dmabuf->cb_out : &dmabuf->cb_in;
- /* Only queue a new one if we are not still waiting for the old one */
- spin_lock_irq(&dmabuf->poll.lock);
- if (dcb->active)
events = 0;
- else
dcb->active = events;
- spin_unlock_irq(&dmabuf->poll.lock);
- if (!events)
return 0;
retry: seq = read_seqcount_begin(&resv->seq); rcu_read_lock();
fobj = rcu_dereference(resv->fence);
- if (fobj)
- if (fobj && events & EPOLLOUT) shared_count = fobj->shared_count; else shared_count = 0;
fence_excl = dma_resv_excl_fence(resv);
if (read_seqcount_retry(&resv->seq, seq)) {
rcu_read_unlock();
goto retry;
}
if (fence_excl && (!(events & EPOLLOUT) || shared_count == 0)) {
struct dma_buf_poll_cb_t *dcb = &dmabuf->cb_excl;
__poll_t pevents = EPOLLIN;
if (shared_count == 0)
pevents |= EPOLLOUT;
spin_lock_irq(&dmabuf->poll.lock);
if (dcb->active) {
dcb->active |= pevents;
events &= ~pevents;
} else
dcb->active = pevents;
spin_unlock_irq(&dmabuf->poll.lock);
if (events & pevents) {
if (!dma_fence_get_rcu(fence_excl)) {
/* force a recheck */
events &= ~pevents;
dma_buf_poll_cb(NULL, &dcb->cb);
} else if (!dma_fence_add_callback(fence_excl, &dcb->cb,
dma_buf_poll_cb)) {
events &= ~pevents;
dma_fence_put(fence_excl);
} else {
/*
* No callback queued, wake up any additional
* waiters.
*/
dma_fence_put(fence_excl);
dma_buf_poll_cb(NULL, &dcb->cb);
}
- for (i = 0; i < shared_count; ++i) {
fence = rcu_dereference(fobj->shared[i]);
fence = dma_fence_get_rcu(fence);
if (!fence || read_seqcount_retry(&resv->seq, seq)) {
/* Concurrent modify detected, force re-check */
dma_fence_put(fence);
rcu_read_unlock();
}goto retry;
}
if ((events & EPOLLOUT) && shared_count > 0) {
struct dma_buf_poll_cb_t *dcb = &dmabuf->cb_shared;
int i;
/* Only queue a new callback if no event has fired yet */
spin_lock_irq(&dmabuf->poll.lock);
if (dcb->active)
events &= ~EPOLLOUT;
else
dcb->active = EPOLLOUT;
spin_unlock_irq(&dmabuf->poll.lock);
if (!(events & EPOLLOUT))
r = dma_fence_add_callback(fence, &dcb->cb, dma_buf_poll_cb);
dma_fence_put(fence);
if (!r) {
/* Callback queued */
events = 0; goto out;
}
- }
for (i = 0; i < shared_count; ++i) {
struct dma_fence *fence = rcu_dereference(fobj->shared[i]);
if (!dma_fence_get_rcu(fence)) {
/*
* fence refcount dropped to zero, this means
* that fobj has been freed
*
* call dma_buf_poll_cb and force a recheck!
*/
events &= ~EPOLLOUT;
dma_buf_poll_cb(NULL, &dcb->cb);
break;
}
if (!dma_fence_add_callback(fence, &dcb->cb,
dma_buf_poll_cb)) {
dma_fence_put(fence);
events &= ~EPOLLOUT;
break;
}
- fence = dma_resv_excl_fence(resv);
- if (fence) {
fence = dma_fence_get_rcu(fence);
if (!fence || read_seqcount_retry(&resv->seq, seq)) {
/* Concurrent modify detected, force re-check */ dma_fence_put(fence);
rcu_read_unlock();
goto retry;
- }
/* No callback queued, wake up any additional waiters. */
if (i == shared_count)
dma_buf_poll_cb(NULL, &dcb->cb);
r = dma_fence_add_callback(fence, &dcb->cb, dma_buf_poll_cb);
dma_fence_put(fence_excl);
if (!r) {
/* Callback queued */
events = 0;
goto out;
}
}
/* No callback queued, wake up any additional waiters. */
dma_buf_poll_cb(NULL, &dcb->cb);
out: rcu_read_unlock(); return events; @@ -562,8 +536,8 @@ struct dma_buf *dma_buf_export(const struct dma_buf_export_info *exp_info) dmabuf->owner = exp_info->owner; spin_lock_init(&dmabuf->name_lock); init_waitqueue_head(&dmabuf->poll);
- dmabuf->cb_excl.poll = dmabuf->cb_shared.poll = &dmabuf->poll;
- dmabuf->cb_excl.active = dmabuf->cb_shared.active = 0;
dmabuf->cb_in.poll = dmabuf->cb_out.poll = &dmabuf->poll;
dmabuf->cb_in.active = dmabuf->cb_out.active = 0;
if (!resv) { resv = (struct dma_resv *)&dmabuf[1];
diff --git a/include/linux/dma-buf.h b/include/linux/dma-buf.h index efdc56b9d95f..7e747ad54c81 100644 --- a/include/linux/dma-buf.h +++ b/include/linux/dma-buf.h @@ -329,7 +329,7 @@ struct dma_buf { wait_queue_head_t *poll;
__poll_t active;
- } cb_excl, cb_shared;
- } cb_in, cb_out;
};
/**
2.25.1