On 14.05.22 17:55, Demi Marie Obenour wrote:
In https://github.com/QubesOS/qubes-issues/issues/7481, a user reported that Xorg locked up when resizing a VM window. While I do not have the same hardware the user does and thus cannot reproduce the bug, the stack trace seems to indicate a deadlock between xen_gntdev and i915. It appears that gnttab_unmap_refs_sync() is waiting for i915 to free the pages, while i915 is waiting for the MMU notifier that called gnttab_unmap_refs_sync() to return. Result: deadlock.
The problem appears to be that a mapped grant in PV mode will stay in the “invalidating” state until it is freed. While MMU notifiers are allowed to sleep, it appears that they cannot wait for the page to be freed, as is happening here. That said, I am not very familiar with this code, so my diagnosis might be incorrect.
All I can say for now is that your patch seems to be introducing a use after free issue, as the parameters of the delayed work might get freed now before the delayed work is being executed.
I don't know why this is happening only with rather recent kernels, as the last gntdev changes in this area have been made in kernel 4.13.
I'd suggest to look at i915, as quite some work has happened in the code visible in your stack backtraces rather recently. Maybe it would be possible to free the pages in i915 before calling the MMU notifier?
Adding the i915 maintainers.
Juergen
On Mon, May 16, 2022 at 08:48:17AM +0200, Juergen Gross wrote:
On 14.05.22 17:55, Demi Marie Obenour wrote:
In https://github.com/QubesOS/qubes-issues/issues/7481, a user reported that Xorg locked up when resizing a VM window. While I do not have the same hardware the user does and thus cannot reproduce the bug, the stack trace seems to indicate a deadlock between xen_gntdev and i915. It appears that gnttab_unmap_refs_sync() is waiting for i915 to free the pages, while i915 is waiting for the MMU notifier that called gnttab_unmap_refs_sync() to return. Result: deadlock.
The problem appears to be that a mapped grant in PV mode will stay in the “invalidating” state until it is freed. While MMU notifiers are allowed to sleep, it appears that they cannot wait for the page to be freed, as is happening here. That said, I am not very familiar with this code, so my diagnosis might be incorrect.
All I can say for now is that your patch seems to be introducing a use after free issue, as the parameters of the delayed work might get freed now before the delayed work is being executed.
I figured it was wrong, not least because I don’t think it compiles (invalid use of void value). That said, the current behavior is quite suspicious to me. For one, it appears that munmap() on a grant in a PV domain will not return until nobody else is using the page. This is not what I would expect, and I can easily imagine it causing deadlocks in userspace. Instead, I would expect for gntdev to automatically release the grant when the reference count hits zero. This would also allow for the same grant to be mapped in multiple processes, and might even unlock DMA-BUF support.
I don't know why this is happening only with rather recent kernels, as the last gntdev changes in this area have been made in kernel 4.13.
I'd suggest to look at i915, as quite some work has happened in the code visible in your stack backtraces rather recently. Maybe it would be possible to free the pages in i915 before calling the MMU notifier?
While I agree that the actual problem is almost certainly in i915, the gntdev code does appear rather fragile. Since so few people use i915 + Xen, problems with the combination generally don’t show up until some Qubes user makes a bug report, which isn’t great. It would be better if Xen didn’t introduce requirements on other kernel code that did not hold when not running on Xen.
In this case, if it is actually an invariant that one must not call MMU notifiers for pages that are still in use, it would be better if this was caught by a WARN_ON() or BUG_ON() in the core memory management code. That would have found the bug instantly and deterministically on all platforms, whereas the current failure is nondeterministic and only happens under Xen.
I also wonder if this is a bug in the core MMU notifier infrastructure. My reading of the mmu_interval_notifier_remove() documentation is that it should only wait for the specific notifier being removed to finish, not for all notifiers to finish. Adding the memory management maintainers.
On Mon, May 16, 2022 at 10:00:07AM -0400, Demi Marie Obenour wrote:
On Mon, May 16, 2022 at 08:48:17AM +0200, Juergen Gross wrote:
On 14.05.22 17:55, Demi Marie Obenour wrote:
In https://github.com/QubesOS/qubes-issues/issues/7481, a user reported that Xorg locked up when resizing a VM window. While I do not have the same hardware the user does and thus cannot reproduce the bug, the stack trace seems to indicate a deadlock between xen_gntdev and i915. It appears that gnttab_unmap_refs_sync() is waiting for i915 to free the pages, while i915 is waiting for the MMU notifier that called gnttab_unmap_refs_sync() to return. Result: deadlock.
The problem appears to be that a mapped grant in PV mode will stay in the “invalidating” state until it is freed. While MMU notifiers are allowed to sleep, it appears that they cannot wait for the page to be freed, as is happening here. That said, I am not very familiar with this code, so my diagnosis might be incorrect.
All I can say for now is that your patch seems to be introducing a use after free issue, as the parameters of the delayed work might get freed now before the delayed work is being executed.
I figured it was wrong, not least because I don’t think it compiles (invalid use of void value). That said, the current behavior is quite suspicious to me. For one, it appears that munmap() on a grant in a PV domain will not return until nobody else is using the page. This is not what I would expect, and I can easily imagine it causing deadlocks in userspace. Instead, I would expect for gntdev to automatically release the grant when the reference count hits zero. This would also allow for the same grant to be mapped in multiple processes, and might even unlock DMA-BUF support.
I don't know why this is happening only with rather recent kernels, as the last gntdev changes in this area have been made in kernel 4.13.
I'd suggest to look at i915, as quite some work has happened in the code visible in your stack backtraces rather recently. Maybe it would be possible to free the pages in i915 before calling the MMU notifier?
While I agree that the actual problem is almost certainly in i915, the gntdev code does appear rather fragile. Since so few people use i915 + Xen, problems with the combination generally don’t show up until some Qubes user makes a bug report, which isn’t great. It would be better if Xen didn’t introduce requirements on other kernel code that did not hold when not running on Xen.
In this case, if it is actually an invariant that one must not call MMU notifiers for pages that are still in use, it would be better if this was caught by a WARN_ON() or BUG_ON() in the core memory management code. That would have found the bug instantly and deterministically on all platforms, whereas the current failure is nondeterministic and only happens under Xen.
I also wonder if this is a bug in the core MMU notifier infrastructure. My reading of the mmu_interval_notifier_remove() documentation is that it should only wait for the specific notifier being removed to finish, not for all notifiers to finish. Adding the memory management maintainers.
Also adding the kernel regression tracker.
#regzbot introduced v5.16..v5.17.4
On Mon, May 16, 2022 at 08:48:17AM +0200, Juergen Gross wrote:
On 14.05.22 17:55, Demi Marie Obenour wrote:
In https://github.com/QubesOS/qubes-issues/issues/7481, a user reported that Xorg locked up when resizing a VM window. While I do not have the same hardware the user does and thus cannot reproduce the bug, the stack trace seems to indicate a deadlock between xen_gntdev and i915. It appears that gnttab_unmap_refs_sync() is waiting for i915 to free the pages, while i915 is waiting for the MMU notifier that called gnttab_unmap_refs_sync() to return. Result: deadlock.
The problem appears to be that a mapped grant in PV mode will stay in the “invalidating” state until it is freed. While MMU notifiers are allowed to sleep, it appears that they cannot wait for the page to be freed, as is happening here. That said, I am not very familiar with this code, so my diagnosis might be incorrect.
All I can say for now is that your patch seems to be introducing a use after free issue, as the parameters of the delayed work might get freed now before the delayed work is being executed.
I don't know why this is happening only with rather recent kernels, as the last gntdev changes in this area have been made in kernel 4.13.
I'd suggest to look at i915, as quite some work has happened in the code visible in your stack backtraces rather recently. Maybe it would be possible to free the pages in i915 before calling the MMU notifier?
Honestly, I would rather fix this in gntdev, regardless of where the actual bug lies. GPU drivers get little testing under Xen, so if something like this happens again, it is likely to remain undiscovered until a Qubes user files a bug report. This results in a bad experience for Qubes users. I would much rather code that works on bare metal work in Xen dom0 too. I have had random hangs in the past (with various kernel versions) that might be due to similar problems.
Furthermore, similar problems can arise whenever a driver removes an MMU notifier on userspace pages that someone else has references to. It is hard for me to see how this is the fault of the driver that removed the MMU notifier. I find it much more plausible that the correct fix is on the Xen side: allocate the delayed work parameters on the heap, and free them after the work is finished. This eliminates this entire class of bugs.
dri-devel@lists.freedesktop.org