From b015cfa3c152146814a8ddbff3e46ba601e58340 Mon Sep 17 00:00:00 2001 From: Taylor R Campbell Date: Fri, 10 Aug 2018 05:51:54 +0000 Subject: [PATCH 1/6] Teach usb_rem_task to return whether removed from queue or not. --- sys/dev/usb/usb.c | 10 +++++++--- sys/dev/usb/usbdi.h | 2 +- 2 files changed, 8 insertions(+), 4 deletions(-) diff --git a/sys/dev/usb/usb.c b/sys/dev/usb/usb.c index 4627a5a322ff..717eb0a1c70b 100644 --- a/sys/dev/usb/usb.c +++ b/sys/dev/usb/usb.c @@ -431,14 +431,16 @@ usb_add_task(struct usbd_device *dev, struct usb_task *task, int queue) /* * usb_rem_task(dev, task) * - * If task is queued to run, remove it from the queue. + * If task is queued to run, remove it from the queue. Return + * true if it successfully removed the task from the queue, false + * if not. * * Caller is _not_ guaranteed that the task is not running when * this is done. * * Never sleeps. */ -void +bool usb_rem_task(struct usbd_device *dev, struct usb_task *task) { unsigned queue; @@ -452,10 +454,12 @@ usb_rem_task(struct usbd_device *dev, struct usb_task *task) TAILQ_REMOVE(&taskq->tasks, task, next); task->queue = USB_NUM_TASKQS; mutex_exit(&taskq->lock); - break; + return true; /* removed from the queue */ } mutex_exit(&taskq->lock); } + + return false; /* was not removed from the queue */ } /* diff --git a/sys/dev/usb/usbdi.h b/sys/dev/usb/usbdi.h index 6c228d6119cf..f837ddeae84b 100644 --- a/sys/dev/usb/usbdi.h +++ b/sys/dev/usb/usbdi.h @@ -216,7 +216,7 @@ struct usb_task { #define USB_TASKQ_MPSAFE 0x80 void usb_add_task(struct usbd_device *, struct usb_task *, int); -void usb_rem_task(struct usbd_device *, struct usb_task *); +bool usb_rem_task(struct usbd_device *, struct usb_task *); bool usb_rem_task_wait(struct usbd_device *, struct usb_task *, int, kmutex_t *); #define usb_init_task(t, f, a, fl) ((t)->fun = (f), (t)->arg = (a), (t)->queue = USB_NUM_TASKQS, (t)->flags = (fl)) From b93327d3ddb79107712d346d61828ec58e47a6d0 Mon Sep 17 00:00:00 2001 From: Taylor R Campbell Date: Fri, 10 Aug 2018 05:56:42 +0000 Subject: [PATCH 2/6] New function usb_task_pending for diagnostic assertions. Usable only for negative diagnostic assertions: KASSERT(!usb_task_pending(dev, task)) If you can think of a better name for this than !usb_task_pending, I'm all ears. --- sys/dev/usb/usb.c | 17 +++++++++++++++++ sys/dev/usb/usbdi.h | 1 + 2 files changed, 18 insertions(+) diff --git a/sys/dev/usb/usb.c b/sys/dev/usb/usb.c index 717eb0a1c70b..b8b0802d302e 100644 --- a/sys/dev/usb/usb.c +++ b/sys/dev/usb/usb.c @@ -530,6 +530,23 @@ usb_rem_task_wait(struct usbd_device *dev, struct usb_task *task, int queue, return removed; } +/* + * usb_task_pending(dev, task) + * + * True if task is queued, false if not. Note that if task is + * already running, it is not considered queued. + * + * For _negative_ diagnostic assertions only: + * + * KASSERT(!usb_task_pending(dev, task)); + */ +bool +usb_task_pending(struct usbd_device *dev, struct usb_task *task) +{ + + return task->queue != USB_NUM_TASKQS; +} + void usb_event_thread(void *arg) { diff --git a/sys/dev/usb/usbdi.h b/sys/dev/usb/usbdi.h index f837ddeae84b..0590af8fae5d 100644 --- a/sys/dev/usb/usbdi.h +++ b/sys/dev/usb/usbdi.h @@ -219,6 +219,7 @@ void usb_add_task(struct usbd_device *, struct usb_task *, int); bool usb_rem_task(struct usbd_device *, struct usb_task *); bool usb_rem_task_wait(struct usbd_device *, struct usb_task *, int, kmutex_t *); +bool usb_task_pending(struct usbd_device *, struct usb_task *); #define usb_init_task(t, f, a, fl) ((t)->fun = (f), (t)->arg = (a), (t)->queue = USB_NUM_TASKQS, (t)->flags = (fl)) struct usb_devno { From 92d4ae4f2d1f1c72791d69309243b8e7ac14baa1 Mon Sep 17 00:00:00 2001 From: Taylor R Campbell Date: Fri, 10 Aug 2018 05:57:48 +0000 Subject: [PATCH 3/6] Nothing guarantees xfer's timeout has completed. Wait for it when we free the xfer. --- sys/dev/usb/usbdi.c | 14 ++++++++------ 1 file changed, 8 insertions(+), 6 deletions(-) diff --git a/sys/dev/usb/usbdi.c b/sys/dev/usb/usbdi.c index 4510d2b5ae39..b1ffeaccbe1f 100644 --- a/sys/dev/usb/usbdi.c +++ b/sys/dev/usb/usbdi.c @@ -494,12 +494,14 @@ usbd_free_xfer(struct usbd_xfer *xfer) if (xfer->ux_buf) { usbd_free_buffer(xfer); } -#if defined(DIAGNOSTIC) - if (callout_pending(&xfer->ux_callout)) { - callout_stop(&xfer->ux_callout); - printf("usbd_free_xfer: timeout_handle pending\n"); - } -#endif + + /* Wait for any straggling timeout to complete. */ + mutex_enter(xfer->ux_bus->ub_lock); + callout_halt(&xfer->ux_callout, xfer->ux_bus->ub_lock); + usb_rem_task_wait(xfer->ux_pipe->up_dev, &xfer->ux_aborttask, + USB_TASKQ_HC, xfer->ux_bus->ub_lock); + mutex_exit(xfer->ux_bus->ub_lock); + cv_destroy(&xfer->ux_cv); xfer->ux_bus->ub_methods->ubm_freex(xfer->ux_bus, xfer); return USBD_NORMAL_COMPLETION; From 47e40f0f4175fea5a4f5c37f5a5dd9eb803efab5 Mon Sep 17 00:00:00 2001 From: Taylor R Campbell Date: Fri, 10 Aug 2018 05:59:03 +0000 Subject: [PATCH 4/6] New xfer state variables ux_timeout_set and ux_timeout_reset. These are needed because: - The host controller interrupt cannot wait for the callout or task to finish running. - Nothing in the USBD API as is waits for the callout or task to finish running. - Callers expect to be able to resubmit USB xfers from xfer callbacks without waiting for anything to finish running. The variable ux_timeout_set can be used by a host controller to decide on submission whether to schedule the callout or to ask an already-scheduled callout or already-queued task to reschedule the callout, by setting the variable ux_timeout_reset to true. When the callout or task runs and sees that ux_timeout_reset is true, rather than queue the task or abort the xfer, it can instead just schedule the callout anew. --- sys/dev/usb/usbdi.c | 1 + sys/dev/usb/usbdivar.h | 13 +++++++++++++ 2 files changed, 14 insertions(+) diff --git a/sys/dev/usb/usbdi.c b/sys/dev/usb/usbdi.c index b1ffeaccbe1f..f641f5d5c9ce 100644 --- a/sys/dev/usb/usbdi.c +++ b/sys/dev/usb/usbdi.c @@ -497,6 +497,7 @@ usbd_free_xfer(struct usbd_xfer *xfer) /* Wait for any straggling timeout to complete. */ mutex_enter(xfer->ux_bus->ub_lock); + xfer->ux_timeout_reset = false; /* do not resuscitate */ callout_halt(&xfer->ux_callout, xfer->ux_bus->ub_lock); usb_rem_task_wait(xfer->ux_pipe->up_dev, &xfer->ux_aborttask, USB_TASKQ_HC, xfer->ux_bus->ub_lock); diff --git a/sys/dev/usb/usbdivar.h b/sys/dev/usb/usbdivar.h index 245cb72dfc9c..a161b504b3d3 100644 --- a/sys/dev/usb/usbdivar.h +++ b/sys/dev/usb/usbdivar.h @@ -288,6 +288,19 @@ struct usbd_xfer { struct usb_task ux_aborttask; struct callout ux_callout; + + /* + * Protected by bus lock. + * + * - ux_timeout_set: The timeout is scheduled as a callout or + * usb task, and has not yet acquired the bus lock. + * + * - ux_timeout_reset: The xfer completed, and was resubmitted + * before the callout or task was able to acquire the bus + * lock, so one or the other needs to schedule a new callout. + */ + bool ux_timeout_set; + bool ux_timeout_reset; }; void usbd_init(void); From 577fe830a370d5f329a22c3daf2e107a0702eae5 Mon Sep 17 00:00:00 2001 From: Taylor R Campbell Date: Fri, 10 Aug 2018 06:04:07 +0000 Subject: [PATCH 5/6] Fix steady state of timeouts in ehci. This is complicated because: 1. There are three ways that an xfer can be completed: (a) hardware interrupt completes xfer (b) software decision aborts xfer with USBD_CANCELLED (c) timeout aborts xfer with USBD_TIMEOUT 2. The timeout abort can't be done in callout because ehci_sync_hc, called unconditionally by ehci_abort_xfer to wait until the device has finished using any references to the xfer, may sleep. So we have to schedule a callout that, when run, will schedule a usb_task. 3. The hardware completion interrupt can't sleep waiting for a callout or task to finish -- can't use callout_halt or usb_rem_task_wait. So the callout and usb_task must be able to run _after_ the hardware completion interrupt, and recognize that they're late to the party. (Note, though, that usbd_free_xfer does wait for the callout and task to complete, so there's no danger they may use themselves after free.) 4. The xfer may resubmitted -- and the timeout may be rescheduled -- immediately after the hardware completion interrupt, _while_ the callout and/or usb_task may still be scheduled. Specifically, we may have the following sequence of events: (a) hardware completion interrupt (b) callout or usb_task fires (c) driver resubmits xfer (d) callout or usb_task acquires lock and looks around dazed and bewildered at the firehose of events like reading the news in 2019 The mechanism for sorting this out is that we have two bits of state: - xfer->ux_timeout_set informs the driver, when submitting an xfer and setting up its timeout, whether either the callout or usb_task is already scheduled or not. - xfer->ux_timeout_reset informs the callout or usb_task whether it should reschedule the callout, because the xfer got resubmitted, or not. --- sys/dev/usb/ehci.c | 326 +++++++++++++++++++++++++++++++++++++++------ 1 file changed, 284 insertions(+), 42 deletions(-) diff --git a/sys/dev/usb/ehci.c b/sys/dev/usb/ehci.c index 028def75a88d..ae8e1dbbdf4c 100644 --- a/sys/dev/usb/ehci.c +++ b/sys/dev/usb/ehci.c @@ -168,6 +168,9 @@ Static void ehci_check_sitd_intr(ehci_softc_t *, struct ehci_xfer *, Static void ehci_idone(struct ehci_xfer *, ex_completeq_t *); Static void ehci_timeout(void *); Static void ehci_timeout_task(void *); +Static bool ehci_probe_timeout(struct usbd_xfer *); +Static void ehci_schedule_timeout(struct usbd_xfer *); +Static void ehci_cancel_timeout_async(struct usbd_xfer *); Static void ehci_intrlist_timeout(void *); Static void ehci_doorbell(void *); Static void ehci_pcd(void *); @@ -1057,13 +1060,11 @@ ehci_idone(struct ehci_xfer *ex, ex_completeq_t *cq) } /* - * Cancel the timeout and the task, which have not yet - * run. If they have already fired, at worst they are - * waiting for the lock. They will see that the xfer - * is no longer in progress and give up. + * We are completing the xfer. Cancel the timeout if we can, + * but only asynchronously. See ehci_cancel_timeout_async for + * why we need not wait for the callout or task here. */ - callout_stop(&xfer->ux_callout); - usb_rem_task(xfer->ux_pipe->up_dev, &xfer->ux_aborttask); + ehci_cancel_timeout_async(xfer); #ifdef DIAGNOSTIC #ifdef EHCI_DEBUG @@ -3235,35 +3236,38 @@ ehci_abort_xfer(struct usbd_xfer *xfer, usbd_status status) KASSERT(mutex_owned(&sc->sc_lock)); ASSERT_SLEEPABLE(); + /* + * Nobody else can set this status: only one caller can time + * out, and only one caller can synchronously abort. So the + * status can't be the status we're trying to set this to. + */ + KASSERT(xfer->ux_status != status); + + /* + * If host controller or timer interrupt has completed it, too + * late to abort. Forget about it. + */ + if (xfer->ux_status != USBD_IN_PROGRESS) + return; + if (status == USBD_CANCELLED) { /* - * We are synchronously aborting. Try to stop the - * callout and task, but if we can't, wait for them to - * complete. + * We are synchronously aborting. Cancel the timeout + * if we can, but only asynchronously. See + * ehci_cancel_timeout_async for why we need to wait + * for the callout or task here. */ - callout_halt(&xfer->ux_callout, &sc->sc_lock); - usb_rem_task_wait(xfer->ux_pipe->up_dev, &xfer->ux_aborttask, - USB_TASKQ_HC, &sc->sc_lock); + ehci_cancel_timeout_async(xfer); } else { - /* Otherwise, we are timing out. */ + /* + * Otherwise, we are timing out. No action needed to + * cancel the timeout because we _are_ the timeout. + * This case should happen only via the timeout task, + * invoked via the callout. + */ KASSERT(status == USBD_TIMEOUT); } - /* - * The xfer cannot have been cancelled already. It is the - * responsibility of the caller of usbd_abort_pipe not to try - * to abort a pipe multiple times, whether concurrently or - * sequentially. - */ - KASSERT(xfer->ux_status != USBD_CANCELLED); - - /* Only the timeout, which runs only once, can time it out. */ - KASSERT(xfer->ux_status != USBD_TIMEOUT); - - /* If anyone else beat us, we're done. */ - if (xfer->ux_status != USBD_IN_PROGRESS) - return; - /* We beat everyone else. Claim the status. */ xfer->ux_status = status; @@ -3473,6 +3477,16 @@ dying: KASSERT(mutex_owned(&sc->sc_lock)); } +/* + * ehci_timeout(xfer) + * + * Called at IPL_SOFTCLOCK when too much time has elapsed waiting + * for xfer to complete. Since we can't abort the xfer at + * IPL_SOFTCLOCK, defer to a usb_task to run it in thread context, + * unless the xfer has completed or aborted concurrently -- and if + * the xfer has also been resubmitted, take care of rescheduling + * the callout. + */ Static void ehci_timeout(void *addr) { @@ -3489,12 +3503,38 @@ ehci_timeout(void *addr) } #endif + /* Acquire the lock so we can transition the timeout state. */ mutex_enter(&sc->sc_lock); - if (!sc->sc_dying && xfer->ux_status == USBD_IN_PROGRESS) + + /* + * Use ehci_probe_timeout to check whether the timeout is still + * valid, or to reschedule the callout if necessary. If it is + * still valid, schedule the task. + */ + if (ehci_probe_timeout(xfer)) usb_add_task(dev, &xfer->ux_aborttask, USB_TASKQ_HC); + + /* + * Notify ehci_cancel_timeout_async that we may have scheduled + * the task. This causes callout_invoking to return false in + * ehci_cancel_timeout_async so that it can tell which stage in + * the callout->task->abort process we're at. + */ + callout_ack(&xfer->ux_callout); + + /* All done -- release the lock. */ mutex_exit(&sc->sc_lock); } +/* + * ehci_timeout_task(xfer) + * + * Called in thread context when too much time has elapsed waiting + * for xfer to complete. Issue ehci_abort_xfer(USBD_TIMEOUT), + * unless the xfer has completed or aborted concurrently -- and if + * the xfer has also been resubmitted, take care of rescheduling + * the callout. + */ Static void ehci_timeout_task(void *addr) { @@ -3505,11 +3545,222 @@ ehci_timeout_task(void *addr) DPRINTF("xfer=%#jx", (uintptr_t)xfer, 0, 0, 0); + /* Acquire the lock so we can transition the timeout state. */ mutex_enter(&sc->sc_lock); - ehci_abort_xfer(xfer, USBD_TIMEOUT); + + /* + * Use ehci_probe_timeout to check whether the timeout is still + * valid, or to reschedule the callout if necessary. If it is + * still valid, schedule the task. + */ + if (ehci_probe_timeout(xfer)) + ehci_abort_xfer(xfer, USBD_TIMEOUT); + + /* All done -- release the lock. */ mutex_exit(&sc->sc_lock); } +/* + * ehci_probe_timeout(xfer) + * + * Probe the status of xfer's timeout. Acknowledge and process a + * request to reschedule. Return true if the timeout is still + * valid and the caller should take further action (queueing a + * task or aborting the xfer), false if it must stop here. + */ +Static bool +ehci_probe_timeout(struct usbd_xfer *xfer) +{ + EHCIHIST_FUNC(); EHCIHIST_CALLED(); + struct ehci_softc *sc = EHCI_XFER2SC(xfer); + bool valid; + + KASSERT(sc->sc_bus.ub_usepolling || mutex_owned(&sc->sc_lock)); + + /* The timeout must be set. */ + KASSERT(xfer->ux_timeout_set); + + /* + * Neither callout nor task may be pending; they execute + * alternately in lock step. + */ + KASSERT(!callout_pending(&xfer->ux_callout)); + KASSERT(!usb_task_pending(xfer->ux_pipe->up_dev, &xfer->ux_aborttask)); + + /* There are a few cases... */ + if (sc->sc_dying) { + /* Host controller dying. Drop it all on the floor. */ + xfer->ux_timeout_set = false; + xfer->ux_timeout_reset = false; + valid = false; + } else if (xfer->ux_timeout_reset) { + /* + * The xfer completed _and_ got resubmitted while we + * waited for the lock. Acknowledge the request to + * reschedule, and reschedule it if there is a timeout + * and the bus is not polling. + */ + xfer->ux_timeout_reset = false; + if (xfer->ux_timeout && !sc->sc_bus.ub_usepolling) { + KASSERT(xfer->ux_timeout_set); + callout_schedule(&xfer->ux_callout, + mstohz(xfer->ux_timeout)); + } else { + /* No more callout or task scheduled. */ + xfer->ux_timeout_set = false; + } + valid = false; + } else if (xfer->ux_status != USBD_IN_PROGRESS) { + /* + * The xfer has completed by hardware completion or by + * software abort, and has not been resubmitted, so the + * timeout must be unset, and is no longer valid for + * the caller. + */ + xfer->ux_timeout_set = false; + valid = false; + } else { + /* + * The xfer has not yet completed, so the timeout is + * valid. + */ + valid = true; + } + + /* Any reset must have been processed. */ + KASSERT(!xfer->ux_timeout_reset); + + /* + * Either we claim the timeout is set, or the callout is idle. + * If the timeout is still set, we may be handing off to the + * task instead, so this is an if but not an iff. + */ + KASSERT(xfer->ux_timeout_set || !callout_pending(&xfer->ux_callout)); + + /* + * The task must be idle now. + * + * - If the caller is the callout, _and_ the timeout is still + * valid, the caller will schedule it, but it hasn't been + * scheduled yet. (If the timeout is not valid, the task + * should not be scheduled.) + * + * - If the caller is the task, it cannot be scheduled again + * until the callout runs again, which won't happen until we + * next release the lock. + */ + KASSERT(!usb_task_pending(xfer->ux_pipe->up_dev, &xfer->ux_aborttask)); + + KASSERT(sc->sc_bus.ub_usepolling || mutex_owned(&sc->sc_lock)); + + return valid; +} + +/* + * ehci_schedule_timeout(xfer) + * + * Ensure that xfer has a timeout. If the callout is already + * queued or the task is already running, request that they + * reschedule the callout. If not, and if we're not polling, + * schedule the callout anew. + */ +Static void +ehci_schedule_timeout(struct usbd_xfer *xfer) +{ + EHCIHIST_FUNC(); EHCIHIST_CALLED(); + ehci_softc_t *sc = EHCI_XFER2SC(xfer); + + KASSERT(sc->sc_bus.ub_usepolling || mutex_owned(&sc->sc_lock)); + + if (xfer->ux_timeout_set) { + /* + * Callout or task has fired from a prior completed + * xfer but has not yet noticed that the xfer is done. + * Ask it to reschedule itself to ux_timeout. + */ + xfer->ux_timeout_reset = true; + } else if (xfer->ux_timeout && !sc->sc_bus.ub_usepolling) { + /* Callout is not scheduled. Schedule it. */ + KASSERT(!callout_pending(&xfer->ux_callout)); + callout_reset(&xfer->ux_callout, mstohz(xfer->ux_timeout), + ehci_timeout, xfer); + xfer->ux_timeout_set = true; + } + + KASSERT(sc->sc_bus.ub_usepolling || mutex_owned(&sc->sc_lock)); +} + +/* + * ehci_cancel_timeout_async(xfer) + * + * Cancel the callout and the task of xfer, which have not yet run + * to completion, but don't wait for the callout or task to finish + * running. + * + * If they have already fired, at worst they are waiting for the + * bus lock. They will see that the xfer is no longer in progress + * and give up, or they will see that the xfer has been + * resubmitted with a new timeout and reschedule the callout. + * + * If a resubmitted request completed so fast that the callout + * didn't have time to process a timer reset, just cancel the + * timer reset. + */ +Static void +ehci_cancel_timeout_async(struct usbd_xfer *xfer) +{ + EHCIHIST_FUNC(); EHCIHIST_CALLED(); + struct ehci_softc *sc = EHCI_XFER2SC(xfer); + + KASSERT(sc->sc_bus.ub_usepolling || mutex_owned(&sc->sc_lock)); + + KASSERT(xfer->ux_timeout_set); + xfer->ux_timeout_reset = false; + if (!callout_stop(&xfer->ux_callout)) { + /* + * We stopped the callout before it ran. The timeout + * is no longer set. + */ + xfer->ux_timeout_set = false; + } else if (callout_invoking(&xfer->ux_callout)) { + /* + * The callout has begun to run but it has not yet + * acquired the lock. The task cannot be queued yet, + * and the callout cannot have been rescheduled yet. + * + * By the time the callout acquires the lock, we will + * have transitioned from USBD_IN_PROGRESS to a + * completed status, and possibly also resubmitted the + * xfer and set xfer->ux_timeout_reset = true. In both + * cases, the callout will DTRT, so no further action + * is needed here. + */ + } else if (usb_rem_task(xfer->ux_pipe->up_dev, &xfer->ux_aborttask)) { + /* + * The callout had fired and scheduled the task, but we + * stopped the task before it could run. The timeout + * is therefore no longer set -- the next resubmission + * of the xfer must schedule a new timeout. + * + * The callout should not be be pending at this point: + * it is scheduled only under the lock, and only when + * xfer->ux_timeout_set is false, or by the callout or + * task itself when xfer->ux_timeout_reset is true. + */ + xfer->ux_timeout_set = false; + } + + /* + * The callout cannot be scheduled and the task cannot be + * queued at this point. Either we cancelled them, or they are + * already running and waiting for the bus lock. + */ + KASSERT(!callout_pending(&xfer->ux_callout)); + KASSERT(!usb_task_pending(xfer->ux_pipe->up_dev, &xfer->ux_aborttask)); + + KASSERT(sc->sc_bus.ub_usepolling || mutex_owned(&sc->sc_lock)); +} + /************************/ Static int @@ -3751,10 +4002,7 @@ ehci_device_ctrl_start(struct usbd_xfer *xfer) /* Insert qTD in QH list - also does usb_syncmem(sqh) */ ehci_set_qh_qtd(sqh, setup); - if (xfer->ux_timeout && !sc->sc_bus.ub_usepolling) { - callout_reset(&xfer->ux_callout, mstohz(xfer->ux_timeout), - ehci_timeout, xfer); - } + ehci_schedule_timeout(xfer); ehci_add_intr_list(sc, exfer); xfer->ux_status = USBD_IN_PROGRESS; if (!polling) @@ -3951,10 +4199,7 @@ ehci_device_bulk_start(struct usbd_xfer *xfer) /* also does usb_syncmem(sqh) */ ehci_set_qh_qtd(sqh, exfer->ex_sqtdstart); - if (xfer->ux_timeout && !sc->sc_bus.ub_usepolling) { - callout_reset(&xfer->ux_callout, mstohz(xfer->ux_timeout), - ehci_timeout, xfer); - } + ehci_schedule_timeout(xfer); ehci_add_intr_list(sc, exfer); xfer->ux_status = USBD_IN_PROGRESS; if (!polling) @@ -4168,10 +4413,7 @@ ehci_device_intr_start(struct usbd_xfer *xfer) /* also does usb_syncmem(sqh) */ ehci_set_qh_qtd(sqh, exfer->ex_sqtdstart); - if (xfer->ux_timeout && !sc->sc_bus.ub_usepolling) { - callout_reset(&xfer->ux_callout, mstohz(xfer->ux_timeout), - ehci_timeout, xfer); - } + ehci_schedule_timeout(xfer); ehci_add_intr_list(sc, exfer); xfer->ux_status = USBD_IN_PROGRESS; if (!polling) From 74eb173669b7567fd59bb353a414f7c640e51a53 Mon Sep 17 00:00:00 2001 From: Taylor R Campbell Date: Fri, 9 Aug 2019 06:55:22 +0000 Subject: [PATCH 6/6] Factor out HCI-independent xfer completion logic. New API for HCI drivers: - When submitting an xfer to hardware, call usbd_xfer_schedule_timeout(xfer). - On HCI completion interrupt for xfer completion: if (!usbd_xfer_trycomplete(xfer)) return; /* timed out or aborted, ignore it */ - In upm_abort methods, call usbd_xfer_abort(xfer). - New ubm_abortx method serves role of former *hci_abort_xfer, but without any logic for wrangling timeouts/callouts/tasks -- caller in usbd_xfer_abort has already handled them. - New ubm_dying method, returns true if the device is in the process of detaching, used by the timeout logic. Converted and compile-tested: - ehci - motg (XXX missing usbd_xfer_schedule_timeout?) - ohci - uhci - xhci Missing: - ahci (sys/arch/mips/adm5120/dev/ahci.c) - slhci (sys/dev/ic/sl811hs.c) - ugenhc (sys/rump/dev/lib/libugenhc/ugenhc.c) --- sys/dev/usb/ehci.c | 388 +++-------------------------------------- sys/dev/usb/motg.c | 66 ++++--- sys/dev/usb/ohci.c | 156 +++-------------- sys/dev/usb/uhci.c | 140 +++------------ sys/dev/usb/usbdi.c | 383 ++++++++++++++++++++++++++++++++++++++++ sys/dev/usb/usbdi.h | 5 + sys/dev/usb/usbdivar.h | 2 + sys/dev/usb/xhci.c | 166 ++++-------------- 8 files changed, 546 insertions(+), 760 deletions(-) diff --git a/sys/dev/usb/ehci.c b/sys/dev/usb/ehci.c index ae8e1dbbdf4c..d51315d79b0c 100644 --- a/sys/dev/usb/ehci.c +++ b/sys/dev/usb/ehci.c @@ -166,11 +166,6 @@ Static void ehci_check_itd_intr(ehci_softc_t *, struct ehci_xfer *, Static void ehci_check_sitd_intr(ehci_softc_t *, struct ehci_xfer *, ex_completeq_t *); Static void ehci_idone(struct ehci_xfer *, ex_completeq_t *); -Static void ehci_timeout(void *); -Static void ehci_timeout_task(void *); -Static bool ehci_probe_timeout(struct usbd_xfer *); -Static void ehci_schedule_timeout(struct usbd_xfer *); -Static void ehci_cancel_timeout_async(struct usbd_xfer *); Static void ehci_intrlist_timeout(void *); Static void ehci_doorbell(void *); Static void ehci_pcd(void *); @@ -180,6 +175,7 @@ Static struct usbd_xfer * Static void ehci_freex(struct usbd_bus *, struct usbd_xfer *); Static void ehci_get_lock(struct usbd_bus *, kmutex_t **); +Static bool ehci_dying(struct usbd_bus *); Static int ehci_roothub_ctrl(struct usbd_bus *, usb_device_request_t *, void *, int); @@ -281,7 +277,7 @@ Static void ehci_set_qh_qtd(ehci_soft_qh_t *, ehci_soft_qtd_t *); Static void ehci_sync_hc(ehci_softc_t *); Static void ehci_close_pipe(struct usbd_pipe *, ehci_soft_qh_t *); -Static void ehci_abort_xfer(struct usbd_xfer *, usbd_status); +Static void ehci_abortx(struct usbd_xfer *); #ifdef EHCI_DEBUG Static ehci_softc_t *theehci; @@ -322,6 +318,8 @@ Static const struct usbd_bus_methods ehci_bus_methods = { .ubm_dopoll = ehci_poll, .ubm_allocx = ehci_allocx, .ubm_freex = ehci_freex, + .ubm_abortx = ehci_abortx, + .ubm_dying = ehci_dying, .ubm_getlock = ehci_get_lock, .ubm_rhctrl = ehci_roothub_ctrl, }; @@ -1049,22 +1047,11 @@ ehci_idone(struct ehci_xfer *ex, ex_completeq_t *cq) DPRINTF("ex=%#jx", (uintptr_t)ex, 0, 0, 0); /* - * If software has completed it, either by cancellation - * or timeout, drop it on the floor. + * Try to claim this xfer for completion. If it has already + * completed or aborted, drop it on the floor. */ - if (xfer->ux_status != USBD_IN_PROGRESS) { - KASSERT(xfer->ux_status == USBD_CANCELLED || - xfer->ux_status == USBD_TIMEOUT); - DPRINTF("aborted xfer=%#jx", (uintptr_t)xfer, 0, 0, 0); + if (!usbd_xfer_trycomplete(xfer)) return; - } - - /* - * We are completing the xfer. Cancel the timeout if we can, - * but only asynchronously. See ehci_cancel_timeout_async for - * why we need not wait for the callout or task here. - */ - ehci_cancel_timeout_async(xfer); #ifdef DIAGNOSTIC #ifdef EHCI_DEBUG @@ -1539,9 +1526,6 @@ ehci_allocx(struct usbd_bus *bus, unsigned int nframes) if (xfer != NULL) { memset(xfer, 0, sizeof(struct ehci_xfer)); - /* Initialise this always so we can call remove on it. */ - usb_init_task(&xfer->ux_aborttask, ehci_timeout_task, xfer, - USB_TASKQ_MPSAFE); #ifdef DIAGNOSTIC struct ehci_xfer *ex = EHCI_XFER2EXFER(xfer); ex->ex_isdone = true; @@ -1569,6 +1553,14 @@ ehci_freex(struct usbd_bus *bus, struct usbd_xfer *xfer) pool_cache_put(sc->sc_xferpool, xfer); } +Static bool +ehci_dying(struct usbd_bus *bus) +{ + struct ehci_softc *sc = EHCI_BUS2SC(bus); + + return sc->sc_dying; +} + Static void ehci_get_lock(struct usbd_bus *bus, kmutex_t **lock) { @@ -3201,22 +3193,12 @@ ehci_close_pipe(struct usbd_pipe *pipe, ehci_soft_qh_t *head) } /* - * Cancel or timeout a device request. We have two cases to deal with - * - * 1) A driver wants to stop scheduled or inflight transfers - * 2) A transfer has timed out - * - * have (partially) happened since the hardware runs concurrently. - * - * Transfer state is protected by the bus lock and we set the transfer status - * as soon as either of the above happens (with bus lock held). - * - * Then we arrange for the hardware to tells us that it is not still + * Arrrange for the hardware to tells us that it is not still * processing the TDs by setting the QH halted bit and wait for the ehci * door bell */ Static void -ehci_abort_xfer(struct usbd_xfer *xfer, usbd_status status) +ehci_abortx(struct usbd_xfer *xfer) { EHCIHIST_FUNC(); EHCIHIST_CALLED(); struct ehci_pipe *epipe = EHCI_XFER2EPIPE(xfer); @@ -3228,48 +3210,14 @@ ehci_abort_xfer(struct usbd_xfer *xfer, usbd_status status) uint32_t qhstatus; int hit; - KASSERTMSG((status == USBD_CANCELLED || status == USBD_TIMEOUT), - "invalid status for abort: %d", (int)status); - DPRINTF("xfer=%#jx pipe=%#jx", (uintptr_t)xfer, (uintptr_t)epipe, 0, 0); KASSERT(mutex_owned(&sc->sc_lock)); ASSERT_SLEEPABLE(); - /* - * Nobody else can set this status: only one caller can time - * out, and only one caller can synchronously abort. So the - * status can't be the status we're trying to set this to. - */ - KASSERT(xfer->ux_status != status); - - /* - * If host controller or timer interrupt has completed it, too - * late to abort. Forget about it. - */ - if (xfer->ux_status != USBD_IN_PROGRESS) - return; - - if (status == USBD_CANCELLED) { - /* - * We are synchronously aborting. Cancel the timeout - * if we can, but only asynchronously. See - * ehci_cancel_timeout_async for why we need to wait - * for the callout or task here. - */ - ehci_cancel_timeout_async(xfer); - } else { - /* - * Otherwise, we are timing out. No action needed to - * cancel the timeout because we _are_ the timeout. - * This case should happen only via the timeout task, - * invoked via the callout. - */ - KASSERT(status == USBD_TIMEOUT); - } - - /* We beat everyone else. Claim the status. */ - xfer->ux_status = status; + KASSERTMSG((xfer->ux_status == USBD_CANCELLED || + xfer->ux_status == USBD_TIMEOUT), + "bad abort status: %d", xfer->ux_status); /* * If we're dying, skip the hardware action and just notify the @@ -3477,290 +3425,6 @@ dying: KASSERT(mutex_owned(&sc->sc_lock)); } -/* - * ehci_timeout(xfer) - * - * Called at IPL_SOFTCLOCK when too much time has elapsed waiting - * for xfer to complete. Since we can't abort the xfer at - * IPL_SOFTCLOCK, defer to a usb_task to run it in thread context, - * unless the xfer has completed or aborted concurrently -- and if - * the xfer has also been resubmitted, take care of rescheduling - * the callout. - */ -Static void -ehci_timeout(void *addr) -{ - EHCIHIST_FUNC(); EHCIHIST_CALLED(); - struct usbd_xfer *xfer = addr; - ehci_softc_t *sc = EHCI_XFER2SC(xfer); - struct usbd_device *dev = xfer->ux_pipe->up_dev; - - DPRINTF("xfer %#jx", (uintptr_t)xfer, 0, 0, 0); -#ifdef EHCI_DEBUG - if (ehcidebug >= 2) { - struct usbd_pipe *pipe = xfer->ux_pipe; - usbd_dump_pipe(pipe); - } -#endif - - /* Acquire the lock so we can transition the timeout state. */ - mutex_enter(&sc->sc_lock); - - /* - * Use ehci_probe_timeout to check whether the timeout is still - * valid, or to reschedule the callout if necessary. If it is - * still valid, schedule the task. - */ - if (ehci_probe_timeout(xfer)) - usb_add_task(dev, &xfer->ux_aborttask, USB_TASKQ_HC); - - /* - * Notify ehci_cancel_timeout_async that we may have scheduled - * the task. This causes callout_invoking to return false in - * ehci_cancel_timeout_async so that it can tell which stage in - * the callout->task->abort process we're at. - */ - callout_ack(&xfer->ux_callout); - - /* All done -- release the lock. */ - mutex_exit(&sc->sc_lock); -} - -/* - * ehci_timeout_task(xfer) - * - * Called in thread context when too much time has elapsed waiting - * for xfer to complete. Issue ehci_abort_xfer(USBD_TIMEOUT), - * unless the xfer has completed or aborted concurrently -- and if - * the xfer has also been resubmitted, take care of rescheduling - * the callout. - */ -Static void -ehci_timeout_task(void *addr) -{ - struct usbd_xfer *xfer = addr; - ehci_softc_t *sc = EHCI_XFER2SC(xfer); - - EHCIHIST_FUNC(); EHCIHIST_CALLED(); - - DPRINTF("xfer=%#jx", (uintptr_t)xfer, 0, 0, 0); - - /* Acquire the lock so we can transition the timeout state. */ - mutex_enter(&sc->sc_lock); - - /* - * Use ehci_probe_timeout to check whether the timeout is still - * valid, or to reschedule the callout if necessary. If it is - * still valid, schedule the task. - */ - if (ehci_probe_timeout(xfer)) - ehci_abort_xfer(xfer, USBD_TIMEOUT); - - /* All done -- release the lock. */ - mutex_exit(&sc->sc_lock); -} - -/* - * ehci_probe_timeout(xfer) - * - * Probe the status of xfer's timeout. Acknowledge and process a - * request to reschedule. Return true if the timeout is still - * valid and the caller should take further action (queueing a - * task or aborting the xfer), false if it must stop here. - */ -Static bool -ehci_probe_timeout(struct usbd_xfer *xfer) -{ - EHCIHIST_FUNC(); EHCIHIST_CALLED(); - struct ehci_softc *sc = EHCI_XFER2SC(xfer); - bool valid; - - KASSERT(sc->sc_bus.ub_usepolling || mutex_owned(&sc->sc_lock)); - - /* The timeout must be set. */ - KASSERT(xfer->ux_timeout_set); - - /* - * Neither callout nor task may be pending; they execute - * alternately in lock step. - */ - KASSERT(!callout_pending(&xfer->ux_callout)); - KASSERT(!usb_task_pending(xfer->ux_pipe->up_dev, &xfer->ux_aborttask)); - - /* There are a few cases... */ - if (sc->sc_dying) { - /* Host controller dying. Drop it all on the floor. */ - xfer->ux_timeout_set = false; - xfer->ux_timeout_reset = false; - valid = false; - } else if (xfer->ux_timeout_reset) { - /* - * The xfer completed _and_ got resubmitted while we - * waited for the lock. Acknowledge the request to - * reschedule, and reschedule it if there is a timeout - * and the bus is not polling. - */ - xfer->ux_timeout_reset = false; - if (xfer->ux_timeout && !sc->sc_bus.ub_usepolling) { - KASSERT(xfer->ux_timeout_set); - callout_schedule(&xfer->ux_callout, - mstohz(xfer->ux_timeout)); - } else { - /* No more callout or task scheduled. */ - xfer->ux_timeout_set = false; - } - valid = false; - } else if (xfer->ux_status != USBD_IN_PROGRESS) { - /* - * The xfer has completed by hardware completion or by - * software abort, and has not been resubmitted, so the - * timeout must be unset, and is no longer valid for - * the caller. - */ - xfer->ux_timeout_set = false; - valid = false; - } else { - /* - * The xfer has not yet completed, so the timeout is - * valid. - */ - valid = true; - } - - /* Any reset must have been processed. */ - KASSERT(!xfer->ux_timeout_reset); - - /* - * Either we claim the timeout is set, or the callout is idle. - * If the timeout is still set, we may be handing off to the - * task instead, so this is an if but not an iff. - */ - KASSERT(xfer->ux_timeout_set || !callout_pending(&xfer->ux_callout)); - - /* - * The task must be idle now. - * - * - If the caller is the callout, _and_ the timeout is still - * valid, the caller will schedule it, but it hasn't been - * scheduled yet. (If the timeout is not valid, the task - * should not be scheduled.) - * - * - If the caller is the task, it cannot be scheduled again - * until the callout runs again, which won't happen until we - * next release the lock. - */ - KASSERT(!usb_task_pending(xfer->ux_pipe->up_dev, &xfer->ux_aborttask)); - - KASSERT(sc->sc_bus.ub_usepolling || mutex_owned(&sc->sc_lock)); - - return valid; -} - -/* - * ehci_schedule_timeout(xfer) - * - * Ensure that xfer has a timeout. If the callout is already - * queued or the task is already running, request that they - * reschedule the callout. If not, and if we're not polling, - * schedule the callout anew. - */ -Static void -ehci_schedule_timeout(struct usbd_xfer *xfer) -{ - EHCIHIST_FUNC(); EHCIHIST_CALLED(); - ehci_softc_t *sc = EHCI_XFER2SC(xfer); - - KASSERT(sc->sc_bus.ub_usepolling || mutex_owned(&sc->sc_lock)); - - if (xfer->ux_timeout_set) { - /* - * Callout or task has fired from a prior completed - * xfer but has not yet noticed that the xfer is done. - * Ask it to reschedule itself to ux_timeout. - */ - xfer->ux_timeout_reset = true; - } else if (xfer->ux_timeout && !sc->sc_bus.ub_usepolling) { - /* Callout is not scheduled. Schedule it. */ - KASSERT(!callout_pending(&xfer->ux_callout)); - callout_reset(&xfer->ux_callout, mstohz(xfer->ux_timeout), - ehci_timeout, xfer); - xfer->ux_timeout_set = true; - } - - KASSERT(sc->sc_bus.ub_usepolling || mutex_owned(&sc->sc_lock)); -} - -/* - * ehci_cancel_timeout_async(xfer) - * - * Cancel the callout and the task of xfer, which have not yet run - * to completion, but don't wait for the callout or task to finish - * running. - * - * If they have already fired, at worst they are waiting for the - * bus lock. They will see that the xfer is no longer in progress - * and give up, or they will see that the xfer has been - * resubmitted with a new timeout and reschedule the callout. - * - * If a resubmitted request completed so fast that the callout - * didn't have time to process a timer reset, just cancel the - * timer reset. - */ -Static void -ehci_cancel_timeout_async(struct usbd_xfer *xfer) -{ - EHCIHIST_FUNC(); EHCIHIST_CALLED(); - struct ehci_softc *sc = EHCI_XFER2SC(xfer); - - KASSERT(sc->sc_bus.ub_usepolling || mutex_owned(&sc->sc_lock)); - - KASSERT(xfer->ux_timeout_set); - xfer->ux_timeout_reset = false; - if (!callout_stop(&xfer->ux_callout)) { - /* - * We stopped the callout before it ran. The timeout - * is no longer set. - */ - xfer->ux_timeout_set = false; - } else if (callout_invoking(&xfer->ux_callout)) { - /* - * The callout has begun to run but it has not yet - * acquired the lock. The task cannot be queued yet, - * and the callout cannot have been rescheduled yet. - * - * By the time the callout acquires the lock, we will - * have transitioned from USBD_IN_PROGRESS to a - * completed status, and possibly also resubmitted the - * xfer and set xfer->ux_timeout_reset = true. In both - * cases, the callout will DTRT, so no further action - * is needed here. - */ - } else if (usb_rem_task(xfer->ux_pipe->up_dev, &xfer->ux_aborttask)) { - /* - * The callout had fired and scheduled the task, but we - * stopped the task before it could run. The timeout - * is therefore no longer set -- the next resubmission - * of the xfer must schedule a new timeout. - * - * The callout should not be be pending at this point: - * it is scheduled only under the lock, and only when - * xfer->ux_timeout_set is false, or by the callout or - * task itself when xfer->ux_timeout_reset is true. - */ - xfer->ux_timeout_set = false; - } - - /* - * The callout cannot be scheduled and the task cannot be - * queued at this point. Either we cancelled them, or they are - * already running and waiting for the bus lock. - */ - KASSERT(!callout_pending(&xfer->ux_callout)); - KASSERT(!usb_task_pending(xfer->ux_pipe->up_dev, &xfer->ux_aborttask)); - - KASSERT(sc->sc_bus.ub_usepolling || mutex_owned(&sc->sc_lock)); -} - /************************/ Static int @@ -4002,7 +3666,7 @@ ehci_device_ctrl_start(struct usbd_xfer *xfer) /* Insert qTD in QH list - also does usb_syncmem(sqh) */ ehci_set_qh_qtd(sqh, setup); - ehci_schedule_timeout(xfer); + usbd_xfer_schedule_timeout(xfer); ehci_add_intr_list(sc, exfer); xfer->ux_status = USBD_IN_PROGRESS; if (!polling) @@ -4053,7 +3717,7 @@ ehci_device_ctrl_abort(struct usbd_xfer *xfer) EHCIHIST_FUNC(); EHCIHIST_CALLED(); DPRINTF("xfer=%#jx", (uintptr_t)xfer, 0, 0, 0); - ehci_abort_xfer(xfer, USBD_CANCELLED); + usbd_xfer_abort(xfer); } /* Close a device control pipe. */ @@ -4199,7 +3863,7 @@ ehci_device_bulk_start(struct usbd_xfer *xfer) /* also does usb_syncmem(sqh) */ ehci_set_qh_qtd(sqh, exfer->ex_sqtdstart); - ehci_schedule_timeout(xfer); + usbd_xfer_schedule_timeout(xfer); ehci_add_intr_list(sc, exfer); xfer->ux_status = USBD_IN_PROGRESS; if (!polling) @@ -4230,7 +3894,7 @@ ehci_device_bulk_abort(struct usbd_xfer *xfer) EHCIHIST_FUNC(); EHCIHIST_CALLED(); DPRINTF("xfer %#jx", (uintptr_t)xfer, 0, 0, 0); - ehci_abort_xfer(xfer, USBD_CANCELLED); + usbd_xfer_abort(xfer); } /* @@ -4413,7 +4077,7 @@ ehci_device_intr_start(struct usbd_xfer *xfer) /* also does usb_syncmem(sqh) */ ehci_set_qh_qtd(sqh, exfer->ex_sqtdstart); - ehci_schedule_timeout(xfer); + usbd_xfer_schedule_timeout(xfer); ehci_add_intr_list(sc, exfer); xfer->ux_status = USBD_IN_PROGRESS; if (!polling) @@ -4447,7 +4111,7 @@ ehci_device_intr_abort(struct usbd_xfer *xfer) * async doorbell. That's dependent on the async list, wheras * intr xfers are periodic, should not use this? */ - ehci_abort_xfer(xfer, USBD_CANCELLED); + usbd_xfer_abort(xfer); } Static void diff --git a/sys/dev/usb/motg.c b/sys/dev/usb/motg.c index f36b92cc0a53..255f0858440e 100644 --- a/sys/dev/usb/motg.c +++ b/sys/dev/usb/motg.c @@ -144,6 +144,7 @@ static void motg_softintr(void *); static struct usbd_xfer * motg_allocx(struct usbd_bus *, unsigned int); static void motg_freex(struct usbd_bus *, struct usbd_xfer *); +static bool motg_dying(struct usbd_bus *); static void motg_get_lock(struct usbd_bus *, kmutex_t **); static int motg_roothub_ctrl(struct usbd_bus *, usb_device_request_t *, void *, int); @@ -174,7 +175,7 @@ static void motg_device_data_read(struct usbd_xfer *); static void motg_device_data_write(struct usbd_xfer *); static void motg_device_clear_toggle(struct usbd_pipe *); -static void motg_device_xfer_abort(struct usbd_xfer *); +static void motg_abortx(struct usbd_xfer *); #define UBARR(sc) bus_space_barrier((sc)->sc_iot, (sc)->sc_ioh, 0, (sc)->sc_size, \ BUS_SPACE_BARRIER_READ|BUS_SPACE_BARRIER_WRITE) @@ -233,6 +234,8 @@ const struct usbd_bus_methods motg_bus_methods = { .ubm_dopoll = motg_poll, .ubm_allocx = motg_allocx, .ubm_freex = motg_freex, + .ubm_abortx = motg_abortx, + .ubm_dying = motg_dying, .ubm_getlock = motg_get_lock, .ubm_rhctrl = motg_roothub_ctrl, }; @@ -770,6 +773,14 @@ motg_freex(struct usbd_bus *bus, struct usbd_xfer *xfer) pool_cache_put(sc->sc_xferpool, xfer); } +static bool +motg_dying(struct usbd_bus *bus) +{ + struct motg_softc *sc = MOTG_BUS2SC(bus); + + return sc->sc_dying; +} + static void motg_get_lock(struct usbd_bus *bus, kmutex_t **lock) { @@ -1387,6 +1398,13 @@ motg_device_ctrl_intr_rx(struct motg_softc *sc) KASSERT(mutex_owned(&sc->sc_lock)); + /* + * Try to claim this xfer for completion. If it has already + * completed or aborted, drop it on the floor. + */ + if (!usbd_xfer_trycomplete(xfer)) + return; + KASSERT(ep->phase == DATA_IN || ep->phase == STATUS_IN); /* select endpoint 0 */ UWRITE1(sc, MUSB2_REG_EPINDEX, 0); @@ -1501,6 +1519,14 @@ motg_device_ctrl_intr_tx(struct motg_softc *sc) MOTGHIST_FUNC(); MOTGHIST_CALLED(); KASSERT(mutex_owned(&sc->sc_lock)); + + /* + * Try to claim this xfer for completion. If it has already + * completed or aborted, drop it on the floor. + */ + if (!usbd_xfer_trycomplete(xfer)) + return; + if (ep->phase == DATA_IN || ep->phase == STATUS_IN) { motg_device_ctrl_intr_rx(sc); return; @@ -1633,7 +1659,7 @@ motg_device_ctrl_abort(struct usbd_xfer *xfer) { MOTGHIST_FUNC(); MOTGHIST_CALLED(); - motg_device_xfer_abort(xfer); + usbd_xfer_abort(xfer); } /* Close a device control pipe */ @@ -2019,6 +2045,14 @@ motg_device_intr_tx(struct motg_softc *sc, int epnumber) MOTGHIST_FUNC(); MOTGHIST_CALLED(); KASSERT(mutex_owned(&sc->sc_lock)); + + /* + * Try to claim this xfer for completion. If it has already + * completed or aborted, drop it on the floor. + */ + if (!usbd_xfer_trycomplete(xfer)) + return; + KASSERT(ep->ep_number == epnumber); DPRINTFN(MD_BULK, " on ep %jd", epnumber, 0, 0, 0); @@ -2102,7 +2136,7 @@ motg_device_data_abort(struct usbd_xfer *xfer) MOTGHIST_FUNC(); MOTGHIST_CALLED(); - motg_device_xfer_abort(xfer); + usbd_xfer_abort(xfer); } /* Close a device control pipe */ @@ -2151,7 +2185,7 @@ motg_device_clear_toggle(struct usbd_pipe *pipe) /* Abort a device control request. */ static void -motg_device_xfer_abort(struct usbd_xfer *xfer) +motg_abortx(struct usbd_xfer *xfer) { MOTGHIST_FUNC(); MOTGHIST_CALLED(); uint8_t csr; @@ -2161,30 +2195,6 @@ motg_device_xfer_abort(struct usbd_xfer *xfer) KASSERT(mutex_owned(&sc->sc_lock)); ASSERT_SLEEPABLE(); - /* - * We are synchronously aborting. Try to stop the - * callout and task, but if we can't, wait for them to - * complete. - */ - callout_halt(&xfer->ux_callout, &sc->sc_lock); - usb_rem_task_wait(xfer->ux_pipe->up_dev, &xfer->ux_aborttask, - USB_TASKQ_HC, &sc->sc_lock); - - /* - * The xfer cannot have been cancelled already. It is the - * responsibility of the caller of usbd_abort_pipe not to try - * to abort a pipe multiple times, whether concurrently or - * sequentially. - */ - KASSERT(xfer->ux_status != USBD_CANCELLED); - - /* If anyone else beat us, we're done. */ - if (xfer->ux_status != USBD_IN_PROGRESS) - return; - - /* We beat everyone else. Claim the status. */ - xfer->ux_status = USBD_CANCELLED; - /* * If we're dying, skip the hardware action and just notify the * software that we're done. diff --git a/sys/dev/usb/ohci.c b/sys/dev/usb/ohci.c index de964ff98931..3855a41b1d6e 100644 --- a/sys/dev/usb/ohci.c +++ b/sys/dev/usb/ohci.c @@ -169,6 +169,7 @@ Static void ohci_device_isoc_enter(struct usbd_xfer *); Static struct usbd_xfer * ohci_allocx(struct usbd_bus *, unsigned int); Static void ohci_freex(struct usbd_bus *, struct usbd_xfer *); +Static bool ohci_dying(struct usbd_bus *); Static void ohci_get_lock(struct usbd_bus *, kmutex_t **); Static int ohci_roothub_ctrl(struct usbd_bus *, usb_device_request_t *, void *, int); @@ -213,12 +214,10 @@ Static void ohci_device_isoc_done(struct usbd_xfer *); Static usbd_status ohci_device_setintr(ohci_softc_t *, struct ohci_pipe *, int); -Static void ohci_timeout(void *); -Static void ohci_timeout_task(void *); Static void ohci_rhsc_enable(void *); Static void ohci_close_pipe(struct usbd_pipe *, ohci_soft_ed_t *); -Static void ohci_abort_xfer(struct usbd_xfer *, usbd_status); +Static void ohci_abortx(struct usbd_xfer *); Static void ohci_device_clear_toggle(struct usbd_pipe *); Static void ohci_noop(struct usbd_pipe *); @@ -287,6 +286,8 @@ Static const struct usbd_bus_methods ohci_bus_methods = { .ubm_dopoll = ohci_poll, .ubm_allocx = ohci_allocx, .ubm_freex = ohci_freex, + .ubm_abortx = ohci_abortx, + .ubm_dying = ohci_dying, .ubm_getlock = ohci_get_lock, .ubm_rhctrl = ohci_roothub_ctrl, }; @@ -1068,9 +1069,6 @@ ohci_allocx(struct usbd_bus *bus, unsigned int nframes) if (xfer != NULL) { memset(xfer, 0, sizeof(struct ohci_xfer)); - /* Initialise this always so we can call remove on it. */ - usb_init_task(&xfer->ux_aborttask, ohci_timeout_task, xfer, - USB_TASKQ_MPSAFE); #ifdef DIAGNOSTIC xfer->ux_state = XFER_BUSY; #endif @@ -1092,6 +1090,14 @@ ohci_freex(struct usbd_bus *bus, struct usbd_xfer *xfer) pool_cache_put(sc->sc_xferpool, xfer); } +Static bool +ohci_dying(struct usbd_bus *bus) +{ + ohci_softc_t *sc = OHCI_BUS2SC(bus); + + return sc->sc_dying; +} + Static void ohci_get_lock(struct usbd_bus *bus, kmutex_t **lock) { @@ -1463,23 +1469,11 @@ ohci_softintr(void *v) } /* - * If software has completed it, either by cancellation - * or timeout, drop it on the floor. + * Try to claim this xfer for completion. If it has + * already completed or aborted, drop it on the floor. */ - if (xfer->ux_status != USBD_IN_PROGRESS) { - KASSERT(xfer->ux_status == USBD_CANCELLED || - xfer->ux_status == USBD_TIMEOUT); + if (!usbd_xfer_trycomplete(xfer)) continue; - } - - /* - * Cancel the timeout and the task, which have not yet - * run. If they have already fired, at worst they are - * waiting for the lock. They will see that the xfer - * is no longer in progress and give up. - */ - callout_stop(&xfer->ux_callout); - usb_rem_task(xfer->ux_pipe->up_dev, &xfer->ux_aborttask); len = std->len; if (std->td.td_cbp != 0) @@ -1552,23 +1546,11 @@ ohci_softintr(void *v) continue; /* - * If software has completed it, either by cancellation - * or timeout, drop it on the floor. + * Try to claim this xfer for completion. If it has + * already completed or aborted, drop it on the floor. */ - if (xfer->ux_status != USBD_IN_PROGRESS) { - KASSERT(xfer->ux_status == USBD_CANCELLED || - xfer->ux_status == USBD_TIMEOUT); + if (!usbd_xfer_trycomplete(xfer)) continue; - } - - /* - * Cancel the timeout and the task, which have not yet - * run. If they have already fired, at worst they are - * waiting for the lock. They will see that the xfer - * is no longer in progress and give up. - */ - callout_stop(&xfer->ux_callout); - usb_rem_task(xfer->ux_pipe->up_dev, &xfer->ux_aborttask); KASSERT(!sitd->isdone); #ifdef DIAGNOSTIC @@ -1908,37 +1890,6 @@ ohci_hash_find_itd(ohci_softc_t *sc, ohci_physaddr_t a) return NULL; } -void -ohci_timeout(void *addr) -{ - OHCIHIST_FUNC(); OHCIHIST_CALLED(); - struct usbd_xfer *xfer = addr; - ohci_softc_t *sc = OHCI_XFER2SC(xfer); - struct usbd_device *dev = xfer->ux_pipe->up_dev; - - DPRINTF("xfer=%#jx", (uintptr_t)xfer, 0, 0, 0); - - mutex_enter(&sc->sc_lock); - if (!sc->sc_dying && xfer->ux_status == USBD_IN_PROGRESS) - usb_add_task(dev, &xfer->ux_aborttask, USB_TASKQ_HC); - mutex_exit(&sc->sc_lock); -} - -void -ohci_timeout_task(void *addr) -{ - struct usbd_xfer *xfer = addr; - ohci_softc_t *sc = OHCI_XFER2SC(xfer); - - OHCIHIST_FUNC(); OHCIHIST_CALLED(); - - DPRINTF("xfer=%#jx", (uintptr_t)xfer, 0, 0, 0); - - mutex_enter(&sc->sc_lock); - ohci_abort_xfer(xfer, USBD_TIMEOUT); - mutex_exit(&sc->sc_lock); -} - #ifdef OHCI_DEBUG void ohci_dump_tds(ohci_softc_t *sc, ohci_soft_td_t *std) @@ -2211,19 +2162,8 @@ ohci_close_pipe(struct usbd_pipe *pipe, ohci_soft_ed_t *head) } /* - * Cancel or timeout a device request. We have two cases to deal with - * - * 1) A driver wants to stop scheduled or inflight transfers - * 2) A transfer has timed out - * - * It's impossible to guarantee that the requested transfer will not - * have (partially) happened since the hardware runs concurrently. - * - * Transfer state is protected by the bus lock and we set the transfer status - * as soon as either of the above happens (with bus lock held). - * - * Then we arrange for the hardware to tells us that it is not still - * processing the TDs by setting the sKip bit and requesting a SOF interrupt + * Arrange for the hardware to tells us that it is not still processing + * the TDs by setting the sKip bit and requesting a SOF interrupt * * Once we see the SOF interrupt we can check the transfer TDs/iTDs to see if * they've been processed and either @@ -2232,7 +2172,7 @@ ohci_close_pipe(struct usbd_pipe *pipe, ohci_soft_ed_t *head) * used. The softint handler will free the old ones. */ void -ohci_abort_xfer(struct usbd_xfer *xfer, usbd_status status) +ohci_abortx(struct usbd_xfer *xfer) { OHCIHIST_FUNC(); OHCIHIST_CALLED(); struct ohci_pipe *opipe = OHCI_PIPE2OPIPE(xfer->ux_pipe); @@ -2242,46 +2182,15 @@ ohci_abort_xfer(struct usbd_xfer *xfer, usbd_status status) ohci_physaddr_t headp; int hit; - KASSERTMSG((status == USBD_CANCELLED || status == USBD_TIMEOUT), - "invalid status for abort: %d", (int)status); - DPRINTF("xfer=%#jx pipe=%#jx sed=%#jx", (uintptr_t)xfer, (uintptr_t)opipe, (uintptr_t)sed, 0); KASSERT(mutex_owned(&sc->sc_lock)); ASSERT_SLEEPABLE(); - if (status == USBD_CANCELLED) { - /* - * We are synchronously aborting. Try to stop the - * callout and task, but if we can't, wait for them to - * complete. - */ - callout_halt(&xfer->ux_callout, &sc->sc_lock); - usb_rem_task_wait(xfer->ux_pipe->up_dev, &xfer->ux_aborttask, - USB_TASKQ_HC, &sc->sc_lock); - } else { - /* Otherwise, we are timing out. */ - KASSERT(status == USBD_TIMEOUT); - } - - /* - * The xfer cannot have been cancelled already. It is the - * responsibility of the caller of usbd_abort_pipe not to try - * to abort a pipe multiple times, whether concurrently or - * sequentially. - */ - KASSERT(xfer->ux_status != USBD_CANCELLED); - - /* Only the timeout, which runs only once, can time it out. */ - KASSERT(xfer->ux_status != USBD_TIMEOUT); - - /* If anyone else beat us, we're done. */ - if (xfer->ux_status != USBD_IN_PROGRESS) - return; - - /* We beat everyone else. Claim the status. */ - xfer->ux_status = status; + KASSERTMSG((xfer->ux_status == USBD_CANCELLED || + xfer->ux_status == USBD_TIMEOUT), + "bad abort status: %d", xfer->ux_status); /* * If we're dying, skip the hardware action and just notify the @@ -2872,10 +2781,7 @@ ohci_device_ctrl_start(struct usbd_xfer *xfer) sizeof(sed->ed.ed_tailp), BUS_DMASYNC_PREWRITE | BUS_DMASYNC_PREREAD); OWRITE4(sc, OHCI_COMMAND_STATUS, OHCI_CLF); - if (xfer->ux_timeout && !polling) { - callout_reset(&xfer->ux_callout, mstohz(xfer->ux_timeout), - ohci_timeout, xfer); - } + usbd_xfer_schedule_timeout(xfer); DPRINTF("done", 0, 0, 0, 0); @@ -2896,7 +2802,7 @@ ohci_device_ctrl_abort(struct usbd_xfer *xfer) OHCIHIST_FUNC(); OHCIHIST_CALLED(); DPRINTF("xfer=%#jx", (uintptr_t)xfer, 0, 0, 0); - ohci_abort_xfer(xfer, USBD_CANCELLED); + usbd_xfer_abort(xfer); } /* Close a device control pipe. */ @@ -3087,11 +2993,7 @@ ohci_device_bulk_start(struct usbd_xfer *xfer) usb_syncmem(&sed->dma, sed->offs, sizeof(sed->ed), BUS_DMASYNC_PREWRITE | BUS_DMASYNC_PREREAD); OWRITE4(sc, OHCI_COMMAND_STATUS, OHCI_BLF); - if (xfer->ux_timeout && !sc->sc_bus.ub_usepolling) { - callout_reset(&xfer->ux_callout, mstohz(xfer->ux_timeout), - ohci_timeout, xfer); - } - + usbd_xfer_schedule_timeout(xfer); xfer->ux_status = USBD_IN_PROGRESS; if (!polling) mutex_exit(&sc->sc_lock); @@ -3109,7 +3011,7 @@ ohci_device_bulk_abort(struct usbd_xfer *xfer) KASSERT(mutex_owned(&sc->sc_lock)); DPRINTF("xfer=%#jx", (uintptr_t)xfer, 0, 0, 0); - ohci_abort_xfer(xfer, USBD_CANCELLED); + usbd_xfer_abort(xfer); } /* @@ -3297,7 +3199,7 @@ ohci_device_intr_abort(struct usbd_xfer *xfer) KASSERT(mutex_owned(&sc->sc_lock)); KASSERT(xfer->ux_pipe->up_intrxfer == xfer); - ohci_abort_xfer(xfer, USBD_CANCELLED); + usbd_xfer_abort(xfer); } /* Close a device interrupt pipe. */ diff --git a/sys/dev/usb/uhci.c b/sys/dev/usb/uhci.c index 70d8549cf6a2..19c473402392 100644 --- a/sys/dev/usb/uhci.c +++ b/sys/dev/usb/uhci.c @@ -194,10 +194,8 @@ Static void uhci_check_intr(uhci_softc_t *, struct uhci_xfer *, ux_completeq_t *); Static void uhci_idone(struct uhci_xfer *, ux_completeq_t *); -Static void uhci_abort_xfer(struct usbd_xfer *, usbd_status); +Static void uhci_abortx(struct usbd_xfer *); -Static void uhci_timeout(void *); -Static void uhci_timeout_task(void *); Static void uhci_add_ls_ctrl(uhci_softc_t *, uhci_soft_qh_t *); Static void uhci_add_hs_ctrl(uhci_softc_t *, uhci_soft_qh_t *); Static void uhci_add_bulk(uhci_softc_t *, uhci_soft_qh_t *); @@ -212,6 +210,7 @@ Static usbd_status uhci_setup_isoc(struct usbd_pipe *); Static struct usbd_xfer * uhci_allocx(struct usbd_bus *, unsigned int); Static void uhci_freex(struct usbd_bus *, struct usbd_xfer *); +Static bool uhci_dying(struct usbd_bus *); Static void uhci_get_lock(struct usbd_bus *, kmutex_t **); Static int uhci_roothub_ctrl(struct usbd_bus *, usb_device_request_t *, void *, int); @@ -330,6 +329,8 @@ const struct usbd_bus_methods uhci_bus_methods = { .ubm_dopoll = uhci_poll, .ubm_allocx = uhci_allocx, .ubm_freex = uhci_freex, + .ubm_abortx = uhci_abortx, + .ubm_dying = uhci_dying, .ubm_getlock = uhci_get_lock, .ubm_rhctrl = uhci_roothub_ctrl, }; @@ -657,9 +658,6 @@ uhci_allocx(struct usbd_bus *bus, unsigned int nframes) if (xfer != NULL) { memset(xfer, 0, sizeof(struct uhci_xfer)); - /* Initialise this always so we can call remove on it. */ - usb_init_task(&xfer->ux_aborttask, uhci_timeout_task, xfer, - USB_TASKQ_MPSAFE); #ifdef DIAGNOSTIC struct uhci_xfer *uxfer = UHCI_XFER2UXFER(xfer); uxfer->ux_isdone = true; @@ -686,6 +684,14 @@ uhci_freex(struct usbd_bus *bus, struct usbd_xfer *xfer) pool_cache_put(sc->sc_xferpool, xfer); } +Static bool +uhci_dying(struct usbd_bus *bus) +{ + struct uhci_softc *sc = UHCI_BUS2SC(bus); + + return sc->sc_dying; +} + Static void uhci_get_lock(struct usbd_bus *bus, kmutex_t **lock) { @@ -1565,24 +1571,11 @@ uhci_idone(struct uhci_xfer *ux, ux_completeq_t *cqp) DPRINTFN(12, "ux=%#jx", (uintptr_t)ux, 0, 0, 0); /* - * If software has completed it, either by cancellation - * or timeout, drop it on the floor. + * Try to claim this xfer for completion. If it has already + * completed or aborted, drop it on the floor. */ - if (xfer->ux_status != USBD_IN_PROGRESS) { - KASSERT(xfer->ux_status == USBD_CANCELLED || - xfer->ux_status == USBD_TIMEOUT); - DPRINTF("aborted xfer=%#jx", (uintptr_t)xfer, 0, 0, 0); + if (!usbd_xfer_trycomplete(xfer)) return; - } - - /* - * Cancel the timeout and the task, which have not yet - * run. If they have already fired, at worst they are - * waiting for the lock. They will see that the xfer - * is no longer in progress and give up. - */ - callout_stop(&xfer->ux_callout); - usb_rem_task(xfer->ux_pipe->up_dev, &xfer->ux_aborttask); #ifdef DIAGNOSTIC #ifdef UHCI_DEBUG @@ -1712,40 +1705,6 @@ uhci_idone(struct uhci_xfer *ux, ux_completeq_t *cqp) DPRINTFN(12, "ux=%#jx done", (uintptr_t)ux, 0, 0, 0); } -/* - * Called when a request does not complete. - */ -void -uhci_timeout(void *addr) -{ - UHCIHIST_FUNC(); UHCIHIST_CALLED(); - struct usbd_xfer *xfer = addr; - uhci_softc_t *sc = UHCI_XFER2SC(xfer); - struct usbd_device *dev = xfer->ux_pipe->up_dev; - - DPRINTF("xfer %#jx", (uintptr_t)xfer, 0, 0, 0); - - mutex_enter(&sc->sc_lock); - if (!sc->sc_dying && xfer->ux_status == USBD_IN_PROGRESS) - usb_add_task(dev, &xfer->ux_aborttask, USB_TASKQ_HC); - mutex_exit(&sc->sc_lock); -} - -void -uhci_timeout_task(void *addr) -{ - struct usbd_xfer *xfer = addr; - uhci_softc_t *sc = UHCI_XFER2SC(xfer); - - UHCIHIST_FUNC(); UHCIHIST_CALLED(); - - DPRINTF("xfer=%#jx", (uintptr_t)xfer, 0, 0, 0); - - mutex_enter(&sc->sc_lock); - uhci_abort_xfer(xfer, USBD_TIMEOUT); - mutex_exit(&sc->sc_lock); -} - void uhci_poll(struct usbd_bus *bus) { @@ -2314,11 +2273,7 @@ uhci_device_bulk_start(struct usbd_xfer *xfer) uhci_add_bulk(sc, sqh); uhci_add_intr_list(sc, ux); - - if (xfer->ux_timeout && !polling) { - callout_reset(&xfer->ux_callout, mstohz(xfer->ux_timeout), - uhci_timeout, xfer); - } + usbd_xfer_schedule_timeout(xfer); xfer->ux_status = USBD_IN_PROGRESS; if (!polling) mutex_exit(&sc->sc_lock); @@ -2336,25 +2291,14 @@ uhci_device_bulk_abort(struct usbd_xfer *xfer) UHCIHIST_FUNC(); UHCIHIST_CALLED(); - uhci_abort_xfer(xfer, USBD_CANCELLED); + usbd_xfer_abort(xfer); } /* - * Cancel or timeout a device request. We have two cases to deal with - * - * 1) A driver wants to stop scheduled or inflight transfers - * 2) A transfer has timed out - * - * It's impossible to guarantee that the requested transfer will not - * have (partially) happened since the hardware runs concurrently. - * - * Transfer state is protected by the bus lock and we set the transfer status - * as soon as either of the above happens (with bus lock held). - * * To allow the hardware time to notice we simply wait. */ -void -uhci_abort_xfer(struct usbd_xfer *xfer, usbd_status status) +static void +uhci_abortx(struct usbd_xfer *xfer) { UHCIHIST_FUNC(); UHCIHIST_CALLED(); struct uhci_xfer *ux = UHCI_XFER2UXFER(xfer); @@ -2362,45 +2306,14 @@ uhci_abort_xfer(struct usbd_xfer *xfer, usbd_status status) uhci_softc_t *sc = UHCI_XFER2SC(xfer); uhci_soft_td_t *std; - KASSERTMSG((status == USBD_CANCELLED || status == USBD_TIMEOUT), - "invalid status for abort: %d", (int)status); - DPRINTFN(1,"xfer=%#jx, status=%jd", (uintptr_t)xfer, status, 0, 0); KASSERT(mutex_owned(&sc->sc_lock)); ASSERT_SLEEPABLE(); - if (status == USBD_CANCELLED) { - /* - * We are synchronously aborting. Try to stop the - * callout and task, but if we can't, wait for them to - * complete. - */ - callout_halt(&xfer->ux_callout, &sc->sc_lock); - usb_rem_task_wait(xfer->ux_pipe->up_dev, &xfer->ux_aborttask, - USB_TASKQ_HC, &sc->sc_lock); - } else { - /* Otherwise, we are timing out. */ - KASSERT(status == USBD_TIMEOUT); - } - - /* - * The xfer cannot have been cancelled already. It is the - * responsibility of the caller of usbd_abort_pipe not to try - * to abort a pipe multiple times, whether concurrently or - * sequentially. - */ - KASSERT(xfer->ux_status != USBD_CANCELLED); - - /* Only the timeout, which runs only once, can time it out. */ - KASSERT(xfer->ux_status != USBD_TIMEOUT); - - /* If anyone else beat us, we're done. */ - if (xfer->ux_status != USBD_IN_PROGRESS) - return; - - /* We beat everyone else. Claim the status. */ - xfer->ux_status = status; + KASSERTMSG((xfer->ux_status == USBD_CANCELLED || + xfer->ux_status == USBD_TIMEOUT), + "bad abort status: %d", xfer->ux_status); /* * If we're dying, skip the hardware action and just notify the @@ -2672,10 +2585,7 @@ uhci_device_ctrl_start(struct usbd_xfer *xfer) DPRINTF("--- dump end ---", 0, 0, 0, 0); } #endif - if (xfer->ux_timeout && !polling) { - callout_reset(&xfer->ux_callout, mstohz(xfer->ux_timeout), - uhci_timeout, xfer); - } + usbd_xfer_schedule_timeout(xfer); xfer->ux_status = USBD_IN_PROGRESS; if (!polling) mutex_exit(&sc->sc_lock); @@ -2834,7 +2744,7 @@ uhci_device_ctrl_abort(struct usbd_xfer *xfer) KASSERT(mutex_owned(&sc->sc_lock)); UHCIHIST_FUNC(); UHCIHIST_CALLED(); - uhci_abort_xfer(xfer, USBD_CANCELLED); + usbd_xfer_abort(xfer); } /* Close a device control pipe. */ @@ -2862,7 +2772,7 @@ uhci_device_intr_abort(struct usbd_xfer *xfer) UHCIHIST_FUNC(); UHCIHIST_CALLED(); DPRINTF("xfer=%#jx", (uintptr_t)xfer, 0, 0, 0); - uhci_abort_xfer(xfer, USBD_CANCELLED); + usbd_xfer_abort(xfer); } /* Close a device interrupt pipe. */ diff --git a/sys/dev/usb/usbdi.c b/sys/dev/usb/usbdi.c index f641f5d5c9ce..157a4aa2ae5d 100644 --- a/sys/dev/usb/usbdi.c +++ b/sys/dev/usb/usbdi.c @@ -71,6 +71,11 @@ static void usbd_free_buffer(struct usbd_xfer *); static struct usbd_xfer *usbd_alloc_xfer(struct usbd_device *, unsigned int); static usbd_status usbd_free_xfer(struct usbd_xfer *); static void usbd_request_async_cb(struct usbd_xfer *, void *, usbd_status); +static void usbd_xfer_abort1(struct usbd_xfer *, usbd_status); +static void usbd_xfer_timeout(void *); +static void usbd_xfer_timeout_task(void *); +static bool usbd_xfer_probe_timeout(struct usbd_xfer *); +static void usbd_xfer_cancel_timeout_async(struct usbd_xfer *); #if defined(USB_DEBUG) void @@ -479,6 +484,8 @@ usbd_alloc_xfer(struct usbd_device *dev, unsigned int nframes) xfer->ux_bus = dev->ud_bus; callout_init(&xfer->ux_callout, CALLOUT_MPSAFE); cv_init(&xfer->ux_cv, "usbxfer"); + usb_init_task(&xfer->ux_aborttask, usbd_xfer_timeout_task, xfer, + USB_TASKQ_MPSAFE); USBHIST_LOG(usbdebug, "returns %#jx", (uintptr_t)xfer, 0, 0, 0); @@ -1363,3 +1370,379 @@ usbd_get_string0(struct usbd_device *dev, int si, char *buf, int unicode) #endif return USBD_NORMAL_COMPLETION; } + +/* + * usbd_xfer_trycomplete(xfer) + * + * Try to claim xfer for completion. To be used in a host + * controller interrupt handler. + * + * Caller must either hold the bus lock or have the bus in polling + * mode. + */ +bool +usbd_xfer_trycomplete(struct usbd_xfer *xfer) +{ + struct usbd_bus *bus __diagused = xfer->ux_bus; + + KASSERT(bus->ub_usepolling || mutex_owned(bus->ub_lock)); + + /* + * If software has completed it, either by explicit abort or by + * timeout, too late. + */ + if (xfer->ux_status != USBD_IN_PROGRESS) + return false; + + /* + * We are completing the xfer. Cancel the timeout if we can, + * but only asynchronously. See usbd_xfer_cancel_timeout_async + * for why we need not wait for the callout or task here. + */ + usbd_xfer_cancel_timeout_async(xfer); + + /* Success! Note: Caller must set xfer->ux_status. */ + return true; +} + +/* + * usbd_xfer_abort(xfer) + * + * Try to claim xfer to abort. If successful, mark it completed + * with USBD_CANCELLED and call the bus-specific method to abort + * at the hardware level. + * + * To be called in thread context from struct + * usbd_pipe_methods::upm_abort. + */ +void +usbd_xfer_abort(struct usbd_xfer *xfer) +{ + + usbd_xfer_abort1(xfer, USBD_CANCELLED); +} + +/* + * usbd_xfer_abort1(xfer, status) + * + * Try to claim xfer to abort. If successful, mark it completed + * with status and call the bus-specific method to abort at the + * hardware level. + * + * To be used in thread context. The status is USBD_TIMEOUT if + * this is the unique timeout thread, and must be USBD_CANCELLED + * for explicit cancellation. + */ +static void +usbd_xfer_abort1(struct usbd_xfer *xfer, usbd_status status) +{ + struct usbd_bus *bus = xfer->ux_bus; + + KASSERT(mutex_owned(bus->ub_lock)); + + /* + * Nobody else can set this status: only one caller can time + * out, and only one caller can synchronously abort. So the + * status can't be the status we're trying to set this to. + */ + KASSERT(xfer->ux_status != status); + + /* + * If host controller interrupt or timer interrupt has + * completed it, too late to abort. + */ + if (xfer->ux_status != USBD_IN_PROGRESS) + return; + + if (status == USBD_CANCELLED) { + /* + * We are synchronously aborting. Cancel the timeout + * if we can, but only asynchronously. See + * usbd_xfer_cancel_timeout_async for why we need to + * wait for the callout or task here. + */ + usbd_xfer_cancel_timeout_async(xfer); + } else { + /* + * Otherwise, we are timing out. No action needed to + * cancel the timeout because we _are_ the timeout. + * This case should happen only via the timeout task, + * invoked via the callout. + */ + KASSERT(status == USBD_TIMEOUT); + } + + /* We beat everyone else. Claim the status. */ + xfer->ux_status = status; + + /* Do the bus-specific dance to abort the hardware. */ + bus->ub_methods->ubm_abortx(xfer); +} + +/* + * usbd_xfer_timeout(xfer) + * + * Called at IPL_SOFTCLOCK when too much time has elapsed waiting + * for xfer to complete. Since we can't abort the xfer at + * IPL_SOFTCLOCK, defer to a usb_task to run it in thread context, + * unless the xfer has completed or aborted concurrently -- and if + * the xfer has also been resubmitted, take care of rescheduling + * the callout. + */ +static void +usbd_xfer_timeout(void *cookie) +{ + struct usbd_xfer *xfer = cookie; + struct usbd_bus *bus = xfer->ux_bus; + struct usbd_device *dev = xfer->ux_pipe->up_dev; + + /* Acquire the lock so we can transition the timeout state. */ + mutex_enter(bus->ub_lock); + + /* + * Use usbd_xfer_probe_timeout to check whether the timeout is + * still valid, or to reschedule the callout if necessary. If + * it is still valid, schedule the task. + */ + if (usbd_xfer_probe_timeout(xfer)) + usb_add_task(dev, &xfer->ux_aborttask, USB_TASKQ_HC); + + /* + * Notify usbd_xfer_cancel_timeout_async that we may have + * scheduled the task. This causes callout_invoking to return + * false in usbd_xfer_cancel_timeout_async so that it can tell + * which stage in the callout->task->abort process we're at. + */ + callout_ack(&xfer->ux_callout); + + /* All done -- release the lock. */ + mutex_exit(bus->ub_lock); +} + +/* + * usbd_xfer_timeout_task(xfer) + * + * Called in thread context when too much time has elapsed waiting + * for xfer to complete. Issue usbd_abort_xfer(USBD_TIMEOUT), + * unless the xfer has completed or aborted concurrently -- and if + * the xfer has also been resubmitted, take care of rescheduling + * the callout. + */ +static void +usbd_xfer_timeout_task(void *cookie) +{ + struct usbd_xfer *xfer = cookie; + struct usbd_bus *bus = xfer->ux_bus; + + /* Acquire the lock so we can transition the timeout state. */ + mutex_enter(bus->ub_lock); + + /* + * Use usbd_xfer_probe_timeout to check whether the timeout is + * still valid, or to reschedule the callout if necessary. If + * it is still valid, schedule the task. + */ + if (usbd_xfer_probe_timeout(xfer)) + usbd_xfer_abort1(xfer, USBD_TIMEOUT); + + /* All done -- release the lock. */ + mutex_exit(bus->ub_lock); +} + +/* + * usbd_xfer_probe_timeout(xfer) + * + * Probe the status of xfer's timeout. Acknowledge and process a + * request to reschedule. Return true if the timeout is still + * valid and the caller should take further action (queueing a + * task or aborting the xfer), false if it must stop here. + */ +static bool +usbd_xfer_probe_timeout(struct usbd_xfer *xfer) +{ + struct usbd_bus *bus = xfer->ux_bus; + bool valid; + + KASSERT(bus->ub_usepolling || mutex_owned(bus->ub_lock)); + + /* The timeout must be set. */ + KASSERT(xfer->ux_timeout_set); + + /* + * Neither callout nor task may be pending; they execute + * alternately in lock step. + */ + KASSERT(!callout_pending(&xfer->ux_callout)); + KASSERT(!usb_task_pending(xfer->ux_pipe->up_dev, &xfer->ux_aborttask)); + + /* There are a few cases... */ + if (bus->ub_methods->ubm_dying(bus)) { + /* Host controller dying. Drop it all on the floor. */ + xfer->ux_timeout_set = false; + xfer->ux_timeout_reset = false; + valid = false; + } else if (xfer->ux_timeout_reset) { + /* + * The xfer completed _and_ got resubmitted while we + * waited for the lock. Acknowledge the request to + * reschedule, and reschedule it if there is a timeout + * and the bus is not polling. + */ + xfer->ux_timeout_reset = false; + if (xfer->ux_timeout && !bus->ub_usepolling) { + KASSERT(xfer->ux_timeout_set); + callout_schedule(&xfer->ux_callout, + mstohz(xfer->ux_timeout)); + } else { + /* No more callout or task scheduled. */ + xfer->ux_timeout_set = false; + } + valid = false; + } else if (xfer->ux_status != USBD_IN_PROGRESS) { + /* + * The xfer has completed by hardware completion or by + * software abort, and has not been resubmitted, so the + * timeout must be unset, and is no longer valid for + * the caller. + */ + xfer->ux_timeout_set = false; + valid = false; + } else { + /* + * The xfer has not yet completed, so the timeout is + * valid. + */ + valid = true; + } + + /* Any reset must have been processed. */ + KASSERT(!xfer->ux_timeout_reset); + + /* + * Either we claim the timeout is set, or the callout is idle. + * If the timeout is still set, we may be handing off to the + * task instead, so this is an if but not an iff. + */ + KASSERT(xfer->ux_timeout_set || !callout_pending(&xfer->ux_callout)); + + /* + * The task must be idle now. + * + * - If the caller is the callout, _and_ the timeout is still + * valid, the caller will schedule it, but it hasn't been + * scheduled yet. (If the timeout is not valid, the task + * should not be scheduled.) + * + * - If the caller is the task, it cannot be scheduled again + * until the callout runs again, which won't happen until we + * next release the lock. + */ + KASSERT(!usb_task_pending(xfer->ux_pipe->up_dev, &xfer->ux_aborttask)); + + KASSERT(bus->ub_usepolling || mutex_owned(bus->ub_lock)); + + return valid; +} + +/* + * usbd_xfer_schedule_timeout(xfer) + * + * Ensure that xfer has a timeout. If the callout is already + * queued or the task is already running, request that they + * reschedule the callout. If not, and if we're not polling, + * schedule the callout anew. + */ +void +usbd_xfer_schedule_timeout(struct usbd_xfer *xfer) +{ + struct usbd_bus *bus = xfer->ux_bus; + + KASSERT(bus->ub_usepolling || mutex_owned(bus->ub_lock)); + + if (xfer->ux_timeout_set) { + /* + * Callout or task has fired from a prior completed + * xfer but has not yet noticed that the xfer is done. + * Ask it to reschedule itself to ux_timeout. + */ + xfer->ux_timeout_reset = true; + } else if (xfer->ux_timeout && !bus->ub_usepolling) { + /* Callout is not scheduled. Schedule it. */ + KASSERT(!callout_pending(&xfer->ux_callout)); + callout_reset(&xfer->ux_callout, mstohz(xfer->ux_timeout), + usbd_xfer_timeout, xfer); + xfer->ux_timeout_set = true; + } + + KASSERT(bus->ub_usepolling || mutex_owned(bus->ub_lock)); +} + +/* + * usbd_xfer_cancel_timeout_async(xfer) + * + * Cancel the callout and the task of xfer, which have not yet run + * to completion, but don't wait for the callout or task to finish + * running. + * + * If they have already fired, at worst they are waiting for the + * bus lock. They will see that the xfer is no longer in progress + * and give up, or they will see that the xfer has been + * resubmitted with a new timeout and reschedule the callout. + * + * If a resubmitted request completed so fast that the callout + * didn't have time to process a timer reset, just cancel the + * timer reset. + */ +static void +usbd_xfer_cancel_timeout_async(struct usbd_xfer *xfer) +{ + struct usbd_bus *bus __diagused = xfer->ux_bus; + + KASSERT(bus->ub_usepolling || mutex_owned(bus->ub_lock)); + + KASSERT(xfer->ux_timeout_set); + xfer->ux_timeout_reset = false; + if (!callout_stop(&xfer->ux_callout)) { + /* + * We stopped the callout before it ran. The timeout + * is no longer set. + */ + xfer->ux_timeout_set = false; + } else if (callout_invoking(&xfer->ux_callout)) { + /* + * The callout has begun to run but it has not yet + * acquired the lock. The task cannot be queued yet, + * and the callout cannot have been rescheduled yet. + * + * By the time the callout acquires the lock, we will + * have transitioned from USBD_IN_PROGRESS to a + * completed status, and possibly also resubmitted the + * xfer and set xfer->ux_timeout_reset = true. In both + * cases, the callout will DTRT, so no further action + * is needed here. + */ + } else if (usb_rem_task(xfer->ux_pipe->up_dev, &xfer->ux_aborttask)) { + /* + * The callout had fired and scheduled the task, but we + * stopped the task before it could run. The timeout + * is therefore no longer set -- the next resubmission + * of the xfer must schedule a new timeout. + * + * The callout should not be be pending at this point: + * it is scheduled only under the lock, and only when + * xfer->ux_timeout_set is false, or by the callout or + * task itself when xfer->ux_timeout_reset is true. + */ + xfer->ux_timeout_set = false; + } + + /* + * The callout cannot be scheduled and the task cannot be + * queued at this point. Either we cancelled them, or they are + * already running and waiting for the bus lock. + */ + KASSERT(!callout_pending(&xfer->ux_callout)); + KASSERT(!usb_task_pending(xfer->ux_pipe->up_dev, &xfer->ux_aborttask)); + + KASSERT(bus->ub_usepolling || mutex_owned(bus->ub_lock)); +} diff --git a/sys/dev/usb/usbdi.h b/sys/dev/usb/usbdi.h index 0590af8fae5d..365ba172558f 100644 --- a/sys/dev/usb/usbdi.h +++ b/sys/dev/usb/usbdi.h @@ -185,6 +185,11 @@ int usbd_ratecheck(struct timeval *); usbd_status usbd_get_string(struct usbd_device *, int, char *); usbd_status usbd_get_string0(struct usbd_device *, int, char *, int); +/* For use by HCI drivers, not USB device drivers */ +void usbd_xfer_schedule_timeout(struct usbd_xfer *); +bool usbd_xfer_trycomplete(struct usbd_xfer *); +void usbd_xfer_abort(struct usbd_xfer *); + /* An iterator for descriptors. */ typedef struct { const uByte *cur; diff --git a/sys/dev/usb/usbdivar.h b/sys/dev/usb/usbdivar.h index a161b504b3d3..d0194732e0fb 100644 --- a/sys/dev/usb/usbdivar.h +++ b/sys/dev/usb/usbdivar.h @@ -96,6 +96,8 @@ struct usbd_bus_methods { void (*ubm_dopoll)(struct usbd_bus *); struct usbd_xfer *(*ubm_allocx)(struct usbd_bus *, unsigned int); void (*ubm_freex)(struct usbd_bus *, struct usbd_xfer *); + void (*ubm_abortx)(struct usbd_xfer *); + bool (*ubm_dying)(struct usbd_bus *); void (*ubm_getlock)(struct usbd_bus *, kmutex_t **); usbd_status (*ubm_newdev)(device_t, struct usbd_bus *, int, int, int, struct usbd_port *); diff --git a/sys/dev/usb/xhci.c b/sys/dev/usb/xhci.c index aaf3daa9787b..d3a2103c460b 100644 --- a/sys/dev/usb/xhci.c +++ b/sys/dev/usb/xhci.c @@ -140,6 +140,8 @@ static void xhci_softintr(void *); static void xhci_poll(struct usbd_bus *); static struct usbd_xfer *xhci_allocx(struct usbd_bus *, unsigned int); static void xhci_freex(struct usbd_bus *, struct usbd_xfer *); +static void xhci_abortx(struct usbd_xfer *); +static bool xhci_dying(struct usbd_bus *); static void xhci_get_lock(struct usbd_bus *, kmutex_t **); static usbd_status xhci_new_device(device_t, struct usbd_bus *, int, int, int, struct usbd_port *); @@ -205,15 +207,14 @@ static void xhci_device_bulk_abort(struct usbd_xfer *); static void xhci_device_bulk_close(struct usbd_pipe *); static void xhci_device_bulk_done(struct usbd_xfer *); -static void xhci_timeout(void *); -static void xhci_timeout_task(void *); - static const struct usbd_bus_methods xhci_bus_methods = { .ubm_open = xhci_open, .ubm_softint = xhci_softintr, .ubm_dopoll = xhci_poll, .ubm_allocx = xhci_allocx, .ubm_freex = xhci_freex, + .ubm_abortx = xhci_abortx, + .ubm_dying = xhci_dying, .ubm_getlock = xhci_get_lock, .ubm_newdev = xhci_new_device, .ubm_rhctrl = xhci_roothub_ctrl, @@ -1714,53 +1715,22 @@ xhci_close_pipe(struct usbd_pipe *pipe) * Should be called with sc_lock held. */ static void -xhci_abort_xfer(struct usbd_xfer *xfer, usbd_status status) +xhci_abortx(struct usbd_xfer *xfer) { XHCIHIST_FUNC(); XHCIHIST_CALLED(); struct xhci_softc * const sc = XHCI_XFER2SC(xfer); struct xhci_slot * const xs = xfer->ux_pipe->up_dev->ud_hcpriv; const u_int dci = xhci_ep_get_dci(xfer->ux_pipe->up_endpoint->ue_edesc); - KASSERTMSG((status == USBD_CANCELLED || status == USBD_TIMEOUT), - "invalid status for abort: %d", (int)status); - DPRINTFN(4, "xfer %#jx pipe %#jx status %jd", (uintptr_t)xfer, (uintptr_t)xfer->ux_pipe, status, 0); KASSERT(mutex_owned(&sc->sc_lock)); ASSERT_SLEEPABLE(); - if (status == USBD_CANCELLED) { - /* - * We are synchronously aborting. Try to stop the - * callout and task, but if we can't, wait for them to - * complete. - */ - callout_halt(&xfer->ux_callout, &sc->sc_lock); - usb_rem_task_wait(xfer->ux_pipe->up_dev, &xfer->ux_aborttask, - USB_TASKQ_HC, &sc->sc_lock); - } else { - /* Otherwise, we are timing out. */ - KASSERT(status == USBD_TIMEOUT); - } - - /* - * The xfer cannot have been cancelled already. It is the - * responsibility of the caller of usbd_abort_pipe not to try - * to abort a pipe multiple times, whether concurrently or - * sequentially. - */ - KASSERT(xfer->ux_status != USBD_CANCELLED); - - /* Only the timeout, which runs only once, can time it out. */ - KASSERT(xfer->ux_status != USBD_TIMEOUT); - - /* If anyone else beat us, we're done. */ - if (xfer->ux_status != USBD_IN_PROGRESS) - return; - - /* We beat everyone else. Claim the status. */ - xfer->ux_status = status; + KASSERTMSG((xfer->ux_status == USBD_CANCELLED || + xfer->ux_status == USBD_TIMEOUT), + "bad abort status: %d", xfer->ux_status); /* * If we're dying, skip the hardware action and just notify the @@ -1904,6 +1874,13 @@ xhci_rhpsc(struct xhci_softc * const sc, u_int ctlrport) if (xfer == NULL) return; + /* + * Try to claim this xfer for completion. If it has already + * completed or aborted, drop it on the floor. + */ + if (!usbd_xfer_trycomplete(xfer)) + return; + uint8_t *p = xfer->ux_buf; memset(p, 0, xfer->ux_length); p[rhp / NBBY] |= 1 << (rhp % NBBY); @@ -1992,6 +1969,13 @@ xhci_event_transfer(struct xhci_softc * const sc, return; } + /* + * Try to claim this xfer for completion. If it has already + * completed or aborted, drop it on the floor. + */ + if (!usbd_xfer_trycomplete(xfer)) + return; + /* 4.11.5.2 Event Data TRB */ if ((trb_3 & XHCI_TRB_3_ED_BIT) != 0) { DPRINTFN(14, "transfer Event Data: 0x%016jx 0x%08jx" @@ -2040,17 +2024,7 @@ xhci_event_transfer(struct xhci_softc * const sc, case XHCI_TRB_ERROR_STOPPED: case XHCI_TRB_ERROR_LENGTH: case XHCI_TRB_ERROR_STOPPED_SHORT: - /* - * don't complete the transfer being aborted - * as abort_xfer does instead. - */ - if (xfer->ux_status == USBD_CANCELLED || - xfer->ux_status == USBD_TIMEOUT) { - DPRINTFN(14, "ignore aborting xfer %#jx", - (uintptr_t)xfer, 0, 0, 0); - return; - } - err = USBD_CANCELLED; + err = USBD_IOERROR; break; case XHCI_TRB_ERROR_STALL: case XHCI_TRB_ERROR_BABBLE: @@ -2073,15 +2047,6 @@ xhci_event_transfer(struct xhci_softc * const sc, /* Override the status. */ xfer->ux_status = USBD_STALLED; - /* - * Cancel the timeout and the task, which have not yet - * run. If they have already fired, at worst they are - * waiting for the lock. They will see that the xfer - * is no longer in progress and give up. - */ - callout_stop(&xfer->ux_callout); - usb_rem_task(xfer->ux_pipe->up_dev, &xfer->ux_aborttask); - xhci_clear_endpoint_stall_async(xfer); return; default: @@ -2090,29 +2055,9 @@ xhci_event_transfer(struct xhci_softc * const sc, break; } - /* - * If software has completed it, either by cancellation - * or timeout, drop it on the floor. - */ - if (xfer->ux_status != USBD_IN_PROGRESS) { - KASSERTMSG((xfer->ux_status == USBD_CANCELLED || - xfer->ux_status == USBD_TIMEOUT), - "xfer %p status %x", xfer, xfer->ux_status); - return; - } - - /* Otherwise, set the status. */ + /* Set the status. */ xfer->ux_status = err; - /* - * Cancel the timeout and the task, which have not yet - * run. If they have already fired, at worst they are - * waiting for the lock. They will see that the xfer - * is no longer in progress and give up. - */ - callout_stop(&xfer->ux_callout); - usb_rem_task(xfer->ux_pipe->up_dev, &xfer->ux_aborttask); - if ((trb_3 & XHCI_TRB_3_ED_BIT) == 0 || (trb_0 & 0x3) == 0x0) { usb_transfer_complete(xfer); @@ -2279,8 +2224,6 @@ xhci_allocx(struct usbd_bus *bus, unsigned int nframes) xfer = pool_cache_get(sc->sc_xferpool, PR_WAITOK); if (xfer != NULL) { memset(xfer, 0, sizeof(struct xhci_xfer)); - usb_init_task(&xfer->ux_aborttask, xhci_timeout_task, xfer, - USB_TASKQ_MPSAFE); #ifdef DIAGNOSTIC xfer->ux_state = XFER_BUSY; #endif @@ -2307,6 +2250,14 @@ xhci_freex(struct usbd_bus *bus, struct usbd_xfer *xfer) pool_cache_put(sc->sc_xferpool, xfer); } +static bool +xhci_dying(struct usbd_bus *bus) +{ + struct xhci_softc * const sc = XHCI_BUS2SC(bus); + + return sc->sc_dying; +} + static void xhci_get_lock(struct usbd_bus *bus, kmutex_t **lock) { @@ -3882,11 +3833,7 @@ xhci_device_ctrl_start(struct usbd_xfer *xfer) mutex_exit(&tr->xr_lock); xhci_db_write_4(sc, XHCI_DOORBELL(xs->xs_idx), dci); - - if (xfer->ux_timeout && !xhci_polling_p(sc)) { - callout_reset(&xfer->ux_callout, mstohz(xfer->ux_timeout), - xhci_timeout, xfer); - } + usbd_xfer_schedule_timeout(xfer); return USBD_IN_PROGRESS; } @@ -3909,7 +3856,7 @@ xhci_device_ctrl_abort(struct usbd_xfer *xfer) { XHCIHIST_FUNC(); XHCIHIST_CALLED(); - xhci_abort_xfer(xfer, USBD_CANCELLED); + usbd_xfer_abort(xfer); } static void @@ -4003,11 +3950,7 @@ xhci_device_bulk_start(struct usbd_xfer *xfer) mutex_exit(&tr->xr_lock); xhci_db_write_4(sc, XHCI_DOORBELL(xs->xs_idx), dci); - - if (xfer->ux_timeout && !xhci_polling_p(sc)) { - callout_reset(&xfer->ux_callout, mstohz(xfer->ux_timeout), - xhci_timeout, xfer); - } + usbd_xfer_schedule_timeout(xfer); return USBD_IN_PROGRESS; } @@ -4035,7 +3978,7 @@ xhci_device_bulk_abort(struct usbd_xfer *xfer) { XHCIHIST_FUNC(); XHCIHIST_CALLED(); - xhci_abort_xfer(xfer, USBD_CANCELLED); + usbd_xfer_abort(xfer); } static void @@ -4115,11 +4058,7 @@ xhci_device_intr_start(struct usbd_xfer *xfer) mutex_exit(&tr->xr_lock); xhci_db_write_4(sc, XHCI_DOORBELL(xs->xs_idx), dci); - - if (xfer->ux_timeout && !polling) { - callout_reset(&xfer->ux_callout, mstohz(xfer->ux_timeout), - xhci_timeout, xfer); - } + usbd_xfer_schedule_timeout(xfer); return USBD_IN_PROGRESS; } @@ -4155,7 +4094,7 @@ xhci_device_intr_abort(struct usbd_xfer *xfer) KASSERT(mutex_owned(&sc->sc_lock)); DPRINTFN(15, "%#jx", (uintptr_t)xfer, 0, 0, 0); KASSERT(xfer->ux_pipe->up_intrxfer == xfer); - xhci_abort_xfer(xfer, USBD_CANCELLED); + usbd_xfer_abort(xfer); } static void @@ -4168,32 +4107,3 @@ xhci_device_intr_close(struct usbd_pipe *pipe) xhci_close_pipe(pipe); } - -/* ------------ */ - -static void -xhci_timeout(void *addr) -{ - XHCIHIST_FUNC(); XHCIHIST_CALLED(); - struct xhci_xfer * const xx = addr; - struct usbd_xfer * const xfer = &xx->xx_xfer; - struct xhci_softc * const sc = XHCI_XFER2SC(xfer); - struct usbd_device *dev = xfer->ux_pipe->up_dev; - - mutex_enter(&sc->sc_lock); - if (!sc->sc_dying && xfer->ux_status == USBD_IN_PROGRESS) - usb_add_task(dev, &xfer->ux_aborttask, USB_TASKQ_HC); - mutex_exit(&sc->sc_lock); -} - -static void -xhci_timeout_task(void *addr) -{ - XHCIHIST_FUNC(); XHCIHIST_CALLED(); - struct usbd_xfer * const xfer = addr; - struct xhci_softc * const sc = XHCI_XFER2SC(xfer); - - mutex_enter(&sc->sc_lock); - xhci_abort_xfer(xfer, USBD_TIMEOUT); - mutex_exit(&sc->sc_lock); -}