diff --git a/sys/dev/usb/ehci.c b/sys/dev/usb/ehci.c index 028def75a88d..d51315d79b0c 100644 --- a/sys/dev/usb/ehci.c +++ b/sys/dev/usb/ehci.c @@ -166,8 +166,6 @@ Static void ehci_check_itd_intr(ehci_softc_t *, struct ehci_xfer *, Static void ehci_check_sitd_intr(ehci_softc_t *, struct ehci_xfer *, ex_completeq_t *); Static void ehci_idone(struct ehci_xfer *, ex_completeq_t *); -Static void ehci_timeout(void *); -Static void ehci_timeout_task(void *); Static void ehci_intrlist_timeout(void *); Static void ehci_doorbell(void *); Static void ehci_pcd(void *); @@ -177,6 +175,7 @@ Static struct usbd_xfer * Static void ehci_freex(struct usbd_bus *, struct usbd_xfer *); Static void ehci_get_lock(struct usbd_bus *, kmutex_t **); +Static bool ehci_dying(struct usbd_bus *); Static int ehci_roothub_ctrl(struct usbd_bus *, usb_device_request_t *, void *, int); @@ -278,7 +277,7 @@ Static void ehci_set_qh_qtd(ehci_soft_qh_t *, ehci_soft_qtd_t *); Static void ehci_sync_hc(ehci_softc_t *); Static void ehci_close_pipe(struct usbd_pipe *, ehci_soft_qh_t *); -Static void ehci_abort_xfer(struct usbd_xfer *, usbd_status); +Static void ehci_abortx(struct usbd_xfer *); #ifdef EHCI_DEBUG Static ehci_softc_t *theehci; @@ -319,6 +318,8 @@ Static const struct usbd_bus_methods ehci_bus_methods = { .ubm_dopoll = ehci_poll, .ubm_allocx = ehci_allocx, .ubm_freex = ehci_freex, + .ubm_abortx = ehci_abortx, + .ubm_dying = ehci_dying, .ubm_getlock = ehci_get_lock, .ubm_rhctrl = ehci_roothub_ctrl, }; @@ -1046,24 +1047,11 @@ ehci_idone(struct ehci_xfer *ex, ex_completeq_t *cq) DPRINTF("ex=%#jx", (uintptr_t)ex, 0, 0, 0); /* - * If software has completed it, either by cancellation - * or timeout, drop it on the floor. + * Try to claim this xfer for completion. If it has already + * completed or aborted, drop it on the floor. */ - if (xfer->ux_status != USBD_IN_PROGRESS) { - KASSERT(xfer->ux_status == USBD_CANCELLED || - xfer->ux_status == USBD_TIMEOUT); - DPRINTF("aborted xfer=%#jx", (uintptr_t)xfer, 0, 0, 0); + if (!usbd_xfer_trycomplete(xfer)) return; - } - - /* - * Cancel the timeout and the task, which have not yet - * run. If they have already fired, at worst they are - * waiting for the lock. They will see that the xfer - * is no longer in progress and give up. - */ - callout_stop(&xfer->ux_callout); - usb_rem_task(xfer->ux_pipe->up_dev, &xfer->ux_aborttask); #ifdef DIAGNOSTIC #ifdef EHCI_DEBUG @@ -1538,9 +1526,6 @@ ehci_allocx(struct usbd_bus *bus, unsigned int nframes) if (xfer != NULL) { memset(xfer, 0, sizeof(struct ehci_xfer)); - /* Initialise this always so we can call remove on it. */ - usb_init_task(&xfer->ux_aborttask, ehci_timeout_task, xfer, - USB_TASKQ_MPSAFE); #ifdef DIAGNOSTIC struct ehci_xfer *ex = EHCI_XFER2EXFER(xfer); ex->ex_isdone = true; @@ -1568,6 +1553,14 @@ ehci_freex(struct usbd_bus *bus, struct usbd_xfer *xfer) pool_cache_put(sc->sc_xferpool, xfer); } +Static bool +ehci_dying(struct usbd_bus *bus) +{ + struct ehci_softc *sc = EHCI_BUS2SC(bus); + + return sc->sc_dying; +} + Static void ehci_get_lock(struct usbd_bus *bus, kmutex_t **lock) { @@ -3200,22 +3193,12 @@ ehci_close_pipe(struct usbd_pipe *pipe, ehci_soft_qh_t *head) } /* - * Cancel or timeout a device request. We have two cases to deal with - * - * 1) A driver wants to stop scheduled or inflight transfers - * 2) A transfer has timed out - * - * have (partially) happened since the hardware runs concurrently. - * - * Transfer state is protected by the bus lock and we set the transfer status - * as soon as either of the above happens (with bus lock held). - * - * Then we arrange for the hardware to tells us that it is not still + * Arrrange for the hardware to tells us that it is not still * processing the TDs by setting the QH halted bit and wait for the ehci * door bell */ Static void -ehci_abort_xfer(struct usbd_xfer *xfer, usbd_status status) +ehci_abortx(struct usbd_xfer *xfer) { EHCIHIST_FUNC(); EHCIHIST_CALLED(); struct ehci_pipe *epipe = EHCI_XFER2EPIPE(xfer); @@ -3227,45 +3210,14 @@ ehci_abort_xfer(struct usbd_xfer *xfer, usbd_status status) uint32_t qhstatus; int hit; - KASSERTMSG((status == USBD_CANCELLED || status == USBD_TIMEOUT), - "invalid status for abort: %d", (int)status); - DPRINTF("xfer=%#jx pipe=%#jx", (uintptr_t)xfer, (uintptr_t)epipe, 0, 0); KASSERT(mutex_owned(&sc->sc_lock)); ASSERT_SLEEPABLE(); - if (status == USBD_CANCELLED) { - /* - * We are synchronously aborting. Try to stop the - * callout and task, but if we can't, wait for them to - * complete. - */ - callout_halt(&xfer->ux_callout, &sc->sc_lock); - usb_rem_task_wait(xfer->ux_pipe->up_dev, &xfer->ux_aborttask, - USB_TASKQ_HC, &sc->sc_lock); - } else { - /* Otherwise, we are timing out. */ - KASSERT(status == USBD_TIMEOUT); - } - - /* - * The xfer cannot have been cancelled already. It is the - * responsibility of the caller of usbd_abort_pipe not to try - * to abort a pipe multiple times, whether concurrently or - * sequentially. - */ - KASSERT(xfer->ux_status != USBD_CANCELLED); - - /* Only the timeout, which runs only once, can time it out. */ - KASSERT(xfer->ux_status != USBD_TIMEOUT); - - /* If anyone else beat us, we're done. */ - if (xfer->ux_status != USBD_IN_PROGRESS) - return; - - /* We beat everyone else. Claim the status. */ - xfer->ux_status = status; + KASSERTMSG((xfer->ux_status == USBD_CANCELLED || + xfer->ux_status == USBD_TIMEOUT), + "bad abort status: %d", xfer->ux_status); /* * If we're dying, skip the hardware action and just notify the @@ -3473,43 +3425,6 @@ dying: KASSERT(mutex_owned(&sc->sc_lock)); } -Static void -ehci_timeout(void *addr) -{ - EHCIHIST_FUNC(); EHCIHIST_CALLED(); - struct usbd_xfer *xfer = addr; - ehci_softc_t *sc = EHCI_XFER2SC(xfer); - struct usbd_device *dev = xfer->ux_pipe->up_dev; - - DPRINTF("xfer %#jx", (uintptr_t)xfer, 0, 0, 0); -#ifdef EHCI_DEBUG - if (ehcidebug >= 2) { - struct usbd_pipe *pipe = xfer->ux_pipe; - usbd_dump_pipe(pipe); - } -#endif - - mutex_enter(&sc->sc_lock); - if (!sc->sc_dying && xfer->ux_status == USBD_IN_PROGRESS) - usb_add_task(dev, &xfer->ux_aborttask, USB_TASKQ_HC); - mutex_exit(&sc->sc_lock); -} - -Static void -ehci_timeout_task(void *addr) -{ - struct usbd_xfer *xfer = addr; - ehci_softc_t *sc = EHCI_XFER2SC(xfer); - - EHCIHIST_FUNC(); EHCIHIST_CALLED(); - - DPRINTF("xfer=%#jx", (uintptr_t)xfer, 0, 0, 0); - - mutex_enter(&sc->sc_lock); - ehci_abort_xfer(xfer, USBD_TIMEOUT); - mutex_exit(&sc->sc_lock); -} - /************************/ Static int @@ -3751,10 +3666,7 @@ ehci_device_ctrl_start(struct usbd_xfer *xfer) /* Insert qTD in QH list - also does usb_syncmem(sqh) */ ehci_set_qh_qtd(sqh, setup); - if (xfer->ux_timeout && !sc->sc_bus.ub_usepolling) { - callout_reset(&xfer->ux_callout, mstohz(xfer->ux_timeout), - ehci_timeout, xfer); - } + usbd_xfer_schedule_timeout(xfer); ehci_add_intr_list(sc, exfer); xfer->ux_status = USBD_IN_PROGRESS; if (!polling) @@ -3805,7 +3717,7 @@ ehci_device_ctrl_abort(struct usbd_xfer *xfer) EHCIHIST_FUNC(); EHCIHIST_CALLED(); DPRINTF("xfer=%#jx", (uintptr_t)xfer, 0, 0, 0); - ehci_abort_xfer(xfer, USBD_CANCELLED); + usbd_xfer_abort(xfer); } /* Close a device control pipe. */ @@ -3951,10 +3863,7 @@ ehci_device_bulk_start(struct usbd_xfer *xfer) /* also does usb_syncmem(sqh) */ ehci_set_qh_qtd(sqh, exfer->ex_sqtdstart); - if (xfer->ux_timeout && !sc->sc_bus.ub_usepolling) { - callout_reset(&xfer->ux_callout, mstohz(xfer->ux_timeout), - ehci_timeout, xfer); - } + usbd_xfer_schedule_timeout(xfer); ehci_add_intr_list(sc, exfer); xfer->ux_status = USBD_IN_PROGRESS; if (!polling) @@ -3985,7 +3894,7 @@ ehci_device_bulk_abort(struct usbd_xfer *xfer) EHCIHIST_FUNC(); EHCIHIST_CALLED(); DPRINTF("xfer %#jx", (uintptr_t)xfer, 0, 0, 0); - ehci_abort_xfer(xfer, USBD_CANCELLED); + usbd_xfer_abort(xfer); } /* @@ -4168,10 +4077,7 @@ ehci_device_intr_start(struct usbd_xfer *xfer) /* also does usb_syncmem(sqh) */ ehci_set_qh_qtd(sqh, exfer->ex_sqtdstart); - if (xfer->ux_timeout && !sc->sc_bus.ub_usepolling) { - callout_reset(&xfer->ux_callout, mstohz(xfer->ux_timeout), - ehci_timeout, xfer); - } + usbd_xfer_schedule_timeout(xfer); ehci_add_intr_list(sc, exfer); xfer->ux_status = USBD_IN_PROGRESS; if (!polling) @@ -4205,7 +4111,7 @@ ehci_device_intr_abort(struct usbd_xfer *xfer) * async doorbell. That's dependent on the async list, wheras * intr xfers are periodic, should not use this? */ - ehci_abort_xfer(xfer, USBD_CANCELLED); + usbd_xfer_abort(xfer); } Static void diff --git a/sys/dev/usb/motg.c b/sys/dev/usb/motg.c index f36b92cc0a53..255f0858440e 100644 --- a/sys/dev/usb/motg.c +++ b/sys/dev/usb/motg.c @@ -144,6 +144,7 @@ static void motg_softintr(void *); static struct usbd_xfer * motg_allocx(struct usbd_bus *, unsigned int); static void motg_freex(struct usbd_bus *, struct usbd_xfer *); +static bool motg_dying(struct usbd_bus *); static void motg_get_lock(struct usbd_bus *, kmutex_t **); static int motg_roothub_ctrl(struct usbd_bus *, usb_device_request_t *, void *, int); @@ -174,7 +175,7 @@ static void motg_device_data_read(struct usbd_xfer *); static void motg_device_data_write(struct usbd_xfer *); static void motg_device_clear_toggle(struct usbd_pipe *); -static void motg_device_xfer_abort(struct usbd_xfer *); +static void motg_abortx(struct usbd_xfer *); #define UBARR(sc) bus_space_barrier((sc)->sc_iot, (sc)->sc_ioh, 0, (sc)->sc_size, \ BUS_SPACE_BARRIER_READ|BUS_SPACE_BARRIER_WRITE) @@ -233,6 +234,8 @@ const struct usbd_bus_methods motg_bus_methods = { .ubm_dopoll = motg_poll, .ubm_allocx = motg_allocx, .ubm_freex = motg_freex, + .ubm_abortx = motg_abortx, + .ubm_dying = motg_dying, .ubm_getlock = motg_get_lock, .ubm_rhctrl = motg_roothub_ctrl, }; @@ -770,6 +773,14 @@ motg_freex(struct usbd_bus *bus, struct usbd_xfer *xfer) pool_cache_put(sc->sc_xferpool, xfer); } +static bool +motg_dying(struct usbd_bus *bus) +{ + struct motg_softc *sc = MOTG_BUS2SC(bus); + + return sc->sc_dying; +} + static void motg_get_lock(struct usbd_bus *bus, kmutex_t **lock) { @@ -1387,6 +1398,13 @@ motg_device_ctrl_intr_rx(struct motg_softc *sc) KASSERT(mutex_owned(&sc->sc_lock)); + /* + * Try to claim this xfer for completion. If it has already + * completed or aborted, drop it on the floor. + */ + if (!usbd_xfer_trycomplete(xfer)) + return; + KASSERT(ep->phase == DATA_IN || ep->phase == STATUS_IN); /* select endpoint 0 */ UWRITE1(sc, MUSB2_REG_EPINDEX, 0); @@ -1501,6 +1519,14 @@ motg_device_ctrl_intr_tx(struct motg_softc *sc) MOTGHIST_FUNC(); MOTGHIST_CALLED(); KASSERT(mutex_owned(&sc->sc_lock)); + + /* + * Try to claim this xfer for completion. If it has already + * completed or aborted, drop it on the floor. + */ + if (!usbd_xfer_trycomplete(xfer)) + return; + if (ep->phase == DATA_IN || ep->phase == STATUS_IN) { motg_device_ctrl_intr_rx(sc); return; @@ -1633,7 +1659,7 @@ motg_device_ctrl_abort(struct usbd_xfer *xfer) { MOTGHIST_FUNC(); MOTGHIST_CALLED(); - motg_device_xfer_abort(xfer); + usbd_xfer_abort(xfer); } /* Close a device control pipe */ @@ -2019,6 +2045,14 @@ motg_device_intr_tx(struct motg_softc *sc, int epnumber) MOTGHIST_FUNC(); MOTGHIST_CALLED(); KASSERT(mutex_owned(&sc->sc_lock)); + + /* + * Try to claim this xfer for completion. If it has already + * completed or aborted, drop it on the floor. + */ + if (!usbd_xfer_trycomplete(xfer)) + return; + KASSERT(ep->ep_number == epnumber); DPRINTFN(MD_BULK, " on ep %jd", epnumber, 0, 0, 0); @@ -2102,7 +2136,7 @@ motg_device_data_abort(struct usbd_xfer *xfer) MOTGHIST_FUNC(); MOTGHIST_CALLED(); - motg_device_xfer_abort(xfer); + usbd_xfer_abort(xfer); } /* Close a device control pipe */ @@ -2151,7 +2185,7 @@ motg_device_clear_toggle(struct usbd_pipe *pipe) /* Abort a device control request. */ static void -motg_device_xfer_abort(struct usbd_xfer *xfer) +motg_abortx(struct usbd_xfer *xfer) { MOTGHIST_FUNC(); MOTGHIST_CALLED(); uint8_t csr; @@ -2161,30 +2195,6 @@ motg_device_xfer_abort(struct usbd_xfer *xfer) KASSERT(mutex_owned(&sc->sc_lock)); ASSERT_SLEEPABLE(); - /* - * We are synchronously aborting. Try to stop the - * callout and task, but if we can't, wait for them to - * complete. - */ - callout_halt(&xfer->ux_callout, &sc->sc_lock); - usb_rem_task_wait(xfer->ux_pipe->up_dev, &xfer->ux_aborttask, - USB_TASKQ_HC, &sc->sc_lock); - - /* - * The xfer cannot have been cancelled already. It is the - * responsibility of the caller of usbd_abort_pipe not to try - * to abort a pipe multiple times, whether concurrently or - * sequentially. - */ - KASSERT(xfer->ux_status != USBD_CANCELLED); - - /* If anyone else beat us, we're done. */ - if (xfer->ux_status != USBD_IN_PROGRESS) - return; - - /* We beat everyone else. Claim the status. */ - xfer->ux_status = USBD_CANCELLED; - /* * If we're dying, skip the hardware action and just notify the * software that we're done. diff --git a/sys/dev/usb/ohci.c b/sys/dev/usb/ohci.c index de964ff98931..3855a41b1d6e 100644 --- a/sys/dev/usb/ohci.c +++ b/sys/dev/usb/ohci.c @@ -169,6 +169,7 @@ Static void ohci_device_isoc_enter(struct usbd_xfer *); Static struct usbd_xfer * ohci_allocx(struct usbd_bus *, unsigned int); Static void ohci_freex(struct usbd_bus *, struct usbd_xfer *); +Static bool ohci_dying(struct usbd_bus *); Static void ohci_get_lock(struct usbd_bus *, kmutex_t **); Static int ohci_roothub_ctrl(struct usbd_bus *, usb_device_request_t *, void *, int); @@ -213,12 +214,10 @@ Static void ohci_device_isoc_done(struct usbd_xfer *); Static usbd_status ohci_device_setintr(ohci_softc_t *, struct ohci_pipe *, int); -Static void ohci_timeout(void *); -Static void ohci_timeout_task(void *); Static void ohci_rhsc_enable(void *); Static void ohci_close_pipe(struct usbd_pipe *, ohci_soft_ed_t *); -Static void ohci_abort_xfer(struct usbd_xfer *, usbd_status); +Static void ohci_abortx(struct usbd_xfer *); Static void ohci_device_clear_toggle(struct usbd_pipe *); Static void ohci_noop(struct usbd_pipe *); @@ -287,6 +286,8 @@ Static const struct usbd_bus_methods ohci_bus_methods = { .ubm_dopoll = ohci_poll, .ubm_allocx = ohci_allocx, .ubm_freex = ohci_freex, + .ubm_abortx = ohci_abortx, + .ubm_dying = ohci_dying, .ubm_getlock = ohci_get_lock, .ubm_rhctrl = ohci_roothub_ctrl, }; @@ -1068,9 +1069,6 @@ ohci_allocx(struct usbd_bus *bus, unsigned int nframes) if (xfer != NULL) { memset(xfer, 0, sizeof(struct ohci_xfer)); - /* Initialise this always so we can call remove on it. */ - usb_init_task(&xfer->ux_aborttask, ohci_timeout_task, xfer, - USB_TASKQ_MPSAFE); #ifdef DIAGNOSTIC xfer->ux_state = XFER_BUSY; #endif @@ -1092,6 +1090,14 @@ ohci_freex(struct usbd_bus *bus, struct usbd_xfer *xfer) pool_cache_put(sc->sc_xferpool, xfer); } +Static bool +ohci_dying(struct usbd_bus *bus) +{ + ohci_softc_t *sc = OHCI_BUS2SC(bus); + + return sc->sc_dying; +} + Static void ohci_get_lock(struct usbd_bus *bus, kmutex_t **lock) { @@ -1463,23 +1469,11 @@ ohci_softintr(void *v) } /* - * If software has completed it, either by cancellation - * or timeout, drop it on the floor. + * Try to claim this xfer for completion. If it has + * already completed or aborted, drop it on the floor. */ - if (xfer->ux_status != USBD_IN_PROGRESS) { - KASSERT(xfer->ux_status == USBD_CANCELLED || - xfer->ux_status == USBD_TIMEOUT); + if (!usbd_xfer_trycomplete(xfer)) continue; - } - - /* - * Cancel the timeout and the task, which have not yet - * run. If they have already fired, at worst they are - * waiting for the lock. They will see that the xfer - * is no longer in progress and give up. - */ - callout_stop(&xfer->ux_callout); - usb_rem_task(xfer->ux_pipe->up_dev, &xfer->ux_aborttask); len = std->len; if (std->td.td_cbp != 0) @@ -1552,23 +1546,11 @@ ohci_softintr(void *v) continue; /* - * If software has completed it, either by cancellation - * or timeout, drop it on the floor. + * Try to claim this xfer for completion. If it has + * already completed or aborted, drop it on the floor. */ - if (xfer->ux_status != USBD_IN_PROGRESS) { - KASSERT(xfer->ux_status == USBD_CANCELLED || - xfer->ux_status == USBD_TIMEOUT); + if (!usbd_xfer_trycomplete(xfer)) continue; - } - - /* - * Cancel the timeout and the task, which have not yet - * run. If they have already fired, at worst they are - * waiting for the lock. They will see that the xfer - * is no longer in progress and give up. - */ - callout_stop(&xfer->ux_callout); - usb_rem_task(xfer->ux_pipe->up_dev, &xfer->ux_aborttask); KASSERT(!sitd->isdone); #ifdef DIAGNOSTIC @@ -1908,37 +1890,6 @@ ohci_hash_find_itd(ohci_softc_t *sc, ohci_physaddr_t a) return NULL; } -void -ohci_timeout(void *addr) -{ - OHCIHIST_FUNC(); OHCIHIST_CALLED(); - struct usbd_xfer *xfer = addr; - ohci_softc_t *sc = OHCI_XFER2SC(xfer); - struct usbd_device *dev = xfer->ux_pipe->up_dev; - - DPRINTF("xfer=%#jx", (uintptr_t)xfer, 0, 0, 0); - - mutex_enter(&sc->sc_lock); - if (!sc->sc_dying && xfer->ux_status == USBD_IN_PROGRESS) - usb_add_task(dev, &xfer->ux_aborttask, USB_TASKQ_HC); - mutex_exit(&sc->sc_lock); -} - -void -ohci_timeout_task(void *addr) -{ - struct usbd_xfer *xfer = addr; - ohci_softc_t *sc = OHCI_XFER2SC(xfer); - - OHCIHIST_FUNC(); OHCIHIST_CALLED(); - - DPRINTF("xfer=%#jx", (uintptr_t)xfer, 0, 0, 0); - - mutex_enter(&sc->sc_lock); - ohci_abort_xfer(xfer, USBD_TIMEOUT); - mutex_exit(&sc->sc_lock); -} - #ifdef OHCI_DEBUG void ohci_dump_tds(ohci_softc_t *sc, ohci_soft_td_t *std) @@ -2211,19 +2162,8 @@ ohci_close_pipe(struct usbd_pipe *pipe, ohci_soft_ed_t *head) } /* - * Cancel or timeout a device request. We have two cases to deal with - * - * 1) A driver wants to stop scheduled or inflight transfers - * 2) A transfer has timed out - * - * It's impossible to guarantee that the requested transfer will not - * have (partially) happened since the hardware runs concurrently. - * - * Transfer state is protected by the bus lock and we set the transfer status - * as soon as either of the above happens (with bus lock held). - * - * Then we arrange for the hardware to tells us that it is not still - * processing the TDs by setting the sKip bit and requesting a SOF interrupt + * Arrange for the hardware to tells us that it is not still processing + * the TDs by setting the sKip bit and requesting a SOF interrupt * * Once we see the SOF interrupt we can check the transfer TDs/iTDs to see if * they've been processed and either @@ -2232,7 +2172,7 @@ ohci_close_pipe(struct usbd_pipe *pipe, ohci_soft_ed_t *head) * used. The softint handler will free the old ones. */ void -ohci_abort_xfer(struct usbd_xfer *xfer, usbd_status status) +ohci_abortx(struct usbd_xfer *xfer) { OHCIHIST_FUNC(); OHCIHIST_CALLED(); struct ohci_pipe *opipe = OHCI_PIPE2OPIPE(xfer->ux_pipe); @@ -2242,46 +2182,15 @@ ohci_abort_xfer(struct usbd_xfer *xfer, usbd_status status) ohci_physaddr_t headp; int hit; - KASSERTMSG((status == USBD_CANCELLED || status == USBD_TIMEOUT), - "invalid status for abort: %d", (int)status); - DPRINTF("xfer=%#jx pipe=%#jx sed=%#jx", (uintptr_t)xfer, (uintptr_t)opipe, (uintptr_t)sed, 0); KASSERT(mutex_owned(&sc->sc_lock)); ASSERT_SLEEPABLE(); - if (status == USBD_CANCELLED) { - /* - * We are synchronously aborting. Try to stop the - * callout and task, but if we can't, wait for them to - * complete. - */ - callout_halt(&xfer->ux_callout, &sc->sc_lock); - usb_rem_task_wait(xfer->ux_pipe->up_dev, &xfer->ux_aborttask, - USB_TASKQ_HC, &sc->sc_lock); - } else { - /* Otherwise, we are timing out. */ - KASSERT(status == USBD_TIMEOUT); - } - - /* - * The xfer cannot have been cancelled already. It is the - * responsibility of the caller of usbd_abort_pipe not to try - * to abort a pipe multiple times, whether concurrently or - * sequentially. - */ - KASSERT(xfer->ux_status != USBD_CANCELLED); - - /* Only the timeout, which runs only once, can time it out. */ - KASSERT(xfer->ux_status != USBD_TIMEOUT); - - /* If anyone else beat us, we're done. */ - if (xfer->ux_status != USBD_IN_PROGRESS) - return; - - /* We beat everyone else. Claim the status. */ - xfer->ux_status = status; + KASSERTMSG((xfer->ux_status == USBD_CANCELLED || + xfer->ux_status == USBD_TIMEOUT), + "bad abort status: %d", xfer->ux_status); /* * If we're dying, skip the hardware action and just notify the @@ -2872,10 +2781,7 @@ ohci_device_ctrl_start(struct usbd_xfer *xfer) sizeof(sed->ed.ed_tailp), BUS_DMASYNC_PREWRITE | BUS_DMASYNC_PREREAD); OWRITE4(sc, OHCI_COMMAND_STATUS, OHCI_CLF); - if (xfer->ux_timeout && !polling) { - callout_reset(&xfer->ux_callout, mstohz(xfer->ux_timeout), - ohci_timeout, xfer); - } + usbd_xfer_schedule_timeout(xfer); DPRINTF("done", 0, 0, 0, 0); @@ -2896,7 +2802,7 @@ ohci_device_ctrl_abort(struct usbd_xfer *xfer) OHCIHIST_FUNC(); OHCIHIST_CALLED(); DPRINTF("xfer=%#jx", (uintptr_t)xfer, 0, 0, 0); - ohci_abort_xfer(xfer, USBD_CANCELLED); + usbd_xfer_abort(xfer); } /* Close a device control pipe. */ @@ -3087,11 +2993,7 @@ ohci_device_bulk_start(struct usbd_xfer *xfer) usb_syncmem(&sed->dma, sed->offs, sizeof(sed->ed), BUS_DMASYNC_PREWRITE | BUS_DMASYNC_PREREAD); OWRITE4(sc, OHCI_COMMAND_STATUS, OHCI_BLF); - if (xfer->ux_timeout && !sc->sc_bus.ub_usepolling) { - callout_reset(&xfer->ux_callout, mstohz(xfer->ux_timeout), - ohci_timeout, xfer); - } - + usbd_xfer_schedule_timeout(xfer); xfer->ux_status = USBD_IN_PROGRESS; if (!polling) mutex_exit(&sc->sc_lock); @@ -3109,7 +3011,7 @@ ohci_device_bulk_abort(struct usbd_xfer *xfer) KASSERT(mutex_owned(&sc->sc_lock)); DPRINTF("xfer=%#jx", (uintptr_t)xfer, 0, 0, 0); - ohci_abort_xfer(xfer, USBD_CANCELLED); + usbd_xfer_abort(xfer); } /* @@ -3297,7 +3199,7 @@ ohci_device_intr_abort(struct usbd_xfer *xfer) KASSERT(mutex_owned(&sc->sc_lock)); KASSERT(xfer->ux_pipe->up_intrxfer == xfer); - ohci_abort_xfer(xfer, USBD_CANCELLED); + usbd_xfer_abort(xfer); } /* Close a device interrupt pipe. */ diff --git a/sys/dev/usb/uhci.c b/sys/dev/usb/uhci.c index 70d8549cf6a2..19c473402392 100644 --- a/sys/dev/usb/uhci.c +++ b/sys/dev/usb/uhci.c @@ -194,10 +194,8 @@ Static void uhci_check_intr(uhci_softc_t *, struct uhci_xfer *, ux_completeq_t *); Static void uhci_idone(struct uhci_xfer *, ux_completeq_t *); -Static void uhci_abort_xfer(struct usbd_xfer *, usbd_status); +Static void uhci_abortx(struct usbd_xfer *); -Static void uhci_timeout(void *); -Static void uhci_timeout_task(void *); Static void uhci_add_ls_ctrl(uhci_softc_t *, uhci_soft_qh_t *); Static void uhci_add_hs_ctrl(uhci_softc_t *, uhci_soft_qh_t *); Static void uhci_add_bulk(uhci_softc_t *, uhci_soft_qh_t *); @@ -212,6 +210,7 @@ Static usbd_status uhci_setup_isoc(struct usbd_pipe *); Static struct usbd_xfer * uhci_allocx(struct usbd_bus *, unsigned int); Static void uhci_freex(struct usbd_bus *, struct usbd_xfer *); +Static bool uhci_dying(struct usbd_bus *); Static void uhci_get_lock(struct usbd_bus *, kmutex_t **); Static int uhci_roothub_ctrl(struct usbd_bus *, usb_device_request_t *, void *, int); @@ -330,6 +329,8 @@ const struct usbd_bus_methods uhci_bus_methods = { .ubm_dopoll = uhci_poll, .ubm_allocx = uhci_allocx, .ubm_freex = uhci_freex, + .ubm_abortx = uhci_abortx, + .ubm_dying = uhci_dying, .ubm_getlock = uhci_get_lock, .ubm_rhctrl = uhci_roothub_ctrl, }; @@ -657,9 +658,6 @@ uhci_allocx(struct usbd_bus *bus, unsigned int nframes) if (xfer != NULL) { memset(xfer, 0, sizeof(struct uhci_xfer)); - /* Initialise this always so we can call remove on it. */ - usb_init_task(&xfer->ux_aborttask, uhci_timeout_task, xfer, - USB_TASKQ_MPSAFE); #ifdef DIAGNOSTIC struct uhci_xfer *uxfer = UHCI_XFER2UXFER(xfer); uxfer->ux_isdone = true; @@ -686,6 +684,14 @@ uhci_freex(struct usbd_bus *bus, struct usbd_xfer *xfer) pool_cache_put(sc->sc_xferpool, xfer); } +Static bool +uhci_dying(struct usbd_bus *bus) +{ + struct uhci_softc *sc = UHCI_BUS2SC(bus); + + return sc->sc_dying; +} + Static void uhci_get_lock(struct usbd_bus *bus, kmutex_t **lock) { @@ -1565,24 +1571,11 @@ uhci_idone(struct uhci_xfer *ux, ux_completeq_t *cqp) DPRINTFN(12, "ux=%#jx", (uintptr_t)ux, 0, 0, 0); /* - * If software has completed it, either by cancellation - * or timeout, drop it on the floor. + * Try to claim this xfer for completion. If it has already + * completed or aborted, drop it on the floor. */ - if (xfer->ux_status != USBD_IN_PROGRESS) { - KASSERT(xfer->ux_status == USBD_CANCELLED || - xfer->ux_status == USBD_TIMEOUT); - DPRINTF("aborted xfer=%#jx", (uintptr_t)xfer, 0, 0, 0); + if (!usbd_xfer_trycomplete(xfer)) return; - } - - /* - * Cancel the timeout and the task, which have not yet - * run. If they have already fired, at worst they are - * waiting for the lock. They will see that the xfer - * is no longer in progress and give up. - */ - callout_stop(&xfer->ux_callout); - usb_rem_task(xfer->ux_pipe->up_dev, &xfer->ux_aborttask); #ifdef DIAGNOSTIC #ifdef UHCI_DEBUG @@ -1712,40 +1705,6 @@ uhci_idone(struct uhci_xfer *ux, ux_completeq_t *cqp) DPRINTFN(12, "ux=%#jx done", (uintptr_t)ux, 0, 0, 0); } -/* - * Called when a request does not complete. - */ -void -uhci_timeout(void *addr) -{ - UHCIHIST_FUNC(); UHCIHIST_CALLED(); - struct usbd_xfer *xfer = addr; - uhci_softc_t *sc = UHCI_XFER2SC(xfer); - struct usbd_device *dev = xfer->ux_pipe->up_dev; - - DPRINTF("xfer %#jx", (uintptr_t)xfer, 0, 0, 0); - - mutex_enter(&sc->sc_lock); - if (!sc->sc_dying && xfer->ux_status == USBD_IN_PROGRESS) - usb_add_task(dev, &xfer->ux_aborttask, USB_TASKQ_HC); - mutex_exit(&sc->sc_lock); -} - -void -uhci_timeout_task(void *addr) -{ - struct usbd_xfer *xfer = addr; - uhci_softc_t *sc = UHCI_XFER2SC(xfer); - - UHCIHIST_FUNC(); UHCIHIST_CALLED(); - - DPRINTF("xfer=%#jx", (uintptr_t)xfer, 0, 0, 0); - - mutex_enter(&sc->sc_lock); - uhci_abort_xfer(xfer, USBD_TIMEOUT); - mutex_exit(&sc->sc_lock); -} - void uhci_poll(struct usbd_bus *bus) { @@ -2314,11 +2273,7 @@ uhci_device_bulk_start(struct usbd_xfer *xfer) uhci_add_bulk(sc, sqh); uhci_add_intr_list(sc, ux); - - if (xfer->ux_timeout && !polling) { - callout_reset(&xfer->ux_callout, mstohz(xfer->ux_timeout), - uhci_timeout, xfer); - } + usbd_xfer_schedule_timeout(xfer); xfer->ux_status = USBD_IN_PROGRESS; if (!polling) mutex_exit(&sc->sc_lock); @@ -2336,25 +2291,14 @@ uhci_device_bulk_abort(struct usbd_xfer *xfer) UHCIHIST_FUNC(); UHCIHIST_CALLED(); - uhci_abort_xfer(xfer, USBD_CANCELLED); + usbd_xfer_abort(xfer); } /* - * Cancel or timeout a device request. We have two cases to deal with - * - * 1) A driver wants to stop scheduled or inflight transfers - * 2) A transfer has timed out - * - * It's impossible to guarantee that the requested transfer will not - * have (partially) happened since the hardware runs concurrently. - * - * Transfer state is protected by the bus lock and we set the transfer status - * as soon as either of the above happens (with bus lock held). - * * To allow the hardware time to notice we simply wait. */ -void -uhci_abort_xfer(struct usbd_xfer *xfer, usbd_status status) +static void +uhci_abortx(struct usbd_xfer *xfer) { UHCIHIST_FUNC(); UHCIHIST_CALLED(); struct uhci_xfer *ux = UHCI_XFER2UXFER(xfer); @@ -2362,45 +2306,14 @@ uhci_abort_xfer(struct usbd_xfer *xfer, usbd_status status) uhci_softc_t *sc = UHCI_XFER2SC(xfer); uhci_soft_td_t *std; - KASSERTMSG((status == USBD_CANCELLED || status == USBD_TIMEOUT), - "invalid status for abort: %d", (int)status); - DPRINTFN(1,"xfer=%#jx, status=%jd", (uintptr_t)xfer, status, 0, 0); KASSERT(mutex_owned(&sc->sc_lock)); ASSERT_SLEEPABLE(); - if (status == USBD_CANCELLED) { - /* - * We are synchronously aborting. Try to stop the - * callout and task, but if we can't, wait for them to - * complete. - */ - callout_halt(&xfer->ux_callout, &sc->sc_lock); - usb_rem_task_wait(xfer->ux_pipe->up_dev, &xfer->ux_aborttask, - USB_TASKQ_HC, &sc->sc_lock); - } else { - /* Otherwise, we are timing out. */ - KASSERT(status == USBD_TIMEOUT); - } - - /* - * The xfer cannot have been cancelled already. It is the - * responsibility of the caller of usbd_abort_pipe not to try - * to abort a pipe multiple times, whether concurrently or - * sequentially. - */ - KASSERT(xfer->ux_status != USBD_CANCELLED); - - /* Only the timeout, which runs only once, can time it out. */ - KASSERT(xfer->ux_status != USBD_TIMEOUT); - - /* If anyone else beat us, we're done. */ - if (xfer->ux_status != USBD_IN_PROGRESS) - return; - - /* We beat everyone else. Claim the status. */ - xfer->ux_status = status; + KASSERTMSG((xfer->ux_status == USBD_CANCELLED || + xfer->ux_status == USBD_TIMEOUT), + "bad abort status: %d", xfer->ux_status); /* * If we're dying, skip the hardware action and just notify the @@ -2672,10 +2585,7 @@ uhci_device_ctrl_start(struct usbd_xfer *xfer) DPRINTF("--- dump end ---", 0, 0, 0, 0); } #endif - if (xfer->ux_timeout && !polling) { - callout_reset(&xfer->ux_callout, mstohz(xfer->ux_timeout), - uhci_timeout, xfer); - } + usbd_xfer_schedule_timeout(xfer); xfer->ux_status = USBD_IN_PROGRESS; if (!polling) mutex_exit(&sc->sc_lock); @@ -2834,7 +2744,7 @@ uhci_device_ctrl_abort(struct usbd_xfer *xfer) KASSERT(mutex_owned(&sc->sc_lock)); UHCIHIST_FUNC(); UHCIHIST_CALLED(); - uhci_abort_xfer(xfer, USBD_CANCELLED); + usbd_xfer_abort(xfer); } /* Close a device control pipe. */ @@ -2862,7 +2772,7 @@ uhci_device_intr_abort(struct usbd_xfer *xfer) UHCIHIST_FUNC(); UHCIHIST_CALLED(); DPRINTF("xfer=%#jx", (uintptr_t)xfer, 0, 0, 0); - uhci_abort_xfer(xfer, USBD_CANCELLED); + usbd_xfer_abort(xfer); } /* Close a device interrupt pipe. */ diff --git a/sys/dev/usb/usb.c b/sys/dev/usb/usb.c index 4627a5a322ff..b8b0802d302e 100644 --- a/sys/dev/usb/usb.c +++ b/sys/dev/usb/usb.c @@ -431,14 +431,16 @@ usb_add_task(struct usbd_device *dev, struct usb_task *task, int queue) /* * usb_rem_task(dev, task) * - * If task is queued to run, remove it from the queue. + * If task is queued to run, remove it from the queue. Return + * true if it successfully removed the task from the queue, false + * if not. * * Caller is _not_ guaranteed that the task is not running when * this is done. * * Never sleeps. */ -void +bool usb_rem_task(struct usbd_device *dev, struct usb_task *task) { unsigned queue; @@ -452,10 +454,12 @@ usb_rem_task(struct usbd_device *dev, struct usb_task *task) TAILQ_REMOVE(&taskq->tasks, task, next); task->queue = USB_NUM_TASKQS; mutex_exit(&taskq->lock); - break; + return true; /* removed from the queue */ } mutex_exit(&taskq->lock); } + + return false; /* was not removed from the queue */ } /* @@ -526,6 +530,23 @@ usb_rem_task_wait(struct usbd_device *dev, struct usb_task *task, int queue, return removed; } +/* + * usb_task_pending(dev, task) + * + * True if task is queued, false if not. Note that if task is + * already running, it is not considered queued. + * + * For _negative_ diagnostic assertions only: + * + * KASSERT(!usb_task_pending(dev, task)); + */ +bool +usb_task_pending(struct usbd_device *dev, struct usb_task *task) +{ + + return task->queue != USB_NUM_TASKQS; +} + void usb_event_thread(void *arg) { diff --git a/sys/dev/usb/usbdi.c b/sys/dev/usb/usbdi.c index 4510d2b5ae39..157a4aa2ae5d 100644 --- a/sys/dev/usb/usbdi.c +++ b/sys/dev/usb/usbdi.c @@ -71,6 +71,11 @@ static void usbd_free_buffer(struct usbd_xfer *); static struct usbd_xfer *usbd_alloc_xfer(struct usbd_device *, unsigned int); static usbd_status usbd_free_xfer(struct usbd_xfer *); static void usbd_request_async_cb(struct usbd_xfer *, void *, usbd_status); +static void usbd_xfer_abort1(struct usbd_xfer *, usbd_status); +static void usbd_xfer_timeout(void *); +static void usbd_xfer_timeout_task(void *); +static bool usbd_xfer_probe_timeout(struct usbd_xfer *); +static void usbd_xfer_cancel_timeout_async(struct usbd_xfer *); #if defined(USB_DEBUG) void @@ -479,6 +484,8 @@ usbd_alloc_xfer(struct usbd_device *dev, unsigned int nframes) xfer->ux_bus = dev->ud_bus; callout_init(&xfer->ux_callout, CALLOUT_MPSAFE); cv_init(&xfer->ux_cv, "usbxfer"); + usb_init_task(&xfer->ux_aborttask, usbd_xfer_timeout_task, xfer, + USB_TASKQ_MPSAFE); USBHIST_LOG(usbdebug, "returns %#jx", (uintptr_t)xfer, 0, 0, 0); @@ -494,12 +501,15 @@ usbd_free_xfer(struct usbd_xfer *xfer) if (xfer->ux_buf) { usbd_free_buffer(xfer); } -#if defined(DIAGNOSTIC) - if (callout_pending(&xfer->ux_callout)) { - callout_stop(&xfer->ux_callout); - printf("usbd_free_xfer: timeout_handle pending\n"); - } -#endif + + /* Wait for any straggling timeout to complete. */ + mutex_enter(xfer->ux_bus->ub_lock); + xfer->ux_timeout_reset = false; /* do not resuscitate */ + callout_halt(&xfer->ux_callout, xfer->ux_bus->ub_lock); + usb_rem_task_wait(xfer->ux_pipe->up_dev, &xfer->ux_aborttask, + USB_TASKQ_HC, xfer->ux_bus->ub_lock); + mutex_exit(xfer->ux_bus->ub_lock); + cv_destroy(&xfer->ux_cv); xfer->ux_bus->ub_methods->ubm_freex(xfer->ux_bus, xfer); return USBD_NORMAL_COMPLETION; @@ -1360,3 +1370,379 @@ usbd_get_string0(struct usbd_device *dev, int si, char *buf, int unicode) #endif return USBD_NORMAL_COMPLETION; } + +/* + * usbd_xfer_trycomplete(xfer) + * + * Try to claim xfer for completion. To be used in a host + * controller interrupt handler. + * + * Caller must either hold the bus lock or have the bus in polling + * mode. + */ +bool +usbd_xfer_trycomplete(struct usbd_xfer *xfer) +{ + struct usbd_bus *bus __diagused = xfer->ux_bus; + + KASSERT(bus->ub_usepolling || mutex_owned(bus->ub_lock)); + + /* + * If software has completed it, either by explicit abort or by + * timeout, too late. + */ + if (xfer->ux_status != USBD_IN_PROGRESS) + return false; + + /* + * We are completing the xfer. Cancel the timeout if we can, + * but only asynchronously. See usbd_xfer_cancel_timeout_async + * for why we need not wait for the callout or task here. + */ + usbd_xfer_cancel_timeout_async(xfer); + + /* Success! Note: Caller must set xfer->ux_status. */ + return true; +} + +/* + * usbd_xfer_abort(xfer) + * + * Try to claim xfer to abort. If successful, mark it completed + * with USBD_CANCELLED and call the bus-specific method to abort + * at the hardware level. + * + * To be called in thread context from struct + * usbd_pipe_methods::upm_abort. + */ +void +usbd_xfer_abort(struct usbd_xfer *xfer) +{ + + usbd_xfer_abort1(xfer, USBD_CANCELLED); +} + +/* + * usbd_xfer_abort1(xfer, status) + * + * Try to claim xfer to abort. If successful, mark it completed + * with status and call the bus-specific method to abort at the + * hardware level. + * + * To be used in thread context. The status is USBD_TIMEOUT if + * this is the unique timeout thread, and must be USBD_CANCELLED + * for explicit cancellation. + */ +static void +usbd_xfer_abort1(struct usbd_xfer *xfer, usbd_status status) +{ + struct usbd_bus *bus = xfer->ux_bus; + + KASSERT(mutex_owned(bus->ub_lock)); + + /* + * Nobody else can set this status: only one caller can time + * out, and only one caller can synchronously abort. So the + * status can't be the status we're trying to set this to. + */ + KASSERT(xfer->ux_status != status); + + /* + * If host controller interrupt or timer interrupt has + * completed it, too late to abort. + */ + if (xfer->ux_status != USBD_IN_PROGRESS) + return; + + if (status == USBD_CANCELLED) { + /* + * We are synchronously aborting. Cancel the timeout + * if we can, but only asynchronously. See + * usbd_xfer_cancel_timeout_async for why we need to + * wait for the callout or task here. + */ + usbd_xfer_cancel_timeout_async(xfer); + } else { + /* + * Otherwise, we are timing out. No action needed to + * cancel the timeout because we _are_ the timeout. + * This case should happen only via the timeout task, + * invoked via the callout. + */ + KASSERT(status == USBD_TIMEOUT); + } + + /* We beat everyone else. Claim the status. */ + xfer->ux_status = status; + + /* Do the bus-specific dance to abort the hardware. */ + bus->ub_methods->ubm_abortx(xfer); +} + +/* + * usbd_xfer_timeout(xfer) + * + * Called at IPL_SOFTCLOCK when too much time has elapsed waiting + * for xfer to complete. Since we can't abort the xfer at + * IPL_SOFTCLOCK, defer to a usb_task to run it in thread context, + * unless the xfer has completed or aborted concurrently -- and if + * the xfer has also been resubmitted, take care of rescheduling + * the callout. + */ +static void +usbd_xfer_timeout(void *cookie) +{ + struct usbd_xfer *xfer = cookie; + struct usbd_bus *bus = xfer->ux_bus; + struct usbd_device *dev = xfer->ux_pipe->up_dev; + + /* Acquire the lock so we can transition the timeout state. */ + mutex_enter(bus->ub_lock); + + /* + * Use usbd_xfer_probe_timeout to check whether the timeout is + * still valid, or to reschedule the callout if necessary. If + * it is still valid, schedule the task. + */ + if (usbd_xfer_probe_timeout(xfer)) + usb_add_task(dev, &xfer->ux_aborttask, USB_TASKQ_HC); + + /* + * Notify usbd_xfer_cancel_timeout_async that we may have + * scheduled the task. This causes callout_invoking to return + * false in usbd_xfer_cancel_timeout_async so that it can tell + * which stage in the callout->task->abort process we're at. + */ + callout_ack(&xfer->ux_callout); + + /* All done -- release the lock. */ + mutex_exit(bus->ub_lock); +} + +/* + * usbd_xfer_timeout_task(xfer) + * + * Called in thread context when too much time has elapsed waiting + * for xfer to complete. Issue usbd_abort_xfer(USBD_TIMEOUT), + * unless the xfer has completed or aborted concurrently -- and if + * the xfer has also been resubmitted, take care of rescheduling + * the callout. + */ +static void +usbd_xfer_timeout_task(void *cookie) +{ + struct usbd_xfer *xfer = cookie; + struct usbd_bus *bus = xfer->ux_bus; + + /* Acquire the lock so we can transition the timeout state. */ + mutex_enter(bus->ub_lock); + + /* + * Use usbd_xfer_probe_timeout to check whether the timeout is + * still valid, or to reschedule the callout if necessary. If + * it is still valid, schedule the task. + */ + if (usbd_xfer_probe_timeout(xfer)) + usbd_xfer_abort1(xfer, USBD_TIMEOUT); + + /* All done -- release the lock. */ + mutex_exit(bus->ub_lock); +} + +/* + * usbd_xfer_probe_timeout(xfer) + * + * Probe the status of xfer's timeout. Acknowledge and process a + * request to reschedule. Return true if the timeout is still + * valid and the caller should take further action (queueing a + * task or aborting the xfer), false if it must stop here. + */ +static bool +usbd_xfer_probe_timeout(struct usbd_xfer *xfer) +{ + struct usbd_bus *bus = xfer->ux_bus; + bool valid; + + KASSERT(bus->ub_usepolling || mutex_owned(bus->ub_lock)); + + /* The timeout must be set. */ + KASSERT(xfer->ux_timeout_set); + + /* + * Neither callout nor task may be pending; they execute + * alternately in lock step. + */ + KASSERT(!callout_pending(&xfer->ux_callout)); + KASSERT(!usb_task_pending(xfer->ux_pipe->up_dev, &xfer->ux_aborttask)); + + /* There are a few cases... */ + if (bus->ub_methods->ubm_dying(bus)) { + /* Host controller dying. Drop it all on the floor. */ + xfer->ux_timeout_set = false; + xfer->ux_timeout_reset = false; + valid = false; + } else if (xfer->ux_timeout_reset) { + /* + * The xfer completed _and_ got resubmitted while we + * waited for the lock. Acknowledge the request to + * reschedule, and reschedule it if there is a timeout + * and the bus is not polling. + */ + xfer->ux_timeout_reset = false; + if (xfer->ux_timeout && !bus->ub_usepolling) { + KASSERT(xfer->ux_timeout_set); + callout_schedule(&xfer->ux_callout, + mstohz(xfer->ux_timeout)); + } else { + /* No more callout or task scheduled. */ + xfer->ux_timeout_set = false; + } + valid = false; + } else if (xfer->ux_status != USBD_IN_PROGRESS) { + /* + * The xfer has completed by hardware completion or by + * software abort, and has not been resubmitted, so the + * timeout must be unset, and is no longer valid for + * the caller. + */ + xfer->ux_timeout_set = false; + valid = false; + } else { + /* + * The xfer has not yet completed, so the timeout is + * valid. + */ + valid = true; + } + + /* Any reset must have been processed. */ + KASSERT(!xfer->ux_timeout_reset); + + /* + * Either we claim the timeout is set, or the callout is idle. + * If the timeout is still set, we may be handing off to the + * task instead, so this is an if but not an iff. + */ + KASSERT(xfer->ux_timeout_set || !callout_pending(&xfer->ux_callout)); + + /* + * The task must be idle now. + * + * - If the caller is the callout, _and_ the timeout is still + * valid, the caller will schedule it, but it hasn't been + * scheduled yet. (If the timeout is not valid, the task + * should not be scheduled.) + * + * - If the caller is the task, it cannot be scheduled again + * until the callout runs again, which won't happen until we + * next release the lock. + */ + KASSERT(!usb_task_pending(xfer->ux_pipe->up_dev, &xfer->ux_aborttask)); + + KASSERT(bus->ub_usepolling || mutex_owned(bus->ub_lock)); + + return valid; +} + +/* + * usbd_xfer_schedule_timeout(xfer) + * + * Ensure that xfer has a timeout. If the callout is already + * queued or the task is already running, request that they + * reschedule the callout. If not, and if we're not polling, + * schedule the callout anew. + */ +void +usbd_xfer_schedule_timeout(struct usbd_xfer *xfer) +{ + struct usbd_bus *bus = xfer->ux_bus; + + KASSERT(bus->ub_usepolling || mutex_owned(bus->ub_lock)); + + if (xfer->ux_timeout_set) { + /* + * Callout or task has fired from a prior completed + * xfer but has not yet noticed that the xfer is done. + * Ask it to reschedule itself to ux_timeout. + */ + xfer->ux_timeout_reset = true; + } else if (xfer->ux_timeout && !bus->ub_usepolling) { + /* Callout is not scheduled. Schedule it. */ + KASSERT(!callout_pending(&xfer->ux_callout)); + callout_reset(&xfer->ux_callout, mstohz(xfer->ux_timeout), + usbd_xfer_timeout, xfer); + xfer->ux_timeout_set = true; + } + + KASSERT(bus->ub_usepolling || mutex_owned(bus->ub_lock)); +} + +/* + * usbd_xfer_cancel_timeout_async(xfer) + * + * Cancel the callout and the task of xfer, which have not yet run + * to completion, but don't wait for the callout or task to finish + * running. + * + * If they have already fired, at worst they are waiting for the + * bus lock. They will see that the xfer is no longer in progress + * and give up, or they will see that the xfer has been + * resubmitted with a new timeout and reschedule the callout. + * + * If a resubmitted request completed so fast that the callout + * didn't have time to process a timer reset, just cancel the + * timer reset. + */ +static void +usbd_xfer_cancel_timeout_async(struct usbd_xfer *xfer) +{ + struct usbd_bus *bus __diagused = xfer->ux_bus; + + KASSERT(bus->ub_usepolling || mutex_owned(bus->ub_lock)); + + KASSERT(xfer->ux_timeout_set); + xfer->ux_timeout_reset = false; + if (!callout_stop(&xfer->ux_callout)) { + /* + * We stopped the callout before it ran. The timeout + * is no longer set. + */ + xfer->ux_timeout_set = false; + } else if (callout_invoking(&xfer->ux_callout)) { + /* + * The callout has begun to run but it has not yet + * acquired the lock. The task cannot be queued yet, + * and the callout cannot have been rescheduled yet. + * + * By the time the callout acquires the lock, we will + * have transitioned from USBD_IN_PROGRESS to a + * completed status, and possibly also resubmitted the + * xfer and set xfer->ux_timeout_reset = true. In both + * cases, the callout will DTRT, so no further action + * is needed here. + */ + } else if (usb_rem_task(xfer->ux_pipe->up_dev, &xfer->ux_aborttask)) { + /* + * The callout had fired and scheduled the task, but we + * stopped the task before it could run. The timeout + * is therefore no longer set -- the next resubmission + * of the xfer must schedule a new timeout. + * + * The callout should not be be pending at this point: + * it is scheduled only under the lock, and only when + * xfer->ux_timeout_set is false, or by the callout or + * task itself when xfer->ux_timeout_reset is true. + */ + xfer->ux_timeout_set = false; + } + + /* + * The callout cannot be scheduled and the task cannot be + * queued at this point. Either we cancelled them, or they are + * already running and waiting for the bus lock. + */ + KASSERT(!callout_pending(&xfer->ux_callout)); + KASSERT(!usb_task_pending(xfer->ux_pipe->up_dev, &xfer->ux_aborttask)); + + KASSERT(bus->ub_usepolling || mutex_owned(bus->ub_lock)); +} diff --git a/sys/dev/usb/usbdi.h b/sys/dev/usb/usbdi.h index 6c228d6119cf..365ba172558f 100644 --- a/sys/dev/usb/usbdi.h +++ b/sys/dev/usb/usbdi.h @@ -185,6 +185,11 @@ int usbd_ratecheck(struct timeval *); usbd_status usbd_get_string(struct usbd_device *, int, char *); usbd_status usbd_get_string0(struct usbd_device *, int, char *, int); +/* For use by HCI drivers, not USB device drivers */ +void usbd_xfer_schedule_timeout(struct usbd_xfer *); +bool usbd_xfer_trycomplete(struct usbd_xfer *); +void usbd_xfer_abort(struct usbd_xfer *); + /* An iterator for descriptors. */ typedef struct { const uByte *cur; @@ -216,9 +221,10 @@ struct usb_task { #define USB_TASKQ_MPSAFE 0x80 void usb_add_task(struct usbd_device *, struct usb_task *, int); -void usb_rem_task(struct usbd_device *, struct usb_task *); +bool usb_rem_task(struct usbd_device *, struct usb_task *); bool usb_rem_task_wait(struct usbd_device *, struct usb_task *, int, kmutex_t *); +bool usb_task_pending(struct usbd_device *, struct usb_task *); #define usb_init_task(t, f, a, fl) ((t)->fun = (f), (t)->arg = (a), (t)->queue = USB_NUM_TASKQS, (t)->flags = (fl)) struct usb_devno { diff --git a/sys/dev/usb/usbdivar.h b/sys/dev/usb/usbdivar.h index 245cb72dfc9c..d0194732e0fb 100644 --- a/sys/dev/usb/usbdivar.h +++ b/sys/dev/usb/usbdivar.h @@ -96,6 +96,8 @@ struct usbd_bus_methods { void (*ubm_dopoll)(struct usbd_bus *); struct usbd_xfer *(*ubm_allocx)(struct usbd_bus *, unsigned int); void (*ubm_freex)(struct usbd_bus *, struct usbd_xfer *); + void (*ubm_abortx)(struct usbd_xfer *); + bool (*ubm_dying)(struct usbd_bus *); void (*ubm_getlock)(struct usbd_bus *, kmutex_t **); usbd_status (*ubm_newdev)(device_t, struct usbd_bus *, int, int, int, struct usbd_port *); @@ -288,6 +290,19 @@ struct usbd_xfer { struct usb_task ux_aborttask; struct callout ux_callout; + + /* + * Protected by bus lock. + * + * - ux_timeout_set: The timeout is scheduled as a callout or + * usb task, and has not yet acquired the bus lock. + * + * - ux_timeout_reset: The xfer completed, and was resubmitted + * before the callout or task was able to acquire the bus + * lock, so one or the other needs to schedule a new callout. + */ + bool ux_timeout_set; + bool ux_timeout_reset; }; void usbd_init(void); diff --git a/sys/dev/usb/xhci.c b/sys/dev/usb/xhci.c index aaf3daa9787b..d3a2103c460b 100644 --- a/sys/dev/usb/xhci.c +++ b/sys/dev/usb/xhci.c @@ -140,6 +140,8 @@ static void xhci_softintr(void *); static void xhci_poll(struct usbd_bus *); static struct usbd_xfer *xhci_allocx(struct usbd_bus *, unsigned int); static void xhci_freex(struct usbd_bus *, struct usbd_xfer *); +static void xhci_abortx(struct usbd_xfer *); +static bool xhci_dying(struct usbd_bus *); static void xhci_get_lock(struct usbd_bus *, kmutex_t **); static usbd_status xhci_new_device(device_t, struct usbd_bus *, int, int, int, struct usbd_port *); @@ -205,15 +207,14 @@ static void xhci_device_bulk_abort(struct usbd_xfer *); static void xhci_device_bulk_close(struct usbd_pipe *); static void xhci_device_bulk_done(struct usbd_xfer *); -static void xhci_timeout(void *); -static void xhci_timeout_task(void *); - static const struct usbd_bus_methods xhci_bus_methods = { .ubm_open = xhci_open, .ubm_softint = xhci_softintr, .ubm_dopoll = xhci_poll, .ubm_allocx = xhci_allocx, .ubm_freex = xhci_freex, + .ubm_abortx = xhci_abortx, + .ubm_dying = xhci_dying, .ubm_getlock = xhci_get_lock, .ubm_newdev = xhci_new_device, .ubm_rhctrl = xhci_roothub_ctrl, @@ -1714,53 +1715,22 @@ xhci_close_pipe(struct usbd_pipe *pipe) * Should be called with sc_lock held. */ static void -xhci_abort_xfer(struct usbd_xfer *xfer, usbd_status status) +xhci_abortx(struct usbd_xfer *xfer) { XHCIHIST_FUNC(); XHCIHIST_CALLED(); struct xhci_softc * const sc = XHCI_XFER2SC(xfer); struct xhci_slot * const xs = xfer->ux_pipe->up_dev->ud_hcpriv; const u_int dci = xhci_ep_get_dci(xfer->ux_pipe->up_endpoint->ue_edesc); - KASSERTMSG((status == USBD_CANCELLED || status == USBD_TIMEOUT), - "invalid status for abort: %d", (int)status); - DPRINTFN(4, "xfer %#jx pipe %#jx status %jd", (uintptr_t)xfer, (uintptr_t)xfer->ux_pipe, status, 0); KASSERT(mutex_owned(&sc->sc_lock)); ASSERT_SLEEPABLE(); - if (status == USBD_CANCELLED) { - /* - * We are synchronously aborting. Try to stop the - * callout and task, but if we can't, wait for them to - * complete. - */ - callout_halt(&xfer->ux_callout, &sc->sc_lock); - usb_rem_task_wait(xfer->ux_pipe->up_dev, &xfer->ux_aborttask, - USB_TASKQ_HC, &sc->sc_lock); - } else { - /* Otherwise, we are timing out. */ - KASSERT(status == USBD_TIMEOUT); - } - - /* - * The xfer cannot have been cancelled already. It is the - * responsibility of the caller of usbd_abort_pipe not to try - * to abort a pipe multiple times, whether concurrently or - * sequentially. - */ - KASSERT(xfer->ux_status != USBD_CANCELLED); - - /* Only the timeout, which runs only once, can time it out. */ - KASSERT(xfer->ux_status != USBD_TIMEOUT); - - /* If anyone else beat us, we're done. */ - if (xfer->ux_status != USBD_IN_PROGRESS) - return; - - /* We beat everyone else. Claim the status. */ - xfer->ux_status = status; + KASSERTMSG((xfer->ux_status == USBD_CANCELLED || + xfer->ux_status == USBD_TIMEOUT), + "bad abort status: %d", xfer->ux_status); /* * If we're dying, skip the hardware action and just notify the @@ -1904,6 +1874,13 @@ xhci_rhpsc(struct xhci_softc * const sc, u_int ctlrport) if (xfer == NULL) return; + /* + * Try to claim this xfer for completion. If it has already + * completed or aborted, drop it on the floor. + */ + if (!usbd_xfer_trycomplete(xfer)) + return; + uint8_t *p = xfer->ux_buf; memset(p, 0, xfer->ux_length); p[rhp / NBBY] |= 1 << (rhp % NBBY); @@ -1992,6 +1969,13 @@ xhci_event_transfer(struct xhci_softc * const sc, return; } + /* + * Try to claim this xfer for completion. If it has already + * completed or aborted, drop it on the floor. + */ + if (!usbd_xfer_trycomplete(xfer)) + return; + /* 4.11.5.2 Event Data TRB */ if ((trb_3 & XHCI_TRB_3_ED_BIT) != 0) { DPRINTFN(14, "transfer Event Data: 0x%016jx 0x%08jx" @@ -2040,17 +2024,7 @@ xhci_event_transfer(struct xhci_softc * const sc, case XHCI_TRB_ERROR_STOPPED: case XHCI_TRB_ERROR_LENGTH: case XHCI_TRB_ERROR_STOPPED_SHORT: - /* - * don't complete the transfer being aborted - * as abort_xfer does instead. - */ - if (xfer->ux_status == USBD_CANCELLED || - xfer->ux_status == USBD_TIMEOUT) { - DPRINTFN(14, "ignore aborting xfer %#jx", - (uintptr_t)xfer, 0, 0, 0); - return; - } - err = USBD_CANCELLED; + err = USBD_IOERROR; break; case XHCI_TRB_ERROR_STALL: case XHCI_TRB_ERROR_BABBLE: @@ -2073,15 +2047,6 @@ xhci_event_transfer(struct xhci_softc * const sc, /* Override the status. */ xfer->ux_status = USBD_STALLED; - /* - * Cancel the timeout and the task, which have not yet - * run. If they have already fired, at worst they are - * waiting for the lock. They will see that the xfer - * is no longer in progress and give up. - */ - callout_stop(&xfer->ux_callout); - usb_rem_task(xfer->ux_pipe->up_dev, &xfer->ux_aborttask); - xhci_clear_endpoint_stall_async(xfer); return; default: @@ -2090,29 +2055,9 @@ xhci_event_transfer(struct xhci_softc * const sc, break; } - /* - * If software has completed it, either by cancellation - * or timeout, drop it on the floor. - */ - if (xfer->ux_status != USBD_IN_PROGRESS) { - KASSERTMSG((xfer->ux_status == USBD_CANCELLED || - xfer->ux_status == USBD_TIMEOUT), - "xfer %p status %x", xfer, xfer->ux_status); - return; - } - - /* Otherwise, set the status. */ + /* Set the status. */ xfer->ux_status = err; - /* - * Cancel the timeout and the task, which have not yet - * run. If they have already fired, at worst they are - * waiting for the lock. They will see that the xfer - * is no longer in progress and give up. - */ - callout_stop(&xfer->ux_callout); - usb_rem_task(xfer->ux_pipe->up_dev, &xfer->ux_aborttask); - if ((trb_3 & XHCI_TRB_3_ED_BIT) == 0 || (trb_0 & 0x3) == 0x0) { usb_transfer_complete(xfer); @@ -2279,8 +2224,6 @@ xhci_allocx(struct usbd_bus *bus, unsigned int nframes) xfer = pool_cache_get(sc->sc_xferpool, PR_WAITOK); if (xfer != NULL) { memset(xfer, 0, sizeof(struct xhci_xfer)); - usb_init_task(&xfer->ux_aborttask, xhci_timeout_task, xfer, - USB_TASKQ_MPSAFE); #ifdef DIAGNOSTIC xfer->ux_state = XFER_BUSY; #endif @@ -2307,6 +2250,14 @@ xhci_freex(struct usbd_bus *bus, struct usbd_xfer *xfer) pool_cache_put(sc->sc_xferpool, xfer); } +static bool +xhci_dying(struct usbd_bus *bus) +{ + struct xhci_softc * const sc = XHCI_BUS2SC(bus); + + return sc->sc_dying; +} + static void xhci_get_lock(struct usbd_bus *bus, kmutex_t **lock) { @@ -3882,11 +3833,7 @@ xhci_device_ctrl_start(struct usbd_xfer *xfer) mutex_exit(&tr->xr_lock); xhci_db_write_4(sc, XHCI_DOORBELL(xs->xs_idx), dci); - - if (xfer->ux_timeout && !xhci_polling_p(sc)) { - callout_reset(&xfer->ux_callout, mstohz(xfer->ux_timeout), - xhci_timeout, xfer); - } + usbd_xfer_schedule_timeout(xfer); return USBD_IN_PROGRESS; } @@ -3909,7 +3856,7 @@ xhci_device_ctrl_abort(struct usbd_xfer *xfer) { XHCIHIST_FUNC(); XHCIHIST_CALLED(); - xhci_abort_xfer(xfer, USBD_CANCELLED); + usbd_xfer_abort(xfer); } static void @@ -4003,11 +3950,7 @@ xhci_device_bulk_start(struct usbd_xfer *xfer) mutex_exit(&tr->xr_lock); xhci_db_write_4(sc, XHCI_DOORBELL(xs->xs_idx), dci); - - if (xfer->ux_timeout && !xhci_polling_p(sc)) { - callout_reset(&xfer->ux_callout, mstohz(xfer->ux_timeout), - xhci_timeout, xfer); - } + usbd_xfer_schedule_timeout(xfer); return USBD_IN_PROGRESS; } @@ -4035,7 +3978,7 @@ xhci_device_bulk_abort(struct usbd_xfer *xfer) { XHCIHIST_FUNC(); XHCIHIST_CALLED(); - xhci_abort_xfer(xfer, USBD_CANCELLED); + usbd_xfer_abort(xfer); } static void @@ -4115,11 +4058,7 @@ xhci_device_intr_start(struct usbd_xfer *xfer) mutex_exit(&tr->xr_lock); xhci_db_write_4(sc, XHCI_DOORBELL(xs->xs_idx), dci); - - if (xfer->ux_timeout && !polling) { - callout_reset(&xfer->ux_callout, mstohz(xfer->ux_timeout), - xhci_timeout, xfer); - } + usbd_xfer_schedule_timeout(xfer); return USBD_IN_PROGRESS; } @@ -4155,7 +4094,7 @@ xhci_device_intr_abort(struct usbd_xfer *xfer) KASSERT(mutex_owned(&sc->sc_lock)); DPRINTFN(15, "%#jx", (uintptr_t)xfer, 0, 0, 0); KASSERT(xfer->ux_pipe->up_intrxfer == xfer); - xhci_abort_xfer(xfer, USBD_CANCELLED); + usbd_xfer_abort(xfer); } static void @@ -4168,32 +4107,3 @@ xhci_device_intr_close(struct usbd_pipe *pipe) xhci_close_pipe(pipe); } - -/* ------------ */ - -static void -xhci_timeout(void *addr) -{ - XHCIHIST_FUNC(); XHCIHIST_CALLED(); - struct xhci_xfer * const xx = addr; - struct usbd_xfer * const xfer = &xx->xx_xfer; - struct xhci_softc * const sc = XHCI_XFER2SC(xfer); - struct usbd_device *dev = xfer->ux_pipe->up_dev; - - mutex_enter(&sc->sc_lock); - if (!sc->sc_dying && xfer->ux_status == USBD_IN_PROGRESS) - usb_add_task(dev, &xfer->ux_aborttask, USB_TASKQ_HC); - mutex_exit(&sc->sc_lock); -} - -static void -xhci_timeout_task(void *addr) -{ - XHCIHIST_FUNC(); XHCIHIST_CALLED(); - struct usbd_xfer * const xfer = addr; - struct xhci_softc * const sc = XHCI_XFER2SC(xfer); - - mutex_enter(&sc->sc_lock); - xhci_abort_xfer(xfer, USBD_TIMEOUT); - mutex_exit(&sc->sc_lock); -}