Lines Matching refs:qh

463 	if (!list_empty(&hwep->qh.queue)) {  in _hardware_enqueue()
470 hwreqprev = list_entry(hwep->qh.queue.prev, in _hardware_enqueue()
489 hwep->qh.ptr->td.next = cpu_to_le32(firstnode->dma); in _hardware_enqueue()
490 hwep->qh.ptr->td.token &= in _hardware_enqueue()
499 hwep->qh.ptr->cap |= mul << __ffs(QH_MULT); in _hardware_enqueue()
526 hwep->qh.ptr->td.next = node->dma; in reprime_dtd()
527 hwep->qh.ptr->td.token &= in reprime_dtd()
630 while (!list_empty(&hwep->qh.queue)) { in _ep_nuke()
633 struct ci_hw_req *hwreq = list_entry(hwep->qh.queue.next, in _ep_nuke()
674 !list_empty(&hwep->qh.queue) && in _ep_set_halt()
809 if (!list_empty(&hwep->qh.queue)) { in _ep_queue()
838 list_add_tail(&hwreq->queue, &hwep->qh.queue); in _ep_queue()
966 list_for_each_entry_safe(hwreq, hwreqtemp, &hwep->qh.queue, in isr_tr_complete_low()
1020 memcpy(&req, &hwep->qh.ptr->setup, sizeof(req)); in isr_setup_packet_handler()
1221 if (!list_empty(&hwep->qh.queue)) { in ep_enable()
1248 hwep->qh.ptr->cap = cpu_to_le32(cap); in ep_enable()
1250 hwep->qh.ptr->td.next |= cpu_to_le32(TD_TERMINATE); /* needed? */ in ep_enable()
1394 list_empty(&hwep->qh.queue)) in ep_dequeue()
1511 hw_device_state(ci, ci->ep0out->qh.dma); in ci_udc_vbus_session()
1638 INIT_LIST_HEAD(&hwep->qh.queue); in init_eps()
1639 hwep->qh.ptr = dma_pool_alloc(ci->qh_pool, GFP_KERNEL, in init_eps()
1640 &hwep->qh.dma); in init_eps()
1641 if (hwep->qh.ptr == NULL) in init_eps()
1644 memset(hwep->qh.ptr, 0, sizeof(*hwep->qh.ptr)); in init_eps()
1675 dma_pool_free(ci->qh_pool, hwep->qh.ptr, hwep->qh.dma); in destroy_eps()
1725 retval = hw_device_state(ci, ci->ep0out->qh.dma); in ci_udc_start()