Lines Matching refs:qh

463 	if (!list_empty(&hwep->qh.queue)) {  in _hardware_enqueue()
470 hwreqprev = list_entry(hwep->qh.queue.prev, in _hardware_enqueue()
489 hwep->qh.ptr->td.next = cpu_to_le32(firstnode->dma); in _hardware_enqueue()
490 hwep->qh.ptr->td.token &= in _hardware_enqueue()
499 hwep->qh.ptr->cap |= mul << __ffs(QH_MULT); in _hardware_enqueue()
526 hwep->qh.ptr->td.next = node->dma; in reprime_dtd()
527 hwep->qh.ptr->td.token &= in reprime_dtd()
630 while (!list_empty(&hwep->qh.queue)) { in _ep_nuke()
633 struct ci_hw_req *hwreq = list_entry(hwep->qh.queue.next, in _ep_nuke()
674 !list_empty(&hwep->qh.queue) && in _ep_set_halt()
809 if (!list_empty(&hwep->qh.queue)) { in _ep_queue()
838 list_add_tail(&hwreq->queue, &hwep->qh.queue); in _ep_queue()
966 list_for_each_entry_safe(hwreq, hwreqtemp, &hwep->qh.queue, in isr_tr_complete_low()
1020 memcpy(&req, &hwep->qh.ptr->setup, sizeof(req)); in isr_setup_packet_handler()
1228 if (!list_empty(&hwep->qh.queue)) { in ep_enable()
1255 hwep->qh.ptr->cap = cpu_to_le32(cap); in ep_enable()
1257 hwep->qh.ptr->td.next |= cpu_to_le32(TD_TERMINATE); /* needed? */ in ep_enable()
1401 list_empty(&hwep->qh.queue)) in ep_dequeue()
1518 hw_device_state(ci, ci->ep0out->qh.dma); in ci_udc_vbus_session()
1659 INIT_LIST_HEAD(&hwep->qh.queue); in init_eps()
1660 hwep->qh.ptr = dma_pool_alloc(ci->qh_pool, GFP_KERNEL, in init_eps()
1661 &hwep->qh.dma); in init_eps()
1662 if (hwep->qh.ptr == NULL) in init_eps()
1665 memset(hwep->qh.ptr, 0, sizeof(*hwep->qh.ptr)); in init_eps()
1696 dma_pool_free(ci->qh_pool, hwep->qh.ptr, hwep->qh.dma); in destroy_eps()
1746 retval = hw_device_state(ci, ci->ep0out->qh.dma); in ci_udc_start()