wa 374 arch/arm/mm/cache-l2x0-pmu.c L220_PLUS_EVENT_ATTR(wa, 0x9), wa 485 drivers/clocksource/arm_arch_timer.c bool arch_timer_check_dt_erratum(const struct arch_timer_erratum_workaround *wa, wa 490 drivers/clocksource/arm_arch_timer.c return of_property_read_bool(np, wa->id); wa 494 drivers/clocksource/arm_arch_timer.c bool arch_timer_check_local_cap_erratum(const struct arch_timer_erratum_workaround *wa, wa 497 drivers/clocksource/arm_arch_timer.c return this_cpu_has_cap((uintptr_t)wa->id); wa 502 drivers/clocksource/arm_arch_timer.c bool arch_timer_check_acpi_oem_erratum(const struct arch_timer_erratum_workaround *wa, wa 506 drivers/clocksource/arm_arch_timer.c const struct ate_acpi_oem_info *info = wa->id; wa 541 drivers/clocksource/arm_arch_timer.c void arch_timer_enable_workaround(const struct arch_timer_erratum_workaround *wa, wa 547 drivers/clocksource/arm_arch_timer.c __this_cpu_write(timer_unstable_counter_workaround, wa); wa 550 drivers/clocksource/arm_arch_timer.c per_cpu(timer_unstable_counter_workaround, i) = wa; wa 553 drivers/clocksource/arm_arch_timer.c if (wa->read_cntvct_el0 || wa->read_cntpct_el0) wa 562 drivers/clocksource/arm_arch_timer.c if (wa->read_cntvct_el0) { wa 571 drivers/clocksource/arm_arch_timer.c const struct arch_timer_erratum_workaround *wa, *__wa; wa 591 drivers/clocksource/arm_arch_timer.c wa = arch_timer_iterate_errata(type, match_fn, arg); wa 592 drivers/clocksource/arm_arch_timer.c if (!wa) wa 596 drivers/clocksource/arm_arch_timer.c if (__wa && wa != __wa) wa 598 drivers/clocksource/arm_arch_timer.c wa->desc, __wa->desc); wa 603 drivers/clocksource/arm_arch_timer.c arch_timer_enable_workaround(wa, local); wa 605 drivers/clocksource/arm_arch_timer.c local ? "local" : "global", wa->desc); wa 63 drivers/crypto/ccp/ccp-ops.c static void ccp_sg_free(struct ccp_sg_workarea *wa) wa 65 drivers/crypto/ccp/ccp-ops.c if (wa->dma_count) wa 66 drivers/crypto/ccp/ccp-ops.c dma_unmap_sg(wa->dma_dev, wa->dma_sg, wa->nents, wa->dma_dir); wa 68 drivers/crypto/ccp/ccp-ops.c wa->dma_count = 0; wa 71 drivers/crypto/ccp/ccp-ops.c static int ccp_init_sg_workarea(struct ccp_sg_workarea *wa, struct device *dev, wa 75 drivers/crypto/ccp/ccp-ops.c memset(wa, 0, sizeof(*wa)); wa 77 drivers/crypto/ccp/ccp-ops.c wa->sg = sg; wa 81 drivers/crypto/ccp/ccp-ops.c wa->nents = sg_nents_for_len(sg, len); wa 82 drivers/crypto/ccp/ccp-ops.c if (wa->nents < 0) wa 83 drivers/crypto/ccp/ccp-ops.c return wa->nents; wa 85 drivers/crypto/ccp/ccp-ops.c wa->bytes_left = len; wa 86 drivers/crypto/ccp/ccp-ops.c wa->sg_used = 0; wa 94 drivers/crypto/ccp/ccp-ops.c wa->dma_sg = sg; wa 95 drivers/crypto/ccp/ccp-ops.c wa->dma_dev = dev; wa 96 drivers/crypto/ccp/ccp-ops.c wa->dma_dir = dma_dir; wa 97 drivers/crypto/ccp/ccp-ops.c wa->dma_count = dma_map_sg(dev, sg, wa->nents, dma_dir); wa 98 drivers/crypto/ccp/ccp-ops.c if (!wa->dma_count) wa 104 drivers/crypto/ccp/ccp-ops.c static void ccp_update_sg_workarea(struct ccp_sg_workarea *wa, unsigned int len) wa 106 drivers/crypto/ccp/ccp-ops.c unsigned int nbytes = min_t(u64, len, wa->bytes_left); wa 108 drivers/crypto/ccp/ccp-ops.c if (!wa->sg) wa 111 drivers/crypto/ccp/ccp-ops.c wa->sg_used += nbytes; wa 112 drivers/crypto/ccp/ccp-ops.c wa->bytes_left -= nbytes; wa 113 drivers/crypto/ccp/ccp-ops.c if (wa->sg_used == wa->sg->length) { wa 114 drivers/crypto/ccp/ccp-ops.c wa->sg = sg_next(wa->sg); wa 115 drivers/crypto/ccp/ccp-ops.c wa->sg_used = 0; wa 119 drivers/crypto/ccp/ccp-ops.c static void ccp_dm_free(struct ccp_dm_workarea *wa) wa 121 drivers/crypto/ccp/ccp-ops.c if (wa->length <= CCP_DMAPOOL_MAX_SIZE) { wa 122 drivers/crypto/ccp/ccp-ops.c if (wa->address) wa 123 drivers/crypto/ccp/ccp-ops.c dma_pool_free(wa->dma_pool, wa->address, wa 124 drivers/crypto/ccp/ccp-ops.c wa->dma.address); wa 126 drivers/crypto/ccp/ccp-ops.c if (wa->dma.address) wa 127 drivers/crypto/ccp/ccp-ops.c dma_unmap_single(wa->dev, wa->dma.address, wa->length, wa 128 drivers/crypto/ccp/ccp-ops.c wa->dma.dir); wa 129 drivers/crypto/ccp/ccp-ops.c kfree(wa->address); wa 132 drivers/crypto/ccp/ccp-ops.c wa->address = NULL; wa 133 drivers/crypto/ccp/ccp-ops.c wa->dma.address = 0; wa 136 drivers/crypto/ccp/ccp-ops.c static int ccp_init_dm_workarea(struct ccp_dm_workarea *wa, wa 141 drivers/crypto/ccp/ccp-ops.c memset(wa, 0, sizeof(*wa)); wa 146 drivers/crypto/ccp/ccp-ops.c wa->dev = cmd_q->ccp->dev; wa 147 drivers/crypto/ccp/ccp-ops.c wa->length = len; wa 150 drivers/crypto/ccp/ccp-ops.c wa->dma_pool = cmd_q->dma_pool; wa 152 drivers/crypto/ccp/ccp-ops.c wa->address = dma_pool_zalloc(wa->dma_pool, GFP_KERNEL, wa 153 drivers/crypto/ccp/ccp-ops.c &wa->dma.address); wa 154 drivers/crypto/ccp/ccp-ops.c if (!wa->address) wa 157 drivers/crypto/ccp/ccp-ops.c wa->dma.length = CCP_DMAPOOL_MAX_SIZE; wa 160 drivers/crypto/ccp/ccp-ops.c wa->address = kzalloc(len, GFP_KERNEL); wa 161 drivers/crypto/ccp/ccp-ops.c if (!wa->address) wa 164 drivers/crypto/ccp/ccp-ops.c wa->dma.address = dma_map_single(wa->dev, wa->address, len, wa 166 drivers/crypto/ccp/ccp-ops.c if (dma_mapping_error(wa->dev, wa->dma.address)) wa 169 drivers/crypto/ccp/ccp-ops.c wa->dma.length = len; wa 171 drivers/crypto/ccp/ccp-ops.c wa->dma.dir = dir; wa 176 drivers/crypto/ccp/ccp-ops.c static int ccp_set_dm_area(struct ccp_dm_workarea *wa, unsigned int wa_offset, wa 180 drivers/crypto/ccp/ccp-ops.c WARN_ON(!wa->address); wa 182 drivers/crypto/ccp/ccp-ops.c if (len > (wa->length - wa_offset)) wa 185 drivers/crypto/ccp/ccp-ops.c scatterwalk_map_and_copy(wa->address + wa_offset, sg, sg_offset, len, wa 190 drivers/crypto/ccp/ccp-ops.c static void ccp_get_dm_area(struct ccp_dm_workarea *wa, unsigned int wa_offset, wa 194 drivers/crypto/ccp/ccp-ops.c WARN_ON(!wa->address); wa 196 drivers/crypto/ccp/ccp-ops.c scatterwalk_map_and_copy(wa->address + wa_offset, sg, sg_offset, len, wa 200 drivers/crypto/ccp/ccp-ops.c static int ccp_reverse_set_dm_area(struct ccp_dm_workarea *wa, wa 209 drivers/crypto/ccp/ccp-ops.c rc = ccp_set_dm_area(wa, wa_offset, sg, sg_offset, len); wa 213 drivers/crypto/ccp/ccp-ops.c p = wa->address + wa_offset; wa 225 drivers/crypto/ccp/ccp-ops.c static void ccp_reverse_get_dm_area(struct ccp_dm_workarea *wa, wa 233 drivers/crypto/ccp/ccp-ops.c p = wa->address + wa_offset; wa 243 drivers/crypto/ccp/ccp-ops.c ccp_get_dm_area(wa, wa_offset, sg, sg_offset, len); wa 411 drivers/crypto/ccp/ccp-ops.c struct ccp_dm_workarea *wa, u32 jobid, u32 sb, wa 427 drivers/crypto/ccp/ccp-ops.c op.dst.u.dma.address = wa->dma.address; wa 428 drivers/crypto/ccp/ccp-ops.c op.dst.u.dma.length = wa->length; wa 431 drivers/crypto/ccp/ccp-ops.c op.src.u.dma.address = wa->dma.address; wa 432 drivers/crypto/ccp/ccp-ops.c op.src.u.dma.length = wa->length; wa 443 drivers/crypto/ccp/ccp-ops.c struct ccp_dm_workarea *wa, u32 jobid, u32 sb, wa 446 drivers/crypto/ccp/ccp-ops.c return ccp_copy_to_from_sb(cmd_q, wa, jobid, sb, byte_swap, false); wa 450 drivers/crypto/ccp/ccp-ops.c struct ccp_dm_workarea *wa, u32 jobid, u32 sb, wa 453 drivers/crypto/ccp/ccp-ops.c return ccp_copy_to_from_sb(cmd_q, wa, jobid, sb, byte_swap, true); wa 226 drivers/gpu/drm/amd/display/dc/core/dc_link_ddc.c ddc_service->wa.raw = 0; wa 165 drivers/gpu/drm/amd/display/dc/dc_ddc_types.h union ddc_wa wa; wa 80 drivers/gpu/drm/amd/display/dc/dce/dce_hwseq.c if (hws->wa.blnd_crtc_trigger) { wa 608 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.c if (dce_mi->wa.single_head_rdreq_dmif_limit) { wa 610 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.c dce_mi->wa.single_head_rdreq_dmif_limit; wa 638 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.c if (dce_mi->wa.single_head_rdreq_dmif_limit) { wa 640 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.c dce_mi->wa.single_head_rdreq_dmif_limit; wa 337 drivers/gpu/drm/amd/display/dc/dce/dce_mem_input.h struct dce_mem_input_wa wa; wa 522 drivers/gpu/drm/amd/display/dc/dce100/dce100_resource.c dce_mi->wa.single_head_rdreq_dmif_limit = 2; wa 522 drivers/gpu/drm/amd/display/dc/dce110/dce110_resource.c hws->wa.blnd_crtc_trigger = true; wa 568 drivers/gpu/drm/amd/display/dc/dce110/dce110_resource.c dce_mi->wa.single_head_rdreq_dmif_limit = 3; wa 634 drivers/gpu/drm/amd/display/dc/dce80/dce80_resource.c dce_mi->wa.single_head_rdreq_dmif_limit = 2; wa 646 drivers/gpu/drm/amd/display/dc/dcn10/dcn10_hw_sequencer.c if (!hws->wa.DEGVIDCN10_253) wa 710 drivers/gpu/drm/amd/display/dc/dcn10/dcn10_hw_sequencer.c if (!dc->hwseq->wa.false_optc_underflow) wa 2665 drivers/gpu/drm/amd/display/dc/dcn10/dcn10_hw_sequencer.c if (dc->hwseq->wa.DEGVIDCN10_254) wa 847 drivers/gpu/drm/amd/display/dc/dcn10/dcn10_resource.c hws->wa.DEGVIDCN10_253 = true; wa 848 drivers/gpu/drm/amd/display/dc/dcn10/dcn10_resource.c hws->wa.false_optc_underflow = true; wa 849 drivers/gpu/drm/amd/display/dc/dcn10/dcn10_resource.c hws->wa.DEGVIDCN10_254 = true; wa 63 drivers/gpu/drm/amd/display/dc/inc/hw_sequencer.h struct dce_hwseq_wa wa; wa 82 drivers/gpu/drm/i915/gt/intel_workarounds.c static void _wa_add(struct i915_wa_list *wal, const struct i915_wa *wa) wa 84 drivers/gpu/drm/i915/gt/intel_workarounds.c unsigned int addr = i915_mmio_reg_offset(wa->reg); wa 94 drivers/gpu/drm/i915/gt/intel_workarounds.c list = kmalloc_array(ALIGN(wal->count + 1, grow), sizeof(*wa), wa 102 drivers/gpu/drm/i915/gt/intel_workarounds.c memcpy(list, wal->list, sizeof(*wa) * wal->count); wa 117 drivers/gpu/drm/i915/gt/intel_workarounds.c if ((wa->mask & ~wa_->mask) == 0) { wa 122 drivers/gpu/drm/i915/gt/intel_workarounds.c wa_->val &= ~wa->mask; wa 126 drivers/gpu/drm/i915/gt/intel_workarounds.c wa_->val |= wa->val; wa 127 drivers/gpu/drm/i915/gt/intel_workarounds.c wa_->mask |= wa->mask; wa 128 drivers/gpu/drm/i915/gt/intel_workarounds.c wa_->read |= wa->read; wa 135 drivers/gpu/drm/i915/gt/intel_workarounds.c *wa_ = *wa; wa 152 drivers/gpu/drm/i915/gt/intel_workarounds.c struct i915_wa wa = { wa 159 drivers/gpu/drm/i915/gt/intel_workarounds.c _wa_add(wal, &wa); wa 620 drivers/gpu/drm/i915/gt/intel_workarounds.c struct i915_wa *wa; wa 637 drivers/gpu/drm/i915/gt/intel_workarounds.c for (i = 0, wa = wal->list; i < wal->count; i++, wa++) { wa 638 drivers/gpu/drm/i915/gt/intel_workarounds.c *cs++ = i915_mmio_reg_offset(wa->reg); wa 639 drivers/gpu/drm/i915/gt/intel_workarounds.c *cs++ = wa->val; wa 938 drivers/gpu/drm/i915/gt/intel_workarounds.c struct i915_wa *wa; wa 941 drivers/gpu/drm/i915/gt/intel_workarounds.c for (i = 0, wa = wal->list; i < wal->count; i++, wa++) wa 943 drivers/gpu/drm/i915/gt/intel_workarounds.c wa->reg, wa 951 drivers/gpu/drm/i915/gt/intel_workarounds.c wa_verify(const struct i915_wa *wa, u32 cur, const char *name, const char *from) wa 953 drivers/gpu/drm/i915/gt/intel_workarounds.c if ((cur ^ wa->val) & wa->read) { wa 955 drivers/gpu/drm/i915/gt/intel_workarounds.c name, from, i915_mmio_reg_offset(wa->reg), wa 956 drivers/gpu/drm/i915/gt/intel_workarounds.c cur, cur & wa->read, wa 957 drivers/gpu/drm/i915/gt/intel_workarounds.c wa->val, wa->mask); wa 970 drivers/gpu/drm/i915/gt/intel_workarounds.c struct i915_wa *wa; wa 981 drivers/gpu/drm/i915/gt/intel_workarounds.c for (i = 0, wa = wal->list; i < wal->count; i++, wa++) { wa 982 drivers/gpu/drm/i915/gt/intel_workarounds.c intel_uncore_rmw_fw(uncore, wa->reg, wa->mask, wa->val); wa 984 drivers/gpu/drm/i915/gt/intel_workarounds.c wa_verify(wa, wa 985 drivers/gpu/drm/i915/gt/intel_workarounds.c intel_uncore_read_fw(uncore, wa->reg), wa 1002 drivers/gpu/drm/i915/gt/intel_workarounds.c struct i915_wa *wa; wa 1006 drivers/gpu/drm/i915/gt/intel_workarounds.c for (i = 0, wa = wal->list; i < wal->count; i++, wa++) wa 1007 drivers/gpu/drm/i915/gt/intel_workarounds.c ok &= wa_verify(wa, wa 1008 drivers/gpu/drm/i915/gt/intel_workarounds.c intel_uncore_read(uncore, wa->reg), wa 1036 drivers/gpu/drm/i915/gt/intel_workarounds.c struct i915_wa wa = { wa 1046 drivers/gpu/drm/i915/gt/intel_workarounds.c wa.reg.reg |= flags; wa 1047 drivers/gpu/drm/i915/gt/intel_workarounds.c _wa_add(wal, &wa); wa 1238 drivers/gpu/drm/i915/gt/intel_workarounds.c struct i915_wa *wa; wa 1244 drivers/gpu/drm/i915/gt/intel_workarounds.c for (i = 0, wa = wal->list; i < wal->count; i++, wa++) wa 1247 drivers/gpu/drm/i915/gt/intel_workarounds.c i915_mmio_reg_offset(wa->reg)); wa 1468 drivers/gpu/drm/i915/gt/intel_workarounds.c const struct i915_wa *wa; wa 1475 drivers/gpu/drm/i915/gt/intel_workarounds.c for (i = 0, wa = wal->list; i < wal->count; i++, wa++) { wa 1476 drivers/gpu/drm/i915/gt/intel_workarounds.c if (!mcr_range(i915, i915_mmio_reg_offset(wa->reg))) wa 1484 drivers/gpu/drm/i915/gt/intel_workarounds.c for (i = 0, wa = wal->list; i < wal->count; i++, wa++) { wa 1485 drivers/gpu/drm/i915/gt/intel_workarounds.c u32 offset = i915_mmio_reg_offset(wa->reg); wa 1504 drivers/gpu/drm/i915/gt/intel_workarounds.c const struct i915_wa *wa; wa 1541 drivers/gpu/drm/i915/gt/intel_workarounds.c for (i = 0, wa = wal->list; i < wal->count; i++, wa++) { wa 1542 drivers/gpu/drm/i915/gt/intel_workarounds.c if (mcr_range(rq->i915, i915_mmio_reg_offset(wa->reg))) wa 1545 drivers/gpu/drm/i915/gt/intel_workarounds.c if (!wa_verify(wa, results[i], wal->name, from)) wa 958 drivers/gpu/drm/i915/gt/selftest_workarounds.c const struct i915_wa *wa = &engine->whitelist.list[i]; wa 960 drivers/gpu/drm/i915/gt/selftest_workarounds.c if (i915_mmio_reg_offset(wa->reg) & wa 964 drivers/gpu/drm/i915/gt/selftest_workarounds.c if (!fn(engine, a[i], b[i], wa->reg)) wa 2880 drivers/gpu/drm/i915/i915_debugfs.c const struct i915_wa *wa; wa 2890 drivers/gpu/drm/i915/i915_debugfs.c for (wa = wal->list; count--; wa++) wa 2892 drivers/gpu/drm/i915/i915_debugfs.c i915_mmio_reg_offset(wa->reg), wa 2893 drivers/gpu/drm/i915/i915_debugfs.c wa->val, wa->mask); wa 1794 drivers/lightnvm/pblk-core.c struct wa_counters *wa = emeta_to_wa(lm, emeta_buf); wa 1800 drivers/lightnvm/pblk-core.c wa->user = cpu_to_le64(atomic64_read(&pblk->user_wa)); wa 1801 drivers/lightnvm/pblk-core.c wa->pad = cpu_to_le64(atomic64_read(&pblk->pad_wa)); wa 1802 drivers/lightnvm/pblk-core.c wa->gc = cpu_to_le64(atomic64_read(&pblk->gc_wa)); wa 592 drivers/lightnvm/pblk-recovery.c struct wa_counters *wa = emeta_to_wa(lm, emeta); wa 596 drivers/lightnvm/pblk-recovery.c u64 user = le64_to_cpu(wa->user); wa 597 drivers/lightnvm/pblk-recovery.c u64 pad = le64_to_cpu(wa->pad); wa 598 drivers/lightnvm/pblk-recovery.c u64 gc = le64_to_cpu(wa->gc); wa 53 drivers/staging/wusbcore/host/hwa-hc.c struct wahc wa; wa 66 drivers/staging/wusbcore/host/hwa-hc.c struct wahc *wa = &hwahc->wa; wa 67 drivers/staging/wusbcore/host/hwa-hc.c struct device *dev = &wa->usb_iface->dev; wa 69 drivers/staging/wusbcore/host/hwa-hc.c result = usb_control_msg(wa->usb_dev, usb_sndctrlpipe(wa->usb_dev, 0), wa 73 drivers/staging/wusbcore/host/hwa-hc.c wa->usb_iface->cur_altsetting->desc.bInterfaceNumber, wa 87 drivers/staging/wusbcore/host/hwa-hc.c struct wahc *wa = &hwahc->wa; wa 89 drivers/staging/wusbcore/host/hwa-hc.c return usb_control_msg(wa->usb_dev, usb_sndctrlpipe(wa->usb_dev, 0), wa 93 drivers/staging/wusbcore/host/hwa-hc.c wa->usb_iface->cur_altsetting->desc.bInterfaceNumber, wa 108 drivers/staging/wusbcore/host/hwa-hc.c struct device *dev = &hwahc->wa.usb_iface->dev; wa 111 drivers/staging/wusbcore/host/hwa-hc.c wa_nep_disarm(&hwahc->wa); wa 112 drivers/staging/wusbcore/host/hwa-hc.c result = __wa_set_feature(&hwahc->wa, WA_RESET); wa 117 drivers/staging/wusbcore/host/hwa-hc.c result = __wa_wait_status(&hwahc->wa, WA_STATUS_RESETTING, 0); wa 187 drivers/staging/wusbcore/host/hwa-hc.c struct wahc *wa = &hwahc->wa; wa 194 drivers/staging/wusbcore/host/hwa-hc.c return usb_get_current_frame_number(wa->usb_dev); wa 203 drivers/staging/wusbcore/host/hwa-hc.c return wa_urb_enqueue(&hwahc->wa, urb->ep, urb, gfp); wa 212 drivers/staging/wusbcore/host/hwa-hc.c return wa_urb_dequeue(&hwahc->wa, urb, status); wa 226 drivers/staging/wusbcore/host/hwa-hc.c rpipe_ep_disable(&hwahc->wa, ep); wa 233 drivers/staging/wusbcore/host/hwa-hc.c struct device *dev = &hwahc->wa.usb_iface->dev; wa 235 drivers/staging/wusbcore/host/hwa-hc.c result = __wa_set_feature(&hwahc->wa, WA_ENABLE); wa 240 drivers/staging/wusbcore/host/hwa-hc.c result = __wa_wait_status(&hwahc->wa, WA_ENABLE, WA_ENABLE); wa 245 drivers/staging/wusbcore/host/hwa-hc.c result = wa_nep_arm(&hwahc->wa, GFP_KERNEL); wa 254 drivers/staging/wusbcore/host/hwa-hc.c if (hwahc->wa.quirks & wa 257 drivers/staging/wusbcore/host/hwa-hc.c hwahc->wa.usb_iface->cur_altsetting; wa 259 drivers/staging/wusbcore/host/hwa-hc.c result = usb_control_msg(hwahc->wa.usb_dev, wa 260 drivers/staging/wusbcore/host/hwa-hc.c usb_sndctrlpipe(hwahc->wa.usb_dev, 0), wa 274 drivers/staging/wusbcore/host/hwa-hc.c result = wa_dti_start(&hwahc->wa); wa 287 drivers/staging/wusbcore/host/hwa-hc.c wa_nep_disarm(&hwahc->wa); wa 289 drivers/staging/wusbcore/host/hwa-hc.c __wa_clear_feature(&hwahc->wa, WA_ENABLE); wa 296 drivers/staging/wusbcore/host/hwa-hc.c struct wahc *wa = &hwahc->wa; wa 297 drivers/staging/wusbcore/host/hwa-hc.c u8 iface_no = wa->usb_iface->cur_altsetting->desc.bInterfaceNumber; wa 300 drivers/staging/wusbcore/host/hwa-hc.c ret = usb_control_msg(wa->usb_dev, usb_sndctrlpipe(wa->usb_dev, 0), wa 309 drivers/staging/wusbcore/host/hwa-hc.c wa_nep_disarm(&hwahc->wa); wa 310 drivers/staging/wusbcore/host/hwa-hc.c __wa_stop(&hwahc->wa); wa 324 drivers/staging/wusbcore/host/hwa-hc.c struct wahc *wa = &hwahc->wa; wa 325 drivers/staging/wusbcore/host/hwa-hc.c struct device *dev = &wa->usb_iface->dev; wa 329 drivers/staging/wusbcore/host/hwa-hc.c result = usb_control_msg(wa->usb_dev, usb_sndctrlpipe(wa->usb_dev, 0), wa 333 drivers/staging/wusbcore/host/hwa-hc.c wa->usb_iface->cur_altsetting->desc.bInterfaceNumber, wa 341 drivers/staging/wusbcore/host/hwa-hc.c result = usb_control_msg(wa->usb_dev, usb_sndctrlpipe(wa->usb_dev, 0), wa 344 drivers/staging/wusbcore/host/hwa-hc.c 0, wa->usb_iface->cur_altsetting->desc.bInterfaceNumber, wa 371 drivers/staging/wusbcore/host/hwa-hc.c struct wahc *wa = &hwahc->wa; wa 372 drivers/staging/wusbcore/host/hwa-hc.c u8 iface_no = wa->usb_iface->cur_altsetting->desc.bInterfaceNumber; wa 374 drivers/staging/wusbcore/host/hwa-hc.c return usb_control_msg(wa->usb_dev, usb_sndctrlpipe(wa->usb_dev, 0), wa 390 drivers/staging/wusbcore/host/hwa-hc.c struct wahc *wa = &hwahc->wa; wa 391 drivers/staging/wusbcore/host/hwa-hc.c u8 iface_no = wa->usb_iface->cur_altsetting->desc.bInterfaceNumber; wa 392 drivers/staging/wusbcore/host/hwa-hc.c return usb_control_msg(wa->usb_dev, usb_sndctrlpipe(wa->usb_dev, 0), wa 409 drivers/staging/wusbcore/host/hwa-hc.c struct wahc *wa = &hwahc->wa; wa 410 drivers/staging/wusbcore/host/hwa-hc.c u8 iface_no = wa->usb_iface->cur_altsetting->desc.bInterfaceNumber; wa 434 drivers/staging/wusbcore/host/hwa-hc.c ret = usb_control_msg(wa->usb_dev, usb_sndctrlpipe(wa->usb_dev, 0), wa 457 drivers/staging/wusbcore/host/hwa-hc.c struct wahc *wa = &hwahc->wa; wa 458 drivers/staging/wusbcore/host/hwa-hc.c u8 iface_no = wa->usb_iface->cur_altsetting->desc.bInterfaceNumber; wa 474 drivers/staging/wusbcore/host/hwa-hc.c result = usb_control_msg(wa->usb_dev, usb_sndctrlpipe(wa->usb_dev, 0), wa 497 drivers/staging/wusbcore/host/hwa-hc.c struct wahc *wa = &hwahc->wa; wa 498 drivers/staging/wusbcore/host/hwa-hc.c u8 iface_no = wa->usb_iface->cur_altsetting->desc.bInterfaceNumber; wa 517 drivers/staging/wusbcore/host/hwa-hc.c result = usb_control_msg(wa->usb_dev, usb_sndctrlpipe(wa->usb_dev, 0), wa 552 drivers/staging/wusbcore/host/hwa-hc.c static int wa_fill_descr(struct wahc *wa) wa 555 drivers/staging/wusbcore/host/hwa-hc.c struct device *dev = &wa->usb_iface->dev; wa 557 drivers/staging/wusbcore/host/hwa-hc.c struct usb_device *usb_dev = wa->usb_dev; wa 589 drivers/staging/wusbcore/host/hwa-hc.c if (hdr->bLength < sizeof(*wa->wa_descr)) { wa 593 drivers/staging/wusbcore/host/hwa-hc.c wa->wa_descr = wa_descr = (struct usb_wa_descriptor *) hdr; wa 626 drivers/staging/wusbcore/host/hwa-hc.c struct usb_device *usb_dev = hwahc->wa.usb_dev; wa 709 drivers/staging/wusbcore/host/hwa-hc.c struct wahc *wa = &hwahc->wa; wa 712 drivers/staging/wusbcore/host/hwa-hc.c wa->usb_dev = usb_get_dev(usb_dev); /* bind the USB device */ wa 713 drivers/staging/wusbcore/host/hwa-hc.c wa->usb_iface = usb_get_intf(iface); wa 718 drivers/staging/wusbcore/host/hwa-hc.c result = wa_fill_descr(wa); /* Get the device descriptor */ wa 721 drivers/staging/wusbcore/host/hwa-hc.c if (wa->wa_descr->bNumPorts > USB_MAXCHILDREN) { wa 723 drivers/staging/wusbcore/host/hwa-hc.c "adapter (%u ports)\n", wa->wa_descr->bNumPorts); wa 726 drivers/staging/wusbcore/host/hwa-hc.c wusbhc->ports_max = wa->wa_descr->bNumPorts; wa 728 drivers/staging/wusbcore/host/hwa-hc.c wusbhc->mmcies_max = wa->wa_descr->bNumMMCIEs; wa 743 drivers/staging/wusbcore/host/hwa-hc.c wa->wusb = wusbhc; /* FIXME: ugly, need to fix */ wa 749 drivers/staging/wusbcore/host/hwa-hc.c result = wa_create(&hwahc->wa, iface, quirks); wa 770 drivers/staging/wusbcore/host/hwa-hc.c __wa_destroy(&hwahc->wa); wa 775 drivers/staging/wusbcore/host/hwa-hc.c usb_put_intf(hwahc->wa.usb_iface); wa 776 drivers/staging/wusbcore/host/hwa-hc.c usb_put_dev(hwahc->wa.usb_dev); wa 782 drivers/staging/wusbcore/host/hwa-hc.c wa_init(&hwahc->wa); wa 22 drivers/staging/wusbcore/wa-hc.c int wa_create(struct wahc *wa, struct usb_interface *iface, wa 31 drivers/staging/wusbcore/wa-hc.c result = wa_rpipes_create(wa); wa 34 drivers/staging/wusbcore/wa-hc.c wa->quirks = quirks; wa 36 drivers/staging/wusbcore/wa-hc.c wa->dti_epd = &iface->cur_altsetting->endpoint[1].desc; wa 37 drivers/staging/wusbcore/wa-hc.c wa->dto_epd = &iface->cur_altsetting->endpoint[2].desc; wa 38 drivers/staging/wusbcore/wa-hc.c wa->dti_buf_size = usb_endpoint_maxp(wa->dti_epd); wa 39 drivers/staging/wusbcore/wa-hc.c wa->dti_buf = kmalloc(wa->dti_buf_size, GFP_KERNEL); wa 40 drivers/staging/wusbcore/wa-hc.c if (wa->dti_buf == NULL) { wa 44 drivers/staging/wusbcore/wa-hc.c result = wa_nep_create(wa, iface); wa 53 drivers/staging/wusbcore/wa-hc.c kfree(wa->dti_buf); wa 55 drivers/staging/wusbcore/wa-hc.c wa_rpipes_destroy(wa); wa 62 drivers/staging/wusbcore/wa-hc.c void __wa_destroy(struct wahc *wa) wa 64 drivers/staging/wusbcore/wa-hc.c if (wa->dti_urb) { wa 65 drivers/staging/wusbcore/wa-hc.c usb_kill_urb(wa->dti_urb); wa 66 drivers/staging/wusbcore/wa-hc.c usb_put_urb(wa->dti_urb); wa 68 drivers/staging/wusbcore/wa-hc.c kfree(wa->dti_buf); wa 69 drivers/staging/wusbcore/wa-hc.c wa_nep_destroy(wa); wa 70 drivers/staging/wusbcore/wa-hc.c wa_rpipes_destroy(wa); wa 80 drivers/staging/wusbcore/wa-hc.c void wa_reset_all(struct wahc *wa) wa 83 drivers/staging/wusbcore/wa-hc.c wusbhc_reset_all(wa->wusb); wa 103 drivers/staging/wusbcore/wa-hc.h struct wahc *wa; wa 235 drivers/staging/wusbcore/wa-hc.h extern int wa_create(struct wahc *wa, struct usb_interface *iface, wa 237 drivers/staging/wusbcore/wa-hc.h extern void __wa_destroy(struct wahc *wa); wa 238 drivers/staging/wusbcore/wa-hc.h extern int wa_dti_start(struct wahc *wa); wa 239 drivers/staging/wusbcore/wa-hc.h void wa_reset_all(struct wahc *wa); wa 256 drivers/staging/wusbcore/wa-hc.h static inline int wa_nep_arm(struct wahc *wa, gfp_t gfp_mask) wa 258 drivers/staging/wusbcore/wa-hc.h struct urb *urb = wa->nep_urb; wa 259 drivers/staging/wusbcore/wa-hc.h urb->transfer_buffer = wa->nep_buffer; wa 260 drivers/staging/wusbcore/wa-hc.h urb->transfer_buffer_length = wa->nep_buffer_size; wa 264 drivers/staging/wusbcore/wa-hc.h static inline void wa_nep_disarm(struct wahc *wa) wa 266 drivers/staging/wusbcore/wa-hc.h usb_kill_urb(wa->nep_urb); wa 271 drivers/staging/wusbcore/wa-hc.h static inline void wa_rpipe_init(struct wahc *wa) wa 273 drivers/staging/wusbcore/wa-hc.h INIT_LIST_HEAD(&wa->rpipe_delayed_list); wa 274 drivers/staging/wusbcore/wa-hc.h spin_lock_init(&wa->rpipe_lock); wa 275 drivers/staging/wusbcore/wa-hc.h mutex_init(&wa->rpipe_mutex); wa 278 drivers/staging/wusbcore/wa-hc.h static inline void wa_init(struct wahc *wa) wa 282 drivers/staging/wusbcore/wa-hc.h edc_init(&wa->nep_edc); wa 283 drivers/staging/wusbcore/wa-hc.h atomic_set(&wa->notifs_queued, 0); wa 284 drivers/staging/wusbcore/wa-hc.h wa->dti_state = WA_DTI_TRANSFER_RESULT_PENDING; wa 285 drivers/staging/wusbcore/wa-hc.h wa_rpipe_init(wa); wa 286 drivers/staging/wusbcore/wa-hc.h edc_init(&wa->dti_edc); wa 287 drivers/staging/wusbcore/wa-hc.h INIT_LIST_HEAD(&wa->xfer_list); wa 288 drivers/staging/wusbcore/wa-hc.h INIT_LIST_HEAD(&wa->xfer_delayed_list); wa 289 drivers/staging/wusbcore/wa-hc.h INIT_LIST_HEAD(&wa->xfer_errored_list); wa 290 drivers/staging/wusbcore/wa-hc.h spin_lock_init(&wa->xfer_list_lock); wa 291 drivers/staging/wusbcore/wa-hc.h INIT_WORK(&wa->xfer_enqueue_work, wa_urb_enqueue_run); wa 292 drivers/staging/wusbcore/wa-hc.h INIT_WORK(&wa->xfer_error_work, wa_process_errored_transfers_run); wa 293 drivers/staging/wusbcore/wa-hc.h wa->dto_in_use = 0; wa 294 drivers/staging/wusbcore/wa-hc.h atomic_set(&wa->xfer_id_count, 1); wa 297 drivers/staging/wusbcore/wa-hc.h usb_init_urb(&(wa->buf_in_urbs[index])); wa 298 drivers/staging/wusbcore/wa-hc.h wa->active_buf_in_urbs = 0; wa 358 drivers/staging/wusbcore/wa-hc.h static inline struct wahc *wa_get(struct wahc *wa) wa 360 drivers/staging/wusbcore/wa-hc.h usb_get_intf(wa->usb_iface); wa 361 drivers/staging/wusbcore/wa-hc.h return wa; wa 364 drivers/staging/wusbcore/wa-hc.h static inline void wa_put(struct wahc *wa) wa 366 drivers/staging/wusbcore/wa-hc.h usb_put_intf(wa->usb_iface); wa 370 drivers/staging/wusbcore/wa-hc.h static inline int __wa_feature(struct wahc *wa, unsigned op, u16 feature) wa 372 drivers/staging/wusbcore/wa-hc.h return usb_control_msg(wa->usb_dev, usb_sndctrlpipe(wa->usb_dev, 0), wa 376 drivers/staging/wusbcore/wa-hc.h wa->usb_iface->cur_altsetting->desc.bInterfaceNumber, wa 381 drivers/staging/wusbcore/wa-hc.h static inline int __wa_set_feature(struct wahc *wa, u16 feature) wa 383 drivers/staging/wusbcore/wa-hc.h return __wa_feature(wa, 1, feature); wa 387 drivers/staging/wusbcore/wa-hc.h static inline int __wa_clear_feature(struct wahc *wa, u16 feature) wa 389 drivers/staging/wusbcore/wa-hc.h return __wa_feature(wa, 0, feature); wa 403 drivers/staging/wusbcore/wa-hc.h s32 __wa_get_status(struct wahc *wa) wa 407 drivers/staging/wusbcore/wa-hc.h wa->usb_dev, usb_rcvctrlpipe(wa->usb_dev, 0), wa 410 drivers/staging/wusbcore/wa-hc.h 0, wa->usb_iface->cur_altsetting->desc.bInterfaceNumber, wa 411 drivers/staging/wusbcore/wa-hc.h &wa->status, sizeof(wa->status), USB_CTRL_GET_TIMEOUT); wa 413 drivers/staging/wusbcore/wa-hc.h result = wa->status; wa 430 drivers/staging/wusbcore/wa-hc.h static inline s32 __wa_wait_status(struct wahc *wa, u32 mask, u32 value) wa 436 drivers/staging/wusbcore/wa-hc.h result = __wa_get_status(wa); wa 449 drivers/staging/wusbcore/wa-hc.h static inline int __wa_stop(struct wahc *wa) wa 452 drivers/staging/wusbcore/wa-hc.h struct device *dev = &wa->usb_iface->dev; wa 454 drivers/staging/wusbcore/wa-hc.h result = __wa_clear_feature(wa, WA_ENABLE); wa 459 drivers/staging/wusbcore/wa-hc.h result = __wa_wait_status(wa, WA_ENABLE, 0); wa 48 drivers/staging/wusbcore/wa-nep.c struct wahc *wa; wa 84 drivers/staging/wusbcore/wa-nep.c struct wahc *wa = nw->wa; wa 88 drivers/staging/wusbcore/wa-nep.c struct device *dev = &wa->usb_iface->dev; wa 95 drivers/staging/wusbcore/wa-nep.c atomic_dec(&wa->notifs_queued); /* Throttling ctl */ wa 115 drivers/staging/wusbcore/wa-nep.c wusbhc_handle_dn(wa->wusb, hwa_dn->bSourceDeviceAddr, wa 121 drivers/staging/wusbcore/wa-nep.c wa_handle_notif_xfer(wa, notif_hdr); wa 138 drivers/staging/wusbcore/wa-nep.c wa_put(wa); wa 166 drivers/staging/wusbcore/wa-nep.c static int wa_nep_queue(struct wahc *wa, size_t size) wa 169 drivers/staging/wusbcore/wa-nep.c struct device *dev = &wa->usb_iface->dev; wa 173 drivers/staging/wusbcore/wa-nep.c BUG_ON(size > wa->nep_buffer_size); wa 176 drivers/staging/wusbcore/wa-nep.c if (atomic_read(&wa->notifs_queued) > 200) { wa 190 drivers/staging/wusbcore/wa-nep.c nw->wa = wa_get(wa); wa 192 drivers/staging/wusbcore/wa-nep.c memcpy(nw->data, wa->nep_buffer, size); wa 193 drivers/staging/wusbcore/wa-nep.c atomic_inc(&wa->notifs_queued); /* Throttling ctl */ wa 209 drivers/staging/wusbcore/wa-nep.c struct wahc *wa = urb->context; wa 210 drivers/staging/wusbcore/wa-nep.c struct device *dev = &wa->usb_iface->dev; wa 214 drivers/staging/wusbcore/wa-nep.c result = wa_nep_queue(wa, urb->actual_length); wa 225 drivers/staging/wusbcore/wa-nep.c if (edc_inc(&wa->nep_edc, EDC_MAX_ERRORS, wa 229 drivers/staging/wusbcore/wa-nep.c wa_reset_all(wa); wa 234 drivers/staging/wusbcore/wa-nep.c result = wa_nep_arm(wa, GFP_ATOMIC); wa 237 drivers/staging/wusbcore/wa-nep.c wa_reset_all(wa); wa 249 drivers/staging/wusbcore/wa-nep.c int wa_nep_create(struct wahc *wa, struct usb_interface *iface) wa 256 drivers/staging/wusbcore/wa-nep.c edc_init(&wa->nep_edc); wa 258 drivers/staging/wusbcore/wa-nep.c wa->nep_buffer_size = 1024; wa 259 drivers/staging/wusbcore/wa-nep.c wa->nep_buffer = kmalloc(wa->nep_buffer_size, GFP_KERNEL); wa 260 drivers/staging/wusbcore/wa-nep.c if (!wa->nep_buffer) wa 262 drivers/staging/wusbcore/wa-nep.c wa->nep_urb = usb_alloc_urb(0, GFP_KERNEL); wa 263 drivers/staging/wusbcore/wa-nep.c if (wa->nep_urb == NULL) wa 265 drivers/staging/wusbcore/wa-nep.c usb_fill_int_urb(wa->nep_urb, usb_dev, wa 267 drivers/staging/wusbcore/wa-nep.c wa->nep_buffer, wa->nep_buffer_size, wa 268 drivers/staging/wusbcore/wa-nep.c wa_nep_cb, wa, epd->bInterval); wa 269 drivers/staging/wusbcore/wa-nep.c result = wa_nep_arm(wa, GFP_KERNEL); wa 277 drivers/staging/wusbcore/wa-nep.c usb_free_urb(wa->nep_urb); wa 279 drivers/staging/wusbcore/wa-nep.c kfree(wa->nep_buffer); wa 284 drivers/staging/wusbcore/wa-nep.c void wa_nep_destroy(struct wahc *wa) wa 286 drivers/staging/wusbcore/wa-nep.c wa_nep_disarm(wa); wa 287 drivers/staging/wusbcore/wa-nep.c usb_free_urb(wa->nep_urb); wa 288 drivers/staging/wusbcore/wa-nep.c kfree(wa->nep_buffer); wa 54 drivers/staging/wusbcore/wa-rpipe.c static int __rpipe_get_descr(struct wahc *wa, wa 58 drivers/staging/wusbcore/wa-rpipe.c struct device *dev = &wa->usb_iface->dev; wa 64 drivers/staging/wusbcore/wa-rpipe.c wa->usb_dev, usb_rcvctrlpipe(wa->usb_dev, 0), wa 92 drivers/staging/wusbcore/wa-rpipe.c static int __rpipe_set_descr(struct wahc *wa, wa 96 drivers/staging/wusbcore/wa-rpipe.c struct device *dev = &wa->usb_iface->dev; wa 102 drivers/staging/wusbcore/wa-rpipe.c wa->usb_dev, usb_sndctrlpipe(wa->usb_dev, 0), wa 134 drivers/staging/wusbcore/wa-rpipe.c static unsigned rpipe_get_idx(struct wahc *wa, unsigned rpipe_idx) wa 138 drivers/staging/wusbcore/wa-rpipe.c spin_lock_irqsave(&wa->rpipe_lock, flags); wa 139 drivers/staging/wusbcore/wa-rpipe.c rpipe_idx = find_next_zero_bit(wa->rpipe_bm, wa->rpipes, rpipe_idx); wa 140 drivers/staging/wusbcore/wa-rpipe.c if (rpipe_idx < wa->rpipes) wa 141 drivers/staging/wusbcore/wa-rpipe.c set_bit(rpipe_idx, wa->rpipe_bm); wa 142 drivers/staging/wusbcore/wa-rpipe.c spin_unlock_irqrestore(&wa->rpipe_lock, flags); wa 147 drivers/staging/wusbcore/wa-rpipe.c static void rpipe_put_idx(struct wahc *wa, unsigned rpipe_idx) wa 151 drivers/staging/wusbcore/wa-rpipe.c spin_lock_irqsave(&wa->rpipe_lock, flags); wa 152 drivers/staging/wusbcore/wa-rpipe.c clear_bit(rpipe_idx, wa->rpipe_bm); wa 153 drivers/staging/wusbcore/wa-rpipe.c spin_unlock_irqrestore(&wa->rpipe_lock, flags); wa 163 drivers/staging/wusbcore/wa-rpipe.c rpipe_put_idx(rpipe->wa, index); wa 164 drivers/staging/wusbcore/wa-rpipe.c wa_put(rpipe->wa); wa 179 drivers/staging/wusbcore/wa-rpipe.c static int rpipe_get_idle(struct wa_rpipe **prpipe, struct wahc *wa, u8 crs, wa 185 drivers/staging/wusbcore/wa-rpipe.c struct device *dev = &wa->usb_iface->dev; wa 193 drivers/staging/wusbcore/wa-rpipe.c for (rpipe_idx = 0; rpipe_idx < wa->rpipes; rpipe_idx++) { wa 194 drivers/staging/wusbcore/wa-rpipe.c rpipe_idx = rpipe_get_idx(wa, rpipe_idx); wa 195 drivers/staging/wusbcore/wa-rpipe.c if (rpipe_idx >= wa->rpipes) /* no more pipes :( */ wa 197 drivers/staging/wusbcore/wa-rpipe.c result = __rpipe_get_descr(wa, &rpipe->descr, rpipe_idx); wa 203 drivers/staging/wusbcore/wa-rpipe.c rpipe_put_idx(wa, rpipe_idx); wa 210 drivers/staging/wusbcore/wa-rpipe.c set_bit(rpipe_idx, wa->rpipe_bm); wa 211 drivers/staging/wusbcore/wa-rpipe.c rpipe->wa = wa_get(wa); wa 216 drivers/staging/wusbcore/wa-rpipe.c static int __rpipe_reset(struct wahc *wa, unsigned index) wa 219 drivers/staging/wusbcore/wa-rpipe.c struct device *dev = &wa->usb_iface->dev; wa 222 drivers/staging/wusbcore/wa-rpipe.c wa->usb_dev, usb_sndctrlpipe(wa->usb_dev, 0), wa 299 drivers/staging/wusbcore/wa-rpipe.c static int rpipe_aim(struct wa_rpipe *rpipe, struct wahc *wa, wa 303 drivers/staging/wusbcore/wa-rpipe.c struct device *dev = &wa->usb_iface->dev; wa 316 drivers/staging/wusbcore/wa-rpipe.c __rpipe_reset(wa, le16_to_cpu(rpipe->descr.wRPipeIndex)); wa 362 drivers/staging/wusbcore/wa-rpipe.c rpipe->descr.bmRetryOptions = (wa->wusb->retry_count & 0xF); wa 365 drivers/staging/wusbcore/wa-rpipe.c result = __rpipe_set_descr(wa, &rpipe->descr, wa 382 drivers/staging/wusbcore/wa-rpipe.c static int rpipe_check_aim(const struct wa_rpipe *rpipe, const struct wahc *wa, wa 387 drivers/staging/wusbcore/wa-rpipe.c struct device *dev = &wa->usb_iface->dev; wa 423 drivers/staging/wusbcore/wa-rpipe.c int rpipe_get_by_ep(struct wahc *wa, struct usb_host_endpoint *ep, wa 427 drivers/staging/wusbcore/wa-rpipe.c struct device *dev = &wa->usb_iface->dev; wa 431 drivers/staging/wusbcore/wa-rpipe.c mutex_lock(&wa->rpipe_mutex); wa 435 drivers/staging/wusbcore/wa-rpipe.c result = rpipe_check_aim(rpipe, wa, ep, urb, gfp); wa 447 drivers/staging/wusbcore/wa-rpipe.c result = rpipe_get_idle(&rpipe, wa, 1 << eptype, gfp); wa 450 drivers/staging/wusbcore/wa-rpipe.c result = rpipe_aim(rpipe, wa, ep, urb, gfp); wa 463 drivers/staging/wusbcore/wa-rpipe.c mutex_unlock(&wa->rpipe_mutex); wa 470 drivers/staging/wusbcore/wa-rpipe.c int wa_rpipes_create(struct wahc *wa) wa 472 drivers/staging/wusbcore/wa-rpipe.c wa->rpipes = le16_to_cpu(wa->wa_descr->wNumRPipes); wa 473 drivers/staging/wusbcore/wa-rpipe.c wa->rpipe_bm = bitmap_zalloc(wa->rpipes, GFP_KERNEL); wa 474 drivers/staging/wusbcore/wa-rpipe.c if (wa->rpipe_bm == NULL) wa 479 drivers/staging/wusbcore/wa-rpipe.c void wa_rpipes_destroy(struct wahc *wa) wa 481 drivers/staging/wusbcore/wa-rpipe.c struct device *dev = &wa->usb_iface->dev; wa 483 drivers/staging/wusbcore/wa-rpipe.c if (!bitmap_empty(wa->rpipe_bm, wa->rpipes)) { wa 486 drivers/staging/wusbcore/wa-rpipe.c wa->rpipes, wa->rpipe_bm); wa 488 drivers/staging/wusbcore/wa-rpipe.c bitmap_free(wa->rpipe_bm); wa 501 drivers/staging/wusbcore/wa-rpipe.c void rpipe_ep_disable(struct wahc *wa, struct usb_host_endpoint *ep) wa 505 drivers/staging/wusbcore/wa-rpipe.c mutex_lock(&wa->rpipe_mutex); wa 511 drivers/staging/wusbcore/wa-rpipe.c wa->usb_dev, usb_sndctrlpipe(wa->usb_dev, 0), wa 517 drivers/staging/wusbcore/wa-rpipe.c mutex_unlock(&wa->rpipe_mutex); wa 522 drivers/staging/wusbcore/wa-rpipe.c void rpipe_clear_feature_stalled(struct wahc *wa, struct usb_host_endpoint *ep) wa 526 drivers/staging/wusbcore/wa-rpipe.c mutex_lock(&wa->rpipe_mutex); wa 532 drivers/staging/wusbcore/wa-rpipe.c wa->usb_dev, usb_sndctrlpipe(wa->usb_dev, 0), wa 537 drivers/staging/wusbcore/wa-rpipe.c mutex_unlock(&wa->rpipe_mutex); wa 139 drivers/staging/wusbcore/wa-xfer.c struct wahc *wa; /* Wire adapter we are plugged to */ wa 207 drivers/staging/wusbcore/wa-xfer.c static inline int __wa_dto_try_get(struct wahc *wa) wa 209 drivers/staging/wusbcore/wa-xfer.c return (test_and_set_bit(0, &wa->dto_in_use) == 0); wa 213 drivers/staging/wusbcore/wa-xfer.c static inline void __wa_dto_put(struct wahc *wa) wa 215 drivers/staging/wusbcore/wa-xfer.c clear_bit_unlock(0, &wa->dto_in_use); wa 219 drivers/staging/wusbcore/wa-xfer.c static void wa_check_for_delayed_rpipes(struct wahc *wa) wa 225 drivers/staging/wusbcore/wa-xfer.c spin_lock_irqsave(&wa->rpipe_lock, flags); wa 226 drivers/staging/wusbcore/wa-xfer.c while (!list_empty(&wa->rpipe_delayed_list) && !dto_waiting) { wa 227 drivers/staging/wusbcore/wa-xfer.c rpipe = list_first_entry(&wa->rpipe_delayed_list, wa 238 drivers/staging/wusbcore/wa-xfer.c spin_unlock_irqrestore(&wa->rpipe_lock, flags); wa 242 drivers/staging/wusbcore/wa-xfer.c static void wa_add_delayed_rpipe(struct wahc *wa, struct wa_rpipe *rpipe) wa 246 drivers/staging/wusbcore/wa-xfer.c spin_lock_irqsave(&wa->rpipe_lock, flags); wa 251 drivers/staging/wusbcore/wa-xfer.c list_add_tail(&rpipe->list_node, &wa->rpipe_delayed_list); wa 253 drivers/staging/wusbcore/wa-xfer.c spin_unlock_irqrestore(&wa->rpipe_lock, flags); wa 270 drivers/staging/wusbcore/wa-xfer.c spin_lock_irqsave(&xfer->wa->xfer_list_lock, flags); wa 272 drivers/staging/wusbcore/wa-xfer.c usb_hcd_unlink_urb_from_ep(&(xfer->wa->wusb->usb_hcd), xfer->urb); wa 273 drivers/staging/wusbcore/wa-xfer.c spin_unlock_irqrestore(&xfer->wa->xfer_list_lock, flags); wa 275 drivers/staging/wusbcore/wa-xfer.c wusbhc_giveback_urb(xfer->wa->wusb, xfer->urb, xfer->result); wa 276 drivers/staging/wusbcore/wa-xfer.c wa_put(xfer->wa); wa 303 drivers/staging/wusbcore/wa-xfer.c xfer->id = atomic_add_return(1, &xfer->wa->xfer_id_count); wa 325 drivers/staging/wusbcore/wa-xfer.c struct device *dev = &xfer->wa->usb_iface->dev; wa 406 drivers/staging/wusbcore/wa-xfer.c static struct wa_xfer *wa_xfer_get_by_id(struct wahc *wa, u32 id) wa 410 drivers/staging/wusbcore/wa-xfer.c spin_lock_irqsave(&wa->xfer_list_lock, flags); wa 411 drivers/staging/wusbcore/wa-xfer.c list_for_each_entry(xfer_itr, &wa->xfer_list, list_node) { wa 419 drivers/staging/wusbcore/wa-xfer.c spin_unlock_irqrestore(&wa->xfer_list_lock, flags); wa 425 drivers/staging/wusbcore/wa-xfer.c struct wahc *wa; wa 432 drivers/staging/wusbcore/wa-xfer.c struct wahc *wa = b->wa; wa 441 drivers/staging/wusbcore/wa-xfer.c struct device *dev = &wa->usb_iface->dev; wa 443 drivers/staging/wusbcore/wa-xfer.c xfer = wa_xfer_get_by_id(wa, le32_to_cpu(b->cmd.dwTransferID)); wa 480 drivers/staging/wusbcore/wa-xfer.c wa_put(wa); /* taken in __wa_xfer_abort */ wa 497 drivers/staging/wusbcore/wa-xfer.c struct device *dev = &xfer->wa->usb_iface->dev; wa 508 drivers/staging/wusbcore/wa-xfer.c b->wa = wa_get(xfer->wa); wa 511 drivers/staging/wusbcore/wa-xfer.c usb_fill_bulk_urb(&b->urb, xfer->wa->usb_dev, wa 512 drivers/staging/wusbcore/wa-xfer.c usb_sndbulkpipe(xfer->wa->usb_dev, wa 513 drivers/staging/wusbcore/wa-xfer.c xfer->wa->dto_epd->bEndpointAddress), wa 522 drivers/staging/wusbcore/wa-xfer.c wa_put(xfer->wa); wa 554 drivers/staging/wusbcore/wa-xfer.c if ((xfer->wa->quirks & WUSB_QUIRK_ALEREON_HWA_CONCAT_ISOC) wa 582 drivers/staging/wusbcore/wa-xfer.c struct device *dev = &xfer->wa->usb_iface->dev; wa 611 drivers/staging/wusbcore/wa-xfer.c * 1 << (xfer->wa->wa_descr->bRPipeBlockSize - 1); wa 736 drivers/staging/wusbcore/wa-xfer.c struct wahc *wa; wa 750 drivers/staging/wusbcore/wa-xfer.c wa = xfer->wa; wa 751 drivers/staging/wusbcore/wa-xfer.c dev = &wa->usb_iface->dev; wa 754 drivers/staging/wusbcore/wa-xfer.c if (wa->quirks & WUSB_QUIRK_ALEREON_HWA_CONCAT_ISOC) wa 806 drivers/staging/wusbcore/wa-xfer.c __wa_dto_put(wa); wa 807 drivers/staging/wusbcore/wa-xfer.c wa_check_for_delayed_rpipes(wa); wa 813 drivers/staging/wusbcore/wa-xfer.c __wa_dto_put(wa); wa 814 drivers/staging/wusbcore/wa-xfer.c wa_check_for_delayed_rpipes(wa); wa 833 drivers/staging/wusbcore/wa-xfer.c if (edc_inc(&wa->nep_edc, EDC_MAX_ERRORS, wa 836 drivers/staging/wusbcore/wa-xfer.c wa_reset_all(wa); wa 846 drivers/staging/wusbcore/wa-xfer.c __wa_dto_put(wa); wa 847 drivers/staging/wusbcore/wa-xfer.c wa_check_for_delayed_rpipes(wa); wa 873 drivers/staging/wusbcore/wa-xfer.c struct wahc *wa; wa 883 drivers/staging/wusbcore/wa-xfer.c wa = xfer->wa; wa 884 drivers/staging/wusbcore/wa-xfer.c dev = &wa->usb_iface->dev; wa 896 drivers/staging/wusbcore/wa-xfer.c wa = xfer->wa; wa 897 drivers/staging/wusbcore/wa-xfer.c dev = &wa->usb_iface->dev; wa 901 drivers/staging/wusbcore/wa-xfer.c if (edc_inc(&wa->nep_edc, EDC_MAX_ERRORS, wa 904 drivers/staging/wusbcore/wa-xfer.c wa_reset_all(wa); wa 946 drivers/staging/wusbcore/wa-xfer.c struct wahc *wa; wa 956 drivers/staging/wusbcore/wa-xfer.c wa = xfer->wa; wa 957 drivers/staging/wusbcore/wa-xfer.c dev = &wa->usb_iface->dev; wa 971 drivers/staging/wusbcore/wa-xfer.c wa = xfer->wa; wa 972 drivers/staging/wusbcore/wa-xfer.c dev = &wa->usb_iface->dev; wa 978 drivers/staging/wusbcore/wa-xfer.c if (edc_inc(&wa->nep_edc, EDC_MAX_ERRORS, wa 982 drivers/staging/wusbcore/wa-xfer.c wa_reset_all(wa); wa 1095 drivers/staging/wusbcore/wa-xfer.c if (xfer->wa->quirks & WUSB_QUIRK_ALEREON_HWA_CONCAT_ISOC) wa 1164 drivers/staging/wusbcore/wa-xfer.c struct usb_device *usb_dev = xfer->wa->usb_dev; wa 1165 drivers/staging/wusbcore/wa-xfer.c const struct usb_endpoint_descriptor *dto_epd = xfer->wa->dto_epd; wa 1299 drivers/staging/wusbcore/wa-xfer.c struct device *dev = &xfer->wa->usb_iface->dev; wa 1412 drivers/staging/wusbcore/wa-xfer.c struct wahc *wa = xfer->wa; wa 1426 drivers/staging/wusbcore/wa-xfer.c if (((wa->quirks & WUSB_QUIRK_ALEREON_HWA_CONCAT_ISOC) == 0) wa 1454 drivers/staging/wusbcore/wa-xfer.c struct device *dev = &rpipe->wa->usb_iface->dev; wa 1464 drivers/staging/wusbcore/wa-xfer.c && (dto_acquired = __wa_dto_try_get(rpipe->wa))) { wa 1478 drivers/staging/wusbcore/wa-xfer.c __wa_dto_put(rpipe->wa); wa 1529 drivers/staging/wusbcore/wa-xfer.c wa_add_delayed_rpipe(rpipe->wa, rpipe); wa 1531 drivers/staging/wusbcore/wa-xfer.c wa_check_for_delayed_rpipes(rpipe->wa); wa 1544 drivers/staging/wusbcore/wa-xfer.c struct wahc *wa = xfer->wa; wa 1545 drivers/staging/wusbcore/wa-xfer.c struct device *dev = &wa->usb_iface->dev; wa 1554 drivers/staging/wusbcore/wa-xfer.c spin_lock_irqsave(&wa->xfer_list_lock, flags); wa 1555 drivers/staging/wusbcore/wa-xfer.c list_add_tail(&xfer->list_node, &wa->xfer_list); wa 1556 drivers/staging/wusbcore/wa-xfer.c spin_unlock_irqrestore(&wa->xfer_list_lock, flags); wa 1572 drivers/staging/wusbcore/wa-xfer.c dto_acquired = __wa_dto_try_get(rpipe->wa); wa 1581 drivers/staging/wusbcore/wa-xfer.c __wa_dto_put(rpipe->wa); wa 1610 drivers/staging/wusbcore/wa-xfer.c wa_add_delayed_rpipe(rpipe->wa, rpipe); wa 1612 drivers/staging/wusbcore/wa-xfer.c wa_check_for_delayed_rpipes(rpipe->wa); wa 1644 drivers/staging/wusbcore/wa-xfer.c struct wahc *wa = xfer->wa; wa 1645 drivers/staging/wusbcore/wa-xfer.c struct wusbhc *wusbhc = wa->wusb; wa 1649 drivers/staging/wusbcore/wa-xfer.c result = rpipe_get_by_ep(wa, xfer->ep, urb, xfer->gfp); wa 1739 drivers/staging/wusbcore/wa-xfer.c struct wahc *wa = container_of(ws, struct wahc, xfer_enqueue_work); wa 1745 drivers/staging/wusbcore/wa-xfer.c spin_lock_irq(&wa->xfer_list_lock); wa 1746 drivers/staging/wusbcore/wa-xfer.c list_cut_position(&tmp_list, &wa->xfer_delayed_list, wa 1747 drivers/staging/wusbcore/wa-xfer.c wa->xfer_delayed_list.prev); wa 1748 drivers/staging/wusbcore/wa-xfer.c spin_unlock_irq(&wa->xfer_list_lock); wa 1770 drivers/staging/wusbcore/wa-xfer.c struct wahc *wa = container_of(ws, struct wahc, xfer_error_work); wa 1777 drivers/staging/wusbcore/wa-xfer.c spin_lock_irq(&wa->xfer_list_lock); wa 1778 drivers/staging/wusbcore/wa-xfer.c list_cut_position(&tmp_list, &wa->xfer_errored_list, wa 1779 drivers/staging/wusbcore/wa-xfer.c wa->xfer_errored_list.prev); wa 1780 drivers/staging/wusbcore/wa-xfer.c spin_unlock_irq(&wa->xfer_list_lock); wa 1797 drivers/staging/wusbcore/wa-xfer.c rpipe_clear_feature_stalled(wa, ep); wa 1820 drivers/staging/wusbcore/wa-xfer.c int wa_urb_enqueue(struct wahc *wa, struct usb_host_endpoint *ep, wa 1824 drivers/staging/wusbcore/wa-xfer.c struct device *dev = &wa->usb_iface->dev; wa 1837 drivers/staging/wusbcore/wa-xfer.c spin_lock_irqsave(&wa->xfer_list_lock, my_flags); wa 1838 drivers/staging/wusbcore/wa-xfer.c result = usb_hcd_link_urb_to_ep(&(wa->wusb->usb_hcd), urb); wa 1839 drivers/staging/wusbcore/wa-xfer.c spin_unlock_irqrestore(&wa->xfer_list_lock, my_flags); wa 1852 drivers/staging/wusbcore/wa-xfer.c xfer->wa = wa_get(wa); wa 1866 drivers/staging/wusbcore/wa-xfer.c spin_lock_irqsave(&wa->xfer_list_lock, my_flags); wa 1867 drivers/staging/wusbcore/wa-xfer.c list_add_tail(&xfer->list_node, &wa->xfer_delayed_list); wa 1868 drivers/staging/wusbcore/wa-xfer.c spin_unlock_irqrestore(&wa->xfer_list_lock, my_flags); wa 1869 drivers/staging/wusbcore/wa-xfer.c queue_work(wusbd, &wa->xfer_enqueue_work); wa 1880 drivers/staging/wusbcore/wa-xfer.c wa_put(xfer->wa); wa 1882 drivers/staging/wusbcore/wa-xfer.c spin_lock_irqsave(&wa->xfer_list_lock, my_flags); wa 1883 drivers/staging/wusbcore/wa-xfer.c usb_hcd_unlink_urb_from_ep(&(wa->wusb->usb_hcd), urb); wa 1884 drivers/staging/wusbcore/wa-xfer.c spin_unlock_irqrestore(&wa->xfer_list_lock, my_flags); wa 1893 drivers/staging/wusbcore/wa-xfer.c spin_lock_irqsave(&wa->xfer_list_lock, my_flags); wa 1894 drivers/staging/wusbcore/wa-xfer.c usb_hcd_unlink_urb_from_ep(&(wa->wusb->usb_hcd), urb); wa 1895 drivers/staging/wusbcore/wa-xfer.c spin_unlock_irqrestore(&wa->xfer_list_lock, my_flags); wa 1919 drivers/staging/wusbcore/wa-xfer.c int wa_urb_dequeue(struct wahc *wa, struct urb *urb, int status) wa 1930 drivers/staging/wusbcore/wa-xfer.c spin_lock_irqsave(&wa->xfer_list_lock, flags); wa 1931 drivers/staging/wusbcore/wa-xfer.c result = usb_hcd_check_unlink_urb(&(wa->wusb->usb_hcd), urb, status); wa 1939 drivers/staging/wusbcore/wa-xfer.c spin_unlock_irqrestore(&wa->xfer_list_lock, flags); wa 1967 drivers/staging/wusbcore/wa-xfer.c spin_lock(&wa->xfer_list_lock); wa 1970 drivers/staging/wusbcore/wa-xfer.c spin_unlock(&wa->xfer_list_lock); wa 2057 drivers/staging/wusbcore/wa-xfer.c spin_unlock(&wa->xfer_list_lock); wa 2162 drivers/staging/wusbcore/wa-xfer.c static int __wa_populate_buf_in_urb_isoc(struct wahc *wa, wa 2169 drivers/staging/wusbcore/wa-xfer.c const int dti_packet_size = usb_endpoint_maxp(wa->dti_epd); wa 2276 drivers/staging/wusbcore/wa-xfer.c static void wa_xfer_result_chew(struct wahc *wa, struct wa_xfer *xfer, wa 2280 drivers/staging/wusbcore/wa-xfer.c struct device *dev = &wa->usb_iface->dev; wa 2289 drivers/staging/wusbcore/wa-xfer.c struct urb *buf_in_urb = &(wa->buf_in_urbs[0]); wa 2333 drivers/staging/wusbcore/wa-xfer.c wa->dti_isoc_xfer_in_progress = wa_xfer_id(xfer); wa 2334 drivers/staging/wusbcore/wa-xfer.c wa->dti_isoc_xfer_seg = seg_idx; wa 2335 drivers/staging/wusbcore/wa-xfer.c wa->dti_state = WA_DTI_ISOC_PACKET_STATUS_PENDING; wa 2344 drivers/staging/wusbcore/wa-xfer.c ++(wa->active_buf_in_urbs); wa 2347 drivers/staging/wusbcore/wa-xfer.c --(wa->active_buf_in_urbs); wa 2364 drivers/staging/wusbcore/wa-xfer.c if (edc_inc(&wa->dti_edc, EDC_MAX_ERRORS, EDC_ERROR_TIMEFRAME)) { wa 2367 drivers/staging/wusbcore/wa-xfer.c wa_reset_all(wa); wa 2392 drivers/staging/wusbcore/wa-xfer.c spin_lock(&wa->xfer_list_lock); wa 2394 drivers/staging/wusbcore/wa-xfer.c list_move_tail(&xfer->list_node, &wa->xfer_errored_list); wa 2395 drivers/staging/wusbcore/wa-xfer.c spin_unlock(&wa->xfer_list_lock); wa 2397 drivers/staging/wusbcore/wa-xfer.c queue_work(wusbd, &wa->xfer_error_work); wa 2410 drivers/staging/wusbcore/wa-xfer.c wa_urb_dequeue(wa, xfer->urb, -ENOENT); wa 2413 drivers/staging/wusbcore/wa-xfer.c if (edc_inc(&wa->dti_edc, EDC_MAX_ERRORS, EDC_ERROR_TIMEFRAME)) { wa 2416 drivers/staging/wusbcore/wa-xfer.c wa_reset_all(wa); wa 2430 drivers/staging/wusbcore/wa-xfer.c static int wa_process_iso_packet_status(struct wahc *wa, struct urb *urb) wa 2432 drivers/staging/wusbcore/wa-xfer.c struct device *dev = &wa->usb_iface->dev; wa 2446 drivers/staging/wusbcore/wa-xfer.c packet_status = (struct wa_xfer_packet_status_hwaiso *)(wa->dti_buf); wa 2452 drivers/staging/wusbcore/wa-xfer.c xfer = wa_xfer_get_by_id(wa, wa->dti_isoc_xfer_in_progress); wa 2455 drivers/staging/wusbcore/wa-xfer.c wa->dti_isoc_xfer_in_progress); wa 2459 drivers/staging/wusbcore/wa-xfer.c if (unlikely(wa->dti_isoc_xfer_seg >= xfer->segs)) wa 2461 drivers/staging/wusbcore/wa-xfer.c seg = xfer->seg[wa->dti_isoc_xfer_seg]; wa 2478 drivers/staging/wusbcore/wa-xfer.c wa->wusb->usb_hcd.driver->get_frame_number(&wa->wusb->usb_hcd); wa 2513 drivers/staging/wusbcore/wa-xfer.c buf_in_urb = &(wa->buf_in_urbs[urb_index]); wa 2514 drivers/staging/wusbcore/wa-xfer.c urb_frame_count = __wa_populate_buf_in_urb_isoc(wa, wa 2520 drivers/staging/wusbcore/wa-xfer.c ++(wa->active_buf_in_urbs); wa 2541 drivers/staging/wusbcore/wa-xfer.c --(wa->active_buf_in_urbs); wa 2544 drivers/staging/wusbcore/wa-xfer.c wa_reset_all(wa); wa 2555 drivers/staging/wusbcore/wa-xfer.c wa->dti_state = WA_DTI_BUF_IN_DATA_PENDING; wa 2557 drivers/staging/wusbcore/wa-xfer.c wa->dti_state = WA_DTI_TRANSFER_RESULT_PENDING; wa 2586 drivers/staging/wusbcore/wa-xfer.c struct wahc *wa; wa 2599 drivers/staging/wusbcore/wa-xfer.c wa = xfer->wa; wa 2600 drivers/staging/wusbcore/wa-xfer.c dev = &wa->usb_iface->dev; wa 2601 drivers/staging/wusbcore/wa-xfer.c --(wa->active_buf_in_urbs); wa 2602 drivers/staging/wusbcore/wa-xfer.c active_buf_in_urbs = wa->active_buf_in_urbs; wa 2639 drivers/staging/wusbcore/wa-xfer.c urb_frame_count = __wa_populate_buf_in_urb_isoc(wa, urb, wa 2643 drivers/staging/wusbcore/wa-xfer.c ++(wa->active_buf_in_urbs); wa 2646 drivers/staging/wusbcore/wa-xfer.c --(wa->active_buf_in_urbs); wa 2649 drivers/staging/wusbcore/wa-xfer.c wa_reset_all(wa); wa 2684 drivers/staging/wusbcore/wa-xfer.c resubmit_dti = wa->dti_state != WA_DTI_TRANSFER_RESULT_PENDING; wa 2690 drivers/staging/wusbcore/wa-xfer.c if (edc_inc(&wa->nep_edc, EDC_MAX_ERRORS, wa 2694 drivers/staging/wusbcore/wa-xfer.c wa_reset_all(wa); wa 2713 drivers/staging/wusbcore/wa-xfer.c wa->dti_state = WA_DTI_TRANSFER_RESULT_PENDING; wa 2715 drivers/staging/wusbcore/wa-xfer.c result = usb_submit_urb(wa->dti_urb, GFP_ATOMIC); wa 2719 drivers/staging/wusbcore/wa-xfer.c wa_reset_all(wa); wa 2753 drivers/staging/wusbcore/wa-xfer.c struct wahc *wa = urb->context; wa 2754 drivers/staging/wusbcore/wa-xfer.c struct device *dev = &wa->usb_iface->dev; wa 2758 drivers/staging/wusbcore/wa-xfer.c BUG_ON(wa->dti_urb != urb); wa 2759 drivers/staging/wusbcore/wa-xfer.c switch (wa->dti_urb->status) { wa 2761 drivers/staging/wusbcore/wa-xfer.c if (wa->dti_state == WA_DTI_TRANSFER_RESULT_PENDING) { wa 2774 drivers/staging/wusbcore/wa-xfer.c xfer_result = (struct wa_xfer_result *)(wa->dti_buf); wa 2794 drivers/staging/wusbcore/wa-xfer.c xfer = wa_xfer_get_by_id(wa, xfer_id); wa 2801 drivers/staging/wusbcore/wa-xfer.c wa_xfer_result_chew(wa, xfer, xfer_result); wa 2803 drivers/staging/wusbcore/wa-xfer.c } else if (wa->dti_state == WA_DTI_ISOC_PACKET_STATUS_PENDING) { wa 2804 drivers/staging/wusbcore/wa-xfer.c dti_busy = wa_process_iso_packet_status(wa, urb); wa 2807 drivers/staging/wusbcore/wa-xfer.c wa->dti_state); wa 2816 drivers/staging/wusbcore/wa-xfer.c if (edc_inc(&wa->dti_edc, EDC_MAX_ERRORS, wa 2820 drivers/staging/wusbcore/wa-xfer.c wa_reset_all(wa); wa 2830 drivers/staging/wusbcore/wa-xfer.c result = usb_submit_urb(wa->dti_urb, GFP_ATOMIC); wa 2834 drivers/staging/wusbcore/wa-xfer.c wa_reset_all(wa); wa 2845 drivers/staging/wusbcore/wa-xfer.c int wa_dti_start(struct wahc *wa) wa 2847 drivers/staging/wusbcore/wa-xfer.c const struct usb_endpoint_descriptor *dti_epd = wa->dti_epd; wa 2848 drivers/staging/wusbcore/wa-xfer.c struct device *dev = &wa->usb_iface->dev; wa 2851 drivers/staging/wusbcore/wa-xfer.c if (wa->dti_urb != NULL) /* DTI URB already started */ wa 2854 drivers/staging/wusbcore/wa-xfer.c wa->dti_urb = usb_alloc_urb(0, GFP_KERNEL); wa 2855 drivers/staging/wusbcore/wa-xfer.c if (wa->dti_urb == NULL) wa 2858 drivers/staging/wusbcore/wa-xfer.c wa->dti_urb, wa->usb_dev, wa 2859 drivers/staging/wusbcore/wa-xfer.c usb_rcvbulkpipe(wa->usb_dev, 0x80 | dti_epd->bEndpointAddress), wa 2860 drivers/staging/wusbcore/wa-xfer.c wa->dti_buf, wa->dti_buf_size, wa 2861 drivers/staging/wusbcore/wa-xfer.c wa_dti_cb, wa); wa 2866 drivers/staging/wusbcore/wa-xfer.c &(wa->buf_in_urbs[index]), wa->usb_dev, wa 2867 drivers/staging/wusbcore/wa-xfer.c usb_rcvbulkpipe(wa->usb_dev, wa 2869 drivers/staging/wusbcore/wa-xfer.c NULL, 0, wa_buf_in_cb, wa); wa 2871 drivers/staging/wusbcore/wa-xfer.c result = usb_submit_urb(wa->dti_urb, GFP_KERNEL); wa 2881 drivers/staging/wusbcore/wa-xfer.c usb_put_urb(wa->dti_urb); wa 2882 drivers/staging/wusbcore/wa-xfer.c wa->dti_urb = NULL; wa 2903 drivers/staging/wusbcore/wa-xfer.c void wa_handle_notif_xfer(struct wahc *wa, struct wa_notif_hdr *notif_hdr) wa 2905 drivers/staging/wusbcore/wa-xfer.c struct device *dev = &wa->usb_iface->dev; wa 2907 drivers/staging/wusbcore/wa-xfer.c const struct usb_endpoint_descriptor *dti_epd = wa->dti_epd; wa 2920 drivers/staging/wusbcore/wa-xfer.c if (wa_dti_start(wa) < 0) wa 2926 drivers/staging/wusbcore/wa-xfer.c wa_reset_all(wa); wa 677 scripts/kallsyms.c int wa, wb; wa 689 scripts/kallsyms.c wa = (sa->sym[0] == 'w') || (sa->sym[0] == 'W'); wa 691 scripts/kallsyms.c if (wa != wb) wa 692 scripts/kallsyms.c return wa - wb; wa 695 scripts/kallsyms.c wa = may_be_linker_script_provide_symbol(sa); wa 697 scripts/kallsyms.c if (wa != wb) wa 698 scripts/kallsyms.c return wa - wb; wa 701 scripts/kallsyms.c wa = prefix_underscores_count((const char *)sa->sym + 1); wa 703 scripts/kallsyms.c if (wa != wb) wa 704 scripts/kallsyms.c return wa - wb;