hw_tail 152 drivers/crypto/cavium/zip/zip_device.c zip_dev->iq[queue].hw_tail); hw_tail 187 drivers/crypto/cavium/zip/zip_device.c zip_dev->iq[queue].hw_tail = zip_dev->iq[queue].sw_head; hw_tail 190 drivers/crypto/cavium/zip/zip_device.c zip_dev->iq[queue].hw_tail += 16; /* 16 64_bit words = 128B */ hw_tail 198 drivers/crypto/cavium/zip/zip_device.c zip_dev->iq[queue].hw_tail); hw_tail 181 drivers/crypto/cavium/zip/zip_main.c zip->iq[q].hw_tail = zip->iq[q].sw_head; hw_tail 199 drivers/crypto/cavium/zip/zip_main.c zip->iq[q].hw_tail); hw_tail 88 drivers/crypto/cavium/zip/zip_main.h u64 *hw_tail; hw_tail 463 drivers/gpu/drm/i915/i915_perf.c u32 head, hw_tail, aged_tail, aging_tail; hw_tail 482 drivers/gpu/drm/i915/i915_perf.c hw_tail = dev_priv->perf.ops.oa_hw_tail_read(stream); hw_tail 487 drivers/gpu/drm/i915/i915_perf.c hw_tail &= ~(report_size - 1); hw_tail 525 drivers/gpu/drm/i915/i915_perf.c OA_TAKEN(hw_tail, aged_tail) >= report_size)) { hw_tail 533 drivers/gpu/drm/i915/i915_perf.c if (hw_tail >= gtt_offset && hw_tail 534 drivers/gpu/drm/i915/i915_perf.c hw_tail < (gtt_offset + OA_BUFFER_SIZE)) { hw_tail 536 drivers/gpu/drm/i915/i915_perf.c aging_tail = hw_tail; hw_tail 540 drivers/gpu/drm/i915/i915_perf.c hw_tail); hw_tail 1683 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c int hw_head, hw_tail; hw_tail 1748 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c hw_tail = readl_relaxed(tx_ring->tqp->io_base + hw_tail 1767 drivers/net/ethernet/hisilicon/hns3/hns3_enet.c bd_num, hw_head, hw_tail, bd_err, hw_tail 5009 drivers/net/wireless/intel/ipw2x00/ipw2200.c u32 hw_tail; hw_tail 5013 drivers/net/wireless/intel/ipw2x00/ipw2200.c hw_tail = ipw_read32(priv, q->reg_r); hw_tail 5014 drivers/net/wireless/intel/ipw2x00/ipw2200.c if (hw_tail >= q->n_bd) { hw_tail 5017 drivers/net/wireless/intel/ipw2x00/ipw2200.c hw_tail, q->n_bd); hw_tail 5020 drivers/net/wireless/intel/ipw2x00/ipw2200.c for (; q->last_used != hw_tail;