itr 546 arch/ia64/include/asm/pal.h itr : 1, /* Fail in inst TR */ itr 730 arch/ia64/include/asm/pal.h #define pmci_tlb_instr_translation_reg_fail pme_tlb.itr itr 634 arch/ia64/kernel/mca_drv.c && !(ptci->itr || ptci->dtc || ptci->itc)) itr 80 arch/x86/platform/ts5500/ts5500.c bool itr; itr 140 arch/x86/platform/ts5500/ts5500.c sbc->itr = tmp & TS5500_ITR; itr 191 arch/x86/platform/ts5500/ts5500.c TS5500_ATTR_BOOL(itr); itr 5206 drivers/android/binder.c struct binder_proc *proc, *itr; itr 5244 drivers/android/binder.c hlist_for_each_entry(itr, &binder_procs, proc_node) { itr 5245 drivers/android/binder.c if (itr->pid == proc->pid) { itr 6004 drivers/android/binder.c struct binder_proc *itr; itr 6008 drivers/android/binder.c hlist_for_each_entry(itr, &binder_procs, proc_node) { itr 6009 drivers/android/binder.c if (itr->pid == pid) { itr 6011 drivers/android/binder.c print_binder_proc(m, itr, 1); itr 1620 drivers/base/power/domain.c struct gpd_link *link, *itr; itr 1650 drivers/base/power/domain.c list_for_each_entry(itr, &genpd->master_links, master_node) { itr 1651 drivers/base/power/domain.c if (itr->slave == subdomain && itr->master == genpd) { itr 80 drivers/crypto/qat/qat_common/adf_dev_mgr.c struct list_head *itr; itr 82 drivers/crypto/qat/qat_common/adf_dev_mgr.c list_for_each(itr, &vfs_table) { itr 84 drivers/crypto/qat/qat_common/adf_dev_mgr.c list_entry(itr, struct vf_id_map, list); itr 94 drivers/crypto/qat/qat_common/adf_dev_mgr.c struct list_head *itr; itr 96 drivers/crypto/qat/qat_common/adf_dev_mgr.c list_for_each(itr, &vfs_table) { itr 98 drivers/crypto/qat/qat_common/adf_dev_mgr.c list_entry(itr, struct vf_id_map, list); itr 144 drivers/crypto/qat/qat_common/adf_dev_mgr.c struct list_head *itr; itr 147 drivers/crypto/qat/qat_common/adf_dev_mgr.c list_for_each(itr, &accel_table) { itr 149 drivers/crypto/qat/qat_common/adf_dev_mgr.c list_entry(itr, struct adf_accel_dev, list); itr 186 drivers/crypto/qat/qat_common/adf_dev_mgr.c struct list_head *itr; itr 202 drivers/crypto/qat/qat_common/adf_dev_mgr.c list_for_each(itr, &accel_table) { itr 204 drivers/crypto/qat/qat_common/adf_dev_mgr.c list_entry(itr, struct adf_accel_dev, list); itr 341 drivers/crypto/qat/qat_common/adf_dev_mgr.c struct list_head *itr; itr 344 drivers/crypto/qat/qat_common/adf_dev_mgr.c list_for_each(itr, &accel_table) { itr 346 drivers/crypto/qat/qat_common/adf_dev_mgr.c list_entry(itr, struct adf_accel_dev, list); itr 360 drivers/crypto/qat/qat_common/adf_dev_mgr.c struct list_head *itr; itr 370 drivers/crypto/qat/qat_common/adf_dev_mgr.c list_for_each(itr, &accel_table) { itr 372 drivers/crypto/qat/qat_common/adf_dev_mgr.c list_entry(itr, struct adf_accel_dev, list); itr 396 drivers/crypto/qat/qat_common/adf_dev_mgr.c struct list_head *itr; itr 400 drivers/crypto/qat/qat_common/adf_dev_mgr.c list_for_each(itr, &vfs_table) { itr 402 drivers/crypto/qat/qat_common/adf_dev_mgr.c list_entry(itr, struct vf_id_map, list); itr 35 drivers/crypto/virtio/virtio_crypto_mgr.c struct list_head *itr; itr 45 drivers/crypto/virtio/virtio_crypto_mgr.c list_for_each(itr, &virtio_crypto_table) { itr 47 drivers/crypto/virtio/virtio_crypto_mgr.c list_entry(itr, struct virtio_crypto, list); itr 223 drivers/gpu/drm/i915/gvt/kvmgt.c struct gvt_dma *itr; itr 226 drivers/gpu/drm/i915/gvt/kvmgt.c itr = rb_entry(node, struct gvt_dma, dma_addr_node); itr 228 drivers/gpu/drm/i915/gvt/kvmgt.c if (dma_addr < itr->dma_addr) itr 230 drivers/gpu/drm/i915/gvt/kvmgt.c else if (dma_addr > itr->dma_addr) itr 233 drivers/gpu/drm/i915/gvt/kvmgt.c return itr; itr 241 drivers/gpu/drm/i915/gvt/kvmgt.c struct gvt_dma *itr; itr 244 drivers/gpu/drm/i915/gvt/kvmgt.c itr = rb_entry(node, struct gvt_dma, gfn_node); itr 246 drivers/gpu/drm/i915/gvt/kvmgt.c if (gfn < itr->gfn) itr 248 drivers/gpu/drm/i915/gvt/kvmgt.c else if (gfn > itr->gfn) itr 251 drivers/gpu/drm/i915/gvt/kvmgt.c return itr; itr 259 drivers/gpu/drm/i915/gvt/kvmgt.c struct gvt_dma *new, *itr; itr 276 drivers/gpu/drm/i915/gvt/kvmgt.c itr = rb_entry(parent, struct gvt_dma, gfn_node); itr 278 drivers/gpu/drm/i915/gvt/kvmgt.c if (gfn < itr->gfn) itr 291 drivers/gpu/drm/i915/gvt/kvmgt.c itr = rb_entry(parent, struct gvt_dma, dma_addr_node); itr 293 drivers/gpu/drm/i915/gvt/kvmgt.c if (dma_addr < itr->dma_addr) itr 1737 drivers/gpu/drm/i915/gvt/kvmgt.c struct intel_vgpu *itr; itr 1743 drivers/gpu/drm/i915/gvt/kvmgt.c for_each_active_vgpu(vgpu->gvt, itr, id) { itr 1744 drivers/gpu/drm/i915/gvt/kvmgt.c if (!handle_valid(itr->handle)) itr 1747 drivers/gpu/drm/i915/gvt/kvmgt.c info = (struct kvmgt_guest_info *)itr->handle; itr 139 drivers/i2c/busses/i2c-qcom-geni.c const struct geni_i2c_clk_fld *itr = geni_i2c_clk_map; itr 141 drivers/i2c/busses/i2c-qcom-geni.c for (i = 0; i < ARRAY_SIZE(geni_i2c_clk_map); i++, itr++) { itr 142 drivers/i2c/busses/i2c-qcom-geni.c if (itr->clk_freq_out == gi2c->clk_freq_out) { itr 143 drivers/i2c/busses/i2c-qcom-geni.c gi2c->clk_fld = itr; itr 152 drivers/i2c/busses/i2c-qcom-geni.c const struct geni_i2c_clk_fld *itr = gi2c->clk_fld; itr 157 drivers/i2c/busses/i2c-qcom-geni.c val = (itr->clk_div << CLK_DIV_SHFT) | SER_CLK_EN; itr 160 drivers/i2c/busses/i2c-qcom-geni.c val = itr->t_high_cnt << HIGH_COUNTER_SHFT; itr 161 drivers/i2c/busses/i2c-qcom-geni.c val |= itr->t_low_cnt << LOW_COUNTER_SHFT; itr 162 drivers/i2c/busses/i2c-qcom-geni.c val |= itr->t_cycle_cnt; itr 293 drivers/net/ethernet/aquantia/atlantic/aq_ethtool.c if (cfg->itr == AQ_CFG_INTERRUPT_MODERATION_ON || itr 294 drivers/net/ethernet/aquantia/atlantic/aq_ethtool.c cfg->itr == AQ_CFG_INTERRUPT_MODERATION_AUTO) { itr 342 drivers/net/ethernet/aquantia/atlantic/aq_ethtool.c cfg->itr = AQ_CFG_INTERRUPT_MODERATION_ON; itr 70 drivers/net/ethernet/aquantia/atlantic/aq_nic.c cfg->itr = aq_itr; itr 29 drivers/net/ethernet/aquantia/atlantic/aq_nic.h u32 itr; itr 816 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c if (self->aq_nic_cfg->itr) { itr 817 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c if (self->aq_nic_cfg->itr != AQ_CFG_INTERRUPT_MODERATION_AUTO) { itr 818 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_a0.c u32 itr_ = (self->aq_nic_cfg->itr >> 1); itr 887 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c switch (self->aq_nic_cfg->itr) { itr 895 drivers/net/ethernet/aquantia/atlantic/hw_atl/hw_atl_b0.c if (self->aq_nic_cfg->itr == AQ_CFG_INTERRUPT_MODERATION_ON) { itr 221 drivers/net/ethernet/intel/e1000/e1000.h u32 itr; itr 1775 drivers/net/ethernet/intel/e1000/e1000_ethtool.c adapter->itr = adapter->itr_setting = 4; itr 1777 drivers/net/ethernet/intel/e1000/e1000_ethtool.c adapter->itr = 20000; itr 1780 drivers/net/ethernet/intel/e1000/e1000_ethtool.c adapter->itr = (1000000 / ec->rx_coalesce_usecs); itr 1781 drivers/net/ethernet/intel/e1000/e1000_ethtool.c adapter->itr_setting = adapter->itr & ~3; itr 1785 drivers/net/ethernet/intel/e1000/e1000_ethtool.c ew32(ITR, 1000000000 / (adapter->itr * 256)); itr 1874 drivers/net/ethernet/intel/e1000/e1000_main.c ew32(ITR, 1000000000 / (adapter->itr * 256)); itr 2530 drivers/net/ethernet/intel/e1000/e1000_main.c u32 itr = goc > 0 ? (dif * 6000 / goc + 2000) : 8000; itr 2532 drivers/net/ethernet/intel/e1000/e1000_main.c ew32(ITR, 1000000000 / (itr * 256)); itr 2622 drivers/net/ethernet/intel/e1000/e1000_main.c u32 new_itr = adapter->itr; itr 2666 drivers/net/ethernet/intel/e1000/e1000_main.c if (new_itr != adapter->itr) { itr 2671 drivers/net/ethernet/intel/e1000/e1000_main.c new_itr = new_itr > adapter->itr ? itr 2672 drivers/net/ethernet/intel/e1000/e1000_main.c min(adapter->itr + (new_itr >> 2), new_itr) : itr 2674 drivers/net/ethernet/intel/e1000/e1000_main.c adapter->itr = new_itr; itr 438 drivers/net/ethernet/intel/e1000/e1000_param.c adapter->itr = InterruptThrottleRate[bd]; itr 439 drivers/net/ethernet/intel/e1000/e1000_param.c switch (adapter->itr) { itr 446 drivers/net/ethernet/intel/e1000/e1000_param.c adapter->itr_setting = adapter->itr; itr 447 drivers/net/ethernet/intel/e1000/e1000_param.c adapter->itr = 20000; itr 452 drivers/net/ethernet/intel/e1000/e1000_param.c adapter->itr_setting = adapter->itr; itr 453 drivers/net/ethernet/intel/e1000/e1000_param.c adapter->itr = 20000; itr 458 drivers/net/ethernet/intel/e1000/e1000_param.c adapter->itr_setting = adapter->itr; itr 461 drivers/net/ethernet/intel/e1000/e1000_param.c e1000_validate_option(&adapter->itr, &opt, itr 468 drivers/net/ethernet/intel/e1000/e1000_param.c adapter->itr_setting = adapter->itr & ~3; itr 473 drivers/net/ethernet/intel/e1000/e1000_param.c adapter->itr = 20000; itr 209 drivers/net/ethernet/intel/e1000e/e1000.h u32 itr; itr 485 drivers/net/ethernet/intel/e1000e/e1000.h void e1000e_write_itr(struct e1000_adapter *adapter, u32 itr); itr 2007 drivers/net/ethernet/intel/e1000e/ethtool.c adapter->itr = adapter->itr_setting; itr 2009 drivers/net/ethernet/intel/e1000e/ethtool.c adapter->itr = 20000; itr 2012 drivers/net/ethernet/intel/e1000e/ethtool.c adapter->itr = (1000000 / ec->rx_coalesce_usecs); itr 2013 drivers/net/ethernet/intel/e1000e/ethtool.c adapter->itr_setting = adapter->itr & ~3; itr 2019 drivers/net/ethernet/intel/e1000e/ethtool.c e1000e_write_itr(adapter, adapter->itr); itr 1947 drivers/net/ethernet/intel/e1000e/netdev.c u32 itr = rx_ring->itr_val ? itr 1950 drivers/net/ethernet/intel/e1000e/netdev.c writel(itr, rx_ring->itr_register); itr 2119 drivers/net/ethernet/intel/e1000e/netdev.c adapter->rx_ring->itr_val = adapter->itr; itr 2135 drivers/net/ethernet/intel/e1000e/netdev.c adapter->tx_ring->itr_val = adapter->itr; itr 2537 drivers/net/ethernet/intel/e1000e/netdev.c u32 new_itr = adapter->itr; itr 2583 drivers/net/ethernet/intel/e1000e/netdev.c if (new_itr != adapter->itr) { itr 2588 drivers/net/ethernet/intel/e1000e/netdev.c new_itr = new_itr > adapter->itr ? itr 2589 drivers/net/ethernet/intel/e1000e/netdev.c min(adapter->itr + (new_itr >> 2), new_itr) : new_itr; itr 2590 drivers/net/ethernet/intel/e1000e/netdev.c adapter->itr = new_itr; itr 2608 drivers/net/ethernet/intel/e1000e/netdev.c void e1000e_write_itr(struct e1000_adapter *adapter, u32 itr) itr 2611 drivers/net/ethernet/intel/e1000e/netdev.c u32 new_itr = itr ? 1000000000 / (itr * 256) : 0; itr 3231 drivers/net/ethernet/intel/e1000e/netdev.c if ((adapter->itr_setting != 0) && (adapter->itr != 0)) itr 3232 drivers/net/ethernet/intel/e1000e/netdev.c e1000e_write_itr(adapter, adapter->itr); itr 4088 drivers/net/ethernet/intel/e1000e/netdev.c adapter->itr = 20000; itr 4089 drivers/net/ethernet/intel/e1000e/netdev.c e1000e_write_itr(adapter, adapter->itr); itr 5380 drivers/net/ethernet/intel/e1000e/netdev.c u32 itr = goc > 0 ? (dif * 6000 / goc + 2000) : 8000; itr 5382 drivers/net/ethernet/intel/e1000e/netdev.c e1000e_write_itr(adapter, itr); itr 337 drivers/net/ethernet/intel/e1000e/param.c adapter->itr = InterruptThrottleRate[bd]; itr 344 drivers/net/ethernet/intel/e1000e/param.c if ((adapter->itr > 4) && itr 345 drivers/net/ethernet/intel/e1000e/param.c e1000_validate_option(&adapter->itr, &opt, adapter)) itr 346 drivers/net/ethernet/intel/e1000e/param.c adapter->itr = opt.def; itr 351 drivers/net/ethernet/intel/e1000e/param.c adapter->itr = opt.def; itr 356 drivers/net/ethernet/intel/e1000e/param.c if (adapter->itr > 4) itr 359 drivers/net/ethernet/intel/e1000e/param.c adapter->itr); itr 362 drivers/net/ethernet/intel/e1000e/param.c adapter->itr_setting = adapter->itr; itr 363 drivers/net/ethernet/intel/e1000e/param.c switch (adapter->itr) { itr 371 drivers/net/ethernet/intel/e1000e/param.c adapter->itr = 20000; itr 383 drivers/net/ethernet/intel/e1000e/param.c adapter->itr = 20000; itr 150 drivers/net/ethernet/intel/fm10k/fm10k.h u16 itr; /* interrupt throttle rate value */ itr 161 drivers/net/ethernet/intel/fm10k/fm10k.h #define ITR_IS_ADAPTIVE(itr) (!!(itr & FM10K_ITR_ADAPTIVE)) itr 187 drivers/net/ethernet/intel/fm10k/fm10k.h u32 __iomem *itr; /* pointer to ITR register for this vector */ itr 680 drivers/net/ethernet/intel/fm10k/fm10k_ethtool.c qv->tx.itr = tx_itr; itr 681 drivers/net/ethernet/intel/fm10k/fm10k_ethtool.c qv->rx.itr = rx_itr; itr 1354 drivers/net/ethernet/intel/fm10k/fm10k_main.c if (!ITR_IS_ADAPTIVE(ring_container->itr)) itr 1403 drivers/net/ethernet/intel/fm10k/fm10k_main.c ring_container->itr = avg_wire_size | FM10K_ITR_ADAPTIVE; itr 1413 drivers/net/ethernet/intel/fm10k/fm10k_main.c u32 itr = FM10K_ITR_ENABLE; itr 1422 drivers/net/ethernet/intel/fm10k/fm10k_main.c itr |= (q_vector->tx.itr & FM10K_ITR_MAX); itr 1425 drivers/net/ethernet/intel/fm10k/fm10k_main.c itr |= (q_vector->rx.itr & FM10K_ITR_MAX) << FM10K_ITR_INTERVAL1_SHIFT; itr 1428 drivers/net/ethernet/intel/fm10k/fm10k_main.c writel(itr, q_vector->itr); itr 1630 drivers/net/ethernet/intel/fm10k/fm10k_main.c q_vector->tx.itr = interface->tx_itr; itr 1659 drivers/net/ethernet/intel/fm10k/fm10k_main.c q_vector->rx.itr = interface->rx_itr; itr 724 drivers/net/ethernet/intel/fm10k/fm10k_pci.c writel(FM10K_ITR_ENABLE | FM10K_ITR_PENDING2, qv->itr); itr 1515 drivers/net/ethernet/intel/fm10k/fm10k_pci.c u32 itr = entry->entry | FM10K_INT_MAP_TIMER0; itr 1532 drivers/net/ethernet/intel/fm10k/fm10k_pci.c fm10k_write_reg(hw, FM10K_VFINT_MAP, itr); itr 1741 drivers/net/ethernet/intel/fm10k/fm10k_pci.c writel(FM10K_ITR_MASK_SET, q_vector->itr); itr 1784 drivers/net/ethernet/intel/fm10k/fm10k_pci.c q_vector->itr = (hw->mac.type == fm10k_mac_pf) ? itr 1802 drivers/net/ethernet/intel/fm10k/fm10k_pci.c writel(FM10K_ITR_ENABLE, q_vector->itr); itr 1825 drivers/net/ethernet/intel/fm10k/fm10k_pci.c writel(FM10K_ITR_MASK_SET, q_vector->itr); itr 1008 drivers/net/ethernet/intel/i40e/i40e_txrx.c unsigned int avg_wire_size, packets, bytes, itr; itr 1020 drivers/net/ethernet/intel/i40e/i40e_txrx.c itr = i40e_container_is_rx(q_vector, rc) ? itr 1039 drivers/net/ethernet/intel/i40e/i40e_txrx.c itr = rc->target_itr; itr 1054 drivers/net/ethernet/intel/i40e/i40e_txrx.c itr = I40E_ITR_ADAPTIVE_LATENCY; itr 1083 drivers/net/ethernet/intel/i40e/i40e_txrx.c itr = rc->target_itr + I40E_ITR_ADAPTIVE_MIN_INC; itr 1084 drivers/net/ethernet/intel/i40e/i40e_txrx.c if ((itr & I40E_ITR_MASK) > I40E_ITR_ADAPTIVE_MAX_USECS) { itr 1085 drivers/net/ethernet/intel/i40e/i40e_txrx.c itr &= I40E_ITR_ADAPTIVE_LATENCY; itr 1086 drivers/net/ethernet/intel/i40e/i40e_txrx.c itr += I40E_ITR_ADAPTIVE_MAX_USECS; itr 1092 drivers/net/ethernet/intel/i40e/i40e_txrx.c itr = min(q_vector->tx.current_itr, q_vector->rx.current_itr); itr 1093 drivers/net/ethernet/intel/i40e/i40e_txrx.c itr &= I40E_ITR_MASK; itr 1107 drivers/net/ethernet/intel/i40e/i40e_txrx.c itr /= 2; itr 1108 drivers/net/ethernet/intel/i40e/i40e_txrx.c itr &= I40E_ITR_MASK; itr 1109 drivers/net/ethernet/intel/i40e/i40e_txrx.c if (itr < I40E_ITR_ADAPTIVE_MIN_USECS) itr 1110 drivers/net/ethernet/intel/i40e/i40e_txrx.c itr = I40E_ITR_ADAPTIVE_MIN_USECS; itr 1121 drivers/net/ethernet/intel/i40e/i40e_txrx.c itr = I40E_ITR_ADAPTIVE_BULK; itr 1169 drivers/net/ethernet/intel/i40e/i40e_txrx.c if (itr & I40E_ITR_ADAPTIVE_LATENCY) itr 1179 drivers/net/ethernet/intel/i40e/i40e_txrx.c itr += DIV_ROUND_UP(avg_wire_size, i40e_itr_divisor(q_vector)) * itr 1182 drivers/net/ethernet/intel/i40e/i40e_txrx.c if ((itr & I40E_ITR_MASK) > I40E_ITR_ADAPTIVE_MAX_USECS) { itr 1183 drivers/net/ethernet/intel/i40e/i40e_txrx.c itr &= I40E_ITR_ADAPTIVE_LATENCY; itr 1184 drivers/net/ethernet/intel/i40e/i40e_txrx.c itr += I40E_ITR_ADAPTIVE_MAX_USECS; itr 1189 drivers/net/ethernet/intel/i40e/i40e_txrx.c rc->target_itr = itr; itr 2456 drivers/net/ethernet/intel/i40e/i40e_txrx.c static inline u32 i40e_buildreg_itr(const int type, u16 itr) itr 2475 drivers/net/ethernet/intel/i40e/i40e_txrx.c itr &= I40E_ITR_MASK; itr 2479 drivers/net/ethernet/intel/i40e/i40e_txrx.c (itr << (I40E_PFINT_DYN_CTLN_INTERVAL_SHIFT - 1)); itr 418 drivers/net/ethernet/intel/iavf/iavf_txrx.c unsigned int avg_wire_size, packets, bytes, itr; itr 430 drivers/net/ethernet/intel/iavf/iavf_txrx.c itr = iavf_container_is_rx(q_vector, rc) ? itr 449 drivers/net/ethernet/intel/iavf/iavf_txrx.c itr = rc->target_itr; itr 464 drivers/net/ethernet/intel/iavf/iavf_txrx.c itr = IAVF_ITR_ADAPTIVE_LATENCY; itr 493 drivers/net/ethernet/intel/iavf/iavf_txrx.c itr = rc->target_itr + IAVF_ITR_ADAPTIVE_MIN_INC; itr 494 drivers/net/ethernet/intel/iavf/iavf_txrx.c if ((itr & IAVF_ITR_MASK) > IAVF_ITR_ADAPTIVE_MAX_USECS) { itr 495 drivers/net/ethernet/intel/iavf/iavf_txrx.c itr &= IAVF_ITR_ADAPTIVE_LATENCY; itr 496 drivers/net/ethernet/intel/iavf/iavf_txrx.c itr += IAVF_ITR_ADAPTIVE_MAX_USECS; itr 502 drivers/net/ethernet/intel/iavf/iavf_txrx.c itr = min(q_vector->tx.current_itr, q_vector->rx.current_itr); itr 503 drivers/net/ethernet/intel/iavf/iavf_txrx.c itr &= IAVF_ITR_MASK; itr 517 drivers/net/ethernet/intel/iavf/iavf_txrx.c itr /= 2; itr 518 drivers/net/ethernet/intel/iavf/iavf_txrx.c itr &= IAVF_ITR_MASK; itr 519 drivers/net/ethernet/intel/iavf/iavf_txrx.c if (itr < IAVF_ITR_ADAPTIVE_MIN_USECS) itr 520 drivers/net/ethernet/intel/iavf/iavf_txrx.c itr = IAVF_ITR_ADAPTIVE_MIN_USECS; itr 531 drivers/net/ethernet/intel/iavf/iavf_txrx.c itr = IAVF_ITR_ADAPTIVE_BULK; itr 579 drivers/net/ethernet/intel/iavf/iavf_txrx.c if (itr & IAVF_ITR_ADAPTIVE_LATENCY) itr 589 drivers/net/ethernet/intel/iavf/iavf_txrx.c itr += DIV_ROUND_UP(avg_wire_size, iavf_itr_divisor(q_vector)) * itr 592 drivers/net/ethernet/intel/iavf/iavf_txrx.c if ((itr & IAVF_ITR_MASK) > IAVF_ITR_ADAPTIVE_MAX_USECS) { itr 593 drivers/net/ethernet/intel/iavf/iavf_txrx.c itr &= IAVF_ITR_ADAPTIVE_LATENCY; itr 594 drivers/net/ethernet/intel/iavf/iavf_txrx.c itr += IAVF_ITR_ADAPTIVE_MAX_USECS; itr 599 drivers/net/ethernet/intel/iavf/iavf_txrx.c rc->target_itr = itr; itr 1597 drivers/net/ethernet/intel/iavf/iavf_txrx.c static inline u32 iavf_buildreg_itr(const int type, u16 itr) itr 1616 drivers/net/ethernet/intel/iavf/iavf_txrx.c itr &= IAVF_ITR_MASK; itr 1620 drivers/net/ethernet/intel/iavf/iavf_txrx.c (itr << (IAVF_VFINT_DYN_CTLN1_INTERVAL_SHIFT - 1)); itr 395 drivers/net/ethernet/intel/ice/ice.h int itr = ICE_ITR_NONE; itr 402 drivers/net/ethernet/intel/ice/ice.h (itr << GLINT_DYN_CTL_ITR_INDX_S); itr 2423 drivers/net/ethernet/intel/ice/ice_switch.c struct ice_fltr_mgmt_list_entry *itr; itr 2443 drivers/net/ethernet/intel/ice/ice_switch.c list_for_each_entry(itr, rule_head, list_entry) { itr 2446 drivers/net/ethernet/intel/ice/ice_switch.c if (!ice_vsi_uses_fltr(itr, vsi_handle)) itr 2450 drivers/net/ethernet/intel/ice/ice_switch.c ice_determine_promisc_mask(&itr->fltr_info); itr 2458 drivers/net/ethernet/intel/ice/ice_switch.c &itr->fltr_info); itr 2726 drivers/net/ethernet/intel/ice/ice_switch.c struct ice_fltr_mgmt_list_entry *itr; itr 2734 drivers/net/ethernet/intel/ice/ice_switch.c list_for_each_entry(itr, list_head, list_entry) { itr 2737 drivers/net/ethernet/intel/ice/ice_switch.c f_entry.fltr_info = itr->fltr_info; itr 2738 drivers/net/ethernet/intel/ice/ice_switch.c if (itr->vsi_count < 2 && recp_id != ICE_SW_LKUP_VLAN && itr 2739 drivers/net/ethernet/intel/ice/ice_switch.c itr->fltr_info.vsi_handle == vsi_handle) { itr 2748 drivers/net/ethernet/intel/ice/ice_switch.c if (!itr->vsi_list_info || itr 2749 drivers/net/ethernet/intel/ice/ice_switch.c !test_bit(vsi_handle, itr->vsi_list_info->vsi_map)) itr 2752 drivers/net/ethernet/intel/ice/ice_switch.c clear_bit(vsi_handle, itr->vsi_list_info->vsi_map); itr 1140 drivers/net/ethernet/intel/ice/ice_txrx.c unsigned int itr) itr 1144 drivers/net/ethernet/intel/ice/ice_txrx.c itr += DIV_ROUND_UP(17 * (avg_pkt_size + 24), itr 1148 drivers/net/ethernet/intel/ice/ice_txrx.c itr += DIV_ROUND_UP(34 * (avg_pkt_size + 24), itr 1152 drivers/net/ethernet/intel/ice/ice_txrx.c itr += DIV_ROUND_UP(43 * (avg_pkt_size + 24), itr 1156 drivers/net/ethernet/intel/ice/ice_txrx.c itr += DIV_ROUND_UP(68 * (avg_pkt_size + 24), itr 1160 drivers/net/ethernet/intel/ice/ice_txrx.c itr += DIV_ROUND_UP(85 * (avg_pkt_size + 24), itr 1166 drivers/net/ethernet/intel/ice/ice_txrx.c itr += DIV_ROUND_UP(170 * (avg_pkt_size + 24), itr 1171 drivers/net/ethernet/intel/ice/ice_txrx.c if ((itr & ICE_ITR_MASK) > ICE_ITR_ADAPTIVE_MAX_USECS) { itr 1172 drivers/net/ethernet/intel/ice/ice_txrx.c itr &= ICE_ITR_ADAPTIVE_LATENCY; itr 1173 drivers/net/ethernet/intel/ice/ice_txrx.c itr += ICE_ITR_ADAPTIVE_MAX_USECS; itr 1176 drivers/net/ethernet/intel/ice/ice_txrx.c return itr; itr 1196 drivers/net/ethernet/intel/ice/ice_txrx.c unsigned int packets, bytes, itr; itr 1209 drivers/net/ethernet/intel/ice/ice_txrx.c itr = rc->target_itr; itr 1217 drivers/net/ethernet/intel/ice/ice_txrx.c itr = container_is_rx ? itr 1242 drivers/net/ethernet/intel/ice/ice_txrx.c itr = ICE_ITR_ADAPTIVE_LATENCY; itr 1271 drivers/net/ethernet/intel/ice/ice_txrx.c itr = rc->target_itr + ICE_ITR_ADAPTIVE_MIN_INC; itr 1272 drivers/net/ethernet/intel/ice/ice_txrx.c if ((itr & ICE_ITR_MASK) > ICE_ITR_ADAPTIVE_MAX_USECS) { itr 1273 drivers/net/ethernet/intel/ice/ice_txrx.c itr &= ICE_ITR_ADAPTIVE_LATENCY; itr 1274 drivers/net/ethernet/intel/ice/ice_txrx.c itr += ICE_ITR_ADAPTIVE_MAX_USECS; itr 1280 drivers/net/ethernet/intel/ice/ice_txrx.c itr = min(q_vector->tx.current_itr, q_vector->rx.current_itr); itr 1281 drivers/net/ethernet/intel/ice/ice_txrx.c itr &= ICE_ITR_MASK; itr 1295 drivers/net/ethernet/intel/ice/ice_txrx.c itr >>= 1; itr 1296 drivers/net/ethernet/intel/ice/ice_txrx.c itr &= ICE_ITR_MASK; itr 1297 drivers/net/ethernet/intel/ice/ice_txrx.c if (itr < ICE_ITR_ADAPTIVE_MIN_USECS) itr 1298 drivers/net/ethernet/intel/ice/ice_txrx.c itr = ICE_ITR_ADAPTIVE_MIN_USECS; itr 1309 drivers/net/ethernet/intel/ice/ice_txrx.c itr = ICE_ITR_ADAPTIVE_BULK; itr 1314 drivers/net/ethernet/intel/ice/ice_txrx.c itr = ice_adjust_itr_by_size_and_speed(q_vector->vsi->port_info, itr 1315 drivers/net/ethernet/intel/ice/ice_txrx.c bytes / packets, itr); itr 1319 drivers/net/ethernet/intel/ice/ice_txrx.c rc->target_itr = itr; itr 1333 drivers/net/ethernet/intel/ice/ice_txrx.c static u32 ice_buildreg_itr(u16 itr_idx, u16 itr) itr 1342 drivers/net/ethernet/intel/ice/ice_txrx.c itr &= ICE_ITR_MASK; itr 1346 drivers/net/ethernet/intel/ice/ice_txrx.c (itr << (GLINT_DYN_CTL_INTERVAL_S - ICE_ITR_GRAN_S)); itr 245 drivers/net/ethernet/intel/igb/igb.h u8 itr; /* current ITR setting for ring */ itr 5542 drivers/net/ethernet/intel/igb/igb_main.c u8 itrval = ring_container->itr; itr 5586 drivers/net/ethernet/intel/igb/igb_main.c ring_container->itr = itrval; itr 5605 drivers/net/ethernet/intel/igb/igb_main.c current_itr = max(q_vector->rx.itr, q_vector->tx.itr); itr 240 drivers/net/ethernet/intel/igc/igc.h u8 itr; /* current ITR setting for ring */ itr 3171 drivers/net/ethernet/intel/igc/igc_main.c u8 itrval = ring_container->itr; itr 3215 drivers/net/ethernet/intel/igc/igc_main.c ring_container->itr = itrval; itr 3315 drivers/net/ethernet/intel/igc/igc_main.c current_itr = max(q_vector->rx.itr, q_vector->tx.itr); itr 433 drivers/net/ethernet/intel/ixgbe/ixgbe.h u8 itr; /* current ITR setting for ring */ itr 455 drivers/net/ethernet/intel/ixgbe/ixgbe.h u16 itr; /* Interrupt throttle rate written to EITR */ itr 2392 drivers/net/ethernet/intel/ixgbe/ixgbe_ethtool.c q_vector->itr = tx_itr_param; itr 2395 drivers/net/ethernet/intel/ixgbe/ixgbe_ethtool.c q_vector->itr = rx_itr_param; itr 887 drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c q_vector->tx.itr = IXGBE_ITR_ADAPTIVE_MAX_USECS | itr 889 drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c q_vector->rx.itr = IXGBE_ITR_ADAPTIVE_MAX_USECS | itr 896 drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c q_vector->itr = IXGBE_12K_ITR; itr 898 drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c q_vector->itr = adapter->tx_itr_setting; itr 902 drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c q_vector->itr = IXGBE_20K_ITR; itr 904 drivers/net/ethernet/intel/ixgbe/ixgbe_lib.c q_vector->itr = adapter->rx_itr_setting; itr 2507 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c unsigned int itr = IXGBE_ITR_ADAPTIVE_MIN_USECS | itr 2537 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c itr = (q_vector->itr >> 2) + IXGBE_ITR_ADAPTIVE_MIN_INC; itr 2538 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c if (itr > IXGBE_ITR_ADAPTIVE_MAX_USECS) itr 2539 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c itr = IXGBE_ITR_ADAPTIVE_MAX_USECS; itr 2540 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c itr += ring_container->itr & IXGBE_ITR_ADAPTIVE_LATENCY; itr 2551 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c itr = IXGBE_ITR_ADAPTIVE_LATENCY; itr 2560 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c itr = (q_vector->itr >> 2) + IXGBE_ITR_ADAPTIVE_MIN_INC; itr 2561 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c if (itr > IXGBE_ITR_ADAPTIVE_MAX_USECS) itr 2562 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c itr = IXGBE_ITR_ADAPTIVE_MAX_USECS; itr 2570 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c itr = q_vector->itr >> 2; itr 2579 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c itr = q_vector->itr >> 3; itr 2580 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c if (itr < IXGBE_ITR_ADAPTIVE_MIN_USECS) itr 2581 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c itr = IXGBE_ITR_ADAPTIVE_MIN_USECS; itr 2591 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c itr = IXGBE_ITR_ADAPTIVE_BULK; itr 2639 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c if (itr & IXGBE_ITR_ADAPTIVE_LATENCY) itr 2653 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c itr += DIV_ROUND_UP(avg_wire_size, itr 2662 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c itr += DIV_ROUND_UP(avg_wire_size, itr 2670 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c ring_container->itr = itr; itr 2692 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c u32 itr_reg = q_vector->itr & IXGBE_MAX_EITR; itr 2724 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c new_itr = min(q_vector->rx.itr, q_vector->tx.itr); itr 2730 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c if (new_itr != q_vector->itr) { itr 2732 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c q_vector->itr = new_itr; itr 3513 drivers/net/ethernet/intel/ixgbe/ixgbe_main.c if (!ring->q_vector || (ring->q_vector->itr < IXGBE_100K_ITR)) itr 858 drivers/net/ethernet/intel/ixgbevf/ethtool.c q_vector->itr = tx_itr_param; itr 861 drivers/net/ethernet/intel/ixgbevf/ethtool.c q_vector->itr = rx_itr_param; itr 221 drivers/net/ethernet/intel/ixgbevf/ixgbevf.h u8 itr; /* current ITR setting for ring */ itr 237 drivers/net/ethernet/intel/ixgbevf/ixgbevf.h u16 itr; /* Interrupt throttle rate written to EITR */ itr 1323 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c u32 itr_reg = q_vector->itr & IXGBE_MAX_EITR; itr 1365 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c q_vector->itr = IXGBE_12K_ITR; itr 1367 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c q_vector->itr = adapter->tx_itr_setting; itr 1371 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c q_vector->itr = IXGBE_20K_ITR; itr 1373 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c q_vector->itr = adapter->rx_itr_setting; itr 1415 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c u8 itr_setting = ring_container->itr; itr 1426 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c timepassed_us = q_vector->itr >> 2; itr 1454 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c ring_container->itr = itr_setting; itr 1459 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c u32 new_itr = q_vector->itr; itr 1465 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c current_itr = max(q_vector->rx.itr, q_vector->tx.itr); itr 1482 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c if (new_itr != q_vector->itr) { itr 1484 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c new_itr = (10 * new_itr * q_vector->itr) / itr 1485 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c ((9 * new_itr) + q_vector->itr); itr 1488 drivers/net/ethernet/intel/ixgbevf/ixgbevf_main.c q_vector->itr = new_itr; itr 140 drivers/net/ethernet/qualcomm/emac/emac-sgmii-fsm9900.c const struct emac_reg_write *itr, size_t size) itr 144 drivers/net/ethernet/qualcomm/emac/emac-sgmii-fsm9900.c for (i = 0; i < size; ++itr, ++i) itr 145 drivers/net/ethernet/qualcomm/emac/emac-sgmii-fsm9900.c writel(itr->val, base + itr->offset); itr 117 drivers/net/ethernet/qualcomm/emac/emac-sgmii-qdf2400.c const struct emac_reg_write *itr, size_t size) itr 121 drivers/net/ethernet/qualcomm/emac/emac-sgmii-qdf2400.c for (i = 0; i < size; ++itr, ++i) itr 122 drivers/net/ethernet/qualcomm/emac/emac-sgmii-qdf2400.c writel(itr->val, base + itr->offset); itr 107 drivers/net/ethernet/qualcomm/emac/emac-sgmii-qdf2432.c const struct emac_reg_write *itr, size_t size) itr 111 drivers/net/ethernet/qualcomm/emac/emac-sgmii-qdf2432.c for (i = 0; i < size; ++itr, ++i) itr 112 drivers/net/ethernet/qualcomm/emac/emac-sgmii-qdf2432.c writel(itr->val, base + itr->offset); itr 1205 drivers/net/wimax/i2400m/control.c void *buf, *itr; itr 1231 drivers/net/wimax/i2400m/control.c itr = buf + sizeof(*cmd); itr 1235 drivers/net/wimax/i2400m/control.c memcpy(itr, tlv_hdr, tlv_size); itr 1236 drivers/net/wimax/i2400m/control.c itr += tlv_size; itr 1259 drivers/net/wimax/i2400m/fw.c const void *itr, *next, *top; itr 1262 drivers/net/wimax/i2400m/fw.c for (itr = bcf, top = itr + bcf_size; itr 1263 drivers/net/wimax/i2400m/fw.c itr < top; itr 1264 drivers/net/wimax/i2400m/fw.c headers++, itr = next) { itr 1267 drivers/net/wimax/i2400m/fw.c leftover = top - itr; itr 1268 drivers/net/wimax/i2400m/fw.c offset = itr - bcf; itr 1275 drivers/net/wimax/i2400m/fw.c bcf_hdr = itr; itr 1281 drivers/net/wimax/i2400m/fw.c next = itr + size; itr 1283 drivers/net/wimax/i2400m/fw.c next = itr + header_len; itr 1549 drivers/net/wimax/i2400m/fw.c int ret, itr; itr 1573 drivers/net/wimax/i2400m/fw.c for (itr = 0, ret = -ENOENT; ; itr++) { itr 1574 drivers/net/wimax/i2400m/fw.c fw_name = i2400m->bus_fw_names[itr]; itr 1579 drivers/net/wimax/i2400m/fw.c d_printf(1, dev, "trying firmware %s (%d)\n", fw_name, itr); itr 929 drivers/net/wimax/i2400m/rx.c unsigned itr; itr 932 drivers/net/wimax/i2400m/rx.c for (itr = 0; itr < I2400M_RO_CIN + 1; itr++) itr 933 drivers/net/wimax/i2400m/rx.c __skb_queue_purge(&i2400m->rx_roq[itr].queue); itr 1351 drivers/net/wimax/i2400m/rx.c unsigned itr; itr 1368 drivers/net/wimax/i2400m/rx.c for(itr = 0; itr < I2400M_RO_CIN + 1; itr++) { itr 1369 drivers/net/wimax/i2400m/rx.c __i2400m_roq_init(&i2400m->rx_roq[itr]); itr 1370 drivers/net/wimax/i2400m/rx.c i2400m->rx_roq[itr].log = &rd[itr]; itr 129 drivers/staging/uwb/drp-avail.c unsigned long get_val(u8 *array, size_t itr, size_t len) itr 132 drivers/staging/uwb/drp-avail.c size_t top = itr + len; itr 136 drivers/staging/uwb/drp-avail.c while (itr < top) { itr 194 drivers/staging/uwb/drp-avail.c size_t itr, len; itr 197 drivers/staging/uwb/drp-avail.c itr = 0; itr 198 drivers/staging/uwb/drp-avail.c while (itr < buffer_size) { itr 199 drivers/staging/uwb/drp-avail.c len = buffer_size - itr >= sizeof(val) ? itr 200 drivers/staging/uwb/drp-avail.c sizeof(val) : buffer_size - itr; itr 201 drivers/staging/uwb/drp-avail.c val = get_val(buffer, itr, len); itr 202 drivers/staging/uwb/drp-avail.c bmp_itr[itr / sizeof(val)] = val; itr 203 drivers/staging/uwb/drp-avail.c itr += sizeof(val); itr 246 drivers/staging/uwb/est.c unsigned itr; itr 256 drivers/staging/uwb/est.c for (itr = 0; itr < uwb_est_used; itr++) itr 257 drivers/staging/uwb/est.c if (uwb_est[itr].type_event_high < type itr 258 drivers/staging/uwb/est.c && uwb_est[itr].vendor < vendor itr 259 drivers/staging/uwb/est.c && uwb_est[itr].product < product) itr 263 drivers/staging/uwb/est.c if (itr < uwb_est_used) itr 264 drivers/staging/uwb/est.c memmove(&uwb_est[itr+1], &uwb_est[itr], uwb_est_used - itr); itr 265 drivers/staging/uwb/est.c uwb_est[itr].type_event_high = type << 8 | event_high; itr 266 drivers/staging/uwb/est.c uwb_est[itr].vendor = vendor; itr 267 drivers/staging/uwb/est.c uwb_est[itr].product = product; itr 268 drivers/staging/uwb/est.c uwb_est[itr].entry = entry; itr 269 drivers/staging/uwb/est.c uwb_est[itr].entries = entries; itr 295 drivers/staging/uwb/est.c unsigned itr; itr 304 drivers/staging/uwb/est.c for (itr = 0; itr < uwb_est_used; itr++) itr 305 drivers/staging/uwb/est.c if (!memcmp(&uwb_est[itr], &est_cmp, sizeof(est_cmp))) itr 311 drivers/staging/uwb/est.c if (itr < uwb_est_used - 1) /* Not last one? move ones above */ itr 312 drivers/staging/uwb/est.c memmove(&uwb_est[itr], &uwb_est[itr+1], uwb_est_used - itr - 1); itr 421 drivers/staging/uwb/est.c unsigned itr; itr 430 drivers/staging/uwb/est.c for (itr = 0; itr < uwb_est_used; itr++) { itr 431 drivers/staging/uwb/est.c if (uwb_est[itr].type_event_high != type_event_high) itr 433 drivers/staging/uwb/est.c size = uwb_est_get_size(rc, &uwb_est[itr], itr 741 drivers/staging/uwb/hwa-rc.c char *itr; itr 748 drivers/staging/uwb/hwa-rc.c itr = usb_dev->rawdescriptors[actconfig_idx]; itr 751 drivers/staging/uwb/hwa-rc.c hdr = (struct usb_descriptor_header *) itr; itr 755 drivers/staging/uwb/hwa-rc.c (itr - usb_dev->rawdescriptors[actconfig_idx]), itr 759 drivers/staging/uwb/hwa-rc.c itr += hdr->bLength; itr 101 drivers/staging/uwb/i1480/dfu/usb.c size_t buffer_size, itr = 0; itr 106 drivers/staging/uwb/i1480/dfu/usb.c memcpy(i1480->cmd_buf, buffer + itr, buffer_size); itr 114 drivers/staging/uwb/i1480/dfu/usb.c itr += result; itr 139 drivers/staging/uwb/i1480/dfu/usb.c size_t itr, read_size = i1480->buf_size; itr 149 drivers/staging/uwb/i1480/dfu/usb.c for (itr = 0; itr < size; itr += read_size) { itr 150 drivers/staging/uwb/i1480/dfu/usb.c size_t itr_addr = addr + itr; itr 151 drivers/staging/uwb/i1480/dfu/usb.c size_t itr_size = min(read_size, size - itr); itr 156 drivers/staging/uwb/i1480/dfu/usb.c i1480->cmd_buf + itr, itr_size, itr 467 drivers/staging/uwb/neh.c void *itr; itr 472 drivers/staging/uwb/neh.c itr = buf; itr 482 drivers/staging/uwb/neh.c rceb = itr; itr 517 drivers/staging/uwb/neh.c itr += real_size; itr 125 drivers/staging/wusbcore/cbaf.c void *itr, *top; itr 160 drivers/staging/wusbcore/cbaf.c itr = cbaf->buffer + sizeof(*assoc_info); itr 165 drivers/staging/wusbcore/cbaf.c while (itr < top) { itr 170 drivers/staging/wusbcore/cbaf.c assoc_request = itr; itr 172 drivers/staging/wusbcore/cbaf.c if (top - itr < sizeof(*assoc_request)) { itr 175 drivers/staging/wusbcore/cbaf.c top - itr, sizeof(*assoc_request)); itr 206 drivers/staging/wusbcore/cbaf.c itr += sizeof(*assoc_request); itr 694 drivers/staging/wusbcore/devconnect.c void *itr, *top; itr 697 drivers/staging/wusbcore/devconnect.c itr = (void *)bos + sizeof(*bos); itr 698 drivers/staging/wusbcore/devconnect.c top = itr + desc_size - sizeof(*bos); itr 699 drivers/staging/wusbcore/devconnect.c while (itr < top) { itr 700 drivers/staging/wusbcore/devconnect.c struct usb_dev_cap_header *cap_hdr = itr; itr 703 drivers/staging/wusbcore/devconnect.c if (top - itr < sizeof(*cap_hdr)) { itr 706 drivers/staging/wusbcore/devconnect.c (int)(itr - (void *)bos), top - itr); itr 714 drivers/staging/wusbcore/devconnect.c if (cap_size > top - itr) { itr 718 drivers/staging/wusbcore/devconnect.c (int)(itr - (void *)bos), itr 719 drivers/staging/wusbcore/devconnect.c cap_type, cap_size, top - itr); itr 731 drivers/staging/wusbcore/devconnect.c wusb_dev->wusb_cap_descr = itr; itr 736 drivers/staging/wusbcore/devconnect.c cap_size, (int)(itr - (void *)bos)); itr 738 drivers/staging/wusbcore/devconnect.c itr += cap_size; itr 556 drivers/staging/wusbcore/host/hwa-hc.c char *itr; itr 564 drivers/staging/wusbcore/host/hwa-hc.c itr = usb_dev->rawdescriptors[actconfig_idx]; itr 567 drivers/staging/wusbcore/host/hwa-hc.c hdr = (struct usb_descriptor_header *) itr; itr 571 drivers/staging/wusbcore/host/hwa-hc.c (itr - usb_dev->rawdescriptors[actconfig_idx]), itr 575 drivers/staging/wusbcore/host/hwa-hc.c itr += hdr->bLength; itr 630 drivers/staging/wusbcore/host/hwa-hc.c void *itr, *top; itr 638 drivers/staging/wusbcore/host/hwa-hc.c itr = usb_dev->rawdescriptors[index]; itr 640 drivers/staging/wusbcore/host/hwa-hc.c top = itr + itr_size; itr 663 drivers/staging/wusbcore/host/hwa-hc.c itr = (void *) secd + sizeof(*secd); itr 667 drivers/staging/wusbcore/host/hwa-hc.c while (itr < top) { itr 668 drivers/staging/wusbcore/host/hwa-hc.c etd = itr; itr 669 drivers/staging/wusbcore/host/hwa-hc.c if (top - itr < sizeof(*etd)) { itr 672 drivers/staging/wusbcore/host/hwa-hc.c top - itr, sizeof(*etd)); itr 682 drivers/staging/wusbcore/host/hwa-hc.c itr += etd->bLength; itr 84 drivers/staging/wusbcore/mmc.c unsigned handle, itr; itr 101 drivers/staging/wusbcore/mmc.c for (itr = 0; itr < wusbhc->mmcies_max - 1; itr++) { itr 102 drivers/staging/wusbcore/mmc.c if (wusbhc->mmcie[itr] == wuie) { itr 103 drivers/staging/wusbcore/mmc.c handle = itr; itr 106 drivers/staging/wusbcore/mmc.c if (wusbhc->mmcie[itr] == NULL) itr 107 drivers/staging/wusbcore/mmc.c handle = itr; itr 130 drivers/staging/wusbcore/mmc.c unsigned handle, itr; itr 133 drivers/staging/wusbcore/mmc.c for (itr = 0; itr < wusbhc->mmcies_max; itr++) { itr 134 drivers/staging/wusbcore/mmc.c if (wusbhc->mmcie[itr] == wuie) { itr 135 drivers/staging/wusbcore/mmc.c handle = itr; itr 145 drivers/staging/wusbcore/mmc.c wusbhc->mmcie[itr] = NULL; itr 411 drivers/staging/wusbcore/rh.c size_t port_size, itr; itr 416 drivers/staging/wusbcore/rh.c for (itr = 0; itr < wusbhc->ports_max; itr++) itr 417 drivers/staging/wusbcore/rh.c wusb_port_init(&wusbhc->port[itr]); itr 209 drivers/staging/wusbcore/security.c const void *itr, *top; itr 242 drivers/staging/wusbcore/security.c itr = &secd[1]; itr 244 drivers/staging/wusbcore/security.c while (itr < top) { itr 245 drivers/staging/wusbcore/security.c etd = itr; itr 246 drivers/staging/wusbcore/security.c if (top - itr < sizeof(*etd)) { itr 249 drivers/staging/wusbcore/security.c top - itr, sizeof(*etd)); itr 259 drivers/staging/wusbcore/security.c itr += etd->bLength; itr 80 drivers/staging/wusbcore/wa-nep.c void *itr; itr 97 drivers/staging/wusbcore/wa-nep.c itr = nw->data; itr 104 drivers/staging/wusbcore/wa-nep.c notif_hdr = itr; itr 107 drivers/staging/wusbcore/wa-nep.c itr += notif_hdr->bLength; itr 252 drivers/staging/wusbcore/wa-rpipe.c void *itr; itr 261 drivers/staging/wusbcore/wa-rpipe.c itr = ep->extra; itr 269 drivers/staging/wusbcore/wa-rpipe.c itr - (void *) ep->extra, itr_size); itr 272 drivers/staging/wusbcore/wa-rpipe.c hdr = itr; itr 274 drivers/staging/wusbcore/wa-rpipe.c epcd = itr; itr 282 drivers/staging/wusbcore/wa-rpipe.c itr - (void *) ep->extra, hdr->bDescriptorType, itr 286 drivers/staging/wusbcore/wa-rpipe.c itr += hdr->bLength; itr 423 include/linux/wimax/debug.h struct d_level *itr, *top; itr 426 include/linux/wimax/debug.h for (itr = d_level, top = itr + d_level_size; itr < top; itr++) { itr 428 include/linux/wimax/debug.h if (itr->name == NULL) { itr 430 include/linux/wimax/debug.h tag, itr, index); itr 433 include/linux/wimax/debug.h if (!strcmp(itr->name, submodule)) { itr 434 include/linux/wimax/debug.h itr->level = level; itr 1034 net/dcb/dcbnl.c struct dcb_app_type *itr; itr 1117 net/dcb/dcbnl.c list_for_each_entry(itr, &dcb_app_list, list) { itr 1118 net/dcb/dcbnl.c if (itr->ifindex == netdev->ifindex) { itr 1119 net/dcb/dcbnl.c err = nla_put(skb, DCB_ATTR_IEEE_APP, sizeof(itr->app), itr 1120 net/dcb/dcbnl.c &itr->app); itr 1230 net/dcb/dcbnl.c struct dcb_app_type *itr; itr 1276 net/dcb/dcbnl.c list_for_each_entry(itr, &dcb_app_list, list) { itr 1277 net/dcb/dcbnl.c if (itr->ifindex == netdev->ifindex) { itr 1284 net/dcb/dcbnl.c itr->app.selector); itr 1289 net/dcb/dcbnl.c itr->app.protocol); itr 1294 net/dcb/dcbnl.c itr->app.priority); itr 1792 net/dcb/dcbnl.c struct dcb_app_type *itr; itr 1794 net/dcb/dcbnl.c list_for_each_entry(itr, &dcb_app_list, list) { itr 1795 net/dcb/dcbnl.c if (itr->app.selector == app->selector && itr 1796 net/dcb/dcbnl.c itr->app.protocol == app->protocol && itr 1797 net/dcb/dcbnl.c itr->ifindex == ifindex && itr 1798 net/dcb/dcbnl.c ((prio == -1) || itr->app.priority == prio)) itr 1799 net/dcb/dcbnl.c return itr; itr 1829 net/dcb/dcbnl.c struct dcb_app_type *itr; itr 1833 net/dcb/dcbnl.c itr = dcb_app_lookup(app, dev->ifindex, -1); itr 1834 net/dcb/dcbnl.c if (itr) itr 1835 net/dcb/dcbnl.c prio = itr->app.priority; itr 1851 net/dcb/dcbnl.c struct dcb_app_type *itr; itr 1862 net/dcb/dcbnl.c itr = dcb_app_lookup(new, dev->ifindex, -1); itr 1863 net/dcb/dcbnl.c if (itr) { itr 1865 net/dcb/dcbnl.c itr->app.priority = new->priority; itr 1867 net/dcb/dcbnl.c list_del(&itr->list); itr 1868 net/dcb/dcbnl.c kfree(itr); itr 1892 net/dcb/dcbnl.c struct dcb_app_type *itr; itr 1896 net/dcb/dcbnl.c itr = dcb_app_lookup(app, dev->ifindex, -1); itr 1897 net/dcb/dcbnl.c if (itr) itr 1898 net/dcb/dcbnl.c prio |= 1 << itr->app.priority; itr 1946 net/dcb/dcbnl.c struct dcb_app_type *itr; itr 1957 net/dcb/dcbnl.c if ((itr = dcb_app_lookup(del, dev->ifindex, del->priority))) { itr 1958 net/dcb/dcbnl.c list_del(&itr->list); itr 1959 net/dcb/dcbnl.c kfree(itr); itr 1980 net/dcb/dcbnl.c struct dcb_app_type *itr; itr 1986 net/dcb/dcbnl.c list_for_each_entry(itr, &dcb_app_list, list) { itr 1987 net/dcb/dcbnl.c if (itr->ifindex == ifindex && itr 1988 net/dcb/dcbnl.c itr->app.selector == IEEE_8021QAZ_APP_SEL_DSCP && itr 1989 net/dcb/dcbnl.c itr->app.protocol < 64 && itr 1990 net/dcb/dcbnl.c itr->app.priority < IEEE_8021QAZ_MAX_TCS) { itr 1991 net/dcb/dcbnl.c prio = itr->app.priority; itr 1992 net/dcb/dcbnl.c p_map->map[prio] |= 1ULL << itr->app.protocol; itr 2010 net/dcb/dcbnl.c struct dcb_app_type *itr; itr 2015 net/dcb/dcbnl.c list_for_each_entry(itr, &dcb_app_list, list) { itr 2016 net/dcb/dcbnl.c if (itr->ifindex == ifindex && itr 2017 net/dcb/dcbnl.c itr->app.selector == IEEE_8021QAZ_APP_SEL_DSCP && itr 2018 net/dcb/dcbnl.c itr->app.protocol < 64 && itr 2019 net/dcb/dcbnl.c itr->app.priority < IEEE_8021QAZ_MAX_TCS) itr 2020 net/dcb/dcbnl.c p_map->map[itr->app.protocol] |= 1 << itr->app.priority; itr 2039 net/dcb/dcbnl.c struct dcb_app_type *itr; itr 2043 net/dcb/dcbnl.c list_for_each_entry(itr, &dcb_app_list, list) { itr 2044 net/dcb/dcbnl.c if (itr->ifindex == ifindex && itr 2045 net/dcb/dcbnl.c itr->app.selector == IEEE_8021QAZ_APP_SEL_ETHERTYPE && itr 2046 net/dcb/dcbnl.c itr->app.protocol == 0 && itr 2047 net/dcb/dcbnl.c itr->app.priority < IEEE_8021QAZ_MAX_TCS) itr 2048 net/dcb/dcbnl.c mask |= 1 << itr->app.priority; itr 37 tools/perf/arch/arm/util/cs-etm.c struct auxtrace_record itr; itr 59 tools/perf/arch/arm/util/cs-etm.c static bool cs_etm_is_etmv4(struct auxtrace_record *itr, int cpu); itr 61 tools/perf/arch/arm/util/cs-etm.c static int cs_etm_set_context_id(struct auxtrace_record *itr, itr 70 tools/perf/arch/arm/util/cs-etm.c ptr = container_of(itr, struct cs_etm_recording, itr); itr 73 tools/perf/arch/arm/util/cs-etm.c if (!cs_etm_is_etmv4(itr, cpu)) itr 110 tools/perf/arch/arm/util/cs-etm.c static int cs_etm_set_timestamp(struct auxtrace_record *itr, itr 119 tools/perf/arch/arm/util/cs-etm.c ptr = container_of(itr, struct cs_etm_recording, itr); itr 122 tools/perf/arch/arm/util/cs-etm.c if (!cs_etm_is_etmv4(itr, cpu)) itr 158 tools/perf/arch/arm/util/cs-etm.c static int cs_etm_set_option(struct auxtrace_record *itr, itr 172 tools/perf/arch/arm/util/cs-etm.c err = cs_etm_set_context_id(itr, evsel, i); itr 177 tools/perf/arch/arm/util/cs-etm.c err = cs_etm_set_timestamp(itr, evsel, i); itr 192 tools/perf/arch/arm/util/cs-etm.c static int cs_etm_parse_snapshot_options(struct auxtrace_record *itr, itr 197 tools/perf/arch/arm/util/cs-etm.c container_of(itr, struct cs_etm_recording, itr); itr 251 tools/perf/arch/arm/util/cs-etm.c static int cs_etm_recording_options(struct auxtrace_record *itr, itr 257 tools/perf/arch/arm/util/cs-etm.c container_of(itr, struct cs_etm_recording, itr); itr 406 tools/perf/arch/arm/util/cs-etm.c err = cs_etm_set_option(itr, cs_etm_evsel, itr 435 tools/perf/arch/arm/util/cs-etm.c static u64 cs_etm_get_config(struct auxtrace_record *itr) itr 439 tools/perf/arch/arm/util/cs-etm.c container_of(itr, struct cs_etm_recording, itr); itr 466 tools/perf/arch/arm/util/cs-etm.c static u64 cs_etmv4_get_config(struct auxtrace_record *itr) itr 477 tools/perf/arch/arm/util/cs-etm.c config_opts = cs_etm_get_config(itr); itr 491 tools/perf/arch/arm/util/cs-etm.c cs_etm_info_priv_size(struct auxtrace_record *itr __maybe_unused, itr 506 tools/perf/arch/arm/util/cs-etm.c if (cs_etm_is_etmv4(itr, i)) itr 517 tools/perf/arch/arm/util/cs-etm.c if (cs_etm_is_etmv4(itr, i)) itr 531 tools/perf/arch/arm/util/cs-etm.c static bool cs_etm_is_etmv4(struct auxtrace_record *itr, int cpu) itr 538 tools/perf/arch/arm/util/cs-etm.c container_of(itr, struct cs_etm_recording, itr); itr 570 tools/perf/arch/arm/util/cs-etm.c struct auxtrace_record *itr, itr 576 tools/perf/arch/arm/util/cs-etm.c container_of(itr, struct cs_etm_recording, itr); itr 580 tools/perf/arch/arm/util/cs-etm.c if (cs_etm_is_etmv4(itr, cpu)) { itr 584 tools/perf/arch/arm/util/cs-etm.c cs_etmv4_get_config(itr); itr 611 tools/perf/arch/arm/util/cs-etm.c info->priv[*offset + CS_ETM_ETMCR] = cs_etm_get_config(itr); itr 634 tools/perf/arch/arm/util/cs-etm.c static int cs_etm_info_fill(struct auxtrace_record *itr, itr 646 tools/perf/arch/arm/util/cs-etm.c container_of(itr, struct cs_etm_recording, itr); itr 649 tools/perf/arch/arm/util/cs-etm.c if (priv_size != cs_etm_info_priv_size(itr, session->evlist)) itr 684 tools/perf/arch/arm/util/cs-etm.c cs_etm_get_metadata(i, &offset, itr, info); itr 767 tools/perf/arch/arm/util/cs-etm.c static int cs_etm_find_snapshot(struct auxtrace_record *itr, itr 775 tools/perf/arch/arm/util/cs-etm.c container_of(itr, struct cs_etm_recording, itr); itr 820 tools/perf/arch/arm/util/cs-etm.c static int cs_etm_snapshot_start(struct auxtrace_record *itr) itr 823 tools/perf/arch/arm/util/cs-etm.c container_of(itr, struct cs_etm_recording, itr); itr 833 tools/perf/arch/arm/util/cs-etm.c static int cs_etm_snapshot_finish(struct auxtrace_record *itr) itr 836 tools/perf/arch/arm/util/cs-etm.c container_of(itr, struct cs_etm_recording, itr); itr 846 tools/perf/arch/arm/util/cs-etm.c static u64 cs_etm_reference(struct auxtrace_record *itr __maybe_unused) itr 852 tools/perf/arch/arm/util/cs-etm.c static void cs_etm_recording_free(struct auxtrace_record *itr) itr 855 tools/perf/arch/arm/util/cs-etm.c container_of(itr, struct cs_etm_recording, itr); itr 861 tools/perf/arch/arm/util/cs-etm.c static int cs_etm_read_finish(struct auxtrace_record *itr, int idx) itr 864 tools/perf/arch/arm/util/cs-etm.c container_of(itr, struct cs_etm_recording, itr); itr 898 tools/perf/arch/arm/util/cs-etm.c ptr->itr.parse_snapshot_options = cs_etm_parse_snapshot_options; itr 899 tools/perf/arch/arm/util/cs-etm.c ptr->itr.recording_options = cs_etm_recording_options; itr 900 tools/perf/arch/arm/util/cs-etm.c ptr->itr.info_priv_size = cs_etm_info_priv_size; itr 901 tools/perf/arch/arm/util/cs-etm.c ptr->itr.info_fill = cs_etm_info_fill; itr 902 tools/perf/arch/arm/util/cs-etm.c ptr->itr.find_snapshot = cs_etm_find_snapshot; itr 903 tools/perf/arch/arm/util/cs-etm.c ptr->itr.snapshot_start = cs_etm_snapshot_start; itr 904 tools/perf/arch/arm/util/cs-etm.c ptr->itr.snapshot_finish = cs_etm_snapshot_finish; itr 905 tools/perf/arch/arm/util/cs-etm.c ptr->itr.reference = cs_etm_reference; itr 906 tools/perf/arch/arm/util/cs-etm.c ptr->itr.free = cs_etm_recording_free; itr 907 tools/perf/arch/arm/util/cs-etm.c ptr->itr.read_finish = cs_etm_read_finish; itr 910 tools/perf/arch/arm/util/cs-etm.c return &ptr->itr; itr 30 tools/perf/arch/arm64/util/arm-spe.c struct auxtrace_record itr; itr 36 tools/perf/arch/arm64/util/arm-spe.c arm_spe_info_priv_size(struct auxtrace_record *itr __maybe_unused, itr 42 tools/perf/arch/arm64/util/arm-spe.c static int arm_spe_info_fill(struct auxtrace_record *itr, itr 48 tools/perf/arch/arm64/util/arm-spe.c container_of(itr, struct arm_spe_recording, itr); itr 63 tools/perf/arch/arm64/util/arm-spe.c static int arm_spe_recording_options(struct auxtrace_record *itr, itr 68 tools/perf/arch/arm64/util/arm-spe.c container_of(itr, struct arm_spe_recording, itr); itr 144 tools/perf/arch/arm64/util/arm-spe.c static u64 arm_spe_reference(struct auxtrace_record *itr __maybe_unused) itr 153 tools/perf/arch/arm64/util/arm-spe.c static void arm_spe_recording_free(struct auxtrace_record *itr) itr 156 tools/perf/arch/arm64/util/arm-spe.c container_of(itr, struct arm_spe_recording, itr); itr 161 tools/perf/arch/arm64/util/arm-spe.c static int arm_spe_read_finish(struct auxtrace_record *itr, int idx) itr 164 tools/perf/arch/arm64/util/arm-spe.c container_of(itr, struct arm_spe_recording, itr); itr 195 tools/perf/arch/arm64/util/arm-spe.c sper->itr.recording_options = arm_spe_recording_options; itr 196 tools/perf/arch/arm64/util/arm-spe.c sper->itr.info_priv_size = arm_spe_info_priv_size; itr 197 tools/perf/arch/arm64/util/arm-spe.c sper->itr.info_fill = arm_spe_info_fill; itr 198 tools/perf/arch/arm64/util/arm-spe.c sper->itr.free = arm_spe_recording_free; itr 199 tools/perf/arch/arm64/util/arm-spe.c sper->itr.reference = arm_spe_reference; itr 200 tools/perf/arch/arm64/util/arm-spe.c sper->itr.read_finish = arm_spe_read_finish; itr 201 tools/perf/arch/arm64/util/arm-spe.c sper->itr.alignment = 0; itr 204 tools/perf/arch/arm64/util/arm-spe.c return &sper->itr; itr 19 tools/perf/arch/s390/util/auxtrace.c static void cpumsf_free(struct auxtrace_record *itr) itr 21 tools/perf/arch/s390/util/auxtrace.c free(itr); itr 24 tools/perf/arch/s390/util/auxtrace.c static size_t cpumsf_info_priv_size(struct auxtrace_record *itr __maybe_unused, itr 31 tools/perf/arch/s390/util/auxtrace.c cpumsf_info_fill(struct auxtrace_record *itr __maybe_unused, itr 41 tools/perf/arch/s390/util/auxtrace.c cpumsf_reference(struct auxtrace_record *itr __maybe_unused) itr 76 tools/perf/arch/s390/util/auxtrace.c cpumsf_parse_snapshot_options(struct auxtrace_record *itr __maybe_unused, itr 40 tools/perf/arch/x86/util/intel-bts.c struct auxtrace_record itr; itr 56 tools/perf/arch/x86/util/intel-bts.c intel_bts_info_priv_size(struct auxtrace_record *itr __maybe_unused, itr 62 tools/perf/arch/x86/util/intel-bts.c static int intel_bts_info_fill(struct auxtrace_record *itr, itr 68 tools/perf/arch/x86/util/intel-bts.c container_of(itr, struct intel_bts_recording, itr); itr 105 tools/perf/arch/x86/util/intel-bts.c static int intel_bts_recording_options(struct auxtrace_record *itr, itr 110 tools/perf/arch/x86/util/intel-bts.c container_of(itr, struct intel_bts_recording, itr); itr 245 tools/perf/arch/x86/util/intel-bts.c static int intel_bts_parse_snapshot_options(struct auxtrace_record *itr, itr 250 tools/perf/arch/x86/util/intel-bts.c container_of(itr, struct intel_bts_recording, itr); itr 268 tools/perf/arch/x86/util/intel-bts.c static u64 intel_bts_reference(struct auxtrace_record *itr __maybe_unused) itr 307 tools/perf/arch/x86/util/intel-bts.c static void intel_bts_recording_free(struct auxtrace_record *itr) itr 310 tools/perf/arch/x86/util/intel-bts.c container_of(itr, struct intel_bts_recording, itr); itr 316 tools/perf/arch/x86/util/intel-bts.c static int intel_bts_snapshot_start(struct auxtrace_record *itr) itr 319 tools/perf/arch/x86/util/intel-bts.c container_of(itr, struct intel_bts_recording, itr); itr 329 tools/perf/arch/x86/util/intel-bts.c static int intel_bts_snapshot_finish(struct auxtrace_record *itr) itr 332 tools/perf/arch/x86/util/intel-bts.c container_of(itr, struct intel_bts_recording, itr); itr 359 tools/perf/arch/x86/util/intel-bts.c static int intel_bts_find_snapshot(struct auxtrace_record *itr, int idx, itr 364 tools/perf/arch/x86/util/intel-bts.c container_of(itr, struct intel_bts_recording, itr); itr 411 tools/perf/arch/x86/util/intel-bts.c static int intel_bts_read_finish(struct auxtrace_record *itr, int idx) itr 414 tools/perf/arch/x86/util/intel-bts.c container_of(itr, struct intel_bts_recording, itr); itr 448 tools/perf/arch/x86/util/intel-bts.c btsr->itr.recording_options = intel_bts_recording_options; itr 449 tools/perf/arch/x86/util/intel-bts.c btsr->itr.info_priv_size = intel_bts_info_priv_size; itr 450 tools/perf/arch/x86/util/intel-bts.c btsr->itr.info_fill = intel_bts_info_fill; itr 451 tools/perf/arch/x86/util/intel-bts.c btsr->itr.free = intel_bts_recording_free; itr 452 tools/perf/arch/x86/util/intel-bts.c btsr->itr.snapshot_start = intel_bts_snapshot_start; itr 453 tools/perf/arch/x86/util/intel-bts.c btsr->itr.snapshot_finish = intel_bts_snapshot_finish; itr 454 tools/perf/arch/x86/util/intel-bts.c btsr->itr.find_snapshot = intel_bts_find_snapshot; itr 455 tools/perf/arch/x86/util/intel-bts.c btsr->itr.parse_snapshot_options = intel_bts_parse_snapshot_options; itr 456 tools/perf/arch/x86/util/intel-bts.c btsr->itr.reference = intel_bts_reference; itr 457 tools/perf/arch/x86/util/intel-bts.c btsr->itr.read_finish = intel_bts_read_finish; itr 458 tools/perf/arch/x86/util/intel-bts.c btsr->itr.alignment = sizeof(struct branch); itr 459 tools/perf/arch/x86/util/intel-bts.c return &btsr->itr; itr 47 tools/perf/arch/x86/util/intel-pt.c struct auxtrace_record itr; itr 235 tools/perf/arch/x86/util/intel-pt.c static int intel_pt_parse_snapshot_options(struct auxtrace_record *itr, itr 240 tools/perf/arch/x86/util/intel-pt.c container_of(itr, struct intel_pt_recording, itr); itr 295 tools/perf/arch/x86/util/intel-pt.c intel_pt_info_priv_size(struct auxtrace_record *itr, struct evlist *evlist) itr 298 tools/perf/arch/x86/util/intel-pt.c container_of(itr, struct intel_pt_recording, itr); itr 316 tools/perf/arch/x86/util/intel-pt.c static int intel_pt_info_fill(struct auxtrace_record *itr, itr 322 tools/perf/arch/x86/util/intel-pt.c container_of(itr, struct intel_pt_recording, itr); itr 574 tools/perf/arch/x86/util/intel-pt.c static int intel_pt_recording_options(struct auxtrace_record *itr, itr 579 tools/perf/arch/x86/util/intel-pt.c container_of(itr, struct intel_pt_recording, itr); itr 809 tools/perf/arch/x86/util/intel-pt.c static int intel_pt_snapshot_start(struct auxtrace_record *itr) itr 812 tools/perf/arch/x86/util/intel-pt.c container_of(itr, struct intel_pt_recording, itr); itr 822 tools/perf/arch/x86/util/intel-pt.c static int intel_pt_snapshot_finish(struct auxtrace_record *itr) itr 825 tools/perf/arch/x86/util/intel-pt.c container_of(itr, struct intel_pt_recording, itr); itr 868 tools/perf/arch/x86/util/intel-pt.c static void intel_pt_recording_free(struct auxtrace_record *itr) itr 871 tools/perf/arch/x86/util/intel-pt.c container_of(itr, struct intel_pt_recording, itr); itr 1025 tools/perf/arch/x86/util/intel-pt.c static int intel_pt_find_snapshot(struct auxtrace_record *itr, int idx, itr 1030 tools/perf/arch/x86/util/intel-pt.c container_of(itr, struct intel_pt_recording, itr); itr 1090 tools/perf/arch/x86/util/intel-pt.c static u64 intel_pt_reference(struct auxtrace_record *itr __maybe_unused) itr 1095 tools/perf/arch/x86/util/intel-pt.c static int intel_pt_read_finish(struct auxtrace_record *itr, int idx) itr 1098 tools/perf/arch/x86/util/intel-pt.c container_of(itr, struct intel_pt_recording, itr); itr 1132 tools/perf/arch/x86/util/intel-pt.c ptr->itr.recording_options = intel_pt_recording_options; itr 1133 tools/perf/arch/x86/util/intel-pt.c ptr->itr.info_priv_size = intel_pt_info_priv_size; itr 1134 tools/perf/arch/x86/util/intel-pt.c ptr->itr.info_fill = intel_pt_info_fill; itr 1135 tools/perf/arch/x86/util/intel-pt.c ptr->itr.free = intel_pt_recording_free; itr 1136 tools/perf/arch/x86/util/intel-pt.c ptr->itr.snapshot_start = intel_pt_snapshot_start; itr 1137 tools/perf/arch/x86/util/intel-pt.c ptr->itr.snapshot_finish = intel_pt_snapshot_finish; itr 1138 tools/perf/arch/x86/util/intel-pt.c ptr->itr.find_snapshot = intel_pt_find_snapshot; itr 1139 tools/perf/arch/x86/util/intel-pt.c ptr->itr.parse_snapshot_options = intel_pt_parse_snapshot_options; itr 1140 tools/perf/arch/x86/util/intel-pt.c ptr->itr.reference = intel_pt_reference; itr 1141 tools/perf/arch/x86/util/intel-pt.c ptr->itr.read_finish = intel_pt_read_finish; itr 1142 tools/perf/arch/x86/util/intel-pt.c return &ptr->itr; itr 80 tools/perf/builtin-record.c struct auxtrace_record *itr; itr 573 tools/perf/builtin-record.c ret = auxtrace_mmap__read(map, rec->itr, &rec->tool, itr 589 tools/perf/builtin-record.c ret = auxtrace_mmap__read_snapshot(map, rec->itr, &rec->tool, itr 627 tools/perf/builtin-record.c if (auxtrace_record__snapshot_finish(rec->itr, on_exit)) itr 640 tools/perf/builtin-record.c auxtrace_record__snapshot_start(rec->itr)) itr 654 tools/perf/builtin-record.c if (!rec->itr) { itr 655 tools/perf/builtin-record.c rec->itr = auxtrace_record__init(rec->evlist, &err); itr 660 tools/perf/builtin-record.c err = auxtrace_parse_snapshot_options(rec->itr, &rec->opts, itr 684 tools/perf/builtin-record.c int auxtrace_record__snapshot_start(struct auxtrace_record *itr __maybe_unused) itr 1276 tools/perf/builtin-record.c err = perf_event__synthesize_auxtrace_info(rec->itr, tool, itr 2446 tools/perf/builtin-record.c err = auxtrace_record__options(rec->itr, rec->evlist, &rec->opts); itr 2478 tools/perf/builtin-record.c auxtrace_record__free(rec->itr); itr 2489 tools/perf/builtin-record.c if (auxtrace_record__snapshot_start(record.itr)) itr 2189 tools/perf/builtin-sched.c struct idle_thread_runtime *itr; itr 2193 tools/perf/builtin-sched.c itr = zalloc(sizeof(*itr)); itr 2194 tools/perf/builtin-sched.c if (itr == NULL) itr 2197 tools/perf/builtin-sched.c init_stats(&itr->tr.run_stats); itr 2198 tools/perf/builtin-sched.c callchain_init(&itr->callchain); itr 2199 tools/perf/builtin-sched.c callchain_cursor_reset(&itr->cursor); itr 2200 tools/perf/builtin-sched.c thread__set_priv(thread, itr); itr 2283 tools/perf/builtin-sched.c struct idle_thread_runtime *itr, itr 2289 tools/perf/builtin-sched.c callchain_cursor__copy(&itr->cursor, &callchain_cursor); itr 2316 tools/perf/builtin-sched.c struct idle_thread_runtime *itr; itr 2324 tools/perf/builtin-sched.c itr = thread__priv(idle); itr 2325 tools/perf/builtin-sched.c if (itr == NULL) itr 2328 tools/perf/builtin-sched.c itr->last_thread = thread; itr 2332 tools/perf/builtin-sched.c save_idle_callchain(sched, itr, sample); itr 2577 tools/perf/builtin-sched.c struct idle_thread_runtime *itr = (void *)tr; itr 2582 tools/perf/builtin-sched.c if (itr->last_thread == NULL) itr 2586 tools/perf/builtin-sched.c last_tr = thread__get_runtime(itr->last_thread); itr 2602 tools/perf/builtin-sched.c if (itr->cursor.nr) itr 2603 tools/perf/builtin-sched.c callchain_append(&itr->callchain, &itr->cursor, t - tprev); itr 2605 tools/perf/builtin-sched.c itr->last_thread = NULL; itr 2869 tools/perf/builtin-sched.c struct idle_thread_runtime *itr; itr 2875 tools/perf/builtin-sched.c itr = thread__priv(t); itr 2876 tools/perf/builtin-sched.c if (itr == NULL) itr 2879 tools/perf/builtin-sched.c callchain_param.sort(&itr->sorted_root.rb_root, &itr->callchain, itr 2883 tools/perf/builtin-sched.c print_sched_time(itr->tr.total_run_time, 6); itr 2885 tools/perf/builtin-sched.c timehist_print_idlehist_callchain(&itr->sorted_root); itr 508 tools/perf/util/auxtrace.c size_t auxtrace_record__info_priv_size(struct auxtrace_record *itr, itr 511 tools/perf/util/auxtrace.c if (itr) itr 512 tools/perf/util/auxtrace.c return itr->info_priv_size(itr, evlist); itr 522 tools/perf/util/auxtrace.c int auxtrace_record__info_fill(struct auxtrace_record *itr, itr 527 tools/perf/util/auxtrace.c if (itr) itr 528 tools/perf/util/auxtrace.c return itr->info_fill(itr, session, auxtrace_info, priv_size); itr 532 tools/perf/util/auxtrace.c void auxtrace_record__free(struct auxtrace_record *itr) itr 534 tools/perf/util/auxtrace.c if (itr) itr 535 tools/perf/util/auxtrace.c itr->free(itr); itr 538 tools/perf/util/auxtrace.c int auxtrace_record__snapshot_start(struct auxtrace_record *itr) itr 540 tools/perf/util/auxtrace.c if (itr && itr->snapshot_start) itr 541 tools/perf/util/auxtrace.c return itr->snapshot_start(itr); itr 545 tools/perf/util/auxtrace.c int auxtrace_record__snapshot_finish(struct auxtrace_record *itr, bool on_exit) itr 547 tools/perf/util/auxtrace.c if (!on_exit && itr && itr->snapshot_finish) itr 548 tools/perf/util/auxtrace.c return itr->snapshot_finish(itr); itr 552 tools/perf/util/auxtrace.c int auxtrace_record__find_snapshot(struct auxtrace_record *itr, int idx, itr 556 tools/perf/util/auxtrace.c if (itr && itr->find_snapshot) itr 557 tools/perf/util/auxtrace.c return itr->find_snapshot(itr, idx, mm, data, head, old); itr 561 tools/perf/util/auxtrace.c int auxtrace_record__options(struct auxtrace_record *itr, itr 565 tools/perf/util/auxtrace.c if (itr) itr 566 tools/perf/util/auxtrace.c return itr->recording_options(itr, evlist, opts); itr 570 tools/perf/util/auxtrace.c u64 auxtrace_record__reference(struct auxtrace_record *itr) itr 572 tools/perf/util/auxtrace.c if (itr) itr 573 tools/perf/util/auxtrace.c return itr->reference(itr); itr 577 tools/perf/util/auxtrace.c int auxtrace_parse_snapshot_options(struct auxtrace_record *itr, itr 593 tools/perf/util/auxtrace.c if (itr) itr 594 tools/perf/util/auxtrace.c return itr->parse_snapshot_options(itr, opts, str); itr 888 tools/perf/util/auxtrace.c int perf_event__synthesize_auxtrace_info(struct auxtrace_record *itr, itr 898 tools/perf/util/auxtrace.c priv_size = auxtrace_record__info_priv_size(itr, session->evlist); itr 906 tools/perf/util/auxtrace.c err = auxtrace_record__info_fill(itr, session, &ev->auxtrace_info, itr 1232 tools/perf/util/auxtrace.c struct auxtrace_record *itr, itr 1245 tools/perf/util/auxtrace.c if (auxtrace_record__find_snapshot(itr, mm->idx, mm, data, itr 1274 tools/perf/util/auxtrace.c ref = auxtrace_record__reference(itr); itr 1301 tools/perf/util/auxtrace.c if (itr->alignment) { itr 1302 tools/perf/util/auxtrace.c unsigned int unwanted = len1 % itr->alignment; itr 1330 tools/perf/util/auxtrace.c if (itr->read_finish) { itr 1333 tools/perf/util/auxtrace.c err = itr->read_finish(itr, mm->idx); itr 1342 tools/perf/util/auxtrace.c int auxtrace_mmap__read(struct mmap *map, struct auxtrace_record *itr, itr 1345 tools/perf/util/auxtrace.c return __auxtrace_mmap__read(map, itr, tool, fn, false, 0); itr 1349 tools/perf/util/auxtrace.c struct auxtrace_record *itr, itr 1353 tools/perf/util/auxtrace.c return __auxtrace_mmap__read(map, itr, tool, fn, true, snapshot_size); itr 318 tools/perf/util/auxtrace.h int (*recording_options)(struct auxtrace_record *itr, itr 321 tools/perf/util/auxtrace.h size_t (*info_priv_size)(struct auxtrace_record *itr, itr 323 tools/perf/util/auxtrace.h int (*info_fill)(struct auxtrace_record *itr, itr 327 tools/perf/util/auxtrace.h void (*free)(struct auxtrace_record *itr); itr 328 tools/perf/util/auxtrace.h int (*snapshot_start)(struct auxtrace_record *itr); itr 329 tools/perf/util/auxtrace.h int (*snapshot_finish)(struct auxtrace_record *itr); itr 330 tools/perf/util/auxtrace.h int (*find_snapshot)(struct auxtrace_record *itr, int idx, itr 333 tools/perf/util/auxtrace.h int (*parse_snapshot_options)(struct auxtrace_record *itr, itr 336 tools/perf/util/auxtrace.h u64 (*reference)(struct auxtrace_record *itr); itr 337 tools/perf/util/auxtrace.h int (*read_finish)(struct auxtrace_record *itr, int idx); itr 452 tools/perf/util/auxtrace.h int auxtrace_mmap__read(struct mmap *map, struct auxtrace_record *itr, itr 456 tools/perf/util/auxtrace.h struct auxtrace_record *itr, itr 497 tools/perf/util/auxtrace.h int auxtrace_parse_snapshot_options(struct auxtrace_record *itr, itr 500 tools/perf/util/auxtrace.h int auxtrace_record__options(struct auxtrace_record *itr, itr 503 tools/perf/util/auxtrace.h size_t auxtrace_record__info_priv_size(struct auxtrace_record *itr, itr 505 tools/perf/util/auxtrace.h int auxtrace_record__info_fill(struct auxtrace_record *itr, itr 509 tools/perf/util/auxtrace.h void auxtrace_record__free(struct auxtrace_record *itr); itr 510 tools/perf/util/auxtrace.h int auxtrace_record__snapshot_start(struct auxtrace_record *itr); itr 511 tools/perf/util/auxtrace.h int auxtrace_record__snapshot_finish(struct auxtrace_record *itr, bool on_exit); itr 512 tools/perf/util/auxtrace.h int auxtrace_record__find_snapshot(struct auxtrace_record *itr, int idx, itr 515 tools/perf/util/auxtrace.h u64 auxtrace_record__reference(struct auxtrace_record *itr); itr 600 tools/perf/util/auxtrace.h void auxtrace_record__free(struct auxtrace_record *itr __maybe_unused) itr 605 tools/perf/util/auxtrace.h int auxtrace_record__options(struct auxtrace_record *itr __maybe_unused, itr 640 tools/perf/util/auxtrace.h int auxtrace_parse_snapshot_options(struct auxtrace_record *itr __maybe_unused, itr 73 tools/perf/util/synthetic-events.h int perf_event__synthesize_auxtrace_info(struct auxtrace_record *itr, struct perf_tool *tool, itr 81 tools/perf/util/synthetic-events.h perf_event__synthesize_auxtrace_info(struct auxtrace_record *itr __maybe_unused, itr 34 tools/testing/selftests/gpio/gpio-mockup-chardev.c struct libmnt_iter *itr = NULL; itr 42 tools/testing/selftests/gpio/gpio-mockup-chardev.c itr = mnt_new_iter(MNT_ITER_FORWARD); itr 43 tools/testing/selftests/gpio/gpio-mockup-chardev.c if (!itr) itr 49 tools/testing/selftests/gpio/gpio-mockup-chardev.c while (mnt_table_next_fs(tb, itr, &fs) == 0) { itr 63 tools/testing/selftests/gpio/gpio-mockup-chardev.c mnt_free_iter(itr);