mw_count 161 drivers/infiniband/hw/bnxt_re/bnxt_re.h atomic_t mw_count; mw_count 130 drivers/infiniband/hw/bnxt_re/hw_counters.c stats->value[BNXT_RE_ACTIVE_MW] = atomic_read(&rdev->mw_count); mw_count 3445 drivers/infiniband/hw/bnxt_re/ib_verbs.c atomic_inc(&rdev->mw_count); mw_count 3466 drivers/infiniband/hw/bnxt_re/ib_verbs.c atomic_dec(&rdev->mw_count); mw_count 744 drivers/infiniband/hw/bnxt_re/main.c atomic_set(&rdev->mw_count, 0); mw_count 78 drivers/ntb/hw/amd/ntb_hw_amd.c if (idx < 0 || idx > ndev->mw_count) mw_count 89 drivers/ntb/hw/amd/ntb_hw_amd.c return ntb_ndev(ntb)->mw_count; mw_count 297 drivers/ntb/hw/amd/ntb_hw_amd.c return ntb_ndev(ntb)->mw_count; mw_count 465 drivers/ntb/hw/amd/ntb_hw_amd.c .mw_count = amd_ntb_mw_count, mw_count 757 drivers/ntb/hw/amd/ntb_hw_amd.c "Memory Window Count -\t%u\n", ndev->mw_count); mw_count 912 drivers/ntb/hw/amd/ntb_hw_amd.c ndev->mw_count = ndev->dev_data->mw_count; mw_count 1130 drivers/ntb/hw/amd/ntb_hw_amd.c .mw_count = 3, mw_count 1134 drivers/ntb/hw/amd/ntb_hw_amd.c .mw_count = 2, mw_count 172 drivers/ntb/hw/amd/ntb_hw_amd.h const unsigned char mw_count; mw_count 192 drivers/ntb/hw/amd/ntb_hw_amd.h unsigned char mw_count; mw_count 2249 drivers/ntb/hw/idt/ntb_hw_idt.c .mw_count = idt_ntb_mw_count, mw_count 178 drivers/ntb/hw/intel/ntb_hw_gen1.c if (idx < 0 || idx >= ndev->mw_count) mw_count 551 drivers/ntb/hw/intel/ntb_hw_gen1.c "Memory Window Count -\t%u\n", ndev->mw_count); mw_count 798 drivers/ntb/hw/intel/ntb_hw_gen1.c return ntb_ndev(ntb)->mw_count; mw_count 1020 drivers/ntb/hw/intel/ntb_hw_gen1.c return ntb_ndev(ntb)->mw_count; mw_count 1323 drivers/ntb/hw/intel/ntb_hw_gen1.c --ndev->mw_count; mw_count 1541 drivers/ntb/hw/intel/ntb_hw_gen1.c ndev->mw_count = HSX_SPLIT_BAR_MW_COUNT; mw_count 1543 drivers/ntb/hw/intel/ntb_hw_gen1.c ndev->mw_count = XEON_MW_COUNT; mw_count 1590 drivers/ntb/hw/intel/ntb_hw_gen1.c ndev->b2b_idx = b2b_mw_idx + ndev->mw_count; mw_count 1594 drivers/ntb/hw/intel/ntb_hw_gen1.c if (ndev->b2b_idx >= ndev->mw_count) { mw_count 1597 drivers/ntb/hw/intel/ntb_hw_gen1.c b2b_mw_idx, ndev->mw_count); mw_count 1838 drivers/ntb/hw/intel/ntb_hw_gen1.c ndev->mw_count = 0; mw_count 2005 drivers/ntb/hw/intel/ntb_hw_gen1.c .mw_count = intel_ntb_mw_count, mw_count 178 drivers/ntb/hw/intel/ntb_hw_gen3.c ndev->mw_count = XEON_MW_COUNT; mw_count 303 drivers/ntb/hw/intel/ntb_hw_gen3.c "Memory Window Count -\t%u\n", ndev->mw_count); mw_count 603 drivers/ntb/hw/intel/ntb_hw_gen3.c .mw_count = intel_ntb_mw_count, mw_count 152 drivers/ntb/hw/intel/ntb_hw_intel.h unsigned char mw_count; mw_count 812 drivers/ntb/hw/mscc/ntb_hw_switchtec.c .mw_count = switchtec_ntb_mw_count, mw_count 999 drivers/ntb/hw/mscc/ntb_hw_switchtec.c u64 *mw_addrs, int mw_count) mw_count 1025 drivers/ntb/hw/mscc/ntb_hw_switchtec.c sndev->nr_direct_mw = min_t(int, sndev->nr_direct_mw, mw_count); mw_count 230 drivers/ntb/ntb_transport.c unsigned int mw_count; mw_count 272 drivers/ntb/ntb_transport.c #define QP_TO_MW(nt, qp) ((qp) % nt->mw_count) mw_count 623 drivers/ntb/ntb_transport.c unsigned int mw_num, mw_count, qp_count; mw_count 627 drivers/ntb/ntb_transport.c mw_count = nt->mw_count; mw_count 636 drivers/ntb/ntb_transport.c if (mw_num < qp_count % mw_count) mw_count 637 drivers/ntb/ntb_transport.c num_qps_mw = qp_count / mw_count + 1; mw_count 639 drivers/ntb/ntb_transport.c num_qps_mw = qp_count / mw_count; mw_count 642 drivers/ntb/ntb_transport.c qp->rx_buff = mw->virt_addr + rx_size * (qp_num / mw_count); mw_count 990 drivers/ntb/ntb_transport.c for (i = 0; i < nt->mw_count; i++) mw_count 1045 drivers/ntb/ntb_transport.c for (i = 0; i < nt->mw_count; i++) { mw_count 1058 drivers/ntb/ntb_transport.c ntb_peer_spad_write(ndev, PIDX, NUM_MWS, nt->mw_count); mw_count 1077 drivers/ntb/ntb_transport.c if (val != nt->mw_count) mw_count 1080 drivers/ntb/ntb_transport.c for (i = 0; i < nt->mw_count; i++) { mw_count 1111 drivers/ntb/ntb_transport.c for (i = 0; i < nt->mw_count; i++) mw_count 1165 drivers/ntb/ntb_transport.c unsigned int mw_num, mw_count, qp_count; mw_count 1168 drivers/ntb/ntb_transport.c mw_count = nt->mw_count; mw_count 1181 drivers/ntb/ntb_transport.c if (mw_num < qp_count % mw_count) mw_count 1182 drivers/ntb/ntb_transport.c num_qps_mw = qp_count / mw_count + 1; mw_count 1184 drivers/ntb/ntb_transport.c num_qps_mw = qp_count / mw_count; mw_count 1193 drivers/ntb/ntb_transport.c qp_offset = tx_size * (qp_num / mw_count); mw_count 1247 drivers/ntb/ntb_transport.c unsigned int mw_count, qp_count, spad_count, max_mw_count_for_spads; mw_count 1252 drivers/ntb/ntb_transport.c mw_count = ntb_peer_mw_count(ndev); mw_count 1281 drivers/ntb/ntb_transport.c if (use_msi && mw_count > 1) { mw_count 1284 drivers/ntb/ntb_transport.c mw_count -= 1; mw_count 1294 drivers/ntb/ntb_transport.c nt->mw_count = 0; mw_count 1300 drivers/ntb/ntb_transport.c nt->mw_count = min(mw_count, max_mw_count_for_spads); mw_count 1302 drivers/ntb/ntb_transport.c nt->msi_spad_offset = nt->mw_count * 2 + MW0_SZ_HIGH; mw_count 1304 drivers/ntb/ntb_transport.c nt->mw_vec = kcalloc_node(mw_count, sizeof(*nt->mw_vec), mw_count 1311 drivers/ntb/ntb_transport.c for (i = 0; i < mw_count; i++) { mw_count 1342 drivers/ntb/ntb_transport.c else if (nt->mw_count < qp_count) mw_count 1343 drivers/ntb/ntb_transport.c qp_count = nt->mw_count; mw_count 1429 drivers/ntb/ntb_transport.c for (i = nt->mw_count; i--; ) { mw_count 272 include/linux/ntb.h int (*mw_count)(struct ntb_dev *ntb, int pidx); mw_count 349 include/linux/ntb.h ops->mw_count && mw_count 804 include/linux/ntb.h return ntb->ops->mw_count(ntb, pidx);