p_mask 1262 arch/x86/include/asm/kvm_host.h u64 dirty_mask, u64 nx_mask, u64 x_mask, u64 p_mask, p_mask 522 arch/x86/kvm/mmu.c u64 dirty_mask, u64 nx_mask, u64 x_mask, u64 p_mask, p_mask 534 arch/x86/kvm/mmu.c shadow_present_mask = p_mask; p_mask 167 drivers/clk/sunxi-ng/ccu_nkmp.c u32 n_mask = 0, k_mask = 0, m_mask = 0, p_mask = 0; p_mask 202 drivers/clk/sunxi-ng/ccu_nkmp.c p_mask = GENMASK(nkmp->p.width + nkmp->p.shift - 1, p_mask 208 drivers/clk/sunxi-ng/ccu_nkmp.c reg &= ~(n_mask | k_mask | m_mask | p_mask); p_mask 213 drivers/clk/sunxi-ng/ccu_nkmp.c reg |= (ilog2(_nkmp.p) << nkmp->p.shift) & p_mask; p_mask 475 drivers/net/ethernet/netronome/nfp/flower/action.c static void nfp_fl_set_helper32(u32 value, u32 mask, u8 *p_exact, u8 *p_mask) p_mask 478 drivers/net/ethernet/netronome/nfp/flower/action.c u32 oldmask = get_unaligned((u32 *)p_mask); p_mask 483 drivers/net/ethernet/netronome/nfp/flower/action.c put_unaligned(oldmask | mask, (u32 *)p_mask); p_mask 3920 drivers/scsi/qla2xxx/qla_nx.c unsigned long p_wait, w_time, p_mask; p_mask 3936 drivers/scsi/qla2xxx/qla_nx.c p_mask = cache_hdr->cache_ctrl.poll_mask; p_mask 3943 drivers/scsi/qla2xxx/qla_nx.c if (p_mask) { p_mask 3947 drivers/scsi/qla2xxx/qla_nx.c if ((c_value_r & p_mask) == 0) p_mask 3954 drivers/scsi/qla2xxx/qla_nx.c c_value_r, p_mask, w_time); p_mask 2477 drivers/scsi/qla2xxx/qla_nx2.c unsigned long p_wait, w_time, p_mask; p_mask 2495 drivers/scsi/qla2xxx/qla_nx2.c p_mask = cache_hdr->cache_ctrl.poll_mask; p_mask 2502 drivers/scsi/qla2xxx/qla_nx2.c if (p_mask) { p_mask 2507 drivers/scsi/qla2xxx/qla_nx2.c if ((c_value_r & p_mask) == 0) { p_mask 2152 drivers/scsi/qla4xxx/ql4_nx.c unsigned long p_wait, w_time, p_mask; p_mask 2170 drivers/scsi/qla4xxx/ql4_nx.c p_mask = cache_hdr->cache_ctrl.poll_mask; p_mask 2178 drivers/scsi/qla4xxx/ql4_nx.c if (p_mask) { p_mask 2183 drivers/scsi/qla4xxx/ql4_nx.c if ((c_value_r & p_mask) == 0) {