new_mask 312 arch/mips/cavium-octeon/smp.c uint32_t mask, new_mask; new_mask 333 arch/mips/cavium-octeon/smp.c new_mask = labi->avail_coremask; new_mask 338 arch/mips/cavium-octeon/smp.c new_mask = *p; new_mask 341 arch/mips/cavium-octeon/smp.c pr_info("Reset core %d. Available Coremask = 0x%x \n", coreid, new_mask); new_mask 68 arch/mips/kernel/mips-mt-fpaff.c cpumask_var_t cpus_allowed, new_mask, effective_mask; new_mask 73 arch/mips/kernel/mips-mt-fpaff.c if (len < sizeof(new_mask)) new_mask 76 arch/mips/kernel/mips-mt-fpaff.c if (copy_from_user(&new_mask, user_mask_ptr, sizeof(new_mask))) new_mask 97 arch/mips/kernel/mips-mt-fpaff.c if (!alloc_cpumask_var(&new_mask, GFP_KERNEL)) { new_mask 115 arch/mips/kernel/mips-mt-fpaff.c cpumask_copy(&p->thread.user_cpus_allowed, new_mask); new_mask 121 arch/mips/kernel/mips-mt-fpaff.c cpumask_intersects(new_mask, &mt_fpu_cpumask)) { new_mask 122 arch/mips/kernel/mips-mt-fpaff.c cpumask_and(effective_mask, new_mask, &mt_fpu_cpumask); new_mask 125 arch/mips/kernel/mips-mt-fpaff.c cpumask_copy(effective_mask, new_mask); new_mask 127 arch/mips/kernel/mips-mt-fpaff.c retval = set_cpus_allowed_ptr(p, new_mask); new_mask 138 arch/mips/kernel/mips-mt-fpaff.c cpumask_copy(new_mask, cpus_allowed); new_mask 145 arch/mips/kernel/mips-mt-fpaff.c free_cpumask_var(new_mask); new_mask 218 arch/x86/hyperv/hv_apic.c struct cpumask new_mask; new_mask 221 arch/x86/hyperv/hv_apic.c cpumask_copy(&new_mask, mask); new_mask 222 arch/x86/hyperv/hv_apic.c cpumask_clear_cpu(this_cpu, &new_mask); new_mask 223 arch/x86/hyperv/hv_apic.c local_mask = &new_mask; new_mask 496 arch/x86/kernel/kvm.c struct cpumask new_mask; new_mask 499 arch/x86/kernel/kvm.c cpumask_copy(&new_mask, mask); new_mask 500 arch/x86/kernel/kvm.c cpumask_clear_cpu(this_cpu, &new_mask); new_mask 501 arch/x86/kernel/kvm.c local_mask = &new_mask; new_mask 3110 drivers/ata/libata-eh.c unsigned int new_mask = 0; new_mask 3175 drivers/ata/libata-eh.c new_mask |= 1 << dev->devno; new_mask 3201 drivers/ata/libata-eh.c if (!(new_mask & (1 << dev->devno))) new_mask 1040 drivers/ata/sata_mv.c u32 old_mask, new_mask; new_mask 1043 drivers/ata/sata_mv.c new_mask = (old_mask & ~disable_bits) | enable_bits; new_mask 1044 drivers/ata/sata_mv.c if (new_mask != old_mask) { new_mask 1045 drivers/ata/sata_mv.c hpriv->main_irq_mask = new_mask; new_mask 1046 drivers/ata/sata_mv.c mv_write_main_irq_mask(new_mask, hpriv); new_mask 2476 drivers/gpu/drm/amd/amdgpu/kv_dpm.c u32 new_mask = (1 << level); new_mask 2480 drivers/gpu/drm/amd/amdgpu/kv_dpm.c new_mask); new_mask 2486 drivers/gpu/drm/amd/amdgpu/kv_dpm.c u32 i, new_mask = 0; new_mask 2489 drivers/gpu/drm/amd/amdgpu/kv_dpm.c new_mask |= (1 << i); new_mask 2493 drivers/gpu/drm/amd/amdgpu/kv_dpm.c new_mask); new_mask 98 drivers/gpu/drm/amd/powerplay/amdgpu_smu.c int smu_sys_set_pp_feature_mask(struct smu_context *smu, uint64_t new_mask) new_mask 112 drivers/gpu/drm/amd/powerplay/amdgpu_smu.c feature_2_enabled = ~feature_enables & new_mask; new_mask 113 drivers/gpu/drm/amd/powerplay/amdgpu_smu.c feature_2_disabled = feature_enables & ~new_mask; new_mask 836 drivers/gpu/drm/amd/powerplay/inc/amdgpu_smu.h int smu_sys_set_pp_feature_mask(struct smu_context *smu, uint64_t new_mask); new_mask 420 drivers/gpu/drm/arm/malidp_crtc.c u32 new_mask = state->connector_mask; new_mask 422 drivers/gpu/drm/arm/malidp_crtc.c if ((old_mask ^ new_mask) == new_mask 2411 drivers/gpu/drm/radeon/kv_dpm.c u32 new_mask = (1 << level); new_mask 2415 drivers/gpu/drm/radeon/kv_dpm.c new_mask); new_mask 2421 drivers/gpu/drm/radeon/kv_dpm.c u32 i, new_mask = 0; new_mask 2424 drivers/gpu/drm/radeon/kv_dpm.c new_mask |= (1 << i); new_mask 2428 drivers/gpu/drm/radeon/kv_dpm.c new_mask); new_mask 15 drivers/infiniband/core/counters.c enum rdma_nl_counter_mask new_mask) new_mask 18 drivers/infiniband/core/counters.c ((new_mask & (~ALL_AUTO_MODE_MASKS)) || new_mask 23 drivers/infiniband/core/counters.c curr->mask = new_mask; new_mask 135 drivers/infiniband/core/counters.c enum rdma_nl_counter_mask new_mask) new_mask 140 drivers/infiniband/core/counters.c counter->mode.mask = new_mask; new_mask 142 drivers/infiniband/core/counters.c if (new_mask & RDMA_COUNTER_MASK_QP_TYPE) new_mask 937 drivers/infiniband/hw/hfi1/sdma.c cpumask_var_t mask, new_mask; new_mask 950 drivers/infiniband/hw/hfi1/sdma.c ret = zalloc_cpumask_var(&new_mask, GFP_KERNEL); new_mask 973 drivers/infiniband/hw/hfi1/sdma.c cpumask_set_cpu(cpu, new_mask); new_mask 1029 drivers/infiniband/hw/hfi1/sdma.c cpumask_set_cpu(cpu, new_mask); new_mask 1077 drivers/infiniband/hw/hfi1/sdma.c cpumask_copy(&sde->cpu_mask, new_mask); new_mask 1082 drivers/infiniband/hw/hfi1/sdma.c free_cpumask_var(new_mask); new_mask 1413 drivers/input/keyboard/atkbd.c DECLARE_BITMAP(new_mask, ATKBD_KEYMAP_SIZE); new_mask 1416 drivers/input/keyboard/atkbd.c err = bitmap_parselist(buf, new_mask, ATKBD_KEYMAP_SIZE); new_mask 1420 drivers/input/keyboard/atkbd.c memcpy(atkbd->force_release_mask, new_mask, sizeof(atkbd->force_release_mask)); new_mask 169 drivers/net/ethernet/brocade/bna/bna_hw_defs.h #define bna_intx_enable(bna, new_mask) \ new_mask 170 drivers/net/ethernet/brocade/bna/bna_hw_defs.h writel((new_mask), (bna)->regs.fn_int_mask) new_mask 1234 drivers/net/ethernet/faraday/ftgmac100.c unsigned int status, new_mask = FTGMAC100_INT_BAD; new_mask 1269 drivers/net/ethernet/faraday/ftgmac100.c new_mask &= ~status; new_mask 1273 drivers/net/ethernet/faraday/ftgmac100.c iowrite32(new_mask, priv->base + FTGMAC100_OFFSET_IER); new_mask 4058 drivers/net/ethernet/intel/i40e/i40e_ethtool.c u64 current_mask, new_mask; new_mask 4090 drivers/net/ethernet/intel/i40e/i40e_ethtool.c new_mask = current_mask; new_mask 4103 drivers/net/ethernet/intel/i40e/i40e_ethtool.c new_mask &= ~I40E_VERIFY_TAG_MASK; new_mask 4111 drivers/net/ethernet/intel/i40e/i40e_ethtool.c new_mask |= I40E_L3_SRC_MASK; new_mask 4113 drivers/net/ethernet/intel/i40e/i40e_ethtool.c new_mask &= ~I40E_L3_SRC_MASK; new_mask 4119 drivers/net/ethernet/intel/i40e/i40e_ethtool.c new_mask |= I40E_L3_DST_MASK; new_mask 4121 drivers/net/ethernet/intel/i40e/i40e_ethtool.c new_mask &= ~I40E_L3_DST_MASK; new_mask 4127 drivers/net/ethernet/intel/i40e/i40e_ethtool.c new_mask |= I40E_L4_SRC_MASK; new_mask 4129 drivers/net/ethernet/intel/i40e/i40e_ethtool.c new_mask &= ~I40E_L4_SRC_MASK; new_mask 4135 drivers/net/ethernet/intel/i40e/i40e_ethtool.c new_mask |= I40E_L4_DST_MASK; new_mask 4137 drivers/net/ethernet/intel/i40e/i40e_ethtool.c new_mask &= ~I40E_L4_DST_MASK; new_mask 4151 drivers/net/ethernet/intel/i40e/i40e_ethtool.c new_mask |= I40E_L3_SRC_MASK; new_mask 4153 drivers/net/ethernet/intel/i40e/i40e_ethtool.c new_mask &= ~I40E_L3_SRC_MASK; new_mask 4159 drivers/net/ethernet/intel/i40e/i40e_ethtool.c new_mask |= I40E_L3_DST_MASK; new_mask 4161 drivers/net/ethernet/intel/i40e/i40e_ethtool.c new_mask &= ~I40E_L3_DST_MASK; new_mask 4167 drivers/net/ethernet/intel/i40e/i40e_ethtool.c new_mask |= I40E_L4_SRC_MASK | I40E_L4_DST_MASK; new_mask 4169 drivers/net/ethernet/intel/i40e/i40e_ethtool.c new_mask &= ~(I40E_L4_SRC_MASK | I40E_L4_DST_MASK); new_mask 4191 drivers/net/ethernet/intel/i40e/i40e_ethtool.c new_mask &= ~I40E_FLEX_INPUT_MASK; new_mask 4274 drivers/net/ethernet/intel/i40e/i40e_ethtool.c new_mask |= i40e_pit_index_to_mask(pit_index); new_mask 4281 drivers/net/ethernet/intel/i40e/i40e_ethtool.c if (new_mask == current_mask && !new_flex_offset) new_mask 4286 drivers/net/ethernet/intel/i40e/i40e_ethtool.c i40e_print_input_set(vsi, current_mask, new_mask); new_mask 4317 drivers/net/ethernet/intel/i40e/i40e_ethtool.c i40e_write_fd_input_set(pf, index, new_mask); new_mask 4327 drivers/net/ethernet/intel/i40e/i40e_ethtool.c new_mask); new_mask 290 drivers/net/fddi/defza.c unsigned int old_mask, new_mask; new_mask 296 drivers/net/fddi/defza.c new_mask = old_mask & ~FZA_MASK_STATE_CHG; new_mask 297 drivers/net/fddi/defza.c writew_u(new_mask, &fp->regs->int_mask); new_mask 299 drivers/net/fddi/defza.c fp->int_mask = new_mask; new_mask 1081 drivers/net/fddi/defza.c unsigned int old_mask, new_mask; new_mask 1123 drivers/net/fddi/defza.c new_mask = old_mask & ~FZA_MASK_SMT_TX_POLL; new_mask 1124 drivers/net/fddi/defza.c writew_u(new_mask, &fp->regs->int_mask); new_mask 1126 drivers/net/fddi/defza.c fp->int_mask = new_mask; new_mask 1521 drivers/net/wireless/ath/ath5k/ath5k.h enum ath5k_int ath5k_hw_set_imr(struct ath5k_hw *ah, enum ath5k_int new_mask); new_mask 755 drivers/net/wireless/ath/ath5k/dma.c ath5k_hw_set_imr(struct ath5k_hw *ah, enum ath5k_int new_mask) new_mask 775 drivers/net/wireless/ath/ath5k/dma.c int_mask = new_mask & AR5K_INT_COMMON; new_mask 783 drivers/net/wireless/ath/ath5k/dma.c if (new_mask & AR5K_INT_FATAL) { new_mask 790 drivers/net/wireless/ath/ath5k/dma.c if (new_mask & AR5K_INT_TIM) new_mask 793 drivers/net/wireless/ath/ath5k/dma.c if (new_mask & AR5K_INT_TIM) new_mask 795 drivers/net/wireless/ath/ath5k/dma.c if (new_mask & AR5K_INT_DTIM) new_mask 797 drivers/net/wireless/ath/ath5k/dma.c if (new_mask & AR5K_INT_DTIM_SYNC) new_mask 799 drivers/net/wireless/ath/ath5k/dma.c if (new_mask & AR5K_INT_BCN_TIMEOUT) new_mask 801 drivers/net/wireless/ath/ath5k/dma.c if (new_mask & AR5K_INT_CAB_TIMEOUT) new_mask 805 drivers/net/wireless/ath/ath5k/dma.c if (new_mask & AR5K_INT_BNR) new_mask 815 drivers/net/wireless/ath/ath5k/dma.c if (new_mask & AR5K_INT_FATAL) new_mask 825 drivers/net/wireless/ath/ath5k/dma.c if (!(new_mask & AR5K_INT_RXNOFRM)) new_mask 829 drivers/net/wireless/ath/ath5k/dma.c ah->ah_imr = new_mask; new_mask 832 drivers/net/wireless/ath/ath5k/dma.c if (new_mask & AR5K_INT_GLOBAL) { new_mask 277 drivers/net/wireless/marvell/libertas/debugfs.c int value, freq, new_mask; new_mask 286 drivers/net/wireless/marvell/libertas/debugfs.c ret = sscanf(buf, "%d %d %d", &value, &freq, &new_mask); new_mask 306 drivers/net/wireless/marvell/libertas/debugfs.c if (new_mask) new_mask 307 drivers/net/wireless/marvell/libertas/debugfs.c new_mask = curr_mask | event_mask; new_mask 309 drivers/net/wireless/marvell/libertas/debugfs.c new_mask = curr_mask & ~event_mask; new_mask 316 drivers/net/wireless/marvell/libertas/debugfs.c events->events = cpu_to_le16(new_mask); new_mask 374 drivers/net/wireless/quantenna/qtnfmac/cfg80211.c u16 new_mask; new_mask 380 drivers/net/wireless/quantenna/qtnfmac/cfg80211.c new_mask = vif->mgmt_frames_bitmask | BIT(mgmt_type); new_mask 382 drivers/net/wireless/quantenna/qtnfmac/cfg80211.c new_mask = vif->mgmt_frames_bitmask & ~BIT(mgmt_type); new_mask 384 drivers/net/wireless/quantenna/qtnfmac/cfg80211.c if (new_mask == vif->mgmt_frames_bitmask) new_mask 415 drivers/net/wireless/quantenna/qtnfmac/cfg80211.c vif->mgmt_frames_bitmask = new_mask; new_mask 3016 drivers/scsi/lpfc/lpfc_ct.c int cmdcode, uint32_t new_mask) new_mask 3105 drivers/scsi/lpfc/lpfc_ct.c if (new_mask) new_mask 3106 drivers/scsi/lpfc/lpfc_ct.c mask = new_mask; new_mask 3150 drivers/scsi/lpfc/lpfc_ct.c if (new_mask) new_mask 3151 drivers/scsi/lpfc/lpfc_ct.c mask = new_mask; new_mask 206 drivers/spi/spi-dw.h u32 new_mask; new_mask 208 drivers/spi/spi-dw.h new_mask = dw_readl(dws, DW_SPI_IMR) & ~mask; new_mask 209 drivers/spi/spi-dw.h dw_writel(dws, DW_SPI_IMR, new_mask); new_mask 215 drivers/spi/spi-dw.h u32 new_mask; new_mask 217 drivers/spi/spi-dw.h new_mask = dw_readl(dws, DW_SPI_IMR) | mask; new_mask 218 drivers/spi/spi-dw.h dw_writel(dws, DW_SPI_IMR, new_mask); new_mask 48 fs/notify/dnotify/dnotify.c __u32 new_mask = 0; new_mask 57 fs/notify/dnotify/dnotify.c new_mask |= (dn->dn_mask & ~FS_DN_MULTISHOT); new_mask 58 fs/notify/dnotify/dnotify.c if (fsn_mark->mask == new_mask) new_mask 60 fs/notify/dnotify/dnotify.c fsn_mark->mask = new_mask; new_mask 192 fs/notify/dnotify/dnotify.c __u32 new_mask = FS_EVENT_ON_CHILD; new_mask 195 fs/notify/dnotify/dnotify.c new_mask |= FS_DN_MULTISHOT; new_mask 197 fs/notify/dnotify/dnotify.c new_mask |= (FS_DELETE | FS_MOVED_FROM); new_mask 199 fs/notify/dnotify/dnotify.c new_mask |= FS_MODIFY; new_mask 201 fs/notify/dnotify/dnotify.c new_mask |= FS_ACCESS; new_mask 203 fs/notify/dnotify/dnotify.c new_mask |= FS_ATTRIB; new_mask 205 fs/notify/dnotify/dnotify.c new_mask |= FS_DN_RENAME; new_mask 207 fs/notify/dnotify/dnotify.c new_mask |= (FS_CREATE | FS_MOVED_TO); new_mask 209 fs/notify/dnotify/dnotify.c return new_mask; new_mask 509 fs/notify/inotify/inotify_user.c __u32 old_mask, new_mask; new_mask 533 fs/notify/inotify/inotify_user.c new_mask = fsn_mark->mask; new_mask 536 fs/notify/inotify/inotify_user.c if (old_mask != new_mask) { new_mask 538 fs/notify/inotify/inotify_user.c int dropped = (old_mask & ~new_mask); new_mask 540 fs/notify/inotify/inotify_user.c int do_inode = (new_mask & ~inode->i_fsnotify_mask); new_mask 121 fs/notify/mark.c u32 new_mask = 0; new_mask 130 fs/notify/mark.c new_mask |= mark->mask; new_mask 132 fs/notify/mark.c *fsnotify_conn_mask_p(conn) = new_mask; new_mask 1574 include/linux/sched.h extern void do_set_cpus_allowed(struct task_struct *p, const struct cpumask *new_mask); new_mask 1575 include/linux/sched.h extern int set_cpus_allowed_ptr(struct task_struct *p, const struct cpumask *new_mask); new_mask 1577 include/linux/sched.h static inline void do_set_cpus_allowed(struct task_struct *p, const struct cpumask *new_mask) new_mask 1580 include/linux/sched.h static inline int set_cpus_allowed_ptr(struct task_struct *p, const struct cpumask *new_mask) new_mask 1582 include/linux/sched.h if (!cpumask_test_cpu(0, new_mask)) new_mask 1853 include/linux/sched.h extern long sched_setaffinity(pid_t pid, const struct cpumask *new_mask); new_mask 199 kernel/compat.c unsigned len, struct cpumask *new_mask) new_mask 204 kernel/compat.c memset(new_mask, 0, cpumask_size()); new_mask 208 kernel/compat.c k = cpumask_bits(new_mask); new_mask 216 kernel/compat.c cpumask_var_t new_mask; new_mask 219 kernel/compat.c if (!alloc_cpumask_var(&new_mask, GFP_KERNEL)) new_mask 222 kernel/compat.c retval = compat_get_user_cpu_mask(user_mask_ptr, len, new_mask); new_mask 226 kernel/compat.c retval = sched_setaffinity(pid, new_mask); new_mask 228 kernel/compat.c free_cpumask_var(new_mask); new_mask 3635 kernel/locking/lockdep.c unsigned int new_mask = 1 << new_bit, ret = 1; new_mask 3646 kernel/locking/lockdep.c if (likely(hlock_class(this)->usage_mask & new_mask)) new_mask 3654 kernel/locking/lockdep.c if (unlikely(hlock_class(this)->usage_mask & new_mask)) { new_mask 3659 kernel/locking/lockdep.c hlock_class(this)->usage_mask |= new_mask; new_mask 1580 kernel/sched/core.c void set_cpus_allowed_common(struct task_struct *p, const struct cpumask *new_mask) new_mask 1582 kernel/sched/core.c cpumask_copy(&p->cpus_mask, new_mask); new_mask 1583 kernel/sched/core.c p->nr_cpus_allowed = cpumask_weight(new_mask); new_mask 1586 kernel/sched/core.c void do_set_cpus_allowed(struct task_struct *p, const struct cpumask *new_mask) new_mask 1607 kernel/sched/core.c p->sched_class->set_cpus_allowed(p, new_mask); new_mask 1625 kernel/sched/core.c const struct cpumask *new_mask, bool check) new_mask 1652 kernel/sched/core.c if (cpumask_equal(p->cpus_ptr, new_mask)) new_mask 1655 kernel/sched/core.c dest_cpu = cpumask_any_and(cpu_valid_mask, new_mask); new_mask 1661 kernel/sched/core.c do_set_cpus_allowed(p, new_mask); new_mask 1668 kernel/sched/core.c WARN_ON(cpumask_intersects(new_mask, cpu_online_mask) && new_mask 1669 kernel/sched/core.c !cpumask_intersects(new_mask, cpu_active_mask) && new_mask 1674 kernel/sched/core.c if (cpumask_test_cpu(task_cpu(p), new_mask)) new_mask 1696 kernel/sched/core.c int set_cpus_allowed_ptr(struct task_struct *p, const struct cpumask *new_mask) new_mask 1698 kernel/sched/core.c return __set_cpus_allowed_ptr(p, new_mask, false); new_mask 2163 kernel/sched/core.c const struct cpumask *new_mask, bool check) new_mask 2165 kernel/sched/core.c return set_cpus_allowed_ptr(p, new_mask); new_mask 5386 kernel/sched/core.c cpumask_var_t cpus_allowed, new_mask; new_mask 5410 kernel/sched/core.c if (!alloc_cpumask_var(&new_mask, GFP_KERNEL)) { new_mask 5430 kernel/sched/core.c cpumask_and(new_mask, in_mask, cpus_allowed); new_mask 5441 kernel/sched/core.c if (!cpumask_subset(task_rq(p)->rd->span, new_mask)) { new_mask 5450 kernel/sched/core.c retval = __set_cpus_allowed_ptr(p, new_mask, true); new_mask 5454 kernel/sched/core.c if (!cpumask_subset(new_mask, cpus_allowed)) { new_mask 5460 kernel/sched/core.c cpumask_copy(new_mask, cpus_allowed); new_mask 5465 kernel/sched/core.c free_cpumask_var(new_mask); new_mask 5474 kernel/sched/core.c struct cpumask *new_mask) new_mask 5477 kernel/sched/core.c cpumask_clear(new_mask); new_mask 5481 kernel/sched/core.c return copy_from_user(new_mask, user_mask_ptr, len) ? -EFAULT : 0; new_mask 5495 kernel/sched/core.c cpumask_var_t new_mask; new_mask 5498 kernel/sched/core.c if (!alloc_cpumask_var(&new_mask, GFP_KERNEL)) new_mask 5501 kernel/sched/core.c retval = get_user_cpu_mask(user_mask_ptr, len, new_mask); new_mask 5503 kernel/sched/core.c retval = sched_setaffinity(pid, new_mask); new_mask 5504 kernel/sched/core.c free_cpumask_var(new_mask); new_mask 2232 kernel/sched/deadline.c const struct cpumask *new_mask) new_mask 2247 kernel/sched/deadline.c if (!cpumask_intersects(src_rd->span, new_mask)) { new_mask 2261 kernel/sched/deadline.c set_cpus_allowed_common(p, new_mask); new_mask 1836 kernel/sched/sched.h extern void set_cpus_allowed_common(struct task_struct *p, const struct cpumask *new_mask); new_mask 170 kernel/time/sched_clock.c u64 res, wrap, new_mask, new_epoch, cyc, ns; new_mask 184 kernel/time/sched_clock.c new_mask = CLOCKSOURCE_MASK(bits); new_mask 188 kernel/time/sched_clock.c wrap = clocks_calc_max_nsecs(new_mult, new_shift, 0, new_mask, NULL); new_mask 200 kernel/time/sched_clock.c rd.sched_clock_mask = new_mask;