snp 244 drivers/infiniband/hw/qib/qib_eeprom.c char *snp = dd->serial; snp 250 drivers/infiniband/hw/qib/qib_eeprom.c memcpy(snp, ifp->if_sprefix, sizeof(ifp->if_sprefix)); snp 251 drivers/infiniband/hw/qib/qib_eeprom.c snp[sizeof(ifp->if_sprefix)] = '\0'; snp 252 drivers/infiniband/hw/qib/qib_eeprom.c len = strlen(snp); snp 253 drivers/infiniband/hw/qib/qib_eeprom.c snp += len; snp 257 drivers/infiniband/hw/qib/qib_eeprom.c memcpy(snp, ifp->if_serial, len); snp 90 kernel/rcu/srcutree.c struct srcu_node *snp; snp 100 kernel/rcu/srcutree.c srcu_for_each_node_breadth_first(ssp, snp) { snp 101 kernel/rcu/srcutree.c spin_lock_init(&ACCESS_PRIVATE(snp, lock)); snp 102 kernel/rcu/srcutree.c WARN_ON_ONCE(ARRAY_SIZE(snp->srcu_have_cbs) != snp 103 kernel/rcu/srcutree.c ARRAY_SIZE(snp->srcu_data_have_cbs)); snp 104 kernel/rcu/srcutree.c for (i = 0; i < ARRAY_SIZE(snp->srcu_have_cbs); i++) { snp 105 kernel/rcu/srcutree.c snp->srcu_have_cbs[i] = 0; snp 106 kernel/rcu/srcutree.c snp->srcu_data_have_cbs[i] = 0; snp 108 kernel/rcu/srcutree.c snp->srcu_gp_seq_needed_exp = 0; snp 109 kernel/rcu/srcutree.c snp->grplo = -1; snp 110 kernel/rcu/srcutree.c snp->grphi = -1; snp 111 kernel/rcu/srcutree.c if (snp == &ssp->node[0]) { snp 113 kernel/rcu/srcutree.c snp->srcu_parent = NULL; snp 118 kernel/rcu/srcutree.c if (snp == ssp->level[level + 1]) snp 120 kernel/rcu/srcutree.c snp->srcu_parent = ssp->level[level - 1] + snp 121 kernel/rcu/srcutree.c (snp - ssp->level[level]) / snp 141 kernel/rcu/srcutree.c for (snp = sdp->mynode; snp != NULL; snp = snp->srcu_parent) { snp 142 kernel/rcu/srcutree.c if (snp->grplo < 0) snp 143 kernel/rcu/srcutree.c snp->grplo = cpu; snp 144 kernel/rcu/srcutree.c snp->grphi = cpu; snp 491 kernel/rcu/srcutree.c static void srcu_schedule_cbs_snp(struct srcu_struct *ssp, struct srcu_node *snp, snp 496 kernel/rcu/srcutree.c for (cpu = snp->grplo; cpu <= snp->grphi; cpu++) { snp 497 kernel/rcu/srcutree.c if (!(mask & (1 << (cpu - snp->grplo)))) snp 523 kernel/rcu/srcutree.c struct srcu_node *snp; snp 543 kernel/rcu/srcutree.c idx = rcu_seq_ctr(gpseq) % ARRAY_SIZE(snp->srcu_have_cbs); snp 544 kernel/rcu/srcutree.c srcu_for_each_node_breadth_first(ssp, snp) { snp 545 kernel/rcu/srcutree.c spin_lock_irq_rcu_node(snp); snp 547 kernel/rcu/srcutree.c last_lvl = snp >= ssp->level[rcu_num_lvls - 1]; snp 549 kernel/rcu/srcutree.c cbs = snp->srcu_have_cbs[idx] == gpseq; snp 550 kernel/rcu/srcutree.c snp->srcu_have_cbs[idx] = gpseq; snp 551 kernel/rcu/srcutree.c rcu_seq_set_state(&snp->srcu_have_cbs[idx], 1); snp 552 kernel/rcu/srcutree.c if (ULONG_CMP_LT(snp->srcu_gp_seq_needed_exp, gpseq)) snp 553 kernel/rcu/srcutree.c snp->srcu_gp_seq_needed_exp = gpseq; snp 554 kernel/rcu/srcutree.c mask = snp->srcu_data_have_cbs[idx]; snp 555 kernel/rcu/srcutree.c snp->srcu_data_have_cbs[idx] = 0; snp 556 kernel/rcu/srcutree.c spin_unlock_irq_rcu_node(snp); snp 558 kernel/rcu/srcutree.c srcu_schedule_cbs_snp(ssp, snp, mask, cbdelay); snp 562 kernel/rcu/srcutree.c for (cpu = snp->grplo; cpu <= snp->grphi; cpu++) { snp 598 kernel/rcu/srcutree.c static void srcu_funnel_exp_start(struct srcu_struct *ssp, struct srcu_node *snp, snp 603 kernel/rcu/srcutree.c for (; snp != NULL; snp = snp->srcu_parent) { snp 605 kernel/rcu/srcutree.c ULONG_CMP_GE(READ_ONCE(snp->srcu_gp_seq_needed_exp), s)) snp 607 kernel/rcu/srcutree.c spin_lock_irqsave_rcu_node(snp, flags); snp 608 kernel/rcu/srcutree.c if (ULONG_CMP_GE(snp->srcu_gp_seq_needed_exp, s)) { snp 609 kernel/rcu/srcutree.c spin_unlock_irqrestore_rcu_node(snp, flags); snp 612 kernel/rcu/srcutree.c WRITE_ONCE(snp->srcu_gp_seq_needed_exp, s); snp 613 kernel/rcu/srcutree.c spin_unlock_irqrestore_rcu_node(snp, flags); snp 636 kernel/rcu/srcutree.c struct srcu_node *snp = sdp->mynode; snp 640 kernel/rcu/srcutree.c for (; snp != NULL; snp = snp->srcu_parent) { snp 641 kernel/rcu/srcutree.c if (rcu_seq_done(&ssp->srcu_gp_seq, s) && snp != sdp->mynode) snp 643 kernel/rcu/srcutree.c spin_lock_irqsave_rcu_node(snp, flags); snp 644 kernel/rcu/srcutree.c if (ULONG_CMP_GE(snp->srcu_have_cbs[idx], s)) { snp 645 kernel/rcu/srcutree.c snp_seq = snp->srcu_have_cbs[idx]; snp 646 kernel/rcu/srcutree.c if (snp == sdp->mynode && snp_seq == s) snp 647 kernel/rcu/srcutree.c snp->srcu_data_have_cbs[idx] |= sdp->grpmask; snp 648 kernel/rcu/srcutree.c spin_unlock_irqrestore_rcu_node(snp, flags); snp 649 kernel/rcu/srcutree.c if (snp == sdp->mynode && snp_seq != s) { snp 656 kernel/rcu/srcutree.c srcu_funnel_exp_start(ssp, snp, s); snp 659 kernel/rcu/srcutree.c snp->srcu_have_cbs[idx] = s; snp 660 kernel/rcu/srcutree.c if (snp == sdp->mynode) snp 661 kernel/rcu/srcutree.c snp->srcu_data_have_cbs[idx] |= sdp->grpmask; snp 662 kernel/rcu/srcutree.c if (!do_norm && ULONG_CMP_LT(snp->srcu_gp_seq_needed_exp, s)) snp 663 kernel/rcu/srcutree.c snp->srcu_gp_seq_needed_exp = s; snp 664 kernel/rcu/srcutree.c spin_unlock_irqrestore_rcu_node(snp, flags); snp 2308 security/smack/smack_lsm.c struct smk_net4addr *snp; snp 2314 security/smack/smack_lsm.c list_for_each_entry_rcu(snp, &smk_net4addr_list, list) snp 2320 security/smack/smack_lsm.c if (snp->smk_host.s_addr == snp 2321 security/smack/smack_lsm.c (siap->s_addr & snp->smk_mask.s_addr)) snp 2322 security/smack/smack_lsm.c return snp->smk_label; snp 2358 security/smack/smack_lsm.c struct smk_net6addr *snp; snp 2369 security/smack/smack_lsm.c list_for_each_entry_rcu(snp, &smk_net6addr_list, list) { snp 2374 security/smack/smack_lsm.c if (snp->smk_label == NULL) snp 2382 security/smack/smack_lsm.c if ((sap->s6_addr16[i] & snp->smk_mask.s6_addr16[i]) != snp 2383 security/smack/smack_lsm.c snp->smk_host.s6_addr16[i]) { snp 2389 security/smack/smack_lsm.c return snp->smk_label; snp 1135 security/smack/smackfs.c struct smk_net4addr *snp; snp 1227 security/smack/smackfs.c list_for_each_entry_rcu(snp, &smk_net4addr_list, list) { snp 1228 security/smack/smackfs.c if (snp->smk_host.s_addr == nsa && snp->smk_masks == masks) { snp 1236 security/smack/smackfs.c snp = kzalloc(sizeof(*snp), GFP_KERNEL); snp 1237 security/smack/smackfs.c if (snp == NULL) snp 1241 security/smack/smackfs.c snp->smk_host.s_addr = newname.sin_addr.s_addr; snp 1242 security/smack/smackfs.c snp->smk_mask.s_addr = mask.s_addr; snp 1243 security/smack/smackfs.c snp->smk_label = skp; snp 1244 security/smack/smackfs.c snp->smk_masks = masks; snp 1245 security/smack/smackfs.c smk_net4addr_insert(snp); snp 1252 security/smack/smackfs.c if (snp->smk_label != NULL) snp 1254 security/smack/smackfs.c &snp->smk_host, &snp->smk_mask, snp 1258 security/smack/smackfs.c snp->smk_label = skp; snp 1268 security/smack/smackfs.c &snp->smk_host, &snp->smk_mask, PF_INET, snp 1269 security/smack/smackfs.c snp->smk_label->smk_secid, &audit_info); snp 1398 security/smack/smackfs.c struct smk_net6addr *snp; snp 1500 security/smack/smackfs.c list_for_each_entry_rcu(snp, &smk_net6addr_list, list) { snp 1501 security/smack/smackfs.c if (mask != snp->smk_masks) snp 1505 security/smack/smackfs.c snp->smk_host.s6_addr16[i]) { snp 1514 security/smack/smackfs.c snp = kzalloc(sizeof(*snp), GFP_KERNEL); snp 1515 security/smack/smackfs.c if (snp == NULL) snp 1518 security/smack/smackfs.c snp->smk_host = newname; snp 1519 security/smack/smackfs.c snp->smk_mask = fullmask; snp 1520 security/smack/smackfs.c snp->smk_masks = mask; snp 1521 security/smack/smackfs.c snp->smk_label = skp; snp 1522 security/smack/smackfs.c smk_net6addr_insert(snp); snp 1525 security/smack/smackfs.c snp->smk_label = skp;