epb 63 arch/x86/kernel/cpu/intel_epb.c u64 epb; epb 65 arch/x86/kernel/cpu/intel_epb.c rdmsrl(MSR_IA32_ENERGY_PERF_BIAS, epb); epb 70 arch/x86/kernel/cpu/intel_epb.c this_cpu_write(saved_epb, (epb & EPB_MASK) | EPB_SAVED); epb 78 arch/x86/kernel/cpu/intel_epb.c u64 epb; epb 80 arch/x86/kernel/cpu/intel_epb.c rdmsrl(MSR_IA32_ENERGY_PERF_BIAS, epb); epb 91 arch/x86/kernel/cpu/intel_epb.c val = epb & EPB_MASK; epb 97 arch/x86/kernel/cpu/intel_epb.c wrmsrl(MSR_IA32_ENERGY_PERF_BIAS, (epb & ~EPB_MASK) | val); epb 125 arch/x86/kernel/cpu/intel_epb.c u64 epb; epb 128 arch/x86/kernel/cpu/intel_epb.c ret = rdmsrl_on_cpu(cpu, MSR_IA32_ENERGY_PERF_BIAS, &epb); epb 132 arch/x86/kernel/cpu/intel_epb.c return sprintf(buf, "%llu\n", epb); epb 140 arch/x86/kernel/cpu/intel_epb.c u64 epb, val; epb 150 arch/x86/kernel/cpu/intel_epb.c ret = rdmsrl_on_cpu(cpu, MSR_IA32_ENERGY_PERF_BIAS, &epb); epb 155 arch/x86/kernel/cpu/intel_epb.c (epb & ~EPB_MASK) | val); epb 524 drivers/cpufreq/intel_pstate.c u64 epb; epb 530 drivers/cpufreq/intel_pstate.c ret = rdmsrl_on_cpu(cpu_data->cpu, MSR_IA32_ENERGY_PERF_BIAS, &epb); epb 534 drivers/cpufreq/intel_pstate.c return (s16)(epb & 0x0f); epb 563 drivers/cpufreq/intel_pstate.c u64 epb; epb 569 drivers/cpufreq/intel_pstate.c ret = rdmsrl_on_cpu(cpu, MSR_IA32_ENERGY_PERF_BIAS, &epb); epb 573 drivers/cpufreq/intel_pstate.c epb = (epb & ~0x0f) | pref; epb 574 drivers/cpufreq/intel_pstate.c wrmsrl_on_cpu(cpu, MSR_IA32_ENERGY_PERF_BIAS, epb); epb 555 fs/ext4/indirect.c unsigned epb = inode->i_sb->s_blocksize / sizeof(u32); epb 566 fs/ext4/indirect.c count = count * epb + (epb - offsets[i] - 1); epb 733 fs/nilfs2/alloc.c const unsigned int epb = NILFS_MDT(inode)->mi_entries_per_block; epb 762 fs/nilfs2/alloc.c entry_start = rounddown(group_offset, epb); epb 781 fs/nilfs2/alloc.c group_offset < entry_start + epb) { epb 788 fs/nilfs2/alloc.c end = entry_start + epb; epb 800 fs/nilfs2/alloc.c entry_start = rounddown(group_offset, epb); epb 58 fs/ocfs2/quota_local.c int epb = ol_quota_entries_per_block(sb); epb 60 fs/ocfs2/quota_local.c return ol_quota_chunk_block(sb, c) + 1 + off / epb; epb 65 fs/ocfs2/quota_local.c int epb = ol_quota_entries_per_block(sb); epb 67 fs/ocfs2/quota_local.c return (off % epb) * sizeof(struct ocfs2_local_disk_dqblk); epb 85 fs/ocfs2/quota_local.c int epb = ol_quota_entries_per_block(sb); epb 88 fs/ocfs2/quota_local.c ol_quota_chunk_block(sb, c) - 1) * epb epb 1101 fs/ocfs2/quota_local.c int epb = ol_quota_entries_per_block(sb); epb 1197 fs/ocfs2/quota_local.c *offset = chunk_blocks * epb; epb 28 fs/quota/quota_tree.c unsigned int epb = info->dqi_usable_bs >> 2; epb 32 fs/quota/quota_tree.c id /= epb; epb 33 fs/quota/quota_tree.c return id % epb; epb 685 fs/quota/quota_tree.c unsigned int epb = info->dqi_usable_bs >> 2; epb 693 fs/quota/quota_tree.c level_inc *= epb; epb 701 fs/quota/quota_tree.c for (i = __get_index(info, *id, depth); i < epb; i++) { epb 714 fs/quota/quota_tree.c if (i == epb) { epb 49 include/linux/dqblk_qtree.h unsigned int epb = info->dqi_usable_bs >> 2; epb 50 include/linux/dqblk_qtree.h unsigned long long entries = epb; epb 54 include/linux/dqblk_qtree.h entries *= epb; epb 509 include/net/sctp/sctp.h #define sctp_for_each_hentry(epb, head) \ epb 510 include/net/sctp/sctp.h hlist_for_each_entry(epb, head, node) epb 723 net/sctp/input.c struct sctp_ep_common *epb; epb 725 net/sctp/input.c epb = &ep->base; epb 726 net/sctp/input.c epb->hashent = sctp_ep_hashfn(net, epb->bind_addr.port); epb 727 net/sctp/input.c head = &sctp_ep_hashtable[epb->hashent]; epb 766 net/sctp/input.c hlist_add_head(&epb->node, &head->chain); epb 788 net/sctp/input.c struct sctp_ep_common *epb; epb 790 net/sctp/input.c epb = &ep->base; epb 792 net/sctp/input.c epb->hashent = sctp_ep_hashfn(sock_net(sk), epb->bind_addr.port); epb 794 net/sctp/input.c head = &sctp_ep_hashtable[epb->hashent]; epb 800 net/sctp/input.c hlist_del_init(&epb->node); epb 833 net/sctp/input.c struct sctp_ep_common *epb; epb 843 net/sctp/input.c sctp_for_each_hentry(epb, &head->chain) { epb 844 net/sctp/input.c ep = sctp_ep(epb); epb 77 net/sctp/proc.c static void sctp_seq_dump_local_addrs(struct seq_file *seq, struct sctp_ep_common *epb) epb 85 net/sctp/proc.c if (epb->type == SCTP_EP_TYPE_ASSOCIATION) { epb 86 net/sctp/proc.c asoc = sctp_assoc(epb); epb 98 net/sctp/proc.c list_for_each_entry_rcu(laddr, &epb->bind_addr.address_list, list) { epb 164 net/sctp/proc.c struct sctp_ep_common *epb; epb 174 net/sctp/proc.c sctp_for_each_hentry(epb, &head->chain) { epb 175 net/sctp/proc.c ep = sctp_ep(epb); epb 176 net/sctp/proc.c sk = epb->sk; epb 181 net/sctp/proc.c epb->bind_addr.port, epb 185 net/sctp/proc.c sctp_seq_dump_local_addrs(seq, epb); epb 235 net/sctp/proc.c struct sctp_ep_common *epb; epb 249 net/sctp/proc.c epb = &assoc->base; epb 250 net/sctp/proc.c sk = epb->sk; epb 262 net/sctp/proc.c epb->bind_addr.port, epb 265 net/sctp/proc.c sctp_seq_dump_local_addrs(seq, epb); epb 5358 net/sctp/socket.c struct sctp_ep_common *epb; epb 5364 net/sctp/socket.c sctp_for_each_hentry(epb, &head->chain) { epb 5365 net/sctp/socket.c err = cb(sctp_ep(epb), p);