bcp 140 arch/x86/platform/uv/tlb_uv.c struct bau_control *bcp; bcp 148 arch/x86/platform/uv/tlb_uv.c bcp = &per_cpu(bau_control, cpu); bcp 149 arch/x86/platform/uv/tlb_uv.c bcp->nobau = false; bcp 159 arch/x86/platform/uv/tlb_uv.c struct bau_control *bcp; bcp 163 arch/x86/platform/uv/tlb_uv.c bcp = &per_cpu(bau_control, cpu); bcp 164 arch/x86/platform/uv/tlb_uv.c bcp->nobau = true; bcp 207 arch/x86/platform/uv/tlb_uv.c static void reply_to_message(struct msg_desc *mdp, struct bau_control *bcp, bcp 226 arch/x86/platform/uv/tlb_uv.c struct bau_control *bcp) bcp 234 arch/x86/platform/uv/tlb_uv.c struct ptc_stats *stat = bcp->statp; bcp 282 arch/x86/platform/uv/tlb_uv.c static void bau_process_message(struct msg_desc *mdp, struct bau_control *bcp, bcp 288 arch/x86/platform/uv/tlb_uv.c struct ptc_stats *stat = bcp->statp; bcp 290 arch/x86/platform/uv/tlb_uv.c struct bau_control *smaster = bcp->socket_master; bcp 310 arch/x86/platform/uv/tlb_uv.c if (msg->msg_type == MSG_RETRY && bcp == bcp->uvhub_master) bcp 311 arch/x86/platform/uv/tlb_uv.c bau_process_retry_msg(mdp, bcp); bcp 322 arch/x86/platform/uv/tlb_uv.c if (socket_ack_count == bcp->cpus_in_socket) { bcp 332 arch/x86/platform/uv/tlb_uv.c if (msg_ack_count == bcp->cpus_in_uvhub) { bcp 337 arch/x86/platform/uv/tlb_uv.c reply_to_message(mdp, bcp, do_acknowledge); bcp 373 arch/x86/platform/uv/tlb_uv.c struct bau_control *bcp = &per_cpu(bau_control, smp_processor_id()); bcp 376 arch/x86/platform/uv/tlb_uv.c struct ptc_stats *stat = bcp->statp; bcp 385 arch/x86/platform/uv/tlb_uv.c for (msg = bcp->queue_first, i = 0; i < DEST_Q_SIZE; msg++, i++) { bcp 419 arch/x86/platform/uv/tlb_uv.c static void reset_with_ipi(struct pnmask *distribution, struct bau_control *bcp) bcp 424 arch/x86/platform/uv/tlb_uv.c int sender = bcp->cpu; bcp 425 arch/x86/platform/uv/tlb_uv.c cpumask_t *mask = bcp->uvhub_master->cpumask; bcp 426 arch/x86/platform/uv/tlb_uv.c struct bau_control *smaster = bcp->socket_master; bcp 438 arch/x86/platform/uv/tlb_uv.c apnode = pnode + bcp->partition_base_pnode; bcp 528 arch/x86/platform/uv/tlb_uv.c struct bau_control *bcp, long try) bcp 532 arch/x86/platform/uv/tlb_uv.c u64 mmr_offset = bcp->status_mmr; bcp 533 arch/x86/platform/uv/tlb_uv.c int right_shift = bcp->status_index; bcp 534 arch/x86/platform/uv/tlb_uv.c struct ptc_stats *stat = bcp->statp; bcp 558 arch/x86/platform/uv/tlb_uv.c if (cycles_2_us(ttm - bcp->send_message) < timeout_us) { bcp 559 arch/x86/platform/uv/tlb_uv.c bcp->conseccompletes = 0; bcp 563 arch/x86/platform/uv/tlb_uv.c bcp->conseccompletes = 0; bcp 573 arch/x86/platform/uv/tlb_uv.c bcp->conseccompletes++; bcp 591 arch/x86/platform/uv/tlb_uv.c static int handle_uv2_busy(struct bau_control *bcp) bcp 593 arch/x86/platform/uv/tlb_uv.c struct ptc_stats *stat = bcp->statp; bcp 596 arch/x86/platform/uv/tlb_uv.c bcp->busy = 1; bcp 601 arch/x86/platform/uv/tlb_uv.c struct bau_control *bcp, long try) bcp 605 arch/x86/platform/uv/tlb_uv.c u64 mmr_offset = bcp->status_mmr; bcp 606 arch/x86/platform/uv/tlb_uv.c int right_shift = bcp->status_index; bcp 607 arch/x86/platform/uv/tlb_uv.c int desc = bcp->uvhub_cpu; bcp 609 arch/x86/platform/uv/tlb_uv.c struct ptc_stats *stat = bcp->statp; bcp 637 arch/x86/platform/uv/tlb_uv.c if (cycles_2_us(ttm - bcp->send_message) < timeout_us) { bcp 638 arch/x86/platform/uv/tlb_uv.c bcp->conseccompletes = 0; bcp 644 arch/x86/platform/uv/tlb_uv.c bcp->conseccompletes = 0; bcp 653 arch/x86/platform/uv/tlb_uv.c if ((ttm - bcp->send_message) > bcp->timeout_interval) bcp 654 arch/x86/platform/uv/tlb_uv.c return handle_uv2_busy(bcp); bcp 663 arch/x86/platform/uv/tlb_uv.c bcp->conseccompletes++; bcp 682 arch/x86/platform/uv/tlb_uv.c struct bau_control *bcp, long try) bcp 684 arch/x86/platform/uv/tlb_uv.c struct ptc_stats *stat = bcp->statp; bcp 686 arch/x86/platform/uv/tlb_uv.c u64 mmr = bcp->status_mmr; bcp 687 arch/x86/platform/uv/tlb_uv.c int index = bcp->status_index; bcp 688 arch/x86/platform/uv/tlb_uv.c int desc = bcp->uvhub_cpu; bcp 701 arch/x86/platform/uv/tlb_uv.c bcp->conseccompletes = 0; bcp 706 arch/x86/platform/uv/tlb_uv.c bcp->conseccompletes = 0; bcp 710 arch/x86/platform/uv/tlb_uv.c bcp->conseccompletes = 0; bcp 719 arch/x86/platform/uv/tlb_uv.c bcp->conseccompletes++; bcp 729 arch/x86/platform/uv/tlb_uv.c struct bau_control *bcp, bcp 732 arch/x86/platform/uv/tlb_uv.c udelay(bcp->plugged_delay); bcp 733 arch/x86/platform/uv/tlb_uv.c bcp->plugged_tries++; bcp 735 arch/x86/platform/uv/tlb_uv.c if (bcp->plugged_tries >= bcp->plugsb4reset) { bcp 736 arch/x86/platform/uv/tlb_uv.c bcp->plugged_tries = 0; bcp 741 arch/x86/platform/uv/tlb_uv.c reset_with_ipi(&bau_desc->distribution, bcp); bcp 746 arch/x86/platform/uv/tlb_uv.c bcp->ipi_attempts++; bcp 752 arch/x86/platform/uv/tlb_uv.c struct bau_control *bcp, struct bau_control *hmaster, bcp 756 arch/x86/platform/uv/tlb_uv.c bcp->timeout_tries++; bcp 757 arch/x86/platform/uv/tlb_uv.c if (bcp->timeout_tries >= bcp->timeoutsb4reset) { bcp 758 arch/x86/platform/uv/tlb_uv.c bcp->timeout_tries = 0; bcp 763 arch/x86/platform/uv/tlb_uv.c reset_with_ipi(&bau_desc->distribution, bcp); bcp 768 arch/x86/platform/uv/tlb_uv.c bcp->ipi_attempts++; bcp 777 arch/x86/platform/uv/tlb_uv.c static void disable_for_period(struct bau_control *bcp, struct ptc_stats *stat) bcp 784 arch/x86/platform/uv/tlb_uv.c hmaster = bcp->uvhub_master; bcp 786 arch/x86/platform/uv/tlb_uv.c if (!bcp->baudisabled) { bcp 794 arch/x86/platform/uv/tlb_uv.c tm1 + bcp->disabled_period; bcp 801 arch/x86/platform/uv/tlb_uv.c static void count_max_concurr(int stat, struct bau_control *bcp, bcp 804 arch/x86/platform/uv/tlb_uv.c bcp->plugged_tries = 0; bcp 805 arch/x86/platform/uv/tlb_uv.c bcp->timeout_tries = 0; bcp 808 arch/x86/platform/uv/tlb_uv.c if (bcp->conseccompletes <= bcp->complete_threshold) bcp 816 arch/x86/platform/uv/tlb_uv.c struct bau_control *bcp, struct ptc_stats *stat, bcp 826 arch/x86/platform/uv/tlb_uv.c bcp->period_requests++; bcp 827 arch/x86/platform/uv/tlb_uv.c bcp->period_time += elapsed; bcp 828 arch/x86/platform/uv/tlb_uv.c if ((elapsed > usec_2_cycles(bcp->cong_response_us)) && bcp 829 arch/x86/platform/uv/tlb_uv.c (bcp->period_requests > bcp->cong_reps) && bcp 830 arch/x86/platform/uv/tlb_uv.c ((bcp->period_time / bcp->period_requests) > bcp 831 arch/x86/platform/uv/tlb_uv.c usec_2_cycles(bcp->cong_response_us))) { bcp 833 arch/x86/platform/uv/tlb_uv.c disable_for_period(bcp, stat); bcp 843 arch/x86/platform/uv/tlb_uv.c if (get_cycles() > bcp->period_end) bcp 844 arch/x86/platform/uv/tlb_uv.c bcp->period_giveups = 0; bcp 845 arch/x86/platform/uv/tlb_uv.c bcp->period_giveups++; bcp 846 arch/x86/platform/uv/tlb_uv.c if (bcp->period_giveups == 1) bcp 847 arch/x86/platform/uv/tlb_uv.c bcp->period_end = get_cycles() + bcp->disabled_period; bcp 848 arch/x86/platform/uv/tlb_uv.c if (bcp->period_giveups > bcp->giveup_limit) { bcp 849 arch/x86/platform/uv/tlb_uv.c disable_for_period(bcp, stat); bcp 877 arch/x86/platform/uv/tlb_uv.c struct bau_control *bcp, struct bau_control *hmaster, bcp 881 arch/x86/platform/uv/tlb_uv.c destination_plugged(bau_desc, bcp, hmaster, stat); bcp 883 arch/x86/platform/uv/tlb_uv.c destination_timeout(bau_desc, bcp, hmaster, stat); bcp 897 arch/x86/platform/uv/tlb_uv.c struct bau_control *bcp, bcp 907 arch/x86/platform/uv/tlb_uv.c struct ptc_stats *stat = bcp->statp; bcp 908 arch/x86/platform/uv/tlb_uv.c struct bau_control *hmaster = bcp->uvhub_master; bcp 912 arch/x86/platform/uv/tlb_uv.c if (bcp->uvhub_version == UV_BAU_V1) { bcp 933 arch/x86/platform/uv/tlb_uv.c seq_number = bcp->message_number++; bcp 946 arch/x86/platform/uv/tlb_uv.c index = (1UL << AS_PUSH_SHIFT) | bcp->uvhub_cpu; bcp 947 arch/x86/platform/uv/tlb_uv.c bcp->send_message = get_cycles(); bcp 952 arch/x86/platform/uv/tlb_uv.c completion_stat = ops.wait_completion(bau_desc, bcp, try); bcp 954 arch/x86/platform/uv/tlb_uv.c handle_cmplt(completion_stat, bau_desc, bcp, hmaster, stat); bcp 956 arch/x86/platform/uv/tlb_uv.c if (bcp->ipi_attempts >= bcp->ipi_reset_limit) { bcp 957 arch/x86/platform/uv/tlb_uv.c bcp->ipi_attempts = 0; bcp 968 arch/x86/platform/uv/tlb_uv.c count_max_concurr(completion_stat, bcp, hmaster); bcp 975 arch/x86/platform/uv/tlb_uv.c record_send_stats(time1, time2, bcp, stat, completion_stat, try); bcp 988 arch/x86/platform/uv/tlb_uv.c static int check_enable(struct bau_control *bcp, struct ptc_stats *stat) bcp 994 arch/x86/platform/uv/tlb_uv.c hmaster = bcp->uvhub_master; bcp 996 arch/x86/platform/uv/tlb_uv.c if (bcp->baudisabled && (get_cycles() >= bcp->set_bau_on_time)) { bcp 1048 arch/x86/platform/uv/tlb_uv.c static int set_distrib_bits(struct cpumask *flush_mask, struct bau_control *bcp, bcp 1063 arch/x86/platform/uv/tlb_uv.c hpp = &bcp->socket_master->thp[cpu]; bcp 1064 arch/x86/platform/uv/tlb_uv.c pnode = hpp->pnode - bcp->partition_base_pnode; bcp 1067 arch/x86/platform/uv/tlb_uv.c if (hpp->uvhub == bcp->uvhub) bcp 1110 arch/x86/platform/uv/tlb_uv.c struct bau_control *bcp; bcp 1113 arch/x86/platform/uv/tlb_uv.c bcp = &per_cpu(bau_control, cpu); bcp 1115 arch/x86/platform/uv/tlb_uv.c if (bcp->nobau) bcp 1118 arch/x86/platform/uv/tlb_uv.c stat = bcp->statp; bcp 1121 arch/x86/platform/uv/tlb_uv.c if (bcp->busy) { bcp 1124 arch/x86/platform/uv/tlb_uv.c status = ((descriptor_status >> (bcp->uvhub_cpu * bcp 1128 arch/x86/platform/uv/tlb_uv.c bcp->busy = 0; bcp 1132 arch/x86/platform/uv/tlb_uv.c if (bcp->baudisabled) { bcp 1133 arch/x86/platform/uv/tlb_uv.c if (check_enable(bcp, stat)) { bcp 1151 arch/x86/platform/uv/tlb_uv.c bau_desc = bcp->descriptor_base; bcp 1152 arch/x86/platform/uv/tlb_uv.c bau_desc += (ITEMS_PER_DESC * bcp->uvhub_cpu); bcp 1154 arch/x86/platform/uv/tlb_uv.c if (set_distrib_bits(flush_mask, bcp, bau_desc, &locals, &remotes)) bcp 1164 arch/x86/platform/uv/tlb_uv.c switch (bcp->uvhub_version) { bcp 1182 arch/x86/platform/uv/tlb_uv.c if (!uv_flush_send_and_wait(flush_mask, bcp, bau_desc)) bcp 1193 arch/x86/platform/uv/tlb_uv.c struct bau_control *bcp) bcp 1198 arch/x86/platform/uv/tlb_uv.c if (msg_next > bcp->queue_last) bcp 1199 arch/x86/platform/uv/tlb_uv.c msg_next = bcp->queue_first; bcp 1205 arch/x86/platform/uv/tlb_uv.c if (msg_next > bcp->queue_last) bcp 1206 arch/x86/platform/uv/tlb_uv.c msg_next = bcp->queue_first; bcp 1216 arch/x86/platform/uv/tlb_uv.c static void process_uv2_message(struct msg_desc *mdp, struct bau_control *bcp) bcp 1236 arch/x86/platform/uv/tlb_uv.c other_msg = find_another_by_swack(msg, bcp); bcp 1242 arch/x86/platform/uv/tlb_uv.c bau_process_message(mdp, bcp, 0); bcp 1256 arch/x86/platform/uv/tlb_uv.c bau_process_message(mdp, bcp, 1); bcp 1280 arch/x86/platform/uv/tlb_uv.c struct bau_control *bcp; bcp 1288 arch/x86/platform/uv/tlb_uv.c bcp = &per_cpu(bau_control, smp_processor_id()); bcp 1289 arch/x86/platform/uv/tlb_uv.c stat = bcp->statp; bcp 1291 arch/x86/platform/uv/tlb_uv.c msgdesc.queue_first = bcp->queue_first; bcp 1292 arch/x86/platform/uv/tlb_uv.c msgdesc.queue_last = bcp->queue_last; bcp 1294 arch/x86/platform/uv/tlb_uv.c msg = bcp->bau_msg_head; bcp 1300 arch/x86/platform/uv/tlb_uv.c if (bcp->uvhub_version == UV_BAU_V2) bcp 1301 arch/x86/platform/uv/tlb_uv.c process_uv2_message(&msgdesc, bcp); bcp 1304 arch/x86/platform/uv/tlb_uv.c bau_process_message(&msgdesc, bcp, 1); bcp 1309 arch/x86/platform/uv/tlb_uv.c bcp->bau_msg_head = msg; bcp 1399 arch/x86/platform/uv/tlb_uv.c struct bau_control *bcp; bcp 1418 arch/x86/platform/uv/tlb_uv.c bcp = &per_cpu(bau_control, cpu); bcp 1419 arch/x86/platform/uv/tlb_uv.c if (bcp->nobau) { bcp 1423 arch/x86/platform/uv/tlb_uv.c stat = bcp->statp; bcp 1427 arch/x86/platform/uv/tlb_uv.c cpu, bcp->nobau, stat->s_requestor, bcp 1558 arch/x86/platform/uv/tlb_uv.c static int parse_tunables_write(struct bau_control *bcp, char *instr, bcp 1592 arch/x86/platform/uv/tlb_uv.c if (val < 1 || val > bcp->cpus_in_uvhub) { bcp 1621 arch/x86/platform/uv/tlb_uv.c struct bau_control *bcp; bcp 1631 arch/x86/platform/uv/tlb_uv.c bcp = &per_cpu(bau_control, cpu); bcp 1632 arch/x86/platform/uv/tlb_uv.c ret = parse_tunables_write(bcp, instr, count); bcp 1638 arch/x86/platform/uv/tlb_uv.c bcp = &per_cpu(bau_control, cpu); bcp 1639 arch/x86/platform/uv/tlb_uv.c bcp->max_concurr = max_concurr; bcp 1640 arch/x86/platform/uv/tlb_uv.c bcp->max_concurr_const = max_concurr; bcp 1641 arch/x86/platform/uv/tlb_uv.c bcp->plugged_delay = plugged_delay; bcp 1642 arch/x86/platform/uv/tlb_uv.c bcp->plugsb4reset = plugsb4reset; bcp 1643 arch/x86/platform/uv/tlb_uv.c bcp->timeoutsb4reset = timeoutsb4reset; bcp 1644 arch/x86/platform/uv/tlb_uv.c bcp->ipi_reset_limit = ipi_reset_limit; bcp 1645 arch/x86/platform/uv/tlb_uv.c bcp->complete_threshold = complete_threshold; bcp 1646 arch/x86/platform/uv/tlb_uv.c bcp->cong_response_us = congested_respns_us; bcp 1647 arch/x86/platform/uv/tlb_uv.c bcp->cong_reps = congested_reps; bcp 1648 arch/x86/platform/uv/tlb_uv.c bcp->disabled_period = sec_2_cycles(disabled_period); bcp 1649 arch/x86/platform/uv/tlb_uv.c bcp->giveup_limit = giveup_limit; bcp 1723 arch/x86/platform/uv/tlb_uv.c struct bau_control *bcp; bcp 1784 arch/x86/platform/uv/tlb_uv.c bcp = &per_cpu(bau_control, cpu); bcp 1785 arch/x86/platform/uv/tlb_uv.c bcp->descriptor_base = bau_desc; bcp 1803 arch/x86/platform/uv/tlb_uv.c struct bau_control *bcp; bcp 1817 arch/x86/platform/uv/tlb_uv.c bcp = &per_cpu(bau_control, cpu); bcp 1818 arch/x86/platform/uv/tlb_uv.c bcp->queue_first = pqp; bcp 1819 arch/x86/platform/uv/tlb_uv.c bcp->bau_msg_head = pqp; bcp 1820 arch/x86/platform/uv/tlb_uv.c bcp->queue_last = pqp + (DEST_Q_SIZE - 1); bcp 1830 arch/x86/platform/uv/tlb_uv.c bcp = &per_cpu(bau_control, smp_processor_id()); bcp 1831 arch/x86/platform/uv/tlb_uv.c if (bcp->uvhub_version <= UV_BAU_V3) { bcp 1910 arch/x86/platform/uv/tlb_uv.c struct bau_control *bcp; bcp 1913 arch/x86/platform/uv/tlb_uv.c bcp = &per_cpu(bau_control, cpu); bcp 1914 arch/x86/platform/uv/tlb_uv.c bcp->baudisabled = 0; bcp 1916 arch/x86/platform/uv/tlb_uv.c bcp->nobau = true; bcp 1917 arch/x86/platform/uv/tlb_uv.c bcp->statp = &per_cpu(ptcstats, cpu); bcp 1919 arch/x86/platform/uv/tlb_uv.c bcp->timeout_interval = usec_2_cycles(2*timeout_us); bcp 1920 arch/x86/platform/uv/tlb_uv.c bcp->max_concurr = max_concurr; bcp 1921 arch/x86/platform/uv/tlb_uv.c bcp->max_concurr_const = max_concurr; bcp 1922 arch/x86/platform/uv/tlb_uv.c bcp->plugged_delay = plugged_delay; bcp 1923 arch/x86/platform/uv/tlb_uv.c bcp->plugsb4reset = plugsb4reset; bcp 1924 arch/x86/platform/uv/tlb_uv.c bcp->timeoutsb4reset = timeoutsb4reset; bcp 1925 arch/x86/platform/uv/tlb_uv.c bcp->ipi_reset_limit = ipi_reset_limit; bcp 1926 arch/x86/platform/uv/tlb_uv.c bcp->complete_threshold = complete_threshold; bcp 1927 arch/x86/platform/uv/tlb_uv.c bcp->cong_response_us = congested_respns_us; bcp 1928 arch/x86/platform/uv/tlb_uv.c bcp->cong_reps = congested_reps; bcp 1929 arch/x86/platform/uv/tlb_uv.c bcp->disabled_period = sec_2_cycles(disabled_period); bcp 1930 arch/x86/platform/uv/tlb_uv.c bcp->giveup_limit = giveup_limit; bcp 1931 arch/x86/platform/uv/tlb_uv.c spin_lock_init(&bcp->queue_lock); bcp 1932 arch/x86/platform/uv/tlb_uv.c spin_lock_init(&bcp->uvhub_lock); bcp 1933 arch/x86/platform/uv/tlb_uv.c spin_lock_init(&bcp->disable_lock); bcp 1948 arch/x86/platform/uv/tlb_uv.c struct bau_control *bcp; bcp 1953 arch/x86/platform/uv/tlb_uv.c bcp = &per_cpu(bau_control, cpu); bcp 1955 arch/x86/platform/uv/tlb_uv.c memset(bcp, 0, sizeof(struct bau_control)); bcp 1965 arch/x86/platform/uv/tlb_uv.c bcp->osnode = cpu_to_node(cpu); bcp 1966 arch/x86/platform/uv/tlb_uv.c bcp->partition_base_pnode = base_pnode; bcp 1978 arch/x86/platform/uv/tlb_uv.c socket = bcp->osnode & 1; bcp 2027 arch/x86/platform/uv/tlb_uv.c struct bau_control *bcp; bcp 2031 arch/x86/platform/uv/tlb_uv.c bcp = &per_cpu(bau_control, cpu); bcp 2032 arch/x86/platform/uv/tlb_uv.c bcp->cpu = cpu; bcp 2034 arch/x86/platform/uv/tlb_uv.c *smasterp = bcp; bcp 2036 arch/x86/platform/uv/tlb_uv.c *hmasterp = bcp; bcp 2038 arch/x86/platform/uv/tlb_uv.c bcp->cpus_in_uvhub = bdp->num_cpus; bcp 2039 arch/x86/platform/uv/tlb_uv.c bcp->cpus_in_socket = sdp->num_cpus; bcp 2040 arch/x86/platform/uv/tlb_uv.c bcp->socket_master = *smasterp; bcp 2041 arch/x86/platform/uv/tlb_uv.c bcp->uvhub = bdp->uvhub; bcp 2043 arch/x86/platform/uv/tlb_uv.c bcp->uvhub_version = UV_BAU_V1; bcp 2045 arch/x86/platform/uv/tlb_uv.c bcp->uvhub_version = UV_BAU_V2; bcp 2047 arch/x86/platform/uv/tlb_uv.c bcp->uvhub_version = UV_BAU_V3; bcp 2049 arch/x86/platform/uv/tlb_uv.c bcp->uvhub_version = UV_BAU_V4; bcp 2054 arch/x86/platform/uv/tlb_uv.c bcp->uvhub_master = *hmasterp; bcp 2056 arch/x86/platform/uv/tlb_uv.c bcp->uvhub_cpu = uvhub_cpu; bcp 2063 arch/x86/platform/uv/tlb_uv.c bcp->status_mmr = UVH_LB_BAU_SB_ACTIVATION_STATUS_0; bcp 2064 arch/x86/platform/uv/tlb_uv.c bcp->status_index = uvhub_cpu * UV_ACT_STATUS_SIZE; bcp 2066 arch/x86/platform/uv/tlb_uv.c bcp->status_mmr = UVH_LB_BAU_SB_ACTIVATION_STATUS_1; bcp 2067 arch/x86/platform/uv/tlb_uv.c bcp->status_index = (uvhub_cpu - UV_CPUS_PER_AS) bcp 2071 arch/x86/platform/uv/tlb_uv.c if (bcp->uvhub_cpu >= MAX_CPUS_PER_UVHUB) { bcp 2073 arch/x86/platform/uv/tlb_uv.c bcp->uvhub_cpu); bcp 1941 fs/xfs/xfs_log_recover.c struct xfs_buf_cancel *bcp; bcp 1956 fs/xfs/xfs_log_recover.c list_for_each_entry(bcp, bucket, bc_list) { bcp 1957 fs/xfs/xfs_log_recover.c if (bcp->bc_blkno == buf_f->blf_blkno && bcp 1958 fs/xfs/xfs_log_recover.c bcp->bc_len == buf_f->blf_len) { bcp 1959 fs/xfs/xfs_log_recover.c bcp->bc_refcount++; bcp 1965 fs/xfs/xfs_log_recover.c bcp = kmem_alloc(sizeof(struct xfs_buf_cancel), 0); bcp 1966 fs/xfs/xfs_log_recover.c bcp->bc_blkno = buf_f->blf_blkno; bcp 1967 fs/xfs/xfs_log_recover.c bcp->bc_len = buf_f->blf_len; bcp 1968 fs/xfs/xfs_log_recover.c bcp->bc_refcount = 1; bcp 1969 fs/xfs/xfs_log_recover.c list_add_tail(&bcp->bc_list, bucket); bcp 1988 fs/xfs/xfs_log_recover.c struct xfs_buf_cancel *bcp; bcp 1997 fs/xfs/xfs_log_recover.c list_for_each_entry(bcp, bucket, bc_list) { bcp 1998 fs/xfs/xfs_log_recover.c if (bcp->bc_blkno == blkno && bcp->bc_len == len) bcp 1999 fs/xfs/xfs_log_recover.c return bcp; bcp 2027 fs/xfs/xfs_log_recover.c struct xfs_buf_cancel *bcp; bcp 2029 fs/xfs/xfs_log_recover.c bcp = xlog_peek_buffer_cancelled(log, blkno, len, flags); bcp 2030 fs/xfs/xfs_log_recover.c if (!bcp) bcp 2040 fs/xfs/xfs_log_recover.c if (--bcp->bc_refcount == 0) { bcp 2041 fs/xfs/xfs_log_recover.c list_del(&bcp->bc_list); bcp 2042 fs/xfs/xfs_log_recover.c kmem_free(bcp); bcp 219 sound/soc/codecs/wm8804.c u16 format, master, bcp, lrp; bcp 261 sound/soc/codecs/wm8804.c bcp = lrp = 0; bcp 266 sound/soc/codecs/wm8804.c bcp = lrp = 1; bcp 269 sound/soc/codecs/wm8804.c bcp = 1; bcp 281 sound/soc/codecs/wm8804.c (bcp << 4) | (lrp << 5)); bcp 283 sound/soc/codecs/wm8804.c (bcp << 4) | (lrp << 5)); bcp 572 sound/soc/codecs/wm8983.c u16 format, master, bcp, lrp; bcp 621 sound/soc/codecs/wm8983.c bcp = lrp = 0; bcp 626 sound/soc/codecs/wm8983.c bcp = lrp = 1; bcp 629 sound/soc/codecs/wm8983.c bcp = 1; bcp 642 sound/soc/codecs/wm8983.c WM8983_BCP_MASK, bcp << WM8983_BCP_SHIFT); bcp 664 sound/soc/codecs/wm8985.c u16 format, master, bcp, lrp; bcp 721 sound/soc/codecs/wm8985.c bcp = lrp = 0; bcp 726 sound/soc/codecs/wm8985.c bcp = lrp = 1; bcp 729 sound/soc/codecs/wm8985.c bcp = 1; bcp 742 sound/soc/codecs/wm8985.c WM8985_BCP_MASK, bcp << WM8985_BCP_SHIFT);