bcp               140 arch/x86/platform/uv/tlb_uv.c 	struct bau_control *bcp;
bcp               148 arch/x86/platform/uv/tlb_uv.c 		bcp = &per_cpu(bau_control, cpu);
bcp               149 arch/x86/platform/uv/tlb_uv.c 		bcp->nobau = false;
bcp               159 arch/x86/platform/uv/tlb_uv.c 	struct bau_control *bcp;
bcp               163 arch/x86/platform/uv/tlb_uv.c 		bcp = &per_cpu(bau_control, cpu);
bcp               164 arch/x86/platform/uv/tlb_uv.c 		bcp->nobau = true;
bcp               207 arch/x86/platform/uv/tlb_uv.c static void reply_to_message(struct msg_desc *mdp, struct bau_control *bcp,
bcp               226 arch/x86/platform/uv/tlb_uv.c 					struct bau_control *bcp)
bcp               234 arch/x86/platform/uv/tlb_uv.c 	struct ptc_stats *stat = bcp->statp;
bcp               282 arch/x86/platform/uv/tlb_uv.c static void bau_process_message(struct msg_desc *mdp, struct bau_control *bcp,
bcp               288 arch/x86/platform/uv/tlb_uv.c 	struct ptc_stats *stat = bcp->statp;
bcp               290 arch/x86/platform/uv/tlb_uv.c 	struct bau_control *smaster = bcp->socket_master;
bcp               310 arch/x86/platform/uv/tlb_uv.c 	if (msg->msg_type == MSG_RETRY && bcp == bcp->uvhub_master)
bcp               311 arch/x86/platform/uv/tlb_uv.c 		bau_process_retry_msg(mdp, bcp);
bcp               322 arch/x86/platform/uv/tlb_uv.c 	if (socket_ack_count == bcp->cpus_in_socket) {
bcp               332 arch/x86/platform/uv/tlb_uv.c 		if (msg_ack_count == bcp->cpus_in_uvhub) {
bcp               337 arch/x86/platform/uv/tlb_uv.c 			reply_to_message(mdp, bcp, do_acknowledge);
bcp               373 arch/x86/platform/uv/tlb_uv.c 	struct bau_control *bcp = &per_cpu(bau_control, smp_processor_id());
bcp               376 arch/x86/platform/uv/tlb_uv.c 	struct ptc_stats *stat = bcp->statp;
bcp               385 arch/x86/platform/uv/tlb_uv.c 	for (msg = bcp->queue_first, i = 0; i < DEST_Q_SIZE; msg++, i++) {
bcp               419 arch/x86/platform/uv/tlb_uv.c static void reset_with_ipi(struct pnmask *distribution, struct bau_control *bcp)
bcp               424 arch/x86/platform/uv/tlb_uv.c 	int sender = bcp->cpu;
bcp               425 arch/x86/platform/uv/tlb_uv.c 	cpumask_t *mask = bcp->uvhub_master->cpumask;
bcp               426 arch/x86/platform/uv/tlb_uv.c 	struct bau_control *smaster = bcp->socket_master;
bcp               438 arch/x86/platform/uv/tlb_uv.c 		apnode = pnode + bcp->partition_base_pnode;
bcp               528 arch/x86/platform/uv/tlb_uv.c 				struct bau_control *bcp, long try)
bcp               532 arch/x86/platform/uv/tlb_uv.c 	u64 mmr_offset = bcp->status_mmr;
bcp               533 arch/x86/platform/uv/tlb_uv.c 	int right_shift = bcp->status_index;
bcp               534 arch/x86/platform/uv/tlb_uv.c 	struct ptc_stats *stat = bcp->statp;
bcp               558 arch/x86/platform/uv/tlb_uv.c 			if (cycles_2_us(ttm - bcp->send_message) < timeout_us) {
bcp               559 arch/x86/platform/uv/tlb_uv.c 				bcp->conseccompletes = 0;
bcp               563 arch/x86/platform/uv/tlb_uv.c 			bcp->conseccompletes = 0;
bcp               573 arch/x86/platform/uv/tlb_uv.c 	bcp->conseccompletes++;
bcp               591 arch/x86/platform/uv/tlb_uv.c static int handle_uv2_busy(struct bau_control *bcp)
bcp               593 arch/x86/platform/uv/tlb_uv.c 	struct ptc_stats *stat = bcp->statp;
bcp               596 arch/x86/platform/uv/tlb_uv.c 	bcp->busy = 1;
bcp               601 arch/x86/platform/uv/tlb_uv.c 				struct bau_control *bcp, long try)
bcp               605 arch/x86/platform/uv/tlb_uv.c 	u64 mmr_offset = bcp->status_mmr;
bcp               606 arch/x86/platform/uv/tlb_uv.c 	int right_shift = bcp->status_index;
bcp               607 arch/x86/platform/uv/tlb_uv.c 	int desc = bcp->uvhub_cpu;
bcp               609 arch/x86/platform/uv/tlb_uv.c 	struct ptc_stats *stat = bcp->statp;
bcp               637 arch/x86/platform/uv/tlb_uv.c 			if (cycles_2_us(ttm - bcp->send_message) < timeout_us) {
bcp               638 arch/x86/platform/uv/tlb_uv.c 				bcp->conseccompletes = 0;
bcp               644 arch/x86/platform/uv/tlb_uv.c 			bcp->conseccompletes = 0;
bcp               653 arch/x86/platform/uv/tlb_uv.c 				if ((ttm - bcp->send_message) > bcp->timeout_interval)
bcp               654 arch/x86/platform/uv/tlb_uv.c 					return handle_uv2_busy(bcp);
bcp               663 arch/x86/platform/uv/tlb_uv.c 	bcp->conseccompletes++;
bcp               682 arch/x86/platform/uv/tlb_uv.c 				struct bau_control *bcp, long try)
bcp               684 arch/x86/platform/uv/tlb_uv.c 	struct ptc_stats *stat = bcp->statp;
bcp               686 arch/x86/platform/uv/tlb_uv.c 	u64 mmr = bcp->status_mmr;
bcp               687 arch/x86/platform/uv/tlb_uv.c 	int index = bcp->status_index;
bcp               688 arch/x86/platform/uv/tlb_uv.c 	int desc = bcp->uvhub_cpu;
bcp               701 arch/x86/platform/uv/tlb_uv.c 			bcp->conseccompletes = 0;
bcp               706 arch/x86/platform/uv/tlb_uv.c 			bcp->conseccompletes = 0;
bcp               710 arch/x86/platform/uv/tlb_uv.c 			bcp->conseccompletes = 0;
bcp               719 arch/x86/platform/uv/tlb_uv.c 	bcp->conseccompletes++;
bcp               729 arch/x86/platform/uv/tlb_uv.c 			struct bau_control *bcp,
bcp               732 arch/x86/platform/uv/tlb_uv.c 	udelay(bcp->plugged_delay);
bcp               733 arch/x86/platform/uv/tlb_uv.c 	bcp->plugged_tries++;
bcp               735 arch/x86/platform/uv/tlb_uv.c 	if (bcp->plugged_tries >= bcp->plugsb4reset) {
bcp               736 arch/x86/platform/uv/tlb_uv.c 		bcp->plugged_tries = 0;
bcp               741 arch/x86/platform/uv/tlb_uv.c 		reset_with_ipi(&bau_desc->distribution, bcp);
bcp               746 arch/x86/platform/uv/tlb_uv.c 		bcp->ipi_attempts++;
bcp               752 arch/x86/platform/uv/tlb_uv.c 			struct bau_control *bcp, struct bau_control *hmaster,
bcp               756 arch/x86/platform/uv/tlb_uv.c 	bcp->timeout_tries++;
bcp               757 arch/x86/platform/uv/tlb_uv.c 	if (bcp->timeout_tries >= bcp->timeoutsb4reset) {
bcp               758 arch/x86/platform/uv/tlb_uv.c 		bcp->timeout_tries = 0;
bcp               763 arch/x86/platform/uv/tlb_uv.c 		reset_with_ipi(&bau_desc->distribution, bcp);
bcp               768 arch/x86/platform/uv/tlb_uv.c 		bcp->ipi_attempts++;
bcp               777 arch/x86/platform/uv/tlb_uv.c static void disable_for_period(struct bau_control *bcp, struct ptc_stats *stat)
bcp               784 arch/x86/platform/uv/tlb_uv.c 	hmaster = bcp->uvhub_master;
bcp               786 arch/x86/platform/uv/tlb_uv.c 	if (!bcp->baudisabled) {
bcp               794 arch/x86/platform/uv/tlb_uv.c 					tm1 + bcp->disabled_period;
bcp               801 arch/x86/platform/uv/tlb_uv.c static void count_max_concurr(int stat, struct bau_control *bcp,
bcp               804 arch/x86/platform/uv/tlb_uv.c 	bcp->plugged_tries = 0;
bcp               805 arch/x86/platform/uv/tlb_uv.c 	bcp->timeout_tries = 0;
bcp               808 arch/x86/platform/uv/tlb_uv.c 	if (bcp->conseccompletes <= bcp->complete_threshold)
bcp               816 arch/x86/platform/uv/tlb_uv.c 		struct bau_control *bcp, struct ptc_stats *stat,
bcp               826 arch/x86/platform/uv/tlb_uv.c 			bcp->period_requests++;
bcp               827 arch/x86/platform/uv/tlb_uv.c 			bcp->period_time += elapsed;
bcp               828 arch/x86/platform/uv/tlb_uv.c 			if ((elapsed > usec_2_cycles(bcp->cong_response_us)) &&
bcp               829 arch/x86/platform/uv/tlb_uv.c 			    (bcp->period_requests > bcp->cong_reps) &&
bcp               830 arch/x86/platform/uv/tlb_uv.c 			    ((bcp->period_time / bcp->period_requests) >
bcp               831 arch/x86/platform/uv/tlb_uv.c 					usec_2_cycles(bcp->cong_response_us))) {
bcp               833 arch/x86/platform/uv/tlb_uv.c 				disable_for_period(bcp, stat);
bcp               843 arch/x86/platform/uv/tlb_uv.c 		if (get_cycles() > bcp->period_end)
bcp               844 arch/x86/platform/uv/tlb_uv.c 			bcp->period_giveups = 0;
bcp               845 arch/x86/platform/uv/tlb_uv.c 		bcp->period_giveups++;
bcp               846 arch/x86/platform/uv/tlb_uv.c 		if (bcp->period_giveups == 1)
bcp               847 arch/x86/platform/uv/tlb_uv.c 			bcp->period_end = get_cycles() + bcp->disabled_period;
bcp               848 arch/x86/platform/uv/tlb_uv.c 		if (bcp->period_giveups > bcp->giveup_limit) {
bcp               849 arch/x86/platform/uv/tlb_uv.c 			disable_for_period(bcp, stat);
bcp               877 arch/x86/platform/uv/tlb_uv.c 			struct bau_control *bcp, struct bau_control *hmaster,
bcp               881 arch/x86/platform/uv/tlb_uv.c 		destination_plugged(bau_desc, bcp, hmaster, stat);
bcp               883 arch/x86/platform/uv/tlb_uv.c 		destination_timeout(bau_desc, bcp, hmaster, stat);
bcp               897 arch/x86/platform/uv/tlb_uv.c 				  struct bau_control *bcp,
bcp               907 arch/x86/platform/uv/tlb_uv.c 	struct ptc_stats *stat = bcp->statp;
bcp               908 arch/x86/platform/uv/tlb_uv.c 	struct bau_control *hmaster = bcp->uvhub_master;
bcp               912 arch/x86/platform/uv/tlb_uv.c 	if (bcp->uvhub_version == UV_BAU_V1) {
bcp               933 arch/x86/platform/uv/tlb_uv.c 			seq_number = bcp->message_number++;
bcp               946 arch/x86/platform/uv/tlb_uv.c 		index = (1UL << AS_PUSH_SHIFT) | bcp->uvhub_cpu;
bcp               947 arch/x86/platform/uv/tlb_uv.c 		bcp->send_message = get_cycles();
bcp               952 arch/x86/platform/uv/tlb_uv.c 		completion_stat = ops.wait_completion(bau_desc, bcp, try);
bcp               954 arch/x86/platform/uv/tlb_uv.c 		handle_cmplt(completion_stat, bau_desc, bcp, hmaster, stat);
bcp               956 arch/x86/platform/uv/tlb_uv.c 		if (bcp->ipi_attempts >= bcp->ipi_reset_limit) {
bcp               957 arch/x86/platform/uv/tlb_uv.c 			bcp->ipi_attempts = 0;
bcp               968 arch/x86/platform/uv/tlb_uv.c 	count_max_concurr(completion_stat, bcp, hmaster);
bcp               975 arch/x86/platform/uv/tlb_uv.c 	record_send_stats(time1, time2, bcp, stat, completion_stat, try);
bcp               988 arch/x86/platform/uv/tlb_uv.c static int check_enable(struct bau_control *bcp, struct ptc_stats *stat)
bcp               994 arch/x86/platform/uv/tlb_uv.c 	hmaster = bcp->uvhub_master;
bcp               996 arch/x86/platform/uv/tlb_uv.c 	if (bcp->baudisabled && (get_cycles() >= bcp->set_bau_on_time)) {
bcp              1048 arch/x86/platform/uv/tlb_uv.c static int set_distrib_bits(struct cpumask *flush_mask, struct bau_control *bcp,
bcp              1063 arch/x86/platform/uv/tlb_uv.c 		hpp = &bcp->socket_master->thp[cpu];
bcp              1064 arch/x86/platform/uv/tlb_uv.c 		pnode = hpp->pnode - bcp->partition_base_pnode;
bcp              1067 arch/x86/platform/uv/tlb_uv.c 		if (hpp->uvhub == bcp->uvhub)
bcp              1110 arch/x86/platform/uv/tlb_uv.c 	struct bau_control *bcp;
bcp              1113 arch/x86/platform/uv/tlb_uv.c 	bcp = &per_cpu(bau_control, cpu);
bcp              1115 arch/x86/platform/uv/tlb_uv.c 	if (bcp->nobau)
bcp              1118 arch/x86/platform/uv/tlb_uv.c 	stat = bcp->statp;
bcp              1121 arch/x86/platform/uv/tlb_uv.c 	if (bcp->busy) {
bcp              1124 arch/x86/platform/uv/tlb_uv.c 		status = ((descriptor_status >> (bcp->uvhub_cpu *
bcp              1128 arch/x86/platform/uv/tlb_uv.c 		bcp->busy = 0;
bcp              1132 arch/x86/platform/uv/tlb_uv.c 	if (bcp->baudisabled) {
bcp              1133 arch/x86/platform/uv/tlb_uv.c 		if (check_enable(bcp, stat)) {
bcp              1151 arch/x86/platform/uv/tlb_uv.c 	bau_desc = bcp->descriptor_base;
bcp              1152 arch/x86/platform/uv/tlb_uv.c 	bau_desc += (ITEMS_PER_DESC * bcp->uvhub_cpu);
bcp              1154 arch/x86/platform/uv/tlb_uv.c 	if (set_distrib_bits(flush_mask, bcp, bau_desc, &locals, &remotes))
bcp              1164 arch/x86/platform/uv/tlb_uv.c 	switch (bcp->uvhub_version) {
bcp              1182 arch/x86/platform/uv/tlb_uv.c 	if (!uv_flush_send_and_wait(flush_mask, bcp, bau_desc))
bcp              1193 arch/x86/platform/uv/tlb_uv.c 						  struct bau_control *bcp)
bcp              1198 arch/x86/platform/uv/tlb_uv.c 	if (msg_next > bcp->queue_last)
bcp              1199 arch/x86/platform/uv/tlb_uv.c 		msg_next = bcp->queue_first;
bcp              1205 arch/x86/platform/uv/tlb_uv.c 		if (msg_next > bcp->queue_last)
bcp              1206 arch/x86/platform/uv/tlb_uv.c 			msg_next = bcp->queue_first;
bcp              1216 arch/x86/platform/uv/tlb_uv.c static void process_uv2_message(struct msg_desc *mdp, struct bau_control *bcp)
bcp              1236 arch/x86/platform/uv/tlb_uv.c 		other_msg = find_another_by_swack(msg, bcp);
bcp              1242 arch/x86/platform/uv/tlb_uv.c 			bau_process_message(mdp, bcp, 0);
bcp              1256 arch/x86/platform/uv/tlb_uv.c 	bau_process_message(mdp, bcp, 1);
bcp              1280 arch/x86/platform/uv/tlb_uv.c 	struct bau_control *bcp;
bcp              1288 arch/x86/platform/uv/tlb_uv.c 	bcp = &per_cpu(bau_control, smp_processor_id());
bcp              1289 arch/x86/platform/uv/tlb_uv.c 	stat = bcp->statp;
bcp              1291 arch/x86/platform/uv/tlb_uv.c 	msgdesc.queue_first = bcp->queue_first;
bcp              1292 arch/x86/platform/uv/tlb_uv.c 	msgdesc.queue_last = bcp->queue_last;
bcp              1294 arch/x86/platform/uv/tlb_uv.c 	msg = bcp->bau_msg_head;
bcp              1300 arch/x86/platform/uv/tlb_uv.c 		if (bcp->uvhub_version == UV_BAU_V2)
bcp              1301 arch/x86/platform/uv/tlb_uv.c 			process_uv2_message(&msgdesc, bcp);
bcp              1304 arch/x86/platform/uv/tlb_uv.c 			bau_process_message(&msgdesc, bcp, 1);
bcp              1309 arch/x86/platform/uv/tlb_uv.c 		bcp->bau_msg_head = msg;
bcp              1399 arch/x86/platform/uv/tlb_uv.c 	struct bau_control *bcp;
bcp              1418 arch/x86/platform/uv/tlb_uv.c 		bcp = &per_cpu(bau_control, cpu);
bcp              1419 arch/x86/platform/uv/tlb_uv.c 		if (bcp->nobau) {
bcp              1423 arch/x86/platform/uv/tlb_uv.c 		stat = bcp->statp;
bcp              1427 arch/x86/platform/uv/tlb_uv.c 			   cpu, bcp->nobau, stat->s_requestor,
bcp              1558 arch/x86/platform/uv/tlb_uv.c static int parse_tunables_write(struct bau_control *bcp, char *instr,
bcp              1592 arch/x86/platform/uv/tlb_uv.c 			if (val < 1 || val > bcp->cpus_in_uvhub) {
bcp              1621 arch/x86/platform/uv/tlb_uv.c 	struct bau_control *bcp;
bcp              1631 arch/x86/platform/uv/tlb_uv.c 	bcp = &per_cpu(bau_control, cpu);
bcp              1632 arch/x86/platform/uv/tlb_uv.c 	ret = parse_tunables_write(bcp, instr, count);
bcp              1638 arch/x86/platform/uv/tlb_uv.c 		bcp = &per_cpu(bau_control, cpu);
bcp              1639 arch/x86/platform/uv/tlb_uv.c 		bcp->max_concurr         = max_concurr;
bcp              1640 arch/x86/platform/uv/tlb_uv.c 		bcp->max_concurr_const   = max_concurr;
bcp              1641 arch/x86/platform/uv/tlb_uv.c 		bcp->plugged_delay       = plugged_delay;
bcp              1642 arch/x86/platform/uv/tlb_uv.c 		bcp->plugsb4reset        = plugsb4reset;
bcp              1643 arch/x86/platform/uv/tlb_uv.c 		bcp->timeoutsb4reset     = timeoutsb4reset;
bcp              1644 arch/x86/platform/uv/tlb_uv.c 		bcp->ipi_reset_limit     = ipi_reset_limit;
bcp              1645 arch/x86/platform/uv/tlb_uv.c 		bcp->complete_threshold  = complete_threshold;
bcp              1646 arch/x86/platform/uv/tlb_uv.c 		bcp->cong_response_us    = congested_respns_us;
bcp              1647 arch/x86/platform/uv/tlb_uv.c 		bcp->cong_reps           = congested_reps;
bcp              1648 arch/x86/platform/uv/tlb_uv.c 		bcp->disabled_period     = sec_2_cycles(disabled_period);
bcp              1649 arch/x86/platform/uv/tlb_uv.c 		bcp->giveup_limit        = giveup_limit;
bcp              1723 arch/x86/platform/uv/tlb_uv.c 	struct bau_control *bcp;
bcp              1784 arch/x86/platform/uv/tlb_uv.c 		bcp = &per_cpu(bau_control, cpu);
bcp              1785 arch/x86/platform/uv/tlb_uv.c 		bcp->descriptor_base = bau_desc;
bcp              1803 arch/x86/platform/uv/tlb_uv.c 	struct bau_control *bcp;
bcp              1817 arch/x86/platform/uv/tlb_uv.c 		bcp = &per_cpu(bau_control, cpu);
bcp              1818 arch/x86/platform/uv/tlb_uv.c 		bcp->queue_first	= pqp;
bcp              1819 arch/x86/platform/uv/tlb_uv.c 		bcp->bau_msg_head	= pqp;
bcp              1820 arch/x86/platform/uv/tlb_uv.c 		bcp->queue_last		= pqp + (DEST_Q_SIZE - 1);
bcp              1830 arch/x86/platform/uv/tlb_uv.c 	bcp = &per_cpu(bau_control, smp_processor_id());
bcp              1831 arch/x86/platform/uv/tlb_uv.c 	if (bcp->uvhub_version <= UV_BAU_V3) {
bcp              1910 arch/x86/platform/uv/tlb_uv.c 	struct bau_control *bcp;
bcp              1913 arch/x86/platform/uv/tlb_uv.c 		bcp = &per_cpu(bau_control, cpu);
bcp              1914 arch/x86/platform/uv/tlb_uv.c 		bcp->baudisabled		= 0;
bcp              1916 arch/x86/platform/uv/tlb_uv.c 			bcp->nobau		= true;
bcp              1917 arch/x86/platform/uv/tlb_uv.c 		bcp->statp			= &per_cpu(ptcstats, cpu);
bcp              1919 arch/x86/platform/uv/tlb_uv.c 		bcp->timeout_interval		= usec_2_cycles(2*timeout_us);
bcp              1920 arch/x86/platform/uv/tlb_uv.c 		bcp->max_concurr		= max_concurr;
bcp              1921 arch/x86/platform/uv/tlb_uv.c 		bcp->max_concurr_const		= max_concurr;
bcp              1922 arch/x86/platform/uv/tlb_uv.c 		bcp->plugged_delay		= plugged_delay;
bcp              1923 arch/x86/platform/uv/tlb_uv.c 		bcp->plugsb4reset		= plugsb4reset;
bcp              1924 arch/x86/platform/uv/tlb_uv.c 		bcp->timeoutsb4reset		= timeoutsb4reset;
bcp              1925 arch/x86/platform/uv/tlb_uv.c 		bcp->ipi_reset_limit		= ipi_reset_limit;
bcp              1926 arch/x86/platform/uv/tlb_uv.c 		bcp->complete_threshold		= complete_threshold;
bcp              1927 arch/x86/platform/uv/tlb_uv.c 		bcp->cong_response_us		= congested_respns_us;
bcp              1928 arch/x86/platform/uv/tlb_uv.c 		bcp->cong_reps			= congested_reps;
bcp              1929 arch/x86/platform/uv/tlb_uv.c 		bcp->disabled_period		= sec_2_cycles(disabled_period);
bcp              1930 arch/x86/platform/uv/tlb_uv.c 		bcp->giveup_limit		= giveup_limit;
bcp              1931 arch/x86/platform/uv/tlb_uv.c 		spin_lock_init(&bcp->queue_lock);
bcp              1932 arch/x86/platform/uv/tlb_uv.c 		spin_lock_init(&bcp->uvhub_lock);
bcp              1933 arch/x86/platform/uv/tlb_uv.c 		spin_lock_init(&bcp->disable_lock);
bcp              1948 arch/x86/platform/uv/tlb_uv.c 	struct bau_control *bcp;
bcp              1953 arch/x86/platform/uv/tlb_uv.c 		bcp = &per_cpu(bau_control, cpu);
bcp              1955 arch/x86/platform/uv/tlb_uv.c 		memset(bcp, 0, sizeof(struct bau_control));
bcp              1965 arch/x86/platform/uv/tlb_uv.c 		bcp->osnode = cpu_to_node(cpu);
bcp              1966 arch/x86/platform/uv/tlb_uv.c 		bcp->partition_base_pnode = base_pnode;
bcp              1978 arch/x86/platform/uv/tlb_uv.c 		socket = bcp->osnode & 1;
bcp              2027 arch/x86/platform/uv/tlb_uv.c 	struct bau_control *bcp;
bcp              2031 arch/x86/platform/uv/tlb_uv.c 		bcp = &per_cpu(bau_control, cpu);
bcp              2032 arch/x86/platform/uv/tlb_uv.c 		bcp->cpu = cpu;
bcp              2034 arch/x86/platform/uv/tlb_uv.c 			*smasterp = bcp;
bcp              2036 arch/x86/platform/uv/tlb_uv.c 				*hmasterp = bcp;
bcp              2038 arch/x86/platform/uv/tlb_uv.c 		bcp->cpus_in_uvhub = bdp->num_cpus;
bcp              2039 arch/x86/platform/uv/tlb_uv.c 		bcp->cpus_in_socket = sdp->num_cpus;
bcp              2040 arch/x86/platform/uv/tlb_uv.c 		bcp->socket_master = *smasterp;
bcp              2041 arch/x86/platform/uv/tlb_uv.c 		bcp->uvhub = bdp->uvhub;
bcp              2043 arch/x86/platform/uv/tlb_uv.c 			bcp->uvhub_version = UV_BAU_V1;
bcp              2045 arch/x86/platform/uv/tlb_uv.c 			bcp->uvhub_version = UV_BAU_V2;
bcp              2047 arch/x86/platform/uv/tlb_uv.c 			bcp->uvhub_version = UV_BAU_V3;
bcp              2049 arch/x86/platform/uv/tlb_uv.c 			bcp->uvhub_version = UV_BAU_V4;
bcp              2054 arch/x86/platform/uv/tlb_uv.c 		bcp->uvhub_master = *hmasterp;
bcp              2056 arch/x86/platform/uv/tlb_uv.c 		bcp->uvhub_cpu = uvhub_cpu;
bcp              2063 arch/x86/platform/uv/tlb_uv.c 			bcp->status_mmr = UVH_LB_BAU_SB_ACTIVATION_STATUS_0;
bcp              2064 arch/x86/platform/uv/tlb_uv.c 			bcp->status_index = uvhub_cpu * UV_ACT_STATUS_SIZE;
bcp              2066 arch/x86/platform/uv/tlb_uv.c 			bcp->status_mmr = UVH_LB_BAU_SB_ACTIVATION_STATUS_1;
bcp              2067 arch/x86/platform/uv/tlb_uv.c 			bcp->status_index = (uvhub_cpu - UV_CPUS_PER_AS)
bcp              2071 arch/x86/platform/uv/tlb_uv.c 		if (bcp->uvhub_cpu >= MAX_CPUS_PER_UVHUB) {
bcp              2073 arch/x86/platform/uv/tlb_uv.c 				bcp->uvhub_cpu);
bcp              1941 fs/xfs/xfs_log_recover.c 	struct xfs_buf_cancel	*bcp;
bcp              1956 fs/xfs/xfs_log_recover.c 	list_for_each_entry(bcp, bucket, bc_list) {
bcp              1957 fs/xfs/xfs_log_recover.c 		if (bcp->bc_blkno == buf_f->blf_blkno &&
bcp              1958 fs/xfs/xfs_log_recover.c 		    bcp->bc_len == buf_f->blf_len) {
bcp              1959 fs/xfs/xfs_log_recover.c 			bcp->bc_refcount++;
bcp              1965 fs/xfs/xfs_log_recover.c 	bcp = kmem_alloc(sizeof(struct xfs_buf_cancel), 0);
bcp              1966 fs/xfs/xfs_log_recover.c 	bcp->bc_blkno = buf_f->blf_blkno;
bcp              1967 fs/xfs/xfs_log_recover.c 	bcp->bc_len = buf_f->blf_len;
bcp              1968 fs/xfs/xfs_log_recover.c 	bcp->bc_refcount = 1;
bcp              1969 fs/xfs/xfs_log_recover.c 	list_add_tail(&bcp->bc_list, bucket);
bcp              1988 fs/xfs/xfs_log_recover.c 	struct xfs_buf_cancel	*bcp;
bcp              1997 fs/xfs/xfs_log_recover.c 	list_for_each_entry(bcp, bucket, bc_list) {
bcp              1998 fs/xfs/xfs_log_recover.c 		if (bcp->bc_blkno == blkno && bcp->bc_len == len)
bcp              1999 fs/xfs/xfs_log_recover.c 			return bcp;
bcp              2027 fs/xfs/xfs_log_recover.c 	struct xfs_buf_cancel	*bcp;
bcp              2029 fs/xfs/xfs_log_recover.c 	bcp = xlog_peek_buffer_cancelled(log, blkno, len, flags);
bcp              2030 fs/xfs/xfs_log_recover.c 	if (!bcp)
bcp              2040 fs/xfs/xfs_log_recover.c 		if (--bcp->bc_refcount == 0) {
bcp              2041 fs/xfs/xfs_log_recover.c 			list_del(&bcp->bc_list);
bcp              2042 fs/xfs/xfs_log_recover.c 			kmem_free(bcp);
bcp               219 sound/soc/codecs/wm8804.c 	u16 format, master, bcp, lrp;
bcp               261 sound/soc/codecs/wm8804.c 	bcp = lrp = 0;
bcp               266 sound/soc/codecs/wm8804.c 		bcp = lrp = 1;
bcp               269 sound/soc/codecs/wm8804.c 		bcp = 1;
bcp               281 sound/soc/codecs/wm8804.c 			    (bcp << 4) | (lrp << 5));
bcp               283 sound/soc/codecs/wm8804.c 			    (bcp << 4) | (lrp << 5));
bcp               572 sound/soc/codecs/wm8983.c 	u16 format, master, bcp, lrp;
bcp               621 sound/soc/codecs/wm8983.c 	bcp = lrp = 0;
bcp               626 sound/soc/codecs/wm8983.c 		bcp = lrp = 1;
bcp               629 sound/soc/codecs/wm8983.c 		bcp = 1;
bcp               642 sound/soc/codecs/wm8983.c 			    WM8983_BCP_MASK, bcp << WM8983_BCP_SHIFT);
bcp               664 sound/soc/codecs/wm8985.c 	u16 format, master, bcp, lrp;
bcp               721 sound/soc/codecs/wm8985.c 	bcp = lrp = 0;
bcp               726 sound/soc/codecs/wm8985.c 		bcp = lrp = 1;
bcp               729 sound/soc/codecs/wm8985.c 		bcp = 1;
bcp               742 sound/soc/codecs/wm8985.c 			    WM8985_BCP_MASK, bcp << WM8985_BCP_SHIFT);