thr              3952 arch/mips/include/asm/octeon/cvmx-mio-defs.h 		uint64_t thr:8;
thr              3954 arch/mips/include/asm/octeon/cvmx-mio-defs.h 		uint64_t thr:8;
thr              4367 arch/mips/include/asm/octeon/cvmx-mio-defs.h 		uint64_t thr:8;
thr              4369 arch/mips/include/asm/octeon/cvmx-mio-defs.h 		uint64_t thr:8;
thr               254 arch/powerpc/kernel/process.c #define loadvec(thr) ((thr).load_vec)
thr               334 arch/powerpc/kernel/process.c #define loadvec(thr) 0
thr               829 arch/powerpc/kernel/process.c static void tm_reclaim_thread(struct thread_struct *thr, uint8_t cause)
thr               849 arch/powerpc/kernel/process.c 	giveup_all(container_of(thr, struct task_struct, thread));
thr               851 arch/powerpc/kernel/process.c 	tm_reclaim(thr, cause);
thr               865 arch/powerpc/kernel/process.c 	if ((thr->ckpt_regs.msr & MSR_FP) == 0)
thr               866 arch/powerpc/kernel/process.c 		memcpy(&thr->ckfp_state, &thr->fp_state,
thr               868 arch/powerpc/kernel/process.c 	if ((thr->ckpt_regs.msr & MSR_VEC) == 0)
thr               869 arch/powerpc/kernel/process.c 		memcpy(&thr->ckvr_state, &thr->vr_state,
thr               891 arch/powerpc/kernel/process.c 	struct thread_struct *thr = &tsk->thread;
thr               893 arch/powerpc/kernel/process.c 	if (!thr->regs)
thr               896 arch/powerpc/kernel/process.c 	if (!MSR_TM_ACTIVE(thr->regs->msr))
thr               903 arch/powerpc/kernel/process.c 		 tsk->pid, thr->regs->nip,
thr               904 arch/powerpc/kernel/process.c 		 thr->regs->ccr, thr->regs->msr,
thr               905 arch/powerpc/kernel/process.c 		 thr->regs->trap);
thr               907 arch/powerpc/kernel/process.c 	tm_reclaim_thread(thr, TM_CAUSE_RESCHED);
thr               918 arch/powerpc/kernel/process.c 	tm_save_sprs(thr);
thr               740 arch/powerpc/kvm/book3s_hv.c 	int thr;
thr               752 arch/powerpc/kvm/book3s_hv.c 	thr = vcpu->vcpu_id - vc->first_vcpuid;
thr               753 arch/powerpc/kvm/book3s_hv.c 	return !!(vc->dpdes & (1 << thr));
thr              1163 arch/powerpc/kvm/book3s_hv.c 	int thr, cpu, pcpu, nthreads;
thr              1170 arch/powerpc/kvm/book3s_hv.c 	for (thr = 0; thr < nthreads; ++thr, ++cpu) {
thr              1183 arch/powerpc/kvm/book3s_hv.c 			dpdes |= 1 << thr;
thr              1196 arch/powerpc/kvm/book3s_hv.c 	u32 inst, rb, thr;
thr              1206 arch/powerpc/kvm/book3s_hv.c 	thr = vcpu->vcpu_id & (kvm->arch.emul_smt_mode - 1);
thr              1215 arch/powerpc/kvm/book3s_hv.c 		tvcpu = kvmppc_find_vcpu(kvm, vcpu->vcpu_id - thr + arg);
thr              1233 arch/powerpc/kvm/book3s_hv.c 			arg = thr;
thr              2661 arch/powerpc/kvm/book3s_hv.c 	int thr;
thr              2667 arch/powerpc/kvm/book3s_hv.c 	thr = 0;
thr              2668 arch/powerpc/kvm/book3s_hv.c 	while (++thr < threads_per_subcore)
thr              2669 arch/powerpc/kvm/book3s_hv.c 		if (cpu_online(cpu + thr))
thr              2673 arch/powerpc/kvm/book3s_hv.c 	for (thr = 1; thr < threads_per_subcore; ++thr) {
thr              2674 arch/powerpc/kvm/book3s_hv.c 		if (kvmppc_grab_hwthread(cpu + thr)) {
thr              2677 arch/powerpc/kvm/book3s_hv.c 				kvmppc_release_hwthread(cpu + thr);
thr              2678 arch/powerpc/kvm/book3s_hv.c 			} while (--thr > 0);
thr              3054 arch/powerpc/kvm/book3s_hv.c 	int pcpu, thr;
thr              3199 arch/powerpc/kvm/book3s_hv.c 	for (thr = 0; thr < controlled_threads; ++thr) {
thr              3200 arch/powerpc/kvm/book3s_hv.c 		struct paca_struct *paca = paca_ptrs[pcpu + thr];
thr              3202 arch/powerpc/kvm/book3s_hv.c 		paca->kvm_hstate.tid = thr;
thr              3245 arch/powerpc/kvm/book3s_hv.c 		thr = is_power8 ? subcore_thread_map[sub] : sub;
thr              3247 arch/powerpc/kvm/book3s_hv.c 		active |= 1 << thr;
thr              3249 arch/powerpc/kvm/book3s_hv.c 		pvc->pcpu = pcpu + thr;
thr              3256 arch/powerpc/kvm/book3s_hv.c 			active |= 1 << (thr + vcpu->arch.ptid);
thr              3281 arch/powerpc/kvm/book3s_hv.c 		for (thr = 1; thr < threads_per_subcore; ++thr)
thr              3282 arch/powerpc/kvm/book3s_hv.c 			if (!(active & (1 << thr)))
thr              3283 arch/powerpc/kvm/book3s_hv.c 				kvmppc_ipi_thread(pcpu + thr);
thr              3342 arch/powerpc/kvm/book3s_hv.c 		for (thr = 1; thr < controlled_threads; ++thr) {
thr              3343 arch/powerpc/kvm/book3s_hv.c 			struct paca_struct *paca = paca_ptrs[pcpu + thr];
thr               692 arch/powerpc/kvm/book3s_hv_builtin.c 	int thr = local_paca->kvm_hstate.tid;
thr               694 arch/powerpc/kvm/book3s_hv_builtin.c 	sip->lpcr_sync.phase[thr] |= phase;
thr               887 arch/powerpc/platforms/powernv/idle.c 	int cpu, cpu0, thr;
thr               894 arch/powerpc/platforms/powernv/idle.c 	for (thr = 0; thr < threads_per_core; ++thr) {
thr               895 arch/powerpc/platforms/powernv/idle.c 		if (cpu != cpu0 + thr)
thr               896 arch/powerpc/platforms/powernv/idle.c 			atomic_inc(&paca_ptrs[cpu0+thr]->dont_stop);
thr               900 arch/powerpc/platforms/powernv/idle.c 	for (thr = 0; thr < threads_per_core; ++thr) {
thr               901 arch/powerpc/platforms/powernv/idle.c 		if (!paca_ptrs[cpu0+thr]->requested_psscr)
thr               904 arch/powerpc/platforms/powernv/idle.c 			poke_threads |= (1 << thr);
thr               910 arch/powerpc/platforms/powernv/idle.c 		for (thr = 0; thr < threads_per_core; ++thr) {
thr               911 arch/powerpc/platforms/powernv/idle.c 			if (poke_threads & (1 << thr)) {
thr               914 arch/powerpc/platforms/powernv/idle.c 					   paca_ptrs[cpu0+thr]->hw_cpu_id);
thr               919 arch/powerpc/platforms/powernv/idle.c 			for (thr = 0; thr < threads_per_core; ++thr) {
thr               920 arch/powerpc/platforms/powernv/idle.c 				if ((poke_threads & (1 << thr)) &&
thr               921 arch/powerpc/platforms/powernv/idle.c 				    !paca_ptrs[cpu0+thr]->requested_psscr) {
thr               923 arch/powerpc/platforms/powernv/idle.c 					poke_threads &= ~(1 << thr);
thr               933 arch/powerpc/platforms/powernv/idle.c 	int cpu, cpu0, thr;
thr               939 arch/powerpc/platforms/powernv/idle.c 	for (thr = 0; thr < threads_per_core; ++thr) {
thr               940 arch/powerpc/platforms/powernv/idle.c 		if (cpu != cpu0 + thr)
thr               941 arch/powerpc/platforms/powernv/idle.c 			atomic_dec(&paca_ptrs[cpu0+thr]->dont_stop);
thr               875 arch/x86/events/intel/uncore_nhmex.c DEFINE_UNCORE_FORMAT_ATTR(thr,			thr,		"config1:0-31");
thr               202 drivers/ipack/devices/ipoctal.c 	iowrite8(value, &channel->regs->w.thr);
thr                33 drivers/ipack/devices/scc2698.h 		u8 d3, thr; /* Transmit holding register */
thr               271 drivers/net/wireless/intel/iwlwifi/mvm/rx.c 	u16 thr;
thr               305 drivers/net/wireless/intel/iwlwifi/mvm/rx.c 		thr = thresh_tpt[rate_n_flags & RATE_HT_MCS_RATE_CODE_MSK];
thr               306 drivers/net/wireless/intel/iwlwifi/mvm/rx.c 		thr *= 1 + ((rate_n_flags & RATE_HT_MCS_NSS_MSK) >>
thr               312 drivers/net/wireless/intel/iwlwifi/mvm/rx.c 		thr = thresh_tpt[rate_n_flags & RATE_VHT_MCS_RATE_CODE_MSK];
thr               313 drivers/net/wireless/intel/iwlwifi/mvm/rx.c 		thr *= 1 + ((rate_n_flags & RATE_VHT_MCS_NSS_MSK) >>
thr               317 drivers/net/wireless/intel/iwlwifi/mvm/rx.c 	thr <<= ((rate_n_flags & RATE_MCS_CHAN_WIDTH_MSK) >>
thr               321 drivers/net/wireless/intel/iwlwifi/mvm/rx.c 	ewma_rate_add(&mdata->uapsd_nonagg_detect.rate, thr);
thr              4308 drivers/net/wireless/marvell/mwifiex/cfg80211.c 	u32 thr, retry;
thr              4451 drivers/net/wireless/marvell/mwifiex/cfg80211.c 			 HostCmd_ACT_GEN_GET, FRAG_THRESH_I, &thr, true);
thr              4452 drivers/net/wireless/marvell/mwifiex/cfg80211.c 	wiphy->frag_threshold = thr;
thr              4454 drivers/net/wireless/marvell/mwifiex/cfg80211.c 			 HostCmd_ACT_GEN_GET, RTS_THRESH_I, &thr, true);
thr              4455 drivers/net/wireless/marvell/mwifiex/cfg80211.c 	wiphy->rts_threshold = thr;
thr              1507 drivers/net/wireless/quantenna/qtnfmac/commands.c 			mac_info->frag_thr = le32_to_cpu(phy_thr->thr);
thr              1511 drivers/net/wireless/quantenna/qtnfmac/commands.c 			mac_info->rts_thr = le32_to_cpu(phy_thr->thr);
thr              1237 drivers/net/wireless/quantenna/qtnfmac/qlink.h 	__le32 thr;
thr               140 drivers/tty/serial/atmel_serial.h #define	ATMEL_US_TXFTHRES(thr)	(((thr) & 0x3f) << 8)	/* TX FIFO Threshold */
thr               141 drivers/tty/serial/atmel_serial.h #define	ATMEL_US_RXFTHRES(thr)	(((thr) & 0x3f) << 16)	/* RX FIFO Threshold */
thr               142 drivers/tty/serial/atmel_serial.h #define	ATMEL_US_RXFTHRES2(thr)	(((thr) & 0x3f) << 24)	/* RX FIFO Threshold2 */
thr               564 drivers/tty/serial/pch_uart.c 	unsigned int thr;
thr               567 drivers/tty/serial/pch_uart.c 		thr = buf[i++];
thr               568 drivers/tty/serial/pch_uart.c 		iowrite8(thr, priv->membase + PCH_UART_THR);
thr              4522 include/net/mac80211.h 					   u32 thr);
thr               577 kernel/power/swap.c 	struct task_struct *thr;                  /* thread */
thr               600 kernel/power/swap.c 			d->thr = NULL;
thr               619 kernel/power/swap.c 	struct task_struct *thr;                  /* thread */
thr               643 kernel/power/swap.c 			d->thr = NULL;
thr               678 kernel/power/swap.c 	unsigned thr, run_threads, nr_threads;
thr               705 kernel/power/swap.c 	for (thr = 0; thr < nr_threads; thr++)
thr               706 kernel/power/swap.c 		memset(&data[thr], 0, offsetof(struct cmp_data, go));
thr               719 kernel/power/swap.c 	for (thr = 0; thr < nr_threads; thr++) {
thr               720 kernel/power/swap.c 		init_waitqueue_head(&data[thr].go);
thr               721 kernel/power/swap.c 		init_waitqueue_head(&data[thr].done);
thr               723 kernel/power/swap.c 		data[thr].thr = kthread_run(lzo_compress_threadfn,
thr               724 kernel/power/swap.c 		                            &data[thr],
thr               725 kernel/power/swap.c 		                            "image_compress/%u", thr);
thr               726 kernel/power/swap.c 		if (IS_ERR(data[thr].thr)) {
thr               727 kernel/power/swap.c 			data[thr].thr = NULL;
thr               742 kernel/power/swap.c 	for (thr = 0; thr < nr_threads; thr++) {
thr               743 kernel/power/swap.c 		crc->unc[thr] = data[thr].unc;
thr               744 kernel/power/swap.c 		crc->unc_len[thr] = &data[thr].unc_len;
thr               747 kernel/power/swap.c 	crc->thr = kthread_run(crc32_threadfn, crc, "image_crc32");
thr               748 kernel/power/swap.c 	if (IS_ERR(crc->thr)) {
thr               749 kernel/power/swap.c 		crc->thr = NULL;
thr               770 kernel/power/swap.c 		for (thr = 0; thr < nr_threads; thr++) {
thr               779 kernel/power/swap.c 				memcpy(data[thr].unc + off,
thr               790 kernel/power/swap.c 			data[thr].unc_len = off;
thr               792 kernel/power/swap.c 			atomic_set(&data[thr].ready, 1);
thr               793 kernel/power/swap.c 			wake_up(&data[thr].go);
thr               796 kernel/power/swap.c 		if (!thr)
thr               799 kernel/power/swap.c 		crc->run_threads = thr;
thr               803 kernel/power/swap.c 		for (run_threads = thr, thr = 0; thr < run_threads; thr++) {
thr               804 kernel/power/swap.c 			wait_event(data[thr].done,
thr               805 kernel/power/swap.c 			           atomic_read(&data[thr].stop));
thr               806 kernel/power/swap.c 			atomic_set(&data[thr].stop, 0);
thr               808 kernel/power/swap.c 			ret = data[thr].ret;
thr               815 kernel/power/swap.c 			if (unlikely(!data[thr].cmp_len ||
thr               816 kernel/power/swap.c 			             data[thr].cmp_len >
thr               817 kernel/power/swap.c 			             lzo1x_worst_compress(data[thr].unc_len))) {
thr               823 kernel/power/swap.c 			*(size_t *)data[thr].cmp = data[thr].cmp_len;
thr               834 kernel/power/swap.c 			     off < LZO_HEADER + data[thr].cmp_len;
thr               836 kernel/power/swap.c 				memcpy(page, data[thr].cmp + off, PAGE_SIZE);
thr               858 kernel/power/swap.c 		if (crc->thr)
thr               859 kernel/power/swap.c 			kthread_stop(crc->thr);
thr               863 kernel/power/swap.c 		for (thr = 0; thr < nr_threads; thr++)
thr               864 kernel/power/swap.c 			if (data[thr].thr)
thr               865 kernel/power/swap.c 				kthread_stop(data[thr].thr);
thr              1104 kernel/power/swap.c 	struct task_struct *thr;                  /* thread */
thr              1127 kernel/power/swap.c 			d->thr = NULL;
thr              1166 kernel/power/swap.c 	unsigned i, thr, run_threads, nr_threads;
thr              1196 kernel/power/swap.c 	for (thr = 0; thr < nr_threads; thr++)
thr              1197 kernel/power/swap.c 		memset(&data[thr], 0, offsetof(struct dec_data, go));
thr              1212 kernel/power/swap.c 	for (thr = 0; thr < nr_threads; thr++) {
thr              1213 kernel/power/swap.c 		init_waitqueue_head(&data[thr].go);
thr              1214 kernel/power/swap.c 		init_waitqueue_head(&data[thr].done);
thr              1216 kernel/power/swap.c 		data[thr].thr = kthread_run(lzo_decompress_threadfn,
thr              1217 kernel/power/swap.c 		                            &data[thr],
thr              1218 kernel/power/swap.c 		                            "image_decompress/%u", thr);
thr              1219 kernel/power/swap.c 		if (IS_ERR(data[thr].thr)) {
thr              1220 kernel/power/swap.c 			data[thr].thr = NULL;
thr              1235 kernel/power/swap.c 	for (thr = 0; thr < nr_threads; thr++) {
thr              1236 kernel/power/swap.c 		crc->unc[thr] = data[thr].unc;
thr              1237 kernel/power/swap.c 		crc->unc_len[thr] = &data[thr].unc_len;
thr              1240 kernel/power/swap.c 	crc->thr = kthread_run(crc32_threadfn, crc, "image_crc32");
thr              1241 kernel/power/swap.c 	if (IS_ERR(crc->thr)) {
thr              1242 kernel/power/swap.c 		crc->thr = NULL;
thr              1335 kernel/power/swap.c 		for (thr = 0; have && thr < nr_threads; thr++) {
thr              1336 kernel/power/swap.c 			data[thr].cmp_len = *(size_t *)page[pg];
thr              1337 kernel/power/swap.c 			if (unlikely(!data[thr].cmp_len ||
thr              1338 kernel/power/swap.c 			             data[thr].cmp_len >
thr              1345 kernel/power/swap.c 			need = DIV_ROUND_UP(data[thr].cmp_len + LZO_HEADER,
thr              1356 kernel/power/swap.c 			     off < LZO_HEADER + data[thr].cmp_len;
thr              1358 kernel/power/swap.c 				memcpy(data[thr].cmp + off,
thr              1366 kernel/power/swap.c 			atomic_set(&data[thr].ready, 1);
thr              1367 kernel/power/swap.c 			wake_up(&data[thr].go);
thr              1383 kernel/power/swap.c 		for (run_threads = thr, thr = 0; thr < run_threads; thr++) {
thr              1384 kernel/power/swap.c 			wait_event(data[thr].done,
thr              1385 kernel/power/swap.c 			           atomic_read(&data[thr].stop));
thr              1386 kernel/power/swap.c 			atomic_set(&data[thr].stop, 0);
thr              1388 kernel/power/swap.c 			ret = data[thr].ret;
thr              1395 kernel/power/swap.c 			if (unlikely(!data[thr].unc_len ||
thr              1396 kernel/power/swap.c 			             data[thr].unc_len > LZO_UNC_SIZE ||
thr              1397 kernel/power/swap.c 			             data[thr].unc_len & (PAGE_SIZE - 1))) {
thr              1404 kernel/power/swap.c 			     off < data[thr].unc_len; off += PAGE_SIZE) {
thr              1406 kernel/power/swap.c 				       data[thr].unc + off, PAGE_SIZE);
thr              1415 kernel/power/swap.c 					crc->run_threads = thr + 1;
thr              1423 kernel/power/swap.c 		crc->run_threads = thr;
thr              1453 kernel/power/swap.c 		if (crc->thr)
thr              1454 kernel/power/swap.c 			kthread_stop(crc->thr);
thr              1458 kernel/power/swap.c 		for (thr = 0; thr < nr_threads; thr++)
thr              1459 kernel/power/swap.c 			if (data[thr].thr)
thr              1460 kernel/power/swap.c 				kthread_stop(data[thr].thr);
thr              2183 net/mac80211/sta_info.c 	u32 thr = 0;
thr              2414 net/mac80211/sta_info.c 	thr = sta_get_expected_throughput(sta);
thr              2416 net/mac80211/sta_info.c 	if (thr != 0) {
thr              2418 net/mac80211/sta_info.c 		sinfo->expected_throughput = thr;
thr              2448 net/mac80211/sta_info.c 	u32 thr = 0;
thr              2455 net/mac80211/sta_info.c 		thr = ref->ops->get_expected_throughput(sta->rate_ctrl_priv);
thr              2457 net/mac80211/sta_info.c 		thr = drv_get_expected_throughput(local, sta);
thr              2459 net/mac80211/sta_info.c 	return thr;
thr              2472 net/mac80211/sta_info.c static void sta_update_codel_params(struct sta_info *sta, u32 thr)
thr              2477 net/mac80211/sta_info.c 	if (thr && thr < STA_SLOW_THRESHOLD * sta->local->num_sta) {
thr              2489 net/mac80211/sta_info.c 					   u32 thr)
thr              2493 net/mac80211/sta_info.c 	sta_update_codel_params(sta, thr);
thr               703 tools/perf/util/svghelper.c 	int thr;
thr               709 tools/perf/util/svghelper.c 		for_each_set_bit(thr, cpumask_bits(&t->sib_thr[i]), nr_cpus)
thr               710 tools/perf/util/svghelper.c 			if (map[thr] == -1)
thr               711 tools/perf/util/svghelper.c 				map[thr] = (*pos)++;
thr                74 tools/testing/selftests/futex/functional/futex_wait_private_mapped_file.c 	pthread_t thr;
thr               102 tools/testing/selftests/futex/functional/futex_wait_private_mapped_file.c 	ret = pthread_create(&thr, NULL, thr_futex_wait, NULL);
thr               120 tools/testing/selftests/futex/functional/futex_wait_private_mapped_file.c 	pthread_join(thr, NULL);
thr                70 tools/testing/selftests/futex/functional/futex_wait_uninitialized_heap.c 	pthread_t thr;
thr               104 tools/testing/selftests/futex/functional/futex_wait_uninitialized_heap.c 	ret = pthread_create(&thr, NULL, wait_thread, NULL);
thr                72 tools/testing/selftests/kvm/x86_64/mmio_warning_test.c 		pthread_create(&th[i], NULL, thr, (void *)(uintptr_t)&tc);