thr 3952 arch/mips/include/asm/octeon/cvmx-mio-defs.h uint64_t thr:8; thr 3954 arch/mips/include/asm/octeon/cvmx-mio-defs.h uint64_t thr:8; thr 4367 arch/mips/include/asm/octeon/cvmx-mio-defs.h uint64_t thr:8; thr 4369 arch/mips/include/asm/octeon/cvmx-mio-defs.h uint64_t thr:8; thr 254 arch/powerpc/kernel/process.c #define loadvec(thr) ((thr).load_vec) thr 334 arch/powerpc/kernel/process.c #define loadvec(thr) 0 thr 829 arch/powerpc/kernel/process.c static void tm_reclaim_thread(struct thread_struct *thr, uint8_t cause) thr 849 arch/powerpc/kernel/process.c giveup_all(container_of(thr, struct task_struct, thread)); thr 851 arch/powerpc/kernel/process.c tm_reclaim(thr, cause); thr 865 arch/powerpc/kernel/process.c if ((thr->ckpt_regs.msr & MSR_FP) == 0) thr 866 arch/powerpc/kernel/process.c memcpy(&thr->ckfp_state, &thr->fp_state, thr 868 arch/powerpc/kernel/process.c if ((thr->ckpt_regs.msr & MSR_VEC) == 0) thr 869 arch/powerpc/kernel/process.c memcpy(&thr->ckvr_state, &thr->vr_state, thr 891 arch/powerpc/kernel/process.c struct thread_struct *thr = &tsk->thread; thr 893 arch/powerpc/kernel/process.c if (!thr->regs) thr 896 arch/powerpc/kernel/process.c if (!MSR_TM_ACTIVE(thr->regs->msr)) thr 903 arch/powerpc/kernel/process.c tsk->pid, thr->regs->nip, thr 904 arch/powerpc/kernel/process.c thr->regs->ccr, thr->regs->msr, thr 905 arch/powerpc/kernel/process.c thr->regs->trap); thr 907 arch/powerpc/kernel/process.c tm_reclaim_thread(thr, TM_CAUSE_RESCHED); thr 918 arch/powerpc/kernel/process.c tm_save_sprs(thr); thr 740 arch/powerpc/kvm/book3s_hv.c int thr; thr 752 arch/powerpc/kvm/book3s_hv.c thr = vcpu->vcpu_id - vc->first_vcpuid; thr 753 arch/powerpc/kvm/book3s_hv.c return !!(vc->dpdes & (1 << thr)); thr 1163 arch/powerpc/kvm/book3s_hv.c int thr, cpu, pcpu, nthreads; thr 1170 arch/powerpc/kvm/book3s_hv.c for (thr = 0; thr < nthreads; ++thr, ++cpu) { thr 1183 arch/powerpc/kvm/book3s_hv.c dpdes |= 1 << thr; thr 1196 arch/powerpc/kvm/book3s_hv.c u32 inst, rb, thr; thr 1206 arch/powerpc/kvm/book3s_hv.c thr = vcpu->vcpu_id & (kvm->arch.emul_smt_mode - 1); thr 1215 arch/powerpc/kvm/book3s_hv.c tvcpu = kvmppc_find_vcpu(kvm, vcpu->vcpu_id - thr + arg); thr 1233 arch/powerpc/kvm/book3s_hv.c arg = thr; thr 2661 arch/powerpc/kvm/book3s_hv.c int thr; thr 2667 arch/powerpc/kvm/book3s_hv.c thr = 0; thr 2668 arch/powerpc/kvm/book3s_hv.c while (++thr < threads_per_subcore) thr 2669 arch/powerpc/kvm/book3s_hv.c if (cpu_online(cpu + thr)) thr 2673 arch/powerpc/kvm/book3s_hv.c for (thr = 1; thr < threads_per_subcore; ++thr) { thr 2674 arch/powerpc/kvm/book3s_hv.c if (kvmppc_grab_hwthread(cpu + thr)) { thr 2677 arch/powerpc/kvm/book3s_hv.c kvmppc_release_hwthread(cpu + thr); thr 2678 arch/powerpc/kvm/book3s_hv.c } while (--thr > 0); thr 3054 arch/powerpc/kvm/book3s_hv.c int pcpu, thr; thr 3199 arch/powerpc/kvm/book3s_hv.c for (thr = 0; thr < controlled_threads; ++thr) { thr 3200 arch/powerpc/kvm/book3s_hv.c struct paca_struct *paca = paca_ptrs[pcpu + thr]; thr 3202 arch/powerpc/kvm/book3s_hv.c paca->kvm_hstate.tid = thr; thr 3245 arch/powerpc/kvm/book3s_hv.c thr = is_power8 ? subcore_thread_map[sub] : sub; thr 3247 arch/powerpc/kvm/book3s_hv.c active |= 1 << thr; thr 3249 arch/powerpc/kvm/book3s_hv.c pvc->pcpu = pcpu + thr; thr 3256 arch/powerpc/kvm/book3s_hv.c active |= 1 << (thr + vcpu->arch.ptid); thr 3281 arch/powerpc/kvm/book3s_hv.c for (thr = 1; thr < threads_per_subcore; ++thr) thr 3282 arch/powerpc/kvm/book3s_hv.c if (!(active & (1 << thr))) thr 3283 arch/powerpc/kvm/book3s_hv.c kvmppc_ipi_thread(pcpu + thr); thr 3342 arch/powerpc/kvm/book3s_hv.c for (thr = 1; thr < controlled_threads; ++thr) { thr 3343 arch/powerpc/kvm/book3s_hv.c struct paca_struct *paca = paca_ptrs[pcpu + thr]; thr 692 arch/powerpc/kvm/book3s_hv_builtin.c int thr = local_paca->kvm_hstate.tid; thr 694 arch/powerpc/kvm/book3s_hv_builtin.c sip->lpcr_sync.phase[thr] |= phase; thr 887 arch/powerpc/platforms/powernv/idle.c int cpu, cpu0, thr; thr 894 arch/powerpc/platforms/powernv/idle.c for (thr = 0; thr < threads_per_core; ++thr) { thr 895 arch/powerpc/platforms/powernv/idle.c if (cpu != cpu0 + thr) thr 896 arch/powerpc/platforms/powernv/idle.c atomic_inc(&paca_ptrs[cpu0+thr]->dont_stop); thr 900 arch/powerpc/platforms/powernv/idle.c for (thr = 0; thr < threads_per_core; ++thr) { thr 901 arch/powerpc/platforms/powernv/idle.c if (!paca_ptrs[cpu0+thr]->requested_psscr) thr 904 arch/powerpc/platforms/powernv/idle.c poke_threads |= (1 << thr); thr 910 arch/powerpc/platforms/powernv/idle.c for (thr = 0; thr < threads_per_core; ++thr) { thr 911 arch/powerpc/platforms/powernv/idle.c if (poke_threads & (1 << thr)) { thr 914 arch/powerpc/platforms/powernv/idle.c paca_ptrs[cpu0+thr]->hw_cpu_id); thr 919 arch/powerpc/platforms/powernv/idle.c for (thr = 0; thr < threads_per_core; ++thr) { thr 920 arch/powerpc/platforms/powernv/idle.c if ((poke_threads & (1 << thr)) && thr 921 arch/powerpc/platforms/powernv/idle.c !paca_ptrs[cpu0+thr]->requested_psscr) { thr 923 arch/powerpc/platforms/powernv/idle.c poke_threads &= ~(1 << thr); thr 933 arch/powerpc/platforms/powernv/idle.c int cpu, cpu0, thr; thr 939 arch/powerpc/platforms/powernv/idle.c for (thr = 0; thr < threads_per_core; ++thr) { thr 940 arch/powerpc/platforms/powernv/idle.c if (cpu != cpu0 + thr) thr 941 arch/powerpc/platforms/powernv/idle.c atomic_dec(&paca_ptrs[cpu0+thr]->dont_stop); thr 875 arch/x86/events/intel/uncore_nhmex.c DEFINE_UNCORE_FORMAT_ATTR(thr, thr, "config1:0-31"); thr 202 drivers/ipack/devices/ipoctal.c iowrite8(value, &channel->regs->w.thr); thr 33 drivers/ipack/devices/scc2698.h u8 d3, thr; /* Transmit holding register */ thr 271 drivers/net/wireless/intel/iwlwifi/mvm/rx.c u16 thr; thr 305 drivers/net/wireless/intel/iwlwifi/mvm/rx.c thr = thresh_tpt[rate_n_flags & RATE_HT_MCS_RATE_CODE_MSK]; thr 306 drivers/net/wireless/intel/iwlwifi/mvm/rx.c thr *= 1 + ((rate_n_flags & RATE_HT_MCS_NSS_MSK) >> thr 312 drivers/net/wireless/intel/iwlwifi/mvm/rx.c thr = thresh_tpt[rate_n_flags & RATE_VHT_MCS_RATE_CODE_MSK]; thr 313 drivers/net/wireless/intel/iwlwifi/mvm/rx.c thr *= 1 + ((rate_n_flags & RATE_VHT_MCS_NSS_MSK) >> thr 317 drivers/net/wireless/intel/iwlwifi/mvm/rx.c thr <<= ((rate_n_flags & RATE_MCS_CHAN_WIDTH_MSK) >> thr 321 drivers/net/wireless/intel/iwlwifi/mvm/rx.c ewma_rate_add(&mdata->uapsd_nonagg_detect.rate, thr); thr 4308 drivers/net/wireless/marvell/mwifiex/cfg80211.c u32 thr, retry; thr 4451 drivers/net/wireless/marvell/mwifiex/cfg80211.c HostCmd_ACT_GEN_GET, FRAG_THRESH_I, &thr, true); thr 4452 drivers/net/wireless/marvell/mwifiex/cfg80211.c wiphy->frag_threshold = thr; thr 4454 drivers/net/wireless/marvell/mwifiex/cfg80211.c HostCmd_ACT_GEN_GET, RTS_THRESH_I, &thr, true); thr 4455 drivers/net/wireless/marvell/mwifiex/cfg80211.c wiphy->rts_threshold = thr; thr 1507 drivers/net/wireless/quantenna/qtnfmac/commands.c mac_info->frag_thr = le32_to_cpu(phy_thr->thr); thr 1511 drivers/net/wireless/quantenna/qtnfmac/commands.c mac_info->rts_thr = le32_to_cpu(phy_thr->thr); thr 1237 drivers/net/wireless/quantenna/qtnfmac/qlink.h __le32 thr; thr 140 drivers/tty/serial/atmel_serial.h #define ATMEL_US_TXFTHRES(thr) (((thr) & 0x3f) << 8) /* TX FIFO Threshold */ thr 141 drivers/tty/serial/atmel_serial.h #define ATMEL_US_RXFTHRES(thr) (((thr) & 0x3f) << 16) /* RX FIFO Threshold */ thr 142 drivers/tty/serial/atmel_serial.h #define ATMEL_US_RXFTHRES2(thr) (((thr) & 0x3f) << 24) /* RX FIFO Threshold2 */ thr 564 drivers/tty/serial/pch_uart.c unsigned int thr; thr 567 drivers/tty/serial/pch_uart.c thr = buf[i++]; thr 568 drivers/tty/serial/pch_uart.c iowrite8(thr, priv->membase + PCH_UART_THR); thr 4522 include/net/mac80211.h u32 thr); thr 577 kernel/power/swap.c struct task_struct *thr; /* thread */ thr 600 kernel/power/swap.c d->thr = NULL; thr 619 kernel/power/swap.c struct task_struct *thr; /* thread */ thr 643 kernel/power/swap.c d->thr = NULL; thr 678 kernel/power/swap.c unsigned thr, run_threads, nr_threads; thr 705 kernel/power/swap.c for (thr = 0; thr < nr_threads; thr++) thr 706 kernel/power/swap.c memset(&data[thr], 0, offsetof(struct cmp_data, go)); thr 719 kernel/power/swap.c for (thr = 0; thr < nr_threads; thr++) { thr 720 kernel/power/swap.c init_waitqueue_head(&data[thr].go); thr 721 kernel/power/swap.c init_waitqueue_head(&data[thr].done); thr 723 kernel/power/swap.c data[thr].thr = kthread_run(lzo_compress_threadfn, thr 724 kernel/power/swap.c &data[thr], thr 725 kernel/power/swap.c "image_compress/%u", thr); thr 726 kernel/power/swap.c if (IS_ERR(data[thr].thr)) { thr 727 kernel/power/swap.c data[thr].thr = NULL; thr 742 kernel/power/swap.c for (thr = 0; thr < nr_threads; thr++) { thr 743 kernel/power/swap.c crc->unc[thr] = data[thr].unc; thr 744 kernel/power/swap.c crc->unc_len[thr] = &data[thr].unc_len; thr 747 kernel/power/swap.c crc->thr = kthread_run(crc32_threadfn, crc, "image_crc32"); thr 748 kernel/power/swap.c if (IS_ERR(crc->thr)) { thr 749 kernel/power/swap.c crc->thr = NULL; thr 770 kernel/power/swap.c for (thr = 0; thr < nr_threads; thr++) { thr 779 kernel/power/swap.c memcpy(data[thr].unc + off, thr 790 kernel/power/swap.c data[thr].unc_len = off; thr 792 kernel/power/swap.c atomic_set(&data[thr].ready, 1); thr 793 kernel/power/swap.c wake_up(&data[thr].go); thr 796 kernel/power/swap.c if (!thr) thr 799 kernel/power/swap.c crc->run_threads = thr; thr 803 kernel/power/swap.c for (run_threads = thr, thr = 0; thr < run_threads; thr++) { thr 804 kernel/power/swap.c wait_event(data[thr].done, thr 805 kernel/power/swap.c atomic_read(&data[thr].stop)); thr 806 kernel/power/swap.c atomic_set(&data[thr].stop, 0); thr 808 kernel/power/swap.c ret = data[thr].ret; thr 815 kernel/power/swap.c if (unlikely(!data[thr].cmp_len || thr 816 kernel/power/swap.c data[thr].cmp_len > thr 817 kernel/power/swap.c lzo1x_worst_compress(data[thr].unc_len))) { thr 823 kernel/power/swap.c *(size_t *)data[thr].cmp = data[thr].cmp_len; thr 834 kernel/power/swap.c off < LZO_HEADER + data[thr].cmp_len; thr 836 kernel/power/swap.c memcpy(page, data[thr].cmp + off, PAGE_SIZE); thr 858 kernel/power/swap.c if (crc->thr) thr 859 kernel/power/swap.c kthread_stop(crc->thr); thr 863 kernel/power/swap.c for (thr = 0; thr < nr_threads; thr++) thr 864 kernel/power/swap.c if (data[thr].thr) thr 865 kernel/power/swap.c kthread_stop(data[thr].thr); thr 1104 kernel/power/swap.c struct task_struct *thr; /* thread */ thr 1127 kernel/power/swap.c d->thr = NULL; thr 1166 kernel/power/swap.c unsigned i, thr, run_threads, nr_threads; thr 1196 kernel/power/swap.c for (thr = 0; thr < nr_threads; thr++) thr 1197 kernel/power/swap.c memset(&data[thr], 0, offsetof(struct dec_data, go)); thr 1212 kernel/power/swap.c for (thr = 0; thr < nr_threads; thr++) { thr 1213 kernel/power/swap.c init_waitqueue_head(&data[thr].go); thr 1214 kernel/power/swap.c init_waitqueue_head(&data[thr].done); thr 1216 kernel/power/swap.c data[thr].thr = kthread_run(lzo_decompress_threadfn, thr 1217 kernel/power/swap.c &data[thr], thr 1218 kernel/power/swap.c "image_decompress/%u", thr); thr 1219 kernel/power/swap.c if (IS_ERR(data[thr].thr)) { thr 1220 kernel/power/swap.c data[thr].thr = NULL; thr 1235 kernel/power/swap.c for (thr = 0; thr < nr_threads; thr++) { thr 1236 kernel/power/swap.c crc->unc[thr] = data[thr].unc; thr 1237 kernel/power/swap.c crc->unc_len[thr] = &data[thr].unc_len; thr 1240 kernel/power/swap.c crc->thr = kthread_run(crc32_threadfn, crc, "image_crc32"); thr 1241 kernel/power/swap.c if (IS_ERR(crc->thr)) { thr 1242 kernel/power/swap.c crc->thr = NULL; thr 1335 kernel/power/swap.c for (thr = 0; have && thr < nr_threads; thr++) { thr 1336 kernel/power/swap.c data[thr].cmp_len = *(size_t *)page[pg]; thr 1337 kernel/power/swap.c if (unlikely(!data[thr].cmp_len || thr 1338 kernel/power/swap.c data[thr].cmp_len > thr 1345 kernel/power/swap.c need = DIV_ROUND_UP(data[thr].cmp_len + LZO_HEADER, thr 1356 kernel/power/swap.c off < LZO_HEADER + data[thr].cmp_len; thr 1358 kernel/power/swap.c memcpy(data[thr].cmp + off, thr 1366 kernel/power/swap.c atomic_set(&data[thr].ready, 1); thr 1367 kernel/power/swap.c wake_up(&data[thr].go); thr 1383 kernel/power/swap.c for (run_threads = thr, thr = 0; thr < run_threads; thr++) { thr 1384 kernel/power/swap.c wait_event(data[thr].done, thr 1385 kernel/power/swap.c atomic_read(&data[thr].stop)); thr 1386 kernel/power/swap.c atomic_set(&data[thr].stop, 0); thr 1388 kernel/power/swap.c ret = data[thr].ret; thr 1395 kernel/power/swap.c if (unlikely(!data[thr].unc_len || thr 1396 kernel/power/swap.c data[thr].unc_len > LZO_UNC_SIZE || thr 1397 kernel/power/swap.c data[thr].unc_len & (PAGE_SIZE - 1))) { thr 1404 kernel/power/swap.c off < data[thr].unc_len; off += PAGE_SIZE) { thr 1406 kernel/power/swap.c data[thr].unc + off, PAGE_SIZE); thr 1415 kernel/power/swap.c crc->run_threads = thr + 1; thr 1423 kernel/power/swap.c crc->run_threads = thr; thr 1453 kernel/power/swap.c if (crc->thr) thr 1454 kernel/power/swap.c kthread_stop(crc->thr); thr 1458 kernel/power/swap.c for (thr = 0; thr < nr_threads; thr++) thr 1459 kernel/power/swap.c if (data[thr].thr) thr 1460 kernel/power/swap.c kthread_stop(data[thr].thr); thr 2183 net/mac80211/sta_info.c u32 thr = 0; thr 2414 net/mac80211/sta_info.c thr = sta_get_expected_throughput(sta); thr 2416 net/mac80211/sta_info.c if (thr != 0) { thr 2418 net/mac80211/sta_info.c sinfo->expected_throughput = thr; thr 2448 net/mac80211/sta_info.c u32 thr = 0; thr 2455 net/mac80211/sta_info.c thr = ref->ops->get_expected_throughput(sta->rate_ctrl_priv); thr 2457 net/mac80211/sta_info.c thr = drv_get_expected_throughput(local, sta); thr 2459 net/mac80211/sta_info.c return thr; thr 2472 net/mac80211/sta_info.c static void sta_update_codel_params(struct sta_info *sta, u32 thr) thr 2477 net/mac80211/sta_info.c if (thr && thr < STA_SLOW_THRESHOLD * sta->local->num_sta) { thr 2489 net/mac80211/sta_info.c u32 thr) thr 2493 net/mac80211/sta_info.c sta_update_codel_params(sta, thr); thr 703 tools/perf/util/svghelper.c int thr; thr 709 tools/perf/util/svghelper.c for_each_set_bit(thr, cpumask_bits(&t->sib_thr[i]), nr_cpus) thr 710 tools/perf/util/svghelper.c if (map[thr] == -1) thr 711 tools/perf/util/svghelper.c map[thr] = (*pos)++; thr 74 tools/testing/selftests/futex/functional/futex_wait_private_mapped_file.c pthread_t thr; thr 102 tools/testing/selftests/futex/functional/futex_wait_private_mapped_file.c ret = pthread_create(&thr, NULL, thr_futex_wait, NULL); thr 120 tools/testing/selftests/futex/functional/futex_wait_private_mapped_file.c pthread_join(thr, NULL); thr 70 tools/testing/selftests/futex/functional/futex_wait_uninitialized_heap.c pthread_t thr; thr 104 tools/testing/selftests/futex/functional/futex_wait_uninitialized_heap.c ret = pthread_create(&thr, NULL, wait_thread, NULL); thr 72 tools/testing/selftests/kvm/x86_64/mmio_warning_test.c pthread_create(&th[i], NULL, thr, (void *)(uintptr_t)&tc);