/linux-4.4.14/tools/perf/util/ |
D | thread_map.c | 52 struct thread_map *threads; in thread_map__new_by_pid() local 63 threads = thread_map__alloc(items); in thread_map__new_by_pid() 64 if (threads != NULL) { in thread_map__new_by_pid() 66 thread_map__set_pid(threads, i, atoi(namelist[i]->d_name)); in thread_map__new_by_pid() 67 threads->nr = items; in thread_map__new_by_pid() 68 atomic_set(&threads->refcnt, 1); in thread_map__new_by_pid() 75 return threads; in thread_map__new_by_pid() 80 struct thread_map *threads = thread_map__alloc(1); in thread_map__new_by_tid() local 82 if (threads != NULL) { in thread_map__new_by_tid() 83 thread_map__set_pid(threads, 0, tid); in thread_map__new_by_tid() [all …]
|
D | thread_map.h | 31 size_t thread_map__fprintf(struct thread_map *threads, FILE *fp); 33 static inline int thread_map__nr(struct thread_map *threads) in thread_map__nr() argument 35 return threads ? threads->nr : 1; in thread_map__nr() 54 void thread_map__read_comms(struct thread_map *threads);
|
D | values.c | 14 values->threads = 0; in perf_read_values_init() 33 for (i = 0; i < values->threads; i++) in perf_read_values_destroy() 62 for (i = 0; i < values->threads; i++) in perf_read_values__findnew_thread() 66 if (values->threads == values->threads_max) in perf_read_values__findnew_thread() 69 i = values->threads++; in perf_read_values__findnew_thread() 91 for (i = 0; i < values->threads; i++) { in perf_read_values__enlarge_counters() 144 for (i = 0; i < values->threads; i++) { in perf_read_values__display_pretty() 165 for (i = 0; i < values->threads; i++) { in perf_read_values__display_pretty() 188 for (i = 0; i < values->threads; i++) { in perf_read_values__display_raw() 204 for (i = 0; i < values->threads; i++) { in perf_read_values__display_raw() [all …]
|
D | evlist.c | 37 struct thread_map *threads) in perf_evlist__init() argument 44 perf_evlist__set_maps(evlist, cpus, threads); in perf_evlist__init() 120 thread_map__put(evlist->threads); in perf_evlist__delete() 122 evlist->threads = NULL; in perf_evlist__delete() 143 thread_map__put(evsel->threads); in __perf_evlist__propagate_maps() 144 evsel->threads = thread_map__get(evlist->threads); in __perf_evlist__propagate_maps() 334 return thread_map__nr(evlist->threads); in perf_evlist__nr_threads() 476 int nr_threads = thread_map__nr(evlist->threads); in perf_evlist__alloc_pollfd() 604 if (!evsel->system_wide && evlist->threads && thread >= 0) in perf_evlist__set_sid_idx() 605 sid->tid = thread_map__pid(evlist->threads, thread); in perf_evlist__set_sid_idx() [all …]
|
D | python.c | 481 struct thread_map *threads; member 494 pthreads->threads = thread_map__new(pid, tid, uid); in pyrf_thread_map__init() 495 if (pthreads->threads == NULL) in pyrf_thread_map__init() 502 thread_map__put(pthreads->threads); in pyrf_thread_map__delete() 510 return pthreads->threads->nr; in pyrf_thread_map__length() 517 if (i >= pthreads->threads->nr) in pyrf_thread_map__item() 520 return Py_BuildValue("i", pthreads->threads->map[i]); in pyrf_thread_map__item() 672 struct thread_map *threads = NULL; in pyrf_evsel__open() local 682 threads = ((struct pyrf_thread_map *)pthreads)->threads; in pyrf_evsel__open() 692 if (perf_evsel__open(evsel, cpus, threads) < 0) { in pyrf_evsel__open() [all …]
|
D | machine.h | 32 struct rb_root threads; member 233 struct target *target, struct thread_map *threads, 238 struct thread_map *threads, bool data_mmap, in machine__synthesize_threads() argument 241 return __machine__synthesize_threads(machine, NULL, target, threads, in machine__synthesize_threads()
|
D | evlist.h | 56 struct thread_map *threads; member 71 struct thread_map *threads); 163 struct thread_map *threads);
|
D | values.h | 7 int threads; member
|
D | evsel.h | 104 struct thread_map *threads; member 235 struct thread_map *threads); 237 struct thread_map *threads);
|
D | evsel.c | 1060 thread_map__put(evsel->threads); in perf_evsel__exit() 1292 struct thread_map *threads) in __perf_evsel__open() argument 1302 nthreads = threads->nr; in __perf_evsel__open() 1343 pid = thread_map__pid(threads, thread); in __perf_evsel__open() 1476 int threads[1]; member 1479 .threads = { -1, }, 1483 struct thread_map *threads) in perf_evsel__open() argument 1490 if (threads == NULL) in perf_evsel__open() 1491 threads = &empty_thread_map.map; in perf_evsel__open() 1493 return __perf_evsel__open(evsel, cpus, threads); in perf_evsel__open() [all …]
|
D | stat.c | 146 int nthreads = thread_map__nr(evsel->threads); in perf_evsel__alloc_stats() 286 int nthreads = thread_map__nr(counter->threads); in process_counter_maps()
|
D | event.c | 485 struct thread_map *threads, in perf_event__synthesize_thread_map() argument 507 for (thread = 0; thread < threads->nr; ++thread) { in perf_event__synthesize_thread_map() 510 thread_map__pid(threads, thread), 0, in perf_event__synthesize_thread_map() 521 if ((int) comm_event->comm.pid != thread_map__pid(threads, thread)) { in perf_event__synthesize_thread_map() 525 for (j = 0; j < threads->nr; ++j) { in perf_event__synthesize_thread_map() 526 if ((int) comm_event->comm.pid == thread_map__pid(threads, j)) { in perf_event__synthesize_thread_map()
|
D | machine.c | 32 machine->threads = RB_ROOT; in machine__init() 113 nd = rb_first(&machine->threads); in machine__delete_threads() 362 struct rb_node **p = &machine->threads.rb_node; in ____machine__findnew_thread() 403 rb_insert_color(&th->rb_node, &machine->threads); in ____machine__findnew_thread() 414 rb_erase_init(&th->rb_node, &machine->threads); in ____machine__findnew_thread() 651 for (nd = rb_first(&machine->threads); nd; nd = rb_next(nd)) { in machine__fprintf() 1371 rb_erase_init(&th->rb_node, &machine->threads); in __machine__remove_thread() 1906 for (nd = rb_first(&machine->threads); nd; nd = rb_next(nd)) { in machine__for_each_thread() 1943 struct target *target, struct thread_map *threads, in __machine__synthesize_threads() argument 1948 …return perf_event__synthesize_thread_map(tool, threads, process, machine, data_mmap, proc_map_time… in __machine__synthesize_threads()
|
D | auxtrace.c | 124 if (evlist->threads) in auxtrace_mmap_params__set_idx() 125 mp->tid = thread_map__pid(evlist->threads, 0); in auxtrace_mmap_params__set_idx() 130 mp->tid = thread_map__pid(evlist->threads, idx); in auxtrace_mmap_params__set_idx()
|
D | event.h | 394 struct thread_map *threads,
|
D | parse-events.c | 1768 int threads[1]; in is_event_supported() member 1771 .threads = { 0 }, in is_event_supported()
|
/linux-4.4.14/tools/perf/tests/ |
D | openat-syscall.c | 13 struct thread_map *threads = thread_map__new(-1, getpid(), UINT_MAX); in test__openat_syscall_event() local 17 if (threads == NULL) { in test__openat_syscall_event() 29 if (perf_evsel__open_per_thread(evsel, threads) < 0) { in test__openat_syscall_event() 54 perf_evsel__close_fd(evsel, 1, threads->nr); in test__openat_syscall_event() 58 thread_map__put(threads); in test__openat_syscall_event()
|
D | task-exit.c | 47 struct thread_map *threads; in test__task_exit() local 64 threads = thread_map__new_by_tid(-1); in test__task_exit() 65 if (!cpus || !threads) { in test__task_exit() 71 perf_evlist__set_maps(evlist, cpus, threads); in test__task_exit() 74 threads = NULL; in test__task_exit() 126 thread_map__put(threads); in test__task_exit()
|
D | sw-clock.c | 38 struct thread_map *threads; in __test__sw_clock_freq() local 56 threads = thread_map__new_by_tid(getpid()); in __test__sw_clock_freq() 57 if (!cpus || !threads) { in __test__sw_clock_freq() 63 perf_evlist__set_maps(evlist, cpus, threads); in __test__sw_clock_freq() 66 threads = NULL; in __test__sw_clock_freq() 119 thread_map__put(threads); in __test__sw_clock_freq()
|
D | mmap-basic.c | 23 struct thread_map *threads; in test__basic_mmap() local 35 threads = thread_map__new(-1, getpid(), UINT_MAX); in test__basic_mmap() 36 if (threads == NULL) { in test__basic_mmap() 62 perf_evlist__set_maps(evlist, cpus, threads); in test__basic_mmap() 79 if (perf_evsel__open(evsels[i], cpus, threads) < 0) { in test__basic_mmap() 142 threads = NULL; in test__basic_mmap() 146 thread_map__put(threads); in test__basic_mmap()
|
D | openat-syscall-all-cpus.c | 17 struct thread_map *threads = thread_map__new(-1, getpid(), UINT_MAX); in test__openat_syscall_event_on_all_cpus() local 21 if (threads == NULL) { in test__openat_syscall_event_on_all_cpus() 41 if (perf_evsel__open(evsel, cpus, threads) < 0) { in test__openat_syscall_event_on_all_cpus() 109 perf_evsel__close_fd(evsel, 1, threads->nr); in test__openat_syscall_event_on_all_cpus() 113 thread_map__put(threads); in test__openat_syscall_event_on_all_cpus()
|
D | mmap-thread-lookup.c | 26 static struct thread_data threads[THREADS]; variable 75 struct thread_data *td = &threads[i]; in thread_create() 95 struct thread_data *td0 = &threads[0]; in threads_create() 112 struct thread_data *td0 = &threads[0]; in threads_destroy() 121 err = pthread_join(threads[i].pt, NULL); in threads_destroy() 182 struct thread_data *td = &threads[i]; in mmap_events()
|
D | keep-tracking.c | 63 struct thread_map *threads = NULL; in test__keep_tracking() local 70 threads = thread_map__new(-1, getpid(), UINT_MAX); in test__keep_tracking() 71 CHECK_NOT_NULL__(threads); in test__keep_tracking() 79 perf_evlist__set_maps(evlist, cpus, threads); in test__keep_tracking() 148 thread_map__put(threads); in test__keep_tracking()
|
D | code-reading.c | 451 struct thread_map *threads = NULL; in do_test_code_reading() local 493 threads = thread_map__new_by_tid(pid); in do_test_code_reading() 494 if (!threads) { in do_test_code_reading() 499 ret = perf_event__synthesize_thread_map(NULL, threads, in do_test_code_reading() 527 perf_evlist__set_maps(evlist, cpus, threads); in do_test_code_reading() 595 thread_map__put(threads); in do_test_code_reading()
|
D | switch-tracking.c | 321 struct thread_map *threads = NULL; in test__switch_tracking() local 329 threads = thread_map__new(-1, getpid(), UINT_MAX); in test__switch_tracking() 330 if (!threads) { in test__switch_tracking() 347 perf_evlist__set_maps(evlist, cpus, threads); in test__switch_tracking() 564 thread_map__put(threads); in test__switch_tracking()
|
D | openat-syscall-tp-fields.c | 49 thread_map__set_pid(evlist->threads, 0, getpid()); in test__syscall_openat_tp_fields()
|
/linux-4.4.14/tools/perf/bench/ |
D | sched-pipe.c | 80 struct thread_data threads[2], *td; in bench_sched_pipe() local 103 td = threads + t; in bench_sched_pipe() 120 td = threads + t; in bench_sched_pipe() 127 td = threads + t; in bench_sched_pipe() 138 worker_thread(threads + 0); in bench_sched_pipe() 141 worker_thread(threads + 1); in bench_sched_pipe()
|
D | numa.c | 140 struct thread_data *threads; member 526 td = g->threads + t; in parse_setup_cpu_list() 648 td = g->threads + t; in parse_setup_node_list() 807 g->threads[task_nr].curr_cpu = cpu; in update_curr_cpu() 833 td = g->threads + task_nr; in count_process_nodes() 869 td = g->threads + task_nr; in count_node_processes() 936 struct thread_data *td = g->threads + t; in calc_convergence() 1163 this_cpu = g->threads[task_nr].curr_cpu; in worker_thread() 1241 td = g->threads + task_nr; in worker_process() 1256 td = g->threads + task_nr; in worker_process() [all …]
|
/linux-4.4.14/tools/perf/python/ |
D | twatch.py | 20 threads = perf.thread_map(thread) 35 evsel.open(cpus = cpus, threads = threads); 36 evlist = perf.evlist(cpus, threads)
|
/linux-4.4.14/tools/usb/ |
D | ffs-test.c | 309 } threads[] = { variable 351 if (t != threads) { in cleanup_thread() 618 init_thread(threads); in main() 619 ep0_init(threads, legacy_descriptors); in main() 621 for (i = 1; i < sizeof threads / sizeof *threads; ++i) in main() 622 init_thread(threads + i); in main() 624 for (i = 1; i < sizeof threads / sizeof *threads; ++i) in main() 625 start_thread(threads + i); in main() 627 start_thread_helper(threads); in main() 629 for (i = 1; i < sizeof threads / sizeof *threads; ++i) in main() [all …]
|
/linux-4.4.14/Documentation/power/ |
D | freezing-of-tasks.txt | 7 kernel threads are controlled during hibernation or system-wide suspend (on some 14 PF_NOFREEZE unset (all user space processes and some kernel threads) are 24 fake signal to all user space processes, and wakes up all the kernel threads. 32 frozen before kernel threads. 39 signal-handling code, but the freezable kernel threads need to call it 59 threads must call try_to_freeze() somewhere or use one of the 75 - freezes all tasks (including kernel threads) because we can't freeze 76 kernel threads without freezing userspace tasks 79 - thaws only kernel threads; this is particularly useful if we need to do 80 anything special in between thawing of kernel threads and thawing of [all …]
|
D | swsusp.txt | 165 kernel threads are controlled during hibernation or system-wide suspend (on some
|
/linux-4.4.14/tools/perf/arch/x86/tests/ |
D | perf-time-to-tsc.c | 50 struct thread_map *threads = NULL; in test__perf_time_to_tsc() local 62 threads = thread_map__new(-1, getpid(), UINT_MAX); in test__perf_time_to_tsc() 63 CHECK_NOT_NULL__(threads); in test__perf_time_to_tsc() 71 perf_evlist__set_maps(evlist, cpus, threads); in test__perf_time_to_tsc()
|
/linux-4.4.14/Documentation/filesystems/nfs/ |
D | nfsd-admin-interfaces.txt | 11 nfsd/threads. 26 nfsd is shut down by a write of 0 to nfsd/threads. All locks and state 29 Between startup and shutdown, the number of threads may be adjusted up 30 or down by additional writes to nfsd/threads or by writes to
|
D | knfsd-stats.txt | 43 which contains all the nfsd threads and all the CPUs in the system, 71 This can happen because there are too few nfsd threads in the thread 73 case configuring more nfsd threads will probably improve the 76 threads-woken 85 threads-timedout 91 threads configured than can be used by the NFS workload. This is 92 a clue that the number of nfsd threads can be reduced without 107 one of three ways. An nfsd thread can be woken (threads-woken counts 114 packets-deferred = packets-arrived - ( sockets-enqueued + threads-woken )
|
/linux-4.4.14/tools/testing/selftests/powerpc/dscr/ |
D | dscr_default_test.c | 65 pthread_t threads[THREADS]; in dscr_default() local 77 if (pthread_create(&threads[i], NULL, do_test, (void *)i)) { in dscr_default() 106 if (pthread_join(threads[i], (void **)&(status[i]))) { in dscr_default()
|
/linux-4.4.14/tools/virtio/virtio-trace/ |
D | README | 14 The read/write threads hold it. 17 the controller wake read/write threads. 18 5) The read/write threads start to read trace data from ring-buffers and 20 6) If the controller receives a stop order from a host, the read/write threads 31 trace-agent-ctl.c: includes controller function for read/write threads 32 trace-agent-rw.c: includes read/write threads function 103 read/write threads in the agent wait for start order from host. If you add -o
|
/linux-4.4.14/drivers/acpi/acpica/ |
D | dbexec.c | 528 if (info->threads && (info->num_created < info->num_threads)) { in acpi_db_method_thread() 529 info->threads[info->num_created++] = acpi_os_get_thread_id(); in acpi_db_method_thread() 676 acpi_gbl_db_method_info.threads = acpi_os_allocate(size); in acpi_db_create_execution_threads() 677 if (acpi_gbl_db_method_info.threads == NULL) { in acpi_db_create_execution_threads() 684 memset(acpi_gbl_db_method_info.threads, 0, size); in acpi_db_create_execution_threads() 762 acpi_os_free(acpi_gbl_db_method_info.threads); in acpi_db_create_execution_threads() 763 acpi_gbl_db_method_info.threads = NULL; in acpi_db_create_execution_threads()
|
D | aclocal.h | 1106 acpi_thread_id *threads; member
|
/linux-4.4.14/Documentation/ |
D | cputopology.txt | 25 internal kernel map of cpuX's hardware threads within the same 30 human-readable list of cpuX's hardware threads within the same 35 internal kernel map of cpuX's hardware threads within the same 40 human-readable list of cpuX's hardware threads within the same 45 internal kernel map of cpuX's hardware threads within the same 50 human-readable list of cpuX's hardware threads within the same
|
D | robust-futex-ABI.txt | 23 threads in the kernel. Options on the sys_futex(2) system call support 31 probably causing deadlock or other such failure of the other threads 40 The pointer 'head' points to a structure in the threads address space 80 waiting for a lock on a threads exit if that next thread used the futex 100 It is anticipated that threads will use robust_futexes embedded in 115 entirely by user level code in the contending threads, and by the 121 There may exist thousands of futex lock structures in a threads shared 127 at different times by any of the threads with access to that region. The 128 thread currently holding such a lock, if any, is marked with the threads 156 exiting threads TID, then the kernel will do two things: [all …]
|
D | unshare.txt | 27 Most legacy operating system kernels support an abstraction of threads 29 special resources and mechanisms to maintain these "threads". The Linux 31 between processes and "threads". The kernel allows processes to share 32 resources and thus they can achieve legacy "threads" behavior without 34 power of implementing threads in this manner comes not only from 37 threads. On Linux, at the time of thread creation using the clone system 39 between threads. 42 allows threads to selectively 'unshare' any resources that were being 45 of the discussion on POSIX threads on Linux. unshare augments the 46 usefulness of Linux threads for applications that would like to control [all …]
|
D | workqueue.txt | 85 Special purpose threads, called worker threads, execute the functions 87 worker threads become idle. These worker threads are managed in so 203 served by worker threads with elevated nice level. 354 Because the work functions are executed by generic worker threads 358 Worker threads show up in the process list as:
|
D | kernel-per-CPU-kthreads.txt | 52 1. Use irq affinity to force the irq threads to execute on 71 both kernel threads and interrupts to execute elsewhere. 121 forcing both kernel threads and interrupts to execute elsewhere. 127 kernel threads and interrupts to execute elsewhere. 145 calls and by forcing both kernel threads and interrupts 157 calls and by forcing both kernel threads and interrupts
|
D | pi-futex.txt | 46 combined with medium-prio construct-audio-data threads and low-prio 47 display-colory-stuff threads. Add video and decoding to the mix and
|
D | kref.txt | 98 This way, it doesn't matter what order the two threads handle the
|
D | robust-futexes.txt | 179 for new threads, without the need of another syscall.]
|
D | assoc_array.txt | 293 It is possible for other threads to iterate over or search the array under
|
D | atomic_ops.txt | 57 all threads are guaranteed to be correct reflecting either the value that has
|
D | kernel-parameters.txt | 3143 the priority of the RCU boost threads (rcub/N) 3226 Set callbacks/threads for rcu_barrier() testing.
|
D | memory-barriers.txt | 301 NOTE 1: Two threads of execution can update and access
|
/linux-4.4.14/arch/powerpc/include/asm/ |
D | cputhreads.h | 42 static inline cpumask_t cpu_thread_mask_to_cores(const struct cpumask *threads) in cpu_thread_mask_to_cores() argument 50 if (cpumask_intersects(threads, &tmp)) { in cpu_thread_mask_to_cores()
|
/linux-4.4.14/Documentation/filesystems/pohmelfs/ |
D | info.txt | 4 All but index, number of crypto threads and maximum IO size can changed via remount. 36 Number of crypto processing threads. Threads are used both for RX and TX traffic. 37 Default is 2, or no threads if crypto operations are not supported. 85 Number of working threads is set to 10. 98 Number of worker threads specifies how many workers will be created for each client.
|
D | design_notes.txt | 45 servers and async threads will pick up replies in parallel, find appropriate transactions in the 70 Crypto performance heavily depends on the number of crypto threads, which asynchronously perform
|
/linux-4.4.14/tools/perf/Documentation/ |
D | perf-lock.txt | 27 'perf lock info' shows metadata like threads or addresses 57 --threads::
|
D | perf-sched.txt | 27 via perf sched record. (this is done by starting up mockup threads 29 threads can then replay the timings (CPU runtime and sleep patterns)
|
D | perf-bench.txt | 99 (20 sender and receiver threads per group) 100 (20 groups == 800 threads run)
|
D | perf-stat.txt | 109 in per-cpu mode. The cgroup filesystem must be mounted. All threads belonging to 157 Aggregate counts per monitored threads, when monitoring threads (-t option)
|
D | examples.txt | 71 Delta compression using up to 2 threads. 125 Delta compression using up to 2 threads.
|
D | perf-trace.txt | 56 Record events in threads owned by uid. Name or number. 124 When processing pre-existing threads /proc/XXX/mmap, it may take a long time,
|
D | perf-top.txt | 84 Record events in threads owned by uid. Name or number. 206 When processing pre-existing threads /proc/XXX/mmap, it may take
|
D | perf-record.txt | 106 Record events in threads owned by uid. Name or number. 219 in per-cpu mode. The cgroup filesystem must be mounted. All threads belonging to 309 When processing pre-existing threads /proc/XXX/mmap, it may take a long time,
|
D | perf-kvm.txt | 155 When processing pre-existing threads /proc/XXX/mmap, it may take
|
D | perf-report.txt | 36 --threads:: 238 zooming into DSOs or threads, among other features. Use of --tui
|
D | intel-pt.txt | 534 In per-thread mode an exact list of threads is traced. There is no inheritance.
|
/linux-4.4.14/Documentation/thermal/ |
D | intel_powerclamp.txt | 42 idle injection across all online CPU threads was introduced. The goal 74 Injection is controlled by high priority kernel threads, spawned for 77 These kernel threads, with SCHED_FIFO class, are created to perform 82 migrated, unless the CPU is taken offline. In this case, threads 214 Per-CPU kernel threads are started/stopped upon receiving 216 keeps track of clamping kernel threads, even after they are migrated 233 case, little can be done from the idle injection threads. In most 257 counter summed over per CPU counting threads spawned for all running 278 will not show idle injection kernel threads. 281 idle time, powerclamp kernel threads will do idle injection, which
|
/linux-4.4.14/Documentation/locking/ |
D | locktorture.txt | 14 This torture test consists of creating a number of kernel threads which 28 nwriters_stress Number of kernel threads that will stress exclusive lock 32 nreaders_stress Number of kernel threads that will stress shared lock 100 shuffle_interval The number of seconds to keep the test threads affinitied 124 (D): Min and max number of times threads failed to acquire the lock.
|
/linux-4.4.14/samples/pktgen/ |
D | README.rst | 23 -t : ($THREADS) threads to start 40 * pg_thread() control the kernel threads and binding to devices
|
/linux-4.4.14/tools/perf/ |
D | builtin-stat.c | 173 return perf_evsel__open_per_thread(evsel, evsel_list->threads); in create_perf_stat_counter() 194 int nthreads = thread_map__nr(evsel_list->threads); in read_counter() 230 thread_map__nr(evsel_list->threads)); in read_counters() 253 nthreads = thread_map__nr(evsel_list->threads); in handle_initial_delay() 481 thread_map__comm(evsel->threads, id), in aggr_printout() 483 thread_map__pid(evsel->threads, id), in aggr_printout() 635 int nthreads = thread_map__nr(counter->threads); in print_aggr_thread() 1380 thread_map__read_comms(evsel_list->threads); in cmd_stat()
|
D | builtin-record.c | 285 if (perf_evsel__open(pos, pos->cpus, pos->threads) < 0) { in record__open() 612 err = __machine__synthesize_threads(machine, tool, &opts->target, rec->evlist->threads, in __cmd_record()
|
D | builtin-top.c | 902 top->evlist->threads) < 0) { in perf_top__start_counters() 971 top->evlist->threads, false, opts->proc_map_timeout); in __cmd_top()
|
D | builtin-trace.c | 1652 evlist->threads, trace__tool_process, false, in trace__symbols_init() 2597 else if (thread_map__pid(evlist->threads, 0) == -1) in trace__run() 2626 trace->multiple_threads = thread_map__pid(evlist->threads, 0) == -1 || in trace__run() 2627 evlist->threads->nr > 1 || in trace__run()
|
D | builtin-kvm.c | 1404 kvm->evlist->threads, false, kvm->opts.proc_map_timeout); in kvm_events_live()
|
/linux-4.4.14/tools/power/x86/turbostat/ |
D | turbostat.c | 198 struct thread_data threads; member 478 if (t == &average.threads) { in format_counters() 667 format_counters(&average.threads, &average.cores, in format_all_counters() 868 average.threads.tsc += t->tsc; in sum_counters() 869 average.threads.aperf += t->aperf; in sum_counters() 870 average.threads.mperf += t->mperf; in sum_counters() 871 average.threads.c1 += t->c1; in sum_counters() 873 average.threads.extra_delta32 += t->extra_delta32; in sum_counters() 874 average.threads.extra_delta64 += t->extra_delta64; in sum_counters() 926 clear_counters(&average.threads, &average.cores, &average.packages); in compute_average() [all …]
|
/linux-4.4.14/arch/arc/ |
D | Kconfig.debug | 11 threads to run on the system and also increase the pressure
|
/linux-4.4.14/net/rds/ |
D | Makefile | 3 recv.o send.o stats.o sysctl.o threads.o transport.o \
|
/linux-4.4.14/tools/perf/scripts/python/ |
D | sched-migration.py | 29 threads = { 0 : "idle"} variable 32 return "%s:%d" % (threads[pid], pid) 342 threads[prev_pid] = prev_comm 343 threads[next_pid] = next_comm
|
/linux-4.4.14/Documentation/scheduler/ |
D | completion.txt | 9 If you have one or more threads of execution that must wait for some process 19 efficient code as both threads can continue until the result is actually 24 struct completion that tells the waiting threads of execution if they 91 work threads remains in-scope, and no references remain to on-stack data 214 (decrementing) the done element of struct completion. Waiting threads
|
D | sched-arch.txt | 29 threads need only ever query need_resched, and may never set or
|
D | sched-bwc.txt | 18 above at each period boundary. As threads consume this bandwidth it is
|
D | sched-deadline.txt | 483 start multiple threads with specific parameters. rt-app supports 496 The above creates 2 threads. The first one, scheduled by SCHED_DEADLINE,
|
/linux-4.4.14/arch/arm/vfp/ |
D | entry.S | 23 @ r10 = this threads thread_info structure
|
/linux-4.4.14/kernel/ |
D | fork.c | 273 u64 threads; in set_max_threads() local 280 threads = MAX_THREADS; in set_max_threads() 282 threads = div64_u64((u64) totalram_pages * (u64) PAGE_SIZE, in set_max_threads() 285 if (threads > max_threads_suggested) in set_max_threads() 286 threads = max_threads_suggested; in set_max_threads() 288 max_threads = clamp_t(u64, threads, MIN_THREADS, MAX_THREADS); in set_max_threads() 2099 int threads = max_threads; in sysctl_max_threads() local 2104 t.data = &threads; in sysctl_max_threads() 2112 set_max_threads(threads); in sysctl_max_threads()
|
/linux-4.4.14/Documentation/powerpc/ |
D | cpu_families.txt | 15 | Old POWER | --------------> | RS64 (threads) |
|
/linux-4.4.14/drivers/dma/ |
D | dmatest.c | 165 struct list_head threads; member 178 list_for_each_entry(thread, &dtc->threads, node) { in is_threaded_test_run() 725 list_for_each_entry_safe(thread, _thread, &dtc->threads, node) { in dmatest_cleanup_channel() 780 list_add_tail(&thread->node, &dtc->threads); in dmatest_add_threads() 802 INIT_LIST_HEAD(&dtc->threads); in dmatest_add_channel()
|
/linux-4.4.14/arch/metag/ |
D | Kconfig.debug | 15 running more threads on a system and also reduces the pressure
|
D | Kconfig | 128 reordering of writes from different hardware threads when SMP is
|
/linux-4.4.14/tools/power/cpupower/utils/helpers/ |
D | helpers.h | 109 unsigned int threads; /* per core */ member
|
/linux-4.4.14/Documentation/vm/ |
D | active_mm.txt | 25 doesn't need any user mappings - all kernel threads basically fall into 26 this category, but even "real" threads can temporarily say that for
|
D | numa_memory_policy.txt | 51 [Linux kernel task] that installs the policy and any threads 52 subsequently created by that thread. Any sibling threads existing 84 space--a.k.a. threads--independent of when the policy is installed; and
|
D | cleancache.txt | 79 different Linux threads are simultaneously putting and invalidating a page
|
/linux-4.4.14/Documentation/ABI/testing/ |
D | sysfs-devices-system-cpu | 79 to other cores and threads in the same physical package. 90 core_siblings: internal kernel map of cpu#'s hardware threads 101 threads within the same core as cpu# 104 threads within the same core as cpu#
|
D | sysfs-devices-power | 60 be executed asynchronously (ie. in separate threads, in parallel
|
/linux-4.4.14/Documentation/networking/ |
D | pktgen.txt | 53 Kernel threads 74 To support adding the same device to multiple threads, which is useful 234 -t : ($THREADS) threads to start 260 to the running threads CPU (directly from smp_processor_id()).
|
/linux-4.4.14/Documentation/sysctl/ |
D | kernel.txt | 86 - threads-max 847 The soft lockup detector monitors CPUs for threads that are hogging the CPUs 848 without rescheduling voluntarily, and thus prevent the 'watchdog/N' threads 850 interrupts which are needed for the 'watchdog/N' threads to be woken up by 887 threads-max 889 This value controls the maximum number of threads that can be created 893 maximum number of threads is created, the thread structures occupy only 896 The minimum value that can be written to threads-max is 20. 897 The maximum value that can be written to threads-max is given by the 899 If a value outside of this range is written to threads-max an error [all …]
|
D | vm.txt | 123 flusher threads will start writeback. 136 flusher threads will start writing out dirty data. 161 for writeout by the kernel flusher threads. It is expressed in 100'ths 179 The kernel flusher threads will periodically wake up and write `old' data 569 Enables a system-wide task dump (excluding kernel threads) to be produced
|
/linux-4.4.14/arch/powerpc/boot/dts/ |
D | ps3.dts | 50 * threads is with an ibm,ppc-interrupt-server#s entry. We'll put one
|
/linux-4.4.14/Documentation/metag/ |
D | kernel-ABI.txt | 36 Extended context registers (EXT) may not be present on all hardware threads and 40 Global registers are shared between threads and are privilege protected. 100 context. A0.15 is global between hardware threads though which means it cannot
|
/linux-4.4.14/Documentation/accounting/ |
D | taskstats.txt | 43 The latter contains the sum of per-pid stats for all threads in the thread 129 When a user queries to get per-tgid data, the sum of all other live threads in 131 threads of the same thread group.
|
D | delay-accounting.txt | 24 delay statistics aggregated for all tasks (or threads) belonging to a
|
/linux-4.4.14/Documentation/RCU/ |
D | torture.txt | 70 nfakewriters This is the number of RCU fake writer threads to run. Fake 71 writer threads repeatedly use the synchronous "wait for 79 nreaders This is the number of RCU reading threads supported. 100 The number of seconds to keep the test threads affinitied
|
D | rcu.txt | 87 needed if you have CPU-bound realtime threads.
|
D | RTFP.txt | 17 destruction until all threads running at that time have terminated, again 19 with short-lived threads, such as the K42 research operating system. 42 In 1990, Pugh [Pugh90] noted that explicitly tracking which threads 44 in the presence of non-terminating threads. However, this explicit 59 which are otherwise required to synchronize the threads at the end 75 time that reading threads can hold references, as there might well be in
|
D | stallwarn.txt | 230 is running at a higher priority than the RCU softirq threads.
|
/linux-4.4.14/drivers/android/ |
D | binder.c | 297 struct rb_root threads; member 2522 struct rb_node **p = &proc->threads.rb_node; in binder_get_thread() 2545 rb_insert_color(&thread->rb_node, &proc->threads); in binder_get_thread() 2560 rb_erase(&thread->rb_node, &proc->threads); in binder_free_thread() 2993 for (n = rb_first(&proc->threads); n != NULL; n = rb_next(n)) { in binder_deferred_flush() 3067 int threads, nodes, incoming_refs, outgoing_refs, buffers, in binder_deferred_release() local 3082 threads = 0; in binder_deferred_release() 3084 while ((n = rb_first(&proc->threads))) { in binder_deferred_release() 3088 threads++; in binder_deferred_release() 3162 __func__, proc->pid, threads, nodes, incoming_refs, in binder_deferred_release() [all …]
|
/linux-4.4.14/Documentation/nvdimm/ |
D | btt.txt | 161 Consider a case where we have two threads, one doing reads and the other, 182 Consider a case where two writer threads are writing to the same LBA. There can 188 Both threads can update their respective free[lane] with the same old, freed
|
D | nvdimm.txt | 144 tagged-command-queuing, and would likely be used by different threads or
|
/linux-4.4.14/arch/sh/ |
D | Kconfig.debug | 37 running more threads on a system and also reduces the pressure
|
/linux-4.4.14/Documentation/device-mapper/ |
D | dm-crypt.txt | 74 encryption threads to a single thread degrades performance
|
/linux-4.4.14/Documentation/scsi/ |
D | scsi-parameters.txt | 106 discovered. async scans them in kernel threads,
|
D | ChangeLog.sym53c8xx_2 | 118 threads. A timer is also used to prevent from waiting indefinitely.
|
/linux-4.4.14/drivers/s390/net/ |
D | lcs.c | 321 lcs_set_allowed_threads(struct lcs_card *card, unsigned long threads) in lcs_set_allowed_threads() argument 326 card->thread_allowed_mask = threads; in lcs_set_allowed_threads() 331 lcs_threads_running(struct lcs_card *card, unsigned long threads) in lcs_threads_running() argument 337 rc = (card->thread_running_mask & threads); in lcs_threads_running() 343 lcs_wait_for_threads(struct lcs_card *card, unsigned long threads) in lcs_wait_for_threads() argument 346 lcs_threads_running(card, threads) == 0); in lcs_wait_for_threads()
|
D | qeth_core_main.c | 207 void qeth_set_allowed_threads(struct qeth_card *card, unsigned long threads, in qeth_set_allowed_threads() argument 213 card->thread_allowed_mask = threads; in qeth_set_allowed_threads() 215 card->thread_start_mask &= threads; in qeth_set_allowed_threads() 221 int qeth_threads_running(struct qeth_card *card, unsigned long threads) in qeth_threads_running() argument 227 rc = (card->thread_running_mask & threads); in qeth_threads_running() 233 int qeth_wait_for_threads(struct qeth_card *card, unsigned long threads) in qeth_wait_for_threads() argument 238 qeth_threads_running(card, threads) == 0); in qeth_wait_for_threads()
|
/linux-4.4.14/Documentation/dmaengine/ |
D | dmatest.txt | 36 Once started a message like "dmatest: Started 1 threads using dma0chan0" is
|
/linux-4.4.14/Documentation/devicetree/bindings/powerpc/fsl/cpm_qe/ |
D | qe.txt | 22 threads.
|
/linux-4.4.14/Documentation/filesystems/ |
D | files.txt | 12 this would be the case for posix threads. As with the common
|
D | btrfs.txt | 221 The number of worker threads to allocate. The default number is equal
|
D | gfs2-glocks.txt | 11 threads from making calls to the DLM, etc. at the same time. If a
|
D | proc.txt | 243 Threads number of threads 301 num_threads number of threads 1317 The "processes" line gives the number of processes and threads created, which 1321 The "procs_running" line gives the total number of threads that are 1322 running or ready to run (i.e., the total number of runnable threads).
|
D | ramfs-rootfs-initramfs.txt | 291 the above threads) is:
|
/linux-4.4.14/drivers/staging/lustre/ |
D | sysfs-fs-lustre | 422 Controls minimum number of ldlm callback threads to start. 428 Controls maximum number of ldlm callback threads to start. 434 Shows actual number of ldlm callback threads running. 440 Controls what percentage of ldlm callback threads is dedicated
|
/linux-4.4.14/Documentation/filesystems/caching/ |
D | object.txt | 130 The work to be done by the various states was given CPU time by the threads of 141 workqueues don't necessarily have the right numbers of threads.
|
D | backend-api.txt | 165 FS-Cache has a pool of threads that it uses to give CPU time to the 716 pool. One of the threads in the pool will invoke the retrieval record's
|
/linux-4.4.14/Documentation/kdump/ |
D | gdbmacros.txt | 176 Run info threads and lookup pid of thread #1
|
/linux-4.4.14/arch/powerpc/kvm/ |
D | Kconfig | 119 and in nap mode due to idle (cede) while other threads are still
|
/linux-4.4.14/fs/ |
D | binfmt_elf.c | 1835 struct elf_thread_core_info *threads = info->thread; in free_note_info() local 1836 while (threads) { in free_note_info() 1838 struct elf_thread_core_info *t = threads; in free_note_info() 1839 threads = t->next; in free_note_info()
|
/linux-4.4.14/Documentation/block/ |
D | cfq-iosched.txt | 284 Hence doing tree idling on threads using REQ_NOIDLE flag on requests 286 time we do not idle on individual threads.
|
/linux-4.4.14/init/ |
D | Kconfig | 655 int "Real-time priority to use for RCU worker threads" 663 assigned to the rcuc/n and rcub/n threads and is also the value 665 real-time application that has one or more CPU-bound threads 670 applications that do not have any CPU-bound threads. 674 multiple real-time threads that, taken together, fully utilize 1076 threads which belong to the cgroup specified and run on the 1610 barriers across all running threads, which can be used to distribute
|
/linux-4.4.14/ |
D | REPORTING-BUGS | 76 threads or separate bugzilla entries. If you report several unrelated
|
/linux-4.4.14/Documentation/devicetree/bindings/arm/ |
D | topology.txt | 23 threads existing in the system and map to the hierarchy level "thread" above.
|
/linux-4.4.14/Documentation/cgroups/ |
D | blkio-controller.txt | 22 You can do a very simple testing of running two dd threads in two different 46 launch two dd threads in different cgroup to read those files.
|
D | cgroups.txt | 239 Writing a thread group ID into this file moves all threads in that 442 threads in a threadgroup at once. Echoing the PID of any task in a
|
D | memory.txt | 57 tasks # attach a task(thread) and show list of threads
|
/linux-4.4.14/Documentation/frv/ |
D | features.txt | 287 Note that kernel threads do not have a userspace context, and so will not
|
/linux-4.4.14/Documentation/laptops/ |
D | laptop-mode.txt | 265 # exceeded, the kernel will wake flusher threads which will then reduce the 387 # exceeded, the kernel will wake flusher threads which will then reduce the
|
/linux-4.4.14/arch/m68k/ |
D | Kconfig.machine | 321 running more threads on a system and also reduces the pressure
|
/linux-4.4.14/Documentation/serial/ |
D | tty.txt | 31 counts the number of threads of execution within an ldisc method (plus those
|
/linux-4.4.14/scripts/ |
D | spelling.txt | 982 threds||threads
|
/linux-4.4.14/Documentation/hid/ |
D | hid-transport.txt | 85 its own worker threads.
|
/linux-4.4.14/Documentation/PCI/ |
D | pci-error-recovery.txt | 169 >>> recovery implementation (e.g. waiting for all notification threads
|
/linux-4.4.14/Documentation/filesystems/configfs/ |
D | configfs.txt | 263 the item in other threads, the memory is safe. It may take some time
|
/linux-4.4.14/kernel/trace/ |
D | Kconfig | 595 to each of the threads, where the IPI handler will also write
|
/linux-4.4.14/drivers/char/ |
D | Kconfig | 178 of threads across a large system which avoids bouncing a cacheline
|
/linux-4.4.14/Documentation/filesystems/cifs/ |
D | CHANGES | 242 on requests on other threads. Improve POSIX locking emulation, 361 kills the cifsd thread (NB: killing the cifs kernel threads is not
|
/linux-4.4.14/Documentation/virtual/kvm/ |
D | api.txt | 216 threads in one or more virtual CPU cores. (This is because the 217 hardware requires all the hardware threads in a CPU core to be in the 552 signal mask temporarily overrides the threads signal mask. Any
|
/linux-4.4.14/Documentation/virtual/uml/ |
D | UserModeLinux-HOWTO.txt | 123 14.3 Case 3 : Tracing thread panics caused by other threads 2474 slightly different because the kernel's threads are already being 4164 14.3. Case 3 : Tracing thread panics caused by other threads
|
/linux-4.4.14/crypto/ |
D | Kconfig | 173 algorithm that executes in kernel threads.
|
/linux-4.4.14/arch/arm/ |
D | Kconfig | 1777 between threads sharing the same address space if they invoke
|
/linux-4.4.14/Documentation/security/ |
D | keys.txt | 280 The keyrings associated with new threads are each labeled with the context of
|
/linux-4.4.14/drivers/scsi/aic7xxx/ |
D | aic79xx.reg | 3724 * The execution head pointer threads the head SCBs for
|
/linux-4.4.14/arch/mips/ |
D | Kconfig | 2191 bool "Dynamic FPU affinity for FP-intensive threads"
|
/linux-4.4.14/Documentation/s390/ |
D | Debugging390.txt | 60 the current running threads private area.
|