Searched refs:cpu_queue (Results 1 - 4 of 4) sorted by relevance
/linux-4.1.27/crypto/ |
H A D | mcryptd.c | 72 struct mcryptd_cpu_queue *cpu_queue; mcryptd_init_queue() local 74 queue->cpu_queue = alloc_percpu(struct mcryptd_cpu_queue); mcryptd_init_queue() 75 pr_debug("mqueue:%p mcryptd_cpu_queue %p\n", queue, queue->cpu_queue); mcryptd_init_queue() 76 if (!queue->cpu_queue) mcryptd_init_queue() 79 cpu_queue = per_cpu_ptr(queue->cpu_queue, cpu); for_each_possible_cpu() 80 pr_debug("cpu_queue #%d %p\n", cpu, queue->cpu_queue); for_each_possible_cpu() 81 crypto_init_queue(&cpu_queue->queue, max_cpu_qlen); for_each_possible_cpu() 82 INIT_WORK(&cpu_queue->work, mcryptd_queue_worker); for_each_possible_cpu() 90 struct mcryptd_cpu_queue *cpu_queue; mcryptd_fini_queue() local 93 cpu_queue = per_cpu_ptr(queue->cpu_queue, cpu); for_each_possible_cpu() 94 BUG_ON(cpu_queue->queue.qlen); for_each_possible_cpu() 96 free_percpu(queue->cpu_queue); 104 struct mcryptd_cpu_queue *cpu_queue; mcryptd_enqueue_request() local 107 cpu_queue = this_cpu_ptr(queue->cpu_queue); mcryptd_enqueue_request() 110 err = crypto_enqueue_request(&cpu_queue->queue, request); mcryptd_enqueue_request() 111 pr_debug("enqueue request: cpu %d cpu_queue %p request %p\n", mcryptd_enqueue_request() 112 cpu, cpu_queue, request); mcryptd_enqueue_request() 113 queue_work_on(cpu, kcrypto_wq, &cpu_queue->work); mcryptd_enqueue_request() 155 struct mcryptd_cpu_queue *cpu_queue; mcryptd_queue_worker() local 164 cpu_queue = container_of(work, struct mcryptd_cpu_queue, work); mcryptd_queue_worker() 173 backlog = crypto_get_backlog(&cpu_queue->queue); mcryptd_queue_worker() 174 req = crypto_dequeue_request(&cpu_queue->queue); mcryptd_queue_worker() 186 if (!cpu_queue->queue.qlen) mcryptd_queue_worker() 190 if (cpu_queue->queue.qlen) mcryptd_queue_worker() 191 queue_work(kcrypto_wq, &cpu_queue->work); mcryptd_queue_worker()
|
H A D | cryptd.c | 42 struct cryptd_cpu_queue __percpu *cpu_queue; member in struct:cryptd_queue 91 struct cryptd_cpu_queue *cpu_queue; cryptd_init_queue() local 93 queue->cpu_queue = alloc_percpu(struct cryptd_cpu_queue); cryptd_init_queue() 94 if (!queue->cpu_queue) cryptd_init_queue() 97 cpu_queue = per_cpu_ptr(queue->cpu_queue, cpu); for_each_possible_cpu() 98 crypto_init_queue(&cpu_queue->queue, max_cpu_qlen); for_each_possible_cpu() 99 INIT_WORK(&cpu_queue->work, cryptd_queue_worker); for_each_possible_cpu() 107 struct cryptd_cpu_queue *cpu_queue; cryptd_fini_queue() local 110 cpu_queue = per_cpu_ptr(queue->cpu_queue, cpu); for_each_possible_cpu() 111 BUG_ON(cpu_queue->queue.qlen); for_each_possible_cpu() 113 free_percpu(queue->cpu_queue); 120 struct cryptd_cpu_queue *cpu_queue; cryptd_enqueue_request() local 123 cpu_queue = this_cpu_ptr(queue->cpu_queue); cryptd_enqueue_request() 124 err = crypto_enqueue_request(&cpu_queue->queue, request); cryptd_enqueue_request() 125 queue_work_on(cpu, kcrypto_wq, &cpu_queue->work); cryptd_enqueue_request() 136 struct cryptd_cpu_queue *cpu_queue; cryptd_queue_worker() local 139 cpu_queue = container_of(work, struct cryptd_cpu_queue, work); cryptd_queue_worker() 148 backlog = crypto_get_backlog(&cpu_queue->queue); cryptd_queue_worker() 149 req = crypto_dequeue_request(&cpu_queue->queue); cryptd_queue_worker() 160 if (cpu_queue->queue.qlen) cryptd_queue_worker() 161 queue_work(kcrypto_wq, &cpu_queue->work); cryptd_queue_worker()
|
/linux-4.1.27/include/crypto/ |
H A D | mcryptd.h | 33 struct mcryptd_cpu_queue __percpu *cpu_queue; member in struct:mcryptd_queue
|
/linux-4.1.27/drivers/xen/events/ |
H A D | events_fifo.c | 65 static DEFINE_PER_CPU(struct evtchn_fifo_queue, cpu_queue); 105 struct evtchn_fifo_queue *q = &per_cpu(cpu_queue, cpu); init_control_block() 286 struct evtchn_fifo_queue *q = &per_cpu(cpu_queue, cpu); consume_one_event()
|
Completed in 286 milliseconds