Searched refs:cpu_queue (Results 1 - 4 of 4) sorted by relevance

/linux-4.1.27/crypto/
H A Dmcryptd.c72 struct mcryptd_cpu_queue *cpu_queue; mcryptd_init_queue() local
74 queue->cpu_queue = alloc_percpu(struct mcryptd_cpu_queue); mcryptd_init_queue()
75 pr_debug("mqueue:%p mcryptd_cpu_queue %p\n", queue, queue->cpu_queue); mcryptd_init_queue()
76 if (!queue->cpu_queue) mcryptd_init_queue()
79 cpu_queue = per_cpu_ptr(queue->cpu_queue, cpu); for_each_possible_cpu()
80 pr_debug("cpu_queue #%d %p\n", cpu, queue->cpu_queue); for_each_possible_cpu()
81 crypto_init_queue(&cpu_queue->queue, max_cpu_qlen); for_each_possible_cpu()
82 INIT_WORK(&cpu_queue->work, mcryptd_queue_worker); for_each_possible_cpu()
90 struct mcryptd_cpu_queue *cpu_queue; mcryptd_fini_queue() local
93 cpu_queue = per_cpu_ptr(queue->cpu_queue, cpu); for_each_possible_cpu()
94 BUG_ON(cpu_queue->queue.qlen); for_each_possible_cpu()
96 free_percpu(queue->cpu_queue);
104 struct mcryptd_cpu_queue *cpu_queue; mcryptd_enqueue_request() local
107 cpu_queue = this_cpu_ptr(queue->cpu_queue); mcryptd_enqueue_request()
110 err = crypto_enqueue_request(&cpu_queue->queue, request); mcryptd_enqueue_request()
111 pr_debug("enqueue request: cpu %d cpu_queue %p request %p\n", mcryptd_enqueue_request()
112 cpu, cpu_queue, request); mcryptd_enqueue_request()
113 queue_work_on(cpu, kcrypto_wq, &cpu_queue->work); mcryptd_enqueue_request()
155 struct mcryptd_cpu_queue *cpu_queue; mcryptd_queue_worker() local
164 cpu_queue = container_of(work, struct mcryptd_cpu_queue, work); mcryptd_queue_worker()
173 backlog = crypto_get_backlog(&cpu_queue->queue); mcryptd_queue_worker()
174 req = crypto_dequeue_request(&cpu_queue->queue); mcryptd_queue_worker()
186 if (!cpu_queue->queue.qlen) mcryptd_queue_worker()
190 if (cpu_queue->queue.qlen) mcryptd_queue_worker()
191 queue_work(kcrypto_wq, &cpu_queue->work); mcryptd_queue_worker()
H A Dcryptd.c42 struct cryptd_cpu_queue __percpu *cpu_queue; member in struct:cryptd_queue
91 struct cryptd_cpu_queue *cpu_queue; cryptd_init_queue() local
93 queue->cpu_queue = alloc_percpu(struct cryptd_cpu_queue); cryptd_init_queue()
94 if (!queue->cpu_queue) cryptd_init_queue()
97 cpu_queue = per_cpu_ptr(queue->cpu_queue, cpu); for_each_possible_cpu()
98 crypto_init_queue(&cpu_queue->queue, max_cpu_qlen); for_each_possible_cpu()
99 INIT_WORK(&cpu_queue->work, cryptd_queue_worker); for_each_possible_cpu()
107 struct cryptd_cpu_queue *cpu_queue; cryptd_fini_queue() local
110 cpu_queue = per_cpu_ptr(queue->cpu_queue, cpu); for_each_possible_cpu()
111 BUG_ON(cpu_queue->queue.qlen); for_each_possible_cpu()
113 free_percpu(queue->cpu_queue);
120 struct cryptd_cpu_queue *cpu_queue; cryptd_enqueue_request() local
123 cpu_queue = this_cpu_ptr(queue->cpu_queue); cryptd_enqueue_request()
124 err = crypto_enqueue_request(&cpu_queue->queue, request); cryptd_enqueue_request()
125 queue_work_on(cpu, kcrypto_wq, &cpu_queue->work); cryptd_enqueue_request()
136 struct cryptd_cpu_queue *cpu_queue; cryptd_queue_worker() local
139 cpu_queue = container_of(work, struct cryptd_cpu_queue, work); cryptd_queue_worker()
148 backlog = crypto_get_backlog(&cpu_queue->queue); cryptd_queue_worker()
149 req = crypto_dequeue_request(&cpu_queue->queue); cryptd_queue_worker()
160 if (cpu_queue->queue.qlen) cryptd_queue_worker()
161 queue_work(kcrypto_wq, &cpu_queue->work); cryptd_queue_worker()
/linux-4.1.27/include/crypto/
H A Dmcryptd.h33 struct mcryptd_cpu_queue __percpu *cpu_queue; member in struct:mcryptd_queue
/linux-4.1.27/drivers/xen/events/
H A Devents_fifo.c65 static DEFINE_PER_CPU(struct evtchn_fifo_queue, cpu_queue);
105 struct evtchn_fifo_queue *q = &per_cpu(cpu_queue, cpu); init_control_block()
286 struct evtchn_fifo_queue *q = &per_cpu(cpu_queue, cpu); consume_one_event()

Completed in 286 milliseconds